Namespaces
Variants
Actions

Difference between revisions of "Ridge function"

From Encyclopedia of Mathematics
Jump to: navigation, search
(Importing text file)
 
m (AUTOMATIC EDIT (latexlist): Replaced 22 formulas out of 22 by TEX code with an average confidence of 2.0 and a minimal confidence of 2.0.)
 
Line 1: Line 1:
 +
<!--This article has been texified automatically. Since there was no Nroff source code for this article,
 +
the semi-automatic procedure described at https://encyclopediaofmath.org/wiki/User:Maximilian_Janisch/latexlist
 +
was used.
 +
If the TeX and formula formatting is correct, please remove this message and the {{TEX|semi-auto}} category.
 +
 +
Out of 22 formulas, 22 were replaced by TEX code.-->
 +
 +
{{TEX|semi-auto}}{{TEX|done}}
 
''plane wave''
 
''plane wave''
  
 
In its simplest form, a ridge function is a multivariate function
 
In its simplest form, a ridge function is a multivariate function
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300901.png" /></td> </tr></table>
+
\begin{equation*} f : \mathbf{R} ^ { n } \rightarrow \mathbf{R} \end{equation*}
  
 
of the form
 
of the form
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300902.png" /></td> </tr></table>
+
\begin{equation*} f ( x _ { 1 } , \dots , x _ { n } ) = g ( a _ { 1 } x _ { 1 } + \ldots + a _ { n } x _ { n } ) = g ( \mathbf{a}\cdot \mathbf{x} ), \end{equation*}
  
where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300903.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300904.png" />. The vector <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300905.png" /> is generally called the direction. In other words, a ridge function is a multivariate function constant on the parallel hyperplanes <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300906.png" />, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300907.png" />.
+
where $g : \mathbf{R} \rightarrow \mathbf{R}$ and ${\bf a} = ( a _ { 1 } , \dots , a _ { n } ) \in {\bf R} ^ { n } \backslash \{ 0 \}$. The vector $\mathbf{a} \in \mathbf{R} ^ { n } \backslash \{ 0 \}$ is generally called the direction. In other words, a ridge function is a multivariate function constant on the parallel hyperplanes $\mathbf{a} \cdot \mathbf{x} = c$, $c \in \mathbf R$.
  
 
Ridge functions appear in various areas and under various guises. In 1975, B.F. Logan and L.A. Shepp coined the name  "ridge function"  in their seminal paper [[#References|[a6]]] in computerized tomography. In [[Tomography|tomography]], or at least in tomography as the theory was initially constructed in the early 1980s, ridge functions were basic. However, these functions have been considered for some time, but under the name of plane waves. See, for example, [[#References|[a5]]] and [[#References|[a1]]]. In general, linear combinations of ridge functions with fixed directions occur in the study of hyperbolic partial differential equations with constant coefficients.
 
Ridge functions appear in various areas and under various guises. In 1975, B.F. Logan and L.A. Shepp coined the name  "ridge function"  in their seminal paper [[#References|[a6]]] in computerized tomography. In [[Tomography|tomography]], or at least in tomography as the theory was initially constructed in the early 1980s, ridge functions were basic. However, these functions have been considered for some time, but under the name of plane waves. See, for example, [[#References|[a5]]] and [[#References|[a1]]]. In general, linear combinations of ridge functions with fixed directions occur in the study of hyperbolic partial differential equations with constant coefficients.
  
Ridge functions and ridge function approximation are studied in statistics. There they often go under the name of projection pursuit, see e.g. [[#References|[a3]]], [[#References|[a4]]], [[#References|[a2]]]. Projection pursuit algorithms approximate a function of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300908.png" /> variables by functions of the form
+
Ridge functions and ridge function approximation are studied in statistics. There they often go under the name of projection pursuit, see e.g. [[#References|[a3]]], [[#References|[a4]]], [[#References|[a2]]]. Projection pursuit algorithms approximate a function of $n$ variables by functions of the form
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r1300909.png" /></td> </tr></table>
+
\begin{equation*} \sum _ { i = 1 } ^ { r } g_i ( \mathbf{a} ^ { i }. \mathbf{x} ), \end{equation*}
  
where the <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009010.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009011.png" /> are the variables. The idea here is to  "reduce dimension"  and thus bypass the [[Curse of dimension|curse of dimension]]. The vector <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009012.png" /> is considered as a projection of <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009013.png" />. The directions <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009014.png" /> are chosen to  "pick out the salient features" .
+
where the $\mathbf{a} ^ { i }$ and $g_i$ are the variables. The idea here is to  "reduce dimension"  and thus bypass the [[Curse of dimension|curse of dimension]]. The vector $\mathbf{a} ^ { i } \mathbf{x}$ is considered as a projection of $\mathbf{x}$. The directions $\mathbf{a}$ are chosen to  "pick out the salient features" .
  
One of the popular models in the theory of neural nets is that of a multi-layer feedforward neural net with input, hidden and output layers (cf. also [[Neural network|Neural network]]). The simplest case (which is that of one hidden layer, <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009015.png" /> processing units and one output) considers, in mathematical terms, functions of the form
+
One of the popular models in the theory of neural nets is that of a multi-layer feedforward neural net with input, hidden and output layers (cf. also [[Neural network|Neural network]]). The simplest case (which is that of one hidden layer, $r$ processing units and one output) considers, in mathematical terms, functions of the form
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009016.png" /></td> </tr></table>
+
\begin{equation*} \sum _ { i = 1 } ^ { r } \alpha _ { i } \sigma ( \mathbf{w} ^ { i } \mathbf{x} + \theta _ { i } ) \end{equation*}
  
where <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009017.png" /> is some given fixed univariate function. In this model, which is just one of many, one is in general permitted to vary over the <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009018.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009019.png" />, in order to approximate an unknown function. Note that for each <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009020.png" /> and <img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009021.png" /> the function
+
where $\sigma : \mathbf{R} \rightarrow \mathbf{R}$ is some given fixed univariate function. In this model, which is just one of many, one is in general permitted to vary over the $\mathbf{w} ^ { i }$ and $\theta _ { i }$, in order to approximate an unknown function. Note that for each $\theta \in \mathbf{R}$ and $\mathbf{w} \in \mathbf{R} ^ { n } \backslash \{ 0 \}$ the function
  
<table class="eq" style="width:100%;"> <tr><td valign="top" style="width:94%;text-align:center;"><img align="absmiddle" border="0" src="https://www.encyclopediaofmath.org/legacyimages/r/r130/r130090/r13009022.png" /></td> </tr></table>
+
\begin{equation*} \sigma ( \mathbf{w}.\mathbf{v}  + \theta ) \end{equation*}
  
 
is also a ridge function, see e.g. [[#References|[a8]]] and references therein.
 
is also a ridge function, see e.g. [[#References|[a8]]] and references therein.
Line 32: Line 40:
  
 
====References====
 
====References====
<table><TR><TD valign="top">[a1]</TD> <TD valign="top">  R. Courant,  D. Hilbert,  "Methods of mathematical physics" , '''II''' , Interscience  (1962)</TD></TR><TR><TD valign="top">[a2]</TD> <TD valign="top">  D.L. Donoho,  I.M. Johnstone,  "Projection-based approximation and a duality method with kernel methods"  ''Ann. Statist.'' , '''17'''  (1989)  pp. 58–106</TD></TR><TR><TD valign="top">[a3]</TD> <TD valign="top">  J.H. Friedman,  W. Stuetzle,  "Projection pursuit regression"  ''J. Amer. Statist. Assoc.'' , '''76'''  (1981)  pp. 817–823</TD></TR><TR><TD valign="top">[a4]</TD> <TD valign="top">  P.J. Huber,  "Projection pursuit"  ''Ann. Statist.'' , '''13'''  (1985)  pp. 435–475</TD></TR><TR><TD valign="top">[a5]</TD> <TD valign="top">  F. John,  "Plane waves and spherical means applied to partial differential equations" , Interscience  (1955)</TD></TR><TR><TD valign="top">[a6]</TD> <TD valign="top">  B.F. Logan,  L.A. Shepp,  "Optimal reconstruction of a function from its projections"  ''Duke Math. J.'' , '''42'''  (1975)  pp. 645–659</TD></TR><TR><TD valign="top">[a7]</TD> <TD valign="top">  A. Pinkus,  "Approximating by ridge functions"  A. Le Méhauté (ed.)  C. Rabut (ed.)  L.L. Schumaker (ed.) , ''Surface Fitting and Multiresolution Methods'' , Vanderbilt Univ. Press  (1997)  pp. 279–292</TD></TR><TR><TD valign="top">[a8]</TD> <TD valign="top">  A. Pinkus,  "Approximation theory of the MLP model in neural networks"  ''Acta Numerica'' , '''8'''  (1999)  pp. 143–195</TD></TR></table>
+
<table><tr><td valign="top">[a1]</td> <td valign="top">  R. Courant,  D. Hilbert,  "Methods of mathematical physics" , '''II''' , Interscience  (1962)</td></tr><tr><td valign="top">[a2]</td> <td valign="top">  D.L. Donoho,  I.M. Johnstone,  "Projection-based approximation and a duality method with kernel methods"  ''Ann. Statist.'' , '''17'''  (1989)  pp. 58–106</td></tr><tr><td valign="top">[a3]</td> <td valign="top">  J.H. Friedman,  W. Stuetzle,  "Projection pursuit regression"  ''J. Amer. Statist. Assoc.'' , '''76'''  (1981)  pp. 817–823</td></tr><tr><td valign="top">[a4]</td> <td valign="top">  P.J. Huber,  "Projection pursuit"  ''Ann. Statist.'' , '''13'''  (1985)  pp. 435–475</td></tr><tr><td valign="top">[a5]</td> <td valign="top">  F. John,  "Plane waves and spherical means applied to partial differential equations" , Interscience  (1955)</td></tr><tr><td valign="top">[a6]</td> <td valign="top">  B.F. Logan,  L.A. Shepp,  "Optimal reconstruction of a function from its projections"  ''Duke Math. J.'' , '''42'''  (1975)  pp. 645–659</td></tr><tr><td valign="top">[a7]</td> <td valign="top">  A. Pinkus,  "Approximating by ridge functions"  A. Le Méhauté (ed.)  C. Rabut (ed.)  L.L. Schumaker (ed.) , ''Surface Fitting and Multiresolution Methods'' , Vanderbilt Univ. Press  (1997)  pp. 279–292</td></tr><tr><td valign="top">[a8]</td> <td valign="top">  A. Pinkus,  "Approximation theory of the MLP model in neural networks"  ''Acta Numerica'' , '''8'''  (1999)  pp. 143–195</td></tr></table>

Latest revision as of 16:45, 1 July 2020

plane wave

In its simplest form, a ridge function is a multivariate function

\begin{equation*} f : \mathbf{R} ^ { n } \rightarrow \mathbf{R} \end{equation*}

of the form

\begin{equation*} f ( x _ { 1 } , \dots , x _ { n } ) = g ( a _ { 1 } x _ { 1 } + \ldots + a _ { n } x _ { n } ) = g ( \mathbf{a}\cdot \mathbf{x} ), \end{equation*}

where $g : \mathbf{R} \rightarrow \mathbf{R}$ and ${\bf a} = ( a _ { 1 } , \dots , a _ { n } ) \in {\bf R} ^ { n } \backslash \{ 0 \}$. The vector $\mathbf{a} \in \mathbf{R} ^ { n } \backslash \{ 0 \}$ is generally called the direction. In other words, a ridge function is a multivariate function constant on the parallel hyperplanes $\mathbf{a} \cdot \mathbf{x} = c$, $c \in \mathbf R$.

Ridge functions appear in various areas and under various guises. In 1975, B.F. Logan and L.A. Shepp coined the name "ridge function" in their seminal paper [a6] in computerized tomography. In tomography, or at least in tomography as the theory was initially constructed in the early 1980s, ridge functions were basic. However, these functions have been considered for some time, but under the name of plane waves. See, for example, [a5] and [a1]. In general, linear combinations of ridge functions with fixed directions occur in the study of hyperbolic partial differential equations with constant coefficients.

Ridge functions and ridge function approximation are studied in statistics. There they often go under the name of projection pursuit, see e.g. [a3], [a4], [a2]. Projection pursuit algorithms approximate a function of $n$ variables by functions of the form

\begin{equation*} \sum _ { i = 1 } ^ { r } g_i ( \mathbf{a} ^ { i }. \mathbf{x} ), \end{equation*}

where the $\mathbf{a} ^ { i }$ and $g_i$ are the variables. The idea here is to "reduce dimension" and thus bypass the curse of dimension. The vector $\mathbf{a} ^ { i } \mathbf{x}$ is considered as a projection of $\mathbf{x}$. The directions $\mathbf{a}$ are chosen to "pick out the salient features" .

One of the popular models in the theory of neural nets is that of a multi-layer feedforward neural net with input, hidden and output layers (cf. also Neural network). The simplest case (which is that of one hidden layer, $r$ processing units and one output) considers, in mathematical terms, functions of the form

\begin{equation*} \sum _ { i = 1 } ^ { r } \alpha _ { i } \sigma ( \mathbf{w} ^ { i } \mathbf{x} + \theta _ { i } ) \end{equation*}

where $\sigma : \mathbf{R} \rightarrow \mathbf{R}$ is some given fixed univariate function. In this model, which is just one of many, one is in general permitted to vary over the $\mathbf{w} ^ { i }$ and $\theta _ { i }$, in order to approximate an unknown function. Note that for each $\theta \in \mathbf{R}$ and $\mathbf{w} \in \mathbf{R} ^ { n } \backslash \{ 0 \}$ the function

\begin{equation*} \sigma ( \mathbf{w}.\mathbf{v} + \theta ) \end{equation*}

is also a ridge function, see e.g. [a8] and references therein.

For a survey on some approximation-theoretic questions concerning ridge functions, see [a7] and references therein.

References

[a1] R. Courant, D. Hilbert, "Methods of mathematical physics" , II , Interscience (1962)
[a2] D.L. Donoho, I.M. Johnstone, "Projection-based approximation and a duality method with kernel methods" Ann. Statist. , 17 (1989) pp. 58–106
[a3] J.H. Friedman, W. Stuetzle, "Projection pursuit regression" J. Amer. Statist. Assoc. , 76 (1981) pp. 817–823
[a4] P.J. Huber, "Projection pursuit" Ann. Statist. , 13 (1985) pp. 435–475
[a5] F. John, "Plane waves and spherical means applied to partial differential equations" , Interscience (1955)
[a6] B.F. Logan, L.A. Shepp, "Optimal reconstruction of a function from its projections" Duke Math. J. , 42 (1975) pp. 645–659
[a7] A. Pinkus, "Approximating by ridge functions" A. Le Méhauté (ed.) C. Rabut (ed.) L.L. Schumaker (ed.) , Surface Fitting and Multiresolution Methods , Vanderbilt Univ. Press (1997) pp. 279–292
[a8] A. Pinkus, "Approximation theory of the MLP model in neural networks" Acta Numerica , 8 (1999) pp. 143–195
How to Cite This Entry:
Ridge function. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Ridge_function&oldid=17815
This article was adapted from an original article by Allan Pinkus (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article