Namespaces
Variants
Actions

Moment matrix

From Encyclopedia of Mathematics
Revision as of 17:13, 7 February 2011 by 127.0.0.1 (talk) (Importing text file)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

A matrix containing the moments of a probability distribution (cf. also Moment; Moments, method of (in probability theory)). For example, if is a probability distribution on a set , then is its th order moment. If and thus the moments are given, then a linear functional is defined on the set of polynomials by , . The inverse problem is called a moment problem (cf. also Moment problem): Given the sequence of moments , , find the necessary and sufficient conditions for the existence of and an expression for a positive distribution (a non-decreasing function with possibly infinitely many points of increase) that gives the integral representation of that linear functional. A positive distribution can only exist if for any polynomial that is positive on .

For the Hamburger moment problem (cf. also Complex moment problem, truncated), is the real axis and the polynomials are real, so the functional is positive if for any non-zero polynomial and this implies that the moment matrices, i.e., the Hankel matrices of the moment sequence, , are positive definite for all (cf. also Hankel matrix). This is a necessary and sufficient condition for the existence of a solution.

For the trigonometric moment problem, is the unit circle in the complex plane and the polynomials are complex, so that "positive definite" here means that for all non-zero polynomials . The linear functional is automatically defined on the space of Laurent polynomials (cf. also Laurent series) since . Positive definite now corresponds to the Toeplitz moment matrices being positive definite for all (cf. also Toeplitz matrix). Again this is the necessary and sufficient condition for the existence of a (unique) solution to the moment problem.

Once the positive-definite linear functional is given, one can define an inner product on the space of polynomials as in the real case or as in the complex case. The moment matrix is then the Gram matrix for the standard basis or .

Generalized moments correspond to the use of non-standard basis functions for the polynomials or for possibly other spaces. Consider a set of basis functions that span the space . The modified or generalized moments are then given by . The moment problem is to find a positive distribution function that gives an integral representation of the linear functional on . However, to define an inner product, one needs the functional to be defined on (in the real case) or on (in the complex case). This requires a doubly indexed sequence of "moments" . Finding a distribution for an integral representation of on is called a strong moment problem.

The solution of moment problems is often obtained using an orthogonal basis. If the are orthonormalized to give the functions , then the moment matrix can be used to give explicit expressions; namely where , and for , with

The leading coefficient in the expansion satisfies .

References

[a1] N.I. Akhiezer, "The classical moment problem" , Oliver & Boyd (1969) (In Russian)
[a2] J.A. Shohat, J.D. Tamarkin, "The problem of moments" , Math. Surveys , 1 , Amer. Math. Soc. (1943) (In Russian)
How to Cite This Entry:
Moment matrix. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Moment_matrix&oldid=15678
This article was adapted from an original article by A. Bultheel (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article