0% found this document useful (0 votes)
229 views3 pages

Hankel Matrix

A Hankel matrix is a square matrix where each ascending skew-diagonal from left to right is constant. Hankel matrices are symmetric and can be used to approximate the action of Hankel operators on Hilbert spaces. Applications of Hankel matrices include realization of state-space models from signal data and decomposition of non-stationary signals.

Uploaded by

besty666
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
229 views3 pages

Hankel Matrix

A Hankel matrix is a square matrix where each ascending skew-diagonal from left to right is constant. Hankel matrices are symmetric and can be used to approximate the action of Hankel operators on Hilbert spaces. Applications of Hankel matrices include realization of state-space models from signal data and decomposition of non-stationary signals.

Uploaded by

besty666
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

Hankel matrix

In linear algebra, a Hankel matrix (or catalecticant matrix), named after Hermann Hankel, is a square
matrix in which each ascending skew-diagonal from left to right is constant, e.g.:

More generally, a Hankel matrix is any matrix of the form

In terms of the components, if the element of is denoted with , and assuming , then we have
for all

Properties
The Hankel matrix is a symmetric matrix.
Let be the exchange matrix. If is a Hankel matrix, then where
is a Toeplitz matrix.

If is real symmetric, then will have the same eigenvalues as up to sign.[1]


The Hilbert matrix is an example of a Hankel matrix.

Hankel operator
A Hankel operator on a Hilbert space is one whose matrix is a (possibly infinite) Hankel matrix with
respect to an orthonormal basis. As indicated above, a Hankel Matrix is a matrix with constant values along
its antidiagonals, which means that a Hankel matrix must satisfy, for all rows and columns ,
. Note that every entry depends only on .
Let the corresponding Hankel Operator be . Given a Hankel matrix , the corresponding Hankel
operator is then defined as .

We are often interested in Hankel operators over the Hilbert


space , the space of square integrable bilateral complex sequences. For any , we have

We are often interested in approximations of the Hankel operators, possibly by low-order operators. In
order to approximate the output of the operator, we can use the spectral norm (operator 2-norm) to measure
the error of our approximation. This suggests singular value decomposition as a possible technique to
approximate the action of the operator.

Note that the matrix does not have to be finite. If it is infinite, traditional methods of computing
individual singular vectors will not work directly. We also require that the approximation is a Hankel
matrix, which can be shown with AAK theory.

The determinant of a Hankel matrix is called a catalecticant.

Hankel matrix transform


The Hankel matrix transform, or simply Hankel transform, produces the sequence of the determinants
of the Hankel matrices formed from the given sequence. Namely, the sequence is the Hankel
transform of the sequence when

The Hankel transform is invariant under the binomial transform of a sequence. That is, if one writes

as the binomial transform of the sequence , then one has

Applications of Hankel matrices


Hankel matrices are formed when, given a sequence of output data, a realization of an underlying state-
space or hidden Markov model is desired.[2] The singular value decomposition of the Hankel matrix
provides a means of computing the A, B, and C matrices which define the state-space realization.[3] The
Hankel matrix formed from the signal has been found useful for decomposition of non-stationary signals
and time-frequency representation.

Method of moments for polynomial distributions

The method of moments applied to polynomial distributions results in a Hankel matrix that needs to be
inverted in order to obtain the weight parameters of the polynomial distribution approximation.[4]

Positive Hankel matrices and the Hamburger moment problems

See also
Toeplitz matrix, an "upside down" (i.e., row-reversed) Hankel matrix
Cauchy matrix
Vandermonde matrix

Notes
1. Yasuda, M. (2003). "A Spectral Characterization of Hermitian Centrosymmetric and
Hermitian Skew-Centrosymmetric K-Matrices". SIAM J. Matrix Anal. Appl. 25 (3): 601–605.
doi:10.1137/S0895479802418835 (https://doi.org/10.1137%2FS0895479802418835).
2. Aoki, Masanao (1983). "Prediction of Time Series" (https://books.google.com/books?id=l_Ls
CAAAQBAJ&pg=PA38). Notes on Economic Time Series Analysis : System Theoretic
Perspectives. New York: Springer. pp. 38–47. ISBN 0-387-12696-1.
3. Aoki, Masanao (1983). "Rank determination of Hankel matrices" (https://books.google.com/b
ooks?id=l_LsCAAAQBAJ&pg=PA67). Notes on Economic Time Series Analysis : System
Theoretic Perspectives. New York: Springer. pp. 67–68. ISBN 0-387-12696-1.
4. J. Munkhammar, L. Mattsson, J. Rydén (2017) "Polynomial probability distribution estimation
using the method of moments". PLoS ONE 12(4): e0174573.
https://doi.org/10.1371/journal.pone.0174573

References
Brent R.P. (1999), "Stability of fast algorithms for structured linear systems", Fast Reliable
Algorithms for Matrices with Structure (editors—T. Kailath, A.H. Sayed), ch.4 (SIAM).
Victor Y. Pan (2001). Structured matrices and polynomials: unified superfast algorithms.
Birkhäuser. ISBN 0817642404.
J.R. Partington (1988). An introduction to Hankel operators. LMS Student Texts. Vol. 13.
Cambridge University Press. ISBN 0-521-36791-3.

Retrieved from "https://en.wikipedia.org/w/index.php?title=Hankel_matrix&oldid=1127582553"

You might also like