Rayleigh quotient
From Wikipedia, the free encyclopedia
In mathematics, for a given complex Hermitian matrix A and nonzero vector x, the Rayleigh quotient R(A,x) is defined as:
For real matrices and vectors, the condition of being Hermitian reduces to that of being symmetric, and the conjugate transpose x * to the usual transpose x'. Note that R(A,c.x) = R(A,x) for any real scalar c. Recall that a Hermitian (or real symmetric) matrix has real eigenvalues. It can be shown that the Rayleigh quotient reaches its minimum value (the smallest eigenvalue of A) when x is (the corresponding eigenvector). Similarly, and . The Rayleigh quotient is used in eigenvalue algorithms to obtain an eigenvalue approximation from an eigenvector approximation. Specifically, this is the basis for Rayleigh quotient iteration.
[edit] Special case of covariance matrices
A covariance matrix Σ can be represented as the product A'A. Its eigenvalues are positive:
- Σvi = λivi
- A'Avi = λivi
- vi'A'Avi = vi'λivi
The eigenvectors are orthogonal to one another:
- A'Avi = λivi
- vj'A'Avi = λivj'vi
- (A'Avj)'vi = λivj'vi
- λjvj'vi = λivj'vi
- (λj − λi)vj'vi = 0
- vj'vi = 0 (different eigenvalues, in case of multiplicity, the basis can be orthogonalized)
The Rayleigh quotient can be expressed as a function of the eigenvalues by decomposing any vector x on the basis of eigenvectors:
Which, by orthogonality of the eigenvectors, becomes:
If a vector x maximizes ρ, then any vector k.x (for ) also maximizes it, one can reduce to the Lagrange problem of maximizing under the constraint that .
Since all the eigenvalues are non-negative, the problem is convex and the maximum occurs on the edges of the domain, namely when α1 = 1 and (when the eigenvalues are ordered in decreasing magnitude).
This property is the basis for principal components analysis and canonical correlation.