Rayleigh quotient

From Wikipedia, the free encyclopedia

In mathematics, for a given complex Hermitian matrix M and nonzero vector x, the Rayleigh quotient[1] R(M,x), is defined as:[2][3]

R(M,x):={x^{{*}}Mx \over x^{{*}}x}.

For real matrices and vectors, the condition of being Hermitian reduces to that of being symmetric, and the conjugate transpose x^{{*}} to the usual transpose x'. Note that R(M,cx)=R(M,x) for any real scalar c\neq 0. Recall that a Hermitian (or real symmetric) matrix has real eigenvalues. It can be shown that, for a given matrix, the Rayleigh quotient reaches its minimum value \lambda _{\min } (the smallest eigenvalue of M) when x is v_{\min } (the corresponding eigenvector). Similarly, R(M,x)\leq \lambda _{\max } and R(M,v_{\max })=\lambda _{\max }. The Rayleigh quotient is used in the min-max theorem to get exact values of all eigenvalues. It is also used in eigenvalue algorithms to obtain an eigenvalue approximation from an eigenvector approximation. Specifically, this is the basis for Rayleigh quotient iteration.

The range of the Rayleigh quotient is called a numerical range.

Special case of covariance matrices

An empirical covariance matrix M can be represented as the product A' A of the data matrix A pre-multiplied by its transpose A'. Being a symmetrical real matrix, M has non-negative eigenvalues, and orthogonal (or othogonalisable) eigenvectors, which can be demonstrated as follows.

Firstly, that the eigenvalues \lambda _{i} are non-negative:

Mv_{i}=A'Av_{i}=\lambda _{i}v_{i}
\Rightarrow v_{i}'A'Av_{i}=v_{i}'\lambda _{i}v_{i}
\Rightarrow \left\|Av_{i}\right\|^{2}=\lambda _{i}\left\|v_{i}\right\|^{2}
\Rightarrow \lambda _{i}={\frac  {\left\|Av_{i}\right\|^{2}}{\left\|v_{i}\right\|^{2}}}\geq 0.

Secondly, that the eigenvectors v_{i} are orthogonal to one another:

Mv_{i}=\lambda _{i}v_{i}
\Rightarrow v_{j}'Mv_{i}=\lambda _{i}v_{j}'v_{i}
\Rightarrow (Mv_{j})'v_{i}=\lambda _{i}v_{j}'v_{i}
\Rightarrow \lambda _{j}v_{j}'v_{i}=\lambda _{i}v_{j}'v_{i}
\Rightarrow (\lambda _{j}-\lambda _{i})v_{j}'v_{i}=0
\Rightarrow v_{j}'v_{i}=0 (if the eigenvalues are different – in the case of multiplicity, the basis can be orthogonalized).

To now establish that the Rayleigh quotient is maximised by the eigenvector with the largest eigenvalue, consider decomposing an arbitrary vector x on the basis of the eigenvectors vi:

x=\sum _{{i=1}}^{n}\alpha _{i}v_{i}, where \alpha _{i}={\frac  {x'v_{i}}{v_{i}'v_{i}}}={\frac  {\langle x,v_{i}\rangle }{\left\|v_{i}\right\|^{2}}} is the coordinate of x orthogonally projected onto v_{i}

so

R(M,x)={\frac  {x'A'Ax}{x'x}}

can be written

R(M,x)={\frac  {(\sum _{{j=1}}^{n}\alpha _{j}v_{j})'A'A(\sum _{{i=1}}^{n}\alpha _{i}v_{i})}{(\sum _{{j=1}}^{n}\alpha _{j}v_{j})'(\sum _{{i=1}}^{n}\alpha _{i}v_{i})}}

which, by orthogonality of the eigenvectors, becomes:

R(M,x)={\frac  {\sum _{{i=1}}^{n}\alpha _{i}^{2}\lambda _{i}}{\sum _{{i=1}}^{n}\alpha _{i}^{2}}}=\sum _{{i=1}}^{n}\lambda _{i}{\frac  {(x'v_{i})^{2}}{(x'x)(v_{i}'v_{i})}}

The last representation establishes that the Rayleigh quotient is the sum of the squared cosines of the angles formed by the vector x and each eigenvector v_{i}, weighted by corresponding eigenvalues.

If a vector x maximizes R(M,x), then any scalar multiple kx (for k\neq 0) also maximizes R, so the problem can be reduced to the Lagrange problem of maximizing \sum _{{i=1}}^{n}\alpha _{i}^{2}\lambda _{i} under the constraint that \sum _{{i=1}}^{n}\alpha _{i}^{2}=1.

Let \beta _{i}{\overset  {{\text{def}}}=}\alpha _{i}^{2}. This then becomes a linear program, which always attains its maximum at one of the corners of the domain. A maximum point will have \alpha _{1}=\pm 1 and \forall i>1,\alpha _{i}=0 (when the eigenvalues are ordered by decreasing magnitude).

Thus, as advertised, the Rayleigh quotient is maximised by the eigenvector with the largest eigenvalue.

Formulation using Lagrange multipliers

Alternatively, this result can be arrived at by the method of Lagrange multipliers. The problem is to find the critical points of the function

R(M,x)=x^{T}Mx,

subject to the constraint \|x\|^{2}=x^{T}x=1. I.e. to find the critical points of

{\mathcal  {L}}(x)=x^{T}Mx-\lambda (x^{T}x-1),

where \lambda is a Lagrange multiplier. The stationary points of {\mathcal  {L}}(x) occur at

{\frac  {d{\mathcal  {L}}(x)}{dx}}=0
\therefore 2x^{T}M^{T}-2\lambda x^{T}=0
\therefore Mx=\lambda x

and R(M,x)={\frac  {x^{T}Mx}{x^{T}x}}=\lambda {\frac  {x^{T}x}{x^{T}x}}=\lambda .

Therefore, the eigenvectors x_{1}\ldots x_{n} of M are the critical points of the Rayleigh Quotient and their corresponding eigenvalues \lambda _{1}\ldots \lambda _{n} are the stationary values of R.

This property is the basis for principal components analysis and canonical correlation.

Use in SturmLiouville theory

Sturm–Liouville theory concerns the action of the linear operator

L(y)={\frac  {1}{w(x)}}\left(-{\frac  {d}{dx}}\left[p(x){\frac  {dy}{dx}}\right]+q(x)y\right)

on the inner product space defined by

\langle {y_{1},y_{2}}\rangle =\int _{a}^{b}w(x)y_{1}(x)y_{2}(x)\,dx

of functions satisfying some specified boundary conditions at a and b. In this case the Rayleigh quotient is

{\frac  {\langle {y,Ly}\rangle }{\langle {y,y}\rangle }}={\frac  {\int _{a}^{b}{y(x)\left(-{\frac  {d}{dx}}\left[p(x){\frac  {dy}{dx}}\right]+q(x)y(x)\right)}dx}{\int _{a}^{b}{w(x)y(x)^{2}}dx}}.

This is sometimes presented in an equivalent form, obtained by separating the integral in the numerator and using integration by parts:

{\frac  {\langle {y,Ly}\rangle }{\langle {y,y}\rangle }}={\frac  {\int _{a}^{b}{y(x)\left(-{\frac  {d}{dx}}\left[p(x)y'(x)\right]\right)}dx+\int _{a}^{b}{q(x)y(x)^{2}}\,dx}{\int _{a}^{b}{w(x)y(x)^{2}}\,dx}}
={\frac  {-y(x)\left[p(x)y'(x)\right]|_{a}^{b}+\int _{a}^{b}{y'(x)\left[p(x)y'(x)\right]}\,dx+\int _{a}^{b}{q(x)y(x)^{2}}\,dx}{\int _{a}^{b}{w(x)y(x)^{2}}\,dx}}
={\frac  {-p(x)y(x)y'(x)|_{a}^{b}+\int _{a}^{b}\left[p(x)y'(x)^{2}+q(x)y(x)^{2}\right]\,dx}{\int _{a}^{b}{w(x)y(x)^{2}}\,dx}}.

Generalization

For a given pair (A,B) of matrices, and a given non-zero vector x, the generalized Rayleigh quotient is defined as:

R(A,B;x):={\frac  {x^{*}Ax}{x^{*}Bx}}.

The Generalized Rayleigh Quotient can be reduced to the Rayleigh Quotient R(D,C^{*}x) through the transformation D=C^{{-1}}A{C^{*}}^{{-1}} where CC^{*} is the Cholesky decomposition of the Hermitian positive-definite matrix B.

See also

  • Field of values

References

  1. Also known as the Rayleigh–Ritz ratio; named after Walther Ritz and Lord Rayleigh.
  2. Horn, R. A. and C. A. Johnson. 1985. Matrix Analysis. Cambridge University Press. pp. 176180.
  3. Parlet B. N. The symmetric eigenvalue problem, SIAM, Classics in Applied Mathematics,1998

Further reading

This article is issued from Wikipedia. The text is available under the Creative Commons Attribution/Share Alike; additional terms may apply for the media files.