Gershgorin circle theorem
From Wikipedia, the free encyclopedia
In mathematics, the Gershgorin circle theorem may be used to bound the spectrum of a square matrix. It was first published by the Belarusian mathematician Semyon Aranovich Gershgorin in 1931. The spelling of S. A. Gershgorin's name has been transliterated in several different ways, including Geršgorin, Gerschgorin and Gershgorin.
Contents |
[edit] Statement and proof
Let A be a complex n×n matrix, with entries (aij). For i ∈ {1, … n} write Ri = ∑j ≠ i |aij| where |aij| denotes the complex norm of aij. Let D(aii, Ri) be the closed disc centered at aii with radius Ri. Such a disc is called a Gershgorin disc.
Theorem: Every eigenvalue of A lies within at least one of the Gershgorin discs D(aii, Ri).
Proof: Let λ be an eigenvalue of A and let x = (xj) be a corresponding eigenvector. Let i ∈ {1, … n} be chosen so that |xi| = maxj |xj|. Then |xi| > 0, otherwise x = 0. Since x is an eigenvector, Ax = λx or equivalent
so, splitting the sum, we get
We may then divide both sides by xi (choosing i as we explained we can be sure that ) and take the norm to obtain
where the last inequality is valid because
Corollary: The eigenvalues of A must also lie within the Gershgorin discs Cj corresponding to the columns of A.
Proof: Apply the Theorem to AT.
Example For a diagonal matrix, the Gershgorin discs coincide with the spectrum. Conversely, if the Gershgorin discs coincide with the spectrum, the matrix is diagonal.
[edit] Discussion
One way to interpret this theorem is that if the off-diagonal entries of a square matrix over the complex numbers have small norms, the eigenvalues of the matrix cannot be "far from" the diagonal entries of the matrix. Therefore, by reducing the norms of off-diagonal entries one can attempt to approximate the eigenvalues of the matrix. Of course, diagonal entries may change in the process of minimizing off-diagonal entries.
[edit] Application
The Gershgorin circle theorem is useful in solving matrix equations of the form Ax = b for x where b is a vector and A is a matrix with a large condition number.
In this kind of problem, the error in the final result is usually of the same order of magnitude as the error in the initial data multiplied by the condition number of A. For instance, if b is known to six decimal places and the condition number of A is 1000 then we can only be confident that x is accurate to three decimal places. For very high condition numbers, even very small errors due to rounding can be magnified to such an extent that the result is meaningless.
It would be good to reduce the condition number of A. This can be done by preconditioning: A matrix P such that P ≈ A−1 is constructed, and then the equation PAx = Pb is solved for x. Using the exact inverse of A would be nice but finding the inverse of a matrix is generally very difficult.
Now, since PA ≈ I where I is the identity matrix, the eigenvalues of PA should all be close to 1. By the Gerschgorin circle theorem, every eigenvalue of PA lies within a known area and so we can form a rough estimate of how good our choice of P was.
[edit] References
- Gerschgorin, S. "Über die Abgrenzung der Eigenwerte einer Matrix." Izv. Akad. Nauk. USSR Otd. Fiz.-Mat. Nauk 7, 749-754, 1931
- Varga, R. S. Geršgorin and His Circles. Berlin: Springer-Verlag, 2004. ISBN 3-540-21100-4. Errata.
[edit] External links
- Gershgorin's circle theorem on PlanetMath
- Eric W. Weisstein. "Gershgorin Circle Theorem." From MathWorld--A Wolfram Web Resource.
- Semyon Aranovich Gershgorin biography at MacTutor