Resonant trans-Neptunian object: Difference between revisions
en>JorisvS |
en>Monkbot |
||
Line 1: | Line 1: | ||
In [[mathematics]], the '''Gershgorin circle theorem''' may be used to bound the [[Eigenvalues and eigenvectors|spectrum]] of a square [[matrix (mathematics)|matrix]]. It was first published by the Soviet mathematician [[Semyon Aranovich Gershgorin]] in 1931. The spelling of S. A. Gershgorin's name has been transliterated in several different ways, including Geršgorin, Gerschgorin, Gershgorin and Hershhorn/Hirschhorn. | |||
==Statement and proof== | |||
Let ''A'' be a [[complex number|complex]] ''n'' × ''n'' matrix, with entries <math>a_{ij}\,</math>. For ''i'' ∈ {1, …, ''n''} let <math>R_i = \sum_{j\neq{i}} \left|a_{ij}\right|</math> be the sum of the [[absolute value]]s of the non-diagonal entries in the ''i''th row. Let ''D''(''a''<sub>''ii''</sub>, ''R''<sub>''i''</sub>) be the closed [[disc (mathematics)|disc]] centered at ''a''<sub>''ii''</sub> with radius ''R''<sub>''i''</sub>. Such a disc is called a '''Gershgorin disc'''. | |||
'''Theorem''': Every [[eigenvalue]] of ''A'' lies within at least one of the Gershgorin discs | |||
<math> D(a_{ii},R_i) </math> | |||
''Proof'': Let λ be an eigenvalue of ''A'' and let '''x''' = (''x''<sub>''j''</sub>) be a corresponding eigenvector. Let ''i'' ∈ {1, …, ''n''} be chosen so that |''x''<sub>''i''</sub>| = max<sub>''j''</sub> |''x''<sub>''j''</sub>|. (That is to say, choose i so that x<sub>i</sub> is the largest (in absolute value) number in the vector '''x''') Then |''x''<sub>''i''</sub>| > 0, otherwise '''x''' = 0. Since '''x''' is an eigenvector, ''A'''''x''' = λ'''x''', and thus: | |||
: <math> \sum_j a_{ij} x_j = \lambda x_i \quad \forall i \in \{1, \ldots, n\}. </math> | |||
So, splitting the sum, we get | |||
: <math> \sum_{j \neq i} a_{ij} x_j = \lambda x_i - a_{ii} x_i. </math> | |||
We may then divide both sides by ''x''<sub>''i''</sub> (choosing ''i'' as we explained we can be sure that ''x''<sub>''i''</sub> ≠ 0) and take the absolute value to obtain | |||
: <math> |\lambda - a_{ii}| = \left|\frac{\sum_{j\ne i} a_{ij} x_j}{x_i}\right| \le \sum_{j\ne i} \left| \frac{a_{ij} x_j}{x_i} \right| \le \sum_{j\ne i} |a_{ij}| = R_i</math> | |||
where the last inequality is valid because | |||
: <math>\left| \frac{x_j}{x_i} \right| \leq 1 \quad \text{for }j \neq i. </math> | |||
'''Corollary''': The eigenvalues of ''A'' must also lie within the Gershgorin discs ''C''<sub>''j''</sub> corresponding to the columns of ''A''. | |||
''Proof'': Apply the Theorem to ''A''<sup>T</sup>. | |||
'''Example''' For a [[diagonal matrix]], the Gershgorin discs coincide with the spectrum. Conversely, if the Gershgorin discs coincide with the spectrum, the matrix is diagonal. | |||
==Discussion== | |||
One way to interpret this theorem is that if the off-diagonal entries of a square matrix over the complex numbers have small norms, the eigenvalues of the matrix cannot be "far from" the diagonal entries of the matrix. Therefore, by reducing the norms of off-diagonal entries one can attempt to approximate the eigenvalues of the matrix. Of course, diagonal entries may change in the process of minimizing off-diagonal entries. | |||
==Strengthening of the theorem== | |||
If one of the discs is disjoint from the others then it contains exactly one eigenvalue. If however it meets another disc it is possible that it contains no eigenvalue (for example, <math> A=\begin{pmatrix}0&1\\4&0\end{pmatrix} </math> or <math> A=\begin{pmatrix}1&-2\\1&-1\end{pmatrix} </math>). In the general case the theorem can be strengthened as follows: | |||
'''Theorem''': If the union of ''k'' discs is disjoint from the union of the other ''n'' − ''k'' discs then the former union contains exactly ''k'' and the latter ''n'' − ''k'' eigenvalues of ''A''. | |||
''Proof'': Let ''D'' be the diagonal matrix with entries equal to the diagonal entries of ''A'' and let | |||
: <math>B(t)=(1-t)D + tA.\,</math> | |||
We will use the fact that the eigenvalues are continuous in <math>t</math>, and show that if any eigenvalue moves from one of the unions to the other, then it must be outside all the discs for some <math>t</math>, which is a contradiction. | |||
The statement is true for <math>D = B(0)</math>. The diagonal entries of <math>B(t)</math> are equal to that of ''A'', thus the centers of the Gershgorin circles are the same, however their radii are ''t'' times that of A. Therefore the union of the corresponding ''k'' discs of <math>B(t)</math> is disjoint from the union of the remaining ''n-k'' for all ''t''. The discs are closed, so the distance of the two unions for ''A'' is <math>d>0</math>. The distance for <math>B(t)</math> is a decreasing function of ''t'', so it is always at least ''d''. Since the eigenvalues of <math>B(t)</math> are a continuous function of ''t'', for any eigenvalue <math>\lambda(t)</math> of <math>B(t)</math> in the union of the ''k'' discs its distance <math>d(t)</math> from the union of the other ''n-k'' discs is also continuous. Obviously <math>d(0)\ge d</math>, and assume <math>\lambda(1)</math> lies in the union of the ''n-k'' discs. Then <math>d(1)=0</math>, so there exists <math>0<t_0<1</math> such that <math>0<d(t_0)<d</math>. But this means <math>\lambda(t_0)</math> lies outside the Gershgorin discs, which is impossible. Therefore <math>\lambda(1)</math> lies in the union of the ''k'' discs, and the theorem is proven. | |||
==Application== | |||
The Gershgorin circle theorem is useful in solving matrix equations of the form ''Ax'' = ''b'' for ''x'' where ''b'' is a vector and ''A'' is a matrix with a large [[condition number]]. | |||
In this kind of problem, the error in the final result is usually of the same [[order of magnitude]] as the error in the initial data multiplied by the condition number of ''A''. For instance, if ''b'' is known to six decimal places and the condition number of ''A'' is 1000 then we can only be confident that ''x'' is accurate to three decimal places. For very high condition numbers, even very small errors due to rounding can be magnified to such an extent that the result is meaningless. | |||
It would be good to reduce the condition number of ''A''. This can be done by [[preconditioning]]: A matrix ''P'' such that ''P'' ≈ ''A''<sup>−1</sup> is constructed, and then the equation ''PAx'' = ''Pb'' is solved for ''x''. Using the ''exact'' [[matrix inverse|inverse]] of ''A'' would be nice but finding the inverse of a matrix is generally very difficult. | |||
Now, since ''PA'' ≈ ''I'' where ''I'' is the identity matrix, the [[eigenvalue]]s of ''PA'' should all be close to 1. By the Gershgorin circle theorem, every eigenvalue of ''PA'' lies within a known area and so we can form a rough estimate of how good our choice of ''P'' was. | |||
== Example == | |||
Use the Gershgorin circle theorem to estimate the eigenvalues of: | |||
[[File:Gershgorin_Disk_Theorem_Example.png|thumb|right|This diagram shows the discs in yellow derived for the eigenvalues. | |||
The first two disks overlap and their union contains two eigenvalues. The third and fourth disks are disjoint from the others and contain one eigenvalue each.]] | |||
:<math> A = | |||
\begin{bmatrix} 10 & -1 & 0 & 1\\ | |||
0.2 & 8 & 0.2 & 0.2\\ | |||
1 & 1 & 2 & 1\\ | |||
-1 & -1 & -1 & -11\\ | |||
\end{bmatrix}.</math> | |||
Starting with row one, we take the element on the diagonal, ''a''<sub>''ii''</sub> as the center for the disc. We then take the remaining elements in the row and apply the formula: | |||
: <math> \sum_{j\ne i} |a_{ij}| = R_i</math> | |||
to obtain the following four discs: | |||
: <math> D(10,2), \; D(8,0.6), \; D(2,3), \; \text{and} \; D(-11,3). </math> | |||
Note that we can improve the accuracy of the last two discs by applying the formula to the corresponding columns of the matrix, obtaining <math> D(2,1.2) </math> and <math> D(-11,2.2) </math>. | |||
The eigenvalues are 9.8218, 8.1478, 1.8995, -10.86 | |||
==See also== | |||
* For matrices with non-negative entries, see [[Perron–Frobenius theorem]]. | |||
* [[Metzler matrix]] | |||
* [[Doubly stochastic matrix]] | |||
* [[Muirhead's inequality]] | |||
* [[Hurwitz matrix]] | |||
==References== | |||
* Gerschgorin, S. "Über die Abgrenzung der Eigenwerte einer Matrix." Izv. Akad. Nauk. USSR Otd. Fiz.-Mat. Nauk 6, 749–754, 1931 [http://mi.mathnet.ru/eng/izv/y1931/i6/p749] | |||
* [[Richard S. Varga|Varga, R. S.]] ''Geršgorin and His Circles.'' Berlin: Springer-Verlag, 2004. ISBN 3-540-21100-4. [http://www.math.kent.edu/~varga/pub/corrections.pdf Errata]. | |||
* [[Richard S. Varga]] 2002 ''Matrix Iterative Analysis'', Second ed. (of 1962 Prentice Hall edition), Springer-Verlag. | |||
* {{cite book |author=[[Gene H. Golub|Golub, G. H.]]; [[Charles F. Van Loan|Van Loan, C. F.]]|title=Matrix Computations |publisher=Johns Hopkins University Press |location=Baltimore |year=1996 |pages=320 |isbn=0-8018-5413-X |oclc= |doi= |accessdate=}} | |||
==External links== | |||
*{{planetmath reference|id=3709|title=Gershgorin's circle theorem}} | |||
* Eric W. Weisstein. "[http://mathworld.wolfram.com/GershgorinCircleTheorem.html Gershgorin Circle Theorem]." From MathWorld—A Wolfram Web Resource. | |||
* Semyon Aranovich Gershgorin biography at [http://www-history.mcs.st-andrews.ac.uk/Mathematicians/Gershgorin.html MacTutor] | |||
[[Category:Theorems in algebra]] | |||
[[Category:Linear algebra]] | |||
[[Category:Matrix theory]] |
Revision as of 09:10, 29 January 2014
In mathematics, the Gershgorin circle theorem may be used to bound the spectrum of a square matrix. It was first published by the Soviet mathematician Semyon Aranovich Gershgorin in 1931. The spelling of S. A. Gershgorin's name has been transliterated in several different ways, including Geršgorin, Gerschgorin, Gershgorin and Hershhorn/Hirschhorn.
Statement and proof
Let A be a complex n × n matrix, with entries . For i ∈ {1, …, n} let be the sum of the absolute values of the non-diagonal entries in the ith row. Let D(aii, Ri) be the closed disc centered at aii with radius Ri. Such a disc is called a Gershgorin disc.
Theorem: Every eigenvalue of A lies within at least one of the Gershgorin discs
Proof: Let λ be an eigenvalue of A and let x = (xj) be a corresponding eigenvector. Let i ∈ {1, …, n} be chosen so that |xi| = maxj |xj|. (That is to say, choose i so that xi is the largest (in absolute value) number in the vector x) Then |xi| > 0, otherwise x = 0. Since x is an eigenvector, Ax = λx, and thus:
So, splitting the sum, we get
We may then divide both sides by xi (choosing i as we explained we can be sure that xi ≠ 0) and take the absolute value to obtain
where the last inequality is valid because
Corollary: The eigenvalues of A must also lie within the Gershgorin discs Cj corresponding to the columns of A.
Proof: Apply the Theorem to AT.
Example For a diagonal matrix, the Gershgorin discs coincide with the spectrum. Conversely, if the Gershgorin discs coincide with the spectrum, the matrix is diagonal.
Discussion
One way to interpret this theorem is that if the off-diagonal entries of a square matrix over the complex numbers have small norms, the eigenvalues of the matrix cannot be "far from" the diagonal entries of the matrix. Therefore, by reducing the norms of off-diagonal entries one can attempt to approximate the eigenvalues of the matrix. Of course, diagonal entries may change in the process of minimizing off-diagonal entries.
Strengthening of the theorem
If one of the discs is disjoint from the others then it contains exactly one eigenvalue. If however it meets another disc it is possible that it contains no eigenvalue (for example, or ). In the general case the theorem can be strengthened as follows:
Theorem: If the union of k discs is disjoint from the union of the other n − k discs then the former union contains exactly k and the latter n − k eigenvalues of A.
Proof: Let D be the diagonal matrix with entries equal to the diagonal entries of A and let
We will use the fact that the eigenvalues are continuous in , and show that if any eigenvalue moves from one of the unions to the other, then it must be outside all the discs for some , which is a contradiction.
The statement is true for . The diagonal entries of are equal to that of A, thus the centers of the Gershgorin circles are the same, however their radii are t times that of A. Therefore the union of the corresponding k discs of is disjoint from the union of the remaining n-k for all t. The discs are closed, so the distance of the two unions for A is . The distance for is a decreasing function of t, so it is always at least d. Since the eigenvalues of are a continuous function of t, for any eigenvalue of in the union of the k discs its distance from the union of the other n-k discs is also continuous. Obviously , and assume lies in the union of the n-k discs. Then , so there exists such that . But this means lies outside the Gershgorin discs, which is impossible. Therefore lies in the union of the k discs, and the theorem is proven.
Application
The Gershgorin circle theorem is useful in solving matrix equations of the form Ax = b for x where b is a vector and A is a matrix with a large condition number.
In this kind of problem, the error in the final result is usually of the same order of magnitude as the error in the initial data multiplied by the condition number of A. For instance, if b is known to six decimal places and the condition number of A is 1000 then we can only be confident that x is accurate to three decimal places. For very high condition numbers, even very small errors due to rounding can be magnified to such an extent that the result is meaningless.
It would be good to reduce the condition number of A. This can be done by preconditioning: A matrix P such that P ≈ A−1 is constructed, and then the equation PAx = Pb is solved for x. Using the exact inverse of A would be nice but finding the inverse of a matrix is generally very difficult.
Now, since PA ≈ I where I is the identity matrix, the eigenvalues of PA should all be close to 1. By the Gershgorin circle theorem, every eigenvalue of PA lies within a known area and so we can form a rough estimate of how good our choice of P was.
Example
Use the Gershgorin circle theorem to estimate the eigenvalues of:
Starting with row one, we take the element on the diagonal, aii as the center for the disc. We then take the remaining elements in the row and apply the formula:
to obtain the following four discs:
Note that we can improve the accuracy of the last two discs by applying the formula to the corresponding columns of the matrix, obtaining and .
The eigenvalues are 9.8218, 8.1478, 1.8995, -10.86
See also
- For matrices with non-negative entries, see Perron–Frobenius theorem.
- Metzler matrix
- Doubly stochastic matrix
- Muirhead's inequality
- Hurwitz matrix
References
- Gerschgorin, S. "Über die Abgrenzung der Eigenwerte einer Matrix." Izv. Akad. Nauk. USSR Otd. Fiz.-Mat. Nauk 6, 749–754, 1931 [1]
- Varga, R. S. Geršgorin and His Circles. Berlin: Springer-Verlag, 2004. ISBN 3-540-21100-4. Errata.
- Richard S. Varga 2002 Matrix Iterative Analysis, Second ed. (of 1962 Prentice Hall edition), Springer-Verlag.
- 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.
My blog: http://www.primaboinca.com/view_profile.php?userid=5889534
External links
- Template:Planetmath reference
- Eric W. Weisstein. "Gershgorin Circle Theorem." From MathWorld—A Wolfram Web Resource.
- Semyon Aranovich Gershgorin biography at MacTutor