Jordan matrix: Difference between revisions
Fix broken link |
m →Linear algebra: changed links to Jordan-Chevalley decomposition to point instead to Jordan normal form |
||
(34 intermediate revisions by 19 users not shown) | |||
Line 1: | Line 1: | ||
{{Short description|Block diagonal matrix of Jordan blocks}} |
|||
In the [[Mathematics|mathematical]] discipline of [[Matrix (mathematics)|matrix theory]], a '''Jordan |
In the [[Mathematics|mathematical]] discipline of [[Matrix (mathematics)|matrix theory]], a '''Jordan matrix''', named after [[Camille Jordan]], is a [[Block matrix|block diagonal matrix]] over a [[Ring (mathematics)|ring]] {{mvar|R}} (whose [[Identity element|identities]] are the [[0 (number)|zero]] 0 and [[1 (number)|one]] 1), where each block along the diagonal, called a Jordan block, has the following form: |
||
<math display="block">\begin{bmatrix} |
|||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | |||
==Definition== |
|||
:<math>\begin{pmatrix} |
|||
Every '''Jordan block''' is specified by its dimension ''n'' and its [[eigenvalue]] <math>\lambda\in R</math>, and is denoted as {{math|''J''<sub>λ,''n''</sub>}}. It is an <math>n\times n</math> matrix of zeroes everywhere except for the diagonal, which is filled with <math>\lambda</math> and for the [[superdiagonal]], which is composed of ones. |
|||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | |||
Any block diagonal matrix whose blocks are Jordan blocks is called a '''Jordan matrix'''. This {{math|(''n''<sub>1</sub> + ⋯ + ''n<sub>r</sub>'') × (''n''<sub>1</sub> + ⋯ + ''n<sub>r</sub>'')}} square matrix, consisting of {{mvar|r}} diagonal blocks, can be compactly indicated as <math>J_{\lambda_1,n_1}\oplus \cdots \oplus J_{\lambda_r,n_r}</math> or <math>\mathrm{diag}\left(J_{\lambda_1,n_1}, \ldots, J_{\lambda_r,n_r}\right)</math>, where the ''i''-th Jordan block is {{math|''J''<sub>λ<sub>i</sub>,''n''<sub>i</sub></sub>}}. |
|||
Every Jordan block is thus specified by its dimension ''n'' and its [[eigenvalue]] <math>\lambda</math> and is indicated as <math>J_{\lambda,n}</math>. |
|||
Any [[Block matrix|block diagonal matrix]] whose blocks are Jordan blocks is called a '''Jordan matrix'''; using either the <math>\oplus</math> or the “<math>\mathrm{diag}</math>” symbol, the <math>(l+m+n)\times (l+m+n)</math> block diagonal square matrix whose first diagonal block is <math>J_{\alpha,l}</math>, whose second diagonal block is <math>J_{\beta,m}</math> and whose third diagonal block is <math>J_{\gamma,n}</math> is compactly indicated as <math>J_{\alpha,l}\oplus J_{\beta,m}\oplus J_{\gamma,n}</math> or <math>\mathrm{diag}\left(J_{\alpha,l}, J_{\beta,m}, J_{\gamma,n}\right)</math>, respectively. |
|||
For example the matrix |
For example, the matrix |
||
<math display="block"> |
|||
J=\left |
J=\left[\begin{array}{ccc|cc|cc|ccc} |
||
0 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ |
0 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ |
||
0 & 0 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ |
0 & 0 & 1 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ |
||
Line 22: | Line 26: | ||
0 & 0 & 0 & 0 & 0 & 0 & 0 & 7 & 1 & 0 \\ |
0 & 0 & 0 & 0 & 0 & 0 & 0 & 7 & 1 & 0 \\ |
||
0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 7 & 1 \\ |
0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 7 & 1 \\ |
||
0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 7 \end{array}\right |
0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 7 \end{array}\right]</math> |
||
is a |
is a {{math|10 × 10}} Jordan matrix with a {{math|3 × 3}} block with [[eigenvalue]] {{math|0}}, two {{math|2 × 2}} blocks with eigenvalue the [[imaginary unit]] {{mvar|i}}, and a {{math|3 × 3}} block with eigenvalue 7. Its Jordan-block structure is written as either <math>J_{0,3}\oplus J_{i,2}\oplus J_{i,2}\oplus J_{7,3}</math> or {{math|diag(''J''<sub>0,3</sub>, ''J''<sub>''i'',2</sub>, ''J''<sub>''i'',2</sub>, ''J''<sub>7,3</sub>)}}. |
||
== Linear algebra == |
== Linear algebra == |
||
Any |
Any {{math|''n'' × ''n''}} square matrix {{mvar|A}} whose elements are in an [[algebraically closed field]] {{mvar|K}} is [[matrix similarity|similar]] to a Jordan matrix {{mvar|J}}, also in <math>\mathbb{M}_n (K)</math>, which is unique up to a permutation of its diagonal blocks themselves. {{mvar|J}} is called the [[Jordan normal form]] of {{mvar|A}} and corresponds to a generalization of the diagonalization procedure.<ref>{{harvtxt|Beauregard|Fraleigh|1973|pp=310–316}}</ref><ref>{{harvtxt|Golub|Van Loan|1996|p=317}}</ref><ref>{{harvtxt|Nering|1970|pp=118–127}}</ref> A [[diagonalizable matrix]] is similar, in fact, to a special case of Jordan matrix: the matrix whose blocks are all {{mvar|1 × 1}}.<ref>{{harvtxt|Beauregard|Fraleigh|1973|pp=270–274}}</ref><ref>{{harvtxt|Golub|Van Loan|1996|p=316}}</ref><ref>{{harvtxt|Nering|1970|pp=113–118}}</ref> |
||
More generally, given a Jordan matrix <math>J=J_{\lambda_1,m_1}\oplus J_{\lambda_2,m_2} \oplus\ |
More generally, given a Jordan matrix <math>J=J_{\lambda_1,m_1}\oplus J_{\lambda_2,m_2} \oplus\cdots\oplus J_{\lambda_N,m_N}</math>, that is, whose {{mvar|k}}th diagonal block, <math>1 \leq k \leq N</math>, is the Jordan block {{math|''J''<sub>λ<sub>''k''</sub>,''m<sub>k</sub>''</sub>}} and whose diagonal elements <math>\lambda_k</math> may not all be distinct, the [[geometric multiplicity]] of <math>\lambda\in K</math> for the matrix {{mvar|J}}, indicated as <math>\operatorname{gmul}_J \lambda</math>, corresponds to the number of Jordan blocks whose eigenvalue is {{math|λ}}. Whereas the '''index''' of an eigenvalue <math>\lambda</math> for {{mvar|J}}, indicated as <math>\operatorname{idx}_J \lambda</math>, is defined as the dimension of the largest Jordan block associated to that eigenvalue. |
||
The same goes for all the matrices |
The same goes for all the matrices {{mvar|A}} similar to {{mvar|J}}, so <math>\operatorname{idx}_A \lambda</math> can be defined accordingly with respect to the [[Jordan normal form]] of {{mvar|A}} for any of its eigenvalues <math>\lambda \in \operatorname{spec}A</math>. In this case one can check that the index of <math>\lambda</math> for {{mvar|A}} is equal to its multiplicity as a [[root]] of the [[minimal polynomial (linear algebra)|minimal polynomial]] of {{mvar|A}} (whereas, by definition, its [[algebraic multiplicity]] for {{mvar|A}}, <math>\operatorname{mul}_A \lambda</math>, is its multiplicity as a root of the [[characteristic polynomial]] of {{mvar|A}}; that is, <math>\det(A-xI)\in K[x]</math>). An equivalent necessary and sufficient condition for {{mvar|A}} to be diagonalizable in {{mvar|K}} is that all of its eigenvalues have index equal to {{math|1}}; that is, its minimal polynomial has only simple roots. |
||
An equivalent necessary and sufficient condition for <math>A</math> to be diagonalizable in <math>K</math> is that all of its eigenvalues have index equal to <math>1</math>, i.e. its minimal polynomial has only simple roots. |
|||
Note that knowing a matrix's spectrum with all of its algebraic/geometric multiplicities and indexes does not always allow for the computation of its [[Jordan normal form]] (this may be a sufficient condition only for spectrally simple, usually low-dimensional matrices) |
Note that knowing a matrix's spectrum with all of its algebraic/geometric multiplicities and indexes does not always allow for the computation of its [[Jordan normal form]] (this may be a sufficient condition only for spectrally simple, usually low-dimensional matrices). Indeed, determining the Jordan normal form is generally a computationally challenging task. From the [[vector space]] point of view, the Jordan normal form is equivalent to finding an orthogonal decomposition (that is, via [[direct sum of vector spaces|direct sums]] of eigenspaces represented by Jordan blocks) of the domain which the associated [[generalized eigenvector]]s make a basis for. |
||
From the [[vector space]] point of view, the [[Jordan–Chevalley decomposition|Jordan decomposition]] is equivalent to finding an orthogonal decomposition (i.e. via [[direct sum of vector spaces|direct sums]] of eigenspaces represented by Jordan blocks) of the domain which the associated [[generalized eigenvector]]s make a basis for. |
|||
== Functions of matrices == |
== Functions of matrices == |
||
Let <math>A\in\mathbb{M}_n (\ |
Let <math>A\in\mathbb{M}_n (\Complex)</math> (that is, a {{math|''n'' × ''n''}} complex matrix) and <math>C\in\mathrm{GL}_n (\Complex)</math> be the [[change of basis]] matrix to the [[Jordan normal form]] of {{mvar|A}}; that is, {{math|1=''A'' = ''C''<sup>−1</sup>''JC''}}. Now let {{math|''f''{{hair space}}(''z'')}} be a [[holomorphic function]] on an open set <math>\Omega</math> such that <math>\mathrm{spec}A \subset \Omega \subseteq \Complex</math>; that is, the spectrum of the matrix is contained inside the [[domain of holomorphy]] of {{mvar|f}}. Let |
||
⚫ | |||
Now let <math>f(z)</math> be a [[holomorphic function]] on an open set <math>\mathit{\Omega}</math> such that <math>\mathrm{spec}A\subset \mathit{\Omega}\subseteq\mathbb{C}</math>, i.e. the spectrum of the matrix is contained inside the [[domain of holomorphy]] of <math>f</math>. Let |
|||
⚫ | |||
⚫ | |||
⚫ | and is [[absolutely convergent]] with respect to the [[Euclidean norm]] of <math>\mathbb{M}_n (\Complex)</math>. To put it another way, {{math|''f''{{hair space}}(''A'')}} converges absolutely for every square matrix whose [[spectral radius]] is less than the [[radius of convergence]] of {{mvar|f}} around {{math|0}} and is [[uniformly convergent]] on any compact subsets of <math>\mathbb{M}_n (\Complex)</math> satisfying this property in the [[matrix Lie group]] topology. |
||
⚫ | The [[Jordan normal form]] allows the computation of functions of matrices without explicitly computing an [[infinite series]], which is one of the main achievements of Jordan matrices. Using the facts that the {{mvar|k}}th power (<math>k\in\N_0</math>) of a diagonal [[block matrix]] is the diagonal block matrix whose blocks are the {{mvar|k}}th powers of the respective blocks; that is, {{nowrap|<math>\left(A_1 \oplus A_2 \oplus A_3 \oplus\cdots\right)^k=A^k_1 \oplus A_2^k \oplus A_3^k \oplus\cdots</math>,}} and that {{math|1=''A<sup>k</sup>'' = ''C''<sup>−1</sup>''J<sup>k</sup>C''}}, the above matrix power series becomes |
||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | where the last series need not be computed explicitly via power series of every Jordan block. In fact, if <math>\lambda\in\Omega</math>, any [[holomorphic function]] of a Jordan block <math>f(J_{\lambda,n}) = f(\lambda I+Z)</math> has a finite power series around <math>\lambda I</math> because <math>Z^n=0</math>. Here, <math>Z</math> is the nilpotent part of <math>J</math> and <math>Z^k</math> has all 0's except 1's along the <math>k^{\text{th}}</math> superdiagonal. Thus it is the following upper [[triangular matrix]]: |
||
⚫ | |||
<math display="block">f(J_{\lambda,n})= \sum_{k=0}^{n-1} \frac{f^{(k)}(\lambda) Z^k}{k!} = |
|||
\begin{bmatrix} |
|||
⚫ | is [[absolutely convergent]] with respect to the [[Euclidean norm]] of <math>\mathbb{M}_n (\ |
||
⚫ | The [[Jordan normal form]] allows the computation of functions of matrices without explicitly computing an [[infinite series]], which is one of the main achievements of Jordan matrices. Using the facts that the |
||
⚫ | |||
⚫ | |||
:<math>f(J_{\lambda,n})=\left(\begin{matrix} |
|||
f(\lambda) & f^\prime (\lambda) & \frac{f^{\prime\prime}(\lambda)}{2} & \cdots & \frac{f^{(n-2)}(\lambda)}{(n-2)!} & \frac{f^{(n-1)}(\lambda)}{(n-1)!} \\ |
f(\lambda) & f^\prime (\lambda) & \frac{f^{\prime\prime}(\lambda)}{2} & \cdots & \frac{f^{(n-2)}(\lambda)}{(n-2)!} & \frac{f^{(n-1)}(\lambda)}{(n-1)!} \\ |
||
0 & f(\lambda) & f^\prime (\lambda) & \cdots & \frac{f^{(n-3)}(\lambda)}{(n-3)!} & \frac{f^{(n-2)}(\lambda)}{(n-2)!} \\ |
0 & f(\lambda) & f^\prime (\lambda) & \cdots & \frac{f^{(n-3)}(\lambda)}{(n-3)!} & \frac{f^{(n-2)}(\lambda)}{(n-2)!} \\ |
||
0 & 0 & f(\lambda) & \cdots & \frac{f^{(n-4)}(\lambda)}{(n-4)!} & \frac{f^{(n-3)}(\lambda)}{(n-3)!} \\ |
0 & 0 & f(\lambda) & \cdots & \frac{f^{(n-4)}(\lambda)}{(n-4)!} & \frac{f^{(n-3)}(\lambda)}{(n-3)!} \\ |
||
\vdots & \vdots & \vdots & \ddots & \vdots & \vdots \\ |
\vdots & \vdots & \vdots & \ddots & \vdots & \vdots \\ |
||
0 & 0 & 0 & \cdots & f(\lambda) & f^\prime (\lambda) \\ |
0 & 0 & 0 & \cdots & f(\lambda) & f^\prime (\lambda) \\ |
||
0 & 0 & 0 & \cdots & 0 & f(\lambda) \\ |
0 & 0 & 0 & \cdots & 0 & f(\lambda) \\ |
||
\end{bmatrix}.</math> |
|||
\end{matrix}\right)=\left(\begin{matrix} |
|||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | |||
0 & 0 & 0 & \cdots & a_0 & a_1 \\ |
|||
0 & 0 & 0 & \cdots & 0 & a_0 \\ |
|||
⚫ | |||
As a consequence of this, the computation of any |
As a consequence of this, the computation of any function of a matrix is straightforward whenever its Jordan normal form and its change-of-basis matrix are known. For example, using <math>f(z)=1/z</math>, the inverse of <math>J_{\lambda,n}</math> is: |
||
<math display="block">J_{\lambda,n}^{-1} = \sum_{k=0}^{n-1}\frac{(-Z)^k}{\lambda^{k+1}} = |
|||
⚫ | Also, |
||
\begin{bmatrix} |
|||
\lambda^{-1} & -\lambda^{-2} & \,\,\,\lambda^{-3} & \cdots & -(-\lambda)^{1-n} & \,-(-\lambda)^{-n} \\ |
|||
0 & \;\;\;\lambda^{-1} & -\lambda^{-2} & \cdots & -(-\lambda)^{2-n} & -(-\lambda)^{1-n} \\ |
|||
⚫ | |||
⚫ | |||
⚫ | |||
⚫ | |||
\end{bmatrix}.</math> |
|||
⚫ | Also, {{math|1=spec{{hair space}}''f''(''A'') = ''f''{{hair space}}(spec{{hair space}}''A'')}}; that is, every eigenvalue <math>\lambda\in\mathrm{spec}A</math> corresponds to the eigenvalue <math>f(\lambda) \in \operatorname{spec}f(A)</math>, but it has, in general, different [[algebraic multiplicity]], geometric multiplicity and index. However, the algebraic multiplicity may be computed as follows: |
||
<math display="block">\text{mul}_{f(A)}f(\lambda)=\sum_{\mu\in\text{spec}A\cap f^{-1}(f(\lambda))}~\text{mul}_A \mu.</math> |
|||
The function |
The function {{math|''f''{{hair space}}(''T'')}} of a [[linear transformation]] {{mvar|T}} between vector spaces can be defined in a similar way according to the [[holomorphic functional calculus]], where [[Banach space]] and [[Riemann surface]] theories play a fundamental role. In the case of finite-dimensional spaces, both theories perfectly match. |
||
== Dynamical systems == |
== Dynamical systems == |
||
Now suppose a (complex) [[dynamical system]] is simply defined by the equation |
Now suppose a (complex) [[dynamical system]] is simply defined by the equation |
||
<math display="block">\begin{align} |
|||
⚫ | |||
\dot{\mathbf{z}}(t)&=A(\mathbf{c})\mathbf{z}(t), \\ |
|||
\mathbf{z}(0) &=\mathbf{z}_0 \in\Complex^n, |
|||
⚫ | where <math>\mathbf{z}:\ |
||
\end{align}</math> |
|||
⚫ | Even if <math>A\in\mathbb{M}_n \left(\mathrm{C}^0(\ |
||
Many aspects of [[bifurcation theory]] for both continuous and discrete dynamical systems can be interpreted with the analysis of functional Jordan matrices. |
|||
⚫ | where <math>\mathbf{z}:\R_+ \to \mathcal{R}</math> is the ({{mvar|n}}-dimensional) curve parametrization of an orbit on the [[Riemann surface]] <math>\mathcal{R}</math> of the dynamical system, whereas {{math|''A''('''c''')}} is an {{math|''n'' × ''n''}} complex matrix whose elements are complex functions of a {{mvar|d}}-dimensional parameter <math>\mathbf{c} \in \Complex^d</math>. |
||
⚫ | Even if <math>A\in\mathbb{M}_n \left(\mathrm{C}^0\left(\Complex^d\right)\right)</math> (that is, {{mvar|A}} continuously depends on the parameter {{math|'''c'''}}) the [[Jordan normal form]] of the matrix is continuously deformed [[almost everywhere]] on <math>\Complex^d</math> but, in general, ''not'' everywhere: there is some critical submanifold of <math>\Complex^d</math> on which the Jordan form abruptly changes its structure whenever the parameter crosses or simply "travels" around it ([[monodromy]]). Such changes mean that several Jordan blocks (either belonging to different eigenvalues or not) join to a unique Jordan block, or vice versa (that is, one Jordan block splits into two or more different ones). Many aspects of [[bifurcation theory]] for both continuous and discrete dynamical systems can be interpreted with the analysis of functional Jordan matrices. |
||
From the [[tangent space]] dynamics, this means that the orthogonal decomposition of the dynamical system's [[phase space]] changes and, for example, different orbits gain periodicity, or lose it, or shift from a certain kind of periodicity to another (such as ''period-doubling'', cfr. [[logistic map]]). |
From the [[tangent space]] dynamics, this means that the orthogonal decomposition of the dynamical system's [[phase space]] changes and, for example, different orbits gain periodicity, or lose it, or shift from a certain kind of periodicity to another (such as ''period-doubling'', cfr. [[logistic map]]). |
||
In a sentence, the qualitative behaviour of such a dynamical system may substantially change as the [[versal deformation]] of the Jordan normal form of |
In a sentence, the qualitative behaviour of such a dynamical system may substantially change as the [[versal deformation]] of the Jordan normal form of {{math|''A''('''c''')}}. |
||
== Linear ordinary differential equations == |
== Linear ordinary differential equations == |
||
The simplest example of a [[dynamical system]] is a system of linear, constant-coefficient, ordinary differential equations |
The simplest example of a [[dynamical system]] is a system of linear, constant-coefficient, ordinary differential equations; that is, let <math>A\in\mathbb{M}_n (\Complex)</math> and <math>\mathbf{z}_0 \in \Complex^n</math>: |
||
<math display="block">\begin{align} |
|||
⚫ | |||
\dot{\mathbf{z}}(t) &= A\mathbf{z}(t), \\ |
|||
⚫ | |||
\end{align}</math> |
|||
whose direct closed-form solution involves computation of the [[matrix exponential]]: |
whose direct closed-form solution involves computation of the [[matrix exponential]]: |
||
<math display="block">\mathbf{z}(t)=e^{tA}\mathbf{z}_0.</math> |
|||
Another way, provided the solution is restricted to the local [[Lp space|Lebesgue space]] of |
Another way, provided the solution is restricted to the local [[Lp space|Lebesgue space]] of {{mvar|n}}-dimensional vector fields <math>\mathbf{z}\in\mathrm{L}_{\mathrm{loc}}^1 (\R_+)^n</math>, is to use its [[Laplace transform]] <math>\mathbf{Z}(s) = \mathcal{L}[\mathbf{z}](s)</math>. In this case |
||
:<math>\mathbf{Z}(s)=\left(sI-A\right)^{-1}\mathbf{z}_0.</math> |
|||
⚫ | |||
⚫ | The matrix function |
||
⚫ | The matrix function {{math|(''A'' − ''sI'')<sup>−1</sup>}} is called the [[resolvent matrix]] of the [[differential operator]] <math display="inline">\frac{\mathrm{d}}{\mathrm{d}t}-A</math>. It is [[meromorphic]] with respect to the complex parameter <math>s \in \Complex</math> since its matrix elements are rational functions whose denominator is equal for all to {{math|det(''A'' − ''sI'')}}. Its polar singularities are the eigenvalues of {{mvar|A}}, whose order equals their index for it; that is, <math>\mathrm{ord}_{(A-sI)^{-1}}\lambda=\mathrm{idx}_A \lambda</math>. |
||
== See also == |
== See also == |
||
Line 113: | Line 119: | ||
==References== |
==References== |
||
* {{ |
* {{citation | first1 = Raymond A. | last1 = Beauregard | first2 = John B. | last2 = Fraleigh | year = 1973 | isbn = 0-395-14017-X | title = A First Course In Linear Algebra: with Optional Introduction to Groups, Rings, and Fields | publisher = [[Houghton Mifflin Co.]] | location = Boston | url-access = registration | url = https://archive.org/details/firstcourseinlin0000beau }} |
||
* {{ citation | first1 = Gene H. | last1 = Golub | first2 = Charles F. | last2 = Van Loan | year = 1996 | isbn = 0-8018-5414-8 | title = Matrix Computations | edition = 3rd | publisher = [[Johns Hopkins University Press]] | location = Baltimore }} |
* {{ citation | first1 = Gene H. | last1 = Golub | first2 = Charles F. | last2 = Van Loan | year = 1996 | isbn = 0-8018-5414-8 | title = Matrix Computations | edition = 3rd | publisher = [[Johns Hopkins University Press]] | location = Baltimore }} |
||
* {{ citation | first1 = Evar D. | last1 = Nering | year = 1970 | title = Linear Algebra and Matrix Theory | edition = 2nd | publisher = [[John Wiley & Sons|Wiley]] | location = New York | lccn = 76091646 }} |
* {{ citation | first1 = Evar D. | last1 = Nering | year = 1970 | title = Linear Algebra and Matrix Theory | edition = 2nd | publisher = [[John Wiley & Sons|Wiley]] | location = New York | lccn = 76091646 }} |
||
Line 119: | Line 125: | ||
[[Category:Matrix theory]] |
[[Category:Matrix theory]] |
||
[[Category:Matrix normal forms]] |
[[Category:Matrix normal forms]] |
||
[[de:Jordansche Normalform]] |
Latest revision as of 15:21, 20 January 2024
In the mathematical discipline of matrix theory, a Jordan matrix, named after Camille Jordan, is a block diagonal matrix over a ring R (whose identities are the zero 0 and one 1), where each block along the diagonal, called a Jordan block, has the following form:
Definition
[edit]Every Jordan block is specified by its dimension n and its eigenvalue , and is denoted as Jλ,n. It is an matrix of zeroes everywhere except for the diagonal, which is filled with and for the superdiagonal, which is composed of ones.
Any block diagonal matrix whose blocks are Jordan blocks is called a Jordan matrix. This (n1 + ⋯ + nr) × (n1 + ⋯ + nr) square matrix, consisting of r diagonal blocks, can be compactly indicated as or , where the i-th Jordan block is Jλi,ni.
For example, the matrix is a 10 × 10 Jordan matrix with a 3 × 3 block with eigenvalue 0, two 2 × 2 blocks with eigenvalue the imaginary unit i, and a 3 × 3 block with eigenvalue 7. Its Jordan-block structure is written as either or diag(J0,3, Ji,2, Ji,2, J7,3).
Linear algebra
[edit]Any n × n square matrix A whose elements are in an algebraically closed field K is similar to a Jordan matrix J, also in , which is unique up to a permutation of its diagonal blocks themselves. J is called the Jordan normal form of A and corresponds to a generalization of the diagonalization procedure.[1][2][3] A diagonalizable matrix is similar, in fact, to a special case of Jordan matrix: the matrix whose blocks are all 1 × 1.[4][5][6]
More generally, given a Jordan matrix , that is, whose kth diagonal block, , is the Jordan block Jλk,mk and whose diagonal elements may not all be distinct, the geometric multiplicity of for the matrix J, indicated as , corresponds to the number of Jordan blocks whose eigenvalue is λ. Whereas the index of an eigenvalue for J, indicated as , is defined as the dimension of the largest Jordan block associated to that eigenvalue.
The same goes for all the matrices A similar to J, so can be defined accordingly with respect to the Jordan normal form of A for any of its eigenvalues . In this case one can check that the index of for A is equal to its multiplicity as a root of the minimal polynomial of A (whereas, by definition, its algebraic multiplicity for A, , is its multiplicity as a root of the characteristic polynomial of A; that is, ). An equivalent necessary and sufficient condition for A to be diagonalizable in K is that all of its eigenvalues have index equal to 1; that is, its minimal polynomial has only simple roots.
Note that knowing a matrix's spectrum with all of its algebraic/geometric multiplicities and indexes does not always allow for the computation of its Jordan normal form (this may be a sufficient condition only for spectrally simple, usually low-dimensional matrices). Indeed, determining the Jordan normal form is generally a computationally challenging task. From the vector space point of view, the Jordan normal form is equivalent to finding an orthogonal decomposition (that is, via direct sums of eigenspaces represented by Jordan blocks) of the domain which the associated generalized eigenvectors make a basis for.
Functions of matrices
[edit]Let (that is, a n × n complex matrix) and be the change of basis matrix to the Jordan normal form of A; that is, A = C−1JC. Now let f (z) be a holomorphic function on an open set such that ; that is, the spectrum of the matrix is contained inside the domain of holomorphy of f. Let be the power series expansion of f around , which will be hereinafter supposed to be 0 for simplicity's sake. The matrix f (A) is then defined via the following formal power series and is absolutely convergent with respect to the Euclidean norm of . To put it another way, f (A) converges absolutely for every square matrix whose spectral radius is less than the radius of convergence of f around 0 and is uniformly convergent on any compact subsets of satisfying this property in the matrix Lie group topology.
The Jordan normal form allows the computation of functions of matrices without explicitly computing an infinite series, which is one of the main achievements of Jordan matrices. Using the facts that the kth power () of a diagonal block matrix is the diagonal block matrix whose blocks are the kth powers of the respective blocks; that is, , and that Ak = C−1JkC, the above matrix power series becomes
where the last series need not be computed explicitly via power series of every Jordan block. In fact, if , any holomorphic function of a Jordan block has a finite power series around because . Here, is the nilpotent part of and has all 0's except 1's along the superdiagonal. Thus it is the following upper triangular matrix:
As a consequence of this, the computation of any function of a matrix is straightforward whenever its Jordan normal form and its change-of-basis matrix are known. For example, using , the inverse of is:
Also, spec f(A) = f (spec A); that is, every eigenvalue corresponds to the eigenvalue , but it has, in general, different algebraic multiplicity, geometric multiplicity and index. However, the algebraic multiplicity may be computed as follows:
The function f (T) of a linear transformation T between vector spaces can be defined in a similar way according to the holomorphic functional calculus, where Banach space and Riemann surface theories play a fundamental role. In the case of finite-dimensional spaces, both theories perfectly match.
Dynamical systems
[edit]Now suppose a (complex) dynamical system is simply defined by the equation
where is the (n-dimensional) curve parametrization of an orbit on the Riemann surface of the dynamical system, whereas A(c) is an n × n complex matrix whose elements are complex functions of a d-dimensional parameter .
Even if (that is, A continuously depends on the parameter c) the Jordan normal form of the matrix is continuously deformed almost everywhere on but, in general, not everywhere: there is some critical submanifold of on which the Jordan form abruptly changes its structure whenever the parameter crosses or simply "travels" around it (monodromy). Such changes mean that several Jordan blocks (either belonging to different eigenvalues or not) join to a unique Jordan block, or vice versa (that is, one Jordan block splits into two or more different ones). Many aspects of bifurcation theory for both continuous and discrete dynamical systems can be interpreted with the analysis of functional Jordan matrices.
From the tangent space dynamics, this means that the orthogonal decomposition of the dynamical system's phase space changes and, for example, different orbits gain periodicity, or lose it, or shift from a certain kind of periodicity to another (such as period-doubling, cfr. logistic map).
In a sentence, the qualitative behaviour of such a dynamical system may substantially change as the versal deformation of the Jordan normal form of A(c).
Linear ordinary differential equations
[edit]The simplest example of a dynamical system is a system of linear, constant-coefficient, ordinary differential equations; that is, let and : whose direct closed-form solution involves computation of the matrix exponential:
Another way, provided the solution is restricted to the local Lebesgue space of n-dimensional vector fields , is to use its Laplace transform . In this case
The matrix function (A − sI)−1 is called the resolvent matrix of the differential operator . It is meromorphic with respect to the complex parameter since its matrix elements are rational functions whose denominator is equal for all to det(A − sI). Its polar singularities are the eigenvalues of A, whose order equals their index for it; that is, .
See also
[edit]- Jordan decomposition
- Jordan normal form
- Holomorphic functional calculus
- Matrix exponential
- Logarithm of a matrix
- Dynamical system
- Bifurcation theory
- State space (controls)
Notes
[edit]- ^ Beauregard & Fraleigh (1973, pp. 310–316)
- ^ Golub & Van Loan (1996, p. 317)
- ^ Nering (1970, pp. 118–127)
- ^ Beauregard & Fraleigh (1973, pp. 270–274)
- ^ Golub & Van Loan (1996, p. 316)
- ^ Nering (1970, pp. 113–118)
References
[edit]- Beauregard, Raymond A.; Fraleigh, John B. (1973), A First Course In Linear Algebra: with Optional Introduction to Groups, Rings, and Fields, Boston: Houghton Mifflin Co., ISBN 0-395-14017-X
- Golub, Gene H.; Van Loan, Charles F. (1996), Matrix Computations (3rd ed.), Baltimore: Johns Hopkins University Press, ISBN 0-8018-5414-8
- Nering, Evar D. (1970), Linear Algebra and Matrix Theory (2nd ed.), New York: Wiley, LCCN 76091646