Jump to content

Elementary matrix: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Korean Version Adding
 
(81 intermediate revisions by 60 users not shown)
Line 1: Line 1:
{{short description|Matrix which differs from the identity matrix by one elementary row operation}}
In [[mathematics]], an '''elementary matrix''' is a simple [[Matrix (mathematics)|matrix]] which differs from the [[identity matrix]] by one single elementary row operation. The elementary matrices generate the [[general linear group]] of [[invertible matrix|invertible matrices]]. Left multiplication (pre-multiplication) by an elementary matrix represents '''elementary row operations''', while right multiplication (post-multiplication) represents '''elementary column operations'''.


In [[mathematics]], an '''elementary matrix''' is a square [[matrix (mathematics)|matrix]] obtained from the application of a single elementary row operation to the [[identity matrix]]. The elementary matrices generate the [[general linear group]] {{math|GL<sub>''n''</sub>('''F''')}} when {{math|'''F'''}} is a [[Field (mathematics)|field]]. Left multiplication (pre-multiplication) by an elementary matrix represents '''elementary row operations''', while right multiplication (post-multiplication) represents '''elementary column operations'''.
In [[algebraic K-theory]], "elementary matrices" refers ''only'' to the row-addition matrices.


Elementary row operations are used in [[Gaussian elimination]] to reduce a matrix to [[row echelon form]]. They are also used in [[Gauss–Jordan elimination]] to further reduce the matrix to [[reduced row echelon form]].
==Use in solving systems of equations==
Elementary ''row'' operations do not change the solution set of the [[system of linear equations]] represented by a matrix, and are used in [[Gaussian elimination]] (respectively, [[Gauss-Jordan elimination]]) to reduce a matrix to [[row echelon form]] (respectively, [[reduced row echelon form]]).


The acronym "ERO" is commonly used for "elementary row operations".
==Elementary row operations==

Elementary row operations do not change the [[kernel (linear operator)|kernel]] of a matrix (and hence do not change the solution set), but they ''do'' change the
[[image (mathematics)|image]]. [[Duality (mathematics)|Dually]], elementary ''column'' operations do not change the ''image'', but they do change the ''kernel''.

There are three types of (n x n) elementary matrices:
1) Permutation Matrix
2) Diagonal Matrix
3) Unipotent Matrix.

==Operations==
There are three types of elementary matrices, which correspond to three types of row operations (respectively, column operations):
There are three types of elementary matrices, which correspond to three types of row operations (respectively, column operations):


Line 22: Line 11:
: <math>R_i \leftrightarrow R_j</math>
: <math>R_i \leftrightarrow R_j</math>


;Row multiplication: Each element in a row can be multiplied by a non-zero constant.
;Row multiplication: Each element in a row can be multiplied by a non-zero constant. It is also known as ''scaling'' a row.
: <math>kR_i \rightarrow R_i,\ \mbox{where } k \neq 0</math>
: <math>kR_i \rightarrow R_i,\ \mbox{where } k \neq 0</math>


Line 28: Line 17:
: <math>R_i + kR_j \rightarrow R_i, \mbox{where } i \neq j </math>
: <math>R_i + kR_j \rightarrow R_i, \mbox{where } i \neq j </math>


The elementary matrix for any row operation is obtained by executing the operation on an [[identity matrix]].
If {{mvar|E}} is an elementary matrix, as described below, to apply the elementary row operation to a matrix {{mvar|A}}, one multiplies {{mvar|A}} by the elementary matrix on the left, {{mvar|EA}}. The elementary matrix for any row operation is obtained by executing the operation on the [[identity matrix]]. This fact can be understood as an instance of the [[Yoneda lemma]] applied to the category of matrices.<ref>{{harvp|Perrone|2024|pages=119-120}}</ref>


===Row-switching transformations===
===Row-switching transformations===
{{See also|Permutation matrix}}
This transformation, ''T<sub>ij</sub>'', switches all matrix elements on row ''i'' with their counterparts on row ''j''. The matrix resulting in this transformation is obtained by swapping row ''i'' and row ''j'' of the [[identity matrix]].
The first type of row operation on a matrix {{mvar|A}} switches all matrix elements on row {{mvar|i}} with their counterparts on a different row {{mvar|j}}. The corresponding elementary matrix is obtained by swapping row {{mvar|i}} and row {{mvar|j}} of the [[identity matrix]].

:<math>T_{i,j} = \begin{bmatrix}
1 & & & & & & \\
& \ddots & & & & & \\
& & 0 & & 1 & & \\
& & & \ddots & & & \\
& & 1 & & 0 & & \\
& & & & & \ddots & \\
& & & & & & 1
\end{bmatrix}</math>

So {{mvar|T<sub>i,j</sub> A}} is the matrix produced by exchanging row {{mvar|i}} and row {{mvar|j}} of {{mvar|A}}.

Coefficient wise, the matrix {{mvar|T<sub>i,j</sub>}} is defined by :


:<math>
:<math>
[T_{i,j}]_{k,l} =
T_{i,j} = \begin{bmatrix} 1 & & & & & & & \\ & \ddots & & & & & & \\ & & 0 & & 1 & & \\ & & & \ddots & & & & \\ & & 1 & & 0 & & \\ & & & & & & \ddots & \\ & & & & & & & 1\end{bmatrix}\quad </math>
\begin{cases}
:That is, ''T<sub>ij</sub>'' is the matrix produced by exchanging row ''i'' and row ''j'' of the identity matrix.
0 & k \neq i, k \neq j ,k \neq l \\
1 & k \neq i, k \neq j, k = l\\
0 & k = i, l \neq j\\
1 & k = i, l = j\\
0 & k = j, l \neq i\\
1 & k = j, l = i\\
\end{cases}</math>


====Properties====
====Properties====
:*The inverse of this matrix is itself: ''T<sub>ij</sub><sup>&minus;1</sup>=T<sub>ij</sub>''.
* The inverse of this matrix is itself: <math>T_{i,j}^{-1} = T_{i,j}.</math>
:*Since the [[determinant]] of the identity matrix is unity, det[''T''<sub>''ij''</sub>] = &minus;1. It follows that for any [[conformable]] square matrix ''A'': det[''T''<sub>''ij''</sub>''A''] = &minus;det[''A''].
* Since the [[determinant]] of the identity matrix is unity, <math>\det(T_{i,j}) = -1.</math> It follows that for any square matrix {{mvar|A}} (of the correct size), we have <math>\det(T_{i,j}A) = -\det(A).</math>
* For theoretical considerations, the row-switching transformation can be obtained from row-addition and row-multiplication transformations introduced below because <math>T_{i,j}=D_i(-1)\,L_{i,j}(-1)\,L_{j,i}(1)\,L_{i,j}(-1).</math>


===Row-multiplying transformations===
===Row-multiplying transformations===
This transformation, ''T<sub>i</sub>''(''m''), multiplies all elements on row ''i'' by ''m'' where ''m'' is non zero. The matrix resulting in this transformation is obtained by multiplying all elements of row ''i'' of the identity matrix by ''m''.
The next type of row operation on a matrix {{mvar|A}} multiplies all elements on row {{mvar|i}} by {{mvar|m}} where {{mvar|m}} is a non-zero [[scalar (mathematics)|scalar]] (usually a real number). The corresponding elementary matrix is a diagonal matrix, with diagonal entries 1 everywhere except in the {{mvar|i}}th position, where it is {{mvar|m}}.

:<math>D_i(m) = \begin{bmatrix}
1 & & & & & & \\
& \ddots & & & & & \\
& & 1 & & & & \\
& & & m & & & \\
& & & & 1 & & \\
& & & & & \ddots & \\
& & & & & & 1
\end{bmatrix}</math>

So {{math|''D<sub>i</sub>''(''m'')''A''}} is the matrix produced from {{mvar|A}} by multiplying row {{mvar|i}} by {{mvar|m}}.

Coefficient wise, the {{math|''D<sub>i</sub>''(''m'')}} matrix is defined by :


:<math>
:<math>
[D_i(m)]_{k,l} = \begin{cases}
T_i(m) = \begin{bmatrix} 1 & & & & & & \\ & \ddots & & & & & \\ & & 1 & & & & \\ & & & m & & & \\ & & & & 1 & & \\ & & & & & \ddots & \\ & & & & & & 1\end{bmatrix}\quad </math>
0 & k \neq l \\
1 & k = l, k \neq i \\
m & k = l, k= i
\end{cases}</math>


====Properties====
====Properties====
:*The inverse of this matrix is: ''T<sub>i</sub>''(''m'')<sup>&minus;1</sup> = ''T<sub>i</sub>''(1/''m'').
* The inverse of this matrix is given by <math>D_i(m)^{-1} = D_i \left(\tfrac 1 m \right).</math>
:*The matrix and its inverse are [[diagonal matrix|diagonal matrices]].
* The matrix and its inverse are [[diagonal matrix|diagonal matrices]].
:*det[''T''<sub>''i''</sub>(m)] = ''m''. Therefore for a conformable square matrix ''A'': det[''T''<sub>''i''</sub>(''m'')''A''] = ''m'' det[''A''].
* <math>\det(D_i(m)) = m.</math> Therefore, for a square matrix {{mvar|A}} (of the correct size), we have <math>\det(D_i(m)A) = m\det(A).</math>


===Row-addition transformations===
===Row-addition transformations===
This transformation, ''T<sub>ij</sub>''(''m''), adds row ''j'' multiplied by ''m'' to row ''i''. The matrix resulting in this transformation is obtained by taking row ''j'' of the identity matrix, and adding it ''m'' times to row ''i''.
The final type of row operation on a matrix {{mvar|A}} adds row {{mvar|j}} multiplied by a scalar {{mvar|m}} to row {{mvar|i}}. The corresponding elementary matrix is the identity matrix but with an {{mvar|m}} in the {{math|(''i, j'')}} position.
:<math>
:<math>L_{ij}(m) = \begin{bmatrix}
T_{i,j}(m) = \begin{bmatrix} 1 & & & & & & & \\ & \ddots & & & & & & \\ & & 1 & & & & & \\ & & & \ddots & & & & \\ & & m & & 1 & & \\ & & & & & & \ddots & \\ & & & & & & & 1\end{bmatrix}
1 & & & & & & \\
& \ddots & & & & & \\
</math>
& & 1 & & & & \\
These are also called '''[[shear mapping]]s''' or '''transvections'''.
& & & \ddots & & & \\
& & m & & 1 & & \\
& & & & & \ddots & \\
& & & & & & 1
\end{bmatrix}</math>

So {{math|''L<sub>ij</sub>''(''m'')''A''}} is the matrix produced from {{mvar|A}} by adding {{mvar|m}} times row {{mvar|j}} to row {{mvar|i}}.
And {{math|''A L<sub>ij</sub>''(''m'')}} is the matrix produced from {{mvar|A}} by adding {{mvar|m}} times column {{mvar|i}} to column {{mvar|j}}.

Coefficient wise, the matrix {{math|''L{{sub|i,j}}''(''m'')}} is defined by :

:<math>[L_{i,j}(m)]_{k,l} = \begin{cases}
0 & k \neq l, k \neq i, l \neq j \\
1 & k = l \\
m & k = i, l = j
\end{cases}</math>


====Properties====
====Properties====
* These transformations are a kind of [[shear mapping]], also known as a ''transvections''.
:*''T<sub>ij</sub>''(''m'')<sup>&minus;1</sup> = ''T<sub>ij</sub>''(&minus;''m'') (inverse matrix).
* The inverse of this matrix is given by <math>L_{ij}(m)^{-1} = L_{ij}(-m).</math>
:*The matrix and its inverse are [[triangular matrix|triangular matrices]].
* The matrix and its inverse are [[triangular matrix|triangular matrices]].
:*det[''T<sub>ij</sub>''(''m'')] = 1. Therefore, for a [[conformable matrix|conformable]] square matrix ''A'': det[''T''<sub>''ij''</sub>(''1'')''A''] = det[''A''].
* <math>\det(L_{ij}(m)) = 1.</math> Therefore, for a square matrix {{mvar|A}} (of the correct size) we have <math>\det(L_{ij}(m)A) = \det(A).</math>
:Row-addition transforms satisfy the [[Steinberg relations]].
* Row-addition transforms satisfy the [[Steinberg relations]].


==See also==
==See also==
*[[Gaussian elimination]]
* [[Gaussian elimination]]
*[[Linear algebra]]
* [[Linear algebra]]
*[[System of linear equations]]
* [[System of linear equations]]
*[[Matrix (mathematics)]]
* [[Matrix (mathematics)]]
* [[LU decomposition]]
* [[Frobenius matrix]]


==References==
==References==
{{reflist}}
{{See also|Linear algebra#Further reading}}
{{See also|Linear algebra#Further reading}}
* {{Citation
* {{Citation
Line 80: Line 130:
| publisher = Springer-Verlag
| publisher = Springer-Verlag
| edition = 2nd
| edition = 2nd
| isbn = 0387982590
| isbn = 0-387-98259-0
}}
}}
* {{Citation
* {{Citation
Line 89: Line 139:
| publisher = Addison Wesley
| publisher = Addison Wesley
| edition = 3rd
| edition = 3rd
| isbn = 978-0321287137
| isbn = 978-0-321-28713-7
}}
}}
* {{Citation
* {{Citation
| last = Meyer
|last = Meyer
| first = Carl D.
|first = Carl D.
| date = February 15, 2001
|date = February 15, 2001
| title = Matrix Analysis and Applied Linear Algebra
|title = Matrix Analysis and Applied Linear Algebra
| publisher = Society for Industrial and Applied Mathematics (SIAM)
|publisher = Society for Industrial and Applied Mathematics (SIAM)
| isbn = 978-0898714548
|isbn = 978-0-89871-454-8
| url = http://www.matrixanalysis.com/DownloadChapters.html
|url = http://www.matrixanalysis.com/DownloadChapters.html
|url-status = dead
|archive-url = https://web.archive.org/web/20091031193126/http://matrixanalysis.com/DownloadChapters.html
|archive-date = 2009-10-31
}}
* {{Citation
|last = Perrone |first = Paolo
|title = Starting Category Theory
|date = 2024
|publisher = World Scientific
|doi = 10.1142/9789811286018_0005
|isbn = 978-981-12-8600-1
|url = https://www.worldscientific.com/worldscibooks/10.1142/13670
}}
}}
* {{Citation
* {{Citation
Line 124: Line 186:
| publisher = Pearson Prentice Hall
| publisher = Pearson Prentice Hall
| edition = 7th
| edition = 7th
}}
* {{Citation
| last = Strang
| first = Gilbert
| author-link = Gilbert Strang
| year = 2016
| title = Introduction to Linear Algebra
| publisher = Wellesley-Cambridge Press
| edition = 5th|isbn=978-09802327-7-6
}}
}}

{{Matrix classes}}


{{DEFAULTSORT:Elementary Matrix}}
{{DEFAULTSORT:Elementary Matrix}}
[[Category:Linear algebra]]
[[Category:Linear algebra]]

[[de:Elementarmatrix]]
[[eo:Rudimenta operacio kun matrico]]
[[fa:عمل سطری مقدماتی]]
[[fr:Matrice élémentaire]]
[[is:Frumfylki]]
[[nl:Elementaire matrix]]
[[ja:行列の基本変形]]
[[pl:Operacje elementarne]]
[[ru:Элементарные преобразования матрицы]]
[[sl:Elementarna matrika]]
[[fi:Alkeismatriisi]]
[[sv:Elementär matris]]
[[uk:Елементарні перетворення матриці]]
[[zh:初等矩阵]]
[[kr:기본행렬]]

Latest revision as of 20:55, 18 October 2024

In mathematics, an elementary matrix is a square matrix obtained from the application of a single elementary row operation to the identity matrix. The elementary matrices generate the general linear group GLn(F) when F is a field. Left multiplication (pre-multiplication) by an elementary matrix represents elementary row operations, while right multiplication (post-multiplication) represents elementary column operations.

Elementary row operations are used in Gaussian elimination to reduce a matrix to row echelon form. They are also used in Gauss–Jordan elimination to further reduce the matrix to reduced row echelon form.

Elementary row operations

[edit]

There are three types of elementary matrices, which correspond to three types of row operations (respectively, column operations):

Row switching
A row within the matrix can be switched with another row.
Row multiplication
Each element in a row can be multiplied by a non-zero constant. It is also known as scaling a row.
Row addition
A row can be replaced by the sum of that row and a multiple of another row.

If E is an elementary matrix, as described below, to apply the elementary row operation to a matrix A, one multiplies A by the elementary matrix on the left, EA. The elementary matrix for any row operation is obtained by executing the operation on the identity matrix. This fact can be understood as an instance of the Yoneda lemma applied to the category of matrices.[1]

Row-switching transformations

[edit]

The first type of row operation on a matrix A switches all matrix elements on row i with their counterparts on a different row j. The corresponding elementary matrix is obtained by swapping row i and row j of the identity matrix.

So Ti,j A is the matrix produced by exchanging row i and row j of A.

Coefficient wise, the matrix Ti,j is defined by :

Properties

[edit]
  • The inverse of this matrix is itself:
  • Since the determinant of the identity matrix is unity, It follows that for any square matrix A (of the correct size), we have
  • For theoretical considerations, the row-switching transformation can be obtained from row-addition and row-multiplication transformations introduced below because

Row-multiplying transformations

[edit]

The next type of row operation on a matrix A multiplies all elements on row i by m where m is a non-zero scalar (usually a real number). The corresponding elementary matrix is a diagonal matrix, with diagonal entries 1 everywhere except in the ith position, where it is m.

So Di(m)A is the matrix produced from A by multiplying row i by m.

Coefficient wise, the Di(m) matrix is defined by :

Properties

[edit]
  • The inverse of this matrix is given by
  • The matrix and its inverse are diagonal matrices.
  • Therefore, for a square matrix A (of the correct size), we have

Row-addition transformations

[edit]

The final type of row operation on a matrix A adds row j multiplied by a scalar m to row i. The corresponding elementary matrix is the identity matrix but with an m in the (i, j) position.

So Lij(m)A is the matrix produced from A by adding m times row j to row i. And A Lij(m) is the matrix produced from A by adding m times column i to column j.

Coefficient wise, the matrix Li,j(m) is defined by :

Properties

[edit]
  • These transformations are a kind of shear mapping, also known as a transvections.
  • The inverse of this matrix is given by
  • The matrix and its inverse are triangular matrices.
  • Therefore, for a square matrix A (of the correct size) we have
  • Row-addition transforms satisfy the Steinberg relations.

See also

[edit]

References

[edit]
  1. ^ Perrone (2024), pp. 119–120
  • Axler, Sheldon Jay (1997), Linear Algebra Done Right (2nd ed.), Springer-Verlag, ISBN 0-387-98259-0
  • Lay, David C. (August 22, 2005), Linear Algebra and Its Applications (3rd ed.), Addison Wesley, ISBN 978-0-321-28713-7
  • Meyer, Carl D. (February 15, 2001), Matrix Analysis and Applied Linear Algebra, Society for Industrial and Applied Mathematics (SIAM), ISBN 978-0-89871-454-8, archived from the original on 2009-10-31
  • Perrone, Paolo (2024), Starting Category Theory, World Scientific, doi:10.1142/9789811286018_0005, ISBN 978-981-12-8600-1
  • Poole, David (2006), Linear Algebra: A Modern Introduction (2nd ed.), Brooks/Cole, ISBN 0-534-99845-3
  • Anton, Howard (2005), Elementary Linear Algebra (Applications Version) (9th ed.), Wiley International
  • Leon, Steven J. (2006), Linear Algebra With Applications (7th ed.), Pearson Prentice Hall
  • Strang, Gilbert (2016), Introduction to Linear Algebra (5th ed.), Wellesley-Cambridge Press, ISBN 978-09802327-7-6