Elementary matrix

Matrix which differs from the identity matrix by one elementary row operation

In mathematics, an elementary matrix is a matrix which differs from the identity matrix by one single elementary row operation. The elementary matrices generate the general linear group GLn(F) when F is a field. Left multiplication (pre-multiplication) by an elementary matrix represents elementary row operations, while right multiplication (post-multiplication) represents elementary column operations.

Elementary row operations are used in Gaussian elimination to reduce a matrix to row echelon form. They are also used in Gauss–Jordan elimination to further reduce the matrix to reduced row echelon form.

Elementary row operations

There are three types of elementary matrices, which correspond to three types of row operations (respectively, column operations):

Row switching
A row within the matrix can be switched with another row.
R i R j {\displaystyle R_{i}\leftrightarrow R_{j}}
Row multiplication
Each element in a row can be multiplied by a non-zero constant. It is also known as scaling a row.
k R i R i ,   where  k 0 {\displaystyle kR_{i}\rightarrow R_{i},\ {\mbox{where }}k\neq 0}
Row addition
A row can be replaced by the sum of that row and a multiple of another row.
R i + k R j R i , where  i j {\displaystyle R_{i}+kR_{j}\rightarrow R_{i},{\mbox{where }}i\neq j}

If E is an elementary matrix, as described below, to apply the elementary row operation to a matrix A, one multiplies A by the elementary matrix on the left, EA. The elementary matrix for any row operation is obtained by executing the operation on the identity matrix. This fact can be understood as an instance of the Yoneda lemma applied to the category of matrices.

Row-switching transformations

The first type of row operation on a matrix A switches all matrix elements on row i with their counterparts on a different row j. The corresponding elementary matrix is obtained by swapping row i and row j of the identity matrix.

T i , j = [ 1 0 1 1 0 1 ] {\displaystyle T_{i,j}={\begin{bmatrix}1&&&&&&\\&\ddots &&&&&\\&&0&&1&&\\&&&\ddots &&&\\&&1&&0&&\\&&&&&\ddots &\\&&&&&&1\end{bmatrix}}}

So Ti,j A is the matrix produced by exchanging row i and row j of A.

Coefficient wise, the matrix Ti,j is defined by :

[ T i , j ] k , l = { 0 k i , k j , k l 1 k i , k j , k = l 0 k = i , l j 1 k = i , l = j 0 k = j , l i 1 k = j , l = i {\displaystyle [T_{i,j}]_{k,l}={\begin{cases}0&k\neq i,k\neq j,k\neq l\\1&k\neq i,k\neq j,k=l\\0&k=i,l\neq j\\1&k=i,l=j\\0&k=j,l\neq i\\1&k=j,l=i\\\end{cases}}}

Properties

  • The inverse of this matrix is itself: T i , j 1 = T i , j . {\displaystyle T_{i,j}^{-1}=T_{i,j}.}
  • Since the determinant of the identity matrix is unity, det ( T i , j ) = 1. {\displaystyle \det(T_{i,j})=-1.} It follows that for any square matrix A (of the correct size), we have det ( T i , j A ) = det ( A ) . {\displaystyle \det(T_{i,j}A)=-\det(A).}
  • For theoretical considerations, the row-switching transformation can be obtained from row-addition and row-multiplication transformations introduced below because T i , j = D i ( 1 ) L i , j ( 1 ) L j , i ( 1 ) L i , j ( 1 ) . {\displaystyle T_{i,j}=D_{i}(-1)\,L_{i,j}(-1)\,L_{j,i}(1)\,L_{i,j}(-1).}

Row-multiplying transformations

The next type of row operation on a matrix A multiplies all elements on row i by m where m is a non-zero scalar (usually a real number). The corresponding elementary matrix is a diagonal matrix, with diagonal entries 1 everywhere except in the ith position, where it is m.

D i ( m ) = [ 1 1 m 1 1 ] {\displaystyle D_{i}(m)={\begin{bmatrix}1&&&&&&\\&\ddots &&&&&\\&&1&&&&\\&&&m&&&\\&&&&1&&\\&&&&&\ddots &\\&&&&&&1\end{bmatrix}}}

So Di(m)A is the matrix produced from A by multiplying row i by m.

Coefficient wise, the Di(m) matrix is defined by :

[ D i ( m ) ] k , l = { 0 k l 1 k = l , k i m k = l , k = i {\displaystyle [D_{i}(m)]_{k,l}={\begin{cases}0&k\neq l\\1&k=l,k\neq i\\m&k=l,k=i\end{cases}}}

Properties

  • The inverse of this matrix is given by D i ( m ) 1 = D i ( 1 m ) . {\displaystyle D_{i}(m)^{-1}=D_{i}\left({\tfrac {1}{m}}\right).}
  • The matrix and its inverse are diagonal matrices.
  • det ( D i ( m ) ) = m . {\displaystyle \det(D_{i}(m))=m.} Therefore for a square matrix A (of the correct size), we have det ( D i ( m ) A ) = m det ( A ) . {\displaystyle \det(D_{i}(m)A)=m\det(A).}

Row-addition transformations

The final type of row operation on a matrix A adds row j multiplied by a scalar m to row i. The corresponding elementary matrix is the identity matrix but with an m in the (i, j) position.

L i j ( m ) = [ 1 1 m 1 1 ] {\displaystyle L_{ij}(m)={\begin{bmatrix}1&&&&&&\\&\ddots &&&&&\\&&1&&&&\\&&&\ddots &&&\\&&m&&1&&\\&&&&&\ddots &\\&&&&&&1\end{bmatrix}}}

So Lij(m)A is the matrix produced from A by adding m times row j to row i. And A Lij(m) is the matrix produced from A by adding m times column i to column j.

Coefficient wise, the matrix Li,j(m) is defined by :

[ L i , j ( m ) ] k , l = { 0 k l , k i , l j 1 k = l m k = i , l = j {\displaystyle [L_{i,j}(m)]_{k,l}={\begin{cases}0&k\neq l,k\neq i,l\neq j\\1&k=l\\m&k=i,l=j\end{cases}}}

Properties

  • These transformations are a kind of shear mapping, also known as a transvections.
  • The inverse of this matrix is given by L i j ( m ) 1 = L i j ( m ) . {\displaystyle L_{ij}(m)^{-1}=L_{ij}(-m).}
  • The matrix and its inverse are triangular matrices.
  • det ( L i j ( m ) ) = 1. {\displaystyle \det(L_{ij}(m))=1.} Therefore, for a square matrix A (of the correct size) we have det ( L i j ( m ) A ) = det ( A ) . {\displaystyle \det(L_{ij}(m)A)=\det(A).}
  • Row-addition transforms satisfy the Steinberg relations.

See also

References

  • Axler, Sheldon Jay (1997), Linear Algebra Done Right (2nd ed.), Springer-Verlag, ISBN 0-387-98259-0
  • Lay, David C. (August 22, 2005), Linear Algebra and Its Applications (3rd ed.), Addison Wesley, ISBN 978-0-321-28713-7
  • Meyer, Carl D. (February 15, 2001), Matrix Analysis and Applied Linear Algebra, Society for Industrial and Applied Mathematics (SIAM), ISBN 978-0-89871-454-8, archived from the original on 2009-10-31
  • Poole, David (2006), Linear Algebra: A Modern Introduction (2nd ed.), Brooks/Cole, ISBN 0-534-99845-3
  • Anton, Howard (2005), Elementary Linear Algebra (Applications Version) (9th ed.), Wiley International
  • Leon, Steven J. (2006), Linear Algebra With Applications (7th ed.), Pearson Prentice Hall
  • Strang, Gilbert (2016), Introduction to Linear Algebra (5th ed.), Wellesley-Cambridge Press, ISBN 978-09802327-7-6
  • v
  • t
  • e
Matrix classes
Explicitly constrained entriesConstantConditions on eigenvalues or eigenvectorsSatisfying conditions on products or inversesWith specific applicationsUsed in statisticsUsed in graph theoryUsed in science and engineeringRelated terms