Matrix: Difference between revisions

From Citizendium
Jump to navigation Jump to search
imported>Richard Pinch
mNo edit summary
 
(8 intermediate revisions by 3 users not shown)
Line 1: Line 1:
{{subpages}}
{{subpages}}


A '''matrix''' (plural "matrices") is a mathematical construct generally represented as a rectangular array of elements.  These elements are usually numbers, but can be other matrices, and other mathematical structures. <!-- Geeknote: any ring with inverses --> Matrices with real elements are called "real matrices" Matrices with complex elements are called "complex matrices", and so on. Matrices with matrices as elements are called "partitioned matrices".
A '''matrix''' (plural "matrices") is a mathematical construct generally represented as a rectangular array of elements.  These elements are usually numbers, but can be other matrices, and other mathematical structures. <!-- Geeknote: any ring with inverses --> Matrices with real elements are called "real matrices", matrices with complex elements are called "complex matrices", and so on. Matrices with matrices as elements are called "partitioned matrices".


Among their other applications, matrices are fundamental to the study of [[linear algebra]], since every [[linear transformation]] can be represented by a [[uniqueness|unique]] matrix, and every matrix represents a unique linear transformation.
Among their other applications, matrices are fundamental to the study of [[linear algebra]], since every [[linear transformation]] can be represented by a [[uniqueness|unique]] matrix, and every matrix represents a unique linear transformation.
Line 136: Line 136:
=== Identity matrix ===
=== Identity matrix ===
{{main|Identity matrix}}
{{main|Identity matrix}}
We denote by <math>\scriptstyle \mathbb{I}_n</math> the [[multiplicative identity]] for matrix multiplication; that is, the matrix such that
We denote by <math>\scriptstyle \mathbb{I}_n</math> the [[multiplicative identity]] for matrix multiplication; that is, the matrix such that


:<math>\scriptstyle A \mathbb{I}_n = A = \mathbb{I}_m A</math> for any ''m''&times;''n'' matrix ''A''.
:<math>\scriptstyle A \mathbb{I}_n = A = \mathbb{I}_m A</math> for any ''m''&times;''n'' matrix ''A''.


<math>\scriptstyle \mathbb{I}_n</math> takes the form of an ''n''&times;''n'' square matrix with ones down its main diagonal -- the diagonal starting at (1,1) -- and zeroes everywhere else.  So
<math>\scriptstyle \mathbb{I}_n</math> takes the form of a ''n''&times;''n'' square matrix with the number 1 down its main diagonal, starting from element (1, 1), and the number 0 everywhere else.  So


:<math>\mathbb{I}_1 = \begin{pmatrix}1\end{pmatrix}</math>, <math>\mathbb{I}_2 = \begin{pmatrix}
:<math>\mathbb{I}_1 = \begin{pmatrix}1\end{pmatrix}</math>, <math>\mathbb{I}_2 = \begin{pmatrix}
Line 177: Line 176:


=== Invertible matrix ===
=== Invertible matrix ===
 
{{main|Matrix inverse}}
Some, but not all, matrices have a multiplicative inverse.  That is, for a matrix ''A'', there may exist a matrix <math>\scriptstyle A^{-1}</math> such that
Some, but not all, matrices have a multiplicative inverse.  That is, for a matrix ''A'', there may exist a matrix <math>\scriptstyle A^{-1}</math> such that


Line 203: Line 202:
=== Antisymmetric matrix ===
=== Antisymmetric matrix ===


An antisymmetric matrix is the additive inverse of its transpose.  It must also therefore be a square matix.  If ''A'' is an ''n''&times;''n'' matrix, ''A'' is antisymmetric if and only if
An antisymmetric or skew-symmetric matrix is the additive inverse of its transpose.  It must also therefore be a square matrix.  If ''A'' is an ''n''&times;''n'' matrix, ''A'' is antisymmetric if and only if


:<math>\scriptstyle A_{i,j} = -A_{j,i}</math> for all <math>\scriptstyle 1 \leq i, j \leq n</math>
:<math>\scriptstyle A_{i,j} = -A_{j,i}</math> for all <math>\scriptstyle 1 \leq i, j \leq n</math>
Line 223: Line 222:
=== Diagonal matrix ===
=== Diagonal matrix ===
{{main|Diagonal matrix}}
{{main|Diagonal matrix}}
 
A matrix that has zeros everywhere other than the main (upper left to lower right) diagonal is a ''Diagonal'' matrix.
A matrix that has zeros everywhere other than the main (upper left to lower right) diagonal is called a [[Diagonal Matrix]].


Example:
Example:
Line 235: Line 233:


Diagonal matrices are especially easy to invert or operate with.
Diagonal matrices are especially easy to invert or operate with.
=== Scalar matrix ===
A diagonal matrix all of whose diagonal entries are equal is a ''Scalar'' matrix.  Scalar matrices are so called because their multiplicative action on other matrices is the same as multiplication by a scalar element.  A scalar matrix is just a scalar multiple of the [[identity matrix]].
Example:
:<math>A = \begin{pmatrix}
3 & 0 & 0 & 0 \\
0 & 3 & 0 & 0 \\
0 & 0 & 3 & 0 \\
0 & 0 & 0 & 3
\end{pmatrix}</math>


=== Triangular matrix ===
=== Triangular matrix ===
Line 270: Line 280:
<br/>or in matrix notation:
<br/>or in matrix notation:
<br/><math>\mathbf{A=A^*(=\overline{A'})}</math>
<br/><math>\mathbf{A=A^*(=\overline{A'})}</math>
=== Idempotent Matrix ===
A matrix is ''Idempotent'' if it is equal to its square.
:<math>P^2 = P .\,</math>
An idempotent matrix ''P'' has eigenvalues 0 or 1 and has a basis of eigenvectors: it is [[diagonalizable matrix|diagonalisable]] since its minimal polynomial polynomial ''X''<sup>2</sup>-''X'' has no repeated roots.  The kernel and image of ''P'' are [[complement (linear algebra)|complements]]: they form an [[internal direct sum]].


=== Sparse Matrix ===
=== Sparse Matrix ===


If a matrix has enough zero elements that one can take advantage of the fact, a matrix is called a "[[Sparse Matrix]]".
If a matrix has enough zero elements that one can take advantage of the fact, a matrix is called a "[[Sparse Matrix]]".
Triangular matrices and Diagonal matrices are examples of sparce matrices
Triangular matrices and Diagonal matrices are examples of sparse matrices


== Applications ==
== Applications ==
Line 334: Line 350:
=== Finite Element methods ===
=== Finite Element methods ===


Breaking an object into a mesh of points, and computing the interactions of those points, is used in a number of Engineering fields to model materials.  Various matrix methods are used to deal with the problem.
Breaking an object into a mesh of points, and computing the interactions of those points, is used in a number of Engineering fields to model materials.  Various matrix methods are used to deal with the problem.[[Category:Suggestion Bot Tag]]

Latest revision as of 17:01, 16 September 2024

This article is developing and not approved.
Main Article
Discussion
Related Articles  [?]
Bibliography  [?]
External Links  [?]
Citable Version  [?]
 
This editable Main Article is under development and subject to a disclaimer.

A matrix (plural "matrices") is a mathematical construct generally represented as a rectangular array of elements. These elements are usually numbers, but can be other matrices, and other mathematical structures. Matrices with real elements are called "real matrices", matrices with complex elements are called "complex matrices", and so on. Matrices with matrices as elements are called "partitioned matrices".

Among their other applications, matrices are fundamental to the study of linear algebra, since every linear transformation can be represented by a unique matrix, and every matrix represents a unique linear transformation.

Dimensions and coordinates

A matrix with m rows and n columns is described as an m×n (pronounced "m by n") matrix, with the number of rows always coming first. When one dimension of a matrix is equal to 1 -- that is, in the case of a 1×n or m×1 matrix -- the matrix is a vector. A matrix with one row is a row vector; a matrix with one column is a column vector.

If the m×n matrix is named A, individual entries are named , where and ; again, the row coordinate comes first. For example, suppose M is a 3×4 matrix:

Now we can say that = 42: the element in the second row and the fourth column, counting from the top left.

Notational conventions vary; the comma in the subscript is sometimes omitted, so the same entry would be named ; of course, this notation is only practical when the matrix in question is smaller than 10×10. A superscript-subscript notation is sometimes used, where the row coordinate appears as a superscript and the column coordinate appears as a subscript, thus: . While upper-case letters are almost universally used for matrices themselves, some texts maintain the upper-case letter for the individual elements (e.g. ) while others use lower-case (e.g. ). Finally, in typesetting the matrix itself, some texts place large parentheses around the elements while others use large square brackets.

Operations

Several operations are defined for matrices.

Matrix addition

Two matrices may be added if and only if they have identical dimensions. The sum of the matrices is simply the matrix composed of the sums of the entries. That is, if A and B are m×n matrices, then A + B is an m×n matrix such that

for all i, j with and

For example:

Just as with numeric addition, matrix addition is commutative:

and associative:

Scalar multiplication

Any scalar may be multiplied by any matrix. To obtain the resultant matrix, multiply each entry of the original matrix by the scalar. That is, if c is a scalar and A is an m×n matrix, then cA is an m×n matrix such that

For example:

Matrix multiplication

Two matrices A and B may be multiplied if A has as many columns as B has rows. (Otherwise, they are said to be incompatible and their product is undefined.) That is, an m×n matrix may be multiplied by an n×p matrix. Then the resultant matrix AB is m×p and the (i,j)th entry of AB is the vector dot product of the ith row of A and the jth column of B. Formally:

For example:

Even if AB is defined, BA may not be. If both matrices are defined, they may have different dimensions. Even if they have the same dimensions, they may not be equal. Thus, matrix multiplication is clearly not commutative. It is, however, associative:

and left- and right-distributive:

and

so long as all the relevant products are defined.

Transposition

Given an m×n matrix A, its transpose (denoted ) is an n×m matrix where each row in A is a column in and vice versa. That is:

For example:

Note that the transpose operation is its own inverse; for all matrices A, .

Special matrices

Certain types of matrices prove useful in different contexts.

Square matrices

A square matrix, as the term implies, is any matrix of dimension n×n -- that is, with the same number of rows as columns. Two n×n matrices may always be multiplied, and their product is another n×n matrix.

Identity matrix

For more information, see: Identity matrix.

We denote by the multiplicative identity for matrix multiplication; that is, the matrix such that

for any m×n matrix A.

takes the form of a n×n square matrix with the number 1 down its main diagonal, starting from element (1, 1), and the number 0 everywhere else. So

, , , ...

In general, the n subscript is included only if necessary; if the size of the identity matrix can be deduced from context, we omit the subscript. For example, we would most likely say:

since only one identity matrix is dimensionally compatible for multiplication.

Zero matrix

For more information, see: Zero matrix.

The additive identity under matrix addition is known as a zero matrix, and denoted for an m×n matrix. Its entries are all zeroes, so (for example)

It is evident that for any m×n matrix A,

It is also clear that the product of any matrix with a zero matrix is another zero matrix, which may or may not have the same dimensions. As with the identity matrix, the subscript is omitted if the context admits only one zero matrix. In this example, any other zero matrix could not be added to A, so the subscript is redundant and we could say

Invertible matrix

For more information, see: Matrix inverse.

Some, but not all, matrices have a multiplicative inverse. That is, for a matrix A, there may exist a matrix such that

Only square matrices may be inverted. Furthermore, if the determinant of a square matrix is 0, it is singular -- that is, not invertible.

Some square matrices without inverses may have matrix "pseudoinverses", which have properties extending the concept inverses.

Symmetric matrix

A symmetric matrix is equal to its transpose. It must therefore be a square matrix, with values reflected across the main diagonal. That is, if A is an n×n matrix, A is symmetric if and only if

for all

For example:

Antisymmetric matrix

An antisymmetric or skew-symmetric matrix is the additive inverse of its transpose. It must also therefore be a square matrix. If A is an n×n matrix, A is antisymmetric if and only if

for all

Therefore, it is a requirement that all entries on the main diagonal of an antisymmetric matrix equal zero. For example:

Diagonal matrix

For more information, see: Diagonal matrix.

A matrix that has zeros everywhere other than the main (upper left to lower right) diagonal is a Diagonal matrix.

Example:

Diagonal matrices are especially easy to invert or operate with.

Scalar matrix

A diagonal matrix all of whose diagonal entries are equal is a Scalar matrix. Scalar matrices are so called because their multiplicative action on other matrices is the same as multiplication by a scalar element. A scalar matrix is just a scalar multiple of the identity matrix.

Example:

Triangular matrix

A matrix that has only zeros above (below) the main diagonal is called a lower (upper) triangular matrix. Examples:

Upper triangular:

Lower triangular:

Triangular matrices are especially easy to invert or operate with.

Orthogonal matrix

A matrix is Orthogonal if the transpose of the matrix is equal to its inverse.

Hermitian Matrix

For more information, see: Hermitian matrix.

A Hermitian matrix (or self-adjoint matrix) is one which is equal to its Hermitian adjoint (also known as its conjugate transpose). That is to say that every entry in the transposed matrix is replaced by its complex conjugate:
,
or in matrix notation:

Idempotent Matrix

A matrix is Idempotent if it is equal to its square.

An idempotent matrix P has eigenvalues 0 or 1 and has a basis of eigenvectors: it is diagonalisable since its minimal polynomial polynomial X2-X has no repeated roots. The kernel and image of P are complements: they form an internal direct sum.

Sparse Matrix

If a matrix has enough zero elements that one can take advantage of the fact, a matrix is called a "Sparse Matrix". Triangular matrices and Diagonal matrices are examples of sparse matrices

Applications

Systems of linear equations

Matrix techniques are often used to solve systems of equations in several variables, because any system of linear equations may be represented in matrix form. For example, the system

is equivalent to the equation

where the unknowns are entirely within the second matrix. Then, if the first matrix is invertible, x, y, and z can be recovered:

Solving linear systems is extensively in Physics, Mechanics, and many other fields.

Linear transformations

If f is a linear mapping from to , then there exists a unique m×n matrix F such that for any vector x in ,

These transformations have extensive use in Computer Graphics, and in mechanics.

Least Squares problems

Surveying problems with redundant (and not quite consistant) measurements were first studied by Gauss and he developed "Least Squares" to find solutions that were "close" to the systems the data described. The exact sense of "close" was that the solution had the lowest possible sum of the squares of the differences between the data at hand and the solution given. He pioneered the use of matrix "Normal Equations" for the problem, although today matrix "Orthogonalization" methods such as QR factorization are more commonly used.

Finite Element methods

Breaking an object into a mesh of points, and computing the interactions of those points, is used in a number of Engineering fields to model materials. Various matrix methods are used to deal with the problem.