Jump to content

Linear Algebra/Introduction to Matrices and Determinants

From Wikibooks, open books for an open world
(Redirected from Linear Algebra/Introduction to Determinants)

The determinant is a function which associates to a square matrix an element of the field on which it is defined (commonly the real or complex numbers).

Matrices

[edit | edit source]
Organization of a matrix

Informally an m×n matrix (plural matrices) is a rectangular table of entries from a field (that is to say that each entry is an element of a field). Here m is the number of rows and n the number of the columns in the table. Those unfamiliar with the concept of a field, can for now assume that by a field of characteristic 0 (which we will denote by F) we are referring to a particular subset of the set of complex numbers.

An m×n matrix (read as m by n matrix), is usually written as:

The row is an element of , showing the n components . Similary the column is an element of showing the m components .

Here m and n are called the dimensions of the matrix. The dimensions of a matrix are always given with the number of rows first, then the number of columns. It is also said that an m by n matrix has an order of m×n.

Formally, an m×n matrix M is a function where A = {1,2...m} × {1,2...n} and F is the field under consideration. It is almost always better to visualize a matrix as a rectangular table (or array) then as a function.

A matrix having only one row is called a row matrix (or a row vector) and a matrix having only one column is called a column matrix (or a column vector). Two matrices of the same order whose corresponding entries are equal are considered equal. The (i,j)-entry of the matrix (often written as or ) is the element at the intersection of the row (from the top) and the column (from the left).

For example,

is a 3×3 matrix (said 3 by 3). The 2nd row is and the 3rd column is . The (2,3) entry is the entry at intersection of the 2nd row and the 3rd column, that is 11.

Some special kinds of matrices are:

  • A square matrix is a matrix which has the same number of rows and columns. A diagonal matrix is a matrix with non zero entries only on the main diagonal (ie at positions).
  • The unit matrix or identity matrix In, is the matrix with elements on the diagonal set to 1 and all other elements set to 0. Mathematically, we may say that for the identity matrix (which is usually written as and called Kronecker's delta) is given by:

For example, if n = 3:

  • The transpose of an m-by-n matrix A is the n-by-m matrix AT formed by turning rows into columns and columns into rows, i.e. . An example is
  • A square matrix whose transpose is equal to itself is called a symmetric matrix; that is, A is symmetric if . An example is
  • A square matrix whose transpose is equal to its negative is called skew-symmetric matrix; that is, A is skew-symmetric if . An example is

Properties of these matrices are developed in the exercises.

Determinants

[edit | edit source]

To define a determinant of order n, suppose there are n2 elements of a field sij where i and j are less than or equal to n. Define the following function (this function is important in the definition):

S(a1,a2,a3,...,an)=# of reversals, meaning the number of times an1<an2 when n1>n2, for each possible combination.

Suppose you have a permutation of numbers from 1 to n {a1,a2,a3,...,an). Then define a term of the determinant to be equal to (-1)S(a1,a2,a3,...,an)s1a1,s2a2,s3a3,...,snan. The sum of all possible terms (i. e. through all possible permutations) is called the determinant.

Theorem

[edit | edit source]

Definition: The transpose of a matrix A, AT is the matrix resulting when the columns and rows are interchanged i. e. the matrix sji when A is the matrix sij A matrix and its transpose have the same determinant:

Proof

[edit | edit source]

All terms are the same, and the signs of the terms are also unchanged since all reversals remain reversals. Thus, the sum is the same.

Theorem

[edit | edit source]

Interchanging two rows (or columns) changes the sign of the determinant:

.

Proof

[edit | edit source]

To show this, suppose two adjacent rows (or columns) are interchanged. Then any reversals in a term would not be affected except for the reversal of the elements of that term within that row (or column), in which case adds or subtracts a reversal, thus changing the signs of all terms, and thus the sign of the matrix. Now, if two rows, the ath row and the (a+n)th are interchanged, then interchange successively the ath row and (a+1)th row, and then the (a+1)th row and (a+2)th row, and continue in this fashion until one reaches the (a+n-1)th row. Then go backwards until one goes back to the ath row. This has the same effect as switching the ath row and the (a+n)th rows, and takes n-1 switches for going forwards, and n-2 switches for going backwards, and their sum must then be an odd number, so it multiplies by -1 an odd number of times, so that its total effect is to multiply by -1.

Corollary

[edit | edit source]

A determinant with two rows (or columns) that are the same has the value 0. Proof: This determinant would be the additive inverse of itself since interchanging the rows (or columns) does not change the determinant, but still changes the sign of the determinant. The only number for which it is possible is when it is equal to 0.

Theorem

[edit | edit source]

It is linear on the rows and columns of the matrix.

Proof

[edit | edit source]

The terms are of the form a1......an. Using the distributive law of fields, this comes out to be a1......an + a1......an, an thus its sum of such terms is the sum of the two determinants:

Corollary

[edit | edit source]

Adding a row (or column) times a number to another row (or column) does not affect the value of a determinant.

Proof

[edit | edit source]

Suppose you have a determinant A with the kth column added by another column times a number:

where akb are elements of another column. By the linear property, this is equal to

The second number is equal to 0 because it has two columns that are the same. Thus, it is equal to

which is the same as the matrix A.



  • It is easy to see that and thus
for all -by- matrices and all scalars .
  • A matrix over a commutative ring R is invertible if and only if its determinant is a unit in R. In particular, if A is a matrix over a field such as the real or complex numbers, then A is invertible if and only if det(A) is not zero. In this case we have

Expressed differently: the vectors v1,...,vn in Rn form a basis if and only if det(v1,...,vn) is non-zero.

The determinants of a complex matrix and of its conjugate transpose are conjugate: