Notes on Mathematical Methods for Physicists Chapter2

本文最后更新于 2024年5月22日 晚上

Notes on Mathematical Methods for Physicists §2

Notes on Mathematical Methods for Physicists

Chapter2 Determinants & Matrices

Determinants

We begin the study of matrices by solving linear equations that will lead us to determinants and matrices. The concept of determinant and the notation were introduced by the renowned German mathematician and philosopher.

Homogeneous Linear Equations

Suppose three unknowns(orequations withunknowns) :

The problem is to determine under what conditions there is any solution , apart from the trivial one. Using vectors , we have. These three vector equations have the geometrical intepretation thatis orthogonal to.

If the volume spanned bygiven by determinant (or triple scalar product)

is not zero , then there is the only trivial solution.

Conversely , if the aforementional determinant of the coefficient vanishes , then one of the row vectors is a combination of the other two., only ratios ofare relevant.

This is Cramer's Rule for homogeneous linear equation.

Inhomogeneous Linear Equation

Simple example :

This is Cramer's Rule for inhomogeneous linear equation.

Definitions

Before defining a determinant , we need to introduce some related concepts and definitions.

When we write two-dimensional () arrays of items , we identify the item in theth horizontal row and theth vertical column by the index set; note that the row index is conventionally written first.

Starting from a set ofobjects in some reference order (e.g. , the number sequence) , we can make a permutation of them to some other order ; the total number of distinct permutations that are possible is(choose the first objectways , then choose the second inways , etc.).

Every permutation ofobjects can be reached from the reference order by a succession of pairwise interchanges (e.g. ,can be reached by the successive steps). Although the number of pairwise interchanges needed for a given permutation depends on the path (compare the above example with) , for a given permutation the number of interchanges will always either be even or odd. Thus a permutation can be identified as having either even or odd parity.

It is convenient to introduce the Levi-Civita symbol , which for an-object system is denoted by, wherehassubscripts , each of which identifies one of the objects.

We now define a determinant of orderto be ansquare array of numbers (or functions) , with the array conventionally written within vertical bars (not parentheses , braces , or any other type of brackets) , as follows :

The determinanthas a value.

Properties of Determinants

Take determinants of orderfor example.

Any determinant with two rows equal , or two columns equal , has the value zero. To prove this , interchange the two identical rows or columns ; the determinant both remains the same and changes sign , and therefore must have the value zero.

An extension of the above is that if two rows (or columns) are proportional , the determinant is zero.

The value of a determinant is unchanged if a multiple of one row is added (column by column) to another row or if a multiple of one column is added (row by row) to another column.

If each element in a row or each element in a column is zero , the determinant has the value zero.

Laplacian Development by Minor

The fact that a determinant of orderexpands intoterms means that it is important to identify efficient means for determinant evaluation. One approach is to expand in terms of minors. The minor corresponding to, denoted, orif we need to identifyas coming from the, is the determinant (of order) produced by stiking out rowand columnof the original determinant. And we get

Linear Equation Systems

For equation

We define

Then we have

This is the Cramer's Rule.

Ifis nonzero , the above construction of theis definitive and unique , so that there will be exactly one solution to the equation set.

Determinants & Linear Dependence

If the coefficients oflinear forms invariables form a nonzero determinant , the forms are linearly independent ; if the determinant of the coefficients is zero , the forms exhibit linear dependence.

Linearly Dependent Equations

Situation

All the equations are homogeneous (which means all the right hand side quantitiesare zero). Then , one or more of the equations in the set will be equivalent to linear combinations of others , and we will have less thanequations in ourvariables. We can then assign one (or in some cases , more than one) variable an arbitrary value , obtaining the others as functions of the assigned variables. We thus have a manifold (i.e. , a parameterized set) of solutions to our equation system.

Situation

A second case is where we have (or combine equations so that we have) the same linear form in two equations , but with different values of the right-hand quantities. In that case the equations are mutually inconsistent , and the equation system has no solution.

Situation

A third , related case , is where we have a duplicated linear form , but with a common value of. This also leads to a solution manifold.

Numerical Evaluation

There are many methods to evaluate determinants , even using computers. We use the Gauss Elimination to calculate determinants , which is a versatile procedure that can be used for evaluating determinants, for solving linear equation systems, and (as we will see later) even for matrix inversion.

Gauss Elimination : make the determinant into a form that all the entries in the lower triangle of the determinant. Then the only effective part is the product of thediagonal elements.

Matrices

Matrices arearrays of numbers or functions that obey the laws that define matrix algebra.

Basic Definitions

A matrix is a set of numbers or functions in asquare or rectangular array. A matrix with(horizontal) rows and(vertical) columns is known as anmatrix. When we introduced determinants , when row and column indices or dimensions are mentioned together , it is customary to write the row indicaters first.

A matrix for whichis termed square; One consisting of a single column (anmatrix) is often called a column vector , while a matrix with only one row (therefore) is a row vector.

Equality

Ifandare matrices ,only iffor all values ofand. A necessary but not sufficient condition for equality is that both matrices have the same dimensions.

Addition , Subtraction

Addition and subtraction are defined only for matricesandof the same dimensions , in which case, withfor all values ofand. Addition is commutative () and also associative (). A matrix with all elements zero , called a null matrix or zero matrix , can either be written asor as a simple zero. Thus for all,

Multiplication (by a Scalar)

Here we have, withfor all values ofand. This operation is commutative , with.

Note that the definition of multiplication by a scalar causes each element of marixto be multiplied by the scalar factor , so there is

Matrix Multiplication (Inner Product)

Matrix multiplication is not an element-by-element operation like addition or multiplication by a scalar. The inner product of matricesandis defined as

This definition causes theelement ofto be formed from the entireth row ofand the entireth column of. And as you can realize ,.

It is useful to define the commutator ofand,

which , as stated above , will in many cases be nonzero.

But , matrix multiplication is associative , meaning that.

Unit Matrix

By direct matrix multiplication , it is possible to show that a square matrix with elements of value unity on its principal diagonal (the elementswith) , and zeros everywhere else , will leave unchanged any matrix with which it can be multiplied. For example , theunit matrix has the form

note that it is not a matrix all of whose elements are unity. Giving such a matrix the name,

Remember thatmust be.

The previously introduced null matrices have only zero elements , so it is also obvious that for all,

 

Diagonal Matrices

If a matrixhas nonzero elementsonly for, it is said to be diagonal.

Matrix Inverse

It will often be the case that given a square matrix, there will be a square matrixsuch that. A matrixwith this property is called the inverse ofand is given the name. Ifexists , it must be unique.

Every nonezero real (or complex) numberhas a nonzero multiplicative inverse , often written. But the corresponding property does not hold for matrices ; there exist nonzero matrices that do not have inverses. To demonstrate this , consider the following :

Ifhas an inverse , we can multiply the equationon the left by, thereby obtaining

Since we started with a matrixthat was nonzero , this is an inconsistency , and we are forced to conclude thatdoes not exist. A matrix without an inverse is said to be singular , so our conclusion is thatis singular. Note that in our derivation , we had to be careful to multiply both members offrom the left , because multiplication is noncommutative. Alternatively , assumingto exist , we could multiply this equation on the right by, obtaining

This is inconsistent with the nonzerowith which we started ; we conclude thatis also singular. Summerizing , there are nonzero matrices that do not have inverses and are identified as singular.

The algebraic properties of real and complex numbers (including the existence of inverses for all nonzero numbers) define what mathematicians call a field. The properties we have identified for matrices are different ; they form what is called a ring.

A closed , but cumber-some formula for the inverse of a matrix exists ; it expresses the elements ofin terms of the determinants that are the minors of. That formula , the derivation of which is in several of the Additional Readings , is

We describe here a well-known method that is computationally more efficient than the equation above , namely the Gauss-Jordan procedure.

Example Gauss-Jordan Matrix Inversion

The Gauss-Jordan method is based on the fact that there exist matricessuch that the productwill leave an arbitrary matrixunchanged , except with ​ (a) one row multiplied by a constant , or ​ (b) one row replaced by the original row minus a multiple of another row , or ​ (c) the interchange of two rows.

By using these transformations , the rows of a matrix can be altered (by matrix multiplication) in the same way as we did to the elements of determinants. Ifis nonsingular , we change both side of the equation to reduceto, then we get:

What we need to do is to find out how to reducetousing the method which we have used to determinants. Here is a concrete example :

Write , side by side , the matrixand a unit matrix of the same size , and perform the same operations on each untilhas been converted to a unit matrix , which means that the unit matrix will have been changed into:

Multiply the rows as necessary to set to unity all elments of the first column of the left matrix ,

Subtracting the first row from the second an third rows , we obtain

Divide the second row byand subtracttimes it from the first row andtimes it from the third row ,

Divide the third row by. Then as the last step ,times the third row is subtracted from each of the first two rows. Our final pair is

Derivatives of Determinants

The formula giving the inverse of a matrix in terms of its minors enables us to write a compact formula for the derivative of a determinantwhere the matrixhas elements that depend on some variable. To carry out the differentiation with respect to thedependence of its element, we writeas its expansion in minorsabout the elements of row, so we have

Applying now the chain rule to allow for thedependence of all elements of, we get

 

Systems of Linear Equations

Note that ifis asquare matrix , andandarecolumn vectors. Consider the matrix equation, which is equivalent to a system oflenear equations.

This tells us two things : ​ (a) that if we can evaluate, we can compute the solution; ​ (b) and that the existence ofmeans that this equation system has a unique solution.

Then the result is important enough to be emphasized : A square matrixis singular if and only if.

Determinant Product Theorem

The Product Theorem is that. The proof of this theorem is dull , I would just skip it.

Note that.

Rank of a Matrix

The concept of a matrix singularity can be refined by introducing the notion of the rank of a matrix. If the elements of a matrix are viewed as the coefficients of a set of linear forms , a square matrix is assigned a rank equal to the number of linearly independent forms that its elements describe. Thus , a nonsingularmatrix will have rank, while asingular matrix will have a rankless. The rank provides a measure of the extent of the singularity ; if, the matrix describes one linear form that is dependent on the others , etc. Further and more systematical discussion will be seen in Chapter 6.

Transpose , Adjoint , Trace

Transpose

The transpose of a matrix is the matrix that results from interchanging its row and column indices. This operation corresponds to subjecting the array to reflection about its principal diagonal. If a matrix is not square , its transpose will not even have the same shape as the original matrix. The transpose of, denotedor sometimes, thus has elements

Note that transposition will convert a column vector into a row vector. A matrix that is unchanged by transposition is called symmetric.

Adjoint

The adjoint of a matrix, denoted, is obtained by both complex conjugating and transposing it. Thus ,

 

Trace

The trace , a quantity defined for square matrices , is the sum of the elements on the principal diagonal. Thus , for anmatrix,

Some properties of the trace :

The second property holds even if, which meansandare not commute.

Considering the trace of the matrix product, if we group the factors as, we easily see that

Repeating this process , we also find. Note , however , that we cannot equate any of these quantities toor to the trace of any other noncyclic permutation of these matrces.

Operations on Matrix Products

There are some properties of the determinant and trace :

whether or notandcommute. The properties above establish that the trace is a linear operator. Since similar relations do not exist for the determinant , it is not a linear operator.

For other operations on matrix products , there are

 

Matrix Representation of Vectors

 

I have nothing to say , because it is easy to understand. (I am going to useto represent vectors , and may sometimes useas well.)

Orthogonal Matrices

A real matrix is termed orthogonal if its transpose is equal to its inverse. Thus , ifis orthogonal , we may write

Since , fororthogonal ,, we see that. It is easy to prove that ifandare each orthogonal , then so areand.

Unitary Matrices

The definition is matrix which the adjoint is also the inverse is identified as unitary. One way of expressing this relationship is

If all the elements of a unitary matrix are real , the matrix is also orthogonal.

Since for any matrix, and therefore, application of the determinant product theorem to a unitary matrixleads to

We observe that ifandare both unitary , thenandwill be unitary as well. This is a generalization of earlier result that the matrix product of two orthogonal matrices is also orthogonal.

Hermitian Matrices

A matrix is identified as Hermitian , or , synonymously , self-adjoint , if it is equal to its adjoint. To be self-adjoint , a matrixmust be square , and in addition , its elements must satisfy

We see that the principal diagonal elements must be real.

Note that if two matricesandare Hermitian , it is not necessarily true thatoris Hermitian ; however ,, if nonzero , will be Hermitian , and, if nonzero , will be anti-Hermitian , meaning that.

Extraction of a Row or Column

It is useful to define column vectorswhich are zero except for the