Packet 12

Matrices III: Inner Products

In this packet we study some interrelated concepts: orthogonal matrices, symmetric matrices, inner products, quadratic forms, and positive definite matrices.

Orthogonal matrices

A matrix Q is called orthogonal when it satisfies A๐–ณA=In.

โ€˜Orthogonalโ€™ = โ€˜orthonormalโ€™ for matrices

The term orthonormal is also used for the same class of matrices. (Warning! Not so for vectors! Orthogonal vectors are merely perpendicular, while orthonormal vectors are also unit vectors.)

Now consider the meaning of the formula Q๐–ณQ=In. Remember the rule for matrix products (e.g. in 3D):

BA=(B๐š1B๐š2B๐š3).

What is Q๐–ณ๐ชi? Each row of this vector is the dot product of the same row of Q๐–ณ with the vector ๐ชi. In turn, each row of Q๐–ณ is from a column of Q. Combining all these claims with the product rule, column i of Q๐–ณQ has rows equal to the dot products of the various columns of Q with the specific column i of Q. Specifically, (Q๐–ณQ)ij=๐ชjโ‹…๐ชi. In conclusion, Q๐–ณQ=In means that ๐ชjโ‹…๐ชi={1i=j0iโ‰ j.

It follows that Q๐–ณQ=In is another way of saying that ๐ชj and ๐ชi are mutually orthonormal. So, if you put an orthonormal basis into the column vectors of a matrix, the resulting matrix is orthogonal. Conversely, if a matrix is orthogonal, then its columns form an orthonormal basis. (They are orthonormal by the definition of orthogonal, and they are a basis because they are perpendicular and thus independent.)

An important fact about orthogonal matrices is that they correspond geometrically to rotations and reflections (and combinations of these actions called โ€˜rotoreflectionsโ€™).

Orthogonal matrices preserve the dot product, preserving lengths and angles

The action of an orthogonal matrix preserves the dot product:

(Q๐ฎ)โ‹…(Q๐ฏ)=๐ฎโ‹…๐ฏ.

Why (Q๐ฎ)โ‹…(Q๐ฏ)=๐ฎโ‹…๐ฏ

To see this, recall (1) our interpretation of row vectors as 1ร—n matrices, and (2) the dot product with ๐ฑ as the action of the transpose row vector ๐ฑ๐–ณ acting by matrix multiplication on the left, so e.g. ๐ฑโ‹…๐ฐ=๐ฑ๐–ณ๐ฐ. Using this, we derive:

(Q๐ฎ)โ‹…(Q๐ฏ)=(Q๐ฎ)๐–ณ(Q๐ฏ)=(๐ฎ๐–ณQ๐–ณ)(Q๐ฏ)=๐ฎ๐–ณ(Q๐–ณQ)๐ฏ=๐ฎ๐–ณ(In)๐ฏ=๐ฎ๐–ณ๐ฏ=๐ฎโ‹…๐ฏ.

Symmetric matrices

A symmetric matrix A is a matrix that satisfies A๐–ณ=A. In terms of coefficients, this means it satisfies aij=aji. A symmetric matrix is symmetric about reflection in a mirror that runs down the main diagonal:

Non-symmetric:(1234),(12โˆ’21) Symmetric:(133โˆ’2),antisymmetric:(03โˆ’30)

An antisymmetric matrix (also called skew-symmetric) satisfies A๐–ณ=โˆ’A, in other words aij=โˆ’aji.

Symmetric matrices do arise naturally in many applications. This can happen for a variety of reasons. One example is PCA (principal component analysis): here the matrix in question arises as a โ€œcovariance matrix,โ€ and it is symmetric because covariance is symmetric: Cov(X,Y)=Cov(Y,X). Antisymmetric matrices are less common.

Symmetric matrices have some very nice properties:

Spectral theorem

Let A be any nร—n symmetric matrix, so A๐–ณ=A. Then:

  • (a) A has the max number of n eigenvalues (some might be repeated, but they still exist).
  • (b) All eigenvalues of A are real numbers.
  • (c) Eigenvectors with distinct eigenvalues are perpendicular to each other.
  • (d) There is an orthonormal basis ๐’ฌ={๐ช1,โ€ฆ,๐ชn} such that A becomes diagonal in this basis: D=Qโˆ’1AQ,ย whereย Q=(๐ช1โ‹ฏ๐ชn)=โ„ฌT๐’ฌ in other words A=QDQโˆ’1. Notice that Q is an orthogonal matrix, and ๐ชi are eigenvectors of A.

Proofs of most of these facts are just a little beyond the scope of a first applied course. Part (c) is not hard though:

Orthogonal eigenvectors of a symmetric matrix

Suppose A๐ช1=ฮป1๐ช1 and A๐ช2=ฮป2๐ช2 and ฮป1โ‰ ฮป2. Assume, as we do for eigenvectors, that ฮป1โ‰ 0. Then consider the sequence:

๐ช1โ‹…๐ช2=ฮป1โˆ’1(ฮป1๐ช1โ‹…๐ช2)=ฮป1โˆ’1(A๐ช1)โ‹…๐ช2=ฮป1โˆ’1(A๐ช1)๐–ณ๐ช2=ฮป1โˆ’1(๐ช1๐–ณA๐–ณ)๐ช2=ฮป1โˆ’1๐ช1๐–ณ(A๐ช2)=ฮป1โˆ’1๐ช1๐–ณ(ฮป2๐ช2)=(ฮป1โˆ’1ฮป2)๐ช1๐–ณ๐ช2=(ฮป1โˆ’1ฮป2)๐ช1โ‹…๐ช2.

Therefore, we find that ๐ช1โ‹…๐ช2=(ฮป1โˆ’1ฮป2)๐ช1โ‹…๐ช2. This can only happen if ฮป1=ฮป2, which we have denied by hypothesis, or else if ๐ช1โ‹…๐ช2=0, which must therefore be true.

The equation A=QDQ๐–ณ is equivalent to a very nice formula for A called the spectral decomposition formula:

A=ฮป1(๐ช1๐ช1๐–ณ)+ฮป2(๐ช2๐ช2๐–ณ)+โ‹ฏ+ฮปn(๐ชn๐ชn๐–ณ).

In this formula, the ๐ชi are the orthogonal basis given by the spectral theorem (remember they are also eigenvectors), and the ฮปi are the corresponding eigenvalues.

SVD prelude

The spectral decomposition is an instance of the singular value decomposition (SVD) which works for general matrices. Stay tuned for more in the next Packet!

A great thing about this formula is that each ๐ชi๐ชi๐–ณ is actually a projection matrix onto ๐ชi:

๐ชi๐ชi๐–ณ=[proj๐ชi].

Why? Of course ๐ชi๐ชi๐–ณ๐ฏ=๐ชi(๐ชiโ‹…๐ฏ). Since ๐ชi are unit vectors, we have proj๐ชi(๐ฏ)=(๐ฏโ‹…๐ชi)๐ชi. This is the same thing, just written with a different order for the scalar factor (result of the dot product).

Explaining the spectral decomposition formula

The spectral decomposition formula itself comes from writing out the matrices in A=QDQ๐–ณ:

A=QDQ๐–ณ=(๐ช1โ‹ฏ๐ชn)(ฮป10โ‹ฑ0ฮปn)(๐ช1๐–ณโ‹ฎ๐ชn๐–ณ)=(๐ช1โ‹ฏ๐ชn)(ฮป1๐ช1๐–ณโ‹ฎฮปn๐ชn๐–ณ)=(๐ช1โ‹ฏ๐ชn)((ฮป1๐ช1๐–ณ๐ŸŽโ‹ฎ๐ŸŽ)+(๐ŸŽฮป2๐ช2๐–ณโ‹ฎ๐ŸŽ)+(๐ŸŽ๐ŸŽโ‹ฎฮปn๐ชn๐–ณ))=๐ช1ฮป1๐ช1๐–ณ+โ‹ฏ+๐ชnฮปn๐ชn๐–ณ=ฮป1(๐ช1๐ช1๐–ณ)+ฮป2(๐ช2๐ช2๐–ณ)+โ‹ฏ+ฮปn(๐ชn๐ชn๐–ณ).

Quadratic forms and symmetric matrices

A quadratic form is a function Q:โ„nโ†’โ„ from vectors to scalars that can be written as a polynomial with all terms having degree 2. For example:

Inย 2D:Q((xy))=ax2+by2+cxyInย 3D:Q((xyz))=ax2+by2+cz2+dxy+exz+fyz

In short, a quadratic form is just a polynomial in several variables that is homogeneous of degree 2. (โ€˜Homogeneousโ€™ just means that all terms have the same degree.) The difference between quadratic forms and quadratic polynomials is that the latter could also have terms of degree zero and degree one.

Applications of quadratic forms

The simplest possible quadratic form is the sum of squares: Q(x,y,z)=x2+y2+z2. Notice that this form gives the square distance from (x,y,z) to (0,0,0). One strong application of linear algebra is the ability to extremize quadratric forms. For example, in Linear Least Squares, we extremize the distance from a point to a certain span. We will shortly see how to represent this distance as a quadratic form. In Weighted LLS, we minimize the distance with inequal weights applied to each term, such as in Q(x,y,z)=4x2+y2+9z2. This kind of weighting is useful when some data is more reliable than other data, so it should be given more weight.

Question 12-01

[Removed!]

Key fact:

Matrix of a quadratic form

Every quadratic form Q can be represented uniquely by a symmetric matrix, meaning there is some symmetric A such that Q(๐ฑ)=๐ฑ๐–ณA๐ฑ.

So there is a perfect correspondence between quadratic forms and symmetric matrices.

Example

Quadratic form

Q(๐ฑ)=๐ฑ๐–ณA๐ฑ=(x1x2x3)(5โˆ’1/20โˆ’1/234042)(x1x2x3)=5x12+3x22+2x32โˆ’x1x2+8x2x3.
Question 12-02

Matrix of a quadratic form

What is the matrix A such that Q(๐ฑ)=3x12+7x22โˆ’4x1x2 is the quadratic form Q(๐ฑ)=๐ฑ๐–ณA๐ฑ?

A quadratic form can always be converted into a new quadratic form that has no cross terms by a suitable linear change of variables. If the original variables are the components of ๐ฑ, then a linear change of variables would be a vector ๐ฒ defined by ๐ฒ=โ„ฌโ€ฒTโ„ฌ(๐ฑ) for some change of basis transfer matrix โ„ฌโ€ฒTโ„ฌ.

Principal axes theorem: eliminating cross terms

Given a quadratic form Q(๐ฑ) defined by the symmetric matrix A, suppose A=PDP๐–ณ. Set ๐ฒ=P๐–ณ๐ฑ, so also ๐ฑ=P๐ฒ. Then Q(P๐ฒ) has no cross terms in the ๐ฒ variables.

Derivation

First plug in the new variables and manipulate:

๐ฑ๐–ณA๐ฑ=(P๐ฒ)๐–ณA(P๐ฒ)=๐ฒ๐–ณ(P๐–ณAP)๐ฒ=๐ฒ๐–ณD๐ฒ

This last expression has no cross terms:

๐ฒ๐–ณD๐ฒ=(y1โ‹ฏyn)(ฮป10โ‹ฑ0ฮปn)(y1โ‹ฎyn)=ฮป1y12+โ‹ฏ+ฮปnyn2.

Example

Eliminating cross terms

Problem: Eliminate the cross terms in the quadratic form Q(๐ฑ)=2x12+6x1x2โˆ’6x22.

Solution: First compute the matrix of Q. By reading off the components (dividing in half the cross term), we see

A=(233โˆ’6).

Now we must write A=PDP๐–ณ for P an orthogonal matrix. We find the eigenvalues and then seek eigenvectors that are orthonormal. Observe that detโกAฮป=ฮป2+4ฮปโˆ’21=(ฮป+7)(ฮปโˆ’3), so the eigenvalues are ฮป=โˆ’7,3.

Then we have:

Aโˆ’7=(2โˆ’โˆ’733โˆ’6โˆ’โˆ’7)=(9331)

so 3x1+x2=0 gives the relation satisfied by the rows of an eigenvector. Choose (1โˆ’3) and then normalize this to the unit vector ๐ฏโˆ’7=(1/10โˆ’3/10).

Then we also have:

A3=(2โˆ’333โˆ’6โˆ’3)=(โˆ’133โˆ’9)

so โˆ’x1+3x2=0 and we choose (31) as an eigenvector, and normalize to the unit vector ๐ฏ3=(3/101/10). Notice that ๐ฏโˆ’7โ‹…๐ฏ3=0.

Now we set ๐ช1=๐ฏโˆ’7 and ๐ช2=๐ฏ3 and write our expression for A:

A=(233โˆ’6)=PDP๐–ณ=(1/103/10โˆ’3/101/10)(โˆ’7003)(1/10โˆ’3/103/101/10).

In terms of the ๐ฒ variables, we have Q(๐ฑ)=Q(P๐ฒ)=โˆ’7y12+3y22. The formulas relating the variables to each other are given by interpreting P๐–ณ as the change of variable transfer matrix:

y1=110x1โˆ’310x2y2=310x1+110x2.
Exercise 12-01

Eliminating cross terms.

Repeat the example to eliminate cross terms for the quadratic form Q(๐ฑ)=2x12โˆ’4x1x2โˆ’x22.

Constrained optimization, min-max principle

One simple yet important application of elimination of cross terms is constrained optimization. The goal of constrained optimization is to find the extreme values of a quadratic form Q(๐ฑ) where ๐ฑ is constrained to be a unit vector, so ๐ฑ๐–ณ๐ฑ=1. Geometrically, this means: find the direction of fastest increase or direction of fastest decrease of Q, and find the rate of increase / decrease in those directions.

If we have written Q(๐ฒ) so as to eliminate cross terms:

Q(๐ฒ)=ฮป1y12+โ‹ฏ+ฮปnyn2

then the max / min of Q occurs for ๐ฒ=๐ži where ฮปi is the max / min coefficient.

If Q(๐ฑ) is not written without cross terms, then we can first eliminate cross terms completely, and find the max / min among the set of ฮปi. This will give the max / min value, and the input where this occurs in terms of the y variables at ๐ฒ=๐ži. The change of variables must be made in reverse to find the corresponding x input values where the max / min occurs. If the eigenvector ๐ชi corresponding to ฮปi has already been found, its components are by definition the components of P๐ži that we seek.

Example

Optimizing Q(๐ฑ)=2x12+6x1x2โˆ’6x22

In the Example of the previous section on eliminating cross terms, we obtained the formula โˆ’7y12+3y22 for Q in terms of the variables y1,y2. We can see directly from this formula that the min eigenvalue is โˆ’7 and the max eigenvalue is 3.

The min of โˆ’7 occurs when y1=1,y2=0. In other words it is at the first eigenvector ๐ช1. In terms of the x1,x2 variables, this occurs at x1=1/10 and x2=โˆ’3/10, the components of ๐ช1=๐ฏโˆ’7 in the standard basis.

The max of 3 occurs when y1=0,y2=1. In other words at ๐ช2, which corresponds to x1=3/10 and x2=1/10.

Exercise 12-02

Optimizing Q(๐ฑ)=2x12โˆ’4x1x2โˆ’x22

Repeat the example on this quadratic form, using your work in Exercise 12-01, to find the min and max of Q(๐ฑ)=2x12โˆ’4x1x2โˆ’x22 for ๐ฑ a unit vector, and find the unit vectors that produce these extreme values.

However, it is not necessary to find all of the ฮปi. These ฮปi are the eigenvalues / eigenvectors, and there are some methods of computing eigenvalues that locate the largest / smallest values by a numerical procedure that does not waste time finding the other values. The power method is an example of such a method. (Read more about the power method online, if you like.)

Min-max principle and Rayleigh quotient

A closely related topic is the min-max principle for finding eigenvalues using the Rayleigh quotient. Given any symmetric matrix A, its Rayleigh quotient is:

R(A,๐ฑ)=๐ฑ๐–ณA๐ฑ๐ฑ๐–ณ๐ฑ.

Notice that the value of R(A,๐ฑ) is the same as the value of the quadratic form generated by A, evaluated at the unit vector given by first normalizing the input ๐ฑ. Therefore, maximizing R(A,๐ฑ) over all possible ๐ฑโ‰ 0 is the same as maximizing the quadratic form of A over all possible unit vectors.

The min-max principle is simply this: the min / max eigenvalues of A are the min / max values of R(A,๐ฑ), and these occur when ๐ฑ is an eigenvector for that eigenvalue. The use of this principle is to find min / max eigenvalues / eigenvectors of a symmetric A by extremizing R(A,๐ฑ) with numerical methods (like the power method and its variants).

The min-max principle and Rayleigh quotient is essentially just a notation and terminology for the process of normalizing ๐ฑ and plugging it into the quadratic form Q arising from the symmetric matrix A.

Positive definite symmetric matrices

  • A positive definite matrix is by definition a symmetric matrix all of whose eigenvalues are greater than zero.
  • A negative definite matrix is by definition a symmetric matrix all of whose eigenvalues are less than zero.
  • An indefinite matrix is by definition a symmetric matrix whose eigenvalues have mixed signs.

For example, in terms of spectral decompositions, we have:

Aย isย positiveย definite:A=7๐ช1๐ช1๐–ณ+๐ช2๐ช2๐–ณ+3๐ช3๐ช3๐–ณAย isย negativeย definite:A=โˆ’7๐ช1๐ช1๐–ณโˆ’๐ช2๐ช2๐–ณโˆ’3๐ช3๐ช3๐–ณAย isย indefinite:A=7๐ช1๐ช1๐–ณโˆ’๐ช2๐ช2๐–ณ+3๐ช3๐ช3๐–ณ.

Since the spectral decomposition reveals the quadratic form in the new variables, where ๐ชi=P๐ži, but these input variables cover the same set of possibilities, we also know:

Quadratic forms definiteness

  • When A is positive definite, then its form Q(๐ฑ) is always positive for ๐ฑโ‰ ๐ŸŽ.
  • When A is negative definite, then its form Q(๐ฑ) is always negative for ๐ฑโ‰ ๐ŸŽ.

Inner products and symmetric matrices

An inner product is a function sending two vector inputs to one scalar output:

โŸจโˆ’,โˆ’โŸฉ:(๐ฎ,๐ฏ)โ†ฆโŸจ๐ฎ,๐ฏโŸฉโˆˆโ„,

which is linear in both inputs considered separately.

Rules of inner products:

  • โŸจ๐ฎ,๐ฏโŸฉ=โŸจ๐ฏ,๐ฎโŸฉ Symmetry.
  • โŸจ๐ฎ1+๐ฎ2,๐ฏโŸฉ=โŸจ๐ฎ1,๐ฏโŸฉ+โŸจ๐ฎ2,๐ฏโŸฉ Respects sums (also in ๐ฏ using symmetry).
  • โŸจฮป๐ฎ,๐ฏโŸฉ=ฮปโŸจ๐ฎ,๐ฏโŸฉ Respects scalars (also in ๐ฏ using symmetry).
  • โŸจ๐ฎ,๐ฎโŸฉ>0 for all ๐ฎโ‰ ๐ŸŽ, and โŸจ๐ฎ,๐ฎโŸฉ=0 only occurs when ๐ฎ=๐ŸŽ.

We are very familiar with the dot product, which is a valid inner product:

โŸจ๐ฎ,๐ฏโŸฉdot=๐ฎโ‹…๐ฏ=(u1u2u3)โ‹…(v1v2v3)=๐ฎ๐–ณ๐ฏ=(u1u2u3)(v1v2v3)=u1v1+u2v2+u3v3.

Using the linearity of matrix actions, we can create many candidate inner products by first acting by a symmetric matrix A=A๐–ณ and then taking the usual dot product:

โŸจ๐ฎ,๐ฏโŸฉA=โŸจ๐ฎ,A๐ฏโŸฉdot=๐ฎโ‹…A๐ฏ.

This definition of โŸจ๐ฎ,๐ฏโŸฉA is automatically symmetric (i.e. it satisfies the symmetry rule) because A is symmetric:

โŸจ๐ฎ,๐ฏโŸฉA=๐ฎ๐–ณA๐ฏ=(๐ฎ๐–ณA๐ฏ)๐–ณ(ฮป๐–ณ=ฮปย forย scalars)=๐ฏ๐–ณA๐–ณ(๐ฎ๐–ณ)๐–ณ=๐ฏ๐–ณA๐ฎ=โŸจ๐ฏ,A๐ฎโŸฉA.

If you take any inner product given by a symmetric matrix, and plug the input ๐ฑ into both slots, you get a quadratic form:

โŸจ๐ฑ,๐ฑโŸฉA=๐ฑ๐–ณA๐ฑ=Q(๐ฑ).

The last property of inner products (namely that โŸจ๐ฎ,๐ฎโŸฉ>0 etc.) means that a quadratic form created in this way by setting Q(๐ฑ)=โŸจ๐ฑ,๐ฑโŸฉA is always a positive definite quadratic form.

Conversely, any inner product is given by some positive definite symmetric matrix. To find the matrix corresponding to an inner product, simply set aij=โŸจ๐ži,๐žjโŸฉ. This works because A๐žj gives the jth column vector of A, and then ๐ži๐–ณA๐žj gives the ith row entry of this column, which is therefore the ith row and jth column entry of A. By linearity, โŸจโ‹…,โ‹…โŸฉ is then determined on every other pair of vectors by this data.

Addendum: determinants and eigenvectors

The determinant of a 2ร—2 matrix is given by the formula:

A=(abcd),detโกA=|abcd|=adโˆ’bc.

The determinant of a 3ร—3 matrix is given by the formula:

A=(abcdefghi),detโกA=|abcdefghi|=a|efhi|โˆ’b|dfgi|+c|degh|.

Higher determinants

You may notice a pattern here that hold true in general and allows calculation of nร—n determinants: traverse any row or column, using these entries as weights (with alternating signs), writing a weighted sum of the smaller determinants of the minor matrices (which are given by deleting the row and column in which the given entry in our traversal process lives). This process therefore reduces an nร—n determinant to a linear combination of (nโˆ’1)ร—(nโˆ’1) determinants, and the reduction can be iterated until the 2ร—2 case is reached.

The characteristic polynomial chA(ฮป) of a matrix A is defined as the determinant detโกAฮป in which ฮป is considered an indefinite variable.

For example: if A=(1โˆ’2โˆ’4โˆ’1) then Aฮป=(1โˆ’ฮปโˆ’2โˆ’4โˆ’1โˆ’ฮป) and (using the determinant formula) we have:

chA(ฮป)=|1โˆ’ฮปโˆ’2โˆ’4โˆ’1โˆ’ฮป|=(1โˆ’ฮป)(โˆ’1โˆ’ฮป)โˆ’8=ฮป2โˆ’9.

Key fact for finding eigenvalues

The eigenvalues of a matrix A are the roots of its characteristic polynomial chA(ฮป).

In the previous example, therefore, ฮป=ยฑ3 are the eigenvalues.

Once the eigenvalues are found by solving chA(ฮป)=0 for ฮป, the solutions can be plugged back in for ฮป in the equation Aฮป๐ฑ=๐ŸŽ, and then we can try to find the eigenvectors ๐ฑ by solving this equation. For the 2ร—2 case, you can solve it using the fact that one row is a multiple of the other. In 3ร—3 and higher, you may need to perform row reduction on Aฮป (with ฮป=specificย eigenvalue) in order to find solutions ๐ฑ.

Notice that if ๐ฑ is an eigenvector, then ฮฑ๐ฑ is also an eigenvector for any scalar ฮฑโ‰ 0. So really we are studying eigenlines, namely the spans of eigenvectors. The relation of proportionality between x1 and x2 gives the direction of this line.

For ฮป=+3 in the previous example, solving Aฮป๐ฑ=๐ŸŽ would mean solving (โˆ’2โˆ’2โˆ’4โˆ’4)(x1x2)=(00). Therefore โˆ’x1โˆ’x2=0 and thus x1=โˆ’x2, so an eigenvector is given by (1โˆ’1) and a line of eigenvectors is given by the span ฮฑ(1โˆ’1)=(ฮฑโˆ’ฮฑ) for all ฮฑ. For ฮป=โˆ’3 we find (4โˆ’2โˆ’42)(x1x2)=(00), and solving this we find 2x1=x2 and therefore (12) is an eigenvector, as is everything in the span (ฮฑ2ฮฑ).

Problems due 10 Apr 2024 by 12:00pm

Exercises for submission with HW (if not in class)

Quadratic form: eliminating cross terms and optimizing

  • (a) Eliminate cross terms in the quadratic form Q(๐ฑ)=2x12โˆ’4x1x2โˆ’x22.
  • (b) Find the min and max values of Q(๐ฑ) from (a) where ๐ฑ is a unit vector, and find the unit vectors ๐ฑ that produce these extreme values. (Use your result in (a) to minimize in changed variables, and then change variables back again.)
Problem 12-01

Finding a spectral decomposition

Let A=(422242224).

  • (a) Find the eigenvalues of A. There will be two distinct numbers, ฮป1,ฮป2, not three.
  • (b) Row reduce Aฮป1 and Aฮป2 to solve the two equations Aฮป1๐ฑ=๐ŸŽ and Aฮป2๐ฑ=๐ŸŽ. In this way, write down a set of 3 eigenvectors, two of which have one of the eigenvalues, and the third has the other eigenvalue.
  • (c) For the โ€˜doubleโ€™ eigenvalue, convert its two eigenvectors to orthogonal vectors (subtract a projection like in Gram-Schmidt, unless they are already perpendicular).
  • (d) Normalize all your orthogonal eigenvectors to get unit eigenvectors. Call them ๐ช1,๐ช2,๐ช3. Now the matrix Q=(๐ช1๐ช2๐ช3) is an orthogonal matrix, its inverse is Qโˆ’1=Q๐–ณ, and we know that A=QDQ๐–ณ where D has the eigenvectors ฮป1,ฮป2,ฮป3 along the diagonal. (In order, so A๐ชi=ฮปi๐ชi for i=1,2,3.)
  • (e) Write out the spectral decomposition: A=ฮป1๐ช1๐ช1๐–ณ+ฮป2๐ช2๐ช2๐–ณ+ฮป3๐ช3๐ช3๐–ณ. In other words, simply compute the three projection matrices ๐ชi๐ชi๐–ณ, and plug them into this formula. Check that if you were to take the sum, you would get A.
Problem 12-02

Quadratic form

Let A be the matrix as in Problem 12-01. Use results of 12-01 for this problem: you do not need to find the eigenvalues / eigenvectors again.

  • (a) What is the quadratic form Q(๐ฑ) associated to A?
  • (b) What is the quadratic form Q(๐ฒ) in the variables ๐ฒ that eliminate cross terms?
  • (c) What is the max and min of Q(๐ฑ) for |๐ฑ|=1?
  • (d) What are the vectors ๐ฑ that yield the max and min found in (d)?
Problem 12-03

Functions of symmetric matrices

  • (a) Write out in coefficients the matrix A=7๐ช1๐ช1๐–ณ+๐ช2๐ช2๐–ณ+3๐ช3๐ช3๐–ณ when ๐ชi=๐ži.
  • (b) Compute the matrix product A2 when ๐ชi=๐ži. (Notice that you have found B=A2.)
  • (c) Write out in coefficients the matrix A=49๐ช1๐ช1๐–ณ+๐ช2๐ช2๐–ณ+9๐ช3๐ช3๐–ณ when ๐ชi=๐ži.
  • (d) Guess the coefficients ฮป1,ฮป2,ฮป3 in B=ฮป1๐ช1๐ช1๐–ณ+ฮป2๐ช2๐ช2๐–ณ+ฮป3๐ช3๐ช3๐–ณ such that B2=A for A as in (a). Verify your guess by writing out B in coefficients and then calculating B2. (Notice that you have found B=A.)
  • (e) Now drop the assumption that ๐ชi=๐ži, and yet find A2 by FOILing the product (7๐ช1๐ช1๐–ณ+๐ช2๐ช2๐–ณ+3๐ช3๐ช3๐–ณ)(7๐ช1๐ช1๐–ณ+๐ช2๐ช2๐–ณ+3๐ช3๐ช3๐–ณ). To simplify the result, figure out what ๐ชi๐ชi๐–ณ๐ชj๐ชj๐–ณ is for i=j and iโ‰ j.
  • (f) Based on (a)-(e), can you guess what ฮป1,ฮป2,ฮป3 should be in the matrix eA? (Hint: consider that ex has a power series expansion: ex=1+x+x22!+x33!+x44!+โ‹ฏ.)
Problem 12-04

Positive definite matrices

Suppose A๐–ณ=A and Aโˆ’1 exists. (Symmetric and invertible.) Notice that the Spectral Theorem applies.

  • (a) What are the eigenvalues of Aโˆ’1 in terms of the eigenvalues of A?
  • (b) Suppose that Q(๐ฑ)=๐ฑ๐–ณA๐ฑ is positive definite. Show that Qโ€ฒ(๐ฑ)=๐ฑ๐–ณAโˆ’1๐ฑ is also positive definite. (Hint: make use of (a).)