Skip to main content

Subsection 9.2.5 Diagonalizing a matrix

fit width

The algebraic eigenvalue problem or, more simply, the computation of eigenvalues and eigenvectors is often presented as the problem of diagonalizing a matrix. We make that link in this unit.

Definition 9.2.5.1. Diagonalizable matrix.

A matrix ACm×m is said to be diagonalizable if and only if there exists a nonsingular matrix X and diagonal matrix D such that X1AX=D.

fit width

Why is this important? Consider the equality w=Av. Notice that we can write w as a linear combination of the columns of X:

w=X(X1w)˜w.

In other words, X1w is the vector of coefficients when w is writen in terms of the basis that consists of the columns of X. Similarly, we can write v as a linear combination of the columns of X:

v=X(X1v)˜v.

Now, since X is nonsingular, w=Av is equivalent to X1w=X1AXX1v, and hence X1w=D(X1v).

Remark 9.2.5.2.

We conclude that if we view the matrices in the right basis (namely the basis that consists of the columns of X), then the transformation w:=Av simplifies to ˜w:=D˜v. This is a really big deal.

How is diagonalizing a matrix related to eigenvalues and eigenvectors? Let's assume that X1AX=D. We can rewrite this as

AX=XD

and partition

A(x0x1xm1)=(x0x1xm1)(δ0000δ1000δm1).

Multiplying this out yields

(Ax0Ax1Axm1)=(δ0x0δ1x1δm1xm1).

We conclude that

Axj=δjxj

which means that the entries on the diagonal of D are the eigenvalues of A and the corresponding eigenvectors are found as columns of X.

Homework 9.2.5.1.

In Homework 9.2.3.3, we computed the eigenvalues and corresponding eigenvectors of

A=(237011002).

Use the answer to that question to give a matrix X such that X1AX=Λ. Check that AX=XΛ.

Solution

The eigenpairs computed for Homework 9.2.3.3 were

\begin{equation*} ( -2, \left( \begin{array}{c} 1 \\ 0 \\ 0 \end{array} \right) ) , ( 1, \left( \begin{array}{r} 1 \\ 1 \\ 0 \end{array} \right) ) , \mbox{ and } ( 2, \left( \begin{array}{r} -1 \\ 1 \\ 1 \end{array} \right) ). \end{equation*}

Hence

\begin{equation*} \left(\begin{array}{rrr} 1 \amp 1 \amp -1 \\ 0 \amp 1 \amp 1 \\ 0 \amp 0 \amp 1 \\ \end{array}\right)^{-1} \left(\begin{array}{rrr} -2 \amp 3 \amp -7 \\ 0 \amp 1 \amp 1 \\ 0 \amp 0 \amp 2 \\ \end{array}\right) \left(\begin{array}{rrr} 1 \amp 1 \amp -1 \\ 0 \amp 1 \amp 1 \\ 0 \amp 0 \amp 1 \\ \end{array}\right) = \left(\begin{array}{rrr} -2 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 2 \\ \end{array}\right) . \end{equation*}

We can check this:

\begin{equation*} \begin{array}[t]{c} \underbrace{ \left(\begin{array}{rrr} -2 \amp 3 \amp -7 \\ 0 \amp 1 \amp 1 \\ 0 \amp 0 \amp 2 \\ \end{array}\right) \left(\begin{array}{rrr} 1 \amp 1 \amp -1 \\ 0 \amp 1 \amp 1 \\ 0 \amp 0 \amp 1 \\ \end{array}\right) } \\ \left(\begin{array}{rrr} -2 \amp 1 \amp -2 \\ 0 \amp 1 \amp 2 \\ 0 \amp 0 \amp 2 \\ \end{array}\right) \end{array} = \begin{array}[t]{c} \underbrace{ \left(\begin{array}{rrr} 1 \amp 1 \amp -1 \\ 0 \amp 1 \amp 1 \\ 0 \amp 0 \amp 1 \\ \end{array}\right) \left(\begin{array}{rrr} -2 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 2 \\ \end{array}\right) . } \\ \left(\begin{array}{rrr} -2 \amp 1 \amp -2 \\ 0 \amp 1 \amp 2 \\ 0 \amp 0 \amp 2 \\ \end{array}\right) \end{array} \end{equation*}

Now assume that the eigenvalues of ACm×m are given by {λ0,λ1,,λm1}, where eigenvalues are repeated according to their algebraic multiplicity. Assume that there are m linearly independent vectors {x0,x1,,xm1} such that Axj=λjxj. Then

A(x0x1xm1)=(x0x1xm1)(λ0000λ1000λm1).

Hence, if X=(x0x1xm1) and D=diag(λ0,λ1,,λm1) then X1AX=D. In other words, if A has m linearly independent eigenvectors, then A is diagonalizable.

These insights are summarized in the following theorem:

Here are some classes of matrices that are diagonalizable:

  • Diagonal matrices.

    If A is diagonal, then choosing X=I and A=D yields X1AX=D.

  • Hermitian matrices.

    If A is Hermitian, then the spectral decomposition theorem tells us that there exists unitary matrix Q and diagonal matrix D such that A=QDQH. Choosing X=Q yields X1AX=D.

  • Triangular matrices with distinct diagonal elements.

    If U is upper triangular and has distinct diagonal elements, then by Homework 9.2.3.4 we know we can find an eigenvector associated with each diagonal element and by design those eigenvectors are linearly independent. Obviously, this can be extended to lower triangular matrices as well.

Homework 9.2.5.2.

Let ACm×m have distinct eigenvalues.

ALWAYS/SOMETIMES/NEVER: A is diagonalizable.

Answer

ALWAYS

Now prove it!

Solution

Let \(A = Q U Q^H \) be the Schur decomposition of matrix \(A \text{.}\) Since \(U \) is upper triangular, and has the same eigenvalues as \(A \text{,}\) it has distinct entries along its diagonal. Hence, by our earlier observations, there exists a nonsingular matrix \(X \) such that \(X^{-1} U X = D \text{,}\) a diagonal matrix. Now,

\begin{equation*} X^{-1} Q^H A Q X = X^{-1} U X = D \end{equation*}

and hence \(Y = Q X \) is the nonsingular matrix that diagonalizes \(A \text{.}\)

fit width