Subsection 9.2.1 Singular matrices and the eigenvalue problem
¶fit widthDefinition 9.2.1.1. Eigenvalue, eigenvector, and eigenpair.
Let A∈Cm×m. Then λ∈C and nonzero x∈Cm are said to be an eigenvalue and corresponding eigenvector if Ax=λx. The tuple (λ,x) is said to be an eigenpair.
A is nonsingular.
A has linearly independent columns.
There does not exist a nonzero vector x such that Ax=0.
N(A)={0}. (The null space of A is trivial.)
dim(N(A))=0.
det(A)≠0.
There exists a vector x≠0 such that (λI−A)x=0.
(λI−A) is singular.
(λI−A) has linearly dependent columns.
The null space of λI−A is nontrivial.
dim(N(λI−A))>0.
det(λI−A)=0.
Definition 9.2.1.2. Spectrum of a matrix.
The set of all eigenvalues of A is denoted by Λ(A) and is called the spectrum of A.Definition 9.2.1.3. Spectral radius.
The spectral radius of A, ρ(A), equals the absolute value of the eigenvalue with largest magnitude:
Theorem 9.2.1.4. Gershgorin Disk Theorem.
Let A∈Cm×m,
and
In other words, ρi(A) equals the sum of the absolute values of the off diagonal elements of A in row i and Ri(A) equals the set of all points in the complex plane that are within a distance ρi of diagonal element αi,i. Then
In other words, the eigenvalues lie in the union of these disks.
Proof.
Let \(\lambda \in \Lambda( A ) \text{.}\) Then \(( \lambda I - A ) x = 0 \) for some nonzero vector \(x \text{.}\) W.l.o.g. assume that index \(i \) has the property that \(1 = \chi_i \geq \vert \chi_j \vert \) for \(j \neq i \text{.}\) Then
or, equivalently,
Hence
Corollary 9.2.1.5.
Let A and Ri(A) be as defined in Theorem 9.2.1.4. Let K and KC be disjoint subsets of {0,…,m−1} such that K∪KC={0,…,m−1}. In other words, let K and KC partition {0,…,m−1}. If
then ∪k∈KRk(A) contains exactly |K| eigenvalues of A (multiplicity counted). In other words, if ∪k∈KRk(A) does not intersect with any of the other disks, then it contains as many eigenvalues of A (multiplicity counted) as there are elements of K.
Proof.
The proof splits \(A = D + ( A - D ) \) where \(D\) equals the diagonal of \(A \) and considers \(A_\omega = D + \omega ( A - D ) \text{,}\) which varies continuously with \(\omega \text{.}\) One can argue that the disks \(R_i( A_0 ) \) start with only one eigenvalue each and only when they start intersecting can an eigenvalue "escape" the disk in which it started. We skip the details since we won't need this result in this course.
Homework 9.2.1.1.
Let A∈Cm×m.
TRUE/FALSE: 0∈Λ(A) if and only A is singular.
TRUE
Now prove it!
\((\Rightarrow) \text{:}\) Assume \(0 \in \Lambda( A ) \text{.}\) Let \(x\) be an eigenvector associated with eigenvalue \(0 \text{.}\) Then \(A x = 0 x = 0 \text{.}\) Hence there exists a nonzero vector \(x \) such that \(A x = 0 \text{.}\) This implies \(A \) is singular.
\((\Leftarrow) \text{:}\) Assume \(A \) is singular. Then there exists \(x \neq 0 \) such that \(A x = 0 \text{.}\) Hence \(A x = 0 x \) and \(0 \) is an eigenvalue of \(A \text{.}\)
Homework 9.2.1.2.
Let A∈Cm×m be Hermitian.
ALWAYS/SOMETIMES/NEVER: All eigenvalues of A are real-valued.
ALWAYS
Now prove it!
Let \(( \lambda, x ) \) be an eigenpair of \(A \text{.}\) Then
and hence
If we now conjugate both sides we find that
which is equivalent to
which is equivalent to
since \(A \) is Hermitian. We conclude that
(since \(x^H x \neq 0 \)).
Homework 9.2.1.3.
Let A∈Cm×m be Hermitian positive definite (HPD).
ALWAYS/SOMETIMES/NEVER: All eigenvalues of A are positive.
ALWAYS
Now prove it!
Let \(( \lambda, x ) \) be an eigenpair of \(A \text{.}\) Then
and hence
and finally (since \(x \neq 0 \))
Since \(A \) is HPD, both \(x^H A x \) and \(x^H x \) are positive, which means \(\lambda\) is positive.
Homework 9.2.1.4.
Let A∈Cm×m be Hermitian, (λ,x) and (μ,y) be eigenpairs associated with A, and λ≠μ.
ALWAYS/SOMETIMES/NEVER: xHy=0
ALWAYS
Now prove it!
Since
we know that
and hence (remembering that the eigenvalues are real-valued)
We can rewrite this as
Since \(\lambda \neq \mu \) this implies that \(y^H x = 0 \) and hence \(x^H y = 0 \text{.}\)
Homework 9.2.1.5.
Let A∈Cm×m, (λ,x) and (μ,y) be eigenpairs, and λ≠μ. Prove that x and y are linearly independent.
Proof by contradiction: Under the assumptions of the homework, we will show that assuming that \(x \) and \(y \) are linearly dependent leads to a contradiction.
If nonzero \(x \) and nonzero \(y \) are linearly dependent, then there exists \(\gamma \neq 0\) such that \(y = \gamma x \text{.}\) Then
implies that
and hence
Rewriting this we get that
Since \(\lambda \neq \mu \) and \(\gamma \neq 0\) this means that \(x = 0 \) which contradicts that \(x \) is an eigenvector.
We conclude that \(x \) and \(y \) are linearly independent.
Homework 9.2.1.6.
Let A∈Cm×m, k≤m, and (λi,xi) for 1≤i<k be eigenpairs of this matrix. Prove that if λi≠λj when i≠j then the eigenvectors xi are linearly independent. In other words, given a set of distinct eigenvalues, a set of vectors created by taking one eigenvector per eigenvalue is linearly independent.
Prove by induction.
Proof by induction on \(k \text{.}\)
Base Case: \(k = 1 \text{.}\) This is trivially.
-
Assume the result holds for \(1 \leq k \lt m \text{.}\) Show it holds for \(k+1 \text{.}\)
The I.H. means that \(x_0, \ldots, x_{k-1} \) are linearly independent. We need to show that \(x_k \) is not a linear combination of \(x_0, \ldots , x_{k-1} \text{.}\) We will do so via a proof by contradiction.
Assume \(x_k \) is a linear combination of \(x_0, \ldots , x_{k-1} \) so that
\begin{equation*} x_k = \gamma_0 x_0 + \cdots + \gamma_{k-1} x_{k-1} \end{equation*}with at least one \(\gamma_j \neq 0 \text{.}\) We know that \(A x_k = \lambda_k x_k\) and hence
\begin{equation*} A ( \gamma_0 x_0 + \cdots + \gamma_{k-1} x_{k-1} ) = \lambda_k ( \gamma_0 x_0 + \cdots + \gamma_{k-1} x_{k-1} ). \end{equation*}Since \(A x_i = \lambda_i x_i \text{,}\) we conclude that
\begin{equation*} \gamma_0 \lambda_0 x_0 + \cdots + \gamma_{k-1} \lambda_{k-1} x_{k-1} = \gamma_0 \lambda_{xk-1} x_0 + \cdots + \gamma_{k-1} \lambda_k x_{k-1} \end{equation*}or, equivalently,
\begin{equation*} \gamma_0 ( \lambda_0 - \lambda_k ) x_0 + \cdots + \gamma_{k-1} ( \lambda_{k-1} - \lambda_k ) x_{k-1} = 0. \end{equation*}Since at least one \(\gamma_i \neq 0 \) and \(\lambda_i \neq \lambda_k \) for \(0 \leq i \lt k \text{,}\) we conclude that \(x_0, \ldots , x_{k-1} \) are linearly dependent, which is a contradiction.
Hence, \(x_0, \ldots , x_{k} \) are linearly independent.
By the Principle of Mathematical Induction, the result holds for \(1 \leq k \leq m \text{.}\)
Homework 9.2.1.7.
Consider the matrices
and
ALWAYS/SOMETIMES/NEVER: All eigenvalues of these matrices are nonnegative.
ALWAYS/SOMETIMES/NEVER: All eigenvalues of the first matrix are positive.
ALWAYS: All eigenvalues of these matrices are nonnegative.
ALWAYS: All eigenvalues of the first matrix are positive. (So are all the eigenvalues of the second matrix, but proving that is a bit trickier.)
Now prove it!
For the first matrix, we can use the Gershgorin disk theorem to conclude that all eigenvalues of the matrix lie in the set \(\{ x \mbox{ s.t. } \vert x - 2 \vert \leq 2 \text{.}\) We also notice that the matrix is symmetric, which means that its eigenvalues are real-valued. Hence the eigenvalues are nonnegative. A similar argument can be used for the second matrix.
Now, in Homework 7.2.1.1 we showed that the first matrix is nonsingular. Hence, it cannot have an eigenvalue equal to zero. We conclude that its eigenvalues are all positive.
It can be shown that the second matrix is also nonsingular, and hence has positive eigenvalues. However, that is a bit nasty to prove...