Section 25.4 Concepts
In this section.
Subsection 25.4.1 The transition matrix and the diagonal form
The columns of the transition matrix.
In Discovery 25.1, we transformed the equation \(\inv{P}AP = D\) into the equivalent equation \(AP = PD\text{.}\) Thinking of \(P\) as being made up of column vectors, multiplying \(P\) on the left by \(A\) multiplies each column of \(P\) by \(A\text{,}\) and multiplying \(P\) on the right by \(D\) multiplies each column of \(P\) by the corresponding diagonal entry. So if we view \(P\) and \(D\) as having forms
then we can view \(AP\) and \(PD\) as having forms
The only way these two matrices can be equal is if they have equal columns, so that
These column vector equalities exhibit the eigenvector-eigenvalue pattern. That is, the only way to make \(\inv{P}AP\) diagonal is to use eigenvectors of \(A\) as the columns of the transition matrix \(P\).
Moreover, \(P\) needs to be invertible, so the columns of \(P\) need to be linearly independent (Theorem 21.5.5).
The diagonal form matrix \(\inv{P}AP\).
In Discovery 25.3, we analyzed the pattern of the diagonal matrix \(D = \inv{P}AP\text{.}\) If \(\lambda_j\) is its \(\nth[j]\) diagonal entry, the condition \(A\uvec{p}_j = \lambda_j\uvec{p}_j\) from our analysis above says that \(\lambda_j\) is an eigenvalue for \(A\text{,}\) and the \(\nth[j]\) column of \(P\) is a corresponding eigenvector. So
- \(D\) will have the eigenvalues of \(A\) for its diagonal entries,
- the number of times an eigenvalue of \(A\) is repeated as a diagonal entry in \(D\) will correspond to the number of linearly independent eigenvectors for that eigenvalue that were used in the columns of \(P\text{,}\) and
- the order of the entries down the diagonal of \(D\) corresponds to the order of eigenvectors in the columns of \(P\text{.}\)
Subsection 25.4.2 Diagonalizable matrices
Is every \(n\times n\) matrix similar to a diagonal one? In Discovery 25.4, we discovered that the answer is no. For some matrices, it will not be possible to collect together enough linearly independent eigenvectors to fill all \(n\) columns of the transition matrix \(P\text{.}\) The largest number of linearly independent eigenvectors we can obtain for a particular eigenvalue is the dimension of the corresponding eigenspace. In Discovery 25.6, we discovered that eigenvectors from different eigenspaces of the same matrix are automatically linearly independent. So the limiting factor is the dimension of each eigenspace, and whether these dimensions add up to \(n\text{,}\) the required number of linearly independent columns in \(P\text{.}\)
An eigenvalue of an \(n \times n\) matrix \(A\) has two important numbers attached to it — its algebraic multiplicity and its geometric multiplicity.
If the roots of the characteristic polynomial are all real numbers, then the characteristic polynomial will factor completely as
where the \(\lambda_j\) are the distinct eigenvalues of \(A\) and the \(m_j\) are the corresponding algebraic multiplicities. Since \(c_A(\lambda)\) is always a degree \(n\) polynomial, the algebraic multiplicities will add up to \(n\text{.}\) To obtain enough linearly independent eigenvectors for \(A\) to fill the columns of \(P\text{,}\) we also need the geometric multiplicities to add up to \(n\text{.}\) We will learn in Subsection 25.6.3 that somehow, the algebraic multiplicity of each eigenvalue is the “best-case scenario” — the geometric multiplicity for an eigenvalue can be no greater than its algebraic multiplicity. Thus, if any eigenvalue for \(A\) is “defective” in the sense that its geometric multiplicity is strictly less than its algebraic multiplicity, we will not obtain enough linearly independent eigenvectors for that eigenvalue to fill up its “portion” of the required eigenvectors. To summarize, a square matrix is diagonalizable precisely when each of its eigenvalues has geometric multiplicity equal to its algebraic multiplicity.
Subsection 25.4.3 Diagonalization procedure
Procedure 25.4.1. To diagonalize an \(n \times n\) matrix \(A\text{,}\) if possible.
- Compute the characteristic polynomial \(c_A(\lambda)\) of \(A\) by computing \(\det (\lambda I - A)\text{,}\) then determine the eigenvalues of \(A\) by solving the characteristic equation \(c_A(\lambda) = 0\text{.}\) Make note of the algebraic multiplicity of each eigenvalue.
- For each eigenvalue \(\lambda_j\) of \(A\text{,}\) determine a basis for the correponding eigenspace \(E_{\lambda_j}(A)\) by solving the homogeneous linear system \((\lambda_j I - A)\uvec{x} = \zerovec\text{.}\) Make note of the geometric multiplicity of each eigenvalue.
- If any eigenvalue has geometric multiplicity strictly less than its algebraic multiplicity, then \(A\) is not diagonalizable. On the other hand, if each eigenvalue has equal geometric and algebraic multiplicities, then you can obtain \(n\) linearly independent eigenvectors to make up the columns of \(P\) by taking together all the eigenspace basis vectors you found in the previous step.
If the matrix \(P\) has successfully been constructed, then \(D = \inv{P}AP\) will be in diagonal form, with eigenvalues of \(A\) in the diagonal entries of \(D\text{,}\) in order corresponding to the order of placement of eigenvectors in the columns of \(P\text{.}\)