Introduction to Diagonalization
Start your free 7-days trial now!
Diagonalizable matrix
An $n\times{n}$ matrix $\boldsymbol{A}$ is said to be diagonalizable if there exists an invertible matrix $\boldsymbol{P}$ and a diagonal matrixlink $\boldsymbol{D}$ such that:
Note that $\boldsymbol{P}$ is said to diagonalize $\boldsymbol{A}$. We say that $\boldsymbol{PDP}^{-1}$ is an eigen-decomposition of $\boldsymbol{A}$.
2x2 diagonalizable matrix
Here's an example of a diagonalizable $2\times2$ matrix:
Diagonalizable matrices and the corresponding diagonal matrices are similar
If $\boldsymbol{A}$ is a diagonalizable matrix and $\boldsymbol{D}$ is the corresponding diagonal matrix, then $\boldsymbol{A}$ and $\boldsymbol{D}$ are similarlink.
Proof. Let $\boldsymbol{A}$ be a diagonalizable matrix:
Where $\boldsymbol{P}$ is an invertible matrix and $\boldsymbol{D}$ is a diagonal matrix. By definitionlink of similar matrices, $\boldsymbol{A}$ and $\boldsymbol{D}$ are similar.
This means that a diagonalizable matrix and its associated diagonal matrix enjoy the properties of similar matrices.
Diagonalization Theorem
Suppose we have an $n\times{n}$ matrix $\boldsymbol{A}$ that is diagonalizable, which means $\boldsymbol{A}$ can be written as:
Where $\boldsymbol{P}$ is some invertible matrix and $\boldsymbol{D}$ is a diagonal matrix.
The diagonalization theorem states that:
$\boldsymbol{D}$ is an $n\times{n}$ diagonal matrix with the eigenvalues of $\boldsymbol{A}$ as its diagonal entries.
$\boldsymbol{P}$ is an invertible $n\times{n}$ matrix with the eigenvectors of $\boldsymbol{A}$ as its columns.
Proof. Suppose we have an $n\times{n}$ diagonalizable matrix $\boldsymbol{A}$ as follows:
Where $\boldsymbol{P}$ is an invertible $n\times{n}$ matrix and $\boldsymbol{D}$ is an $n\times{n}$ diagonal matrix.
Suppose $\boldsymbol{P}$ has columns $\boldsymbol{x}_1$, $\boldsymbol{x}_2$, $\cdots$, $\boldsymbol{x}_n$ like so:
Suppose the diagonal matrix $\boldsymbol{D}$ is:
Where the diagonal entries $\lambda_1$, $\lambda_2$, $\cdots$, $\lambda_n$ are some real numbers. Note that we have defined $\boldsymbol{P}$ and $\boldsymbol{D}$ as above but we haven't imposed any conditions on them - we're simply expressing them as generic matrices.
By theoremlink, the matrix product $\boldsymbol{PD}$ is:
By theoremlink, the matrix product $\boldsymbol{AP}$ is:
Now recall that $\boldsymbol{A}=\boldsymbol{PDP}^{-1}$. Multiplying both sides by $\boldsymbol{P}$ would give us $\boldsymbol{AP}=\boldsymbol{PD}$. Using equations \eqref{eq:LDiIcKRQr8b5m9eHnuG} and \eqref{eq:JWfoh7K8qb73Xx4MAR1}, we have that:
Equating the columns gives us:
By theoremlink, because $\boldsymbol{P}$ is invertible, its column vectors $\boldsymbol{x}_i$ are non-zero vectors. Therefore, by the definitionlink of eigenvalues and eigenvectors:
$\lambda_1$, $\lambda_2$, $\cdots$, $\lambda_n$ must represent the eigenvalues.
$\boldsymbol{x}_1$, $\boldsymbol{x}_2$, $\cdots$, $\boldsymbol{x}_n$ must represent the corresponding eigenvectors!
This completes the proof.
Diagonalizing a 2x2 matrix
Diagonalize the following matrix:
That is, find the diagonal matrix $\boldsymbol{D}$ and invertible matrix $\boldsymbol{P}$ such that $\boldsymbol{A}=\boldsymbol{PDP}^{-1}$.
Solution. We will use the diagonalization theorem to diagonalize matrix $\boldsymbol{A}$. We know that the columns of matrix $\boldsymbol{P}$ are the eigenvectors of $\boldsymbol{A}$ whereas the diagonal entries of matrix $\boldsymbol{D}$ are the eigenvalues of $\boldsymbol{A}$.
Therefore, we must compute the eigenvalues and eigenvectors of $\boldsymbol{A}$. We already went through the step-by-step calculations for this specific matrix $\boldsymbol{A}$ in this sectionlink of our guide on eigenvalues and eigenvectors, so we will just skip the calculations here. The eigenvalues of $\boldsymbol{A}$ are as follows:
The corresponding eigenvectors are:
We can now fully define matrices $\boldsymbol{P}$ and $\boldsymbol{D}$ like so:
Matrix $\boldsymbol{A}$ can therefore be written as:
Taking the power of a diagonalized matrix
If $\boldsymbol{A}$ is a diagonalizable $n\times{n}$ matrix, then $\boldsymbol{A}$ raised to the power of $k$ is:
Proof. Since $\boldsymbol{A}$ is assumed to be a diagonalizable matrix, we have that:
Taking the power of $k$ on both sides yields:
This completes the proof.
Computing the power of a diagonalizable matrix
Consider the same matrix as earlier:
Compute $\boldsymbol{A}^{3}$.
Solution. Recall that the diagonalized form of $\boldsymbol{A}$ is:
We now apply the theorem:
This theorem is particularly useful when computing much higher powers. For instance, suppose we wanted to compute $\boldsymbol{A}^{100}$. Without the theorem, this would involve computing matrix products $99$ times:
In contrast, using our theorem involves computing the following:
Computing powers of several numbers is computationally much cheaper than performing matrix multiplication $99$ times!
Matrix of size n is diagonalizable if and only if it has n linearly independent eigenvectors
Let $\boldsymbol{A}$ be an $n\times{n}$ matrix. $\boldsymbol{A}$ is diagonalizable if and only if $\boldsymbol{A}$ has $n$ linearly independent eigenvectors.
Proof. We start by proving the forward proposition. Assume $\boldsymbol{A}$ is diagonalizable, which means that:
Where:
$\boldsymbol{D}$ is a diagonal matrix with eigenvalues of $\boldsymbol{A}$ as its diagonal entries.
$\boldsymbol{P}$ is an invertible $n\times{n}$ matrix with the eigenvectors of $\boldsymbol{A}$ as its columns.
Since $\boldsymbol{A}$ is diagonalizable, both $\boldsymbol{D}$ and $\boldsymbol{P}$ exist. Since $\boldsymbol{P}$ is invertible, $\boldsymbol{P}$ has $n$ linearly independent columns by theoremlink. Since the columns of $\boldsymbol{P}$ are its eigenvectors, we have that $\boldsymbol{P}$ has $n$ linearly independent eigenvectors.
Let's now prove the converse. We assume that $\boldsymbol{A}$ has $n$ linearly independent eigenvectors $\boldsymbol{x}_1$, $\boldsymbol{x}_2$, $\cdots$, $\boldsymbol{x}_n$ and corresponding eigenvalues $\lambda_1$, $\lambda_2$, $\cdots$, $\lambda_n$. By definitionlink of eigenvectors and eigenvalues, the following equations hold:
In matrix form, we can express this as:
Where \eqref{eq:x4dH09uGAwai0JZqNxh} is obtained by equating the columns. By theoremlink, the left-hand side of \eqref{eq:ymonPOqHL1v7Dg4EZ2F} can be expressed as:
Where $\boldsymbol{P}$ is a matrix whose columns are the $n$ linearly independent eigenvectors.
Next, by theoremlink, the right-hand side of \eqref{eq:ymonPOqHL1v7Dg4EZ2F} can be expressed as:
Where $\boldsymbol{D}$ is a diagonal matrix whose diagonal entries are the eigenvalues of $\boldsymbol{A}$.
Now, \eqref{eq:ymonPOqHL1v7Dg4EZ2F} can be expressed as $\boldsymbol{AP}=\boldsymbol{PD}$. Since the columns of $\boldsymbol{P}$ are linearly independent, $\boldsymbol{P}$ is invertible by theoremlink. This means that $\boldsymbol{P}^{-1}$ exists and thus:
By definitionlink, $\boldsymbol{A}$ is therefore diagonalizable. This completes the proof.
Matrix of size n with n distinct eigenvalues is diagonalizable
If $\boldsymbol{A}$ is an $n\times{n}$ matrix with $n$ distinct eigenvalues, then $\boldsymbol{A}$ is diagonalizable.
Proof. By definitionlink of diagonalizable matrices and the diagonalization theoremlink, we know that $\boldsymbol{A}$ is diagonalizable if we can write $\boldsymbol{A}$ as:
Where:
$\boldsymbol{D}$ is a diagonal matrix with eigenvalues of $\boldsymbol{A}$ as its diagonal entries.
$\boldsymbol{P}$ is an invertible $n\times{n}$ matrix with the eigenvectors of $\boldsymbol{A}$ as its columns.
By theoremlink, if an $n\times{n}$ matrix $\boldsymbol{A}$ has $n$ distinct eigenvalues, then $\boldsymbol{A}$ has $n$ linearly independent eigenvectors. By theoremlink, $\boldsymbol{A}$ is therefore diagonalizable. This completes the proof.
Note that the converse is not necessarily true, that is, if $\boldsymbol{A}$ is diagonalizable, then $n\times{n}$ matrix may not have $n$ distinct eigenvalues.
Matrix is diagonalizable if and only if it has an eigenbasis
Let $\boldsymbol{A}$ be a square matrix. $\boldsymbol{A}$ is diagonalizable if and only if $\boldsymbol{A}$ has an eigenbasis.
Proof. This theorem is analogous to theoremlink. Let's first prove the forward proposition. Assume an $n\times{n}$ matrix $\boldsymbol{A}$ is diagonalizable. By theoremlink, $\boldsymbol{A}$ has $n$ linearly independent eigenvectors, which form an eigenbasis for $\mathbb{R}^n$.
Let's now prove the converse. Assume that $\boldsymbol{A}$ has an eigenbasis. This means that $\boldsymbol{A}$ has $n$ linearly independent eigenvectors. By theoremlink once again, $\boldsymbol{A}$ is diagonalizable.
This completes the proof.