The transpose of a matrix , denoted as , is a matrix obtained by swapping the rows and the columns of . This means that:
the first row of becomes the first column of .
the second row of becomes the second column of .
and so on.
Mathematically, if is an matrix with entries , then the transpose of is an matrix with entries . In other words, the entry located at the -th row -th column in will be located at the -th row -th column in .
Example.
Finding the transpose of matrices
Find the transpose of the following matrices:
Solution. The transpose of these matrices is:
Notice how taking the transpose affects the shape of the matrices:
a matrix remains a matrix.
a matrix becomes a matrix.
a matrix remains a matrix.
In general, if is an matrix, then will be an matrix. Let's also understand the mathematical definition of the matrix transpose. Observe what happens to the entry in after taking the transpose. Since the rows and columns are swapped, this entry is located at in .
Example.
Shape of a matrix after taking its transpose
Suppose we take the transpose of a matrix with rows and columns. What would the shape of be?
Solution. Since we are swapping the rows and columns, the transpose of would have rows and columns, that is, .
Properties of matrix transpose
Theorem.
Transpose of a matrix transpose
Taking the transpose of a matrix transpose results in the matrix itself:
Proof. By definition, is obtained by swapping the rows and columns of . Taking the transpose of will swap back the rows and columns and so we end up with the original matrix .
To be more mathematically precise, suppose is as follows:
The transpose of is:
The transpose of is:
This completes the proof.
Theorem.
Transpose of A+B
If and are matrices, then:
Proof. Consider the following matrices:
The transpose of and is:
The sum is:
Now, the sum is:
Taking the transpose gives:
Notice that this matrix is equal to . Therefore, we have that:
This completes the proof.
Theorem.
Diagonal entries of a square matrix do not change after taking its transpose
Let be a square matrix. The diagonal entries of and are the same.
Proof. Suppose is an matrix. The diagonals of is for . Taking the transpose involves switching the two subscripts, but since they are identical, we also end up with as the diagonal entries of . This completes the proof.
Theorem.
Transpose of kA where k is a scalar
If is a matrix and is any scalar, then:
Proof. Consider the following matrix:
If is any scalar, then the product is:
The transpose of is:
The transpose of is:
The scalar-matrix product is:
This is equal to the matrix in . Therefore, we conclude that:
This completes the proof.
Theorem.
Expressing dot product using matrix notation
If and are vectors, then:
Proof. We will prove the case for but this easily generalizes to . Let vectors and be defined as follows:
Their dot product is:
Similarly, we have that:
Similarly, we have that:
This completes the proof.
Theorem.
Matrix product of a vector and its transpose is equal to the vector's squared magnitude
Proof. From the rules of matrix multiplication, we know that:
Here, the subscript represents the value in the -th row -th column. is true for all and . For instance, represents the entry at the st row rd column of matrix .
We know from the definition of transpose that:
Equating and gives:
Now, consider the following:
Here, we used the fact that the entry in is equal to the entry in .
Equating and gives:
Equation holds true for all , and . Since the values of two matrices and are equivalent, these matrices are equivalent, that is:
This completes the proof.
Theorem.
Transpose of a product of three matrices
If , and are matrices, then:
Proof. The proof makes use of our previous theorem . We start from the left-hand side of our proposition:
This completes the proof.
We will now generalize this theorem for any number of matrices. The proof uses induction and follows a similar logic.
Theorem.
Transpose of a product of n matrices
If for are matrices, then:
Proof. We will prove the theorem by induction. Consider the base case when we have matrices. We have already shown in theoremlink that:
Therefore, the base case holds. We now assume that the theorem holds for matrices:
Our goal is to show that the theorem holds for matrices:
We now use the inductive assumption to get:
By the principle of mathematical induction, the theorem holds for the general case of matrices. This completes the proof.
The next theorem is useful for certain proofs.
Theorem.
Another expression for the dot product of Av and w
If is an matrix and and are vectors, then:
Proof. The matrix-vector product results in a vector. We can use theoremlink to convert the dot product of and into a matrix product: