S6 UNIT 7: Transformation of Matrices.

About Course
In linear algebra, a matrix transformation (often synonymous with linear transformation) is a special type of function that takes a vector as input and produces another vector as output, where this transformation can be represented by multiplication with a matrix. It’s a fundamental concept that bridges abstract vector spaces with concrete matrix operations.
What is a Matrix Transformation?
A transformation
is a matrix transformation if there exists an matrix such that for every vector in ,T(x) = Ax
Here:
- x is a column vector in Rn (n-dimensional input vector).
- A is an m×n matrix (the transformation matrix).
- Ax is the matrix-vector product, resulting in a column vector in Rm (m-dimensional output vector).
This means that any matrix multiplication can be interpreted as a geometric transformation of vectors.
Key Properties of Matrix Transformations (Linear Transformations)
Matrix transformations are a specific type of linear transformation. A transformation
is linear if it satisfies two properties:-
- Additivity: for all vectors in the domain.
- Homogeneity (Scalar Multiplication): (cu)= cT(u) for all vectors in the domain and all scalars .
The kernel and range (also known as the null space and image, respectively) are two fundamental subspaces associated with a linear transformation. They provide crucial insights into how a transformation behaves, particularly concerning its injectivity (one-to-one) and surjectivity (onto) properties.
Let be a linear transformation from a vector space (the domain) to a vector space (the codomain).
The Kernel of a Transformation (Null Space).
The kernel of a linear transformation , denoted as or , is the set of all vectors in the domain that are mapped to the zero vector in the codomain .
Formally:
(T)= {v∈V ∣ T(v)=0W}Where
is the zero vector in the codomain .Key Properties of the Kernel:
- Subspace of the Domain: The kernel of a linear transformation is always a subspace of the domain V. This means it contains the zero vector of V, is closed under vector addition, and is closed under scalar multiplication.
- Injectivity (One-to-One): A linear transformation T is injective (one-to-one) if and only if its kernel contains only the zero vector, i.e., ker(T)={0V}. This means that distinct vectors in the domain always map to distinct vectors in the codomain.
- Relationship to Homogeneous Systems: If the linear transformation T is represented by a matrix A (i.e., T(x)=Ax), then finding the kernel of T is equivalent to solving the homogeneous system of linear equations Ax =0. The kernel is precisely the solution space of this system.
- Nullity: The dimension of the kernel of T is called the nullity of T, denoted as nullity(T) or dim(ker(T)).
How to Find the Kernel (for a matrix transformation T(x)= Ax)?
- Set up the homogeneous equation Ax=0.
- Form the augmented matrix [A∣0] and row-reduce it to its Reduced Row Echelon Form (RREF).
- Write out the general solution to the system. The vectors that form the basis for this solution space constitute the basis for the kernel.
The Range of a Transformation (Image)
The range of a linear transformation
, denoted as (T) or (T), is the set of all possible output vectors in the codomain that can be obtained by applying to vectors from the domain .Formally:
(T)= {w∈W ∣ ∃v∈V such that T(v)=w} This is equivalent to the set (V)={T(v) ∣ v∈V}.Key Properties of the Range:
-
-
- Subspace of the Codomain: The range of a linear transformation is always a subspace of the codomain . It contains the zero vector of , is closed under vector addition, and is closed under scalar multiplication.
Surjectivity (Onto):
-
A linear transformation
is surjective (onto) if and only if its range is equal to the entire codomain , i.e., . This means that every vector in the codomain can be reached by the transformation from some vector in the domain.Relationship to Column Space: If the linear transformation
is represented by a matrix (i.e., ), then the range of is precisely the column space of ( ). This is because any output is a linear combination of the columns of .Rank: The dimension of the range of is called the rank of , denoted as or . For a matrix , .How to Find the Range (for a matrix transformation (x)= Ax):- Identify the columns of the matrix .
- The range of is the span of these column vectors.
- To find a basis for the range, row-reduce the matrix to its RREF. The pivot columns in the original matrix (corresponding to the columns with leading ones in the RREF) form a basis for the column space, and thus for the range.
The Rank-Nullity Theorem
A fundamental theorem in linear algebra connects the dimensions of the kernel and range:
For a linear transformation
: V→W, where is a finite-dimensional vector space:dim(V)=dim(ker(T)) +dim(range(T))
Or, in terms of nullity and rank:
dim(V)=nullity(T)+rank(T)
Diagonalization of a matrix is a powerful concept in linear algebra that simplifies complex matrix operations by transforming a square matrix into a special, much simpler form: a diagonal matrix. This transformation is achieved through a process involving eigenvalues and eigenvectors.
What is Diagonalization?
A square matrix
is said to be diagonalizable if it is similar to a diagonal matrix . This means that there exists an invertible matrix such that:A=PDP−1 or, equivalently: D=P−1AP
Here:
- A is the original square matrix.
- D is a diagonal matrix, meaning all its entries are zero except for those on the main diagonal. The diagonal entries of D are the eigenvalues of A.
- P is an invertible matrix whose columns are the linearly independent eigenvectors of A, corresponding to the eigenvalues in D. This matrix P is often called the modal matrix or eigenvector matrix.
- P−1 is the inverse of matrix P.
The expression P−1AP is called a similarity transformation. If two matrices are similar, they share many important properties, including their eigenvalues, determinant, trace, and rank.
Conditions for Diagonalizability
Not all square matrices are diagonalizable. A matrix
is diagonalizable if and only if:- It has a full set of linearly independent eigenvectors. For an matrix, this means it must have linearly independent eigenvectors.
- Equivalently, for each eigenvalue, its geometric multiplicity (the dimension of its eigenspace) must be equal to its algebraic multiplicity (the number of times it appears as a root of the characteristic polynomial).
Important Note:
- If a matrix has n distinct eigenvalues, it is guaranteed to be diagonalizable.
- If a matrix has repeated eigenvalues, it may or may not be diagonalizable. You need to check if you can find enough linearly independent eigenvectors for each repeated eigenvalue. If not, the matrix is defective and not diagonalizable. In such cases, a matrix can sometimes be put into a Jordan Canonical Form, which is the “next best thing” to a diagonal form.
How to Diagonalize a Matrix (Recipe)
To diagonalize a matrix
:- Find the Eigenvalues of : Solve the characteristic equation (A−λI) =0 for . These values are the eigenvalues.
- Find the Eigenvectors for Each Eigenvalue: For each eigenvalue , solve the system to find the corresponding eigenvectors .
- Form the Matrix : If you found linearly independent eigenvectors, arrange them as the columns of a matrix .
- Form the Diagonal Matrix : Create a diagonal matrix with the eigenvalues on its main diagonal, in the same order as their corresponding eigenvectors appear in .
- Verify (Optional but Recommended): Calculate to confirm that it equals
The Process of Diagonalization:
- The Diagonalization Formula: Students will learn the fundamental relationship A=PDP−1 (or D=P−1AP) and its components:
- A: The original square matrix to be diagonalized.
- D: A diagonal matrix whose diagonal entries are the eigenvalues of A.
- P: An invertible matrix whose columns are the linearly independent eigenvectors of A, ordered to correspond with the eigenvalues in D.
- Steps to Diagonalize a Matrix: A clear, step-by-step procedure:
- Calculate the eigenvalues of A.
- For each eigenvalue, find a basis for its corresponding eigenspace (i.e., find the eigenvectors).
- Check for diagonalizability: If there are n linearly independent eigenvectors for an n×n matrix, it is diagonalizable. Otherwise, it is not.
- Construct P using the eigenvectors as columns.
- Construct D using the eigenvalues on the diagonal (in the same order as their eigenvectors in P).
- (Optional but good for understanding) Verify the relationship A=PDP−1 or D=P−1AP.
Conditions for Diagonalizability:
- Sufficient Condition: If an n×n matrix has n distinct eigenvalues, it is guaranteed to be diagonalizable.
- Necessary and Sufficient Condition: A matrix is diagonalizable if and only if for every eigenvalue, its algebraic multiplicity equals its geometric multiplicity.
- Non-Diagonalizable (Defective) Matrices: Students will learn that some matrices cannot be diagonalized (they are “defective”) because they don’t have enough linearly independent eigenvectors. They might be introduced to the concept of the Jordan Canonical Form as a “next best” decomposition for such matrices.
Course Content
Kernel and Range of a Transformation.
-
Kernel and Range of a Transformation.
32:54