Definition
A linear transformation T : V → W T: V \rightarrow W T : V → W between vector spaces V V V and W W W is a function that preserves vector addition and scalar multiplication :
T ( u + v ) = T ( u ) + T ( v ) T(u + v) = T(u) + T(v) T ( u + v ) = T ( u ) + T ( v ) for all u , v ∈ V u, v \in V u , v ∈ V (preserves addition)
T ( c ⋅ v ) = c ⋅ T ( v ) T(c \cdot v) = c \cdot T(v) T ( c ⋅ v ) = c ⋅ T ( v ) for all v ∈ V v \in V v ∈ V and scalars c c c (preserves scalar multiplication)
Examples
Rotation in R 2 \mathbb{R}^2 R 2
A rotation by angle θ \theta θ in the plane is a linear transformation :
T ( x y ) = ( cos θ − sin θ sin θ cos θ ) ( x y ) T\begin{pmatrix} x \\ y \end{pmatrix} = \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix} \begin{pmatrix} x \\ y \end{pmatrix} T ( x y ) = ( cos θ sin θ − sin θ cos θ ) ( x y )
Differentiation
The differentiation operator D : P n → P n − 1 D: P_n \rightarrow P_{n-1} D : P n → P n − 1 defined by D ( p ( x ) ) = p ′ ( x ) D(p(x)) = p'(x) D ( p ( x )) = p ′ ( x ) is a linear transformation on the space of polynomials.
Projection
The projection of a vector onto a subspace is a linear transformation .
Matrix Representation
Every linear transformation T : V → W T: V \rightarrow W T : V → W between finite-dimensional vector spaces can be represented by a matrix with respect to chosen bases .
If { v 1 , v 2 , … , v n } {v_1, v_2, \ldots, v_n} { v 1 , v 2 , … , v n } is a basis for V V V and { w 1 , w 2 , … , w m } {w_1, w_2, \ldots, w_m} { w 1 , w 2 , … , w m } is a basis for W W W , then:
Compute T ( v j ) T(v_j) T ( v j ) for each basis vector v j v_j v j
Express each T ( v j ) T(v_j) T ( v j ) as a linear combination of the basis vectors of W W W :
T ( v j ) = a 1 j w 1 + a 2 j w 2 + ⋯ + a m j w m T(v_j) = a_{1j}w_1 + a_{2j}w_2 + \cdots + a_{mj}w_m T ( v j ) = a 1 j w 1 + a 2 j w 2 + ⋯ + a mj w m
The matrix representation of T T T is the m × n m \times n m × n matrix :
[ T ] = [ a 11 a 12 ⋯ a 1 n a 21 a 22 ⋯ a 2 n ⋮ ⋮ ⋱ ⋮ a m 1 a m 2 ⋯ a m n ] [T] = \begin{bmatrix}
a_{11} & a_{12} & \cdots & a_{1n} \\
a_{21} & a_{22} & \cdots & a_{2n} \\
\vdots & \vdots & \ddots & \vdots \\
a_{m1} & a_{m2} & \cdots & a_{mn}
\end{bmatrix} [ T ] = a 11 a 21 ⋮ a m 1 a 12 a 22 ⋮ a m 2 ⋯ ⋯ ⋱ ⋯ a 1 n a 2 n ⋮ a mn
Properties
Kernel and Image
The kernel (or null space ) of T T T , denoted ker ( T ) \ker(T) ker ( T ) or Null ( T ) \text{Null}(T) Null ( T ) , is the set of all vectors in V V V that map to the zero vector in W W W :
ker ( T ) = { v ∈ V : T ( v ) = 0 } \ker(T) = \{v \in V : T(v) = 0\} ker ( T ) = { v ∈ V : T ( v ) = 0 }
The image (or range ) of T T T , denoted Im ( T ) \text{Im}(T) Im ( T ) or Range ( T ) \text{Range}(T) Range ( T ) , is the set of all vectors in W W W that are the image of some vector in V V V :
Im ( T ) = { T ( v ) : v ∈ V } \text{Im}(T) = \{T(v) : v \in V\} Im ( T ) = { T ( v ) : v ∈ V }
Rank and Nullity
The rank of T T T is the dimension of its image: rank ( T ) = dim ( Im ( T ) ) \text{rank}(T) = \dim(\text{Im}(T)) rank ( T ) = dim ( Im ( T ))
The nullity of T T T is the dimension of its kernel: nullity ( T ) = dim ( ker ( T ) ) \text{nullity}(T) = \dim(\ker(T)) nullity ( T ) = dim ( ker ( T ))
Rank-Nullity Theorem
For a linear transformation T : V → W T: V \rightarrow W T : V → W where V V V is finite-dimensional:
dim ( V ) = rank ( T ) + nullity ( T ) \dim(V) = \text{rank}(T) + \text{nullity}(T) dim ( V ) = rank ( T ) + nullity ( T )
This fundamental theorem connects the dimension of the domain, the dimension of the image, and the dimension of the kernel .
Composition of Linear Transformations
If S : U → V S: U \rightarrow V S : U → V and T : V → W T: V \rightarrow W T : V → W are linear transformations , then their composition T ∘ S : U → W T \circ S: U \rightarrow W T ∘ S : U → W defined by ( T ∘ S ) ( u ) = T ( S ( u ) ) (T \circ S)(u) = T(S(u)) ( T ∘ S ) ( u ) = T ( S ( u )) is also a linear transformation .
In terms of matrix representations, if [ S ] [S] [ S ] is the matrix of S S S and [ T ] [T] [ T ] is the matrix of T T T , then the matrix of T ∘ S T \circ S T ∘ S is the product [ T ] [ S ] [T][S] [ T ] [ S ] .
Invertible Linear Transformations
A linear transformation T : V → W T: V \rightarrow W T : V → W is invertible if there exists a linear transformation T − 1 : W → V T^{-1}: W \rightarrow V T − 1 : W → V such that T − 1 ∘ T = I V T^{-1} \circ T = I_V T − 1 ∘ T = I V and T ∘ T − 1 = I W T \circ T^{-1} = I_W T ∘ T − 1 = I W , where I V I_V I V and I W I_W I W are the identity transformations on V V V and W W W respectively.
A linear transformation is invertible if and only if:
It is injective (one-to-one): ker ( T ) = { 0 } \ker(T) = {0} ker ( T ) = { 0 }
It is surjective (onto): Im ( T ) = W \text{Im}(T) = W Im ( T ) = W
For finite-dimensional spaces of the same dimension , these conditions are equivalent.
Applications
Linear transformations are fundamental in:
Computer graphics (rotations, scaling, projections)
Quantum mechanics (operators on state spaces)
Signal processing (Fourier transforms)
Machine learning (linear models, dimensionality reduction)
Differential equations (solving systems of linear ODEs)