Linear Transformation
Definition and Linearity
Linearity
Mapping between vector spaces
For any field
Applications, examples
Vector spaces can model many real world things (see vector spaces ref), even functions. In all of these, linear transformations have deep meanings.
Over function spaces: Differentiation, integration, multiplication by fixed polynomial in
Geometric operations:
As a matrix
Use action on basis vectors
Take the linear operation
(So,
Matrix
Now, write a matrix
Matrix * vector multiplication
Define
For other views of
Vector dot product
(This also defines row vector * column vector multiplication!) This is also the standard inner product.
Row view: inner product with rows
Take the standard inner product. Then, Ax is the vector formed by
Changing basis vectors
Representation of the same geometric point can change with the choice of bases.
Take the point
Representation wrt i/p and o/p bases
Similarly, representation of a linear transformation can also change with choice of bases.
When the right input and output basis is chosen, every matrix is actually diagonal : mere scaling: see SVD.
Forward operation: Ax
Spaces related to range(A)
Range space
As
By linearity,
Left null space
Also, orthogonal complement of
Subspaces of the domain
Null space (Kernel)
N(A):
Left null space is
Find null-space
Ax=0; Reduce
Similarly, Find left null-space
Row space
This is the space spanned by rows of
Every x:
SVD view
See elsewhere.
Rank of A
Row and column ranks
The number of linearly independent rows in
Row rank = column rank
Do triangular row elimination to get PA = LU. Then, row rank = number of non-zero rows/ pivots in A.
But, every column
SVD view
Take SVD: \(A = U \SW V^{}\). Rank r of
Invertability
Both row space and range(A) must have dimension r. Every
Else, use pseudoinverse to get left inverse if
Identity operation
Matrix multiplication: AB
Composition of transformations
AB is defined to conform to transformation ABx. So, ABCx is associative.
As sum of rank 1 matrices
Similarly, for diagonal matrix D:
Elementwise description
From sum of rank 1 matrices form: \((AB){i,j} = A{i,:}B_{:,j}\).
Columns of the result
A acts on B’s rows; B acts on A’s columns. Every col of C=AB is a linear combination of A’s cols according to some col of B:
Form of ABC
Consider expansion of quadratic functional (see vector spaces ref). Similarly, in D = ABC has
So,
Computation
Needs
Rank
rank(AB)
Inverse and transpose
So, as \(I = (AA^{-1})^{} = (A^{-1})^{}A^{}\), \((A^{-1})^{} = (A^{})^{-1}\) (aka \(A^{-}\)). Also, \((AA^{})^{} = AA^{}\) and \(A^{}A\) Hermitian.
Block matrices
Block multiplication works.
Other matrix products
Entrywise product
Aka Hadamard product, Schur product.
Kronecker product
Aka outer product.
From defn,
Inverse operation: Solve Ax = b for x
Solvability Assumption
Left and right inverses
Right inverse:
Existance conditions
The left inverse exists exactly when you can solve Ax = b for all
Equivalence if both exist
If both left and right inverses exist, they’re the same:
Solutions
Full column rank
If
Rank Defective matrix
If
Finding solutions
See numerical analysis reference for various techniques.
Finding inverses
Find right inverse: Gauss-Jordan: Make augmented matrix A:I; use row operations to reduce it to
Or make cofactor matrix C, use
Block matrices
Block Gaussian elimination
Take
Block LU
Thence, get block LU, Aka Leibnitz factoriazation: \
Schur complement
S is called the Schur complement of
Inverse of X
Do block back substitution to get
Pseudoinverse for long thin A
Projector + inverse. For rectangular, non column rank deficient matrices:
\(A^{+} = (A^{}A)^{-1}A^{} = \hat{R}^{-1}\hat{Q}^{}\) (as
Restriction to a subspace S
Let Q be the orthogonal matrix formed by an orthonormal basis of S. Then, projection of
Submatrices
Principle submatrices:
Random matrices
Random projections
(Johnson Lindenstrauss) See randomized algorithms survey.