Matrix decompositions (or factorizations) are ways of breaking down a complex matrix into a product of simpler, structured matrices.
Think of it like factoring a number (e.g., \(12 = 2 \times 2 \times 3\)). Factoring a matrix reveals its fundamental properties and makes certain computations much easier.
A matrix decomposition expresses a given matrix \(\mathbf{A}\) as a product of two or more matrices with specific properties (e.g., diagonal, orthogonal, triangular).
$$ \mathbf{A} = \mathbf{F}_1 \mathbf{F}_2 \dots \mathbf{F}_k $$
where each \(\mathbf{F}_i\) has useful characteristics.
Data Compression & Rank Reduction:SVD allows for low-rank approximation, compressing data by retaining essential components.
Revealing Properties:Eigenvalues from Eigendecomposition indicate stability; singular values from SVD relate to matrix rank and energy/variance.
Algorithm Development: Many algorithms in machine learning (e.g., PCA, recommender systems) and numerical analysis rely directly on matrix decompositions.
Applies to: Square matrices with a full set of linearly independent eigenvectors (diagonalizable matrices). Especially useful for symmetric matrices where \(\mathbf{V}\) is orthogonal (\(\mathbf{V}^{-1} = \mathbf{V}^T\)).
Components:
\(\mathbf{V}\): Matrix whose columns are the eigenvectors of \(\mathbf{A}\).
\(\mathbf{\Lambda}\) (Lambda): Diagonal matrix containing the corresponding eigenvalues (\(\lambda_i\)) of \(\mathbf{A}\).
LU Decomposition (\(\mathbf{A = LU}\)): Factors a square matrix \(\mathbf{A}\) (under certain conditions) into a Lower triangular matrix (with 1s on diagonal) and an Upper triangular matrix. Primarily used for efficient solving of \(\mathbf{Ax=b}\) via forward/backward substitution.
QR Decomposition (\(\mathbf{A = QR}\)): Factors any\(m \times n\) matrix \(\mathbf{A}\) (with \(m \ge n\) and linearly independent columns) into an Qrthogonal matrix (\(m \times n\), \(\mathbf{Q}^T\mathbf{Q}=\mathbf{I}\)) and an Right (upper) triangular matrix (\(n \times n\)). Used in solving least-squares problems (\(\mathbf{A}^T\mathbf{Ax} = \mathbf{A}^T\mathbf{b} \implies \mathbf{R}^T\mathbf{Q}^T\mathbf{QRx} = \mathbf{R}^T\mathbf{Q}^T\mathbf{b} \implies \mathbf{Rx} = \mathbf{Q}^T\mathbf{b}\)) and for eigenvalue algorithms (QR algorithm).
Cholesky Decomposition (\(\mathbf{A = LL^T}\) or \(\mathbf{A = R^T R}\)): Factors a symmetric, positive-definite matrix \(\mathbf{A}\) into a Lower triangular matrix \(\mathbf{L}\) and its transpose \(\mathbf{L}^T\) (or an upper triangular \(\mathbf{R}\) and its transpose). Very efficient (\(\sim \frac{1}{2}\) cost of LU) for solving systems with such matrices (common in statistics, e.g., with covariance matrices) and for Monte Carlo simulation of correlated variables.
Eigenvalues/vectors and singular values/vectors are key outputs of specific decompositions.
[[Orthogonal Matrix|Orthogonal matrices]] (like \(\mathbf{U}, \mathbf{V}, \mathbf{Q}\)) play a crucial role as they represent rotations/reflections and preserve lengths and angles (\(\|\mathbf{Qx}\|_2 = \|\mathbf{x}\|_2\)), leading to numerical stability.
[[Eigendecomposition|Eigendecomposition]] (\(\mathbf{A=V\Lambda V^{-1}}\)): For diagonalizable square matrices, reveals invariant directions (eigenvectors) and scaling factors (eigenvalues).
SVD (\(\mathbf{A=U\Sigma V^T}\)): For any matrix, reveals principal directions and scaling factors (singular vectors/values). Foundational for PCA, low-rank approximation.
Other types (LU, QR, Cholesky) exist for specific computational tasks (solving systems, least squares).