Eigenvalue Decomposition
Why Eigenvalues Matter
Eigenvalues and eigenvectors expose the intrinsic geometry of a linear transformation. When you decompose a matrix into its eigenstructure you discover the directions along which the map acts as simple scaling — the natural modes of the system.
In telecommunications this idea is pervasive:
- Channel covariance. The eigenvalues of the spatial covariance matrix quantify how much energy the channel concentrates along each angular direction.
- MIMO capacity. The capacity-achieving input covariance is diagonal in the eigenbasis of ; the water-filling solution allocates power to eigenmodes whose eigenvalues (channel gains) exceed a threshold.
- Vibration and stability. Eigenvalues of a system matrix determine whether feedback loops oscillate, converge, or diverge — a concept inherited directly from the theory of differential equations.
- Principal Component Analysis (PCA). The dominant eigenvectors of a data covariance matrix are the directions of maximum variance, used for dimensionality reduction in signal processing and machine learning.
Understanding eigendecomposition is therefore not a detour; it is the very language in which channel capacity, beamforming optimality, and adaptive filtering are expressed.
Definition: Eigenvalue and Eigenvector
Eigenvalue and Eigenvector
Let . A scalar is an eigenvalue of if there exists a nonzero vector such that The vector is called an eigenvector associated with . The set of all eigenvalues is the spectrum of , denoted .
Eigenvalues are the roots of the characteristic polynomial , which is a degree- polynomial in . By the Fundamental Theorem of Algebra, has exactly roots (counted with multiplicity) in .
Definition: Characteristic Polynomial
Characteristic Polynomial
The characteristic polynomial of is It is a monic polynomial of degree . Its roots are precisely the eigenvalues of .
The algebraic multiplicity of an eigenvalue is its multiplicity as a root of . The geometric multiplicity is , i.e., the dimension of the corresponding eigenspace. One always has
When the algebraic and geometric multiplicities coincide for every eigenvalue, the matrix is diagonalizable.
Definition: Eigendecomposition (Diagonalization)
Eigendecomposition (Diagonalization)
A matrix is diagonalizable if there exists an invertible matrix and a diagonal matrix such that The columns of are eigenvectors of , and the diagonal entries of are the corresponding eigenvalues.
A matrix is diagonalizable if and only if it possesses linearly independent eigenvectors.
Not every matrix is diagonalizable. For example, the matrix has eigenvalue with algebraic multiplicity but geometric multiplicity . However, every Hermitian (or more generally, normal) matrix is diagonalizable, and moreover can be diagonalized by a unitary matrix.
Definition: Hermitian Matrix
Hermitian Matrix
A matrix is Hermitian (or self-adjoint) if where denotes the conjugate transpose.
In particular, the diagonal entries of a Hermitian matrix are real, and the off-diagonal entries satisfy .
A Hermitian matrix with real entries is called symmetric.
In wireless communications, covariance matrices and Gram matrices are always Hermitian (and positive semidefinite).
Theorem: Eigenvalues of Hermitian Matrices Are Real
Let be Hermitian. Then every eigenvalue of is real.
Proof
Let be an eigenvalue with eigenvector , so . Left-multiply both sides by : Since , we have , so Now take the conjugate of both sides. Using : Hence , which means .
Theorem: Spectral Theorem for Hermitian Matrices
Every Hermitian matrix can be decomposed as where is unitary () and with all . Equivalently, , where is an orthonormal basis of eigenvectors.
A Hermitian matrix acts by stretching space along its eigenvectors (which are orthogonal) by real scaling factors (the eigenvalues). It has no "rotational" component. The spectral decomposition separates the geometry (the orthonormal eigenvectors) from the magnitude of the action (the real eigenvalues).
First show eigenvalues of Hermitian matrices must be real.
Then show eigenvectors for distinct eigenvalues are orthogonal.
Use Schur's theorem to guarantee a unitary triangularization, then show the triangular factor must be diagonal.
Alternatively, use induction on matrix dimension via deflation.
Step 1: Eigenvalues are real
This was established in TEigenvalues of Hermitian Matrices Are Real. For completeness: if with , then where the second equality uses and the scalar identity . Hence .
Step 2: Eigenvectors for distinct eigenvalues are orthogonal
Let and with . Compute: where the last equality uses (Step 1). Rearranging: Since , we conclude , i.e., .
Step 3: Induction on dimension via deflation
We prove the full spectral decomposition by strong induction on .
Base case (). A Hermitian matrix has , and the decomposition is trivially .
Inductive step. Assume the theorem holds for all Hermitian matrices of size . Let be Hermitian.
Existence of an eigenvalue. The characteristic polynomial has degree and therefore possesses at least one root . By Step 1, . Let be a corresponding unit eigenvector: , .
Deflation. Extend to an orthonormal basis of (e.g., via Gram--Schmidt). Form the unitary matrix . Then for some and .
Because is Hermitian, so is (since ). The Hermitian property forces (already known) and . It also forces to equal the conjugate transpose of , which is automatically satisfied. However, a sharper constraint comes from the first column: Comparing with the first column of the block form gives . Therefore
Apply the inductive hypothesis. Since is an Hermitian matrix, by the induction hypothesis there exists a unitary and a real diagonal such that .
Step 4: Assembling the full decomposition
Define Then is unitary and which is real and diagonal.
Setting (a product of unitary matrices, hence unitary), we obtain
Writing , this becomes which completes the proof.
Step 5: Handling repeated eigenvalues
The induction above does not assume that eigenvalues are distinct. If has algebraic multiplicity , then the deflated matrix still has as an eigenvalue (with multiplicity ), and the inductive hypothesis produces orthonormal eigenvectors for it.
In the case of a repeated eigenvalue, the eigenspace has dimension equal to the algebraic multiplicity (for Hermitian matrices, geometric and algebraic multiplicities always coincide). Any orthonormal basis of serves as a valid set of eigenvectors.
This non-uniqueness of eigenvectors within a repeated eigenspace is a feature, not a deficiency: it grants freedom in choosing convenient bases — a freedom exploited, for instance, when jointly diagonalizing commuting Hermitian matrices.
Definition: Rayleigh Quotient
Rayleigh Quotient
For a Hermitian matrix and nonzero , the Rayleigh quotient is The Rayleigh quotient is scale-invariant: for any .
The Rayleigh quotient is bounded between the smallest and largest eigenvalues of (see TExtremal Properties of the Rayleigh Quotient). In wireless communications, SNR expressions such as (signal-to-noise ratio after beamforming with weight vector ) are generalized Rayleigh quotients. Maximizing the SNR thus reduces to an eigenvalue problem.
Theorem: Extremal Properties of the Rayleigh Quotient
Let be Hermitian with eigenvalues and corresponding orthonormal eigenvectors . Then for all nonzero : Moreover,
Since eigenvectors form an orthonormal basis, any vector is a weighted combination of eigenvectors. The Rayleigh quotient computes a weighted average of the eigenvalues, where the weights are the squared magnitudes of the expansion coefficients. A weighted average always lies between the minimum and maximum values being averaged.
Step 1: Expand in the eigenbasis
By the spectral theorem (TSpectral Theorem for Hermitian Matrices), write . Any nonzero can be expressed as where . Define .
Since is unitary, .
Step 2: Express the Rayleigh quotient as a weighted average
Compute: Therefore Define weights with . Then .
Step 3: Establish the bounds
Since and , the Rayleigh quotient is a convex combination of the eigenvalues. Therefore: and similarly
Step 4: The bounds are attained
For : , for , giving .
For : , for , giving .
Hence the upper and lower bounds are both achieved.
Theorem: Courant–Fischer Min-Max Theorem
Let be Hermitian with eigenvalues . Then for each :
The -th eigenvalue is the "best worst case" of the Rayleigh quotient over all -dimensional subspaces. This characterization is extremely powerful because it does not reference eigenvectors — it describes eigenvalues purely through optimization.
Step 1: Upper bound via a specific subspace
Let . For any with : since for . Thus , and equality is achieved at . This gives
Step 2: Matching upper bound for an arbitrary $k$-dimensional subspace
Let be any -dimensional subspace. Define , which has dimension . By a dimension argument: so contains a nonzero vector .
Since , write . Then since for .
Since , we have .
As this holds for every -dimensional :
Step 3: Conclusion
Combining Steps 1 and 2: The dual (min-max) characterization follows by applying the max-min result to (whose eigenvalues are ) and substituting .
Eigenvectors as Fixed Directions of a Linear Map
Visualize how a 2×2 Hermitian matrix transforms vectors. Eigenvectors (shown in red) maintain their direction, only getting scaled by their eigenvalues. The unit circle is mapped to an ellipse whose semi-axes align with the eigenvectors; the lengths of the semi-axes equal the absolute values of the eigenvalues.
Parameters
Top-left entry
Real part of off-diagonal
Imaginary part of off-diagonal
Bottom-right entry
Power Iteration Converging to Dominant Eigenvector
Watch how the power iteration converges to the dominant eigenvector. At each step, the iterate is shown as a blue arrow; the true dominant eigenvector is shown in red. A bar chart tracks the Rayleigh quotient converging to .
Parameters
Ratio of dominant to subdominant eigenvalue (controls convergence speed)
Example: Eigendecomposition of a 2×2 Hermitian Matrix
Find the eigendecomposition of
Full Solution
Step 1 — Verify Hermitian. Check: , and diagonal entries are real. So .
Step 2 — Characteristic polynomial. Note: .
Step 3 — Eigenvalues. So and . Both real, as guaranteed by the spectral theorem.
Step 4 — Eigenvectors. For : From the first row: , so . Setting and normalizing:
For , a similar computation yields the orthogonal eigenvector .
Step 5 — Verify orthogonality. One can confirm directly, consistent with the spectral theorem.
Step 6 — Decomposition.
Sanity checks:
- .
- .
Example: Eigendecomposition of a 3×3 Real Symmetric Matrix
Find the eigenvalues of
Full Solution
Step 1 — Characteristic polynomial. Expanding along the first row:
Step 2 — Eigenvalues. , , .
Step 3 — Sanity checks.
- .
- .
Step 4 — Eigenvectors. For : solve :
For : solve :
For : solve :
One may verify for and .
Historical Note: From Vibrating Strings to Abstract Algebra: The History of Eigenvalues
1740s–1910The concept of eigenvalues emerged gradually over two centuries of mathematics:
Euler and d'Alembert (1740s–1750s). The study of coupled oscillations and the rotation of rigid bodies led Euler to consider equations of the form . The term "secular equation" (from the Latin saecularis, relating to long-period astronomical perturbations) was used for the characteristic polynomial.
Cauchy (1829). Augustin-Louis Cauchy proved that the eigenvalues of a real symmetric matrix are real, establishing the first rigorous version of what would become the spectral theorem. He also showed that symmetric matrices have orthogonal eigenvectors.
Sylvester (1852). James Joseph Sylvester introduced the term "matrix" (from the Latin for "womb") and developed much of the algebraic machinery for determinants and characteristic polynomials.
Hilbert (1904–1910). David Hilbert extended the spectral theorem to infinite-dimensional spaces (integral operators), founding the field of functional analysis. His spectral theory is the mathematical backbone of quantum mechanics and modern signal processing.
The word "Eigenwert" (own value) was coined by Hilbert in 1904. The English hybrid "eigenvalue" preserves the German prefix, a testament to the concept's Germanic origins.
Common Mistake: Common Misconceptions About Eigendecomposition
Mistake:
Several common misconceptions about eigendecomposition can lead to errors in analysis and computation.
Correction:
1. Eigendecomposition requires a square matrix. Only square matrices () have eigenvalues. For rectangular matrices, use the singular value decomposition (SVD) instead (see §Singular Value Decomposition).
2. Eigenvectors are not always orthogonal. For a general (non-Hermitian) matrix, eigenvectors corresponding to distinct eigenvalues need not be orthogonal. Orthogonality is guaranteed only for normal matrices (), of which Hermitian matrices are a special case.
3. Not every matrix is diagonalizable. Defective matrices (where geometric multiplicity < algebraic multiplicity for some eigenvalue) cannot be diagonalized. The Jordan normal form handles such matrices but is numerically unstable and rarely used in practice.
4. Eigenvalues of a product products of eigenvalues (in general). unless and share a common eigenbasis (e.g., when they commute and are both diagonalizable).
5. Numerical eigenvalue computation never uses . Finding roots of the characteristic polynomial is numerically unstable. Practical algorithms (QR iteration, divide-and-conquer) work directly with the matrix and achieve backward-stable results.
Practical Eigenvalue Algorithms and Their Costs
Never compute eigenvalues via the characteristic polynomial — it is numerically catastrophic even for matrices. Production algorithms:
- QR iteration (general matrices): per step, converges in steps. Total: worst case, typical.
- Divide-and-conquer (symmetric/Hermitian): average, fastest for moderate . Used by
numpy.linalg.eigh. - Lanczos/Arnoldi (large sparse): for the top eigenvalues, where is the number of nonzero entries. Used by
scipy.sparse.linalg.eigsh. For MIMO covariance matrices (),numpy.linalg.eigh(LAPACK's divide-and-conquer) is optimal. For massive MIMO spatial correlation matrices (), use Lanczos.
- •
LAPACK dsyevd (divide-and-conquer): ~4n^3 flops for all eigenvalues
- •
For channel covariance estimation in 5G NR: → eigendecomposition costs ~130K flops (negligible)
- •
For XL-MIMO with : eigendecomposition costs ~4G flops (requires optimization)
Power Iteration: When Simple Beats Sophisticated
The power iteration converges to the dominant eigenvector at rate per iteration. Each iteration costs (one matrix-vector product). Despite its simplicity:
- It is the method of choice when only the largest eigenvalue is needed (e.g., spectral radius for stability analysis, dominant singular value for condition estimation).
- Google's PageRank is essentially power iteration on a sparse matrix.
- In wireless: the dominant eigenvector of gives the optimal beamforming direction. Power iteration computes it with per step — much cheaper than full eigendecomposition. Convergence is slow when . In this case, use inverse iteration (converges as for the smallest eigenvalue) or Rayleigh quotient iteration (cubic convergence).
- •
Power iteration per step: one matrix-vector product ( dense, sparse)
- •
Typical convergence: 10–50 iterations for
Key Takeaway
A Hermitian matrix is nothing more than scaling along orthogonal directions: says that stretches each eigenvector by the real factor and does nothing else.
Implications for computation: Matrix powers become trivial (), the inverse is (when it exists), and any analytic function is simply .
Why This Matters: Eigenvalues and MIMO Channel Capacity
Consider a narrowband MIMO channel with transmit and receive antennas, modeled as , where is the channel matrix and .
When channel state information is available at the transmitter, the capacity is The Hermitian matrix has eigendecomposition , where . The optimal input covariance is , and the capacity simplifies to where and the powers are chosen by water-filling: with set to satisfy the power constraint .
Key insight: The eigenvalues of determine the channel gains of the independent spatial modes. Large eigenvalues correspond to strong modes that receive more power; weak modes (small ) may be shut off entirely. The eigendecomposition transforms the coupled MIMO channel into a set of parallel, independent scalar channels — the conceptual simplification that makes MIMO tractable.
Quick Check
Let . What are the eigenvalues of ?
First verify is Hermitian: and . The characteristic polynomial is . So and , both real as expected.
Quick Check
A Hermitian matrix has eigenvalues . What is the maximum value of over all nonzero ?
By the extremal property of the Rayleigh quotient (TExtremal Properties of the Rayleigh Quotient), , achieved at the eigenvector corresponding to .
Quick Check
Which of the following matrices is guaranteed to be diagonalizable by a unitary matrix?
Any invertible matrix
Any upper triangular matrix
Any Hermitian matrix
Any matrix with distinct eigenvalues
The spectral theorem (TSpectral Theorem for Hermitian Matrices) guarantees that every Hermitian matrix has a unitary eigendecomposition . While a matrix with distinct eigenvalues is diagonalizable, the diagonalizing matrix need not be unitary. More generally, a matrix is unitarily diagonalizable if and only if it is normal ().
Quick Check
If has eigenvalues , what is ?
3
5
15
19
The eigenvalues of are the squares of the eigenvalues of . Therefore .
Eigenvalue
A scalar such that for some nonzero vector . Equivalently, a root of the characteristic polynomial . For Hermitian matrices, all eigenvalues are real.
Spectral Theorem
The result stating that every Hermitian matrix admits a decomposition where is unitary and is real diagonal. This guarantees the existence of an orthonormal eigenbasis with real eigenvalues.
Rayleigh Quotient
For a Hermitian matrix and nonzero vector , the ratio . It is bounded by the extreme eigenvalues of and equals an eigenvalue when is a corresponding eigenvector.