Trace, Determinant, and Matrix Inequalities
Why Trace and Determinant Are Everywhere in Capacity Expressions
Every capacity formula in multi-antenna communications rests on two matrix functionals: the trace and the determinant.
The MIMO capacity formula. For an channel with input covariance and noise variance , the mutual information is The capacity-achieving is found by maximizing this subject to a trace constraint: The trace constrains total transmit power; the log-determinant measures information rate. Understanding the interplay between the two β through trace identities, Hadamard's inequality, Fischer's inequality, and the concavity of β is essential for deriving capacity bounds, water-filling solutions, and beamforming strategies.
Where these tools appear:
- Trace identities simplify covariance manipulations: .
- Hadamard inequality bounds the determinant by the product of diagonal entries, yielding capacity upper bounds when the channel decomposes into independent sub-channels.
- Fischer inequality provides tighter bounds when the channel has block structure (e.g., multi-user MIMO).
- Log-det concavity guarantees that the MIMO capacity optimization is a convex program (maximizing a concave function over a convex set), ensuring that water-filling is globally optimal.
Definition: Trace of a Matrix
Trace of a Matrix
The trace of a square matrix is the sum of its diagonal entries:
The trace satisfies the following fundamental properties:
-
Linearity. For any and :
-
Cyclic property. For and : More generally, the trace is invariant under cyclic permutations of a matrix product (see TCyclic Property of the Trace).
-
Relation to eigenvalues. If are the eigenvalues of (counted with algebraic multiplicity), then
-
Similarity invariance. For any invertible : This follows immediately from the cyclic property.
-
Transpose and conjugate transpose. and .
In telecommunications, represents total transmit power when is the input covariance matrix. The power constraint is therefore a trace constraint.
Definition: Determinant of a Matrix
Determinant of a Matrix
The determinant of a square matrix , denoted or , is the unique multilinear, alternating function on the columns of satisfying . Equivalently, via Leibniz's formula: where is the symmetric group on .
Fundamental properties:
-
Product of eigenvalues. If are the eigenvalues of , then
-
Multiplicativity. For :
-
Transpose and conjugate transpose. and . In particular, when is Hermitian.
-
Invertibility. is invertible if and only if .
-
Determinant of a block triangular matrix. If , then .
-
Positive definiteness. If , then (since all eigenvalues are strictly positive).
In capacity formulas, converts a product of eigenvalue contributions into a sum: . This is why the MIMO capacity decomposes into a sum of per-eigenmode rates.
Theorem: Cyclic Property of the Trace
Let , , , be matrices such that the product is square. Then the trace is invariant under cyclic permutations: In particular, for three square matrices: .
Note: The trace is not invariant under arbitrary permutations. In general, .
A cyclic permutation moves the first matrix to the end. The trace "sees" a product of matrices as a closed loop of index contractions, and rotating the starting point of a loop does not change the loop itself.
Step 1: Prove the two-matrix case
Let and . Then is and is . We compute each trace using the definition: Both double sums run over the same set of index pairs and involve the same summands . By commutativity of scalar multiplication and the fact that the order of a finite sum is immaterial, the two expressions are equal:
Step 2: Extend to $k$ matrices by induction
For matrices, define and . Then and . By the two-matrix case: Repeating this argument times produces all cyclic permutations.
Theorem: Hadamard's Inequality
Let be positive definite () with diagonal entries . Then with equality if and only if is diagonal.
Equivalently, if are the columns of , then , with equality if and only if the columns are mutually orthogonal.
The determinant measures the volume of the parallelepiped spanned by the columns. If columns are not orthogonal, the parallelepiped "collapses" partially, reducing its volume below the product of the column lengths. For a positive definite matrix, the diagonal entries are the squared norms of the "contributions" in each coordinate direction, and any off-diagonal correlation reduces the determinant.
Step 1: Cholesky factorization setup
Since , it admits a unique Cholesky factorization , where is lower triangular with strictly positive diagonal entries . By the multiplicativity of determinants:
Step 2: Relate diagonal entries of $\mathbf{A}$ to $\mathbf{L}$
The entry of is (Here is real and positive by the Cholesky construction.)
Since all terms in the sum are nonneg-ative and is one of them, we obtain with equality if and only if for all , i.e., if and only if the -th row of has only the diagonal entry nonzero.
Step 3: Combine to get the inequality
Taking the product over all : This proves .
Step 4: Equality condition
Equality holds if and only if for every , which by Step 2 occurs if and only if for all and all . This means is diagonal, so is also diagonal.
Conversely, if is diagonal (with positive diagonal entries, since ), then trivially.
Hadamard's Inequality: Volume of a Parallelepiped
Theorem: Concavity of on the Positive Definite Cone
The function is concave on the cone of positive definite matrices . That is, for any and :
Since , the log-det is a sum of concave () functions of the eigenvalues. However, eigenvalues of a convex combination are not simply convex combinations of eigenvalues, so the proof requires more care. The key idea is to reduce to a one-dimensional argument by examining and showing it is concave in .
Step 1: Reduce to a one-variable problem
A function on is concave if and only if it is concave along every line segment in . Fix and a Hermitian matrix , and define for all such that . It suffices to show that is concave in .
Step 2: Factor out $\mathbf{A}$
Write Taking the log-det: where is Hermitian. Since is a constant in , it suffices to show that is concave in .
Step 3: Diagonalize $\mathbf{C}$
Since is Hermitian, by the spectral theorem there exists a unitary such that . Then and (This is valid for in the range where all .)
Step 4: Verify concavity of each summand
Each function has second derivative Since for all in the domain, each is concave.
Step 5: Conclude
A sum of concave functions is concave. Therefore is concave in , which implies is concave in . Since this holds for every line in , is concave on .
Theorem: Fischer's Inequality
Let be positive definite and partitioned as where and are the diagonal blocks. Then with equality if and only if (i.e., the two blocks are uncorrelated).
Fischer's inequality says that correlations between two groups of variables (captured by the off-diagonal block ) can only reduce the determinant, never increase it. Hadamard's inequality is the special case where each block is .
Step 1: Schur complement factorization
Since , the diagonal block is itself positive definite (as a principal submatrix of a positive definite matrix). The Schur complement of in is
The block LDU factorization of gives
We verify this by direct multiplication. The block: . The block: . The block: . The block: .
Step 2: Take determinants
The two outer matrices in the LDU factorization are unit lower and upper triangular, so each has determinant . Therefore:
Step 3: Bound the Schur complement
Since , the Schur complement is also positive definite (a standard result: if and only if and ).
Now, , so (it is positive semidefinite for any ). Therefore:
For positive definite matrices, implies . (Proof of this auxiliary fact: write ; the matrix in the middle has eigenvalues in , so its determinant is .)
Step 4: Combine and state the equality condition
From Steps 2 and 3:
Equality holds if and only if , which occurs if and only if . Since , this holds if and only if .
Example: Verifying Trace Identities and Hadamard's Inequality for a Matrix
Let (a) Verify the cyclic property of the trace for . (b) Compute and verify Hadamard's inequality. (c) Verify Fisher's inequality for the / partition.
Step 1: Basic quantities
First, is real symmetric with positive diagonal entries. To confirm , we check leading principal minors:
- .
- .
- .
So by Sylvester's criterion.
. .
Step 2: Cyclic property verification
Compute :
.
Alternatively, by the eigenvalue relation: . The eigenvalues of satisfy , , and . However, a direct computation from the coefficient of in the characteristic polynomial confirms .
Now, the cyclic property says , which is trivially true for a product of the same matrix. For a non-trivial check, let . Then: . .
Step 3: Hadamard's inequality
Hadamard's inequality states .
We have:
- .
- .
Indeed .
The inequality is strict because is not diagonal (it has nonzero off-diagonal entries and ). The ratio quantifies how much the off-diagonal correlations reduce the determinant.
Step 4: Fischer's inequality
Partition with and , so .
Fischer's inequality gives: Indeed .
Note that Fischer's bound () is tighter than Hadamard's bound () because Fischer accounts for the correlation within the top-left block, while Hadamard treats every entry independently.
The Schur complement is , and indeed .
of a Parameterized Positive Definite Matrix
Explore how varies with , demonstrating the concavity of the log-det function. The plot shows as a function of for several matrix types. Observe that the curve always bends downward (concavity), and that ill-conditioned matrices exhibit more pronounced concavity due to the spread of their eigenvalues.
Parameters
Scaling parameter
Why This Matters: Log-Det in the MIMO Capacity Formula
The central object in MIMO information theory is the log-det capacity formula:
The results of this section connect directly to this formula:
-
Log-det concavity ( on the Positive Definite Cone" data-ref-type="theorem">TConcavity of on the Positive Definite Cone) guarantees that the capacity optimization is a concave maximization over the convex set . Any local maximum is therefore the global maximum, and the water-filling solution is provably optimal.
-
Hadamard's inequality (THadamard's Inequality) yields the bound , showing that capacity is maximized when the effective channel sub-channels are uncorrelated β achieved by diagonalizing through the SVD of .
-
Fischer's inequality (TFischer's Inequality) provides capacity bounds for block-structured channels, such as multi-user MIMO systems where the channel matrix has a natural block partition corresponding to different users.
-
Trace identities (TCyclic Property of the Trace) are used repeatedly in manipulating covariance expressions: .
See full treatment in Physical MIMO Channel Modeling
Key Takeaway
The trace constrains power () while the log-determinant measures information rate (). The inequalities of this section β Hadamard, Fischer, and log-det concavity β are the analytical backbone of every capacity bound in MIMO communications. Hadamard bounds capacity from above by the sum of independent sub-channel rates; Fischer refines this for block-structured channels; and log-det concavity ensures that capacity optimization is a tractable convex program with a unique global optimum (water-filling).
Common Mistake: Trace of a Product Is Not the Product of Traces
Mistake:
A common error is to assume that . This is analogous to confusing "the sum of products" with "the product of sums." Students who have internalized the linearity of trace sometimes extend it incorrectly to multiplicativity.
Correction:
The trace is linear but NOT multiplicative. In general,
Counterexample. Let , . Then , so , but .
What is true is the cyclic property: . See TCyclic Property of the Trace.
Quick Check
Let be a positive definite matrix with , , . By Hadamard's inequality, what is the tightest upper bound on that the inequality provides?
(Hadamard gives no finite bound)
Hadamard's inequality states . Equality holds if and only if is diagonal.
Quick Check
Which of the following statements about on the positive definite cone is TRUE?
is convex.
is concave.
is neither convex nor concave.
is linear.
This is precisely the content of on the Positive Definite Cone" data-ref-type="theorem">TConcavity of on the Positive Definite Cone. The concavity of is what makes MIMO capacity optimization a tractable convex program (maximizing a concave function).
Trace
The sum of the diagonal entries of a square matrix: . Equivalently, the sum of the eigenvalues. The trace is linear, similarity-invariant, and satisfies the cyclic property . In wireless communications, the trace of the input covariance matrix equals total transmit power.
Related: Determinant of a Matrix, Eigenvalue and Eigenvector, power constraint
Hadamard Inequality
For a positive definite matrix , the bound , with equality if and only if is diagonal. Geometrically, the volume of the parallelepiped spanned by the columns of is maximized when the columns are orthogonal. In MIMO systems, this bounds the capacity by the sum of independent sub-channel capacities.
Related: Determinant of a Matrix, Fischer's Inequality, Positive Definite Matrix
Log-Det Concavity
The property that is a concave function on the cone of positive definite matrices. This ensures that MIMO capacity optimization (maximizing subject to a convex power constraint) is a convex program, guaranteeing global optimality of the water-filling solution.
Related: concave function, convex optimization, MIMO Capacity (Deterministic Channel)
Computing log-det in Practice: Cholesky, Not Eigendecomposition
Never compute by finding all eigenvalues first β that costs for the eigendecomposition plus the risk of numerical overflow/underflow in the product. Instead, use the Cholesky factorization (which requires ):
Cholesky costs flops β half the cost of a full eigendecomposition β and the sum of logarithms avoids overflow. In MIMO capacity computation, this is the standard approach: numpy.linalg.slogdet(A) uses Cholesky internally.
- β’
Cholesky: flops vs eigendecomposition: flops
- β’
For MIMO: Cholesky takes ~90K flops, eigendecomposition ~1M flops
- β’
numpy.linalg.slogdet returns (sign, log|det|) β numerically stable for any matrix
Historical Note: Hadamard's Inequality and the Maximum Determinant Problem
1893Jacques Hadamard proved his determinant inequality in 1893, establishing that the determinant of a positive definite matrix is bounded by the product of its diagonal entries. The geometric interpretation β the volume of a parallelepiped is maximized when its edges are orthogonal β was already implicit in the work of Gram (1883), but Hadamard gave the sharp algebraic bound.
The related Hadamard maximum determinant problem β finding the matrix with entries in that maximizes β remains open for most . Hadamard matrices (those achieving equality with entries ) exist only when , or is a multiple of 4, and their existence for all such is a famous unsolved conjecture.
Why This Matters: Trace and Determinant in Information-Theoretic Capacity Bounds
The trace and log-determinant are the two fundamental matrix functionals in information theory. The trace constrains power (), while the log-determinant measures rate (). The inequalities of this section β Hadamard, Fischer, and log-det concavity β are the analytical backbone of capacity analysis in Book ITA (Chapters 13-16) and Book MIMO (Chapters 1-5).