Adaptive Equalization

Learning the Channel on the Fly

All equalizers discussed so far assume the channel h[n]h[n] is known. In practice, the channel must be estimated and may change over time (e.g., due to mobility). Adaptive equalization algorithms update the equalizer coefficients iteratively, either using a known training sequence or operating in decision-directed mode where past symbol decisions serve as the reference. The two dominant algorithms are the least mean squares (LMS) algorithm, prized for its simplicity, and the recursive least squares (RLS) algorithm, which converges faster at the cost of higher complexity.

Definition:

Least Mean Squares (LMS) Algorithm

The LMS algorithm is a stochastic gradient descent method that adapts the equalizer tap vector w\mathbf{w} to minimise E[∣ek∣2]E[|e_k|^2], where ek=dkβˆ’wHyke_k = d_k - \mathbf{w}^H \mathbf{y}_k is the error between the desired output dkd_k and the equalizer output.

At each time step kk:

  1. Compute the output: a^k=wkHyk\hat{a}_k = \mathbf{w}_k^H \mathbf{y}_k
  2. Compute the error: ek=dkβˆ’a^ke_k = d_k - \hat{a}_k
  3. Update the taps: wk+1=wk+μ ekβˆ—β€‰yk\mathbf{w}_{k+1} = \mathbf{w}_k + \mu\, e_k^*\, \mathbf{y}_k

where ΞΌ>0\mu > 0 is the step size and dkd_k is either the known training symbol (training mode) or the hard decision dec(a^k)\text{dec}(\hat{a}_k) (decision-directed mode).

The LMS algorithm has complexity O(Nf)O(N_f) per symbol --- one multiply-accumulate per tap.

The LMS update is the instantaneous (sample-by-sample) approximation to the steepest descent algorithm wk+1=wk+ΞΌ(pβˆ’Ryywk)\mathbf{w}_{k+1} = \mathbf{w}_k + \mu (\mathbf{p} - \mathbf{R}_{yy} \mathbf{w}_k). The noise in the gradient estimate is the price paid for avoiding the computation of Ryy\mathbf{R}_{yy} and p\mathbf{p}.

Definition:

Recursive Least Squares (RLS) Algorithm

The RLS algorithm minimises the exponentially weighted least-squares cost

Jk(w)=βˆ‘i=0kΞ»kβˆ’iβ€‰βˆ£diβˆ’wHyi∣2J_k(\mathbf{w}) = \sum_{i=0}^{k} \lambda^{k-i}\, |d_i - \mathbf{w}^H \mathbf{y}_i|^2

where 0<λ≀10 < \lambda \leq 1 is the forgetting factor.

The RLS update is:

  1. Gain vector: kk=Pkβˆ’1 ykΞ»+ykHPkβˆ’1 yk\mathbf{k}_k = \frac{\mathbf{P}_{k-1}\, \mathbf{y}_k}{\lambda + \mathbf{y}_k^H \mathbf{P}_{k-1}\, \mathbf{y}_k}

  2. Error: ek=dkβˆ’wkβˆ’1Hyke_k = d_k - \mathbf{w}_{k-1}^H \mathbf{y}_k

  3. Tap update: wk=wkβˆ’1+kk ekβˆ—\mathbf{w}_k = \mathbf{w}_{k-1} + \mathbf{k}_k\, e_k^*

  4. Inverse correlation update: Pk=1Ξ»(Pkβˆ’1βˆ’kk ykH Pkβˆ’1)\mathbf{P}_k = \frac{1}{\lambda}(\mathbf{P}_{k-1} - \mathbf{k}_k\, \mathbf{y}_k^H\, \mathbf{P}_{k-1})

RLS converges much faster than LMS (independent of the eigenvalue spread of Ryy\mathbf{R}_{yy}) but has complexity O(Nf2)O(N_f^2) per symbol.

Definition:

Training Mode

In training mode, the receiver knows the transmitted symbols dk=akd_k = a_k during a preamble or midamble period. The adaptive algorithm uses these known symbols as the desired output to converge the equalizer taps. Training mode provides reliable convergence but reduces throughput because the training symbols carry no user data.

Definition:

Decision-Directed Mode

In decision-directed (DD) mode, the desired output is set to the hard decision on the equalizer output: dk=dec(a^k)d_k = \text{dec}(\hat{a}_k). This allows the equalizer to track slow channel variations without interrupting data transmission.

Decision-directed mode works reliably only when the BER is already low (typically below 10βˆ’210^{-2}), because incorrect decisions corrupt the adaptation. In practice, the equalizer first converges using a training sequence, then switches to decision-directed mode for tracking.

LMS Adaptive Equalizer

Complexity: Time: O(Nf)O(N_f) per symbol (one complex multiply-accumulate per tap). Memory: O(Nf)O(N_f) for the tap vector and input buffer.
Input: Received samples {yk}\{y_k\}, training symbols {dk}\{d_k\}, step size ΞΌ\mu, filter length NfN_f
Output: Equalizer coefficients w\mathbf{w}, symbol estimates {a^k}\{\hat{a}_k\}
1. Initialise: w0=[0,…,0,1,0,…,0]T\mathbf{w}_0 = [0, \ldots, 0, 1, 0, \ldots, 0]^T (centre-spike or all-zeros)
2. For k=0,1,2,…k = 0, 1, 2, \ldots:
a. Form input vector: yk=[yk,ykβˆ’1,…,ykβˆ’Nf+1]T\mathbf{y}_k = [y_k, y_{k-1}, \ldots, y_{k-N_f+1}]^T
b. Compute output: a^k=wkHyk\hat{a}_k = \mathbf{w}_k^H \mathbf{y}_k
c. Determine desired signal:
- Training mode: dk=akd_k = a_k (known)
- Decision-directed: dk=dec(a^k)d_k = \text{dec}(\hat{a}_k)
d. Compute error: ek=dkβˆ’a^ke_k = d_k - \hat{a}_k
e. Update taps: wk+1=wk+μ ekβˆ—β€‰yk\mathbf{w}_{k+1} = \mathbf{w}_k + \mu\, e_k^*\, \mathbf{y}_k
3. Output final w\mathbf{w} and decisions {a^k}\{\hat{a}_k\}.

The step size ΞΌ\mu must satisfy 0<ΞΌ<2/(Ξ»max⁑Nf)0 < \mu < 2/(\lambda_{\max} N_f) for convergence, where Ξ»max⁑\lambda_{\max} is the largest eigenvalue of Ryy\mathbf{R}_{yy}. A common rule of thumb is ΞΌβ‰ˆ1/(5NfΟƒy2)\mu \approx 1/(5 N_f \sigma_y^2).

LMS Convergence Animation

Watch the LMS algorithm converge in real time. The animation shows the equalizer tap values and the MSE learning curve as a function of iteration number. Adjust the step size to see the trade-off between convergence speed and steady-state misadjustment.

Parameters
0.02
0.5
0
20
11

Example: LMS Step Size and Convergence

An LMS equalizer with Nf=7N_f = 7 taps operates on a channel with autocorrelation eigenvalues λmax⁑=3.2\lambda_{\max} = 3.2 and λmin⁑=0.4\lambda_{\min} = 0.4.

(a) Determine the range of stable step sizes.

(b) Compute the convergence time constant for ΞΌ=0.05\mu = 0.05.

(c) Compute the steady-state excess MSE (misadjustment).

Quick Check

What is the main advantage of the RLS algorithm over LMS for adaptive equalization?

RLS has lower computational complexity per symbol

RLS converges faster, especially for channels with large eigenvalue spread

RLS achieves lower steady-state MSE

RLS does not require a training sequence

Common Mistake: Choosing the LMS Step Size

Mistake:

Setting the LMS step size ΞΌ\mu too large for fast convergence without checking the stability condition, causing the algorithm to diverge.

Correction:

The step size must satisfy ΞΌ<2/tr(Ryy)\mu < 2 / \text{tr}(\mathbf{R}_{yy}). A practical guideline is ΞΌβ‰ˆ1/(5NfΟƒy2)\mu \approx 1 / (5 N_f \sigma_y^2), which provides a good trade-off between convergence speed and misadjustment. If convergence is too slow, increase ΞΌ\mu cautiously and monitor the learning curve. If the MSE starts increasing, ΞΌ\mu is too large.

Common Mistake: Premature Switch to Decision-Directed Mode

Mistake:

Switching from training mode to decision-directed mode before the equalizer has sufficiently converged, causing the equalizer to lock onto a wrong solution or diverge.

Correction:

Decision-directed mode requires the BER to be below approximately 10βˆ’210^{-2} for reliable adaptation. Always verify that the training-mode MSE has converged to a sufficiently low level before switching. In fast-fading environments, use longer training sequences or pilot-aided adaptation rather than pure decision-directed tracking.

LMS vs. RLS Adaptive Algorithms

PropertyLMSRLS
Complexity per symbolO(Nf)O(N_f)O(Nf2)O(N_f^2)
Convergence speedDepends on eigenvalue spread Ο‡\chiIndependent of Ο‡\chi
Tuning parameterStep size ΞΌ\muForgetting factor Ξ»\lambda
Tracking abilityGood for slow variationsBetter for fast variations
Numerical stabilityVery stableCan suffer from finite-precision issues
MemoryO(Nf)O(N_f)O(Nf2)O(N_f^2) (stores Pk\mathbf{P}_k)
Typical applicationLow-complexity receiversFast convergence scenarios

Deeper Treatment in the FSI Book

The Wiener filter and its adaptive implementations (LMS, RLS) are covered in depth in the FSI book (Chapters 6–8), which treats the general LMMSE estimation framework, Kalman filtering for time-varying channels, and convergence analysis with full measure-theoretic rigor. The FSP book (Chapter 8) covers the spectral factorisation underlying the MMSE-DFE from the stochastic processes perspective.

Least Mean Squares (LMS)

A stochastic gradient descent algorithm that adapts filter coefficients by updating in the direction of the instantaneous gradient of the squared error. Complexity is O(Nf)O(N_f) per sample.

Related: Recursive Least Squares (RLS), Equalization

Recursive Least Squares (RLS)

An adaptive filtering algorithm that recursively minimises a weighted least-squares cost function. Converges faster than LMS but requires O(Nf2)O(N_f^2) operations per sample.

Related: Least Mean Squares (LMS), Equalization

Training Mode

An operating mode of an adaptive equalizer in which known pilot/training symbols are used as the desired output for coefficient adaptation.

Related: Decision-Directed Mode

Decision-Directed Mode

An operating mode of an adaptive equalizer in which the hard decision on the equalizer output is used as the desired signal for continued adaptation, enabling tracking without training overhead.

Related: Training Mode