Stationarity

Why Stationarity Matters

The full statistical description of a process requires the infinite family of fdds β€” an unwieldy object. Stationarity is the assumption that the statistical properties of the process do not change with time. Under this assumption, a single first-order distribution suffices (it is the same at all times), a single second-order distribution characterized by the lag Ο„=t1βˆ’t2\tau = t_1 - t_2 suffices, and so on. This enormous simplification is what makes statistical signal processing tractable. In communications, stationarity holds (at least approximately) for thermal noise, for a fading channel observed within its coherence time, and for the output of any time-invariant system driven by a stationary input.

Definition:

Strict-Sense Stationarity (Definition 49)

A stochastic process {X(t):t∈T}\{X(t) : t \in \mathcal{T}\} is strict-sense stationary (SSS) if its finite-dimensional distributions are invariant under time shifts: for every Nβ‰₯1N \geq 1, every (t1,…,tN)∈TN(t_1, \ldots, t_N) \in \mathcal{T}^N, and every shift Ο„\tau such that (t1+Ο„,…,tN+Ο„)∈TN(t_1 + \tau, \ldots, t_N + \tau) \in \mathcal{T}^N, Ft1,…,tN(x1,…,xN)=Ft1+Ο„,…,tN+Ο„(x1,…,xN).F_{t_1, \ldots, t_N}(x_1, \ldots, x_N) = F_{t_1 + \tau, \ldots, t_N + \tau}(x_1, \ldots, x_N).

Strict stationarity is a very strong condition: it requires every aspect of the joint distribution to be time-invariant. In practice, we rarely verify SSS directly because it involves all orders of the fdds. Instead, we work with the weaker but far more practical notion of wide-sense stationarity.

Consequences of SSS

If X(t)X(t) is SSS, then:

  1. First-order distribution is constant: FX(t)(x)=FX(0)(x)F_{X(t)}(x) = F_{X(0)}(x) for all tt. In particular, ΞΌX(t)=ΞΌ\mu_X(t) = \mu and Var(X(t))=Οƒ2\text{Var}(X(t)) = \sigma^2 are constants.

  2. Second-order distribution depends only on lag: FX(t1),X(t2)(x1,x2)F_{X(t_1), X(t_2)}(x_1, x_2) depends only on t1βˆ’t2t_1 - t_2.

  3. All NNth-order statistics depend only on time differences: the NN-point statistics are invariant to a common time shift.

SSS β‡’\Rightarrow WSS (but not conversely, in general).

Definition:

Wide-Sense Stationarity (Definition 52)

A second-order process {X(t):t∈T}\{X(t) : t \in \mathcal{T}\} is wide-sense stationary (WSS) if:

  1. ΞΌX(t)=E[X(t)]=ΞΌ\mu_X(t) = \mathbb{E}[X(t)] = \mu is constant (independent of tt), and
  2. rXX(t1,t2)=E[X(t1)Xβˆ—(t2)]r_{XX}(t_1, t_2) = \mathbb{E}[X(t_1)X^*(t_2)] depends only on the time difference Ο„=t1βˆ’t2\tau = t_1 - t_2.

We write rXX(Ο„)β‰œrXX(t,tβˆ’Ο„)=E[X(t)Xβˆ—(tβˆ’Ο„)]r_{XX}(\tau) \triangleq r_{XX}(t, t - \tau) = \mathbb{E}[X(t)X^*(t - \tau)].

For a discrete-time process, {Xn}\{X_n\} is WSS if ΞΌ[n]=ΞΌ\mu[n] = \mu (constant) and rxx[n,m]=rxx[nβˆ’m]r_{xx}[n, m] = r_{xx}[n - m] depends only on the lag k=nβˆ’mk = n - m.

WSS Is the Engineer's Stationarity

Wide-sense stationarity involves only two conditions β€” constant mean and lag-dependent autocorrelation β€” both of which can be estimated from data. This is why WSS is the default assumption in signal processing and communications: it is strong enough to enable powerful tools (power spectral density, Wiener filtering, matched filtering) but weak enough to be approximately satisfied in many practical scenarios.

The gap between WSS and SSS is bridged for Gaussian processes, as Lemma 43 below shows.

Theorem: SSS Implies WSS (for Second-Order Processes)

If {X(t)}\{X(t)\} is strict-sense stationary and E[∣X(t)∣2]<∞\mathbb{E}[|X(t)|^2] < \infty, then {X(t)}\{X(t)\} is wide-sense stationary.

SSS makes all fdds time-shift invariant. In particular, the first-order moment (from the first-order distribution) and the second-order product moment (from the second-order distribution) inherit this invariance, which is exactly the WSS condition.

Definition:

Gaussian Process (Definition 51)

A process {X(t):t∈T}\{X(t) : t \in \mathcal{T}\} is Gaussian if for every Nβ‰₯1N \geq 1 and every (t1,…,tN)∈TN(t_1, \ldots, t_N) \in \mathcal{T}^N, the random vector (X(t1),…,X(tN))⊀(X(t_1), \ldots, X(t_N))^\top is jointly Gaussian.

A Gaussian process is completely determined by its mean function ΞΌX(t)\mu_X(t) and its autocovariance function cXX(t1,t2)=Cov(X(t1),X(t2))c_{XX}(t_1, t_2) = \text{Cov}(X(t_1), X(t_2)).

Theorem: WSS Gaussian Process Is SSS (Lemma 43)

If {X(t)}\{X(t)\} is a Gaussian process and wide-sense stationary, then {X(t)}\{X(t)\} is strict-sense stationary.

A Gaussian distribution is completely determined by its mean vector and covariance matrix. If the mean is constant and the covariance depends only on time differences (WSS), then the entire joint distribution is shift-invariant β€” which is SSS.

Example: WSS but Not SSS

Let XX be a random variable with E[X]=0\mathbb{E}[X] = 0, E[X2]=1\mathbb{E}[X^2] = 1, but XX is not Gaussian (e.g., XX takes values Β±1\pm 1 each with probability 1/21/2). Define the process Yn=XY_n = X for all nn. Is {Yn}\{Y_n\} WSS? Is it SSS?

WSS vs. Non-WSS Processes

Compare WSS and non-WSS processes. The WSS process has constant mean and lag-dependent autocorrelation. The non-WSS process has time-varying statistics.

Parameters
5
7

Strict-Sense vs. Wide-Sense Stationarity

PropertySSSWSS
ConditionAll fdds shift-invariantConstant mean + lag-dependent rXXr_{XX}
InvolvesAll moments and distributionsOnly 1st and 2nd moments
Testable from data?Generally noYes β€” estimate mean and autocorrelation
SSS β‡’\Rightarrow WSS?Yes (if 2nd moment exists)β€”
WSS β‡’\Rightarrow SSS?No (in general)Yes, if Gaussian
Practical useTheoretical idealDefault assumption in signal processing

Quick Check

Let X(t)=B+W(t)X(t) = B + W(t) where B∼N(0,ΟƒB2)B \sim \mathcal{N}(0, \sigma_B^2) is a constant random bias and W(t)W(t) is a zero-mean WSS process independent of BB. Is X(t)X(t) WSS?

Yes, because the sum of WSS processes is WSS

Yes: the mean is constant and the autocorrelation depends only on Ο„\tau

No, because BB is a random constant that biases each realization differently

Common Mistake: Assuming WSS Implies SSS

Mistake:

Concluding that a WSS process has time-invariant distributions of all orders.

Correction:

WSS guarantees only that the first two moments are time-invariant. Higher-order statistics may still vary with time. The only general case where WSS implies SSS is for Gaussian processes (Lemma 43), because their distributions are fully determined by the first two moments.

Common Mistake: Forgetting That WSS Implies Constant Variance

Mistake:

Computing Var(X(t))\text{Var}(X(t)) for a WSS process and getting a time-dependent answer.

Correction:

For a WSS process, Var(X(t))=rXX(0)βˆ’βˆ£ΞΌβˆ£2\text{Var}(X(t)) = r_{XX}(0) - |\mu|^2 is the same for all tt. If your calculation gives a time-dependent variance, check whether the process is truly WSS.

⚠️Engineering Note

WSS and the Coherence Time

In mobile wireless communications, the channel is modeled as WSS only over a time interval called the coherence time TcT_c. Beyond TcT_c, the channel statistics change due to mobility and environmental changes. A typical design rule is to place pilot symbols at intervals shorter than TcT_c so that channel estimation can exploit the WSS assumption. The wide-sense stationary uncorrelated scattering (WSSUS) model, introduced by Bello (1963), formalizes this for doubly-selective (time-frequency) channels.

Wide-Sense Stationary (WSS)

A process with constant mean and autocorrelation that depends only on the time difference. The standard assumption in linear signal processing and communications.

Related: Strict-Sense Stationary (SSS)

Strict-Sense Stationary (SSS)

A process whose finite-dimensional distributions are invariant under time shifts. Stronger than WSS; equivalent to WSS for Gaussian processes.

Related: Wide-Sense Stationary (WSS)

Key Takeaway

Wide-sense stationarity β€” constant mean and lag-dependent autocorrelation β€” is the practical form of stationarity used throughout signal processing and communications. For Gaussian processes, WSS and SSS are equivalent, which is why the Gaussian assumption is so powerful: second-order statistics tell the whole story.