Random Variables and Distributions

Why Random Variables?

A probability space (Ξ©,F,P)(\Omega, \mathcal{F}, P) gives us the formal machinery to talk about random experiments, but the sample space Ξ©\Omega can be unwieldy --- outcomes might be waveforms, sequences of channel states, or abstract protocol events. To compute, we need numbers.

A random variable is the bridge: it maps every outcome Ο‰βˆˆΞ©\omega \in \Omega to a real number, converting abstract randomness into something we can integrate, differentiate, and optimise.

In wireless communications the idea is immediate. The received baseband sample at a single antenna is

r=s+n,r = s + n,

where ss is the transmitted symbol (deterministic, once the codebook is fixed) and nn is additive noise --- a Gaussian random variable. Understanding the distribution of nn lets us compute bit-error rates, design detectors, and prove capacity theorems. Every subsequent chapter builds on this section's definitions.

Definition:

Random Variable

Let (Ξ©,F,P)(\Omega, \mathcal{F}, P) be a probability space. A random variable is a measurable function

X:Ξ©β†’R,X : \Omega \to \mathbb{R},

meaning that for every Borel set BβŠ†RB \subseteq \mathbb{R} the pre-image Xβˆ’1(B)={Ο‰βˆˆΞ©:X(Ο‰)∈B}X^{-1}(B) = \{\omega \in \Omega : X(\omega) \in B\} belongs to the Οƒ\sigma-algebra F\mathcal{F}.

In particular, the set {X≀x}β‰œ{Ο‰βˆˆΞ©:X(Ο‰)≀x}\{X \le x\} \triangleq \{\omega \in \Omega : X(\omega) \le x\} is an event (element of F\mathcal{F}) for every x∈Rx \in \mathbb{R}, so P(X≀x)P(X \le x) is well-defined.

The measurability condition is automatically satisfied for every function encountered in engineering practice. Its role is to prevent pathological constructions that would make probabilities undefined.

We write "XX" (uppercase italic) for a random variable and "xx" (lowercase italic) for a particular realisation.

,

Definition:

Cumulative Distribution Function (CDF)

The cumulative distribution function (CDF) of a random variable XX is the function FX:R→[0,1]F_X : \mathbb{R} \to [0, 1] defined by

FX(x)=P(X≀x),x∈R.F_X(x) = P(X \le x), \qquad x \in \mathbb{R}.

Properties. Every CDF satisfies:

  1. Right-continuity: FXF_X is right-continuous at every point, i.e., lim⁑h↓0FX(x+h)=FX(x)\lim_{h \downarrow 0} F_X(x + h) = F_X(x).
  2. Monotonicity: FXF_X is non-decreasing: x1<x2β€…β€ŠβŸΉβ€…β€ŠFX(x1)≀FX(x2)x_1 < x_2 \;\Longrightarrow\; F_X(x_1) \le F_X(x_2).
  3. Boundary limits: lim⁑xβ†’βˆ’βˆžFX(x)=0\displaystyle\lim_{x \to -\infty} F_X(x) = 0 and lim⁑xβ†’+∞FX(x)=1\displaystyle\lim_{x \to +\infty} F_X(x) = 1.

Conversely, any function satisfying properties 1--3 is the CDF of some random variable (the Skorokhod construction).

The CDF completely characterises the distribution of XX. Two random variables with the same CDF are said to be equal in distribution, written X=dYX \stackrel{d}{=} Y.

,

Definition:

Probability Density Function (PDF)

A random variable XX is called continuous (or, more precisely, absolutely continuous) if there exists a non-negative function fX:Rβ†’[0,∞)f_X : \mathbb{R} \to [0, \infty) such that

FX(x)=βˆ«βˆ’βˆžxfX(t) dt,βˆ€β€‰x∈R.F_X(x) = \int_{-\infty}^{x} f_X(t)\,dt, \qquad \forall\, x \in \mathbb{R}.

The function fXf_X is called the probability density function (PDF) of XX. At every point where FXF_X is differentiable,

fX(x)=dFX(x)dx.f_X(x) = \frac{dF_X(x)}{dx}.

Properties.

  1. fX(x)β‰₯0f_X(x) \ge 0 for all xx.
  2. βˆ«βˆ’βˆžβˆžfX(x) dx=1\displaystyle\int_{-\infty}^{\infty} f_X(x)\,dx = 1.
  3. P(a<X≀b)=∫abfX(x) dx\displaystyle P(a < X \le b) = \int_a^b f_X(x)\,dx.

The PDF is not a probability: fX(x)f_X(x) can exceed 11. Only when integrated over an interval does it yield a probability.

,

Definition:

Probability Mass Function (PMF)

A random variable XX is discrete if it takes values in a countable set X={x1,x2,… }\mathcal{X} = \{x_1, x_2, \dots\}. Its probability mass function (PMF) is

pX(x)=P(X=x),x∈X.p_X(x) = P(X = x), \qquad x \in \mathcal{X}.

Properties.

  1. pX(x)β‰₯0p_X(x) \ge 0 for all x∈Xx \in \mathcal{X}.
  2. βˆ‘x∈XpX(x)=1\displaystyle\sum_{x \in \mathcal{X}} p_X(x) = 1.

The CDF of a discrete random variable is a staircase function with jumps of size pX(xk)p_X(x_k) at each point xk∈Xx_k \in \mathcal{X}.

In digital communications, discrete random variables arise naturally: the transmitted bit b∈{0,1}b \in \{0, 1\}, the number of errors in a codeword, and the number of packet arrivals in a time slot are all discrete.

,

Definition:

Expectation (Mean)

The expectation (or mean) of a random variable XX is defined as

E[X]=βˆ«βˆ’βˆžβˆžx fX(x) dxE[X] = \int_{-\infty}^{\infty} x\,f_X(x)\,dx

for continuous XX (provided the integral converges absolutely), and

E[X]=βˆ‘x∈Xx pX(x)E[X] = \sum_{x \in \mathcal{X}} x\,p_X(x)

for discrete XX. We also write ΞΌX=E[X]\mu_X = E[X].

LOTUS (Law of the Unconscious Statistician). For any (measurable) function g:R→Rg : \mathbb{R} \to \mathbb{R},

E[g(X)]=βˆ«βˆ’βˆžβˆžg(x) fX(x) dxE[g(X)] = \int_{-\infty}^{\infty} g(x)\,f_X(x)\,dx

without needing to derive the distribution of g(X)g(X) first.

Linearity of expectation. For any random variables XX and YY (not necessarily independent) and constants a,b∈Ra, b \in \mathbb{R},

E[aX+bY]=a E[X]+b E[Y].E[aX + bY] = a\,E[X] + b\,E[Y].

LOTUS is one of the most practically useful results in probability. For example, the average received power is E[∣r∣2]E[|r|^2], which LOTUS lets us compute directly from the distribution of rr without first finding the distribution of ∣r∣2|r|^2.

,

Definition:

Variance and Standard Deviation

The variance of a random variable XX is

Var(X)=E[(Xβˆ’ΞΌX)2]=E[X2]βˆ’(E[X])2,\mathrm{Var}(X) = E\bigl[(X - \mu_X)^2\bigr] = E[X^2] - \bigl(E[X]\bigr)^2,

where ΞΌX=E[X]\mu_X = E[X]. We write ΟƒX2=Var(X)\sigma_X^2 = \mathrm{Var}(X).

The standard deviation is ΟƒX=Var(X)\sigma_X = \sqrt{\mathrm{Var}(X)}.

Properties.

  1. Var(X)β‰₯0\mathrm{Var}(X) \ge 0, with equality iff X=ΞΌXX = \mu_X a.s.
  2. Var(aX+b)=a2 Var(X)\mathrm{Var}(aX + b) = a^2\,\mathrm{Var}(X) for constants a,ba, b.
  3. If XX and YY are independent, then Var(X+Y)=Var(X)+Var(Y)\mathrm{Var}(X + Y) = \mathrm{Var}(X) + \mathrm{Var}(Y).

The identity Var(X)=E[X2]βˆ’(E[X])2\mathrm{Var}(X) = E[X^2] - (E[X])^2 is the computational workhorse. In noise analysis, if nn is zero-mean (E[n]=0E[n] = 0), then Var(n)=E[n2]\mathrm{Var}(n) = E[n^2], which is the noise power.

,

Theorem: Properties of Expectation

Let XX and YY be random variables on (Ω,F,P)(\Omega, \mathcal{F}, P) with finite expectations, and let a,b,c∈Ra, b, c \in \mathbb{R}. Then:

  1. Linearity: E[aX+bY+c]=a E[X]+b E[Y]+cE[aX + bY + c] = a\,E[X] + b\,E[Y] + c.

  2. Monotonicity: If X≀YX \le Y almost surely, then E[X]≀E[Y]E[X] \le E[Y].

  3. Triangle inequality: ∣E[X]βˆ£β‰€E[∣X∣]|E[X]| \le E[|X|].

  4. Product rule for independent RVs: If XX and YY are independent, then E[XY]=E[X] E[Y]E[XY] = E[X]\,E[Y].

Linearity is the single most important property of expectation and holds without any independence assumption. It is the reason we can decompose the mean of a sum of interference terms without knowing their joint distribution.

Monotonicity formalises the obvious: if one random variable always dominates another, its average must be at least as large.

The product rule for independent RVs is not a consequence of linearity; it requires the factorisation of the joint PDF.

,

Definition:

Bernoulli Distribution

A random variable XX has the Bernoulli distribution with parameter p∈[0,1]p \in [0, 1], written X∼Bernoulli(p)X \sim \mathrm{Bernoulli}(p), if

pX(x)={1βˆ’p,x=0,p,x=1,0,otherwise.p_X(x) = \begin{cases} 1 - p, & x = 0, \\ p, & x = 1, \\ 0, & \text{otherwise}. \end{cases}

Mean: E[X]=pE[X] = p.

Variance: Var(X)=p(1βˆ’p)\mathrm{Var}(X) = p(1 - p).

The Bernoulli distribution is the natural model for a single bit decision. In a binary symmetric channel with crossover probability pp, the error indicator e∼Bernoulli(p)e \sim \mathrm{Bernoulli}(p) takes the value 11 when the received bit differs from the transmitted bit.

Definition:

Binomial Distribution

A random variable XX has the binomial distribution with parameters n∈Nn \in \mathbb{N} and p∈[0,1]p \in [0,1], written X∼Binomial(n,p)X \sim \mathrm{Binomial}(n, p), if

pX(k)=(nk)pk(1βˆ’p)nβˆ’k,k=0,1,…,n.p_X(k) = \binom{n}{k} p^k (1-p)^{n-k}, \qquad k = 0, 1, \dots, n.

Mean: E[X]=npE[X] = np.

Variance: Var(X)=np(1βˆ’p)\mathrm{Var}(X) = np(1 - p).

If a block of nn coded bits is sent through a BSC with crossover probability pp, and errors are independent across bits, then the total number of bit errors K∼Binomial(n,p)K \sim \mathrm{Binomial}(n, p). The binomial distribution thus governs block error analysis for memoryless channels.

,

Definition:

Poisson Distribution

A random variable XX has the Poisson distribution with rate parameter λ>0\lambda > 0, written X∼Poisson(λ)X \sim \mathrm{Poisson}(\lambda), if

pX(k)=Ξ»keβˆ’Ξ»k!,k=0,1,2,…p_X(k) = \frac{\lambda^k e^{-\lambda}}{k!}, \qquad k = 0, 1, 2, \dots

Mean: E[X]=Ξ»E[X] = \lambda.

Variance: Var(X)=Ξ»\mathrm{Var}(X) = \lambda.

The Poisson distribution arises as the limit of Binomial(n,p)\mathrm{Binomial}(n, p) when nβ†’βˆžn \to \infty, pβ†’0p \to 0, and npβ†’Ξ»np \to \lambda. In telecommunications it models the number of packet arrivals in a time slot, the number of interferers in a stochastic geometry model, and the number of photons detected in an optical receiver.

,

Definition:

Continuous Uniform Distribution

A random variable XX has the uniform distribution on [a,b][a, b], written X∼Uniform(a,b)X \sim \mathrm{Uniform}(a, b), if its PDF is

fX(x)={1bβˆ’a,a≀x≀b,0,otherwise.f_X(x) = \begin{cases} \dfrac{1}{b - a}, & a \le x \le b, \\[6pt] 0, & \text{otherwise}. \end{cases}

Mean: E[X]=a+b2\displaystyle E[X] = \frac{a + b}{2}.

Variance: Var(X)=(bβˆ’a)212\displaystyle \mathrm{Var}(X) = \frac{(b-a)^2}{12}.

In wireless communications, the carrier phase offset Θ∼Uniform(0,2Ο€)\Theta \sim \mathrm{Uniform}(0, 2\pi) models the fact that the receiver has no a priori knowledge of the absolute phase. This assumption underlies non-coherent detection and is implicit in the derivation of the Rayleigh fading model.

Definition:

Exponential Distribution

A random variable XX has the exponential distribution with rate parameter λ>0\lambda > 0, written X∼Exp(λ)X \sim \mathrm{Exp}(\lambda), if

fX(x)={Ξ»eβˆ’Ξ»x,xβ‰₯0,0,x<0.f_X(x) = \begin{cases} \lambda e^{-\lambda x}, & x \ge 0, \\ 0, & x < 0. \end{cases}

CDF: FX(x)=1βˆ’eβˆ’Ξ»xF_X(x) = 1 - e^{-\lambda x} for xβ‰₯0x \ge 0.

Mean: E[X]=1Ξ»\displaystyle E[X] = \frac{1}{\lambda}.

Variance: Var(X)=1Ξ»2\displaystyle \mathrm{Var}(X) = \frac{1}{\lambda^2}.

Memoryless property. The exponential is the only continuous distribution with the memoryless property:

P(X>s+t∣X>s)=P(X>t),βˆ€β€‰s,tβ‰₯0.P(X > s + t \mid X > s) = P(X > t), \qquad \forall\, s, t \ge 0.

The exponential distribution models inter-arrival times in a Poisson process. In teletraffic engineering, the time between successive packet arrivals at a router is often modelled as Exp(Ξ»)\mathrm{Exp}(\lambda). Additionally, in Rayleigh fading the instantaneous SNR Ξ³=∣h∣2Ξ³Λ‰\gamma = |h|^2 \bar{\gamma} is exponentially distributed when h∼CN(0,1)h \sim \mathcal{CN}(0, 1).

,

Definition:

Gaussian (Normal) Distribution

A random variable XX has the Gaussian (or normal) distribution with mean μ∈R\mu \in \mathbb{R} and variance Οƒ2>0\sigma^2 > 0, written X∼N(ΞΌ,Οƒ2)X \sim \mathcal{N}(\mu, \sigma^2), if its PDF is

fX(x)=12πσ2exp⁑ ⁣(βˆ’(xβˆ’ΞΌ)22Οƒ2),x∈R.f_X(x) = \frac{1}{\sqrt{2\pi\sigma^2}} \exp\!\left(-\frac{(x - \mu)^2}{2\sigma^2}\right), \qquad x \in \mathbb{R}.

Mean: E[X]=ΞΌE[X] = \mu.

Variance: Var(X)=Οƒ2\mathrm{Var}(X) = \sigma^2.

The standard normal Z∼N(0,1)Z \sim \mathcal{N}(0, 1) is obtained by setting ΞΌ=0\mu = 0, Οƒ2=1\sigma^2 = 1. Any Gaussian can be standardised: Z=(Xβˆ’ΞΌ)/ΟƒZ = (X - \mu)/\sigma.

The Q-function. The tail probability of the standard normal is

Q(x)=P(Z>x)=∫x∞12Ο€eβˆ’t2/2 dt=12 erfc ⁣(x2),Q(x) = P(Z > x) = \int_x^{\infty} \frac{1}{\sqrt{2\pi}} e^{-t^2/2}\,dt = \frac{1}{2}\,\mathrm{erfc}\!\left(\frac{x}{\sqrt{2}}\right),

where erfc\mathrm{erfc} is the complementary error function. This function appears in virtually every error-probability expression in digital communications.

The Gaussian distribution owes its central role to the Central Limit Theorem (Section 2.5): the sum of many independent, identically distributed random variables converges in distribution to a Gaussian, regardless of the individual distributions. Thermal noise, which is the aggregate effect of countless electron vibrations, is therefore accurately modelled as Gaussian.

The complex Gaussian CN(ΞΌ,Οƒ2)\mathcal{CN}(\mu, \sigma^2) used for baseband noise has independent real and imaginary parts, each N(ΞΌr,Οƒ2/2)\mathcal{N}(\mu_r, \sigma^2/2) and N(ΞΌi,Οƒ2/2)\mathcal{N}(\mu_i, \sigma^2/2) respectively, so that the total power (variance) is Οƒ2\sigma^2.

, ,

Gaussian PDF and CDF Explorer

Visualise how the Gaussian PDF and CDF change as you vary the mean ΞΌ\mu and standard deviation Οƒ\sigma. Use the dropdown to overlay Rayleigh, Ricean, or Nakagami-mm densities --- these fading distributions, derived from the Gaussian in Section 2.3, are previewed here for comparison.

Parameters
0
1

Example: Bit-Error Rate of BPSK in AWGN

A binary phase-shift keying (BPSK) system transmits s∈{+ ⁣Eb,β€…β€Šβˆ’β€‰β£Eb}s \in \{+\!\sqrt{E_b},\;-\!\sqrt{E_b}\} with equal probability over an AWGN channel. The received signal is

r=s+n,n∼N ⁣(0,N02).r = s + n, \qquad n \sim \mathcal{N}\!\left(0, \frac{N_0}{2}\right).

The receiver decides s^=+ ⁣Eb\hat{s} = +\!\sqrt{E_b} if r>0r > 0 and s^=βˆ’β€‰β£Eb\hat{s} = -\!\sqrt{E_b} if r≀0r \le 0. Derive the bit-error rate (BER) as a function of Eb/N0E_b/N_0.

,

The Gaussian Distribution: Shape, Scale, and the 68-95-99.7 Rule

Watch how the Gaussian PDF transforms as Οƒ\sigma sweeps from 0.5 to 3.0, with the 68-95-99.7 rule illustrated as shaded regions.
The Gaussian N(ΞΌ,Οƒ2)\mathcal{N}(\mu, \sigma^2) concentrates probability near the mean. As Οƒ\sigma grows, the bell curve flattens and spreads, but the area under it remains 1.

Why This Matters: The Q-Function and Error Probability in Digital Communications

The Q-function is the universal currency for error-rate computation in AWGN channels. For any modulation scheme whose conditional error event reduces to a Gaussian tail probability, the BER or symbol-error rate (SER) can be expressed in terms of Q(β‹…)Q(\cdot).

Examples:

Modulation Error probability
BPSK Pb=Q ⁣(2Eb/N0)P_b = Q\!\left(\sqrt{2E_b/N_0}\right)
QPSK Psβ‰ˆ2 Q ⁣(2Es/N0)P_s \approx 2\,Q\!\left(\sqrt{2E_s/N_0}\right)
MM-QAM Psβ‰ˆ4 ⁣(1βˆ’1M)Q ⁣(3Es(Mβˆ’1)N0)P_s \approx 4\!\left(1 - \frac{1}{\sqrt{M}}\right) Q\!\left(\sqrt{\frac{3E_s}{(M-1)N_0}}\right)

The Q-function decreases faster than any polynomial but slower than a pure exponential. Its asymptotic approximation for large xx is

Q(x)β‰ˆ12π x eβˆ’x2/2,x≫1,Q(x) \approx \frac{1}{\sqrt{2\pi}\,x}\,e^{-x^2/2}, \qquad x \gg 1,

which is useful for high-SNR analysis and diversity-order calculations.

Relation to erfc. The Q-function and complementary error function are related by

Q(x)=12 erfc ⁣(x2),erfc(x)=2 Q ⁣(x2).Q(x) = \frac{1}{2}\,\mathrm{erfc}\!\left(\frac{x}{\sqrt{2}}\right), \qquad \mathrm{erfc}(x) = 2\,Q\!\left(x\sqrt{2}\right).

Most numerical libraries provide erfc\mathrm{erfc} directly, making this conversion essential for simulation code.

See full treatment in Chapter 5، Section 2

Quick Check

If fX(x)=c x eβˆ’x2f_X(x) = c\,x\,e^{-x^2} for xβ‰₯0x \ge 0 (and fX(x)=0f_X(x) = 0 for x<0x < 0), what value of cc makes fXf_X a valid PDF?

c=1c = 1

c=2c = 2

c=1/2c = 1/2

c=4c = 4

Quick Check

Let X∼N(3,4)X \sim \mathcal{N}(3, 4). What is E[2X+5]E[2X + 5]?

1111

66

1313

1616

Quick Check

If Var(X)=Οƒ2\mathrm{Var}(X) = \sigma^2, what is Var(3Xβˆ’7)\mathrm{Var}(3X - 7)?

3Οƒ2βˆ’73\sigma^2 - 7

3Οƒ23\sigma^2

9Οƒ29\sigma^2

9Οƒ2βˆ’79\sigma^2 - 7

Random variable

A measurable function X:Ξ©β†’RX : \Omega \to \mathbb{R} mapping outcomes of a random experiment to real numbers. Enables computation of probabilities, expectations, and distributions.

Related: Random Variable, Cumulative Distribution Function (CDF), Probability Density Function (PDF)

Probability density function (PDF)

A non-negative function fX(x)f_X(x) whose integral over any interval [a,b][a, b] gives P(a<X≀b)P(a < X \le b). It is the derivative of the CDF wherever the derivative exists.

Related: Probability Density Function (PDF), Cumulative Distribution Function (CDF)

Cumulative distribution function (CDF)

The function FX(x)=P(X≀x)F_X(x) = P(X \le x), which completely characterises the distribution of a random variable. It is non-decreasing, right-continuous, with limits 00 and 11.

Related: Cumulative Distribution Function (CDF), Probability Density Function (PDF)

Expectation

The weighted average E[X]=∫x fX(x) dxE[X] = \int x\,f_X(x)\,dx (continuous) or E[X]=βˆ‘xx pX(x)E[X] = \sum_x x\,p_X(x) (discrete). A linear functional of the random variable's distribution.

Related: Expectation (Mean), Properties of Expectation

Variance

Var(X)=E[(Xβˆ’E[X])2]\mathrm{Var}(X) = E[(X - E[X])^2], a measure of the spread of XX around its mean. Equals E[X2]βˆ’(E[X])2E[X^2] - (E[X])^2. Units are the square of the units of XX.

Related: Variance and Standard Deviation

Q-function

The tail probability of the standard normal distribution: Q(x)=P(Z>x)=12 erfc(x/2)Q(x) = P(Z > x) = \frac{1}{2}\,\mathrm{erfc}(x/\sqrt{2}) where Z∼N(0,1)Z \sim \mathcal{N}(0,1). Ubiquitous in error-probability expressions for digital modulation over AWGN channels.

Related: Gaussian (Normal) Distribution, The Q-Function and Error Probability in Digital Communications, Bit-Error Rate of BPSK in AWGN

Key Takeaway

The core message of this section in three bullets:

  1. Random variables turn experiments into numbers. The measurable function X:Ξ©β†’RX : \Omega \to \mathbb{R} is the formal link between abstract probability spaces and the integrals, expectations, and variances that drive engineering design.

  2. Six distributions cover most of wireless. Bernoulli and binomial for error counting, Poisson for arrivals, uniform for unknown phase, exponential for inter-arrival times and Rayleigh SNR, and Gaussian for thermal noise --- these six appear on almost every page of a communications textbook.

  3. The Gaussian is the workhorse noise model. Because thermal noise is the sum of myriad independent microscopic contributions, the Central Limit Theorem guarantees it is Gaussian. The Q-function, the tail integral of the standard normal, is the single most important special function in digital communications: every BER formula you will encounter in Chapters 5--8 is built from it.

Common Mistake: Forgetting the 12\frac{1}{2} in the Q-function vs erfc Relationship

Mistake:

Writing Q(x)=erfc(x/2)Q(x) = \mathrm{erfc}(x/\sqrt{2}) without the factor of 1/21/2, or equivalently erfc(x)=Q(x2)\mathrm{erfc}(x) = Q(x\sqrt{2}) instead of erfc(x)=2 Q(x2)\mathrm{erfc}(x) = 2\,Q(x\sqrt{2}).

This error is extremely common in homework and even published papers. It leads to BER predictions that are off by exactly a factor of 22, which can silently propagate through simulation comparisons.

Correction:

The correct relations are:

Q(x)=12 erfc ⁣(x2),erfc(x)=2 Q ⁣(x2).Q(x) = \frac{1}{2}\,\mathrm{erfc}\!\left(\frac{x}{\sqrt{2}}\right), \qquad \mathrm{erfc}(x) = 2\,Q\!\left(x\sqrt{2}\right).

A useful sanity check: Q(0)=1/2Q(0) = 1/2 and erfc(0)=1\mathrm{erfc}(0) = 1. Without the factor of 1/21/2 one would get Q(0)=1Q(0) = 1, which is clearly wrong since it would imply P(Z>0)=1P(Z > 0) = 1 for a symmetric distribution.

Tip: Always verify your formula at x=0x = 0 as a quick consistency test.

Common Mistake: Interpreting the PDF Value as a Probability

Mistake:

Claiming that fX(x0)f_X(x_0) is the probability that X=x0X = x_0. For continuous random variables, P(X=x0)=0P(X = x_0) = 0 for every individual point. Yet it is common to see statements like "the probability of receiving r=1.5r = 1.5 is fR(1.5)=0.35f_R(1.5) = 0.35."

Correction:

The PDF is a density, not a probability. Only when integrated over an interval does it yield a probability:

P(a<X≀b)=∫abfX(x) dx.P(a < X \le b) = \int_a^b f_X(x)\,dx.

Informally, fX(x0) dxf_X(x_0)\,dx is the probability that XX falls in an infinitesimally small interval around x0x_0. In particular, fX(x)f_X(x) can exceed 11 (e.g., Uniform(0,0.5)\mathrm{Uniform}(0, 0.5) has fX(x)=2f_X(x) = 2 on its support).