Joint PMFs and PDFs
Why Joint Distributions?
In Chapters 5 and 6 we studied a single random variable at a time. But in virtually every engineering problem, multiple quantities interact: the signal and the noise, the channel gain and the interference, the transmit power and the received SNR. To reason about how two or more random variables relate to each other β whether they are dependent, how one conditions the other, what happens when we add or transform them β we need the joint distribution.
The marginal distributions and alone do not determine the joint distribution . The joint distribution is a strictly richer object: it encodes all marginals, all conditionals, and all dependence structure.
Definition: Joint Cumulative Distribution Function
Joint Cumulative Distribution Function
Let and be random variables defined on a common probability space . The joint CDF of is the function defined by
The joint CDF is the fundamental object from which all other joint distributional quantities are derived. It always exists, regardless of whether the RVs are discrete, continuous, or mixed.
Theorem: Properties of the Joint CDF
Let be the joint CDF of . Then:
-
Limits: for all , for all , and .
-
Monotonicity: is non-decreasing in each argument.
-
Right-continuity: is right-continuous in each argument.
-
Marginals: and .
-
Rectangle probability: For and ,
Limits and monotonicity
Property 1 follows from continuity of probability applied to increasing/decreasing sequences of events. For instance, when , so .
Monotonicity in : if then , so .
Right-continuity
Fix and let . Then , so by continuity of probability from above. The argument for is identical.
Marginals and rectangle formula
Letting , the event , so .
The rectangle formula follows by inclusion-exclusion: is the intersection of and , and expanding via the CDF gives the alternating sum. Non-negativity follows because this is a probability.
Definition: Joint Probability Mass Function
Joint Probability Mass Function
Let and be discrete random variables with supports and . The joint PMF is
satisfying for all and .
The joint PMF can be displayed as a table (or matrix) indexed by the values of and . Row sums give the marginal ; column sums give the marginal .
Definition: Marginal PMF from Joint PMF
Marginal PMF from Joint PMF
Given the joint PMF , the marginal PMFs are obtained by summing over the other variable:
The marginal PMFs are proper PMFs: each is non-negative and sums to 1.
Example: Joint PMF β Weather in Two Cities
Let and denote the weather in Los Angeles and San Francisco, respectively, where = sunny and = cloudy. The joint PMF is given by the table:
| 0 | 1 | |
|---|---|---|
| 0 | 0.2 | 0.5 |
| 1 | 0.1 | 0.2 |
Find the marginal PMFs and compute .
Marginal of $X$
, . So LA is sunny with probability 0.7.
Marginal of $Y$
, .
At least one sunny
The complement is "both cloudy": . Therefore .
Definition: Joint Probability Density Function
Joint Probability Density Function
Two random variables and are jointly continuous if their joint CDF can be expressed as
for some non-negative function called the joint probability density function. Equivalently,
wherever the mixed partial derivative exists.
The joint PDF satisfies and .
The value is not a probability β it is a density. The probability of falling in a region is .
Definition: Marginal PDF from Joint PDF
Marginal PDF from Joint PDF
Given the joint PDF , the marginal PDFs are
Geometrically, the marginal is obtained by integrating ("projecting") the joint density along the -axis.
Example: Bivariate Uniform on a Triangle
Let be uniformly distributed on the triangle , which has area . Find the joint PDF, the marginal PDFs, and .
Joint PDF
The area of the triangle is , so the uniform density is for and zero otherwise.
Marginal of $X$
for .
Marginal of $Y$
for .
Compute $\mathbb{P}(Y > X/2)$
The region is , i.e., :
Joint probability density function
A non-negative function whose double integral over any region gives .
Related: Marginal distribution
Marginal distribution
The distribution of a single random variable obtained from a joint distribution by integrating (or summing) over all other variables.
Related: Joint probability density function
Joint PDF and Marginal Projections
Joint PDF Contour Plot with Marginals
Explore the joint density of a bivariate Gaussian with adjustable means, variances, and correlation coefficient . The marginal densities are displayed on the side panels.
Parameters
Common Mistake: Marginals Do Not Determine the Joint Distribution
Mistake:
Assuming that knowing and is enough to determine .
Correction:
Infinitely many joint distributions share the same marginals. The joint distribution encodes the dependence structure between and , which the marginals alone cannot capture. For instance, two standard Gaussian marginals can be paired with any correlation to produce different bivariate Gaussian distributions.
Quick Check
If for and zero otherwise, what is for ?
.
Historical Note: The Origins of Multivariate Distributions
1880sβ1933The study of joint distributions began in earnest with Francis Galton's work on regression and correlation in the 1880s. Galton noticed that the heights of fathers and sons formed an elliptical scatter pattern β the hallmark of a bivariate Gaussian. Karl Pearson formalized this observation into the multivariate normal distribution and introduced the correlation coefficient that we still use today. The generalization to arbitrary joint distributions, via the joint CDF, came later with Kolmogorov's axiomatization of probability in 1933.
Key Takeaway
The joint distribution determines the marginals and (by integration), but the converse is false. The joint distribution is a strictly richer object that encodes the full dependence structure between the random variables.