Exercises
ex-ch34-01
EasyApply the three-pass reading method to any recent paper from IEEE Transactions on Wireless Communications (or IEEE Communications Letters). After Pass 1 only, answer:
- What is the paper's claimed contribution in one sentence?
- What performance metric is used?
- How many baselines are compared in the main results figure?
- Based on Pass 1 alone, is the paper relevant to MIMO precoding research? Justify your answer.
Focus on the abstract, contributions list (end of introduction), and figure captions.
You should be able to answer all four questions in under 5 minutes.
Grading rubric
This exercise has no single correct answer β it depends on the chosen paper. The grading criteria are:
- Contribution sentence should capture the novelty, not just the topic (e.g., "proposes an iterative algorithm for RIS phase optimization" not "studies RIS").
- Metric should be specific: "ergodic sum rate" not just "performance."
- Baseline count should be verifiable from the figure legend.
- Relevance judgment should reference specific aspects (system model, antenna configuration, metric) rather than vague statements.
ex-ch34-02
EasyA Monte Carlo BER simulation of BPSK over AWGN at dB produces the following results from 5 independent runs (each with transmitted bits):
| Run | Bit errors | Estimated BER |
|---|---|---|
| 1 | 23 | |
| 2 | 18 | |
| 3 | 25 | |
| 4 | 15 | |
| 5 | 29 |
- Compute the pooled BER estimate (combining all 5 runs).
- Compute the 95% confidence interval.
- The theoretical BER of BPSK at dB is . Does your CI contain the theoretical value? What might explain the discrepancy?
Pool the errors: total errors = sum of all runs, total bits = .
CI formula: .
Pooled estimate
Total errors: . Total bits: . .
Confidence interval
.
95% CI: , i.e., .
Comparison with theory
The theoretical value lies outside the 95% CI. Possible explanations:
- The simulation has a bug (most likely: incorrect SNR normalization or using instead of ).
- The random number generator is not producing true Gaussian noise.
- Rounding in the SNR conversion.
This demonstrates how CIs can catch simulation errors: when the CI does not contain the known analytical answer, something is wrong.
ex-ch34-03
EasyList the five essential components of a wireless system model (from Definition 34.1.3). For each component, give one example of a common assumption and one example of an unrealistic assumption that should be flagged in a paper review.
The five components are: signal model, channel model, CSI assumption, design variable, performance metric.
System model components with examples
| Component | Common assumption | Unrealistic flag |
|---|---|---|
| Signal model | Linear precoding: | Noiseless: |
| Channel model | i.i.d. Rayleigh fading | Purely LoS for urban micro at 2 GHz |
| CSI assumption | MMSE-estimated with pilot overhead | Perfect CSI at TX for FDD massive MIMO |
| Design variable | ZF precoder | Joint precoder + user location optimization |
| Metric | Ergodic sum rate | "Capacity" without specifying assumptions |
The "unrealistic flag" column shows assumptions that should prompt skepticism during a paper review.
ex-ch34-04
MediumA paper claims that its proposed NOMA scheme achieves 40% higher throughput than OFDMA. Upon reading the simulation setup, you discover:
- NOMA: 2 users per subcarrier, perfect SIC, perfect CSI at TX/RX
- OFDMA: 1 user per subcarrier, no CSI at TX, equal power allocation
- Both: 10 subcarriers, users, Rayleigh fading
Identify at least three sources of unfairness and explain how each one biases the comparison in favor of NOMA. Propose a fair experimental setup.
Consider CSI, power allocation, and user scheduling separately.
What would OFDMA performance look like with the same CSI that NOMA has?
Sources of unfairness
1. CSI asymmetry: NOMA uses perfect CSIT for power allocation and SIC ordering; OFDMA has no CSIT. With CSIT, OFDMA can perform frequency-selective scheduling (assigning each subcarrier to the user with the best channel), which dramatically improves throughput.
2. Power allocation asymmetry: NOMA presumably uses optimized power allocation (needed for SIC); OFDMA uses equal power. With CSIT, OFDMA can use water-filling across subcarriers.
3. Perfect SIC assumption: In practice, SIC suffers from error propagation, especially with imperfect CSI. This assumption inflates NOMA performance.
4. User pairing advantage: NOMA pairs users with different channel strengths on the same subcarrier. The comparison does not account for the scheduling gain that OFDMA would get from multi-user diversity.
Fair experimental setup
- Give both schemes equal CSIT (e.g., MMSE-estimated channels with the same pilot overhead).
- Allow OFDMA to use frequency-selective scheduling and water-filling power allocation.
- Model imperfect SIC for NOMA with realistic error propagation.
- Compare at the same total power and same number of served users.
- Include an upper bound (e.g., DPC capacity) to gauge how close either scheme comes to optimal.
ex-ch34-05
MediumYou are simulating a MIMO system with ZF detection. Your simulation produces the following BER curve:
| (dB) | BER (simulated) | BER (expected for ZF, 4x4 Rayleigh) |
|---|---|---|
| 0 | ||
| 5 | ||
| 10 | ||
| 15 | ||
| 20 |
The simulated values are close to expected but consistently slightly lower (better). What could cause this systematic bias? List at least two possible explanations and describe how to diagnose each one.
Think about what could give a small but systematic improvement: array gain, normalization, or channel model.
Check the noise variance normalization carefully.
Possible causes of systematic bias
1. Channel normalization error: If instead of 1 (i.e., each entry has unit variance but the column norm is not normalized), the effective SNR is times higher than intended. For a 4x4 system this is a 6 dB shift β but if partially corrected, it might manifest as a small consistent bias.
Diagnosis: Print from the simulation and verify it matches the intended normalization.
2. SNR definition mismatch: If the code defines but the -axis plots , the conversion might be slightly off (e.g., forgetting the code rate for uncoded transmission where , or mishandling MIMO degrees of freedom).
Diagnosis: Verify the BER at one SNR point analytically. For BPSK in AWGN (SISO), the BER should be exactly .
3. Rounding in noise generation:
Using randn with incorrect variance scaling (e.g.,
per real dimension but applying it as )
would reduce noise power and improve BER.
Diagnosis: Generate noise samples, compute their empirical variance, and compare to .
ex-ch34-06
MediumDesign a complete simulation parameter table for a paper that studies downlink massive MIMO with ZF precoding. The system has BS antennas, single-antenna users, and operates at 3.5 GHz. Your table should include every parameter from the checklist in Section 34.2 (Remark 34.2.11).
Additionally, specify: (a) what BER/rate curves you would plot, (b) what baselines you would include, and (c) how many Monte Carlo realizations you would use and why.
Use the parameter checklist from Section 34.2 as a template.
For massive MIMO, channel hardening reduces variance, so fewer realizations may suffice for rate β but not for BER at low error rates.
Complete parameter table
| Parameter | Value |
|---|---|
| BS antennas | |
| Users | , single-antenna |
| Channel model | i.i.d. Rayleigh, |
| Path loss | 3GPP UMi, |
| User distribution | Uniform in cell, m |
| Channel estimation | MMSE from uplink pilots, |
| SNR definition | (total TX power to noise) |
| SNR range | to dB |
| Modulation | QPSK (for BER), Gaussian signaling (for rate) |
| Coding | Uncoded (for BER), capacity formula (for rate) |
| Frame length | symbols (coherence block) |
| Carrier frequency | GHz |
| Bandwidth | MHz |
| MC realizations (rate) | |
| MC realizations (BER) | Until 100 errors per SNR point |
| Random seed | 2024 |
Plots, baselines, and MC justification
(a) Plots:
- Ergodic sum rate vs. SNR (all baselines)
- Per-user rate CDF at SNR = 10 dB
- BER vs. for QPSK
(b) Baselines:
- MRT precoding (lower complexity baseline)
- MMSE precoding (stronger baseline)
- Single-user bound:
- Perfect CSI genie bound (for estimated-CSI case)
(c) MC justification:
- Rate: 1,000 realizations. With , channel hardening ensures low variance. Doubling to 2,000 changes the mean by less than 0.5%.
- BER: Adaptive, minimum 100 errors. At BER = with QPSK ( bits/frame), need frames.
ex-ch34-07
MediumA colleague shows you the following Python snippet for a BPSK BER simulation and asks you to find the bug:
import numpy as np
def simulate_bpsk_ber(snr_db, n_bits=100000):
snr_lin = 10**(snr_db / 10)
bits = np.random.randint(0, 2, n_bits)
symbols = 2 * bits - 1 # BPSK: {0,1} -> {-1,+1}
noise = np.random.randn(n_bits) + 1j * np.random.randn(n_bits)
noise_power = 1 / snr_lin
received = symbols + np.sqrt(noise_power) * noise
detected = (np.real(received) > 0).astype(int)
ber = np.mean(bits != detected)
return ber
- Identify the bug.
- Explain the quantitative impact on the BER curve.
- Write the corrected code.
Check the noise power: what is the variance of noise before scaling?
For , each real/imaginary part has variance .
Identify the bug
The noise is generated as
np.random.randn(n_bits) + 1j * np.random.randn(n_bits),
which has variance 2 (each component has variance 1).
After scaling by np.sqrt(noise_power), the total noise
power is instead of
.
The effective SNR is halved (3 dB lower than intended).
Quantitative impact
The BER curve is shifted right by 3 dB. For example, at a plotted dB, the actual simulation runs at 7 dB, producing a higher BER than expected.
Corrected code
Two valid fixes:
Fix 1: Scale noise correctly for complex baseband:
noise = (np.random.randn(n_bits)
+ 1j * np.random.randn(n_bits)) / np.sqrt(2)
Fix 2: Since BPSK is real-valued, use only real noise:
noise = np.random.randn(n_bits)
received = symbols + np.sqrt(noise_power) * noise
Fix 2 is preferred for BPSK because the imaginary noise component is irrelevant after taking the real part for detection.
ex-ch34-08
Hard(Simulation design and validation.) Design and implement (in MATLAB or Python pseudocode) a complete BER simulation for QPSK transmission over a Rayleigh fading channel with MMSE detection. Your implementation must:
- Use the Monte Carlo template from Algorithm 34.2.2 with a minimum of 100 errors per SNR point.
- Correctly normalize noise for .
- Report 95% confidence intervals at each SNR point.
- Validate against the known analytical BER for MMSE detection (or at least verify that the diversity order is 1 for a system with 2 streams).
Provide the pseudocode and explain each normalization step.
For QPSK, .
The MMSE filter for stream is the -th column of .
At high SNR, the BER slope on a log-log plot reveals the diversity order.
Signal model and normalization
Signal model: where , , (QPSK, unit energy), .
Per-stream SNR: (since ). . So .
MMSE detection
MMSE filter: . Estimated symbols: . Slice to nearest QPSK constellation point. Count bit errors using Gray-coded bit mapping.
Confidence intervals
At each SNR point, after accumulating errors from bits: , CI: .
Validation
The diversity order of a system with 2 spatial streams and MMSE detection is . At high SNR, the BER should decay as , i.e., a slope of on a log-log plot (or dB/decade on a log-linear BER vs. dB plot). Verify this by fitting a line to the high-SNR portion of the simulated curve.
ex-ch34-09
Hard(Critical paper analysis.) Consider a hypothetical paper abstract:
"We propose a deep-learning-based channel estimator for massive MIMO that achieves 5 dB gain over MMSE estimation. The network is trained on channels generated from the i.i.d. Rayleigh model and tested on the same distribution. We consider antennas, users, and perfect knowledge of the noise variance."
Perform a critical analysis:
- Is the 5 dB gain claim plausible? Why or why not?
- Identify at least three methodological concerns.
- What additional experiments would you request as a reviewer?
MMSE estimation is optimal for Gaussian channels when the channel statistics are known. What does this imply?
Think about the training/testing distribution and generalization.
Plausibility assessment
The claim is highly suspicious. MMSE estimation is the optimal linear estimator for Gaussian channels, and for i.i.d. Rayleigh fading (which is Gaussian), MMSE is actually the optimal estimator (not just optimal among linear ones). A 5 dB gain over the optimal estimator is mathematically impossible unless:
- The MMSE baseline is implemented incorrectly.
- The comparison metric is not MSE (e.g., BER with mismatched detection).
- The "MMSE" baseline does not use the true channel statistics.
Methodological concerns
-
Training = testing distribution: The DL estimator is trained and tested on i.i.d. Rayleigh channels. This is the easiest possible setting and tells us nothing about real-world performance. The estimator may have simply memorized the distribution.
-
MMSE baseline likely wrong: For i.i.d. Rayleigh with known noise variance, MMSE estimation has a closed form. If the implementation uses LS estimation as the "MMSE" baseline, the 5 dB gap is explained by the LS-to-MMSE gap, not by the neural network.
-
No complexity comparison: Deep learning estimators require significant computation (matrix multiplications through many layers). Without a complexity comparison, the practical value is unclear.
-
No generalization test: What happens when the channel model changes (e.g., correlated fading, Rician)?
-
Perfect noise variance knowledge: In practice, noise variance must also be estimated, which affects MMSE performance.
Requested experiments
As a reviewer, request:
- MSE learning curves showing training vs. validation loss
- Comparison with correctly implemented MMSE using true channel covariance
- Testing on a different channel model (generalization)
- Complexity table (flops per estimation, inference time)
- Performance with estimated (not perfect) noise variance
ex-ch34-10
Hard(Importance sampling for rare events.) Estimating BER by brute-force Monte Carlo requires approximately bit decisions. Importance sampling (IS) reduces this dramatically.
Consider BPSK over AWGN. The optimal IS distribution shifts the noise mean by (biasing toward errors).
- Derive the IS weight (likelihood ratio) for this shifted distribution.
- Show that the IS estimator is unbiased.
- Estimate how many samples are needed to achieve 10% relative error at BER using IS, and compare with brute force.
The IS weight is where is the original noise PDF and is the shifted PDF.
For Gaussian distributions, the likelihood ratio has a simple exponential form.
IS weight derivation
Original noise: , PDF . Shifted noise: , PDF .
IS weight: .
With and : .
Unbiasedness
.
The IS estimator is unbiased.
Sample count comparison
With the optimal IS shift, nearly every sample produces an error (the noise is biased to the wrong decision region). The variance of the IS estimator is approximately , so for 10% relative error: .
Brute force: .
Speedup: . This is why importance sampling is essential for simulating coded systems at operational BER levels.
ex-ch34-11
Hard(Reproducibility audit.) You receive a paper for review. The numerical results section states: "We simulate a 16-antenna BS serving 4 users over Rayleigh fading channels. Results are averaged over sufficient channel realizations."
The paper provides three figures: sum rate vs. SNR, per-user rate CDF, and convergence of the proposed algorithm.
- List every piece of missing information that prevents you from reproducing these results.
- Write a concrete reviewer comment requesting the necessary details (aim for 150--200 words).
- Propose a "minimum reproducibility standard" that IEEE could adopt for wireless papers (5--7 bullet points).
Use the simulation parameter table from Section 34.2 as your checklist.
Consider what "sufficient channel realizations" actually means quantitatively.
Missing information
- SNR definition (per-antenna? total? ?)
- SNR range and step size
- Channel model details (i.i.d.? correlated? path loss?)
- CSI assumption (perfect? estimated? what estimator?)
- Number of Monte Carlo realizations ("sufficient" is vague)
- Noise variance normalization
- Modulation and coding scheme
- Frame/block length
- Power constraint type (total? per-antenna?)
- Algorithm initialization and stopping criterion
- Baseline implementation details
- Computational platform and software version
- Random seed or number of independent runs
Reviewer comment
"The simulation setup lacks critical details for reproducibility. Specifically: (1) The SNR definition is not stated β is this total transmit power to noise ratio, per-antenna SNR, or ? (2) 'Sufficient channel realizations' is not quantitative; please state the exact number and provide 95% confidence intervals on all curves. (3) The channel model says 'Rayleigh fading' but does not specify correlation, path loss, or whether the channel is frequency-flat or frequency-selective. (4) The CSI assumption is not stated. (5) The baselines are not described in enough detail to re-implement. I recommend adding a simulation parameter table (see IEEE JSAC guidelines) and consider releasing the simulation code."
Minimum reproducibility standard
- All papers must include a simulation parameter table listing at minimum: antenna configuration, channel model, CSI assumption, SNR definition, modulation/coding, and MC trial count.
- SNR must be defined by an explicit equation.
- Number of Monte Carlo realizations must be stated numerically.
- 95% confidence intervals must be shown on all stochastic curves.
- Baseline implementations must be described or cited with specific equation/algorithm numbers.
- Computational complexity must be compared in FLOPs or runtime.
- Authors are encouraged to provide code via a public repository, with a "Reproducibility" section in the paper.
ex-ch34-12
Hard(End-to-end simulation project.) Design and conduct (or describe in complete pseudocode) a simulation study comparing MRT, ZF, and MMSE precoding for a massive MIMO downlink with antennas and users. Your study must:
- Use i.i.d. Rayleigh fading with proper normalization.
- Define SNR as total transmit power to noise ratio, .
- Compute ergodic sum rate with SINR expressions for each precoder.
- Average over channel realizations and report 95% CIs.
- Include a complexity comparison table.
- Discuss when the three precoders converge (massive MIMO regime) and when they diverge.
Present your results as you would in a paper: parameter table, one figure description, and a one-paragraph discussion.
SINR for ZF: , where with equal power.
SINR for MRT: for large .
The three precoders converge when because inter-user interference vanishes.
Parameter table
| Parameter | Value |
|---|---|
| BS antennas | |
| Users | 8, single-antenna |
| Channel model | i.i.d. Rayleigh, |
| CSI | Perfect |
| SNR definition | |
| SNR range | to dB, 2 dB step |
| Power allocation | Equal: |
| MC realizations | 1,000 |
| Metric | Ergodic sum rate (bits/s/Hz) |
SINR expressions and complexity
MRT: , normalized. . Complexity: .
ZF: , normalized. . Complexity: .
MMSE: , . Complexity: .
| Precoder | Flops per coherence block |
|---|---|
| MRT | |
| ZF | |
| MMSE |
Discussion
At (only overprovisioning), ZF and MMSE significantly outperform MRT at high SNR because MRT's inter-user interference floor limits the sum rate. MMSE outperforms ZF at low SNR by trading off interference suppression for noise enhancement. At ( overprovisioning), all three precoders converge because the channel vectors become nearly orthogonal (channel hardening and favorable propagation). The 95% CIs are narrow (< 1% relative) due to channel hardening at , but wider at . This confirms the massive MIMO benefit: simple MRT is near-optimal when the antenna-to-user ratio is large, avoiding the matrix inversion of ZF/MMSE.
ex-ch34-13
MediumA paper reports BER results for a 4Γ4 MIMO system using 16-QAM. The x-axis is labeled "SNR (dB)" and shows a range of 0--25 dB. The paper defines where is the total transmit power across 4 antennas.
You want to compare these results against BPSK SISO curves from Proakis which use .
(a) Express the paper's SNR in terms of for 16-QAM with code rate and 4 transmit antennas. (b) At the paper's dB, what is the equivalent in dB?
per antenna, and .
There are bits per symbol and code rate , so 2 information bits per symbol.
SNR to $E_b/\ntn{n0}$ conversion
Per-antenna symbol energy: .
With : .
.
With , , : .
Numerical evaluation
At dB ( linear): ( dB).
The paper's 20 dB SNR is equivalent to only 11 dB .
ex-ch34-14
MediumYou receive a reviewer comment: "The authors should compare against MMSE precoding, not just ZF, as MMSE is the stronger baseline at low-to-moderate SNR."
(a) Explain why MMSE is a stronger baseline than ZF. (b) Give the MMSE (regularized ZF) precoding formula and explain the role of the regularization parameter. (c) Under what conditions does ZF become equivalent to MMSE?
MMSE balances noise enhancement against interference suppression.
The regularization parameter depends on and .
Why MMSE is stronger
ZF completely nulls inter-user interference but at the cost of noise enhancement (amplifying noise to satisfy the zero-forcing constraint). MMSE minimizes the total MSE by balancing interference suppression against noise amplification. At low SNR, the noise dominates, and ZF's noise enhancement is costly. MMSE adapts: at low SNR it behaves like MRT (ignores interference, maximizes received power), at high SNR it behaves like ZF.
MMSE precoding formula
\alpha = K\sigma^2/P\alpha = 0\alpha \to \infty$: reduces to MRT (up to scaling).
ZF-MMSE equivalence
ZF and MMSE become equivalent as (i.e., ), or when (massive MIMO regime where by channel hardening).
ex-ch34-15
EasyIdentify the type of each SNR definition in the following system model descriptions:
(a) "We define SNR where is the total transmit power." (b) "The received SNR is ." (c) "BER is plotted against ."
Per-antenna, total receive, and energy-per-bit are the three main types.
Classification
(a) Per-antenna transmit SNR β the power is divided by to give the SNR per antenna element. No channel gain is included.
(b) Instantaneous total receive SNR β includes the channel gain . This is a random variable that depends on the fading realization.
(c) Energy per bit to noise density β normalized by both the modulation order and code rate. The standard for comparing different modulation/coding combinations on a common basis.