Exercises
ex-ch06-01
EasyLet be i.i.d. Bernoulli. Derive the MLE of and verify that it is unbiased.
Write the log-likelihood as where .
Set the derivative to zero.
Log-likelihood
with .
Solve score equation
gives .
Unbiasedness
.
ex-ch06-02
EasyLet be i.i.d. Poisson. Find the MLE of and compute its Fisher information.
Recall .
MLE
, giving .
Fisher information
, so . The asymptotic variance of the MLE is , matching exactly.
ex-ch06-03
MediumLet be i.i.d. with (mean and variance equal). Derive the score equation and find the MLE in closed form.
Both and equal , so there is a single parameter.
Use the Gaussian log-density and differentiate carefully in .
Log-likelihood
.
Score
. Simplify to (collect like terms carefully).
Quadratic in theta
After simplification, the score equation is (using the identity ). The positive root is with .
ex-ch06-04
EasyUse the invariance property to find the MLE of the standard deviation from i.i.d. observations, given the MLE of .
and the square root is one-to-one on .
Apply invariance
With and , .
ex-ch06-05
MediumFor i.i.d. exponential , compute the finite-sample bias of and show it is .
, so is inverse-Gamma distributed.
Distribution of $\bar Y$
, so and . For , (standard identity, ).
Bias computation
. Therefore bias . The bias-corrected estimator is .
ex-ch06-06
MediumShow that for i.i.d. the MLE has asymptotic variance , matching the CRLB for .
Compute the per-sample Fisher information for treating as known; then check the block-diagonal FIM.
Fisher information for $\sigma^2$
. Taking negative expectation at the truth gives .
Asymptotic variance
By asymptotic normality, , attaining the CRLB.
ex-ch06-07
MediumLet with and of full column rank. Write the MLE and its covariance, and identify it as a BLUE.
Apply Theorem thm-linear-gaussian-mle with .
MLE formula
(ordinary least squares).
Covariance
, which equals the CRLB and makes the estimator both MVUE and BLUE (Gauss-Markov).
ex-ch06-08
MediumFor i.i.d. Laplace observations , show that the MLE of the location is the sample median.
Minimize ; the derivative (where it exists) is .
Log-likelihood
. Maximizing is equivalent to minimizing .
Median minimizes L1
The function is convex and piecewise linear, minimized at the sample median (any median when is even). Thus .
ex-ch06-09
Hard(Pareto tail index) i.i.d. with density for , known. Find the MLE of and its asymptotic distribution.
Differentiate in .
.
Score equation
gives .
Asymptotic normality
. Hence .
ex-ch06-10
MediumImplement Fisher scoring for Poisson GLM: . Derive the score, the FIM, and the scoring update.
Let .
Score
.
FIM
.
Scoring update
, a weighted least squares step — the IRLS algorithm.
ex-ch06-11
MediumDerive the Cramer-Rao bound for the frequency of a single complex sinusoid , , , treating and as known.
Compute the score in and the Fisher information via .
Use for large .
Log-likelihood
.
Fisher information
.
CRLB
— the super-efficient scaling peculiar to frequency estimation.
ex-ch06-12
EasyShow that for the Gaussian mean model with known , the MLE is efficient (achieves CRLB) for every finite .
Compute and compare to .
CRLB
, so .
MLE variance
. The bound is attained exactly for all , not just asymptotically.
ex-ch06-13
Hard(Consistency via Jensen) For i.i.d. observations, show that using Jensen's inequality, and interpret the difference as the KL divergence.
Apply Jensen to .
Jensen
.
KL identification
Negating, , with equality iff a.s. By identifiability this forces , so is the unique population maximizer.
ex-ch06-14
Medium(Profile likelihood) In the Gaussian AR(1) model with i.i.d., , find the MLE of by profiling out .
Conditional on , the conditional likelihood of is Gaussian with mean .
Conditional log-likelihood
.
Profile $\sigma^2$
. Substituting makes the profiled log-likelihood .
MLE of $\rho$
Minimizing in yields the least-squares estimator .
ex-ch06-15
Hard(Two-sinusoid MLE is non-convex) For with , explain why the joint MLE in has multiple local maxima and recommend a practical algorithm.
Consider label symmetry and near-collision behaviour.
Identify non-convexity
The likelihood is invariant under permutation — at least two global maxima. Additionally, when the periodogram has two dominant peaks, the likelihood has a ridge along the line that creates a saddle.
Practical algorithm
(i) Compute the periodogram and identify the largest peaks as initializations. (ii) Run alternating Newton updates: fix , refine ; fix , refine . (iii) Alternatively apply ESPRIT/MUSIC to the sample covariance for a closed-form starting point. Without good initialization, pure Newton-Raphson will often return a wrong local maximum.
ex-ch06-16
MediumShow that the Kaczmarz/normal-equation step for the Gaussian linear model with is the orthogonal projection of onto the column space of .
The fitted values are with .
Compute the fitted values
.
Show projection properties
and , so is an orthogonal projector onto . The residual is orthogonal to every column of .
ex-ch06-17
Challenge(Regularity failure) For the shifted exponential , find the MLE of , its exact distribution, and its convergence rate. Is it asymptotically Gaussian?
The MLE is ; compute the distribution of the minimum.
MLE
The likelihood is zero unless , then , increasing in . Hence .
Distribution
, so , independent of .
Convergence rate
Rate is (super-efficient), limit is exponential — not Gaussian. CRLB machinery does not apply: the support boundary breaks regularity.
ex-ch06-18
MediumShow that if is an asymptotically efficient estimator, then so is for any fixed constant .
Compute and take the limit.
Reduce to original estimator
. The second term goes to zero, so by Slutsky the limit is the same Gaussian.
Interpret
Any shift of an efficient estimator is asymptotically equivalent — bias correction at the level does not change asymptotic variance.
ex-ch06-19
MediumUse the delta method with invariance to compute the asymptotic variance of in the exponential model, and verify it equals plus higher-order terms.
Delta method: if , then .
Variance of the rate MLE
(CRLB).
Delta method with $g(v)=1/v$
. Asymptotic variance of is , which is the CRLB for the mean parameter .
ex-ch06-20
Hard(Tone in colored noise) Let with , where is known. Derive the MLE of and interpret it as a generalized periodogram.
Whiten the observation: and apply Theorem thm-linear-gaussian-mle as a profile over .
Whiten
, . The whitened model has i.i.d. noise, so Theorem TGaussian Linear Model: Closed-Form MLE applies.
Profile $A$
For fixed , .
Generalized periodogram
The MLE of is , the peak of the noise-whitened ("generalized") periodogram.