Lecture 27
April 6, 2026
We’re 12 weeks in and haven’t said anything about how to quantify uncertainties in model estimation.
Let’s talk about that.
But for both of these, we need to know the sampling distribution: the underlying distribution of estimates across different data replications.
The sampling distribution of a statistic captures the uncertainty associated with random samples.

Theory around linear regression: if data is truly generated by some linear model (or something reasonable close), then
\[\frac{\hat{\beta} - \beta}{\hat{\text{se}}\left[\hat{\beta}\right]} \sim t_{n-2}\]
This means that even if the model is correctly specified, the standardized estimates of a coefficient \(\beta\) can be expected to follow a \(t\) distribution based on sampling variability.
Fisher Information: \[\mathcal{I}_x(\theta) = -\mathbb{E}\left[\frac{\partial^2}{\partial \theta_i \partial \theta_j} \log \mathcal{L}(\theta | x)\right]\]
Observed Fisher Information (uses observed data and calculated at the MLE): \(\mathcal{I}_\tilde{x}(\hat{\theta})\)
Asymptotic result: \[\sqrt{n}(\theta_\text{MLE} - \theta^*) \to N(0, \left(\mathcal{I}_x(\hat{\theta^*})\right)^{-1}\]
Sampling distribution based on observed data: \[\theta \sim N\left(\hat{\theta}_\text{MLE}, \left(n\mathcal{I}_\tilde{x}(\theta_\text{MLE})\right)^{-1}\right)\]
Wednesday: The Nonparametric Bootstrap
Friday: The Parametric Bootstrap
Homework 5: Due next Friday (4/17)
Project Updates: Due Friday (4/10)
Quiz 3: Friday (4/10) through pre-break material on Monte Carlo.