Logo

Resources for students in Stat 111 (Spring 2023). Managed by aL Xin.

View files on GitHub awqx/stat111-2023

Sections
Lectures
Exam materials
R Bootcamp
Syllabus

Lecture 05 - Asymptotics

07 Feb 2023


Creating estimators

Empirical CDF

\[\hat F(y) = \frac{1}{n} \sum_{j = 1}^n \mathbb{1}(Y_j \leq y)\]

Asymptotics

LLN

Ex: Let $Y_1, Y_2, \dots$ be i.i.d. with mean $\mu$ and variance $\sigma^2$. MoM for $\mu$ is $\bar Y = \hat\theta$ (this is unbiased). The variance of $\bar Y$ is $\sigma^2 / n$.

CLT

Reminder: Expression for CLT. $d$ or $D$ indicates convergence in distribution.

\[\sqrt{n} (\bar Y - \mu) \overset{d}{\sim} \mathcal{N}(0, \sigma^2)\]

Equivalently,

\[\frac{\bar Y - \mu}{\sigma / \sqrt{n}} \overset{d}{\sim} \mathcal{N}(0, 1),\]

which may be preferable to use the PDF $\phi$ or CDF $\Phi$ of the standard Normal.

We can also use the statement

\[\bar Y \; \dot\sim \; \mathcal{N}\left(\mu, \frac{\sigma^2}{n}\right),\]

which is NOT the same as the above. This approximate distribution is NOT a limit statement; the $n$ appears on the RHS, which would not work at all with a limit.

Thm: Let $Y_1, Y_2, \dots$ be i.i.d. continuous r.v.s with PDF $f$, CDF $F$, and quantile function $Q = F^{-1}$. We have

\[\hat Q_n(p) = Y_{(\lceil{np} \rceil)}, \quad \sqrt{n} (\hat Q_n(p) - Q(p)) \overset{d}{\sim} \mathcal{N}\left(0, \frac{p(1 - p)}{f(p)^2} \right)\]

Ex: $Y_1, Y_2, \dots$ is i.i.d. $\mathcal{N}(\theta, \sigma^2)$ and our estimand is $\theta$. Let $M_n$ be the sample median (an order statistic). Which is better?

From the above theorem and CLT, we have

\[M_n \; \dot\sim \; \mathcal{N} \left( \theta, \frac{\pi\sigma^2}{2n} \right), \quad \bar Y \sim \mathcal{N} (\theta, \sigma^2 /n).\]

We can see in the approximate distributions that $M_n$ has higher variance (but the same bias). So by MSE $M_n$ would be worse.

Forms of convergence

Def: $X_n$ converges to $X$ in probability if for any $\epsilon > 0$, we have $P(\mid X_n - X \mid \geq \epsilon) \to 0$ as $n \to \infty$.

CMT

Thm (Continuous mapping theorem): If $g$ is continuous and $X_n \to X$, then $g(X_n) \to g(X)$ in the same form of convergence.

Slutsky’s theorem

Thm (Slutsky’s theorem): Suppose $X_n \overset{d}{\to} X, Y_n \overset{d}{\to} Y$.

Does $X_n + Y_n \overset{d}{\to} X + Y$? No, not generally. For example, let $X_n, Y_n$ be i.i.d. $\mathcal{N}(0, 1)$ where $X = Y \sim \mathcal{N}(0, 1)$. The LHS converges to $\mathcal{N}(0, 2)$ and the RHS is $\mathcal{N}(0, 4)$.

Now, suppose that $Y_n \overset{d}{\to} c$ where $c$ is a constant. We have

\[\begin{aligned} X_n + Y_n &\overset{d}{\to} X + c \\ X_n - Y_n &\overset{d}{\to} X - c \\ X_n Y_n &\overset{d}{\to} cX \\ X_n / Y_n &\overset{d}{\to} X / c: c \neq 0, Y_n \neq 0 \end{aligned}\]