Resources for students in Stat 111 (Spring 2023). Managed by aL Xin.
View files on GitHub awqx/stat111-2023
SectionsEx: Let $Y_1, Y_2, \dots$ be i.i.d. with mean $\mu$ and variance $\sigma^2$. MoM for $\mu$ is $\bar Y = \hat\theta$ (this is unbiased). The variance of $\bar Y$ is $\sigma^2 / n$.
Reminder: Expression for CLT. $d$ or $D$ indicates convergence in distribution.
\[\sqrt{n} (\bar Y - \mu) \overset{d}{\sim} \mathcal{N}(0, \sigma^2)\]Equivalently,
\[\frac{\bar Y - \mu}{\sigma / \sqrt{n}} \overset{d}{\sim} \mathcal{N}(0, 1),\]which may be preferable to use the PDF $\phi$ or CDF $\Phi$ of the standard Normal.
We can also use the statement
\[\bar Y \; \dot\sim \; \mathcal{N}\left(\mu, \frac{\sigma^2}{n}\right),\]which is NOT the same as the above. This approximate distribution is NOT a limit statement; the $n$ appears on the RHS, which would not work at all with a limit.
Thm: Let $Y_1, Y_2, \dots$ be i.i.d. continuous r.v.s with PDF $f$, CDF $F$, and quantile function $Q = F^{-1}$. We have
\[\hat Q_n(p) = Y_{(\lceil{np} \rceil)}, \quad \sqrt{n} (\hat Q_n(p) - Q(p)) \overset{d}{\sim} \mathcal{N}\left(0, \frac{p(1 - p)}{f(p)^2} \right)\]Ex: $Y_1, Y_2, \dots$ is i.i.d. $\mathcal{N}(\theta, \sigma^2)$ and our estimand is $\theta$. Let $M_n$ be the sample median (an order statistic). Which is better?
From the above theorem and CLT, we have
\[M_n \; \dot\sim \; \mathcal{N} \left( \theta, \frac{\pi\sigma^2}{2n} \right), \quad \bar Y \sim \mathcal{N} (\theta, \sigma^2 /n).\]We can see in the approximate distributions that $M_n$ has higher variance (but the same bias). So by MSE $M_n$ would be worse.
Def: $X_n$ converges to $X$ in probability if for any $\epsilon > 0$, we have $P(\mid X_n - X \mid \geq \epsilon) \to 0$ as $n \to \infty$.
Thm (Continuous mapping theorem): If $g$ is continuous and $X_n \to X$, then $g(X_n) \to g(X)$ in the same form of convergence.
Thm (Slutsky’s theorem): Suppose $X_n \overset{d}{\to} X, Y_n \overset{d}{\to} Y$.
Does $X_n + Y_n \overset{d}{\to} X + Y$? No, not generally. For example, let $X_n, Y_n$ be i.i.d. $\mathcal{N}(0, 1)$ where $X = Y \sim \mathcal{N}(0, 1)$. The LHS converges to $\mathcal{N}(0, 2)$ and the RHS is $\mathcal{N}(0, 4)$.
Now, suppose that $Y_n \overset{d}{\to} c$ where $c$ is a constant. We have
\[\begin{aligned} X_n + Y_n &\overset{d}{\to} X + c \\ X_n - Y_n &\overset{d}{\to} X - c \\ X_n Y_n &\overset{d}{\to} cX \\ X_n / Y_n &\overset{d}{\to} X / c: c \neq 0, Y_n \neq 0 \end{aligned}\]