Sampling Distributions
Central Limit Theorem
Order Statistics
Let \(Z\sim N(0,1)\), \(W\sim \chi^2_\nu\), \(Z\perp W\); therefore:
\[ T=\frac{Z}{\sqrt{W/\nu}} \sim t_\nu \]
Let \(W_1\sim\chi^2_{\nu_1}\) \(W_2\sim\chi^2_{\nu_2}\), and \(W_1\perp W_2\); therefore:
\[ F = \frac{W_1/\nu_1}{W_2/\nu_2}\sim F_{\nu_1,\nu_2} \]
Sampling Distributions
Central Limit Theorem
Order Statistics
Let \(X_1, X_2, \ldots, X_n\) be identical and independent distributed random variables with \(E(X_i)=\mu\) and \(Var(X_i) = \sigma²\). We define
\[ Y_n = \sqrt n \left(\frac{\bar X-\mu}{\sigma}\right) \mathrm{ where }\ \bar X = x\frac{1}{n}\sum^n_{i=1}X_i. \]
Then, the distribution of the function \(Y_n\) converges to a standard normal distribution function as \(n\rightarrow \infty\).
\[ \bar X \sim N\left(\mu, \frac{\sigma^2}{n}\right) \]
Let \(X_1, \ldots, X_n \overset{iid}{\sim} \chi^2_p\), the MGF is \(M(t)=(1-2t)^{-p/2}\). Find the distribution of \(\bar X\) as \(n \rightarrow \infty\).
Sampling Distributions
Central Limit Theorem
Order Statistics
Order statistics are a fundamental concept in statistics and probability, dealing with the properties of sorted random variables. They provide insights into the distribution and behavior of sample data, such as minimum, maximum, and quantiles. Understanding order statistics is crucial in various fields such as risk management, quality control, and data analysis.
Let \(X_1, X_2, \ldots, X_n\) be a sample of \(n\) independent and identically distributed (i.i.d.) random variables with a common probability density function \(f(x)\). The order statistics are the sorted values of this sample, denoted as:
\[ X_{(1)} \leq X_{(2)} \leq \cdots \leq X_{(n)} \]
Here, \(X_{(1)}\) is the minimum, and \(X_{(n)}\) is the maximum of the sample.
The distribution of the \(k\)-th order statistic \(X_{(k)}\) can be derived using combinatorial arguments. Its PDF is given by: \[ f_{X_{(k)}}(x) = \frac{n!}{(k-1)!(n-k)!} [F(x)]^{k-1} [1 - F(x)]^{n-k} f(x) \]
This formula shows how the distribution of \(X_{(k)}\) depends on the underlying distribution of the sample and its position \(k\).