Rescaling Errors

We investigate the appearance of the scaling factor √n / σ in the central limit theorem.

29 March 2025

Recall the strong law of large numbers.

{}

Theorem [SLLN]. Let X1,...,XnX_1,...,X_n be a sequence of i.i.d. random variables with finite mean μ\mu and variance σ2.\sigma^2. Then, the sequence of sample means Xˉn:=1ni=1mXi\bar{X}_n:=\frac{1}{n}\sum_{i=1}^m X_i converges almost surely to μ\mu. In other words, almost surely

εn:=Xˉnμo(1).\begin{equation} \varepsilon_n:=\bar{X}_n-\mu \to o(1). \end{equation}

Under a signal-noise interpretation, SLLN tells the signal (mean) dominates asymptotically. This result is sharpened by the central limit theorem.

{}

Theorem [CLT]. Let X1,...,XnX_1,...,X_n be a sequence of i.i.d. random variables with finite mean μ\mu and variance σ2.\sigma^2. Then the sequence of properly normalized sample means Xˉn:=1ni=1mXi\bar{X}_n:=\frac{1}{n}\sum_{i=1}^mX_i convergences in distribution to the Gaussian

nσ(Xˉnμ)=nσεndN(0,1).\begin{equation} \frac{\sqrt{n}}{\sigma}(\bar{X}_n - \mu)=\frac{\sqrt{n}}{\sigma}\varepsilon_n\xrightarrow{d} \mathcal{N}(0, 1). \end{equation}

The signal-noise interpretation is that amplifying the signal by the noise (standard deviation) reveals the Gaussian. We try to elucidate the normalization of n\sqrt{n} in (2). We begin by asking for a quantitative, pre-asymptotic version of SLLN in order to quantify the decay on the error εn\varepsilon_n. This is the goal of Chebyshev's inequality. Since the sample variance of Xˉn\bar{X}_n is σ2n\frac{\sigma^2}{n}, we compute

P(εnt)σ2nt2.\begin{equation} \mathbb{P}(|\varepsilon_n|\geq t)\leq \frac{\sigma^2}{nt^2}. \end{equation}

Taking the change of variables ttσnt\mapsto \frac{t\sigma}{\sqrt{n}},

P(nσεnt)=P(εntσn)1t2,\begin{equation} \mathbb{P}(\frac{\sqrt{n}}{\sigma}|\varepsilon_n|\geq t)=\mathbb{P}\left(|\varepsilon_n|\geq \frac{t \sigma}{\sqrt{n}}\right)\leq \frac{1}{t^2}, \end{equation}

which pins down the scale in natural units on the tolerance as the standard deviation σn\frac{\sigma}{\sqrt{n}}, since the right hand side is no longer artifically dependent on the sample size nn. Notice the appearce of nσεn\frac{\sqrt{n}}{\sigma}\varepsilon_n in both the CLT and Chebyshev's inequality. Both cases point out that amplifying by the standard deviation stablizes the error εn\varepsilon_n. In particular, the decay rate on ϵn\epsilon_n is of order O(n1/2)O(n^{-1/2}).