Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

🎥 See More
Loading...

Let ZZ be standard normal. Then the mgf of ZZ is given by

MZ(t) = et2/2   for all tM_Z(t) ~ = ~ e^{t^2/2} ~~~ \text{for all } t

To see this, just work out the integral:

MZ(t) = etz12πe12z2dz= 12πe12(z22tz)dz= et2/212πe12(z22tz+t2)dz= et2/212πe12(zt)2dz= et2/2\begin{align*} M_Z(t) ~ &= ~ \int_{-\infty}^\infty e^{tz} \frac{1}{\sqrt{2\pi}} e^{-\frac{1}{2}z^2} dz \\ \\ &= ~ \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}} e^{-\frac{1}{2}(z^2 - 2tz)} dz \\ \\ &= ~ e^{t^2/2} \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}} e^{-\frac{1}{2}(z^2 - 2tz + t^2)} dz \\ \\ &= ~ e^{t^2/2} \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}} e^{-\frac{1}{2}(z- t)^2} dz \\ \\ &= ~ e^{t^2/2} \end{align*}

because the integral is 1. It is the normal (t,1)(t, 1) density integrated over the whole real line.

19.3.1Linear Transformation

It’s handy to note that moment generating functions behave well under linear transformation.

MaX+b(t) = E(et(aX+b)) = ebtE(eatX) = ebtMX(at)M_{aX+b}(t) ~ = ~ E(e^{t(aX + b)}) ~ = ~ e^{bt}E(e^{atX}) ~ = ~ e^{bt}M_X(at)

19.3.2Normal (μ,σ2)(\mu, \sigma^2)

Since a normal (μ,σ2)(\mu, \sigma^2) variable can be written as σZ+μ\sigma Z + \mu where ZZ is standard normal, its m.g.f. is

MσZ+μ(t) = eμtMZ(σt) = eμt+σ2t2/2M_{\sigma Z + \mu} (t) ~ = ~ e^{\mu t}M_Z(\sigma t) ~ = ~ e^{\mu t +\sigma^2 t^2/2}

Details aside, what this formula is saying is that if a moment generating function is exp(c1t+c2t2)\exp(c_1t + c_2t^2) for any constant c1c_1 and any positive constant c2c_2, then it is the moment generating function of a normally distributed random variable.

19.3.3Sums of Independent Normal Variables

We can now show that sums of independent normal variables are normal.

Let XX have normal (μX,σX2)(\mu_X, \sigma_X^2) distribution, and let YY independent of XX have normal (μY,σY2)(\mu_Y, \sigma_Y^2) distribution. Then

MX+Y(t)=eμXt+σX2t2/2eμYt+σY2t2/2=e(μX+μY)t+(σX2+σY2)t2/2M_{X+Y} (t) = e^{\mu_X t + \sigma_X^2 t^2/2} \cdot e^{\mu_Y t + \sigma_Y^2 t^2/2} = e^{(\mu_X + \mu_Y)t + (\sigma_X^2 + \sigma_Y^2)t^2/2}

That’s the m.g.f. of the normal distribution with mean μX+μY\mu_X + \mu_Y and variance σX2+σY2\sigma_X^2 + \sigma_Y^2.

19.3.4“Proof” of the Central Limit Theorem

Another important reason for studying mgf’s is that they can help us identify the limit of a sequence of distributions.

The main example of convergence that we have seen is the Central Limit Theorem. Now we can indicate a proof.

Let X1,X2,X_1, X_2, \ldots be i.i.d. random variables with expectation μ\mu and SD σ\sigma. For every n1n \ge 1 let Sn=X1+X2++XnS_n = X_1 + X_2 + \cdots + X_n.

The Central Limit Theorem says that for large nn, the distribution of the standardized sum

Sn=SnnμnσS_n^* = \frac{S_n - n\mu}{\sqrt{n}\sigma}

is approximately standard normal.

To show this, we will assume a major result whose proof is well beyond the scope of this class. Suppose Y1,Y2,Y_1, Y_2, \ldots are random variables and we want to show that the the distribution of the YnY_n’s converges to the distribution of some random variable YY. The result says that it is enough to show that the mgf’s of the YnY_n’s converge to the mgf of YY.

The result requires a careful statement and the proof requires considerable attention to detail. We won’t go into that in this course. Instead we’ll just point out that it should seem reasonable. Since mgf’s determine distributions, it’s not difficult to accept that if two mgf’s are close to each other then the corresponding distributions should also be close to each other.

Let’s use this result to “prove” the CLT. The quotes are because we will use the above result without proof, and also because the argument below involves some hand-waving about approximations.

First, write the standardized sum in terms of the standardized XX’s.

Sn=Snnμnσ=i=1n1n(Xiμσ)=i=1n1nXi\begin{align*} S_n^* = \frac{S_n - n\mu}{\sqrt{n}\sigma} = \sum_{i=1}^n \frac{1}{\sqrt{n}} \big( \frac{X_i - \mu}{\sigma} \big) = \sum_{i=1}^n \frac{1}{\sqrt{n}} X_i^* \end{align*}

where for each ii, the random variable XiX_i^* is XiX_i in standard units.

The random variables XiX_i^* are i.i.d., so let MXM_{X^*} denote the mgf of any one of them. By the linear transformation property proved above, the mgf of each 1nXi\frac{1}{\sqrt{n}}X_i^* is given by

M1nXi(t)=MX(t/n)M_{\frac{1}{\sqrt{n}}X_i^*} (t) = M_{X^*} (t/\sqrt{n})

Therefore

MSn(t)=(MX(t/n))n=(1+tnE(X)1!+t2nE(X2)2!+t3n3/2E(X3)3!+)n (1+t22n)n   for large n\begin{align*} M_{S_n^*} (t) &= \big( M_{X^*}(t/\sqrt{n}) \big)^n \\ \\ &= \Big( 1 + \frac{t}{\sqrt{n}} \cdot \frac{E(X^*)}{1!} + \frac{t^2}{n} \cdot \frac{E({X^*}^2)}{2!} + \frac{t^3}{n^{3/2}} \cdot \frac{E({X^*}^3)}{3!} + \cdots \Big)^n \\ \\ &\approx ~ \Big( 1 + \frac{t^2}{2n}\Big)^n ~~~ \text{for large } n\\ \\ \end{align*}

by ignoring small terms and using the fact that for any standardized random variable XX^* we have E(X)=0E(X^*) = 0 and E(X2)=1E({X^*}^2) = 1.

Thus for large nn,

MSn(t)(1+t22n)net2/2as nM_{S_n^*} (t) \approx \Big( 1 + \frac{t^2}{2n}\Big)^n \to e^{t^2/2} \text{as } n \to \infty

The limit is the moment generating function of the standard normal distribution.