Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

If the elements of X\mathbf{X} are mutually independent then Cov(Xi,Xj)=0Cov(X_i, X_j) = 0 for all iji \ne j and hence the covariance matrix Σ\boldsymbol{\Sigma} is a diagonal matrix with the iith diagonal element equal to Var(Xi)Var(X_i).

In the other direction, zero covariance doesn’t in general imply independence, and pairwise independence doesn’t imply mutual independence. But once again, the multivariate normal turns out to have a wonderful property:

If X\mathbf{X} has a multivariate normal density and its elements are pairwise uncorrelated – that is, Cov(Xi,Xj)=0Cov(X_i, X_j) = 0 for all iji \ne j – then the elements of X\mathbf{X} are mutually independent.

That is, multivariate normal random variables are independent if and only if they are pairwise uncorrelated.

🎥 See More
Loading...

This is easy to see from the form of the density of X\mathbf{X}. If Σ\boldsymbol{\Sigma} is a diagonal matrix then so is Σ1\boldsymbol{\Sigma}^{-1}. The iith diagonal element of Σ1\boldsymbol{\Sigma}^{-1} is 1/σi21/\sigma_i^2 where σi2=Var(Xi)\sigma_i^2 = Var(X_i). So

(xμ)TΣ1(xμ) = i=1n(xiμ(i))2σi2(\mathbf{x} - \boldsymbol{\mu})^T\boldsymbol{\Sigma}^{-1} (\mathbf{x} - \boldsymbol{\mu}) ~ = ~ \sum_{i=1}^n \frac{(x_i - \boldsymbol{\mu}(i))^2}{\sigma_i^2}

and therefore

exp(12(xμ)TΣ1(xμ)) = i=1nexp(12(xiμ(i)σi)2)\exp\big( -\frac{1}{2} (\mathbf{x} - \boldsymbol{\mu})^T\boldsymbol{\Sigma}^{-1} (\mathbf{x} - \boldsymbol{\mu}) \big) ~ = ~ \prod_{i=1}^n \exp\big(-\frac{1}{2} \big(\frac{x_i - \boldsymbol{\mu}(i)}{\sigma_i}\big)^2\big)

In the constant of integration, det(Σ)=σ12σ22σn2\det(\boldsymbol{\Sigma}) = \sigma_1^2 \sigma_2^2 \cdots \sigma_n^2.

Therefore the density of X\mathbf{X} is the product of the marginal normal densities.

Thus for a multivariate normal random vector, “pairwise uncorrelated” is equivalent to “mutually independent” and is a much easier condition to check.

The result makes it easy to see if two coordinates of a multivariate normal vector are independent. All you have to do is find the covariance between the two. If the covariance is 0, they are independent.

23.4.1Sum and Difference, Revisited

Let X=[X1,X2]T\mathbf{X} = [X_1, X_2]^T have a bivariate normal distribution. Let S=X1+X2S = X_1 + X_2 and D=X1X2D = X_1 - X_2. We know that SS and DD have a bivariate normal distribution and that

Cov(S,D) = Var(X1)Var(X2)Cov(S, D) ~ = ~ Var(X_1) - Var(X_2)

If X1X_1 and X2X_2 have the same variance then SS and DD are uncorrelated, and hence also independent by what we have just proved.

Thus for example the sum and difference of two i.i.d. normal random variables are independent.

You have shown in exercises that the sum and differences of any two i.i.d. random variables are uncorrelated. If in addition the two variables are normal, then their sum and difference are independent, not just uncorrelated.