Normal Distribution
A Random Variable is said to be normally distributed if it has
Because of the Central Limit Theorem, in practice, many random phenomena obey, at least approximately, a normal probability distribution.
Equivalent Definitions for Multivariate Normal Distribution
We can define a random vector to be (nondegenerate multivariate) normal if if has a PDF specified above.
- Or, if it has the form:
for any matrix and vector , where is a random vector whose components are independent standard normal random variables . ^7bb02c 2. Or, if for any real vector , the random variable is normal.
These two alternative definitions cover degenerate normal distribution, i.e., some components are constant/ the distribution is concentrated on a proper subspace of .
Equivalence
First Alternative
By change of variables, we can see that the definition via PDF implies the first alternative definition. We now show that if is not singular, is a nondegenerate normal r.v. with a PDF of the form above. By the relationship of Derived Distribution, we have
Plugging in the PDF of standard normal distribution, we get
and is the covariance of .
Second Alternative
The relationship from the first alternative definition to the second is direct. Now suppose is normal for any vector .
Wrong
We inspect the MGF of :
where we use the fact that is normal with mean and variance . On the other hand, let , where is a standard normal random vector. Then,
is multivariate normal andhas the same MGF as . By the inversion theorem of MGF, and have the same distribution.
- ❗️ The above proof is wrong, as we want to show but not only .
To rigorously prove the equivalence, we consider two cases.
Case I. . Thus, is invertible and we let . We want to show that is standard normal, and thus satisfies the first alternative definition. We first have and . We then inspect the MGF (transform) of :
which is the MGF of a standard normal random variable. Therefore, .
Case II. is singular. Then, there exists a vector such that . For simplicity, suppose . Note that
Thus, , indicating the linear dependence of . WLOG, suppose and the first components of are linearly independent and is a linear combination of . Since satisfies the second alternative definition, also satisfies it. By the Case I, we know for some positive definite and standard normal . Let be another -dimensional standard normal r.v. independent of . Then, we can write
Properties
-
(Sufficiency) The mean and covariance of a multivariate normal distribution consist of a Sufficient Statistic.
-
In other words, the distribution of a multivariate normal random vector is completely determined by its mean and covariance
-
📎 See Sufficiency for proof
-
-
(Affine transformation). The Affine Transformation of a normal random variable : is also a normal random variable
-
As a special case, any sub-vector of a normal random vector is also normal
- As a special case, any component of a normal random vector is also normal
-
If , then
-
📎 The proof follows the alternative definition 1 above
-
-
(Independent Gaussians are jointly Gaussian). The sum of independent normal random variables is also a normal random variable
-
📎 Prove this using the Inversion Theorem
-
❗️ Note that this is generally not true for dependent random variables
-
❗️ More generally, if a random vector with normal components is not jointly normal, then its affine transformation is not necessarily normal! See also Independence, Correlation, and Jointly Normal
-
-
(Independent iff Uncorrelated). For a multivariate normal random vector, its components are independent if and only if they are uncorrelated
-
📎 We can use the sufficiency property to prove this. For with uncorrelated components, we can construct with independent components that have the same mean and variance. Then and thus has independent components. See also Independence, Correlation, and Jointly Normal
-
📎 For nondegenerate normal random vector, we can also factorize the PDF to show independence.
-
❗️ The statement is not true for general Random Variables for which the mean and variance are not sufficient; see Independence, Correlation, and Jointly Normal
-
-
Hence, if , then is normal with mean 0 and variance 1; is said to have a standard or unit normal distribution
- We write the CDF of a standard normal distribution
-
(Symmetry).
-
Let ; then, (see Chi-Square Distribution)
Proofs
Sufficiency
Note that PDF or CDF completely determines the distribution of a random variable. Therefore, by the definition for nondegenerate multivariate normal distribution via PDF, we can see that the mean and covariance matrix are sufficient.
For general normal random vectors, we can seek help from the MGF. Note that a MGF also completely determines a distribution. We use the second alternative definition. For a random vector, its MGF is the multivariate transform:
where we use the definition that is normal with mean and variance . As we can see, the mean and variance are sufficient to determine the MGF, and thus the distribution.
Independence, Correlation, and Jointly Normal
Normal components does not imply jointly normal.
It is not true that if and are both normal, then the joint distribution of is normal. For example, let and , where , i.e., with equal probability. Then, it is easy to verify that is also the CDF of , and thus . However, if is jointly normal, we would have is normal, which is not true because .
Independent normal components implies jointly normal.
The above statement becomes true once we impose the independence condition. We use the second alternative definition above to prove this. Let with normal components. Then, for any vector , we have . Note that are independent normal random variables, and thus their sum is normal by Property ^prop-ind-joint, or the Inversion Theorem. By the alternative definition, is jointly normal.
Joint normal with zero correlation implies independence.
Suppose that the components of are uncorrelated, i.e., its covariance matrix is a diagonal. Consider another random vector such that and are independent. By Property ^prop-ind-joint, is jointly normal. Since and have the same mean and covariance, by Property ^prop-suff, and have the same distribution, and thus the components of are independent.
Zero correlation does not imply independence for general random variables.
For example, let and . Certainly, and are not independent, but they are uncorrelated:
Sample Mean and Sample Variance
Thm
If is a sample from a normal population having mean and variance , then and are independent random variables, with , and . Then we have .
- ❗️ This independence of and is a unique property of the normal distribution.
General Bivariate Normal Distribution
We skipped the introduction of standard bivariate normal random variable , as it is more convenient, and not conceptually harder, to directly deal with the general bivariate normal distribution.
Let and be two random vectors with a joint normal distribution. That is
Suppose (PSD). We have
- .
- is independent of , and thus independent of any function of .
- .
Before proving the above results, we recover the standard bivariate normal distribution from them. Let , , . Then, we have
Proofs
WLOG, we assume for simplicity. Let . We have
Note that is a linear transformation of , and thus the above uncorrelatedness implies independence. This also implies, for any function :
By the General Definition of Conditional Expectation, the above condition says that
Finally, we compute . Since is also a function of , it is independent of . Thus, we have