Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2013 Nov 22.
Published in final edited form as: J Multivar Anal. 2010 Nov;101(10):10.1016/j.jmva.2010.04.015. doi: 10.1016/j.jmva.2010.04.015

A characterization of multivariate normality through univariate projections

Yongzhao Shao a,b,, Ming Zhou b
PMCID: PMC3837532  NIHMSID: NIHMS213961  PMID: 24273352

Abstract

This paper introduces a new characterization of multivariate normality of a random vector based on univariate normality of linear combinations of its components.

Keywords: Goodness of fit, Linear combination of components, Marginal distribution, Multivariate normal distribution, Non-normality

1. Introduction

As is well known, the multivariate normal distribution is central to multivariate analysis. Therefore, characterizations and assessments of multivariate normality have attracted sustained interests from researchers as demonstrated in the monographs and papers by [1], [2], [3], [4] and others.

Commonly used assessments of multivariate normality or non-normality of a random vector include a variety of approaches based on linear combinations of variates. In particular, many types of univariate-based plots are both easy to make and simple to use for detecting skewness, outliers, and other departures from multivariate normality [3]. In addition, there exist many formal tests for multivariate normality of a random vector based on examination of selected linear combinations of its components [3, 5]. Indeed, as pointed out by Anderson [1, pp. 23], “One of the reasons that the study of normal multivariate distributions is so useful is that marginal distributions and conditional distributions derived from multivariate normal distributions are also normal distributions. Moreover, linear combinations of multivariate normal variates are again normally distributed.”

One the other hand, it is well known that a non-normal random vector may have some normally distributed linear combinations of its components [2, 5]. This does raise a serious question concerning the effectiveness of the common statistical practice for assessing multivariate normality by examining a few linear combinations of components. After all, only a few of the infinitely many linear combinations can be plotted or tested in practice. Therefore, it is of theoretical interest to characterize or measure the size of the set of normally-distributed linear combinations. Probabilistically, one might ask how large is the chance that a randomly selected linear combination of components from a non-normal random vector is normally distributed? Indeed, this problem has attracted attention of many researchers for a long time [6], [7], [8]. Remarkably, Hamedani and Tata [7] proved that a bivariate random variable is normally distributed if it has a infinite collection of distinct linear combinations of its components that are normally distributed. In particular, this result implies that, a non-normal bivariate random vector can only have finitely many normally distributed linear combinations of its components. However, this characterization of bivariate normality cannot be extended to the multivariate case in a straightforward way [8]. The main objective of this paper is to introduce a new characterization of multivariate normality through univariate projections that holds in all dimensions. We show that, for any multivariate random variable, the set of normally-distributed linear combinations of its components is negligible among all possible linear combinations. In particular, in any dimensions, the probability is zero that a randomly selected linear combination of components of a non-normal random vector is normally distributed. This finding includes the existing bivariate result of Hamedani and Tata [7] as a corollary (see Remark 2 in Section 2 for more details). Given the eminent role of normal distributions in multivariate statistical analysis [1], [2], [3], the finding of this paper bears certain significance to the assessment of multivariate normality, thus might be of interest to many researchers.

In the next section, we establish a new characterization of multivariate normality for a random vector by assessing normality of linear combinations of its components The linear combinations will also be called projections on the vector of coefficients. Section 3 contains concluding remarks.

2. Main Results

The first subsection introduces the basic notation, the multivariate normal distribution, the normal directions, and a few lemmas. The proofs of these lemmas are rather elementary, but are included for completeness. The new characterization of multivariate normality can be found in the second subsection.

2.1. Notation and Lemmas

Let ℝp (p ≥ 1) be the p-dimensional Euclidean space. The inner-product of two vectors x = (x1, …, xp)T and y = (y1, …, yp)T ∈ ℝp is denoted as xTy=i=1pxiyi. We use S = {u ∈ ℝpuTu = 1} to denote the unit sphere, m the Lebesgue measure in ℝp, i.e., m(A) denotes the Lebesgue measure of a measurable set A. Also, let Π denote the uniform measure on the unit sphere S, and ℕ the set of natural numbers.

We will say that a random vector X = (X1, …, Xp) has a multivariate normal distribution if the support of X is the entire space ℝp and there exist a p-vector μ and a symmetric, positive-definite p × p matrix Σ, such that the probability density function of X can be expressed as

fx(x)=1(2π)p/2||1/2exp(12(xμ)T1(xμ)),

where |Σ| is the determinant of Σ. The vector μ is the expected value and the matrix Σ is the covariance matrix of X. If a random vector has a p-variate normal distribution, by the above definition, it must have a density function and a non-singular covariance matrix. As is well known, given independent and identically distributed random observations, the mean vector μ and the covariance matrix Σ can be consistently estimated by their sample counterparts, i.e., the sample mean and sample covariance matrix Sn2, respectively. Moreover, given the existence of a p-variate Lebesgue density of X, the sample covariance matrix Sn2 is non-singular almost surely [9] [10]. Therefore, it is not essential to know the mean vector μ and the covariance matrix Σ. Indeed, without loss of generality, both the mean vector μ and the covariance matrix Σ are commonly assumed unknown in statistics and many other applications. Throughout this paper, we consider a given random vector X = (X1, …, Xp)T ∈ ℝp possessing a density function f(x) relative to the Lebesgue measure m. In particular, we call a vector u = (u1, …, up)T ∈ S a normal direction of X (or fX) if its one dimensional projection on u, uTX, has a univariate normal distribution.

Note that when uTX is normally distributed, its moment generating function exists. Then we can denote its mean and variance by μu and σu2, respectively. Therefore uTX is normally distributed if and only if E{exp(tuTX)}=exp(μut+t2σu2/2), σu2>0, for all t ∈ ℝ. Or equivalently, in terms of the density f of X,

pexp(tuTx)f(x)dx=exp(μut+t2σu2/2),σu2>0,for allt. (1)

Let G be the set of lines in ℝp that lie on normal directions of X and pass through the origin, that is,

G={upuTXis normally distributed}. (2)

We assume 0 ∈ G. Also, denote U as the set of normal directions of X, then U = G ∩ S. Since univariate normal distribution is completely determined by its moments [11, pp. 389], U can be written in terms of moment equations. Let ϕ be the density of the standard normal distribution, then

U={uS|p(uTx)nf(x)dxtn1σuϕ(tμuσu)dt=0,for alln}. (3)

With the above notation, it is well-known that X is normally distributed if and only if G = ℝp or U = S. Next we are going to show that X is normally distributed as long as G has positive Lebesgue measure. In the first lemma, we will show that G is a closed set thus Lebesgue measurable.

Lemma 1

The set G = {u ∈ ℝpuTX is normally distributed} is closed if X has a density inp.

Proof

It suffices to show that the set G contains all its limiting points. If a non-zero sequence {un}n≥1 ⊂ G converges to u00, then unTX converges to u0TX in distribution, where u0TX is non-degenerate because X has a Lebesgue density by assumption. Let αn=E(unTX), βn2=Var(unTX), then βn1(unTXαn) has a standard normal distribution. By the convergence of types theorem [11, pp. 193], there exist real numbers β > 0 and α such that limn→∞ αn = α, limn→∞ βn = β and u0TX has a normal distribution. Thus u0 ∈ G and G is a closed set in ℝp.

Before proving that X is normally distributed, it is necessary to show that all moments of X exist, which is true if G has positive Lebesgue measure, i.e. m(G) > 0, as asserted by the next lemma.

Lemma 2

For a random vector X with a Lebesgue density inp, all moments of X exist if the set G = {u ∈ ℝpuTX is normally distributed} has positive Lebesgue measure.

Proof

Let m be the Lebesgue measure in ℝp. Since m(G) > 0, there exists a basis of ℝp, {u1,…, up} ⊂ G. Otherwise there exists {ui1,…, uir} ⊂ G with r < p, such that any element in G is a linear combination of ui1,…, uir. Then G would be a subset of the linear vector space spanned by ui1,…, uir, who has Lebesgue measure 0 in ℝp. Consequently m(G) = 0, which is a contradiction. Now we can assume that {u1,…, up} can be chosen as a basis in ℝp, let Y = (Y1,…, Yp)T = (u1,…, up)TX, i = 1,…, n, then E|Yi|m < ∞ for all m ∈ ℕ because Yi=uiTX is normal. Moreover, X = {(u1,…, up)T}−1Y, that is, each Xi is a linear combination of normal random variables. Thus for each i, E|Xi|m < ∞ for all m ∈ ℕ, or equivalently, E{|X1|r1 ⋯ |Xp|rp} < ∞ for all r1,…, rp ∈ ℕ.

Remark 1

It is clear that m(G) = 0 if and only if Π(U) = 0, where m and Π are the Lebesgue measures in ℝp and on the unit sphere S, respectively.

When all moments of X exist, let W = (W1,…, Wp)T be a normal random vector having the same mean and covariance matrix as X and define the following moment equations

gn(u)=E{(uTX)n}E{(uTW)n},u=(u1,,up)Tp,n. (4)

Let Hn be the set of solutions to the above moment equations gn = 0, that is,

Hn={upgn(u)=0},n. (5)

Lemma 3

Using the notation in (2), (4), (5), if all moments of X exist, then G = ∩n≥1Hn. Moreover, for each n, either m(Hn) = 0 or Hn = ℝp.

Proof

G = ∩n≥1Hn follows from the fact that univariate normal distribution is determined by its moments. When all moments of X exist, gn(u) is a homogenous multivariate polynomial about u1,…, up with degrees at most n. If gn is the zero function, then Hn = ℝp. If gn is not the zero function, then for any fixed (u1,…, up−1)T, there are at most n values of up such that (u1,…, up) ∈ Hn by the fundamental theorem of algebra (i.e. a polynomial of degree n has at most n solutions). Thus m(Hn) = 0. Because, if we denote Hn(u1,…, up−1) = {up ∈ ℝ ∣ (u1,…, up)T ∈ Hn}, which is a finite set in this case. Let m1 be the Lebesgue measure in ℝ, then the Lebesgue measure m in ℝp is the product measure m1p=m1××m1. By Tonelli’s theorem [12, pp. 152], m(Hn)=p1m1{Hn(u1,,up1)}dm1p1=0.

2.2. A New Characterization of Multivariate Normality

If the set G = {u ∈ ℝpuTX is normally distributed} has positive Lebesgue measure, then, by Lemma 2, all moments of X exist, and then G = ℝp by Lemma 3. On the other hand, if G has zero measure, then clearly X can not be normally distributed. This yields the following theorem.

Theorem 1

A random vector X ∈ ℝp with a Lebesgue density f is not normally distributed if and only if the set of normal directions, U = {u ∈ S ∣ uTX is normally distributed}, has measure 0, i.e., Π(U) = 0.

One might think that a set with Lebesgue measure zero is not necessarily small. For example, the set of rational numbers has Lebesgue measure zero but is dense in Rp. However, G here is a nowhere dense set. In particular, in the bivariate case, if X is not normally distributed, U not only has measure zero, but also is a finite set, as claimed by the next corollary.

Corollary 1

If a bivariate random vector X (or its density) is not normal, then X has at most finitely many normal directions, i.e., U = {u ∈ S ∣ uTX is normally distributed} is a finite set.

Proof

Suppose U has two or more points, same arguments as in Lemma 2 yield that X has finite moments of all orders and U satisfies all the moment equations gn = 0 by Lemma 3. However, if gn is not the zero function, gn(u) is essentially a univariate polynomial (due to homogeneity of gn), which has finitely many solutions on the unit circle. Thus U is a finite set if X is not normal.

Remark 2

A result equivalent to the above Corollary for the bivariate case was established previously by Hamedani and Tata [7] and also claimed as part of the results in Ferguson [6]. While Ferguson [6] did not give a proof, Hamedani and Tata [7] proved the fact using characteristic functions. In particular, Theorem 3 of Hamedani and Tata [7] asserts that, given {(ak, bk), k = 1, 2, ⋯}, a countable distinct sequence in ℝ2, such that for each k, akX1 + bkX2 is a normal random variable, then X = (X1, X2)T is a bivariate normal random variable. To see that this fact directly follows from the above Corollary, it suffices to take uk = (u1k, u2k)T where u1k=ak/ak2+bk2 and u2k=bk/ak2+bk2. Then akX1 + bkX2 is a normal random variable if and only if uk = (u1k, u2k)T is a normal direction of X = (X1, X2)T. However, the above result as stated in Hamedani and Tata [7] for the bivariate case does not hold in three or higher dimensions as pointed out in Hamedani [8]. Thus, Theorem 1 of this paper, which holds for any dimension p ≥ 2, provides a non-straightforward generalization to the existing result for the bivariate case.

Suppose Y is another random vector with Lebesgue density. If X is not normally distributed, then m(G) = m({uRpuTX is normally distributed}) = 0 by Theorem 1. Thus P(Y ∈ G) = 0 or P{(YTY)12YU}=0, since the probability measure of Y is dominated by m. Therefore we obtain the following corollary.

Corollary 2

If a random vector X is not normally distributed, then for any other random vector Y ∈ ℝp with a Lebesgue density, the probability of (YTY)12Y taking values of normal directions of X is zero.

Remark 3

Formal tests for multivariate normality of a random vector might be constructed based on randomly selected linear combinations of its components. Suppose X, X1, …, Xn is an independent random sample from an unknown density f. Then we can consider univariate data XTXi, i = 1, …, n, which can be viewed as projections of X1, …, Xn on X. If f is not normal, then each XTXi, conditioned on X, are not normally distributed almost surely, thus can be tested using a consistent univariate test for normality such as the [3, 5] test. By Corollary 2, such a univariate-based test would have power against any non-normal alternative density. Thus one may construct univariate tests for multivariate normality based on a randomly selected direction. Tests based such univariate projections might be found in [3] and others.

3. Concluding Remarks

This paper establishes that a multivariate density is not normal if and only if its set of normal directions has Lebesgue measure zero. Consequently, the normal directions of a non-normal density are indeed quite rare. Note that this characterization of a non-normal multivariate density holds in any fixed dimension. Moreover, this new characterization is not an asymptotic result thus its validity does not depend on typical assumptions such as a large sample size. The main finding of this paper may have some significance for the assessment of multivariate normality which is of great relevance in multivariate analysis.

Acknowledgments

The authors thank the editor and the reviewers for their valuable comments and suggestions. This research is partially supported by a research grant from the Stony Wold-Herbert Foundation (YS) and by a translational research grant NIH/NCI P30 CA 16087-24 (YS).

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

Contributor Information

Yongzhao Shao, Email: shaoy01@nyu.edu.

Ming Zhou, Email: mingzhou@iastate.edu.

References

  • 1.Anderson TW. An Introduction to Multivariate Statistical Analysis. 3. New York: Wiley; 2003. [Google Scholar]
  • 2.Tong YL. The Multivariate Normal Distribution. New York: Springer-Verlag; 1990. [Google Scholar]
  • 3.Thode HC., Jr . Testing for Normality. New York: Marcel Dekker, Inc.; 2002. [Google Scholar]
  • 4.Sinz F, Gerwinn S, Bethge M. Characterization of the p-generalized normal distribution. Journal of Multivariate Analysis. 2009;100:817–820. [Google Scholar]
  • 5.Looney SW. How to Use Tests for Univariate Normality to Assess Multivariate Normality. American Statistician. 1995;49:64–70. [Google Scholar]
  • 6.Ferguson T. On the Determination of the Joint Distributions from the Marginal Distributions of Linear Combinations (Abstract) Annals of Mathematical Statistics. 1959;30:255. [Google Scholar]
  • 7.Hamedani GG, Tata MN. On the Determination of the Bivariate Normal Distribution from Distributions of Linear Combinations of the Variables. The American Mathematical Monthly. 1975;82:913–915. [Google Scholar]
  • 8.Hamedani GG. Nonnormality of Linear Combinations of Normal Random Variables. American Statistician. 1984;38:295–296. [Google Scholar]
  • 9.Eaton ML, Perlman MD. The non-singularity of generalized sample covariance matrices. The Annals of Statistics. 1973;1:710–717. [Google Scholar]
  • 10.Dykstra RL. Establishing the positive definiteness of the sample covariance matrix. The Annals of Mathematical Statistics. 1970;41:2153–2154. [Google Scholar]
  • 11.Billingsley P. Probability and Measure. 3. New York: Wiley; 1995. [Google Scholar]
  • 12.Arthreya K, Lahiri S. Measure Theory and Probability Theory. New York: Springer; 2006. [Google Scholar]

RESOURCES