Skip to main content

Gaussian vector

Definition

Definition

XX is said to be a Gaussian Vector if any linear combination of its components is Gaussian (possibly degenerate i.e. almost sure constant). This means that :

aRd,aTXN(aTmx,aTΓXa).\forall a \in \mathbb R^d, \quad a^TX \sim \mathcal N (a^Tm_x, a^T\Gamma_Xa).

Results

Theorem

X is said to be a Gaussian vector if and only if its characteristic function is written as :

ξRd,ΦX(ξ)=exp(iξTm12ξTΓξ)\forall \xi \in \mathbb R^d, \quad \Phi_X(\xi) = exp(i\xi^Tm-\frac{1}{2}\xi^T\Gamma\xi)
Proof

It suffices to apply the definition, remembering that the characteristic function of the Gaussian distribution N(m,σ2)\mathcal N(m, \sigma^2) is ξeimξσ2ξ22\xi \rightarrow e^{im\xi-\frac{\sigma^2\xi^2}{2}}. For the converse, with the characteristic function that ξTX\xi^TX follows a Gaussian distribution for all ξRd\xi \in \mathbb R^d.


Theorem

Let XX be a Gaussian vector of Rd\mathbb R^d with expectation mm and covariance matrix Γ\Gamma. Then XX has a density if and only if Γ\Gamma is invertible, in which case the density is written as :

fX(x)=1det(2πΓ)exp(12(xm)TΓ1(xm))f_X(x) = \frac{1}{\sqrt{det(2\pi\Gamma)}}exp \left (-\frac{1}{2}(x-m)^T\Gamma^{-1}(x-m)\right )
Proof

It exists UO(n)U \in \mathcal O(n) (i.e, UUT=UTU=IUU^T = U^TU = I) such that UTΓU=D=diag(σ12,,σd2)U^T \Gamma U = D = diag(\sigma^2_1,\dots,\sigma^2_d). Since Γ\Gamma is invertible, we have σi2>0\sigma^2_i > 0. Let R=Udiag(σ1,,σd)UTR = Udiag(\sigma_1,\dots,\sigma_d)U^T that verify R2=ΓR^2 = \Gamma. Let ZN(0,Id)Z \sim \mathcal N(0, I_d) (ZZ is the vector with its coordinates i.i.d. normal and centered and with variance 11), m+RZN(m,Γ)m + RZ \sim \mathcal N(m, \Gamma) and then XX and m+RZm+RZ have the same law. Thus, for any function φ:RdR\varphi : \mathbb R^d \rightarrow \mathbb R mesurable 0\geq 0,

E[φ(X)]=E[φ(m+RZ)]=1(2π)d2Rdφ(m+Rz)ez22dz.\mathbb E[\varphi(X)] = \mathbb E[\varphi(m+RZ)] = \frac{1}{(2\pi)^{\frac{d}{2}}} \int_{\mathbb R^d} \varphi(m + Rz)e^{-\frac{||z||^2}{2}}dz.

Change the affine variable x=m+Rzx = m+Rz. We have

dz=det(R1)dx=dxdet(R)=dxdet(D)=dxdet(Γ)dz = |det(R^-1)|dx = \frac{dx}{|det(R)|} = \frac{dx}{\sqrt{det(D)}} = \frac{dx}{\sqrt{det(\Gamma)}}

that gives

E[φ(X)]=1(2π)d21det(Γ)Rdφ(x)exp(R1(xm22)dx.\mathbb E[\varphi(X)] = \frac{1}{(2\pi)^{\frac{d}{2}}}\frac{1}{\sqrt{det(\Gamma)}}\int_{\mathbb R^d} \varphi(x)exp\left (- \frac{||R^{-1}(x-m||^2}{2} \right )dx.

However

R1(xm)2=(xm)TR1R1(xm)=(xm)T(R2)1(xm)=(xm)TΓ1(xm).\begin{equation} \begin{split}||R^{-1}(x-m)||^2 & = (x-m)^TR^{-1}R^{-1}(x-m) \\ & = (x-m)^T(R^2)^{-1}(x-m) \\ & = (x-m)^T\Gamma^{-1}(x-m). \end{split} \end{equation}

Theorem

Let XX and YY be gaussian vectors of Rp\mathbb R^p and Rq\mathbb R^q respectively. We assume that the vector Z=(XT,YT)Z = (X^T, Y^T) is gaussian. Then X,YX, Y are independant if and only if for all i,j,Cov(Xi,Yj)=0i,j, Cov(X_i, Y_j) = 0

Proof

The direct implication is obvious. For the converse, if X,YX, Y are decorrelated, then the matrix of covariance ZZ is written as

ΓZ=(ΓX00ΓY)\Gamma_Z = \begin{pmatrix} \Gamma_X & 0\\ 0 & \Gamma_Y \end{pmatrix}

But as ZZ is Gaussian, its characteristic function is written as

E[eiζTZ]=eimZζ12ζTΓZζ\mathbb E[e^{i\zeta^TZ}] = e^{im_Z\zeta-\frac{1}{2}\zeta^T\Gamma_Z\zeta}

or also

E[ei(ξTX+ηTY)]=ei(mXξ+mYη)12(ξTΓXξ+ηtΓYη)=E[eiξX]E[eiηY]\mathbb E[e^{i(\xi^TX + \eta^TY)}] = e^{i(m_X\xi+m_Y\eta)-\frac{1}{2}(\xi^T\Gamma_X\xi+\eta^t\Gamma_Y\eta)} = \mathbb E[e^{i\xi X}]\mathbb E[e^{i\eta Y}]