04 Random Vector properties

Mean

E[X]:=(E[Xi]).

Linearity

If X is a random matrix, A, B, C are constant matrices: E[AXB+C]=AE[X]B+C. Proof: by using \((AXB){i, j} = A{i,:} X B_{:, j}\), which is a linear combination of Xk,l.

Also, if X is random vector, E[aTX]=aTE[X].

Covariance

Definition

How correlated are deviations of X, Y from their means?\ cov(X,Y)=Ex,y[(XE[X])(YE[Y])].

Extension to vectors

cov(X,Y)=Ex,y[(XE[X])T(YE[Y])]:\ corresponds to measuring cov(Xi,Yj).

Correlation

(Pearson) correlation coefficient: corr(X,Y)=cov(X,Y)σXσY: normalized covariance.

Correlation vs Independence

If XiXj,Cov[Xi,Xj]=0: even if they are only pairwise independent. But, \ cov(X,X2)=0 even if (X,X2) not .

If Cov[Xi,Xj]=0 holds, then Xi and Xj are uncorrelated. If they are independent, they are uncorrelated; but not necessarily vice versa.

Covariance matrix

Σ=var[X]=cov(X,X)=E[(Xμ)(Xμ)T]=E[XXT]μμT.

Diagonal has variances of (Xi). It is diagonal if (Xi) are independent.

Effect of linear transformation

Var[BX+a]=E[(BXBE[X])(BXBE[X])T]=BVar[X]BT. As in the scalar case, constant shifts have no effect.

Special case: var[aTX]=aTvar(X)a.

Nonnegative definiteness

Σ0 as aTE[(Xμ)(Xμ)T]a0.

If aTΣa=0, with probability 1, aTXaTμ=0, so some {Xi} are linearly dependent. So, X lies on the hyperplane/ subspace with normal a.

Precision matrix

V=Σ1. Consider partial correlation deduced in case of multidimensional normal distribution.

Moment generating function

E[etTX] is the moment generating function.