Moments and Distributions of Random Vectors

Một phần của tài liệu Introductory econometrics (Trang 822 - 827)

In order to derive the expected value and variance of the OLS estimators using matrices, we need to define the expected value and variance of a random vector. As its name sug- gests, a random vector is simply a vector of random variables. We also need to define the multivariate normal distribution. These concepts are simply extensions of those covered in Appendix B.

1 0 0

0 0 0

0 0 1

Expected Value

DEFINITION D.15 (EXPECTED VALUE)

(i) If y is an n 1 random vector, the expected value of y, denoted E( y), is the vec- tor of expected values: E( y) [E(y1), E(y2), …, E(yn)].

(ii) If Z is an n m random matrix, E(Z) is the n m matrix of expected values:

E(Z)[E(zij)].

PROPERTIES OF EXPECTED VALUE:(1) If A is an m n matrix and b is an n 1 vector, where both are nonrandom, then E(Ayb) AE( y) b; and (2) If A is p n and B is m k, where both are nonrandom, then E(AZB) AE(Z)B.

Variance-Covariance Matrix

DEFINITION D.16 (VARIANCE-COVARIANCE MATRIX)

If y is an n 1 random vector, its variance-covariance matrix, denoted Var( y), is defined as

Var( y) ,

where sj2Var(yj) and sijCov(yi,yj). In other words, the variance-covariance matrix has the variances of each element of y down its diagonal, with covariance terms in the off diagonals. Because Cov(yi,yj) Cov(yj,yi), it immediately follows that a variance- covariance matrix is symmetric.

PROPERTIES OF VARIANCE: (1) If a is an n 1 nonrandom vector, then Var(ay) a[Var(y)]a 0; (2) If Var(ay) 0 for all a 0, Var( y) is positive definite; (3) Var( y) E[( yM)( y M)], where ME( y); (4) If the elements of y are uncorre- lated, Var( y) is a diagonal matrix. If, in addition, Var(yj) s2 for j 1,2, …, n, then Var( y) s2In; and (5) If A is an m n nonrandom matrix and b is an n 1 nonrandom vector, then Var(Ayb) A[Var( y)] A.

Multivariate Normal Distribution

The normal distribution for a random variable was discussed at some length in Appendix B. We need to extend the normal distribution to random vectors. We will not provide an expression for the probability distribution function, as we do not need it. It is important to know that a multivariate normal random vector is completely characterized by its mean and its variance-covariance matrix. Therefore, if y is an n 1 multivariate normal ran- dom vector with mean Mand variance-covariance matrix , we write y ~ Normal(M,).

We now state several useful properties of the multivariate normal distribution.

PROPERTIES OF THE MULTIVARIATE NORMAL DISTRIBUTION:(1) If y ~ Nor- mal(M,), then each element of y is normally distributed; (2) If y ~ Normal(M,), then yi

s12 s12 . . . s1n s21 s22 . . . s2n

.. .

sn1 sn2 . . . sn2

and yj, any two elements of y, are independent if, and only if, they are uncorrelated, that is,sij0; (3) If y ~ Normal(M,), then Ayb ~ Normal(AMb, AA), where A and b are nonrandom; (4) If y ~ Normal(0,), then, for nonrandom matrices A and B, Ay and By are independent if, and only if, AB 0. In particular, if s2In, then AB 0 is necessary and sufficient for independence of Ay and By; (5) If y ~ Normal(0,s2In), A is a k n nonrandom matrix, and B is an n n symmetric, idempotent matrix, then Ay and yBy are independent if, and only if, AB 0; and (6) If y ~ Normal(0,s2In) and A and B are nonrandom symmetric, idempotent matrices, then yAy and yBy are indepen- dent if, and only if, AB0.

Chi-Square Distribution

In Appendix B, we defined a chi-square random variable as the sum of squared inde- pendent standard normal random variables. In vector notation, if u ~ Normal(0,In), then uu ~ xn2.

PROPERTIES OF THE CHI-SQUARE DISTRIBUTION: (1) If u ~ Normal(0,In) and A is an n n symmetric, idempotent matrix with rank(A) q, then uAu ~ xq2; (2) If u

~ Normal(0,In) and A and B are n n symmetric, idempotent matrices such that AB 0, then uAu and uBu are independent, chi-square random variables; and (3) If z ~ Nor- mal (0,C) where C is an m m nonsingular matrix, then zC1z ~ xm2.

t Distribution

We also defined the t distribution in Appendix B. Now we add an important property.

PROPERTY OF THE t DISTRIBUTION:If u ~ Normal(0,In), c is an n 1 nonrandom vector, A is a nonrandom n n symmetric, idempotent matrix with rank q, and Ac 0, then {cu/(cc)1/ 2}/(uAu)1/ 2~ tq.

F Distribution

Recall that an F random variable is obtained by taking two independent chi-square ran- dom variables and finding the ratio of each, standardized by degrees of freedom.

PROPERTY OF THE F DISTRIBUTION:If u ~ Normal(0,In) and A and B are n n nonrandom symmetric, idempotent matrices with rank(A) k1, rank(B) k2, and AB 0, then (uAu/k1)/(uBu/k2) ~ Fk1,k2.

S U M M A R Y

This appendix contains a condensed form of the background information needed to study the classical linear model using matrices. Although the material here is self-contained, it is primarily intended as a review for readers who are familiar with matrix algebra and multi- variate statistics, and it will be used extensively in Appendix E.

K E Y T E R M S

Chi-Square Random Variable

Column Vector Diagonal Matrix Expected Value F Random Variable Idempotent Matrix Identity Matrix Inverse

Linearly Independent Vectors

Matrix

Matrix Multiplication Multivariate Normal

Distribution Positive Definite (p.d.) Positive Semi-Definite

(p.s.d.) Quadratic Form Random Vector Rank of a Matrix Row Vector

Scalar Multiplication Square Matrix Symmetric Matrix t Distribution Trace of a Matrix Transpose

Variance-Covariance Matrix

Zero Matrix

P R O B L E M S

D.1 i(i) Find the product AB using

A , B .

(ii) Does BA exist?

D.2 If A and B are n n diagonal matrices, show that ABBA.

D.3 Let X be any n k matrix. Show that XX is a symmetric matrix.

D.4 (i)i Use the properties of trace to argue that tr(AA) tr(AA) for any n m matrix A.

(ii) For A 20 03 10, verify that tr(AA) tr(AA).

0 1 6

1 8 0

3 0 0

2 1 7

4 5 0

D.5 (i)i Use the definition of inverse to prove the following: if A and B are n n nonsingular matrices, then (AB)1B1A1.

(ii) If A, B, and C are all n n nonsingular matrices, find (ABC)1in terms of A1, B1, and C1.

D.6 (i)i Show that if A is an n n symmetric, positive definite matrix, then A must have strictly positive diagonal elements.

(ii) Write down a 22 symmetric matrix with strictly positive diagonal ele- ments that is not positive definite.

D.7 Let A be an n n symmetric, positive definite matrix. Show that if P is any n n nonsingular matrix, then PAP is positive definite.

D.8 Prove Property 5 of variances for vectors, using Property 3.

The Linear Regression Model in Matrix Form

This appendix derives various results for ordinary least squares estimation of the multiple linear regression model using matrix notation and matrix algebra (see Appendix D for a summary). The material presented here is much more advanced than that in the text.

Một phần của tài liệu Introductory econometrics (Trang 822 - 827)

Tải bản đầy đủ (PDF)

(878 trang)