πŸ“ˆ
Preparation Minor Applied Econometrics
  • πŸ““Preparation module
  • πŸ“‹Table of contents
  • πŸ“šImportant concepts
    • πŸ“‰What is econometrics?
    • ♦️Univariate random variables
    • ♠️Multivariate random variables
    • πŸ§ͺEstimation and inference
    • ♾️Asymptotics
  • πŸ“–Literature
Powered by GitBook
On this page
  1. Important concepts

Multivariate random variables

As alluded to on the previous page, econometricians are often interested in the relationship between two or more random variables. Therefore, we need measures that give us insight into the joint probability distribution of these random variables.

An important quantity is the covariance, which is a measure of joint variability between two random variables. The sign of the covariance (i.e. + or -) indicates whether the random variables generally move in the same or opposite direction. The interpretation of a covariance is generally not straightforward, because the values it can take are not bounded. For that reason, the measure correlation which only takes values between -1 and +1 is often preferred.

Whereas the previous measures are useful, they only tell part of the story. More specifically, they only provide information on how two random variables are related linearly. Correlation is however just one way for random variables to be dependent. Note for example that it is possible for two random variables to be uncorrelated, yet dependent.

Example. Suppose we have two random variables XXXand YYY, such that Y=X2Y=X^2Y=X2. Assume that XXXcan only take values βˆ’1,0-1, 0βˆ’1,0 and 111 each with equal probability 13\frac{1}{3}31​(alternatively, we say that XXXis uniformly distributed on {βˆ’1,0,1}\{ -1, 0 , 1\}{βˆ’1,0,1}). Note that, using elementary results from probability theory:

  • E(X)=(βˆ’1)β‹…13+0β‹…13+1β‹…13=βˆ’13+13=0\mathbb{E}(X) = (-1) \cdot \frac{1}{3} + 0 \cdot \frac{1}{3} + 1 \cdot \frac{1}{3} = -\frac{1}{3} + \frac{1}{3} = 0E(X)=(βˆ’1)β‹…31​+0β‹…31​+1β‹…31​=βˆ’31​+31​=0,

  • E(XY)=E(X3)=(βˆ’1)3β‹…13+03β‹…13+13β‹…13=βˆ’13+13=0=E(X)E(Y)\mathbb{E}(XY) = \mathbb{E}(X^{3}) = (-1)^3 \cdot \frac{1}{3} + 0^3 \cdot \frac{1}{3} + 1^3 \cdot \frac{1}{3} = -\frac{1}{3} + \frac{1}{3} = 0 = \mathbb{E}(X)\mathbb{E}(Y)E(XY)=E(X3)=(βˆ’1)3β‹…31​+03β‹…31​+13β‹…31​=βˆ’31​+31​=0=E(X)E(Y).

Because E(XY)=E(X)E(Y)\mathbb{E}(XY) = \mathbb{E}(X)\mathbb{E}(Y)E(XY)=E(X)E(Y), XXXand YYYare uncorrelated. We can however show that they are dependent:

  • P(X=βˆ’1Β andΒ Y=1)=P(X=βˆ’1)=13\mathbb{P}(X = -1 \text{ and } Y=1) = \mathbb{P}( X = -1) = \frac{1}{3}P(X=βˆ’1Β andΒ Y=1)=P(X=βˆ’1)=31​,

  • P(X=βˆ’1)P(Y=1)=13β‹…23=29\mathbb{P}(X=-1)\mathbb{P}(Y=1) = \frac{1}{3} \cdot \frac{2}{3} = \frac{2}{9}P(X=βˆ’1)P(Y=1)=31​⋅32​=92​.

Because (for example) P(X=βˆ’1,Y=1)β‰ P(X=βˆ’1)P(Y=1)\mathbb{P}(X=-1,Y=1) \neq \mathbb{P}(X=-1)\mathbb{P}(Y=1)P(X=βˆ’1,Y=1)ξ€ =P(X=βˆ’1)P(Y=1), XXXand YYY are dependent.

So far we have focused on the relation between only two random variables. However, often we deal with (many) more random variables. The multivariate generalizations of the concepts we discussed above are the covariance matrix and the correlation matrix. We characterize one of these in the example below.

Example. Suppose we are given three random variables X,Y,ZX,Y,ZX,Y,Z. Then we can define the covariance matrix of these three random variables as

(Var(X)Cov(X,Y)Cov(X,Z)Cov(Y,X)Var(Y)Cov(Y,Z)Cov(Z,X)Cov(Z,Y)Var(Z))\begin{pmatrix} \mathbb{V}\text{ar}(X) & \mathbb{C}\text{ov}(X,Y) & \mathbb{C}\text{ov}(X,Z) \\ \mathbb{C}\text{ov}(Y,X) & \mathbb{V}\text{ar}(Y) & \mathbb{C}\text{ov}(Y,Z) \\ \mathbb{C}\text{ov}(Z,X) & \mathbb{C}\text{ov}(Z,Y) & \mathbb{V}\text{ar}(Z) \end{pmatrix}​Var(X)Cov(Y,X)Cov(Z,X)​Cov(X,Y)Var(Y)Cov(Z,Y)​Cov(X,Z)Cov(Y,Z)Var(Z)​​.

In words: the matrix has the respective variances of the three random variables on the diagonal, while the off-diagonal elements consist of the covariances.

Thus, in the case of multivariate random variables we often use vectors and matrices to "summarize" the relationships between random variables. More generally, we use concepts from linear algebra to investigate the properties of statistical quantities. For example, one could check whether the covariance matrix above has full rank (which is a necessary condition for the matrix to be invertible).

Let us now go back to the concept of joint distribution. Consider the case of two random variables XXXand YYY. The joint distribution, denoted fX,Y(x,y)f_{X,Y}(x,y)fX,Y​(x,y), then determines the probability distribution on all pairs of outcomes. Simply said, it considers all possible combinations of outcomes of (X,Y)(X,Y)(X,Y)and assigns a probability to all of these outcomes. However, the random variables XXXand YYYalso still have their own, separate distributions, denoted fX(x)f_{X}(x)fX​(x) andfY(y)f_{Y}(y)fY​(y) which we call marginal distributions.

The joint and marginal distributions are related through conditional distributions, denoted fX∣Y(x∣y)f_{X|Y}(x|y)fX∣Y​(x∣y) and fY∣X(y∣x)f_{Y|X}(y|x)fY∣X​(y∣x). The concept of conditioning plays a central role in econometrics and allows us to ask questions such as "What is the probability of YYY given that XXXtakes on a certain value xxx?". That is, knowledge of the outcome of a related variable XXXmight influence the likelihood of YYY. It is generally not straightforward to work with conditional distributions, conditional means and conditional variances. But despair not, there are useful theorems out there, such as the law of total expectation and the law of total variance, that aid in simplifying difficult computations.

PreviousUnivariate random variablesNextEstimation and inference

Last updated 1 year ago

πŸ“š
♠️
Page cover image