0% found this document useful (0 votes)
14 views3 pages

Midterm2 Cheatsheet Annotated

Uploaded by

guoj310
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
Download as pdf or txt
0% found this document useful (0 votes)
14 views3 pages

Midterm2 Cheatsheet Annotated

Uploaded by

guoj310
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 3

The University of Texas at Austin

ECO 329 (Economic Statistics)


MIDTERM #2 “CHEAT SHEET”

Discrete random variables: univariate

Random variable: maps each outcome of the sample space S to a number


Discrete random variable: random variable with finite or countable set of possible values {x∗k }
Probability mass function (pmf): pX (x∗k ) = P (X = x∗k ) for every possible outcome x∗k
Cumulative distribution function (cdf): FX (x0 ) = P (X ≤ x0 ) = pX (x∗k )
P
x∗k ≤x0

Population mean or population average or expected value: µX = E(X) = x∗k pX (x∗k )


P
k
Population variance: σX
2 = V ar(X) = E[(X − µ )2 ] =
k (xk
∗ − µX )2 pX (x∗k )
P
X
q qP
Population standard deviation: σX = 2 =
σX k (xk
∗ − µX )2 pX (x∗k )
Linear transformation of a random variable: Y = a + bX

• Population mean: µY = a + bµX

• Population variance: σY2 = b2 σX


2

• Population standard deviation: σY = |b|σX

X−µX
Standardized random variable: Z = σx has µZ = 0 and σZ2 = σZ = 1
Bernoulli random variable: X ∼ Bernoulli(π), where π = P (X = 1) and 1 − π = P (X = 0)

• µX = π, σX
2 = π(1 − π), σ =
p
X π(1 − π)

Binomial random variable: X ∼ Binomial(n, π), where X is the total number of 1’s (“successes”) from n
independent trials of the Bernoulli(π) random variable

• pmf: P (X = x) = n x
x π (1 − π)n−x for x ∈ {0, 1, 2, . . . , n}

• µX = nπ, σX
2 = nπ(1 − π), σ =
p
X nπ(1 − π)

“Sample proportion” random variable: P = n1 X, where X ∼ Binomial(n, π)


q
π(1−π) π(1−π)
• µP = π, 2
σX = n , σX = n

Discrete random variables: bivariate

Joint probability function: pXY (x, y) = P (X = x ∩ Y = y) = P (X = x, Y = y)

• pXY (x∗k , y`∗ ) ≥ 0 for any possible outcome pair (x∗k , y`∗ )

• pXY (x∗k , y`∗ ) = pXY (x∗k , y`∗ ) = 1


P PP
(k,`) k `

Joint cumulative distribution function: FXY (x, y) = P (X ≤ x ∩ Y ≤ y) = P (X ≤ x, Y ≤ y)

1
Obtaining marginal probability functions from the joint probability distribution:

• pX (x∗k ) = ` pXY (xk , y` )


∗ ∗ and pY (y`∗ ) = pXY (x∗k , y`∗ )
P P
k

pXY (x,y)
Conditional probability function (of X given Y ): pX|Y (x|y) = P (X = x|Y = y) = pY (y)
Conditional mean (of X given Y ): µX|Y =y`∗ = E(X|Y = y`∗ ) = x∗k pX|Y (x∗k |y`∗ )
P
k

Conditional variance (of X given Y ): σX|Y


2
=y ∗ = V ar(X|Y = y` ) =

k (xk
∗ − µX|Y =y`∗ )2 pX|Y (x∗k |y`∗ )
P
`

Population covariance: σXY = Cov(X, Y ) = (k,`) (xk − µX )(y` − µY ) pXY (xk , y` )


∗ ∗ ∗ ∗
P

Population correlation: ρXY = Corr(X, Y ) = σXY


σX σY (−1 ≤ ρXY ≤ 1; sign(ρXY )=sign(σXY ))
Independence:

• X and Y are independent if and only if pXY (x∗k , y`∗ ) = pX (x∗k )pY (y`∗ ) for every possible outcome pair
(x∗k , y`∗ ). If this equality fails for any (x∗k , y`∗ ) pair, then X and Y are dependent.

• X and Y are independent if and only if:

– For any possible value y`∗ , the conditional distribution of X given Y = y`∗ is the same as the
unconditional (marginal) distribution of X. (And the same when switching roles of X and Y .)

Linear transformations of random variables: V = a + bX, W = c + dY

• Covariance: σV W = bdσXY

• Correlation: ρV W = ρXY if bd ≥ 0 and −ρXY if bd < 0

Linear combination of two random variables: V = k + aX + bY

• Mean: µV = k + aµX + bµY

• Variance: σV2 = a2 σX
2 + b2 σ 2 + 2abσ
Y XY

– Special case (a = b = 1; V = X + Y ): σV2 = σX


2 + σ 2 + 2σ
Y XY
– Special case (a = 1, b = −1; V = X − Y ): σV2 = σX
2 + σ 2 − 2σ
Y XY
q q
• Standard deviation: σV = σV2 = 2 + b2 σ 2 + 2abσ
a2 σX Y XY

Continuous random variables: univariate

Continuous random variable: random variable taking on values within an interval of the real line, multiple
intervals of the real line, or the entire real line (with an uncountable set of possible outcomes)
´b
Probability density function (pdf): function fX (x) such that P (a ≤ X ≤ b) = a fX (x) dx
´ x0
Cumulative distribution function (cdf): FX (x0 ) = P (X ≤ x0 ) = −∞ fX (x) dx
Relationship between pdf and cdf: fX (x0 ) = FX0 (x0 )
Population quantile τX,q : for any q between 0 and 1, τX,q is value such that P (X ≤ τX,q ) = FX (τX,q ) = q
´∞
Population mean or population average or expected value: µX = E(X) = −∞ xfX (x) dx
´
2 = V ar(X) = E[(X − µ )2 ] = ∞ (x − µ )2 f (x) dx
Population variance: σX X −∞ X X
q q´

Population standard deviation: σX = sd(X) = σX 2 =
−∞ (x − µX ) fX (x) dx
2

2
Symmetry: X has a symmetric distribution if, for a midpoint x∗ , fX (x∗ − v) = fX (x∗ + v) for all v ≥ 0.
Linear transformations and combinations: same formulas hold as in the discrete X case
Uniform distribution with lower bound a and upper bound b: fX (x) = 1
b−a for a ≤ x ≤ b and 0 otherwise

Normal random variables

√1 e− 2 ( σ )
1 x−µ 2
Normal distribution: X ∼ N (µ, σ 2 ) has pdf fX (x) = σ 2π
for −∞ < x < ∞ with E(X) =
µ, V ar(X) = σ 2
Standard normal: normal with µ = 0 and σ 2 = σ = 1, that is Z ∼ N (0, 1)
X−µ
Standardized normal RV: X ∼ N (µ, σ 2 ) implies σ ∼ N (0, 1) = Z and X = µ + σZ
Probability intervals for the standard normal (Z ∼ N (0, 1)):

• P (−1.96 < Z < 1.96) ≈ 0.95 (that is, z0.025 = 1.96)


• P (−1.645 < Z < 1.645) ≈ 0.90 (that is, z0.05 = 1.645)

Linear combination of normal random variables is a normal random variable (figure out mean and variance
from the linear combination formulas)
Log-normal random variable: X is log-normal if ln(X) ∼ N (µ, σ 2 ) for some µ and σ 2

Continuous random variables: bivariate

´d´b
Joint pdf of X and Y : function fXY (x, y) such that P (a ≤ X ≤ b, c ≤ Y ≤ d) = c a fXY (x, y) dx dy
Population covariance: σXY = Cov(X, Y ) = E[(X − µX )(Y − µY )] (don’t worry about integral form)
Population correlation: ρXY = Corr(X, Y ) = σXY
σX σY (−1 ≤ ρXY ≤ 1; sign(ρXY )=sign(σXY ))
Independence of continuous RV’s: X and Y independent if and only if fXY (x, y) = fX (x)fY (y)

• Independence implies σXY = ρXY = 0 (but not vice versa)


• Independence implies P (a ≤ X ≤ b, c ≤ Y ≤ d) = P (a ≤ X ≤ b) · P (c ≤ Y ≤ d)

Sampling distributions

i.i.d. random variables: X1 , X2 , . . . , Xn are independent and identically distribution (i.i.d.) if they are
mutually independent and have the same distribution function (FXi (x) = FX (x) for all i ∈ {1, 2, . . . , n}
and all x)

Sampling distribution of the sample average (X̄ = n1 (X1 + X2 + . . . + Xn ) for i.i.d. RV’s)

• Xi ∼ Bernoulli(π) implies X̄ ∼ n1 Binomial(n, π)


 2

• Xi ∼ N (µ, σ 2 ) implies X̄ ∼ N µ, σn

s2
Sampling distribution of the sample variance for i.i.d. normal random variables: (n − 1) σX2 ∼ χ2n−1 , where
X
χ2n−1 has a chi-squared distribution with n − 1 degrees of freedom. (If Z1 , . . . , Zk are i.i.d. N (0, 1), then it
is said that Z12 + · · · + Zk2 has a χ2k distribution.)

You might also like