Joint Probability Distribution
Joint Probability Distribution
Joint Probability Distribution
W xi ,=
yj (( X x=
P= i ,Y yi ) )
Note that as usual, the comma means "and,".
y x
F ( x , y ) = P ( X ≤ x ,Y ≤ y ) = ∑ ∑ P(X =
y j = −∞ xi = −∞
xi ,Y = y j )
W ( x , y ) ≥ 0;
+∞ +∞
∑ ∑ P (=
X
y j = −∞ xi = −∞
xi = )
,Y y=
j 1.
The cdf is
y x
F ( x , y ) = P ( X ≤ x ,Y ≤ y ) = ∑ ∑ P(X =
y j = −∞ xi = −∞
xi ,Y = y j )
The marginal PMF of X, denoted by W (x), is given by
both ↑ .
Implies that the volume beneath any joint pdf must be unity.
y x
3. F ( x , y ) = ∫ ∫ f ( u, v ) dudv
−∞ −∞
+∞ +∞
=4. f ( x ) f ( x , y ) dy ; f ( y )
∫= ∫ f ( x , y ) dx .
−∞ −∞
x 2 y2
5. P ( x1 ≤ X ≤ x2 , y1 ≤ Y ≤ y2 ) =∫ ∫ f ( x , y ) dydx
x1 y1
+∞ x
F ( x) = ∫ ∫ f ( u, v ) dudv
−∞ −∞
6. +∞ y
F ( y) = ∫ ∫ f ( u, v ) dvdu
−∞ −∞
Let Z= Φ (X,Y)
Suppose if x ≤ X ≤ x + dx and y ≤ Y ≤ y + dy ,
then, z ≤ Z ≤ z + dz .
Then, P ( z ≤ Z ≤ z + dz )
= P ( x ≤ X ≤ x + dx , y ≤ Y ≤ y + dy )
or ,
f ( z )dz = f ( x , y )dxdy
+∞
+∞ +∞
E(z)
= ∫=
−∞
zf ( z )dz ∫ ∫
−∞ −∞
φ ( x , y ) f ( x , y )dxdy
E[X]
= xf ( x )dx ∫ x ∫ f ( x , y ) dy dx
∫=
−∞ −∞ −∞
&
+∞ +∞ +∞
E [Y ]
= yf ( y )dy ∫ y ∫ f ( x , y ) dx dy
∫=
−∞ −∞ −∞
X & Y are s-independent if the occurrence of any o/c X=x does not
influence the occurrence of any o/c Y=y & vice versa.
Necessary & sufficient condn :
y ) P ({ X ≤ x} {Y ≤ y} )
P ( X ≤ x,Y ≤ =
{ X ≤ x} & {Y ≤ y} are s-independent,
P ({ X ≤ x} {Y ≤ y} )= P ({ X ≤ x} ) . P ({Y ≤ y} )
) P ( X ≤ x ) . P (Y ≤ y )
P ( X ≤ x , Y ≤ y=
F ( x , y ) = F ( x ) .F ( y )
y x x y
∫ ∫ f ( λ , s ) d λ ds = ∫ f ( λ ) d λ . ∫ f ( s ) ds
−∞ −∞ −∞ −∞
Similarly,
W ( x , y ) = W ( x ) .W ( y ) for drv.
∫ xf ( x )dx ×
=
−∞
∫
−∞
yf ( y )dy =
E ( X ) × E (Y )
Cov ( X , Y ) = E ( X − µ X )( Y − µY )
+∞ +∞
=∫ ∫ ( x − µ )( y − µ )
−∞ −∞
X Y f ( x , y )dxdy
Cov ( X , Y ) =E [ XY ] − µ X E [Y ] − µY E [ X ] + µ X µY
= E [ XY ] − µ X µY
0 ≤ ρ ≤1
f ( x, y ) =
6 × 10−6 exp ( −0.001 x − 0.002 y ) ; for x < y
Reasonable assumptions can be used to develop such a distribution, but for
now, our focus is only on the joint probability density function. The region
with nonzero probability is shaded in Fig 5.8.
The property that this joint pdf integrates to 1 can be verified by the
integral of f(x, y) over this region as follows:
The probability that X<1000 & Y < 2000 is determined as the integral over
the darkly shaded region in Fig. 5-9.
The probability that Y exceeds 2000 milliseconds is determined as the
integral of f (x, y) over the darkly shaded region in Fig. 5-10.
The region is partitioned into two parts & different limits of integration
are determined for each part.
Jointly Gaussian Random Variables:
As with single RVs, the most common & important example of a
two-dimensional probability distribution is that of a joint Gaussian
distribution.
.
DEFINITION: A pair of RVs X & Y are said to be jointly Gaussian if their
joint pdf is of the general form
where μX & μY are the expectations of X& Y, respectively; σX and σY are the
standard deviations of X & Y, respectively; and ρXY is the correlation
coefficient of X & Y.
It is left as an exercise for the reader to verify that this joint PDF results in
marginal pdfs that are Gaussian. That is
For a set of N number of RVs X1,X2, . . . ,XN, the Nth order joint
PMF, CDF, and PDF are given respectively by
W ( x K 1 , x K 2 , x K 3 , .....,=
xkN ) P=
( X 1 x K=
1, X2 2, X3
x K= x K 3 , .....,
= X N xkN )
F ( x1 , x2 , x3 , ....., x N ) = P ( X 1 ≤ x1 , X 2 ≤ x2 , X 3 ≤ x3 , ....., X N ≤ x N )
and
∂ N F ( x1 , x2 , x3 , ....., x N )
f ( x1 , x2 , x3 , ....., x N ) =
∂x1∂x2 ∂x3 .....∂x N
Marginal CDFs can be found for a subset of the variables by evaluating the
joint CDF at infinity for the unwanted variables. For example,
F (=
x1 , x2 , x3 , ....., x M ) F ( x1 , x2 , x3 , ....., x M , ∞ , ∞ , ...., ∞ )
Marginal PDFs are found from the joint PDF by integrating out the
unwanted variables. Similarly, marginal PMFs are obtained from the joint
PMF by summing out the unwanted variables:
+∞
implying
f ( x1 , x2 ) ( x1 ) f ( x2 ) ; f ( x2 , x3 )
f= f ( x2 ) f ( x3 )
& f ( x2 , x3 ) = f ( x2 ) f ( x3 )
would be just the marginal distribution function of Y-that is, F(y). From the
basic definition of the conditional distribution function, it would follow that
F ( x, y )
F ( x Y ≤ y )= P ( X ≤ x Y ≤ y )=
F ( y)
Another possible definition of M is that it is the event y1 < Y< y2. The
definition of conditional distribution now leads to
F ( x , y2 ) − F ( x , y1 )
) P ( X ≤ x y1 < Y < y2=)
F ( x y1 < Y < y2=
F ( y2 ) − F ( y1 )
∫ f ( u, y ) du
∴ F ( x y) =
−∞
f ( y)
Hence,
∂F ( x , y )
f ( x, y )
f ( x y=
) f (x Y ) ∂∂x F ( x y=) ∂f x(∂yy) =
= y=
f ( y)
Similarly,
f ( x, y )
f ( y x) = ; and consequently, we have the Bayes’ theorem
f ( x)
f ( x y) f ( y) f ( y x) f ( x)
f ( y x) = & f ( x y)
f ( x) f ( y)
We also have,
+∞ +∞
=f ( x) f ( x , y ) dy
∫=
−∞
∫ f ( x y ) f ( y ) dy
−∞
&
+∞ +∞
=f ( y) f ( x , y ) dx
∫=
−∞
∫ f ( y x ) f ( x ) dx
−∞
Similarly,
+∞
E Y X E Y =
X x=
∫ yf ( y x ) dy
−∞
One of the most common (and probably the simplest) use of conditional
density functions is that in which some observed quantity is the sum of two
quantities-one of which is usually considered to be a signal while the other
is considered to be a noise. Suppose, for example, that a signal X (t ) is
perturbed by additive noise N (t ) that the sum of these two, Y(t) , is the
only quantity that can be observed. Hence, at some time instant, there are
three random variables related by
Y = X+ N
and it is desired to find the conditional probability density function of X
given the observed value of Y-that is, f (x l y ) .
The reason for being interested in this is that the most probable values of
X, given the observed value Y, may be a reasonable guess, or estimate, of
the true value of X when X can be observed only in the presence of noise.
From Bayes' theorem this conditional probability is
f ( y x) f ( x)
f ( x y ) ==
f ( y)
&
f ( x, y) f ( x) f ( y)
( y x)
f= =
f ( x)
=
f ( x)
f ( y)
For a set of N RVs X1,X2, . . . ,XN, the conditional PMF and PDF are
given by
(K1, X2
P X 1 x=
= x K 2 ,=
..., X M x KM= , X M + 2 x K M + 2 ,=
X M + 1 x K M + 1= ..., X N x KN )
=P ( X 1 x=K1, X2 ..., X N x KN )
x K 2 ,=
=
= (
P X M +1 x=K M +1 , X M + 2 x K=M +2
, ..., X N x KN )
(
i .e . W x K 1 , x K 2 , ..., x KM x K M +1 , x K M + 2 , ..., x KN )
W ( x K 1 , x K 2 , ..., x KN )
=
(
W x K M +1 , x K M + 2 , ..., x KN )
and
f ( x1 , x2 , ..., x N )
( )
f x1 , x2 , ..., x M x M +1 , x M + 2 , ..., x N =
f ( x M +1 , x M + 2 , ..., x N )