Var 3 2 3 1 3 Var 3

Download as pdf or txt
Download as pdf or txt
You are on page 1of 4

UC Berkeley

Department of Electrical Engineering and Computer Science

EE 126: Probablity and Random Processes

Problem Set 5
Spring 2007

Issued: Thursday, March 8, 2007 Due: Thursday, March 15, 2007

Problem 5.1

1. N = 200, 000.

2. N = 100, 000.

Problem 5.2

a)This is a straight application of the Chebychev inequality. Chebychev tells us that:


P (|X − 7| ≥ 3) ≤ V3ar 2 1
2 = 3 and therefore that P (4 < X < 10) ≥ 3 .
V ar
b)If the variance = 9, P (|X − 7| ≥ 3) ≤ 32 = 1 and therefore that P (4 < X < 10) ≥ 0.

Problem 5.3

(a) 
1/4k if k = 1, 2, 3, 4 and n = 1, . . . , k
pN,K (n, k) =
0 otherwise

(b) 

 1/4 + 1/8 + 1/12 + 1/16 = 25/48 n=1
 1/8 + 1/12 + 1/16 = 13/48 n=2


pN (n) = 1/12 + 1/16 = 7/48 n=3
 1/16 = 3/48 n=4



0 otherwise

(c) The conditional PMF



 6/13
 k=2
pN,K (2, k)  4/13 k=3
pK|N (k|2) = =
pN (2) 
 3/13 k=4
0 otherwise

(d) Let A be the event that Chuck bought at least 2 but no more than 3 books, E[K|A] = 3
var(K|A) = 35
21
(e) E[T ] = 4

1
Problem 5.4

1.  1

 20 , x=1
3



 20 , x=2

3 13
pX (x) = 10 , x=3 , E[X] =

1
4
, x=4




 2

0 , otherwise

2.
1


 10 , w=2
 1
, w=3




 20
 7

20 , w=4
pW (w) = 3


 10 , w=6
1

, w=8




 5
0 , otherwise

3. E[R] = 34 , var(R) = 63
80

3
4. σR|A = 4

Problem 5.5

1. µ = 3, σ 2 = 2

2. µ = 5, σ 2 = 20.

Problem 5.6

1. To calculate the probability of heads on the next flip, we use the continuous version of
the total probability theorem:

Z 1
P (H) = P (H|P = p) · fP (p)dp
0
Z 1
= p2 ep dp
0
= e−2

2
2. Here we need to use the continuous version of Bayes’ theorem:

P (A|p)fP (p)
fP |A (p|A) = R1
P (A|p)fP (p)dp
(0
p2 ep
e−2 , 0≤p≤1
=
0, otherwise

3. Now we apply the above result to the technique used in (a):

Z 1
P (H) = P (A|P = p) · fP |A (p|A)dp
0
Z 1
1
= p3 ep dp
e−2 0
1
= · (6 − 2e)
e−2
.564
=
.718
≈ .786

Problem 5.7

Note that we can rewrite E[X1 | Sn = sn , Sn+1 = sn+1 , . . .] as follows:

E[X1 | Sn = sn , Sn+1 = sn+1 , . . .]


= E[X1 | Sn = sn , Xn+1 = sn+1 − sn , Xn+2 = sn+2 − sn+1 , . . .]
= E[X1 | Sn = sn ],

where the last equality holds due to the fact that the Xi ’s are independent.
We also note that

E[X1 + · · · + Xn | Sn = sn ] = E[Sn | Sn = sn ] = sn

It follows from the linearity of expectations that

E[X1 + · · · + Xn | Sn = sn ] = E[X1 | Sn = sn ] + · · · + E[Xn | Sn = sn ]

Because the Xi ’s are identically distributed, we have the following relationship:

E[Xi | Sn = sn ] = E[Xj | Sn = sn ], for any 1 ≤ i ≤ n, 1 ≤ j ≤ n.

Therefore,
sn
E[X1 + · · · + Xn | Sn = sn ] = nE[X1 | Sn = sn ] = sn ⇒ E[X1 | Sn = sn ] = .
n

3
Problem 5.8

a) γ is determined because the density function must integrate to 1. Since (X, Y ) uni-
formly distributed in R, we have: ZZ
γdxdy = 1

where the integral is over the area of R. Therefore γ1 =Area R.


b) Showing independence is equivalent to showing that:

fX,Y (x, y) = fX (x) · fY (y)

But this is clear, since:


fX,Y (x, y) = fY (x) = fY (y) = 1
c) Consider a region that consists of the upper half plane, and the point (1, −1). If we are
told that Y < 0, X is determined, and hence X, Y cannot be independent in this region.
d) To find the probability that (X, Y ) lie in the circle C inscribed in the R in part (b) we
could integrate, or observe that the integral will in fact come out to the area of the circle,
and hence the desired probability will be the ratio of the are of the circle to the area of the
square:
.25π
P ((X, Y ) ∈ C) = .
1

You might also like