Escolar Documentos
Profissional Documentos
Cultura Documentos
Probability
(Jim Pitman)
http://www2.imm.dtu.dk/courses/02405/
Question b) 67%.
Question c) 0.667
4
Question a.2) 7
It may be useful to read the definition of Odds and payoff odds in Pitman pp. 6 in
order to solve this exercise
Question b) The average gain pr. game is defined as the profit divided by the number
of games
pr −10
= = −0.1
n 100
1
Denote the fraction the neighbor gets by x. Then your friend gets 2x and you get 4x.
The total is one, thus x = 71 and you get 74 .
1
(A ∩ B c ) ∪ (Ac ∩ B)
(Ac ∩ B c ∩ C c )
It may be useful to make a sketch similar to the one given at page 22 in Pitman.
Question a)
Question b)
Question c)
Question d)
Question e)
Question f )
using exclusion-inclusion.
Question b)
Question c)
P (F c ∩ Gc ∩ H) = P ((F ∪ G)c ∩ H)
P (H) = P ((F ∪ G)c ∩ H) + P ((F ∪ G) ∩ H)
The latter probability is
P (A ∪ B ∪ C) = P (A ∪ (B ∪ C))
now applying inclusion-exclusion
once again we aplly inclusion-exclusion (the second and the third time) to get
We know from exercise 1.3.11 that the formula is valid for n = 3 and consider
P ∪n+1 n
i=1 Ai = P ((∪i=1 Ai ) ∪ An+1 ) .
Using exclusion-inclusion for two events we get the formula stated p.32. Since the
exclusion-inclusion formula is assumed valid for n events we can use this formula for
the first term. To get through we realize that the last term
P (∪ni=1 Ai An+1 )
is of the form
P (∪ni=1 Bi )
with Bi = Ai ∩ An+1 , implying that we can use the inclusion-exclusion formula for this
term too. The proof is completed by writing down the expansion explicitly.
1
Question c) True
Question d) True
Question e)
3 1
· 0.92 + · 0.88 = 0.91
4 4
true
1
solution
Question a)
S Chip sold
P (D|H c )P (H c )
P (H c |D) = = 0.8 · 0.010.008 + 0.05 · 0.99 = 0.139
P (D|H c )P (H c ) + P (D|H)P (H)
Denote the event that a shape of type i is picked by Ti , the event that it lands flat by
F and the event that the number rolled is six by S. We have P (T i ) = 31 , i = 1, 2, 3,
P (F |T1 ) = 13 , P (F |T2 ) = 12 , and P (F |T3 ) = 32 P (S|F ) = 21 , and P (S|F c ) = 0.
P (S) = P (S|T1 ∩F )P (T1 ∩F )+P (S|T2 ∩F )P (T2 ∩F )+P (S|T3 ∩F )P (T3 ∩F )+P (S|T1 ∩F c )P (T1 ∩F c )+P
The last three terms are zero. We apply The Multiplication Rule for the proba-
bilities P (Ti ∩ F ) leading to
P (S) = P (S|T1 ∩F )P (F |T1 )P (T1 )+P (S|T2 ∩F )P (F |T2 )P (T2 )+P (S|T3 ∩F )P (F |T3 )P (T3 )
a special case of The Multiplication Rule for n Events page 56. Inserting numbers
111 111 121 1
P (S) = + + =
233 223 233 4
Question b) The probability in question is P (T1 |S). Applying Bayes’ rule page 49
11
P (S|T1 )P (T1 ) 63 2
P (T1 |S) = = 1 =
P (S) 4
9
1
This is another version of the birthday problem. We denote the event that the first n
persons are born under different signs, exactly as in example 5 page 62. Correspond-
ingly, Rn denotes the event that the n’th person is the first person born under the same
sign as one of the previous n − 1 persons. We find
n
Y
i−1
P (Dn ) = 1− , n ≤ 13
i=1
12
Question b)
n−1
364 1 ln (2)
1− ≥ ⇔n≥ + 1 = 253.7
365 2 ln (365) − ln (364)
Question c) In the birthday problem we only ask for two arbitrary birthdays to be
the same, while the question in this exercise is that at least one out of n − 1 has
a certain birthday.
1
P (1) = 0
1
P (2) =
6
52
P (3) =
66
543
P (4) =
666
5434
P (5) =
6666
54325
P (6) =
66666
54321
P (7) =
66666
Question b) The sum of the probabilities p2 to p6 must be one, thus the sum in
question is 1.
Question b)
1
P (Bij ) =
365
It is immediately clear that
1
P (B12 ∩ B23 ) = = P (B12 )P (B23 ).
3652
implying independence. The following is a formal and lengthy argument. Define
Aij as the the event that the i’th person is born the j’th day of the year.
1
We have P (Aij ) = 365 and that A1i , A2,j , A3,k , and A4,l are independent. The
event Bij can be expressed by
Bij = ∪365
k=1 (Ai,k ∩ Aj,k )
1
such that P (Bij ) = 365 by the independence of Ai,k and Aj,k . The event B12 ∩B23
can be expressed by
1
and by the independence of the A’s we get P (B12 ∩ B23 ) = 3652
Question b) The probability in question is given by the binomial distribution, see eg.
page 81. 3 4
5 1 35 · 125
35 = = 0.0156
6 6 67
1
We define the events Gi: i girls in family. The probabilities P (Gi) is given by the
binomial distribution due to the assumptions that the probabilities that each child is
a girl do not change with the number or sexes of previous children.
i 4−1
4 1 1 1 3
P (Gi) = , P (G2) = 6 · =
i 2 2 16 8
5
P (G2c ) = 1 − P (G2) =
8
1
We denote the event that there are 3 sixes in 8 rolls by A, the event that there are 2
sixes in the first 5 rolls by B. The probability in question is P (B|A). Using the general
formula for conditional probabilities page 36
P (B ∩ A)
P (B|A) =
P (A)
P (B ∩ A) P (A|B)P (B)
P (B|A) = =
P (A) P (A)
Now the probability of P (A) is given by the binomial distribution page 81, as is P (B)
and P (A|B) (the latter is the probability of getting 1 six in 3 rolls). Finally
5 53 3 52 5 3
P (2 sixes in 5 rolls)P (1 six in 3 rolls) 2 65 1 63 2 1
P (B|A) = = =
P (3 sixes in 8 rolls) 5 55 8
2 68 3
We define events Bi that the man hits the bull’s eye exactly i times. The probabilities
of the events Bi is given by the Binomial distribution
8
P (Bi) = 0.7i0.38−i
i
Question b)
Question c)
6
0.72 0.34 = 0.0595
2
1
We apply The Normal Approximation to the Binomial Distribution page 99. Note
that b=∞
˜ such that the first term is 1. We have µ = n · p = 300 · 31 = 100 and
q q
σ = 300 3 3 = 10 23 . The value of a in the formula is 121 (more than 120). We get
12
120.5 − 100
P (More than 120 patients helped = 1−Φ = 1−Φ(2.51) = 1−0.994 = 0.006
8.165
1
−P (W 390∪W 391∪W 392∪W 393∪W 394∪W 395∪W 396∪W 397∪W 398∪W 399∪W 400)
Question b)
k + 21 − 400 · 0.95
˜ −Φ
P (at least k)=1 √ ≥ 0.95
400 · 0.95 · 0.05
With
k + 21 − 400 · 0.95
√ ≤ −1.645
400 · 0.95 · 0.05
we find k = 373.
1
Question b)
25
0.12 0.923 = 0.2659
2
Question c)
2 + 21 − 2.5 1 + 12 − 2.5
Φ √ −Φ √ = Φ(0) − Φ(−0.667) = 0.2475
25 · 0.09 25 · 0.09
Question d)
2.52 −2.5
·e = 0.2566
2!
Question e) Normal m is now 250
250 + 21 − 250 250 − 21 − 250 1 1
Φ √ −Φ √ = Φ( ) − Φ(− ) = 0.0266
2500 · 0.09 2500 · 0.09 30 30
µk −
Pµ (k) = e µ
k!
We use odds ratio for the probabilities
µk+1 −
P (k + 1) (k+1)!
e µ µ
= k
µ −
=
P (k) e µ k+1
k!
The ratio is strictly decreasing in k. For µ < 1 maximum will be P µ (0), otherwise the
probabilities will increase for all k such that µ > k, and decrease whenever µ < k. For
non-integer µ the maximum of Pµ (k) (the mode of the distribution) is obtained for the
largest k < µ. For µ intger the value of Pµ (µ) = Pµ (µ + 1).
1
The probability of the event that there is at least one success can be calculated using
the Binomial distribution. The probability of the complentary event that there is no
successes in n trials can be evaluated by the Poisson approximation.
1 2
P (0) = e− N 3 N = 0.5134
Similarly for n = 53 N
1 5
−N N 5
P (0) + P (1) = e 3 1+ = 0.5037
3
1
(the hypergeometric distribution page 125). The probability that the second
sample contains more than one bad item is calculated via the probability of the
complementary event, i.e. that the second sample contains one or two bad items,
which is
9 36 9 36
0 10 1 9
p2 = +
45 45
10 10
The answer to the question is the product of these two probabilities p 1 (1 − p2 ) =
0.2804.
Question b) The lot is accepted if we have no bad items in the first sample or the
event described under a)
10 40 10 40 9 36 9 36
0 5 1 4 0 10 1 9
+ + = 0.4595
50 50 45 45
5 5 10 10
1
The random variable Z = X1 X2 has range {1, 2, 3, 4, 5, 6, 8, 9, 10, 12, 15, 16, 18, 20, 24, 25, 36}.
We find the probability of Z = i by counting the combinations of X 1 , X2 for which
X1 X2 = i. we get:
Z = i P (Z = i)
1
1 36
2
2 36
2
3 36
3
4 36
2
5 36
4
6 36
2
8 36
1
9 36
2
10 36
4
12 36
2
15 36
1
16 36
2
18 36
2
20 36
2
24 36
1
25 36
2
30 36
1
36 36
1
Question b)
7
X
4 g−1
p q g−4
3
g=4
Question c) The easiest way is first answering question d) then using 1−binocdf (3, 7, 2/3)
in MATLAB.
0.8267
Question d) Imagine that all games are played etc. From the binomial formula
Question e)
P (G = 4) = p4 + q 4 P (G = 5) = 4pq(p3 + q 3 )
P (G = 6) = 10p2 q 2 (p2 + q 2 ) P (G = 7) = 20p3 q 3 (p + q)
1
Independence for p = q = 2
1
1 − (p0 p1 + p0 p2 + p1 p2 − 3p0 p1 p2 )
Question b) Let Y denote the number of odd numbers on three rolls, then Y ∈
binomial 3, 12 thus E(Y ) = 3 · 21 = 32 .
1
We define the indicator variables Ii which are 1 of switch i are closed 0 elsewhere. We
have X = I1 + I2 + · · · + In , such that
n
X
E(X) = E(I1 + I2 + · · · + In ) = E(I1 ) + E(I2 ) + · · · + E(In ) = p1 + p2 + · · · + pn = pi
i=0
1
Question b)
3 10 3 10
3 6 3 5
P (D = 9) = P (D ≤ 9) − P (D ≤ 8) = − = 0.2284
13 13
9 8
The computational formula for the variance page 186 is quite useful (important). This
exercise is solved by applying it twice. First we use it once to get:
E((X1 X2 )2 )−(E(X1 X2 ))2 = E(X12 X22 )−(E(X1 )E(X2 ))2 = E(X12 )E(X22 )−(E(X1 )E(X2 ))2
using the multiplication rule for Expectation page.177 valid for independent random
variables. We have also used the fact that if X1 and X2 are independent then f (X1 )
and g(X2 ) are independent too, for arbitrary functions f () and g(). We now use the
computational formula for the variance once more to get
V ar(X1 X2 ) = (V ar(X1 ) + (E(X1 ))2 )(V ar(X2 ) + (E(X2 ))2 ) − (E(X1 )E(X2 ))2
We apply the Normal approximation (the Central Limit Theorem P(p.196). Let X i
30
denote the weight of the i’th passenger. The total load W is W = i=1 Xi .
5000 − 30 · 150
P (W > 5000)=1 ˜ −Φ √ = 1 − Φ(1.66) = 0.0485
55 30
1
First we restate D : number of balls drawn to get two of the same colour. We draw
one ball which is either red or black. Having drawn a ball of some colour the number
of draws to get one of the same colour is geometrically distributed with probability 21 .
Thus D = X + 1 where X is geometrically distributed with p = 12 .
Question a)
P (D = i) = p(1 − p)i−2 , p = 2, 3, . . .
Question b)
1
E(D) = E(X + 1) = E(X) + 1 = +1=3
p
from page 212 or 476,482.
Question c)
1−p √
V (D) = V (X + 1) = V (X) = = 2, SD(D) = 2
p2
from page 213 or 476,482.
1
We define the random variable N as the number of throws to get heads. The pay back
value is N 2 , the expected win from the game can be expressed as
using the rule for the expectation of a linear function of a random variable p. 175
b. We could derive E(N 2 ) from the general rule for expectation of a function of a
random variable p. 175 t. However, it is more convenient to use the fact the N follows
a Geometric distribution and use the Computational Formula for the Variance p. 186.
2
2 2 1−p 1
E(N ) = V ar(N ) + (E(N )) = 2
+ =2+4=6
p p
The values for V ar(N ) and E(N ) can be found p. 476 in the distribution summary.
1
6.023 · 1023
µ(x) = x3 = 2.688 · 1019 x3
22.4 · 103
and p √
σ(x) = µ(x) = 5.1854 · 109 x x
Question b) √
5.1854 · 109 x x
≥ 0.01 → x ≤ 7.1914 · 10−6
2.688 · 1019 x3
1
We assume that the chocolate chips and mashmallows are randomly scattered in the
dough.
Question a) The number of chocoloate chips in one cubic inch is Poisson distributed
with parameter 2 according to our assumptions. The number of chocolate chips
in thre cubic inches is thus Poisson distributed with parameter 6. Let X denote
the number of chocolate chops in a three cubic inch cookie.
−6 36 36 · 6 216 · 6
P (X ≤ 4) = e 1+6+ + + = 115 · e−6 = 0.285
2 6 4·6
Question b) We have three Poisson variates Xi : total number of chocolate chips and
marshmallows in cookie i. According to our assumptions, X1 follows a Poisson
distribution with parameter 6, while X2 and X3 follow a Poisson distribution
with parameter 9. The complementary event is the event that we get two or
three cookies without chocoloate chips and marshmallows.
we are almost certain that we will get at most one cookie without goodies.
1
λn == pλn−1 + µ
λn == pλn−1 + µ
1
such that
f (x) = 30 · x2 (1 − x)2 0<x<1
This is an example of the Beta distribution page 327,328,478.
which we could have stated directly due to the symmetry of f (x) around 12 , or
from page 478.
Question c) We apply the computational formula for variances as restated page 261.
such that
30 1 1
V ar(X) = − =
105 4 28
which can be verified page 478.
3·3 1
SD(X3,3 )2 = =
(3 + 3)2 (3
+ 3 + 1) 28
1
We get
Z 2 Z 0 Z 2
1 1 1
P (−1 ≤ X ≤ 2) = dx = dx + dx
−1 2(1 + |x|)2 −1 2(1 − x)2 0 2(1 + x)2
x=0 x=2
1 1 1 1 1 1 7
= + − = − + − =
2(1 − x) x=−1 2(1 + x) x=0 2 4 2 6 12
h ix=∞
1
Question c) The distribution is symmetric so P (|X| > 1) = 2P (X > 1) = 2 − 2(1+x) =
x=1
1
2
.
R∞ 1
Question d) No. (the integral 0
x 2(1+x) 2 dx does not exist).
1
We first determine S4 and V ar(S4 ). From the distribution summary page 477 we have
E(S4 ) = 4 21 = 2 and due to the independence of the Xi ’s we have V ar(S4 ) = 4 12
1
= 31 .
(the result from the variance follows from the result page 249 for a sum of independent
random variables and the remarks page 261 which states the validity for continuous
distributions). We now have
3−2
P (S4 ≥ 3) = 1 − Φ q = 1 − Φ(1.73) = 1 − 0.9582 = 0.0418
1
3
1
Question b) We define the random variable Y as the length of an item which has
passed the quality inspection. The probability
Question b) The problem is similar to the determination of the half life of a radioac-
tive isotope Example 2. p.281-282. We repeat the derivation
ln 2
P (Ti ≤ t50% ) = 0.5 ⇔ e−λt50% = 0.5 t50% = = 6.93
λ
Question e) The sum of the lifetime of two components is Gamma distributed. From
p.286 (Right tail probability) we get
Question a) The time between two calls in a Poisson process is exponentially dis-
tributed (page 289). Using the notation of page 289 with = 1 we get
P (W4 2) = 1 e 2
= 0:8647
Question b) The distribution of the time to the arrival of the fourth call is a Gamma
(4; ) distribution. We nd the probability using the result (2) on page 286
25 125 118
P (T4 5) = 1 e 5
1+5+ + =1 e 5
= 0:735
2 6 3
Question c)
4
E (T4 ) = =4
using (3) page 286.
1
Question b)
k k+1 λ k
k
−λ m −λ k+1 −m λ
−m
P (Tm = k) = P (T > )−P (T > )=e −e m = e 1−e
m m
λ
pm = e− m .
The mean is measured in m1 time units so we have to multiply with this fraction
to get an approximate value for E(T )
1 1 − pm
E(T )=
˜ = E(Tm ) =
m pm
λ λ λ
1 e− m 1 1− m +o m 1
= = λ λ
→ for m → inf ty
m 1 − e− mλ m1− 1− m +o m λ
1
The relation between the hazard rate λ(t) and the survival function G(t) is given by
(7) page 297 Rt
G(t) = e− 0 λ(u)du
Now inserting λ(u) = λαuα−1
Rt
λαuα−1 du α ]u=t α
G(t) = e− 0 = e−λ[u u=0 = e−λt
dG(t) α α
f (t) = − = −e−λt −λαtα−1 = λαtα−1 e−λt
dt
Finally from (6) page 297
α
λαtα−1 e−λt
λ(t) = = λαtα−1
e−λtα
1
First we introduce Y = g(U ) = U 2 and note that g() is strictly increasing on ]0, 1[. We
then apply the formula in the box on page 304. In our case we have
√ dy √
fX (x) = 1 for 0 < x < 1, y = g(x) = x2 , x= y, = 2x = 2 y
dx
Inserting in the formula
1
fY (y) = √ 0<y<1
2 y
dFU 2 (y) 1
fU 2 (y) = = √ ,0 < y < 1
dy 2 y
1
We have
tan (Φ) = y
and use the change of variable result page 304 to get
dtan (Φ)
= 1 + tan (Φ)2 = 1 + y 2
dΦ
Now inserting into the formula page 304 we get
1 1
fY (y) = , −∞ < y < ∞
π 1 + y2
Question b) Z Z
∞ ∞
2 −λy 2 2
2λy e dy = λy 2 e−λy dy
0 −∞
We note the similarity with the variance of an unbiased (zero mean) normal
variable.
Z ∞ Z ∞ r s 1 2 r Z ∞ 2
2 −λy 2 2 2π 2λ − 12 y1 π 2 1 1 − 21 y1
λy e dy = λ y 1 e
2λ dy = λ y √ e 2λ dy
−∞ −∞ 2π 2λ λ −∞ 2π √12λ
1
the integral is the expected value of Z 2 , where Z is normal 0, 2λ distributed.
1
Thus the value of the integral is 2λ Finally we get
√ √
λπE(Z 2 ) = λπV ar(Z)
E(Y ) =
r
√ 1 1 π
= λπ == = 0.51 with λ = 3
2λ 2 λ
Question c) We apply the inverse distribution function method suggested page 320-
323. Thus
1
U = 1 − e−λX ⇒ X = − ln (1 − U )
λ
Now 1 − U and U are identically distributed such that we can generate an expo-
nential X with X = − λ1 ln (U ). To generate a Weibull (α = 2) distributed Y we
q
take the square root of X, thus Y = − λ1 ln (1 − U ).
1
We let Xi denote the lifetime of the i’th component, and S denote the lifetime of the
system.
Question b) In this case we have S = min(X1 , X2 ) and we apply the result for the
minimum of random variables page 317. The special case of two exponentials is
treated in example 3 page 317
P (S ≤ t) = 1 − e−(λ1 +λ2 )t
Question c) From the system design we deduce S = max (min (X1 , X2 ), min (X3 , X4 ))
such that
P (S ≤ t) = 1 − e−(λ1 +λ2 )t 1 − e−(λ3 +λ4 )t
Question b)
Question c)
Question d)
1 − (1 − x)n − y n + (y − x)n
Question e)
n
xk (1 − y)n−k
k
Question f )
k < x, n − k − 1 > y
one in between
1
Question b) From the boxed result at the bottom of page 327 we have that (X (k) has
beta(k, n − k + 1) distribution. Substituting r = k and s = n − k + 1 we get
X
r+s−1
r+s−1
P (X(k) ≤ x) = xi (1 − x)s+r−i−1
i
i=r
X s−1
1 1 s−1
f (x) = xr−1 (1 − x)s−1 = xr−1 (−x)i
B(r, s) B(r, s) i
i=0
Now
Z x Z x X s−1
1 s−1
P (X(k) ≤ x) = f (x)dx = ur−1 (−u)i du
0 0 B(r, s) i
i=0
s−1 Z s−1
1 X x
s−1 r+i−1 xr X s−1 (−x)i
= (−u) du =
B(r, s) i=0 0 i B(r, s) i=0 i r+i
as was to be proved.
1
Question b) We see that the probability can be rewritten This is example 2 page 343
with different values. We get
13 14 9
1− − =
24 25 40
Question c)
Question a)
P (U > 0.9) = 1 − P (U ≤ 0.9) = 0.9
Question b) The question is Example 3 page 343 the probability of a meeting with
different parameters. Denoting U1 and U2 respectively as the rank of the two
students
P (|U1 − U2 | > 0.1) = 0.92 = 0.81
1
We denote the radius of the circle by ρ. The are of the circle is πρ 2 . If a chosen point
is within radius r it has to be within the circle of radius r with area πr 2 . We find the
probability as the fraction of these two areas
r2
FR (r) = P (R1 ≤ r) =
ρ2
We recognize this as a gamma density (1) page 286 with λ = 1 and r = 4 thus
c = 81
dg(y) Z
13
Question b) With Z = g(Y ) = 4Y 3 , = 12y 2 , Y = , using the boxed
dy 4
result page 304 we get
1
z 3
y 3 −y 1 4
1
−( z4 ) 3
fZ (z) = e = e
6 12y 2 72
Question b)
E(XY ) = E(X)E(Y ) = 0.5
by the independe of X and Y .
Question c)
Question d)
E X 2 e2Y = E(X 2 )E e2Y
We recall the general formula for E(g(Y )) from page 263 or 332
Z
E(g(Y )) = g(y)f (y)dy
y
= P (X ≤ b, Y ≤ d) − P (X ≤ b, Y ≤ c) − (P (X ≤ a, Y ≤ d) − P (X ≤ a, Y ≤ c))
= P (X ≤ b, Y ≤ d) − P (X ≤ b, Y ≤ c) − P (X ≤ a, Y ≤ d) + P (X ≤ a, Y ≤ c)
= F (b, d) − F (b, c) − F (a, d) + F (a, c)
This relation can also be derived from geometric considerations.
Question b) Z Z
x y
F (x, y) = f (u, v)dudv
−∞ −∞
Question c)
d2 F (x, y)
f (x, y) =
dxdy
from the fundamental theorem of calculus.
d2 F (x, y)
= n(n − 1)(y − x)n−2
dxdy
1
Φ(1) − Φ(−1)
Question d)
1 −1
P (1 > max (X, Y ) − min (X, Y ) = P (1 > |X − Y |) = Φ √ −Φ √
2 2
1
Inserting in the boxed formula page 304 and use the many to one extension.
1 y
fY (y) = √ e− 2 0<y<∞
2πy
We recognize the gamma density with scale parameter λ = 12 and shape parameter
r = 21 from the distribution summary page 481. By a slight reformulation we have
y
1 −1
1 2 2 y
fY (y) = √ e− 2
2 π
and we deduce have
1 √
Γ = π
2
Question b) The formula is valid for n = 1. Assuming the formula valid for odd n
we get
n+2 n
Gamma =Γ +1
2 2
The recursive formula for the gamma-function page 191 tells us that Γ(r + 1) =
rΓ(r) and we derive
√
n+2 n π(n − 1)!
Gamma =
2 2 2n−1 n−1
2
!
2
n−1
n Y 1 √
Γ = i− π
2 i=1
2
Question d) From the additivity of the gamma distribution, which we can prove
directly
n
Question f ) The mean of a gamma (r, λ) distribution is λr , thus χn has mean 2
1 = n.
2
r
The variance of a gamma (r, λ) distribution is λ2
, thus the variance of χn is
n
1 = 2n. Skewness bla bla bla
2
4
1
Question a) The distribution of the total waiting time Z is found using the density
convolution formula page 372 for independent variables.
Z t Z t
−αu −β(t−u) −βt αβ
f (t) = αe βe du = αβe eu(β−α) du = e−αt − e−βt
0 0 β−α
Question b)
1 1
E(Z) = E(X1 ) + E(X2 ) = +
α β
See e.g. page 480 for the means E(Xi ) for the exponential variables .
Question c)
1 − e−2λt0.9 (1 + 2λt0.9 ) = 0.9
e−2λt0.9 (1 + 2λt0.9 ) = 0.1
1
Pn Pn
a sum of i=1 ri exponential(λ) random variables. The sum is gamma( i=1 ri , λ)
distributed.
1
P (X1 = x1 , X1 + X2 = n)
P (X1 = x1 |X1 + X2 = n) =
P (X1 + X2 = n)
Question b) Let Xi denote the number of eggs laid by insect i. The probability
in
1
question is P (X1 ≥ 90) = P (X2 ≤ 60). Now Xi ∈ binomial 150, 2 . With the
normal approximation to the binomial distribution page 99 to get
!
60 + 12 − 150 · 12 −29
P (X2 ≤ 60) = Φ 1
√ =Φ √ = Φ(−2.37) = 0.0089
2
150 150
1
P
Now realising that P (N1 = n1 , N2 = n2 , . . . Nm = nm ∩ m PN
i=1 i = n) = P (N1 =
m
n1 , N2 = n2 , . . . Nm = nm ) and using
Pm the fact that N = i=1 Ni has Poisson
distribution with parameter λ = i=1 λi we get
m
Qm n
λi i −λi
X i=1 ni ! e
P (N1 = n1 , N2 = n2 , . . . Nm = nm | Ni = n) = Pm
ni
λ
P i=1
i=1 ( m
e−λ
i=1 i )!
n
Pm
such that with n = i=1 ni
m
X n1 n 2 n m
n! λ1 λ1 λm
P (N1 = n1 , N2 = n2 , . . . Nm = nm | Ni = n) = ···
i=1
n1 !n2 ! · · · nm ! λ λ λ
λi
a multinomial distribution (page 155) with probabilities p i = λ
.
Question b) Using
m
X
P (N1 = n1 , N2 = n2 , . . . Nm = nm ) = P (N = n)P (N1 = n1 , N2 = n2 , . . . Nm = nm | Ni = n)
i=1
By definition
X X X XX
V ar(Y ) = (y−E(Y ))2 f (y) = (y−E(Y ))2 f (x, y) = (y−E(Y ))2 f (x, y)
y y x x y
We now apply the crucial idea of adding 0 in the form of E(Y |x) − E(Y |x) inside the
brackets. XX
V ar(Y ) = (y − E(Y |x) + E(Y |x) − E(Y ))2 f (x, y)
x y
f (x)
Next we multiply with one in the form of f (x)
XX f (x, y)
V ar(Y ) = (y − E(Y |x) + E(Y |x) − E(Y ))2 f (x)
x y
f (x)
f (x,y)
By definition fY (y|x) = f (x)
thus
" #
X X
V ar(Y ) = (y − E(Y |x) + E(Y |x) − E(Y ))2 fY (y|x) f (x)
x y
the inner part of the first term is V ar(Y |X = x) while the inner part of the second
term is constant. Thus
X X
V ar(Y ) = V ar(Y |X = x)f (x) + (E(Y |x) − E(Y ))2 f (x)
x x
We note that Y for given X = x is uniformly distributed, on 1 + x for −1 < x < 0 and
on 1 − x for 0 < x < 1. Thus
y
F (y|x) = P (Y ≤ y|X = x) = , 0 < y < 1 − |x|
1 − |x|
Question a) We have P Y ≥ 21 |X = x = 1 − F 12 |x
Question b) We have P Y ≤ 21 |X = x = F 12 |x
We have immediately
n
Y Pn Pn
P (X1 = x1 , X2 = x2 , . . . , Xn = xn ) = pXi (1 − p)1−Xi = p i=1 Xi
(1 − p)n− i=1 Xi
i=1
f (p; X1 = x1 , X2 = x2 , . . . , Xn = xn )
f (p|X1 = x1 , X2 = x2 , . . . , Xn = xn ) =
f (X1 = x1 , X2 = x2 , . . . , Xn = xn )
We note that if the prior density of p f (p) is a beta(r, s) distribution, then the
posterior distribution is a beta(r + Sn , s + n − Sn ) distribution.
1
Question b) We have
1 1 1
P Y > |X| > = 1 6= P Y >
4 2 4
X2 X3 X2 + X 3 X2 − X 3 Probability
1
0 0 0 0 3
1
0 1 1 -1 6
1
1 0 1 1 3
1
1 1 2 0 6
X2 + X 3 / X 2 − X 3 -1 0 1
1
0 0 3
0
1 1
1 6
0 3
1
2 0 6
0
This is the sum of two independent normal variables which itself is N ormal(0, 2 2 +
√ 2
3 ) distributed. Thus
3
P (X + 2Y ≤ 3) = Φ √ = Φ(1.13) = 0.8708
7
1
Question c)
√ √
P (U 2 + V 2 < 1) = P (3X 2 + Y 2 + 2 3XY + X 2 + 3Y 2 − 2 3XY < 1)
1 1
2
=P X +Y < 2
= 1 − e− 8 = 0.118
4
where we have used X 2 + Y 2 ∈ exponential(0.5) in the last equality (page 360,
364-366, 485).
Question d) √
X =v+ 3Y ∈ normal(v, 3)
1
Question a)
P (E2 ) = P (A2 ) + P (B2 ) + P (C2 ) + P (D2 ) = p2a + p2b + p2c + p2d = 0.3816
(i1 + i2 + i3 + i4 )! i1 i2 i3 i4
P (Ai1 ∩ Bi2 ∩ Ci3 ∩ Di4 ) = pa pb pc pd
i1 !i2 !i3 !i4 !
with i1 + i2 + i3 + i4 = 4, in our case. We have to sum over the appropriate values
of (i1 , i2 , i3 , i4 ).
It is doable but much more cumbersome to use basic rules. We get
p(1) = P (E1 ) = P (A4 ) + P (B4 ) + P (C4 ) + P (D4 ) = p4a + p4b + p4c + p4d = 0.0687
p(4) = P (E4 ) = P (A1 ∩ B1 ∩ C1 ∩ D1 ) = 24pa pb pc pd = 0.0177
To calculate p(3) = P (E3 ) we use the law of averaged conditional probabilities
4
X
p(3) = P (E3 ) = P (E3 |Ai )P (Ai ).
i=0
We immediately have
P (E3 |A4 ) = P (E3 |A3 ) = 0
2
further
4pb pc pd (pb + pc + pd
P (E3 |A0 ) = P (B2 ∩C1 ∩D1 |A0 )+P (B1 ∩C2 ∩D1 |A0 )+P (B1 ∩C1 ∩D2 |A0 ) =
(1 − pa )4
with
3pc pd (pc + pd )
P (E3|A1 ∩ B0 ) =
(1 − pa − pb )3
p2c + p2d
P (E3|A1 ∩ B1 ) =
(1 − pa − pb )2
pc + pd
P (E3|A1 ∩ B2 ) =
1 − pa − pb
P (E3|A1 ∩ B3 ) = 0
and we get
3
3pc pd (pc + pd ) 1 − pa − pb p2c + p2d
P (E3|A1) = +
(1 − pa − pb )3 1 − pa (1 − pa − pb )2
1
Define the events Bi that box i is chosen, and the event G that a gold coin is found.
We have
1
P (G|B1) = 1, P (G|B2) = 0, P (G|B3) =
2
We want to find P (B1|G). The probability is found using Baye’s rule (p.49)
P (G|B1)P (B1) 2
P (B1|G) = =
P (G|B1)P (B1) + P (G|B2)P (B2) + P (G|B3)P (B3) 3
1
gihQj1hkmlZh
npo1hQh&qrhlsn,tNuwv`npomxynzh&{ZxF|n,}~DvB1n,txF+hKnp,xFl1t,<n+nphj
|r+,h|np}~
xyl1j:npo1h
h&qrhlnK
npo1xynPxr+jtznppxyl1t,<n+nphj|r,+h|n,}~F
<%cÊ
²¡³ ³
ËÌ!Ë Á ËZÌ Ë
a´B¥¼ËZÌ ÂrÂFÍÎ Ï ¿ Î Ï ¥ÑËÌ!ÂrÂFÂrÂFÀsÒ
ËÌ!Âr ± ËZÌ ÂrÂÐÐ
1
P (A3 ∩ A) 1
P (A3|A) = =
P (A) 1 + 3q + 6q 2
Question d)
3
8
Question e) Pitman suggests no, which is reasonable. However, the way to assess
whether we can assume independence or not would be to analyze the distribution
of the number of sets played in a large number of matches.
1
Question b) Throw the coin twice, repeat until you get one head and one tail. Then
use HT or T H as the two possibilities.
1
Question b)
1
e−µp = e− 2 = 0.6065
1
Question b)
1 1 1 10
E(3Y1 +2Y2 ) = E(3Y1 )+E(2Y2 ) = 3E(Y1 )+2E(Y2 ) = 5E(Y1 ) = 5 0 · + 1 · + 2 · =
2 3 6 3
The first equality is true due to the addition rule for expectations (page 181), the
second equality is true due to the result for linear functions of random variables
page 175 b., the third equality is true since Y1 and Y2 has the same distribution,
and the fourth equality is obtained from the definition of the mean see page 181.
Question c)
0 for X ≤ 3
f (x) = 1 for 4 ≤ X ≤ 5
2 for X = 6
or something similar.
1
Question c)
n k!(n − k)! 1
=
k (n + 1)! n+1
b
P (X1 = 1) =
b+w
P (Xn+1 = 1) = P (Xn+1 = 1|X1 = 1)P (X1 = 1)+P (Xn+1 = 1|X1 = 0)P (X1 = 0) = P (Xn+1 = 1|X1 =
b+d b b w b
P (Xn+1 = 1) = + =
b+w+db+w b+w+db+w b+w
2
Question f )
using Bayes rule, or from the exchangeability. From the exchangeability we also
have
b+d
P (X10 = 1|X5 = 1) = P (X2 = 1|X1 = 1) =
b+w+d
1
However, this is a power series in z that is absolutely convergent for |z| ≤ 1 and
thus defines a C ∞ function of z for |z| < 1.
Question b) The more elegant and maybe more abstract proof is
GX+Y (z) = E z X+Y = E z X z Y
From the independence of X and Y we get (page 177)
GX+Y (z) == E z X E z Y = GX (z)GY (z)
The more crude analytic proof goes as follows
∞ ∞ k
!
X X X
X+Y
GX+Y (z) = E z = z k P (X+Y = k) = zk P (X = i, Y = k − i)
k=0 k=0 i=0
The interchange of the sums are justified since all terms are positive. The rear-
rangement is a commonly used tool in analytic derivations in probability. It is
quite instructive to draw a small diagram to verify the limits of the sums. We
now make further rearrangements
∞ X
X ∞
GX+Y (z) = z k P (X = i)P (Y = k − i)
i=0 k=i
∞
X ∞
X ∞
X ∞
X
= z i P (X = i) z k−i P (Y = k − i) = z i P (X = i) z m P (Y = m)
i=0 k=i i=0 m=0
Now using the general result for Xi with binomial distribution b(ni , p) we get
i.e. that the sum of independent binomially distributed random variables is itself
binomially distributed provided equality of the p i ’s.
i=1
The result proofs that the sum of independent Poisson random variables is itself
Poisson.
Question e) n
zp zp
GX (z) = G Sn =
1 − z(1 − p) 1 − z(1 − p)
Question f )
Pni=1 ri
zp
G Sn =
1 − z(1 − p)
1
Question c)
Z ∞ Z ∞
β −βt
P (|X| > y) = 2P (X > y) = 2 e dt = βe−βt dt = e−βy
y 2 y
We introduce the random variables Nloc (t) and Ndis (t) as the number of local respec-
tively long distance calls arriving within time t (where t is given in minutes).
Question a)
due to the independence of the Poisson processes. The variables N loc (t) and
Ndis (t) has Poisson distributions (page 289) such that
3 5
(λloc · 1)5 −λ ·1 (λdis · 1)3 −λ ·1 λdis λloc −λ −λ
P (Nloc (1) = 5, Ndis (1) = 3) = e loc · e dis = e loc dis
5! 3! 5!3!
Question b) The sum of two indpendent Poisson random variables is Poisson dis-
tributed (boxed result page 226), leading to
Question c) We now introduce the random variables Siloc and Sidis as the time of
the i’th local and long distance call respectively. These random variables are
Gamma distributed according to the box on the top of page 286 or to 4. page
289 The probability in question can be expressed as The waiting time to the first
long distance in terms of calls are geometrically distributed
10
10 λloc
P (X > 10) = (1 − pdis ) =
λloc + λdis
1
P (Y ≤ y) = P (min (R1 , R2 ) ≤ y) = 1−P (R1 > y, R2 > y) = 1−P (R1 > y)(R2 > y)
1 2 1 2 2
= 1 − e− 2 y e− 2 y = 1 − e−y
a new Weibull distribution with α = 2 and λ = 1. If we did not recognize the
distribution as a Weibull we would derive the survival function of the R i ’s by
Z ∞
1 2 1 2
P (Ri > x) = ue− 2 u du = e− 2 x
x
We find the density using (5) page 297 or directly using E4.3.4 (i)
2
fY (y) = 2ye−y
Question b) This is a special case of E4.4.9 a). We can re-derive this result using the
dg(y)
change of variable formula page 304. With Z = g(Y ) = Y 2 we get = 2y.
dy
Inserting we get
2 1
fZ (z) = 2ye−y = e−z
2y
an exponential(1) distribution.
Question c) We have E(Z) = 1 (see e.g. the mean of an exponential variable page
279 or the distribution summary page 477 or page 480).
1
Question a)
where we have used standard rules for mean and variance see eg. page 249, and
the result page 279 for the variance of the exponential distribution.
Question b) The standard uniform density f (y) = 1 for 0 < y < 1, 0 elsewhere.
Question c)
1 1
2
2
−0 1 2
−0 1
E(Y ) = = , V ar(Y ) = =
2 4 12 48
1
2et t
E(Wt ) = 2 e −1
2
t
where we have used that X has Gamma (4,2) density, and apply the formula for
E(g(X, Y )) page 349.
where E(X 2 ) = V ar(X) + (E(X))2 = 5, see eg. page 481. Next we derive
2 e2t t
E etY = e −1
t
and apply the computational formula for the variance page 261
s
e2t t 2et t 2
SD(Wt ) = 5 (e − 1) − 2 e2 − 1 =
t t
1
The joint density of X and Y is the product of the marginal densities since X and Y
are independent (page 349). We calculate the denominator using the formula for the
probability of a set B page 349
Z 1Z 1 Z 1
2 1 2
P (Y ≥ X ) = 1 · 1 · dydx = (1 − x2 )dx = 1 − =
0 x2 0 3 3
and the numerator
1 2 2 1 2
P Y ≥ ∩ Y ≥ X = P (Y ≥ X ) − P Y < ∩ Y ≥ X
2 2
Question c) Consider
Z ∞
λ1
= λ1 e−λ1 t1 e−λ2 t1 dt1 = e−(λ1 +λ2 )t
t λ1 + λ2
By inserting back we finally get
Question c) Similarly
1
E(XY ) = E(E(XY |Y )) = E(Y E(X|Y )) = E(Y 2 ) = V ar(Y ) + (E(Y ))2 =
2
We have E(X 2 ) = E(E(X 2 |Y )) = E(2Y 2 ) = 1. Thus V ar(X) = SD(X)2 =
1 1 √
−
1 − 14 = 43 and SD(Y ) = 21 . Finally Corr(X, Y ) = 2√3 14 = 33
2 2