Escolar Documentos
Profissional Documentos
Cultura Documentos
x<y
F (x) F (y)
ii)
limx F (x) = 0
limx+ F (x) = 1
Note: The set of discountinuities of F is at most countable (this fact is true for
increasing functions).
1 x1
3e
x
x+1
x<1
is a distribution funcx1
tion.
It is known that
Theorem: If F : R [0, 1] is a distribution function, then there exists a random
variable X such that for all x R we have
F (x) = P (X x)
We denote this F by FX .
Definition: A sequence {Xn } of random variables is said to converge in distribution (or, in law) to a r.v. X provided that
lim FXn (x) = FX (x)
(1)
d
Note: When Xn s and X are discrete with values from {u1 , u2 , ...}, then the
above condition is equivalent to having
k = 1, 2, ...
lim P (Xn = uk ) = P (X = uk )
Then
FXn (u) =
nx2
n
e 2 dx = (u n)
2
So,
limn FXn (u)
()
(0)
=
()
1
2
u<0
u=0
u>0
u<0
u=0
u>0
t=0
x<1
0
So
lim Fn (x) =
n
1
x1
d
Note: We recall from Calculus that for every real number c we have limx+ 1 +
x
ec . To see this, put y = 1 + xc , and note that
limx+ ln(y) = limx+ x ln 1 + xc
=
limx+
limz0+
= limz0+
c
ln(1+ x
)
1
x
ln(1+cz)
z
c
1+cz
= c
Then,
lim
x+
1+
c x
= lim y = lim eln(y) = elimx+ ln(y) = ec
x+
x+
x
3
c x
x
|rn |
So, if we let sn = sup{|rn | , |rn+1 | , |rn+2 | , ...}, then
n n0
0 sn
This shows that lim sn = 0. So, there exists some n1 such that
n n1
|sn | 1
k < n & n n1
Therefore,
n
k
P
n1
= k=0
n
k
rn n
n
1+
c n
n
1+
c k
n
rn nk
n
1+
c k
n
rn nk
n
1+
Pn1
k=0
n
k
1 + c k rn nk
Pn1
k=0
n
k
1 + c k |sn |nk
n
Pn1
n
k
1 + c k |sn |
n
1 nk
n
Pn
n
k
1 + c k |sn |
n
1 nk
n
n
k
1 + c k
1 nk
n
n
k
k=0
k=0
= |sn |
Pn
|sn |
Pn
k=0
k=0
= |sn | 1 +
|c|+1
n
1+
c n
n
|c|
n
k
n
1 nk
n
1 nk
n
(n n0 )
t (, ), then Xn X.
Continuity Theorem (based on characteristic functions): If X, X1 , X2 ,
d
... are random variables such that limn Xn (t) = X (t), then Xn X, and
vice versa.
Example (Poisson approximation to the binomial distribution): Let
Xn Binomial(n, pn = n ), or more generally, assume lim npn = . Show that
5
Xn X, where X Poisson().
Proof: One way of doing this exercise is by showing that
lim P (Xn = k) = P (X = k)
k = 0, 1, 2, ...
h
in
lim 1 + pn (eit 1)
h
lim 1 +
= e(e
it
npn (eit 1)
n
1)
in
= X (t)
d
E|X k | = E(|X|k )
=
|x|k f (x)dx
|x|k f (x)dx +
|x|n f (x)dx +
Ac
Ac
|x|n f (x)dx +
= E(|X|n ) + 1
= E(|X n |) + 1 <
We recall the following fact from Calculus:
|x|k f (x)dx
1 f (x)dx
f (x)dx
Proposition: Let h(x, t) be a function defined for all t in an interval (a, b).
Suppose that an integrable function g(x) exists such that
d
h(x, t) g(x)
t (a, b) x
dt
then
d
dt
Z
h(x, t)dx =
d
h(x, t)dx
dt
t (a, b)
eitx f (x)dx
(t) =
Especially,
(k) (0) = ik E(X k )
Proof:
d itx
e f (x) = (ix)eitx f (x) = |x|f (x)
dt
But, from the preceeding Lemma weR know that the moments E(X), E(X 2 ),
... , E(X n ) all exist, and therefore |x|f (x) = E|X| < , so, using the
preceeding Proposition, we can write
R itx
d
e f (x)dx
0 (t) = dt
d
dt
eitx f (x) dx
(ix)eitx f (x)dx
So,
Z
(ix)eitx f (x)dx
(t) =
(2)
Next Step:
d
(ix)eitx f (x) = (ix)2 eitx f (x) = |x2 |f (x)
dt
and we know that
so
00
(t) =
(3)
Continuing this way, we get the equality (1) for all k = 1, 2, ..., n.
Remark: We have seen smewhere that if the MGF of a random variable X exists
on some neighbourhood (h, h) of the origin, then both MX (t) and X (t) have
power series expansions on that neighbourhood, therefore have derivatives of all
order in that neighbourhood with
(n) (0) = in E(X n )
M (n) (0) = E(X n )
f 0 (x0 )
f 00 (x0 )
f (n) (x0 )
(xx0 )+
(xx0 )2 + +
(xx0 )n +R(x)
1!
2!
n!
where limxx0
R(x)
(xx0 )n
= 0.
Definition: A sequence {Xn } of random variables is called an i.i.d (independent identically distributed) if the xn s share the same distribution and that
the Xn s are independent.
A simple observation is the following, which shows that applying a transformation on an i.i.d. will result in an i.i.d:
Proposition: If {Xn } is an i.i.d., then for every continuous function g : R R,
the sequence {g(Xn )} is an i.i.d. too.
Proof: If Xm and Xn are any two members of the sequence and if O1 = [a, b]
and O2 = [c, d] are two intervals in the real line, then
P (g(Xm ) O1 , g(Xn ) O2 ) = P (Xm g 1 (O1 ) , Xn g 1 (O2 ))
= P (Xm g 1 (O1 ))P (Xn g 1 (O2 ))
= P (g(Xm ) O1 )P (g(Xn ) O2 )
8
This shows that g(Xn )s are independent. Further, for every interval O in the
real line:
P (g(Xm ) O) = P (Xm g 1 (O))
= P (Xn g 1 (O))
= P (g(Xn ) O)
which shows that the Xn s enjoy the same distribution.
Central Limit Theorem: Let {X1 , X2 , ...} be an i.i.d. with finite mean
and finite variance 2 (equivalently, the first and second moments are finite).
n = X1 ++Xn be the sequence of sample means. Then the sequence
Let X
n
n )
n(X
t2
Proof: Let n be the C.F. of the r.v. n(Xn ) . Let (t) = e 2 be the C.F.
f the r.v. N (0, 1). For every t we will show n (t) (t); then in light of the
convergence theorem this will do.
set
Xi
i = 1, ..., n
Zi =
Then form the last proposition, the sequence {Z1 , ..., Zn , ...} is an i.i.d. since it
is found by applying a transformation on the i.i.d. {X1 , ..., Xn }. Since the Zi s
have the same distribution, their C.F.s are
identical; let be their common
Pn
n(Xn )
characteristic function. Note further that
= 1n i=1 Zi as
Pn
n )
)
n n
n(X
n(X
nX
n
=
=
=
n
n
Xi n
1 X
=
n
n
i=1
Xi
Then
n (t) = 1
Pn
Z
i=1 i
= Pn
i=1
= Z1
=
Zi
t
n
(t)
t
n
Zn
n
tn
t
n
(1)
Since the common second moment of Xi s is finite, the common second moment
of Zi s is finite and therefore the second derivative 00 (t) exists for all t. So then
we can apply the Taylors theorem to write:
(s) = (0) + 0 (0)s +
1 00
(0)s2 + R(s)
2!
1 X
=
Zi
n i=1
R(s)
s2
where, lims0
t
n
= 0. By changing s to
t
1
= (0) + 0 (0) +
2!
n
we will have
2
00
(0) + R
(2)
But
(0) = 1
0 (0) = i E(Z1 ) = 0
00 (0) = i2 E(Z12 ) = 1
So, the identity (2) reduces to
t
t2
t
=1
+R
2n
n
n
By putting this back into (1):
n
n (t) =
1
1+
=
t2
2n
+R
2
t2 +nR
t
n
t
n
on
n
(3)
But,
nR
=n
2 R t
t
R
t
n
n
2
=
t
2
2 0
n
t
t
lim n (t) = e 2
as desired.
Definition: Let {Yn } be a sequence of random variables. Suppose that sed
n
N (0, 1). Then we say that {Yn }
quences {an } and {bn } exis such that Ynba
n
is asymptotically normal N (an , b2n ).
(1)
1
n2
{V ar(X1 ) + + V ar(Xn )} =
10
2
n
std(Xn )
(2)
equivalently
Sn n d
N (0, 1)
(3)
n
Sn
n ,
so we actually
= E(X1 ) + + E(Xn ) = n
= V ar(X1 ) + + V ar(Xn ) = n 2
Note: If {Yn } is a sequence that Yn N (0, 1), then for every two real numbers
a < b we have
P (a < Yn b) = P (Yn b) P (Yn a) = FYn (b) FYn (a) (b) (a)
so we use (b) (a) to approximate P (a < Yn b) for large n. When Yn has
a continuous distribution, the probabilities P (Yn = a) and P (Yn = b) are zero,
therefore we may use (b) (a) to approximate all of
P (a Yn b)
P (a Yn < b)
P (a < Yn < b)
But, when the r.v.s Yn are integer-valued, then we apply a continuity correction.
So here is the discussion:
11
= FX (k + 0.5) FX (k 0.5)
FY (k + 0.5) FY (k 0.5)
= P (k 0.5 < Y < k + 0.5)
= P
k0.5
k+0.5
<
<
k+0.5
k0.5
So,
P (X = k)
k + 0.5
k 0.5
Pn
k=m
i
h
k+0.5
k0.5
n+0.5
m0.5
where
X N (, 2 ) Y
This is the continuity correction when both end points are inclusive. But id
the left-end-point is exclusive, then we write
P (m < X n) = P (m+1 X n) P ((m+1)0.5 < Y < n+0.5) = P (m+0.5 < Y < n+0.5)
0.52
3.7
4.52
3.7
<
Y 2
3.7
<
4.52
3.7
0.52
3.7
= 0.6854
Example: A die is rolled thirty times. Find the (approximate) probability that
the total score will be between 90 and 120 inclusive.
Solution: Let Xi be the outcome of the i-th rolling. We want P (90 S30
120). In fact, the mean value of each rolling Xi is
1 + 2 + + 6
= 3.5
6
and the variance is
(1 3.5)2 + (2 3.5)2 + + (6 3.5)2
= 2.9167
6
So then
E(S30 ) =
V ar(S30 ) =
(30)(3.5) = 105
(30)(2.9167) = 87.5
13
89.5105
87.5
<
120.5105
87.5
Y105
87.5
<
120.5105
87.5
89.5105
87.5
So,
X N (6, 4.2) Y
Then
P (X = 7) = P (7 X 7) P (6.5 < Y < 7.5) = P
6.56
4.2
<
Y 6
4.2
<
7.56
4.2
6.56
( 7.56
4.2
4.2
Example: A random vector (X1 , ..., X100 ) (i.e. an i.i.d.) is picked from a
population with distribution uniform(1, 1). Find the approximate probability
that the square of the distance of the vector from the origin is less than or equal
to 40.
P100
Solution: We are looking for P ( i=1 Xi2 40). We have
i1
R1
= 13
E(Xi2 ) = 1 x2 12 dx = 16 x3
1
E(Xi4 )
V ar(Xi2 )
R1
x4 21 dx =
1
1 5
10 x
i1
1
2
= E(Xi4 ) E(Xi2 ) =
14
1
5
1
5
1
9
4
45
So then:
(100)( 13 ) =
P
E( Xi2 ) =
P
V ar( Xi2 )
100
3
4
V ar(Xi2 ) = (100)( 45
)=
Based on CLT:
100
80
9
80
Y
3
9
Since the Xi s are continuously distributed, we dont apply the continuity correction. So,
100
P
Y 100
803 40
803
P ( Xi2 40) P (Y 40) = P
X
Xi2 N
+
2
and
V ar(X) =
( )2
12
E(Xi )
V ar(Xi ) =
48 )
E(X
48 ) =
V ar(X
V ar(X1 )
48
1
144
1
1,
144
Y
P (0.8 < X
<
<
1
1
1
144
15
144
144