Professional Documents
Culture Documents
Contents
1.3 Exercises
2.2 Exercises
2.3 Exercises
2.4 Exercises
2.5 Exercises
2.6 Exercises
2.7 Exercises
10
3.1 Exercises
24
3.2 Exercises
24
3.6 Exercises
25
4.1 Exercises
34
4.3 Exercises
34
4.5. Exercises
36
5.5 Exercises
46
6.2 Exercises
52
6.3 Exercises
53
7.2 Exercises
60
1.3 Exercises
Exercise 1.3.1. Suppose that = {1, 2}, with P() = 0 and P{1, 2} = 1.
Suppose P{1} = 14 . Prove that P is countably additive if and only if P{2} = 34 .
Proof. (=) Assume P is countably additive. Since {1} {2} = , countable additivity implies P{1} + P{2} = P{1, 2}. Then, subtracting P{1} from both sides
of the equation we have P{2} = P{1, 2} P{1}. Finally, substituting in P{1} = 41
and P{1, 2} = 1 we obtain P{2} = 1 41 = 34 .
(=) Assume P{2} = 34 . In order to show that P is countably additive, we
do an exhaustive proof by considering all disjoint subsets of . All subsets A
are disjoint with , and since P() = 0, we have P(A) + P() = P(A) + 0 = P(A).
The only sets left which are disjoint are {1} and {2}. By assumption, we have
P{1} + P{2} = 41 + 34 = 1 = P{1, 2}. Therefore P is countably additive.
Exercise 1.3.2. Suppose = {1, 2, 3} and F is the collection of all subsets of .
Find (with proof) necessary and sufficient conditions on the real numbers x, y, and
z, such that there exists a countably additive probability measure P on F, with
x = P{1, 2}, y = P{2, 3}, and z = P{1, 3}.
Proof. In order for P to be a probability measure we must have P() = 1 and
thus P() = 0. All sets A F are disjoint with , and since P() = 0, we have
P(A) + P() = P(A) + 0 = P(A). We must also have P satisfy the following
conditions on disjoint sets:
P{1} + P{2} + P{3} = P{1, 2, 3} = 1
P{1} + P{2} = P{1, 2} = x
P{2} + P{3} = P{2, 3} = y
P{1} + P{3} = P{1, 3} = z
Substituting the last three equations into the first individually gives all combinations of disjoint sets in F. Adding the last three equations and substituting in the
first gives us x + y + z = 2P{1} + 2P{2} + 2P{3} = 2(P{1} + P{2} + P{3}) = 2.
Since A F, we must have P(A) [0, 1], and thus we have derived necessary and
sufficient conditions on x, y, and z, namely {x, y, z [0, 1]; x + y + z = 2}.
Exercise 1.3.3. Suppose that = N is the set of positive integers, and P is
defined for all A by P(A) = 0 if A is finite, and P(A) = 1 if A is infinite. Is
P finitely additive?
Exercise 1.3.5.
(a) In what step of the proof of Proposition 1.2.6 was (1.2.1) used?
(b) Give an example of a countably additive set function P, defined on all subsets
of [0,1], which satisfies (1.2.3) and (1.2.5), but not (1.2.1).
Proof.
(a) On page 4, (1.2.1) is used to show 1 = P((0, 1]).
(b) Let F be the collection of all subsets of [0, 1]. Consider P : F [0, 1] defined
by P(A) = 0. The set function P clearly satisfies (1.2.3) and (1.2.5), but fails
to satisfty (1.2.1). For example 0 = P([0, 1]) 6= 1 0 = 1.
2.2 Exercises
Exercise 2.2.3. Prove that the above collection J is a semialgebra of subsets of
, meaning that it contains and , it is closed under finite intersection, and the
complement of any element of J is equal to a finite disjoint union of elements of
J.
Proof. We have that = [0, 0) (or any empty interval), and = [0, 1]. Therefore
by definition, , J . Now let A, B J be intervals with endpoints a0 , a1 and
b0 , b1 respectively. Then A and B are intervals contained in [0,1] and A B is
either empty (and hence an empty interval) or an interval with endpoints max{a0 , b0 }
and min{a1 , b1 }. Thus we have A B J . The proof is similar for any finite number of elements A1 , A2 , A3 , J (simply induct on the previous argument) and
therefore J is closed under finite intersection.
Note C = = [0, 1] J and C = = [0, 0) J . Now, let A J be
an interval of the form [a, b). If a = 0, then AC = [b, 1] J and if a 6= 0,
then AC = [0, a) [b, 1] where [0, a), [b, 1] J and thus AC J . The proof for
intervals of the form (a, b), [a, b], (a, b] are similar. Therefore the collection J is a
semialgebra.
Exercise 2.2.5.
(a) Prove that B0 is an algebra (or, field ) of subsets of , meaning that it contains
and , and is closed under the formation of complements and of finite
unions and intersections.
(b) Prove that B0 is not a -algebra.
Proof.
(a) Since , J , we have that , B0 . Since the finite union of a finite union
of intervals is itself a finite union of intervals, we have that B0 is closed under
finite unions.
Similarly, a finite intersection of finite unions of intervals is also a finite union
of intervals. To see this let A, B B0 be finite
S unions
ofS intervals A =
SN A
SNB
NB
NB
i=1 Ai and B =
i=1 Bi . Then A B = A
i=1 Bi =
i=1 (A Bi ) =
SN B
SNA
SNB SNA
i=1
j=1 Aj Bi =
i=1 j=1 (Aj Bi ) and since J is a semialgebra,
we have Aj Bi J , and thus A B B0 . The proof is similar for any finite
intersection of elements A1 , A2 , A3 , B0 (simply induct on the previous
argument) and therefore B0 is closed under finite intersection.
Since J is a semialgebra, we have that if A J thenSAC is a finite disjoint
N
union of intervals and thus AC B0 . Thus if B = i=1 Ai where Ai J
TN
i=1
AC
i is a finite
(b) Note that n1 = [ n1 , n1 ] B0 for all n = 1, 2, 3, . . . . We will show that the set A =
{1, 12 , 13 , . . . }
/ B0 by contradiction, and thus prove that B0 is not a -algebra.
Assume A B0 is a finite union of (WLOG) nonempty disjoint intervals
SNA
Ai , since if Ai Aj 6= for some i, j, are intervals with endpoints
A=
i=1
/ A leading to a
contradiction.
2.3 Exercises
Exercise 2.3.16. Prove that the extension (, M, P ) is a probability measure
on M, and that P is an extension of P.
Proof. Let A M with P (A) = 0 and let B A. Assume E . By subadditivity we have 0 6 P (B E) 6 P (A E) 6 P (A) = 0, so that P (B E) = 0.
Therefore P (B E) + P (B C E) = P (B C E) 6 P (E). Thus (2.3.8) is
satisfied and B M.
2.4 Exercises
Exercise 2.4.3.
Sn
(a) Prove that if I1 , I2 , . . . , In is a finite collection of intervals, and if j=1 Ij I
Pn
for some interval I, then
j=1 P(Ij ) > P(I). [Hint: Imitate the proof of
Proposition 2.4.2.]
(b) Prove
S that if I1 , I2 , . . . is a countable collection
P of open intervals, and if
I
I
for
some
closed
interval
I,
then
j
j=1 P(Ij ) > P(I). [Hint: You
j=1
may use the Heine-Borel Theorem, which says that if a collection of open
intervals contain a closed interval, then some finite sub-collection of the open
intervals also contains the closed interval.]
(c) Verify (2.3.3),
Pcountable collection of intervals,
S i.e. prove that if I1 , I2 , . . . is any
and if j=1 Ij I for any interval I, then j=1 P(Ij ) > P(I). [Hint: Extend
the interval Ij by 2j at each end, and decrease I by at each end, while
making Ij open and I closed. Then use part (b).]
Proof.
(a) Let I1 ,S
. . . , In be intervals contained in [0, 1]. Let I be an interval such that
n
I j=1 Ij . If I = there is nothing to prove. Suppose I 6= . For
1 6 j 6 n, write aj for the left end-point of Ij , bj for the right end-point
of Ij ordered such that a1 6 a2 6 . . . 6 an , and a the left endpoint of
I and b the right endpoint of I. Then for each j = 1, . . . , nP
1, we have
n
(bj aj ) + (bj+1 aj+1 ) > max{bj , bj+1 } aj . Thus we have j=1 P(Ij ) =
Pn
j=1 (bj aj ) > max{b1 , . . . , bn } a1 > b a = P(I).
(b) Let I1 , I2 . . . be a countable
S collection of open intervals. Let I be a closed
interval such that I j=1 Ij . If I = there is nothing to prove. Suppose I 6=
. By the Heine-Borel
Theorem there exists a finiteP
subcollection Ii1 , . . . , Iin
Sn
n
such that I j=1 Iij . By part (a) we have that j=1 P(Iij ) > P(I). Let
Ps
Pn
s = max{i1 , . . . , in }, then j=1 P(Ij ) > j=1 P(Iij ) > P(I). Then for any
Pk
P
k > s we have j=1 P(Ij ) > P(I) and thus j=1 P(Ij ) > P(I)
(c) Let I1 , I2 . .S
. be a countable collection of intervals. Let I be an interval such
Exercise 2.4.5. Let A = {(, x]; x R}. Prove that (A) = B, i.e. that the
smallest -algebra of subsets of R which contains A is equal to the Borel -algebra
of subsets of R. [Hint: Does (A) include all intervals?]
Proof. Let J = {all intervals contained in R} so that the Borel -algebra on R is
B = (J ). Let A = {(, x]; x R}. Clearly A J and thus (A) B. We
will show (A) J and thus (A) B which will complete the proof. We will do
this case by case. Let a, b R such that a 6 b:
Case (, b]: (, b] (A) by definition.
S
Case (, b): (, b) = j=1 (, b 1j ] (A).
Case (a, ): (a, ) = (, a]C (A).
T
Case [a, ): [a, ) = j=1 (a 1j , ) (A).
6
Exercise 2.4.7.
(a) Prove that K, K C B, where B are the Borel subsets of [0, 1].
(b) Prove that K, K C M, where M is the -algebra of Theorem 2.4.4.
(c) Prove that K C B1 , where B1 is defined by (2.2.6).
(d) Prove that K
/ B1 .
(e) Prove that B1 is not a -algebra.
Proof.
(a) At each stage, n, of construction of the Cantor set, removing 2n1 open sets
corresponds to the intersection of the (n1)th stage set with the complement
of the 2n1 open sets. Thus K is the countable intersection of sets in B and
therefore K B. Since B is a -algebra, we have that K C B.
(b) Since M is a -algebra containing all intervals, we have B M. Therefore
K, K C M.
(c) In part (a) we showed that K is the countable intersection of intervals. Therefore, since the complement of an interval is the union of intervals, we have by
de Morgans laws K C is the countable union of intervals. Thus K C B1 .
S
(d) We prove by contradiction. Assume K = j=1 Ij of (possibly empty) intervals
Ij . Since (K) = 0, we have that (Ij ) = 0 for all j = 1, 2, . . . . Therefore Ij
is a singleton set or empty. Therefore K is the countable union of singleton
sets and empty sets and is therefore countable. This is a contradiction since
K is uncountable. Therefore K
/ B1 .
(e) We have that K C B1 by part (c) and K
/ B1 by part (d). Therefore B1 is
not a -algebra.
2.5 Exercises
Exercise 2.5.6. Suppose P satisfies (2.5.5) for finite disjoint collections {Dn }.
Suppose further that, whenever A1 , A2 , . . . are finite disjoint unions of elements of
7
T
J such that An+1 An and n=1 An = , we have limn P(An ) = 0 (where
P is extended to An by finite additivity). Prove
satisfies
S that PSalso
S(2.5.5) for
n
countable collections {Dn }. [Hint: Set An = ( i=1 Di )\( j=1 Dj ) = i=n+1 Di .]
S
Proof. Let D1 , D2 , J be disjoint with n Dn J .
S
S
S
Sn
D
D1C DnC
Set An = ( i=1 Di )\( j=1 Dj ) = i=n+1 Di . Since An =
j
j
and J is a semialgebra, then T
An can be written as the disjoint union of elements of
J . Clearly
A
A
and
An = which implies that
n+1
n
P(An ) = 0.
S limn
S
n=1
Sn
n
D
= P(An ) +
D
=
P(A
)
+
P
Now P
D
=
P
A
j
n
j
n
j=1 j
j
S
j=1
Pn
S
j=1 P(Dj ) for any given n. Thus P
j Dj = limn P
j Dj =
Pn
P
P
limn P(An ) + j=1 P(Dj ) = 0 + j=1 P(Dj ) = j=1 P(Dj ).
2.6 Exercises
Exercise 2.6.1.
(a) Verify that the above J is a semialgebra.
(b) Verify that the above J and P satisfy (2.5.5) for finite collections {Dn }. [Hint:
For a finite collection {Dn } J , there is k N such that the results of only
coins 1 through k are specified by any Dn . Partition into the corresponding
2k subsets.]
Proof.
(a) Let J = {Aa1 a2 ...an ; n N, a1 , a2 , . . . , an {0, 1}} {, }. Note , J
by definition. Now let A, B J . If A B = , then clearly A B J .
Suppose A B 6= , then A = Aa1 a2 ...anA and B = Bb1 b2 ...bnB for some
nA , nB N. Suppose (WLOG) nA 6 nB , then A B 6= imples that
ai = bi , for i = 1, . . . , nA . Therefore A B = B J . To show this
is true for any finite collection of elements, simply induct on the previous
argument. Let S = {c1 . . . cnA ; c1 , . . . , cnA {0, 1}}. Now partition into
S
the 2nA subsets = sS Cs . Therefore AC
a1 a2 ...anA = \Aa1 a2 ...anA =
S
nA
semialgebra.
S
(b) Let (WLOG) D1 , . . . , Dn J \{} be disjoint with j Dj J . For each
j, one can write Dj = Dd1j ...dmj j for some mj . Let k = maxj {mj }. Let
S = {a1 . . . ak ; a1 , . . . , ak {0, 1}}. Now partition into the 2k subsets =
S
Exercise 2.6.4. Verify that the above J is a semialgebra, and that , J with
P() = 0 and P() = 1.
Proof. Let (1 , F1 , P1 ) and (2 , F2 , P2 ) be probability triples and = 1 2 .
Let J = {AB; A F1 , B F2 }. Define P(AB) = P1 (A)P2 (B) for AB J .
We first show that , J with P() = 0 and P() = 1. Clearly = J .
Also for any A, B pair where at least one of A and B is the emptyset we have at least
one of P1 (A) and P2 (B) equal to 0. Therefore P() = P(A B) = P1 (A)P2 (B) =
0. Also = 1 2 J and since both P1 (1 ) = 1 and P2 (2 ) = 1 we have
P() = P(1 2 ) = P1 (1 )P2 (2 ) = 1.
2.7 Exercises
Exercise 2.7.1. Let = {1, 2, 3, 4}. Determine whether or not each of the
following is a -algebra.
(a) F1 = {, {1, 2}, {3, 4}, {1, 2, 3, 4}}
(b) F2 = {, {3}, {4}, {1, 2}, {3, 4}, {1, 2, 3}, {1, 2, 4}, {1, 2, 3, 4}}
(c) F3 = {, {1, 2}, {1, 3}, {1, 4}, {2, 3}, {2, 4}, {3, 4}, {1, 2, 3, 4}}
Proof.
(a) F1 is a -algebra. Since F1 , it is sufficient to show that F1 is closed under
complement and countable unions. Well C = {1, 2, 3, 4} F1 , {1, 2}C =
{3, 4} F1 , and therefore F1 is closed under complement. For each A F1 ,
we have A = A F1 and A {1, 2, 3, 4} = {1, 2, 3, 4} F1 . Also
{1, 2} {3, 4} = {1, 2, 3, 4} F1 , and any countable union of elements will
be equal to one of the previous. Thus F1 is closed under countable unions.
(b) F2 is a -algebra. Since F2 , it is sufficient to show that F2 is closed
under complement and countable unions. Well C = {1, 2, 3, 4} F2 , {3}C =
{1, 2, 4} F2 , {4}C = {1, 2, 3} F2 , and {1, 2}C = {3, 4} F2 , and therefore
F2 is closed under complement. For each A F2 , we have A = A F2
and A {1, 2, 3, 4} = {1, 2, 3, 4} F2 . The rest of the unions are clearly in
F2 , but are too tedious to list here.
(c) F3 is NOT a -algebra. {1, 2} {1, 3} = {1, 2, 3}
/ F3 , and therefore F3 is not
closed under countable union.
Exercise 2.7.2. Let = {1, 2, 3, 4}, and let J = {{1}, {2}}. Describe explicitly
the -algebra (J ) generated by J .
Proof. Let = {1, 2, 3, 4}, and let J = {{1}, {2}}. Any -algebra containing
J must contain and be closed under countable unions, intersections, and complements of elements of J . Therefore (J ), C = (J ), {1}C =
{2, 3, 4} (J ), {2}C = {1, 3, 4} (J ), {1} {2} = {1, 2} (J ) and
10
11
(b) Let = N = {1, 2, 3, . . . }. Let F1 = ({1}) and for each i = 2, 3,S. . . , let
Exercise 2.7.5. Suppose that = N is the set of positive integers, and F is the
set of all subsets A such that either A or AC is finite, and P is defined by P(A) = 0
if A is finite, and P(A) = 1 if AC is finite.
(a) Is F an algebra?
(b) Is F a -algebra?
(c) Is P finitely additive?
(d) Is P countably additive
2 , F are disjoint, and
S on F, meaning that
S if A1 , AP
if it happens that n An F, then P ( n An ) = n P(An )?
Proof.
(a) Yes, F is an algebra. First, F since |C | = || = 0. Let A F, then
by definition either A or AC is finite, and thus AC F. It remains to show
that F is closed under finite intersection. Let A1 , . . . , An F. If Ai is finite
for some i, then A1 An Ai is finite, and thus A1 An F.
If instead, for each j, Aj is infinite, then for each j, AC
j is finite. Thus
C
is
the
finite
union
of
finite
sets and is thus
(A1 An )C = AC
A
n
1
finite. Therefore A1 An F and F is an algebra.
(b) No, F is not a -algebra. For each j,SAj := {1, 3, 5, . . . , 2j 1} F since
Aj is finite. However {1, 3, 5, . . . } = j Aj
/ F since both {1, 3, 5, . . . } and
{1, 3, 5, . . . }C = {2, 4, 6, . . . } are infinite.
(c) Yes, P is finitely additive
S on F. Let (WLOG) A1 , . . . , An F\{, } be
disjoint and let A = j Aj . We will first show that at most one of AC
j is
and
AC
finite where 1 6 j 6 n. Suppose on the contrary that (WLOG) AC
1
2
C
C
C
C
are both finite. Let a = max{A1 A2 }. Then A1 A2 {1, 2, . . . , a} and
C C
thus A1 A2 = (AC
{1, 2, . . . , a}C = {a + 1, a + 2, . . . }. This
1 A2 )
contradicts the fact that A1 A2 = and therefore at most one of AC
j is
finite.
Now if none of AC
are finite, then each Aj is finite and therefore A is finite.
jP
P
Then P(A) = 0 = j 0 = j P(Aj ). If there exists k such that AC
k is finite,
12
P
then
A is infinite. Thus P(A) = 1 = j6=k 0 + 1 =
P Ak is infinite and therefore
P
j6=k P(Aj ) + P(Ak ) =
j P(Aj ). Therefore P is finitely additive on F.
(d) No, P is
S not countably additive on F. For each j N, let Aj = {j} with
A := j Aj F. For each j, Aj F since Aj is
clearly A
P finite, and P
is a disjoint union. Since each Aj is finite, we have j P(Aj ) = j 0 = 0.
However, A = N P
has finite complement AC = and thus P(A) = 1. Therefore
P(A) = 1 6= 0 = j P(Aj ) and P is not countably additive on F.
Exercise 2.7.6. Suppose that = [0, 1] is the unit interval, and F is the set of
all subsets A such that either A or AC is finite, and P is defined by P(A) = 0 if A
is finite, and P(A) = 1 if AC is finite.
(a) Is F an algebra?
(b) Is F a -algebra?
(c) Is P finitely additive?
(d) Is P countably additive on F (as in the previous exercise)?
Proof.
(a) Yes, F is an algebra. First, F since |C | = || = 0. Let A F, then
by definition either A or AC is finite, and thus AC F. It remains to show
that F is closed under finite intersection. Let A1 , . . . , An F. If Ai is finite
for some i, then A1 An Ai is finite, and thus A1 An F.
If instead, for each j, Aj is infinite, then for each j, AC
j is finite. Thus
C
is
the
finite
union
of
finite
sets and is thus
(A1 An )C = AC
A
n
1
finite. Therefore A1 An F and F is an algebra.
(b) No, F is not a -algebra. For each j, Aj := {1, 12 , 13 , . . . , 1j } F since Aj
S
is finite. However {1, 12 , 13 , . . . } = j Aj
/ F since both {1, 12 , 13 , . . . } and
S
1
{1, 12 , 13 , . . . }C = j j+1
, 1j are infinite.
(c) Yes, P is finitely additive
on F. Let (WLOG) A1 , . . . , An F\{, } be
S
disjoint with A := j Aj F. We will first show that at most one of AC
j
is finite where 1 6 j 6 n. Suppose on the contrary that (WLOG) AC
and
1
C
C
C C
C
=
AC
2 are both finite. Since both A1 and A2 are finite, then (A1 A2 )
C
[0, 1]\(AC
A
)
=
A
A
is
infinite
and
nonempty.
This
contradicts
the
1
2
2
1
fact that A1 A2 = and therefore at most one of AC
j is finite.
Now if none of AC
are finite, then each Aj is finite and therefore A is finite.
jP
P
Then P(A) = 0 = j 0 = j P(Aj ). If there exists k such thatP
AC
k is finite,
then
A
is
infinite
and
therefore
A
is
infinite.
Thus
P(A)
=
1
=
k
j6=k 0 + 1 =
P
P
P(A
)
+
P(A
)
=
P(A
).
Therefore
P
is
finitely
additive
on F.
j
k
j
j6=k
j
13
Exercise 2.7.7. Suppose that = [0, 1] is the unit interval, and F is the set of all
subsets A such that either A or AC is countable (i.e., finite or countably infinite),
and P is defined by P(A) = 0 if A is countable, and P(A) = 1 if AC is countable.
(a) Is F an algebra?
(b) Is F a -algebra?
(c) Is P finitely additive?
(d) Is P countably additive on F?
Proof.
(a) Yes, F is an algebra. First, F since C = is countable. Let A F,
then by definition either A or AC is countable, and thus AC F. It remains
to show that F is closed under finite intersection. Let A1 , . . . , An F.
If Ai is countable for some i, then A1 An Ai is countable, and
A1 A2 An F. If for all j, Aj is uncountable, then for all j, AC
j is
C
countable. Thus (A1 An )C = AC
A
is
the
finite
union
of
n
1
countable sets and is thus countable. Therefore A1 An F and F is
an algebra.
(b) Yes, F is a -algebra. The argument is same as in part (a) replacing finite
intersections with countable intersections.
(c) Yes, P is finitely additive
on F. Let (WLOG) A1 , . . . , An F\{, } be
S
disjoint with A := j Aj F. We will first show that at most one of AC
j
is countable where 1 6 j 6 n. Suppose on the contrary that (WLOG) AC
1
C
C
and AC
2 are both countable. Since both A1 and A2 are countable, then
C C
C
C
(AC
A
)
=
[0,
1]\(A
A
)
=
A
A
is
uncountable
and nonempty.
1
2
1
2
1
2
14
Exercise 2.7.8. For the example of Exercise 2.7.7, is P uncountably additive (cf.
page 2)?
Proof. S
No, P is not uncountably additive. For each x , consider
P Ax = {x} F
where
A
=
F.
Then
each
A
is
countable,
and
thus
x
x x
x P(Ax ) = 0, but
S
P ( x Ax ) = P () = 1.
Exercise 2.7.9. Let F be a -algebra, and write |F| for the total number of
subsets in F. Prove that if |F| < (i.e., if F consists of just a finite number of
subsets), then |F| = 2m for some m N. [Hint: Consider those non-empty subsets
in F which do not contain any other non-empty subset in F. How can all subsets
in F be built up from these particular subsets?]
Proof. Let F be a -algebra with |F| < . Let D = {D1 , . . . , Dm } consist of
those non-empty subsets in F whichSdo not contain any other non-empty subset
n
in F. Then for each A F, A = j=1 Dkj , for some Dk1 , . . . , Dkn D. Let
G = {(r1 , r2 , . . . , rm ); ri = 0 or 1}. Consider the function : F G defined by
(A) = (r1 , r2 , . . . , rm ) where for 1 6 j 6 m, we have that rj = 1 if A Dj and
rj = 0 if A ) Dj . We will prove that is a bijection, and thus |F| = |G| = 2m .
15
To show that is onto, let (k1 , . . . , km ) G and letK = {j; kj = 1}. Then
S
S
= (k1 , . . . , km ). To show
jK Dj F since F is a -algebra and
jK Dj
is one-to-one, assume (A) = (B) = (s1 , . . . , sm ) for
S A, B F and let S =
{j; sj = 1}. Then by definition of , we have A = B = jS Dj . Therefore is a
bijection.
Exercise 2.7.10. Prove that the collection J of (2.5.10) is a semialgebra.
Proof. Let J = {(, x] : x R} {(y, ) : y R} {(x, y] : x, y R} {, R}.
Clearly , R J . Next, we show J is closed under finite intersection. Let (WLOG)
A, B J \{, R}. Consider the following cases (WLOG):
Case A = (, x], B = (, y]: Let s = min{x, y}, then A B = (, s] J .
Case A = (, x], B = (y, ): If x 6 y, then A B = J . If x > y, then
A B = (y, x] J .
Case A = (, x], B = (y, z]: Let s = min{x, z}. If x 6 y, then A B = J .
If x > y, then A B = (y, s] J .
Case A = (x, ), B = (y, ): Let s = max{x, y}. Then A B = (s, ) J .
Case A = (x, ), B = (y, z]: Let s = max{x, y}. If x > z, then A B = J . If
x < z, then A B = (s, z] J .
Case A = (w, x], B = (y, z]: Let s = max{w, y} and t = min{x, z}. If w > z or
x 6 y, then A B = J . Otherwise A B = (s, t] J .
For general A1 , . . . , An J simply induct on the above argument. Therefore J
is closed under finite intersection. Now we show the complement of A J is the
finite disjoint union of elements in J . Let (WLOG) A J \{, R}. Consider the
following cases:
Case A = (, x]: AC = (x, ) J .
Case A = (y, ): AC = (, y] J .
16
1] where
{0} (0,
a] where {0}, (0, a] J 0 . If b < 1, then AC = {0} (0,
a] (b,
0
{0}, (0, a], (b, 1] J . Thus the complement of an element is the disjoint union
of elements in J 0 and therefore J 0 is a semialgebra.
(b) Let J = {all intervals contained in [0,1]} and let B = (J ). Clearly B
(J 0 ) since J J 0 . It remains to show that B (J 0 ). We will do this
by showing that (J 0 ) J , since, by definition, all -algebras containing J
contain B = (J ). Let (WLOG) A J \{, [0, 1]}. Then A is an interval
(possibly a singular point or empty). Consider the following cases (WLOG):
Case (a, b]: (a, b] (J 0 ).
0
Case [a, b]: If a = 0, then [0, b] = {0} (0,
let k be large
Tb] (J1 ). If a > 0,
1
such that a k > 0. Then [a, b] = j=k (a j , b] (J 0 ).
Case (a, b): Let k be large such that 0 < b k1 < 1. Then (a, b) =
1
0
k ] (J ).
j=k [a, b
j=k (a, b
(c) Let B00 be the collection of all finite disjoint unions of elements of J 0 . Let
SkA
Sk B
A, B B00 . Then we can write A = j=1 Aj and B = j=1 Bj for some
SkA
kA , kB and Aj , Bj J 0 for each j. Then A B =
Aj B =
j=1
S
kA (Aj B) = S
kA (Aj ( S
kB Bi )) = S
kA S
kB
j=1
i=1
j=1 i=1 (Aj Bi ). Since each
j=1
Aj , Bi J 0 , then so is Aj Bi , and therefore A B B00 . By inducting
on this previous argument we have that B00 is closed under finite intersection.
SkA
C
TkA C
Now AC =
Aj
=
A . Since J 0 is a semialgebra, for each j,
j=1
j=1
AC
j
17
Exercise 2.7.12. Let K be the Cantor set as defined inSSubsection 2.4. Let
1
n
(a)
(b) Since is shift-invariant as in (1.2.5), we must have (Dn ) = (K) = 0 for
all n. In particular, (D3 ) = 0.
(c)
(d) By part P
(b), (Dn ) = 0Pfor all n, and therefore by subadditivity we have
18
Exercise 2.7.15. Let (, M, ) be Lebesgue measure on the interval [0, 1]. Let
0 = {(x, y) R2 ; 0 < x 6 1, 0 < y 6 1}.
Let F be the collection of all subsets of 0 of the form
{(x, y) R2 ; x A, 0 < y 6 1}
for some A M. Finally, define a probability P on F by
P({(x, y) R2 ; x A, 0 < y 6 1}) = (A).
(a) Prove that (0 , F, P) is a probability triple.
(b) Let P be the outer measure corresponding to P amd F. Define the subset
S 0 by
S = (x, y) R2 ; 0 < x 6 1, y = 1/2 .
(Note that S
/ F.) Prove that P (S) = 1 and P (S C ) = 1.
Proof. Let (, M, ) be Lebesgue measure on the interval [0, 1]. Let 0 = (0, 1]2
and F = M {(0, 1]}.
(a) Since M is a -algebra, clearly F is a -algebra. Since is a probability
measure and : F M given by (A (0, 1]) = A is clearly a bijection and
P(A (0, 1]) = ((A (0, 1])) for all A M, we see that P is a probability
measure. Therefore (0 , F, P) is a probability triple.
(b) Let P be the outer measure corresponding to P and F. SLet S = (0, 1]
{1/2}. Let A1 (0, 1], A2 (0, 1], F such that S j=1 (Aj (0, 1]).
S
Then j=1 (Aj (0, 1]) = (0, 1]2 = 0 . Therefore, by additivity, P(0 ) 6
19
Aj (0, 1] . Thus by definition of outer measure, we have P (S) =
P(0 ) = 1. Now we have that S C = (0, 1]C {1/2} (0, 1] {1/2}C
(0, 1]C {1/2}C = (0, 1] {1/2}C . Then the exact same argument as
S
j=1
Exercise 2.7.16.
(a) Where in the proof of Theorem 2.3.1 was assumption (2.3.3) used?
(b) How would the conclusion of Theorem 2.3.1 be modified if assumption (2.3.3)
were dropped (but all other assumptions remained the same)?
Proof.
(a) In Lemma 2.3.5 it is used to show that P (A) = P(A) for A J .
(b) We have that P is not extension of P, but is a probability measure on M
(J ).
Exercise 2.7.17. Let = {1, 2}, and let J be the collection of all subsets of ,
with P() = 0, P() = 1, and P{1} = P{2} = 1/3.
(a) Verify that all assumptions of Theorem 2.3.1 other than (2.3.3) are satisfied.
(b) Verify that assumption (2.3.3) is not satisfied.
(c) Describe precisely the M and P that would result in this example from the
modified version of Theorem 2.3.1 in Exercise 2.7.16(b)
Proof.
(a) J is a semialgebra since it contains ALL subsets of . P() = 0 and P() = 1
by assumption. We have A = A so that P(A ) = P(A) = P(A) + 0 =
P(A) + P(). Also, P({1} {2}) = P{1, 2} = 1 > 2/3 = 1/3 + 1/3 =
P{1} + P{2}. Therefore (2.3.2) is satisfied.
(b) We have that {1} {2} {1, 2}, but P({1} {2}) = P{1, 2} = 1 > 2/3 =
1/3 + 1/3 = P{1} + P{2}. Therefore (2.3.3) is not satisfied.
(c) M = {A ; P (A E) + P (AC E) = P (E) E } = {, } (J ).
20
Exercise 2.7.18.
P({1}) = 1/3.
, , {1} , P() = 0, P() = 1, and
(a) Can Theorem 2.3.1, Corollary 2.5.1, or Corollary 2.5.4 be applied in this case?
Why or why not?
(b) Can this P be extended to a valid probability measure? Explain.
Proof.
(a) They cannot be applied. J is not a semialgebra since {1}C = {2} is not a
disjoint union of elements in J . Therefore the hypotheses are not satisfied.
(b) Yes, we can extend P to P on (J ) by defining P (A) = P(A) on J , and
P {2} = 1 P {1} = 1 1/3 = 2/3.
21
Exercise 2.7.20. Let P and Q be two probability measures defined on the same
sample space and -algebra F.
(a) Suppose that P(A) = Q(A) for all A F with P(A) 6 21 . Prove that P = Q,
i.e. that P(A) = Q(A) for all A F.
(b) Give an example where P(A) = Q for all A F with P(A) <
that P 6= Q, i.e. that P(A) 6= Q(A) for some A F.
1
2,
but such
Proof. Let P and Q be two probability measures defined on the same sample space
and -algebra F.
(a) Suppose that P(A) = Q(A) for all A F with P(A) 6 21 . Let A F with
P(A) > 21 . Then P(AC ) = 1 P(A) 6 12 and thus P(AC ) = Q(AC ). Thus
Q(A) = 1 Q(AC ) = 1 P(AC ) = P(A). Therefore P(A) = Q(A) for all
A F.
(b) Let = {, {1}, {2}, {1, 2}}. Define P by P() = 0, P() = 1, and P{1} =
P{2} = 21 . Define Q by Q() = 0, Q() = 1, Q{1} = 31 , and Q{2} = 23 .
Then P() = Q() = 0 and therefore P(A) = Q(A) for all A F with
P(A) < 21 , but clearly P 6= Q.
R
,
j=1 Aj .
2m
2m
2m 2m Am
S
Therefore A = j=1 Aj and A M and is thus measurable with respect to .
Now, let B1 = A1 and for each j, let Bj+1 = Aj+1 \Aj . Since A1 A2 . . . we
S
can rewrite A as the disjoint union j=1 Bj . Therefore
22
(A) =
(Bj )
= (A1 ) +
((Aj+1 ) (Aj ))
j=1
j+1
X
X
2j
2X
i
i
1
i
1
i
, j
= R
,1
+
R
R
j+1 2j+1
j
2
2
2
2
i=2
j=1
i=2
j+1
j
2
2
1 X X i 1 X i 1
= +
4 j=1 i=2 22j+2 i=2 22j
j+1
j
2X
1
2
1
1 X 1 X
= +
4i
i
4 j=1 22j+2
i=1
i=1
1 X 1
+
(2j+1 1)2j 4(2j 1)2j1
4 j=1 22j+2
1 X 1
+
4 j=1 2j+2
1 1X 1
+
4 4 j=1 2j
1 1
+
4 4
1
= .
2
23
3.1 Exercises
Exercise 3.1.4. For Example 3.1.3, compute P(Z > a) and P(X < a and Y < b)
as functions of a, b R.
Proof. Let (, F, P) be Lebesgue measure on [0,1]. Let X, Y and Z be defined by
X() = , Y () = 2 and Z() = 3 + 4. Then P(Z > a) = P(; 4 + 3 >
1a
b
a) = P ; > a3
= P ( a3
4
4 , 1] =
4 . Let c = min{a, 2 }, then P(X <
a and Y < b) = P(; < a and 2 < b) = P(; < a and < 2b ) = P(; <
c) = P([0, c)) = c.
Exercise 3.1.7. Prove (3.1.6). [Hint: remember the definition of X(w) =
limn Xn (), cf. Subsection A.3.]
Proof. Let Z1 , Z2 , . . . be random variables such that
Zn () exists for each
T lim
Sn
T
1
}.
, and Z() = limn Zn (). Let A = m=1 n=1 k=n {Zk 6 x + m
Let {Z 6 x}. If Z() 6 x, then for each natural number m there exists
1
n(m)Tsuch that for all k > n, we have Zk () 6 x + m
. That is, for each m,
1
k=n(m) {Zk 6 x + m } and therefore A, so that {Z 6 x} A. Now
suppose A. Then for each natural number m there exists n(m) such that for
1
1
. Thus for each m, Z() 6 x + m
. Since
all k > n, we have Zk () 6 x + m
this is true for all m, we must have Z(w) 6 x. Therefore {Z 6 x}, so that
A {Z 6 x}. This implies that {Z 6 x} = A.
3.2 Exercises
Exercise 3.2.2. Suppose (3.2.1) is satisfied.
(a) Show that (3.2.1) is still satisfied if A1 is replaced by AC
1 .
(b) Show that (3.2.1) is still satisfied if each Ai is replaced by the corresponding
AC
i .
(c) Prove that if {A }I is independent, then so is {AC
}I .
24
Proof. Let P be a probability measure. Let {A }I be a possibly-infinite independent collection of events. That is, for each j N and each distinct finite choice
1 , 2 , . . . , j I, we have
P(A1 A2 Aj ) = P(A1 )P(A2 ) . . . P(Aj ).
(a) Let j N and let 1 , 2 , . . . , j I. Since
P(A1 A2 Aj AC
1 A2 Aj ) = P(A2 Aj )
then
P(AC
1 A2 Aj ) = P(A2 Aj ) P(A1 A2 Aj )
= P(A2 ) . . . P(Aj ) P(A1 )P(A2 ) . . . P(Aj )
= (1 P(A1 ))P(A2 ) . . . P(Aj )
= P(AC
1 )P(A2 ) . . . P(Aj )
(b) Let j N and let 1 , 2 , . . . , j I. Suppose
C
C
C
P(AC
1 Ak Ak+1 Aj ) = P(A1 ) . . . P(Ak )P(Ak+1 ) . . . P(Aj ),
3.6 Exercises
Exercise 3.6.1. Let X be a real-valued random variable defined on a probability
triple (, F, P). Fill in the following blanks:
(a) F is a collection of subsets of
25
subset of
, then {X S} is a subset of
which is defined
.
Exercise 3.6.2. Let (, F, P) be Lebesgue measure on [0, 1]. Let A = (1/2, 3/4)
and B = (0, 2/3). Are A and B independent events?
Proof. Yes. Let (, F, P) be Lebesgue measure on [0, 1]. Let A = (1/2, 3/4) and
B = (0, 2/3). Then A B = (1/2, 2/3) and P(A B) = 1/6 = (1/4)(2/3) =
P(A)P(B).
Exercise 3.6.3. Give an example of events A, B, and C, each of probability
strictly between 0 and 1, such that
(a) P(A B) = P(A)P(B), P(A C) = P(A)P(C), and P(B C) = P(B)P(C);
but it is not the case that P(A B C) = P(A)P(B)P(C). [Hint: You can
let be a set of four equally likely points.]
(b) P(A B) = P(A)P(B), P(A C) = P(A)P(C), and P(A B C) =
P(A)P(B)P(C); but it is not the case that P(B C) = P(B)P(C). [Hint:
You can let be a set of eight equally likely points.]
Proof. (a) Let = {1, 2, 3, 4}, F = 2 and P be defined by P(n) = 1/4 for n .
Let A = {1, 2}, B = {2, 3} and C = {1, 3}. Then A B = {2}, B C = {3}
and A C = {1}, thus P(A B) = 1/4 = (1/2)(1/2) = P(A)P(B), P (B
C) = 1/4 = (1/2)(1/2) = P(B)P(C), and P(A C) = 1/4 = (1/2)(1/2) =
P(A)P(C). However, A B C = , and therefore P(A B C) = 0 6=
1/8 = (1/2)(1/2)(1/2) = P(A)P(B)P(C).
(b) Let = {1, 2, 3, 4, 5, 6, 7, 8}, F = 2 and P be defined by P(n) = 1/8 for n
and extended to F. Let A = {1, 2, 3, 4}, B = {3, 4, 5, 6} and C = {1, 3, 7, 8}.
Then A B = {3, 4}, A C = {1, 3} and A B C = {3}, thus P(A B) =
1/4 = (1/2)(1/2) = P(A)P(B), P(A C) = 1/4 = (1/2)(1/2) = P(A)P(C)
26
Ak =
(Bk,n Ck,m ).
n, m Z
(n + m)2k < x
Fix x, z R, and let A = {X + Y < x} = {W < x} and D = {Z < x}.
(a) Prove that {Ak } % A.
27
(Bk,n Ck,m ).
Ak =
n, m Z
(n + m)2k < x
Fix x, z R, and let A = {X + Y < x} = {W < x} and D = {Z < z}.
(a) Since {n, m Z; (n + m)2k < x} {n, m Z; (n + m)2(k+1) < x}, we have
that {Ak } is increasing.
Let Ak . Then we have Bk,n Ck,m for some n, m Z with (n +
m)2k < x. This implies that X() < n2k and Y () < m2k , so that
X() + Y () < n2k + m2k = (n + m)2k < x. Thus A. Therefore,
for each k, Ak A.
It remains to show that for each A, there exists a k0 such that Ak0 .
Let A. For each k, there exists a unique nk , mk such that (nk 1)2k 6
X() < nk 2k and (mk 1)2k 6 Y () < mk 2k . Then for each k, we have
X()+Y () < (nk +mk )2k and in the limit limk (nk +mk )2k = X()+
Y (). Then we can find k0 such that X() + Y () < (nk0 + mk0 )2k0 < x.
Thus Ak0 . Therefore we have {Ak } % A.
(b) Since X,
Y and Z are independent, for each k, n, m, we have P(Bk,n Ck,m
D) = P X (n 1)2k , n2k , Y (m 1)2k , m2k , Z (, z) =
k
k
k
k
P X (n 1)2 , n2
, Y (m 1)2 , m2
P (Z (, z)) =
P(Bk,n Ck,m )P(D). Thus, since Ak is a disjoint union over n, m, we have
X
P(Ak D) =
P(Bk,n Ck,m D)
(n+m)2k <x
(n+m)2k <x
= P(Ak )P(D).
Therefore Ak and D are independent.
(c) Since {Ak } % A, we have that {Ak D} % A D. By continuity of probabilities (Proposition 3.3.1), we have that P(A D) = limk P(Ak D) =
28
29
(b) Give an example where the above inclusion is strict, and another example
where it holds with equality.
Proof. (a) Let A1 , A2 , . . . , B1 , B2 , . . . be
n Bn ), then
Sevents. Let c lim supnS(A
there exists some k0 such that c k=k0 (Ak Bk ). then c k=k0 Ak and c
S
k=k0 Bk . Thus c (lim supn An ) (lim supn Bn ). Therefore (lim supn An )
(lim supn Bn ) lim supn (An Bn ).
(b) Consider the sequences A1 = {1}, A2 = , A3 = {1}, A4 = , . . . and B1 =
, B2 = {1}, B3 = , B4 = {1}, . . . . Then An Bn = for all n so that
lim supn (An Bn ) = , but (lim supn An ) = (lim supn Bn ) = {1}. Therefore
we have that (lim supn An ) (lim supn Bn ) ) lim supn (An Bn ).
Now consider the sequences An = Bn = {1} for all n. Then clearly {1} =
(lim supn An ) (lim supn Bn ) = lim supn (An Bn ).
30
for
i
=
1,
2,
.
.
.
,
n.
Let
A
be
the
event
that
X
=
5.
Then
n
n
n=5 P(An ) =
P
P
1/n
=
.
This
implies
that
P(A
)
=
.
Since
{X
}
n
n
n=1 are inden=5
n
pendent, we have that {An }
are
independent,
and
thus
by
the
Borel-Cantelli
n=1
lemma, P(Xn = 5 i.o.) = 1.
Exercise 3.6.12. Let X be a random variable with P(X > 0) > 0. Prove
that there is > 0 such that P(X > ) > 0. [Hint: Dont forget continuity of
probabilities.]
Proof. Let X be a random variable with P(X > S
0) > 0. Consider the increasing
sequence of events An = {X > 1/n}, with A := n An = {X > 0}. Suppose for
contradiction that for all n we have that P(An ) = 0. By continuity of probability
we have that limn P(An ) = P(A). This implies that for all > 0, there exists an
N such that for all n > N , we have that |P(A) P(An )| = |P(A)| < . Since this is
true for all > 0, then P(A) = 0 which is a contradiction. Thus, there exists some
n0 such that P(An0 ) > 0. Therefore, given = 1/n0 , we have P({X > }) > 0.
Exercise 3.6.13.
P Let X1 , X2 , . . . be defined jointly on some probability space
(, F, P), with i=1 i2 P(i 6 Xn < i + 1) 6 C < for all n. Prove that P[Xn >
n i.o.] = 0.
Proof. Let X1 , X2 , . . . be defined jointly on some probability space (,P
F, P), with
P
2
i
P(i
6
X
<
i
+
1)
6
C
<
for
all
n.
Then
P[X
>
n]
=
n
n
i=1
i=n P(i 6
P 2
P 2
2
Xn < i+1) 6 n P[Xn > n] = i=n n P(i 6 Xn < i+1)
6
i
P(i
6
i=1
P
P Xn < i+
1) 6 C. This implies that P[Xn > n] 6 C/n2 . Then n P[Xn > n] 6 n C/n2 <
. Then by the Borel-Cantelli lemma, we have that P[Xn > n i.o.] = 0.
Exercise 3.6.14. Let > 0, 1 > > 0 and let X1 , X2 , . . . be a sequence of
independent non-negative random
variables such that P(Xi > ) > for all i.
P
Prove that with probability one, i=1 Xi = .
31
1,
and
(ii)
j+1
j
n P(An ) = .
P
P
Then either
P(A
)
=
or
P(A
)
=
,
(if
both
were
finite, then
2n1
2n
n
n
their sum would be finite which contradicts (ii)). By (i) and simple induction, we
have that A1 , A3 , . . . are independent and A2 , A4 , . . . are independent. Thus by
the Borel-Cantelli Lemma, either P(lim supn A2n1 ) = 1 or P(lim supn A2n ) = 1
and therefore by subadditivity P(lim supn An ) = 1.
Exercise 3.6.16. Consider infinite, independent, fair coin tossing as in Subsection
2.6, and let Hn be the event that the nth coin is heads. Determine the following
probabilities.
(a) P(Hn+1 Hn+2 Hn+9 i.o.).
(b) P(Hn+1 Hn+2 H2n i.o.).
(c) P(Hn+1 Hn+2 Hn+[2 log2 n] i.o.).
(d) Prove that P(Hn+1 Hn+2 Hn+[log2 n] i.o.) must equal either 0 or 1.
(e) Determine P(Hn+1 Hn+2 Hn+[log2 n] i.o.). [Hint: Find the right subsequence of indices.]
Proof. Consider infinite, independent, fair coin tossing as in Subsection 2.6, and
let Hn be the event that the nth coin is heads.
(a) Let Bn = Hn+1 Hn+2
the subsequence of indepedent
P Hn+9 and
Pconsider
9
events B9n . Then
P(B
)
=
1/2
=
, and thus by the Borel9n
n
n
Cantelli lemma, P(B9n i.o.) = 1. Therefore by subadditivity, P(Hn+1
Hn+2 Hn+9 i.o.) = 1.
P
P
(b) Let Bn = Hn+1 Hn+2 H2n . Then n P(Bn ) = n 1/2n < , and
therefore by the Borel-Cantelli lemma, P(Bn i.o.) = 0.
P
P
[2 log2 n]
=
(c) Let
Bn = Hn+1 HP
n+2 Hn+[2 log2 n] . Then Pn=2 P(Bn ) =
n=2 1/2
P
n
2n
n
= n=1 1/2 < Therefore n=1 P(Bn ) < and by the
n=1 2 /2
Borel-Cantelli lemma, P(Bn i.o.) = 0.
32
(d),(e) Let Bn = Hn+1 Hn+2 Hn+[log2 n] . Since (n+1)2 > n log2 n2 for all n,
we have that [log2 (n + 1)2 ] > [log2 (n log2 n2 )] for all n. Then (n + 1)[log2 (n +
1)2 ] + 1 > n[log2 n2 ] + [log2 (n[log2 n2 ])] and {Bn[log2 n2 ] }
n=2 is an indepenP
P
1
dent sequence of events with n=2 P(Bn[log2 n2 ] ) > n=2
= .
n[log2 n2 ]
Therefore, by the Borel-Cantelli lemma, P(Bn i.o.) = 1.
Exercise 3.6.17. Show that Lemma 3.5.2 is false if we require only that P(B
Bn ) = P(B)P(Bn ) for each n N, but do not require that the {Bn } be indepedent
of each other. [Hint: Dont forget Exercise 3.6.3(a).]
Proof. From Exercise 3.6.3(a), let = {1, 2, 3, 4}, F = 2 and P be defined by
P(n) = 1/4 for n . Let B = {1, 2}, B1 = {2, 3} and B2 = {1, 3}. Then
B B1 = {2}, B1 C2 = {3} and B B2 = {1}, thus P(B B1 ) = 1/4 =
(1/2)(1/2) = P(B)P(B1 ), P (B1 B2 ) = 1/4 = (1/2)(1/2) = P(B1 )P(B2 ), and
P(B B2 ) = 1/4 = (1/2)(1/2) = P(B)P(B2 ). However, B (B1 B2 ) = , and
therefore P(B (B1 B2 )) = 0 6= 1/8 = (1/2)(1/4) = P(B)P(B1 B2 ).
Therefore P(B Bn ) = P(B)P(Bn ) for each n, but for B1 B2 (B1 , B2 ), we
have that P(B (B1 B2 )) 6= P(B)P(B1 B2 ).
Exercise 3.6.18.PLet A1 , A2 , . . . be any independent sequence of events, and let
n
Sx = {limn n1 i=1 1Ai 6 x}. Prove that for each x R we have P(Sx ) = 0 or
1.
Proof. Let A1 , A2 , . . . be any independent sequence of events, and let
=
(An , An+1 , . . . ).
n=1
Pn
Pn
For each x R we have Sx := {limn n1 i=1 1Ai 6 x} {limn n1 i=k 1Ai 6
x} (Ak , Ak+1 , . . . ) for each k N. Thus Sx and therefore, by the Kolmogorov Zero-One Law, for each x R we have P(Sx ) = 0 or 1.
Exercise 3.6.19. Let A1 , A2 , . . . be independent events. Let Y be a random
variable which is measurable with respect to (An , An+1 , . . . ) for each n N.
Prove that there is a real number a such that P(Y = a) = 1. [Hint: Consider
P(Y 6 x) for x R; what values can it take?]
Proof. Let A1 , A2 , . . . be independent events. Let Y be a random variable which is
measurable with respect to (An , An+1 , . . . ) for each n N. Then Y is measurable
with respect to
\
=
(An , An+1 , . . . ).
n=1
33
Then by the Kolmogorov Zero-One Law (and the defintion of a random variable),
for each x R, we have P(Y 6 x) S
= 0 or P(Y 6 x) = 1. Since Bn = {Y 6 n}
is an increasing sequence such that n Bn = {Y R}, by continuity there exists
some N such that P(Bn ) = 1 for all n > N (since P(YT R) = 1). Similarly, since
Cn = {Y 6 n} is a decreasing sequence such that n Cn = , there exists a K
such that P(Cn ) = 0 for all n > K. Thus the set S := {x R; P(Y 6 x) = 1} is
nonempty and bounded below.
Let a = inf S. Let Dn = {Y 6 aS 1/n} be an increasing sequence of events.
Then P(Dn ) = 0 for each n with n Dn = {Y < a}, and thus by continuity of
probability P(Y < a) = 0. Let En = {Y 6 a + 1/n} be a T
decreasing sequence of
events. Since a + 1/n > a, then P(En ) = 1 for each n with n En = {Y 6 a}, and
thus by continuity of probability P(Y 6 a) = 1. Therefore, P(Y = a) = P(Y 6
a) P(Y < a) = 1 0 = 1.
4.1 Exercises
Exercise 4.1.3. Prove that (4.1.2) is well-defined, in P
the sense thatP
if {Ai } and
n
m
{Bj } are two different finite partitions of , such that i=1 xi 1Ai = j=1 yj 1Bj ,
Pn
Pm
then
i=1 xi P(Ai ) =
j=1 yj P(Bj ). [Hint: collect together those Ai and Bj
corresponding to the same values of xi and yj .]
Pn
Proof. Suppose {Ai } and {Bj } are two finite partitions of , such that i=1 xi 1Ai =
Pm
S
n
I = {j; yj = xi } so that i=1 Ixi = {1, 2, . . . , m}. Then we have
j=1 yj 1Bj . Let
P xi
Pn
Pn P
that xi 1Ai = jIx yj 1Bj for each i. Then i=1 xi 1Ai = i=1 jIx yj 1Bj =
i
iP
Pm
P
y
1
.
Therefore,
since
x
P(A
)
=
E(x
1
y
1
)
=
E
= jIx yj P(Bj ),
j
B
i
i
i
A
j
B
j
i
j
j=1
jIxi
Pn
Pn
Pn P
Pm i
we have that E ( i=1 xi 1Ai ) = i=1 xi P(Ai ) = i=1 jIx yj P(Bj ) = j=1 yj P(Bj ).
i
4.3 Exercises
Exercise 4.3.2. Let X and Y be two general random variables (not necessarily
non-negative) with well-defined means, such that X 6 Y .
(a) Prove that X + 6 Y + and X > Y .
(b) Prove that expectation is still order-preserving, i.e. that E(X) 6 E(Y ) under
these assumptions.
Proof. Let X and Y be two general random variables (not necessarily non-negative)
with well-defined means, such that X 6 Y .
34
Exercise 4.3.3. Let X and Y be two general random variables with finite means,
and let Z = X + Y .
(a) Express Z + Z in terms of X + , X , Y + , and Y .
(b) Prove that E(Z) = E(X) + E(Y ), i.e. that E(Z + ) E(Z ) = E(X + )
E(X ) + E(Y + ) E(Y ). [Hint: Re-arrange the relations of part (a) so that
you can make use of (4.2.6).]
(c) Prove that expectation is still (finitely) linear, for general random variables
with finite means.
Proof. Let X and Y be two general random variables with finite means, and let
Z =X +Y.
(a) Z + Z = Z = X + Y = X + X + Y + Y .
(b) From (a) we have Z + + X + Y = Z + X + + Y + . Then by (4.2.6),
we have E(Z + ) + E(X ) + E(Y ) = E(Z ) + E(X + ) + E(Y + ). Therefore
E(Z) = E(Z + )E(Z ) = E(X + )E(X )+E(Y + )E(Y ) = E(X)+E(Y ).
(c) Let a, b R. Since, by definition, (X)+ = X and (X) = X + , we have
E(X) = E(X ) E(X + ) = E(X). So WLOG assume a, b > 0. Let U =
aX, V = bY and W = U + Y . Then U and V are general random variables
with finite means. Then by (b) and (4.2.6), we have E(W ) = E(U + )
E(U ) + E(V + ) E(V ) = E(aX + ) E(aX ) + E(bY + ) E(bY ) =
aE(X + ) aE(X ) + bE(Y + ) bE(Y ) = aE(X) + bE(Y ).
Exercise 4.3.4. Let X and Y be two independent general random variables with
finite means, and let Z = XY .
(a) Prove that X + and Y + are independent, and similarly for each of X + and Y ,
and X and Y + , and X and Y .
(b) Express Z + and Z in terms of X + , X , Y + , and Y .
(c) Prove that E(XY ) = E(X)E(Y ).
Proof. Let X and Y be two general random variables with finite means, and let
Z =X +Y.
35
(a) Since X + > 0 ,if x > 0, we have that P(X + 6 x) = P(X + = 0)+P(0 < X + 6
x) = P(X 6 0) + P(0 < X 6 x) = P(X 6 x). Also, P(X 6 x) = P(X =
0) + P(0 < X 6 x) = P(X 6 0) + P(0 < X 6 x) = P(X 6 x).
Similarly, if y > 0, we have that P(Y + 6 y) = P(Y 6 y) and P(Y 6 y) =
P(Y 6 y).
Then P(X + 6 x, Y + 6 y) = P(X + = 0, Y + 6 y) + P(0 < X + 6 x, Y + 6
y) = P(X + = 0, Y + = 0) + P(X + = 0, 0 < Y + 6 y) + P(0 < X + 6 x, Y + =
0) + P(0 < X + 6 x, 0 < Y + 6 y) = P(X 6 0, Y 6 0) + P(X 6 0, 0 < Y 6
y) + P(0 < X 6 x, Y 6 0) + P(0 < X 6 x, 0 < Y 6 y) = P(X 6 0)P(Y 6
0)+P(X 6 0)P(0 <Y 6 y)+P(0 < X 6 x)P(Y 6 0)+P(0 < X 6
x)P(0 <
Y 6 y) = P(X 6 0) P(Y 6 0) + P(0 < Y 6 y) + P(0 < X 6 x) P(Y 6
0) + P(0 < Y 6 y) = P(X 6 0)P(Y 6 y) + P(0 < X 6 x)P(Y 6 y) =
P(X 6 x)P(Y 6 y). Since X + > 0, if x < 0, then P(X + 6 x) = 0. So if
x < 0 or y < 0 there is nothing to prove (anything times 0 is 0). Therefore
by Proposition 3.2.4, X + and Y + are independent. The proof is similar for
X + and Y , and X and Y + , and X and Y , using the fact that if X and
Y are independent then each of X and Y are independent.
+
+
(b) Z = XY = (X X )(Y Y ). Therefore Z () = max (X + ()
4.5. Exercises
Exercise 4.5.1. Let (, F, P) be Lebesgue measure on [0, 1], and set
1, 0 6 < 1/4
2 , 3/4 6 6 1.
Compute P(X A) where
36
1, 0 6 < 1/4
2 , 3/4 6 6 1.
(a) Let A = [0, 1]. We have that 0 6 2 2 6 1 if and only if 0 6 6 2/2 < 3/4.
Also 06 2 6 1 if and only if 06 6 1. Therefore P(X A) = P([0, 1/4)
[1/4, 2/2) [3/4, 1]) = 1/4 + 2/2 1/4 + 1/4 = (1 + 2 2)/4.
(b) Let A = [1/2, 1]. Then 1/2 6 2 2 61 if and only if 1/2 6 6 2/2 < 3/4.
Also 1/2 6 2 61 if and only if 2/2 6 6 1. Therefore P(X
A) =
P([0, 1/4) [1/2, 2/2] [3/4, 1]) = 1/4 + 2/2 1/2 + 1/4 = 2/2.
Exercise 4.5.2. Let X be a random variable with finite mean, and let a R
be any real number. Prove that E(max(X, a)) > max(E(X), a). [Hint: Consider
separately the cases E(X) > a and E(X) < a.] (See also Exercise 5.5.7.)
Proof. Let X be a random variable with finite mean, and let a R be any real
number. Let A = {X > a}, then max(X, a) = X1A + a1AC and by linearity
we have E(X1A ) > E(a1A ) and E(a1AC ) > E(X1AC ). Thus E(max(X, a)) =
E(X1A ) + E(a1AC ) > E(X1A ) + E(X1AC ) = E(X) and similarly we have that
E(max(X, a)) > E(a) = a. Therefore E(max(X, a)) > max(E(X), a).
Exercise 4.5.3. Give an example of random variables X and Y defined on
Lebesgue measure on [0,1], such that P(X > Y ) > 21 , but E(X) < E(Y ).
Proof. Let X = 1[0,1] and let Y = 1001[3/4,0] . Then P(X > Y ) = P([0, 3/4)) =
3/4, but E(X) = P([0, 1]) = 1 < E(Y ) = 100P([3/4, 1]) = 25.
Exercise 4.5.4. Let (, F, P) be the uniform distribution on = {1, 2, 3}, as
in Example 2.2.2. Find random variables X, Y, and Z on (, F, P) such that
P(X > Y )P(Y > Z)P(Z > X) > 0, and E(X) = E(Y ) = E(Z).
Proof. Let (, F, P) be the uniform distribution on = {1, 2, 3}. Let X = 1{1,2} ,
Y = 1{2,3} and Z = 1{1,3} . Then we have P(X > Y ) = P({1}) = 1/3, P(Y >
Z) = P({2}) = 1/3, and P(Z > X) = P({3}) = 1/3, so that P(X > Y )P(Y >
Z)P(Z > X) = 1/27 > 0. Also, we have that E(X) = P({1, 2}) = 2/3, E(Y ) =
P({2, 3}) = 2/3, and E(Z) = P({1, 3}) = 2/3, so that E(X) = E(Y ) = E(Z).
37
n
X
P(Ai )
i=1
P(Ai Aj )+
16i<j6n
P(Ai Aj Ak ) P(A1 An ).
16i<j<k6n
[Hint: Expand 1
Qn
i=1 (1
n
Y
(1 1Ai ) = 1 (1n
i=1
n
X
i=1
1Ai +
1Ai 1Aj
16i<j6n
16i<j<k6n
n
X
1Ai
i=1
1Ai Aj
16i<j6n
16i<j<k6n
38
1Ai Aj Ak 1A1 An .
Qn
i=1
C
P(A1 An ) = 1P(AC
C ) = E(11AC AC )
1 An ) = 1E(1AC
n
1 An
1
= E(1
n
Y
(1 1Ai )) =
i=1
E(1Ai )
i=1
E(1Ai Aj )
16i<j6n
E(1Ai Aj Ak ) E(1A1 An )
16i<j<k6n
n
X
P(Ai )
i=1
n
X
P(Ai Aj )
16i<j6n
P(Ai Aj Ak ) P(A1 An ).
16i<j<k6n
39
Exercise 4.5.9. In proving property (4.1.6) of variance, why did we not simply
proceed by induction on n? That is, suppose we know that Var(X+Y ) = Var(X)+
Var(Y ) whenever X and Y are independent. Does it follow easily that Var(X +
Y + Z) = Var(X) + Var(Y ) + Var(Z) whenever X, Y, and Z are independent?
Why or why not? How does Exercise 3.6.6 fit in?
Proof. In (4.1.6) it is only required that X1 , . . . , Xn are pairwise independent
(since we only need Cov(Xi , Xj ) = 0), which is more general. Also, we introduce the concept of covariance. Induction is straight forward. Suppose we
know that Var(X + Y ) = Var(X) + Var(Y ) whenever X and Y are independent. Suppose X, Y, and Z are independent, then letting W = X + Y , by Exercise 3.6.6, W and Z are independent. Therefore by the hypothesis, we have
Var(W + Z) = Var(W ) + Var(Z) = Var(X) + Var(Y ) + Var(Z).
Exercise 4.5.10. Let X1 , X2 , . . . be i.i.d. with mean mean and variance 2 , and
let N be a non-negative integer-valued random variable with mean mPand variance
40
X
k=1
Var(Xk 1N >k ) + 2
i<j
X
2 (E(12N >k ) E(1N >k )2 ) + 2 E(12N >i ) + 2
2 (E(1N >j ) E(1N >i )E(1N >j ))
i<j
k=1
= 2 m 2
= 2 m
i<j
P(N > j) 2
P(N > j)
i<j
k=1
P(N > j)
i<j
i<j
i<j
k=1
= 2 m + 2
= 2 m + 2
P(N > j) 22
i<j
k=1
i<j
!2
P(N > k)
+ 2 m
k=1
= 2 m + 2
i<j
= 2
m+
!
l(l 1)P(N = l) m2
+ 2 m
l=1
2
m+
l P(N = l)
l=1
2
= 2
!
2
lP(N = l) m
+ 2 m
l=1
2
m + E(N ) m m
m2 + v m2 + 2 m
+ 2 m
= 2 v + 2 m.
Exercise 4.5.11. Let X and Z be independent, each with the standard normal
distribution, let a, b R (not both 0), and let Y = aX + bZ.
(a) Compute Corr(X, Y ).
(b) Show that |Corr(X, Y )| 6 1 in this case. (Compare Exercise 5.5.6.)
(c) Give necessary and sufficient conditions on the values of a and b such that
Corr(X, Y ) = 1.
41
(d) Give necessary and sufficient conditions on the values of a and b such that
Corr(X, Y ) = 1.
Proof. Let X and Z be independent, each with the standard normal distribution
(that is, with mean 0 and variance 1), let a, b R (not both 0), and let Y = aX+bZ.
(a) We have that E(Y ) = aE(X) + bE(Z) = 0. Also, E(X 2 )E(X)2 = Var(X) =
1 so that E(X 2 ) = 1. Then by independence, Cov(X, Y ) = E(XY ) =
E(aX 2 + bXZ) = aE(X 2 ) + bE(X)E(Z) = aE(X 2 ) = a. Also by in2
dependence, we have that Var(Y ) = Var(aX) +
Var(bZ) = a Var(X) +
2
2
2
2
2
b Var(Z) = a + b . Therefore Corr(X, Y ) = a/ a + b .
(c) Since
a2 + b2 > |a|, we have that Corr(X, Y ) = 1 if and only if a > 0 and
2
a + b2 = a if and only if a > 0 and b = 0.
(d) Similar to (c) we have that Corr(X, Y ) = 1 if and only if a < 0 and b = 0.
E(Zi+ ) <
+
theorem,
we
have
Z
=
Z
and
Z
=
i iP
i Zi , and we have E(Z ) =
P
+
E(Z ) = i E(Zi ).
P
P
(b) Suppose that at least one of i E(Zi+ ) < orP i E(Zi ) < . If they are
both finite,Pthen by (a) we have
= i E(Zi ). We will prove the
Pthat E(Z)
n=1
n=1
(c) Use this result to prove that the J and P for product measure, presented in
Subsection 2.6, do indeed satisfy (2.5.5).
44
all disjoint, and furthermore that n=1 (An Bn ) = A B for some A F1 and
B F2 .
S
(a) Since
(An Bn ) = A B, we have that
n=1
1AB (1 2 ) = 1A (1 )1B (2 ) =
1An (1 )1Bn (2 )
n=1
Or in otherwords,
n=1
n=1
45
implies that n=1 P(Xn > n) 6 n=1 1/n2 < . Therefore, by the Borel-Cantelli
Lemma, we that P[Xn > n i.o.] = 0.
5.5 Exercises
Exercise 5.5.1. Suppose E(2X ) = 4. Prove that P(X > 3) 6 1/2.
Proof. Suppose E(2X ) = 4. First note that X > 3 if and only if log2 (2X ) > log2 (8)
if and only if 2X > 8. Then by Markovs inequality, P(X > 3) = P(2X > 8) 6
E(2X )/8 = 1/2.
Exercise 5.5.2. Give an example of a random variable X and > 0 such that
P(X > ) > E(X)/. [Hint: Obviously X cannot be non-negative.] Where does
the proof of Markovs inequality break down in this case?
Proof. Let (, F, P) be Lebesgue measure on the interval [0, 1]. Let X be a random
variable defined by X() = 1 for [0, 3/4) and X() = 1 for [3/4, 1]. Then
E(X) = 1(3/4) + 1(1/4) = 1/2. But since probability is non-negative, for any
> 0, we have P(X > ) > 0 > 1/2 = E(X)/. The proof of Markovs
inequality relies on bounding X below by a non-negative simple random variable
which cannot be done if the range of X has negative values. In particular, the
simple random variable takes on the value of 0 if X() < which is necessary to
obtain the inequality.
Exercise 5.5.3. Give examples of random variables Y with mean 0 and variance
1 such that
(a) P(|Y | > 2) = 1/4.
(b) P(|Y | > 2) < 1/4.
Proof. (a) Let Y be defined on = {0, 1}3 by Y () = 2 if = (1, 0, 0), Y () = 2
if = (0, 1, 0), and Y () = 0 otherwise. Then E(Y ) = 2(1/8) + 2(1/8) = 0
and Var(Y ) = E(Y 2 ) = 4(1/8) + 4(1/8) = 1. Also P(|Y | > 2) = P(Y =
2 or Y = 2) = P(Y = 2) + P(Y = 2) = 1/8 + 1/8 = 1/4.
(b) Let Y be defined on {0, 1} by Y () = 1 if = 0 and Y () = 1 if = 1.
Then E(Y ) = 1/2 + 1/2 = 0 and Var(Y ) = E(Y 2 ) = 1/2 + 1/2 = 1. And
clearly P(|Y | > 2) = 0 < 1/4.
46
|Corr(X, Y )| =
In Exercise 4.5.11 we prove this result for a more specific example involving independent standard normals (and linear combinations of them). Here we generalize
the result.
47
Exercise 5.5.7. Let a R, and let (x) = max(x, a) as in Exercise 4.5.2. Prove
that is a convex function. Relate this to Jensens inequality and to Exercise
4.5.2.
Proof. Let a R, and let (x) = max(x, a). Now let x, y, R with 0 6 6 1.
Then
(x + (1 )y) = max(x + (1 )y, a + (1 )a)
6 max(x, a) + max((1 )y, (1 )a)
= (x) + (1 )(y).
Therefore is a convex function. Jensens inequality then says that E((X)) >
(E(X)), which proves Exercise 4.5.2.
Exercise 5.5.8. Let (x) = x2 .
(a) Prove that is a convex function.
(b) What does Jensens inequality say for this choice of ?
(c) Where in the text have we already seen the result of part (b)?
Proof. Let (x) = x2 .
(a) Let x, y, R with 0 6 6 1. First we note that
x2 = y 2 + 2y(x y) + (x y)2 > y 2 + 2y(x y) = y(2x y).
This implies that
(x + (1 )y) = (x + (1 )y)2
= 2 x2 + 2(1 )xy + (1 )2 y 2
= 2 x2 + 2(1 )xy + (1 )y 2 (1 )y 2
= 2 x2 + (1 )y 2 + (1 )y(2x y)
6 2 x2 + (1 )y 2 + (1 )x2
= x2 + (1 )y 2
= (x) + (1 (y).
Therefore is a convex function.
(b) Jensens inequality then says that E(X 2 ) 6 E(X)2 .
(c) This result is seen on page 44 when we defined variance and showed its nonnegativity. That is Var(X) = E(X 2 ) E(X)2 > 0.
48
f 0 (y) =
v
.
If y < v/ we see that f 0 (y) < 0 and if y > v/, we see that f 0 (y) > 0. Therefore
y is a global minimum. Thus we have
v + y2
( + y)2
v + (v/)2
6
( + v/)2
2 + v
=v 2
( + v)2
v
=
.
v + 2
P(X m > ) 6
49
50
51
1 3 n1
(3
+ 1) = .
2
n
(d) Since (1/n)Rn , for all > 0, we have P( n1 |Sn | > ) 1. Thus for all
> 0, it is the case that P( n1 |Sn | > ) 6 0.
(e) For the first version of the weak law of large numbers we must have that {Xn }
has uniformly bounded variance. However, Var(Xn ) = E(Xn2 ) = 32n .
For the first version of the strong law of large numbers we must have that
{Xn } has uniformly bounded fourth moments. However E(Xn4 ) = 34n .
For the second version of both the strong and weak law of large numbers
we must have that {Xn } are identically distributed. However, since E(Xn2 )
depends on n, they are not identically distributed. Therefore the hypotheses
arent satisfied.
6.2 Exercises
Exercise 6.2.4. Why does Proposition 6.2.1 not imply that Var(X) equals the
corresponding linear combination of variances?
Proof. The function f which defined the variance is dependent on the random variable used, i.e. f (x) = (x X )2 . Var(X) does indeed equal the linear combination
of variances using this function, but these are not the variances of the individual
distributions. That is, we want the variance of X with respect to the various distributions, not merely the variances of the distributions. Indeed the following is
true: Let f (x) = (x X )2 and L(X) = 41 1 + 14 2 + 21 N := . Then
Z
Var(X) =
f (X)(dt)
ZR
=
(X X )2 (dt)
R
Z
Z
Z
1
1
1
=
(X X )2 1 (dt) +
(X X )2 2 (dt) +
(X X )2 N (dt)
4 R
4 R
2 R
1 1
1 25
1 25
=
+
+
4 16
4 16
2 16
19
=
.
16
52
6.3 Exercises
Exercise 6.3.1. Let have density 4x3 10<x<1 , and let have density 21 x10<x<2 .
(a) Compute E(X) where L(X) = 31 + 23 .
(b) Compute E(Y 2 ) where L(Y ) = 16 + 13 2 + 12 5 .
(c) Compute E(Z 3 ) where L(Z) = 81 + 18 + 14 3 + 12 4 .
Proof. Let have density 4x3 10<x<1 , and let have density 12 x10<x<2 .
(a) Let L(X) = 31 + 23 . Then
E(X) =
=
=
=
=
2
1
E (X) + E (X)
3Z
3
Z
1
2
1
3
t 4t 10<t<1 (dt) +
t t10<t<2 (dt)
3 R
3 R 2
Z 1
Z 2
4
1
t4 dt +
t2 dt
3 0
3 0
4
8
+
15 9
52
.
45
1
1
1
E (Y 2 ) + E2 (Y 2 ) + E5 (Y 2 )
6Z
3
2 Z
Z
1
1
1
2
3
2
t 4t 10<t<1 (dt) +
t 2 (dt) +
t2 5 (dt)
6 R
3 R
2 R
Z
2 1 5
1
1
t dt + 22 + 52
3 0
3
2
1 4 25
+ +
9 3
2
251
.
18
53
1
1
1
1
E (Z 3 ) + E (Z 3 ) + E3 (Z 3 ) + E4 (Z 3 )
8Z
8
4 Z
2
Z
Z
1
1
1
1
3
3
3 1
t 4t 10<t<1 (dt) +
t t10<t<2 (dt) +
t3 3 (dt) +
t3 4 (dt)
8 R
8 R
2
4 R
2 R
Z
Z 2
1 1 6
1
1
1
t dt +
t4 dt + 33 + 43
2 0
16 0
4
2
1
2 27 64
+ +
+
14 5
4
2
5491
.
140
Exercise 6.3.2. Suppose P(Z = 0) = P(Z = 1) = 21 , that Y N (0, 1), and that
Y and Z are independent. Set X = Y Z. What is the law of X?
Proof. Suppose P(Z = 0) = P(Z = 1) = 21 , that Y N (0, 1), and that Y
and Z are independent. Set X = Y Z and let N = L(Y ). Let B B. Then
X (B) = P(X B) = P(Y Z B).
Suppose 0 6 B, then 0 (B) = 0 and X (B) = P(Y Z B) = P((Z = 1) (Y
B)) = P(Z = 1)P(Y B) = 21 N (B) = 21 0 (B) + 12 N (B). Suppose 0 B, then
X (B) = P[(Z = 0) ((Z = 1) (Y B))] = P(Z = 0) + P(Z = 1)P(Y B) =
1
1
2 0 (B) + 2 N (B).
That is L(X) = 21 0 + 21 N .
Exercise 6.3.3. Let X Poisson(5).
(a) Compute E(X) and Var(X).
(b) Compute E(3X ).
Proof. Let X Poisson(5). Then X =
54
j=0 (e
5 j
5 /j!)j .
X
=
(e5 5j /j!)
tj (dt)
R
j=1
= e5
= 5e
j=1
X
5
j1
= 5e5
= 5e
j5j /j!
j=1
/(j 1)!
5k /k!
k=0
5 5
= 5.
Next we calculate the variance using the formula Var(X) = E(X 2 ) E(X)2 .
First we have that:
Z
E(X 2 ) =
t2 X (dt)
R
(e5 5j /j!)
Z
R
j=1
= e5
= 5e
t2 j (dt)
j 2 5j /j!
j=1
X
5
j=1
X
X
= 5e5
5j1 /(j 1)! +
(j 1)5j1 /(j 1)!
j=1
j=1
X
X
= 5e5
5j1 /(j 1)! + 5
5j2 /(j 2)!
j=1
= 5e
j=2
k
5 /(k)! + 5
k=0
5
X
k=0
= 5e5 (e + 5e5 )
= 5(1 + 5)
= 30.
55
!
k
5 /(k)!
Then finally
Var(X) = E(X 2 ) E(X)2
= 30 52
= 5.
(b)
E(3X ) =
=
3t X (dt)
(e5 5j /j!)
Z
R
j=1
= e5
3t j (dt)
3j 5j /j!
j=1
= e5
15j /j!
j=1
= e5 e15
= e10 .
Exercise 6.3.4.
given by
56
1
2
t2 1 (dt) +
1
= +
2
1
= +
2
= 1.
1
2
1
2
1
2
t2 (dt)
[0,1]
t2 dt
E(X 2 ) =
Exercise 6.3.5. Let X and Z be independent, with X N (0, 1), and with
P(Z = 1) = P(Z = 1) = 1/2. Let Y = XZ (i.e., Y is the product of X and Z).
(a) Prove that Y N (0, 1).
(b) Prove that P(|X| = |Y |) = 1.
(c) Prove that X and Y are not independent.
(d) Prove that Cov(X, Y ) = 0.
(e) It is sometimes claimed that if X and Y are normally distributed random
variables with Cov(X, Y ) = 0, then X and Y must be independent. Is that
claim correct?
Proof. Let X and Z be independent, with X N (0, 1), and with P(Z = 1) =
2
P(Z = 1) = 1/2. Let Y = XZ. Let (t) = 12 et /2 .
(a) Let L(X) := N and L(Y ) := Y . Let B B. Since (t) is an even function,
we have for all x R, that P(X 6 x) = P(X > x) = P(X 6 x). Thus
by Proposition 6.0.2, X N (0, 1) so that N (B) = P(X B). We have
that
Y (B) = P(XZ B)
= P (Z = 1) (X B) (Z = 1) (X B)
= P (Z = 1) (X B) + P (Z = 1) (X B)
= P(Z = 1)P(X B) + P(Z = 1)P(X B)
1
1
= N (B) + N (B)
2
2
= N (B).
Therefore Y N (0, 1).
(b) Clearly
P(|Z| = 1) = P (Z = 1) (Z = 1)
= P(Z = 1) + P(Z = 1)
= 1.
Thus
P(|X| = |Y |) = P(|X| = |X||Z|)
= P (|X| = |X|) (|Z| = 1)
= P(|Z| = 1)
= 1.
58
(c) Suppose on the contrary that X and Y are indepedent. Then by Proposition
3.2.3, X 2 and Y 2 are indepdent so that E(X 2 Y 2 ) = E(X 2 )E(Y 2 ). However, since X and Z are independent, we have E(X 2 Y 2 ) = E(X 4 Z 2 ) =
E(X 4 )E(Z 2 ) = E(X 4 ) = 3, while, E(X 2 )E(Y 2 ) = E(Y 2 ) = E(X 2 Z 2 ) =
E(X 2 )E(Z 2 ) = 1. Thus E(X 2 Y 2 ) 6= E(X 2 )E(Y 2 ) which is a contradiction.
Therefore X and Y are not independent.
(d) Since X and Z are independent, then by Proposition 3.2.3, X 2 and Z are also
independent. Since X N (0, 1) and Y N (0, 1) we have E(X) = E(Y ) =
0. Also E(Z) = 21 12 = 0. Thus E(XY ) = E(X 2 Z) = E(X 2 )E(Z) =
1 0 = 0. We also have E(X)E(Y ) = 0. Therefore Cov(X, Y ) = E(XY )
E(X)E(Y ) = 0 0 = 0.
(e) The claim is not correct. We have shown in parts (a)-(d), that X and Y
are normally distributed random variables with Cov(X, Y ) = 0, but are not
independent.
59
(a) We have that limy%x FX (y) = limy&x FX (y) = FX (x) if and only if FX is
continuous at x.
(b) Since FX is monotonic and bounded, we have that limy%x FX (y) exists. Let
L = limy%x FX (y). Let n N, then by additivity we have
FX (x) FX (x 1/n) = P(x 1/n < X 6 x).
Then
lim FX (x)FX (x1/n) = FX (x)L = lim P(x1/n < X 6 x) = P(X = x).
7.2 Exercises
Exercise 7.2.4. Verify that (7.2.2) and (7.2.3) satisfy (7.2.1), and also satisfy
s(0) = 0 and s(c) = 1.
Proof. First we verify (7.2.2) satisfies (7.2.1) and also satisfies s(0) = 0 and s(c) = 1.
60
Since q + p = 1, we have
q
qs(a 1) + ps(a + 1) =
qa
pa1
c
q
p
q a+1
pa
c
q
p
a+1
q
pa1
q pa
c
q
p
pq a +q a+1
pa
c
pq
q a (p+q)
pa
c
1 pq
a
1 pq
c =
1 pq
s(a).
Also
1
s(0) =
1
=
0
q
p
c
q
p
11
c = 0,
1 pq
and
1
s(c) =
1
c
q
p
c = 1.
q
p
qs(a 1) + ps(a + 1) =
Also
s(0) =
0
= 0,
c
61
and
s(c) =
c
= 1.
c
Pa1 q i
i=0
s(1).
(d) Use the fact that s(c) = 1 to solve for s(1), and verify (7.2.2) and (7.2.3).
Proof.
(a) Suppose that s(a) = qs(a 1) + ps(a + 1) whenever 1 6 a 6 c 1. This implies
1
that s(a + 1) = (s(a) qs(a 1)). Subtracting s(a) from both sides we obtain
p
s(a + 1) s(a) =
=
=
=
=
1
(s(a) qs(a 1)) s(a)
p
1
(s(a) qs(a 1) ps(a))
p
1
((1 p)s(a) qs(a 1))
p
1
(qs(a) qs(a 1))
p
q
(s(a) s(a 1)) .
p
(b) Suppose part (a) is true. If a = 0, then since s(0) = 0 we have s(0 + 1) s(0) =
0
q
s(1). If a = 1, then
s(1) =
p
q
(s(1) s(0))
p
1
q
=
s(1)
p
s(1 + 1) s(1) =
62
s(a + 1) s(a) =
q
s(1). Therefore
p
k+1
q
s(1)
p
a
q
=
s(1).
p
a
q
Therefore for 0 6 a 6 c 1 we have s(a + 1) s(a) =
s(1).
p
s(a + 1) s(a) =
(c) Suppose 0 6 a 6 c. Then by parts (a) and (b) we have for each 0 6 k 6 a, that
k1
q
s(1). Therefore
s(k) s(k 1) =
p
s(a) = s(a) s(0)
a
X
=
(s(k) s(k 1))
k=1
a
X
k=1
a1
X
k=0
q
p
k1
q
p
k
s(1)
s(1).
c1 k
X
q
k=0
= s(1)
s(1)
c1 k
X
q
k=0
= 1.
63
Therefore s(1) = 1/
Pc1 q k
k=0
k
q
k=0
p
s(a) =
Pc1 q k
k=0
p
c
a
p
p
1
1
q
q
=
p /
p
1
1
q
q
a
p
1
q
c
=
p
1
q
Pa1
Exercise 7.2.6. Solve equation (7.2.1) using the theory of difference equations,
as follows.
(a) Show that the corresponding characteristic equation t0 = qt1 + pt1 has two
distinct roots t1 and t2 when p 6= 1/2, and one double root t3 when p = 1/2.
Solve for t1 , t2 and t3 .
(b) When p 6= 1/2, the theory of difference equations says that we must have
sc,p (a) = C1 (t1 )a + C2 (t2 )a for some constants C1 and C2 . Assuming this,
using the boundary conditions sc,p (0) = 0 and sc,p (c) = 1 to solve for C1 and
C2 . Verify (7.2.2).
(c) When p = 1/2, the theory of difference equations says that we must have
sc,p (a) = C3 (t3 )a + C4 a(t3 )a for some constants C3 and C4 . Assuming this,
use the boundary conditions to solve for C3 and C4 . Verify (7.2.3).
64
Proof.
(a) Suppose t0 = qt1 + pt1 . Then pt2 t + q = 0. If p = q = 1/2, then multplying
through by 2, we have t2 2t + 1 = 0 so that (t 1)2 = 0. Therefore we have a
double root t3 = 1. If p 6= 1/2, then note that
1 4pq = 1 4p(1 p)
= 1 4p + 4p2
= 4(p2 p + 1/4)
= 4(p 1/2)2 .
so that pq < 1/4 and we have by the quadratic formula, distinct roots
1 1 4pq
t1,2 =
2p
p
1 4(p 1/2)2
=
2p
1 |2p 1|
.
=
2p
Assume WLOG p > 1/2, then |2p 1| = 2p 1 and
t1,2 =
1 (2p 1)
,
2p
Exercise 7.4.1. For the stochastic process {Xn } given by (7.0.1), compute (for
n, k > 0)
(a) P(Xn = k).
(b) P(k = n).
[Hint: These two questions do not have the same answer.]
Proof. Consider the stochastic process {Xn } given by (7.0.1).
(a) If n < k, then Xn < k so that
P(Xn = k) = 0. Suppose n > k. Then
n
k
.
2n
(b) If n < k, then Xn < k so that P(k = n) = 0. Suppose n > k. Then
P(k = n) = P(inf{m > 0; Xm = k} = n) = P(Xn = k and Xn1 = k 1) =
n1
n1
k1
2n
66