December 2, 2020
December 2, 2020
General recommendations.
• These lectures assume that the audience is familiar with measure theory.
• The videos do not replace the books. I suggest to choose one among the
many listed at the end of these notes and to read the corresponding sections
before or after the videos.
• After the statement of a result, interrupt the video and try to prove the
assertion. It is the only way to understand the difficulty of the problem, to
differentiate simple steps from crucial ones, and to appreciate the ingenuity
of the solution. Sometimes you find an alternative proof of the result.
• You can speed-up or slow-down the video. By pressing settings at the
bottom-right corner, you can modify the playback speed.
• Exercises highlighted in blue present results which will be used later in the
lectures and are highly recommended, as well as those indicated with ∗.
• Send me an e-mail if you find a mistake which is not reported in these notes.
• If you typed in latex, with no personal definitions nor the use of special
packages, solutions to some exercises proposed below, send the file. Hope-
fully, I’ll create a note with solutions to the exercises, acknowledging the
authors of the solutions.
• A note about the methodology. I ask the students to view the video(s)
before the class. In the first part of the lecture, I recall the content of the
video. Sometimes, I ask one of the students to replace me. Occasionally,
the student is randomly chosen. This is the opportunity for the students to
ask questions on the content of the class. In the second part of the lecture, I
present some of the applications included in the “Further Readings” topic.
December 2, 2020
1
2
Lecture 1: Introduction
This
R is [Chung,RTheorem 3.2.1].
37:05 Ω f (X) dP = R f (x) µX (dx), [Chung, Theorem 3.2.2].
40:32 Jensen’s inequality, [Chung, Section 3.2]. See comment below on convex
functions.
47:28 Cebyshev’s inequality, [Chung, Section 3.2].
On convex functions. Let I be an open interval of R (which may be equal to R).
Consider a real-valued convex function ϕ : I → R. Show that ϕ is continuous on I
and that it has left and right-derivatives at every point. Denote by (D+ ϕ)(x) the
right-derivative of ϕ at x. Show that for all x0 ∈ I, (D+ ϕ)(x0 ) (x − x0 ) + ϕ(x0 ) ≤
ϕ(x). This bound is used in the proof of Jensen’s inequality.
Further Readings.
A. [Varadhan, Chapter 1] presents a review on measure theory.
B. [Breiman, Chapter 2] examines the properties of the distribution functions
of random vectors and presents Kolmogorov’s extension theorem. Be aware
that Breiman defined the distribution function as FX (x) = P [X < x], while
we adopt in these lectures the convention that FX (x) = P [X ≤ x].
C. [Durrett, Sections 1.1 – 1.3] has many examples.
Recommended exercises.
a. In Section 3.1 of [Chung], prove Theorems 1 to 6 (that is Theorems 3.1.1
to 3.1.6).
b. Section 3.1 of [Chung], exercises 3, 4, 5, 11.
c. In Section 3.2 of [Chung], prove Theorems 2 and 3.
d. Section 3.2 of [Chung], exercises 2, 5, 6, 7, 11, 12, 13, 14, 16, 19
Suggested exercises.
a. Section 3.1 of [Chung], exercises 6, 10
b. Section 3.2 of [Chung], exercises 1, 4, 8, 10, 15, 17, 18
3
Lecture 2: Independence
Content.
0:00 Definition of independent random variables
4:47 Subfamilies of independent random variables are independent
8:12 Definition of distribution function and probability measure of a random
vector
12:43 Lemma. A finite set of random variables is independent if and only if the
distribution function of the random vector is equal to the product of the
distribution functions.
18:15 Lemma. A finite set of random variables is independent if and only if the
probability measure of the random vector is equal to the product of the
probability measures.
19:12 Theorem. Let X1 , . . . , XN be independent random variables and f1 , . . . , fN ,
fj : R → R, measurable functions. Then, f1 (X1 ), . . . , fN (XN ) are indepen-
dent random variables.
22:52 Theorem. Let X1 , . . . , XN be independent random variables, n0 = 0,
1 ≤ n1 < n2 < · · · < np = N and f1 , . . . , fp , fj : Rnj −nj−1 → R, mea-
surable functions. Then, f1 (X1 , . . . , Xn1 ), . . . , fp (Xnp−1 +1 , · · · , Xnp ) are
independent random variables.
26:19 Theorem. Let X, Y be independent random variables such that E[ |X| ] <
R∞, E[ |Y | ] < ∞.
R Then, E[ X Y ] = E[ X ] E[ Y ].
43:17 Ω f (X) dP = R f (x) µX (dx).
45:53 Second proof of the identity E[ X Y ] = E[ X ] E[ Y ].
52:03 Construction of a product measure on an infinite product space.
Summary. The first two applications of this lecture can be found in Section 1.5
of [Durrett], the last one in Chapter 3 of [Varadhan-LD].
Content and Comments.
0:00 Weak law of large numbers.
8:06 Convergence in Lp implies convergence in probability.
14:25 Bernstein polynomials approximate uniformly continuous functions.
33:57 An upper bound for large deviations. This result is known as Cramer’s
theorem.
Further readings.
A. Interesting examples (coupon collector, random permutation, occupancy
problems, St. Petersburg paradox) can be found in Section 1.5 of [Durrett].
B. More details on large deviations of i.i.d. random variables can be found in
[Varadhan-LD, Deuschel-Stroock, Dembo-Zeitouni].
Suggested exercises.
a. Exercises in Section 1.5 of [Durrett].
6
Summary. This lecture is based on [Chung, Section 4.2] and [Durrett, Section
1.6].
Content and Comments.
0:00 Definition of lim supn En , lim inf n En
1:50 P [lim sup En ] = lim P [∪m≥n Em ]
n n
5:14 lim sup En = { En i. o. } := {ω : ω ∈ En i. o. }
n X
15:07 [Chung, Theorem 4.2.1]. P [En ] < ∞ ⇒ P [ En i. o. ] = 0.
n≥1
17:28 Application. [Durrett, Theorem 1.6.5] Xn i.i.d., E[X14 ] < ∞ ⇒ (X1 + · · · +
Xn )/n → E[X1 ] a. e. X
29:46 [Chung, Theorem 4.2.4]. (En : n ≥ 1), independent, P [En ] = ∞ ⇒
n≥1
P [En i. o. ] = 1.
35:21 Remark: (En : n ≥ 1), independent.
P Then, P [En i. o. ] = 1 or 0. Morever,
P [En i. o. ] = 1 if and only if n≥1 P [En ] = ∞.
37:05 Application. [Durrett, Theorem 1.6.7]. Xn i.i.d., E[ |X1 | ] = ∞ ⇒ P [ |Xn | ≥
n i. o. ] = 1.
38:57 We are using here [Chung, Theorem 3.2.1] (cf. Lecture 1, time 0:00)
39:39 In particular, P lim(X1 + · · · + Xn )/n exists and belongs to R = 0.
n
40:36 To prove that the set {ω ∈ Ω : limn (X1 +· · ·+Xn )/n exists and belongs to R}
is an element of the σ-algebra F, recall that this set corresponds to the set
{ω ∈ Ω : (X1 + · · · + Xn )/n is a Cauchy sequence }.
50:45 Thus, the hypothesis E[ |X1 | ] < ∞ is needed for a strong law of large
numbers for i.i.d random variables.
Further readings.
A. [Chung, Theorem 4.2.5]. Only pairwise independence is needed in [Chung,
Theorem 4.2.4]
B. [Durrett, Examples 1.6.2 and 1.6.3] [If Rn represents the number of records
up to time n, Rn / log n → 1 a. e.] and [if Ln represents the size of largest
sequence of consecutive 1’s in a Bernoulli sequence, Ln / log2 n → 1 a. e.]
C. [Breiman, Propositions 3.16-3.18] investigates the number of returns to the
origin in a coin-tossing problem.
Recommended exercises.
a. [Chung, Section 4.2] 2, 5, 6, 7, 10, 12, 14, 16,
b. [Durrett, Section 1.6] Exercises 2 – 8, 10
c. [Breiman, Section 3.3] Problems 6 [we presented a proof of this result earlier.
Use Borel-Cantelli to derive a second proof], 7, 9, 10
Suggested exercises.
a. [Chung, Section 4.2] 1, 3, 4, 8, 9, 11, 13, 15, 18, 19, 20
b. [Durrett, Section 1.6] Exercises 12, 14, 17, 18
8
Summary. This lecture is based on [Breiman, Sections 8.1 and 8.2] and [Varadhan,
Section 2.3].
Content and Comments.
0:00 Definition of weak convergence and convergence in distribution.
8:55 The space of distributions N and the space of generalized distributions M.
13:38 [Breiman, Theorem 8.6] Helly’s selection Theorem, This result correspond
to steps 1, 2, 3 of [Varadhan, Theorem 2.4]
49:18 Examples where the limit is a generalized distribution and not a distribu-
tion.
51:54 Uniqueness o limit points yields convergence. [Breiman, Corollary 8.8]
1:00:18 Tightness of probability measures [Breiman, Definition 8.9]
1:05:52 A set of distribution funtions {Fα : α ∈ I} is tight if and only if the following
statement holds [Fαn → G ⇒ G ∈ N ]. [Breiman, Proposition 8.10]
1:21:11 Let me clarify. By hypothesis, Fαn (n) − Fαn (−n) ≤ 1 − for all n ≥ 1. We
(1)
introduced a subsequence αn . This means that αn1 = αp(n) , where p(n) ≥
n and p(n + 1) > p(n). By definition of the sequence p(n) and the above
inequality, Fα(1) (p(n)) − Fα(1) (−p(n)) = Fαp(n) (p(n)) − Fαp(n) (−p(n)) ≤
n n
1 − . Hence, instead of writing Fα(1) (n) − Fα(1) (−n), I should have written
n n
Fαp(n) (p(n)) − Fαp(n) (−p(n)). This is what I meant when saying that n is
(1)
the one which corresponds to the subsequence αn .
Further readings.
A. [Chung, Sections 4.3 and 4.4] provides examples and an alternative view on
convergence in distribution.
B. [Durrett, Section 2.2] presents many examples of sequences of random vari-
ables which converge in distribution.
Recommended exercises.
a. Prove [Chung, Theorems 4.3.1 and 4.3.2]
b. [Breiman, Chapter 8] Problems 1, 2, 4, 5, 6, 7, 9, 10.
c. [Durrett, Section 2.2] Exercise 6
Suggested exercises.
a. [Breiman, Chapter 8] Problems 3, 8, 11.
b. [Chung, Section 4.3] Exercises 3, 8
c. [Durrett, Section 2.2] Exercises 2, 3, 7
9
Summary. This lecture is based on [Breiman, Section 8.3], [Chung, Section 4.4]
and [Varadhan, Section 2.3].
Content and Comments.
0:00 Helly-Bray’s theorem, [Breiman, Proposition 8.12] presents a stronger ver-
sion of the first part of the theorem, and [Varadhan, Theorem 2.3].
32:26 Convergence in probability implies convergence in distribution. [Chung,
Theorem 4.4.5].
41:18 Convergence in distribution to a Dirac mass implies convergence in proba-
bility. [Chung, Section 4.4, Exercise 4].
d d d d
46:15 Xn → X and Yn → y ∈ R ⇒ Xn + Yn → X + y and Xn Yn → Xy. [Chung,
Theorem 4.4.6].
Further readings.
A. [Breiman, Section 8.3] presents a stronger version of Helly-Bray’s theorem
[continuity is replaced by the hypothesis that the set of discontinuity points
has measure 0 for the limiting probability measure.
B. [Chung, Theorem 4.4.1] states that to prove convergence in distribution it
is enough to show that E[f (Xn )] → E[f (X)] for all continuous functions f
with compact support. The corollary of [Chung, Theorem 4.4.6] provides
necessary and sufficient conditions for convergence in distribution in terms
of open and closed sets.
Recommended exercises.
*a. Let f : R → R be a bounded function which is continuous everywhere,
except at a finite number of points, represented by D = {x1 , . . . , xp }. Let
µn be a sequence R of probability
R measures which converges weakly to µ. If
µ(D) = 0, then f dµn → f dµ.
b. [Breiman, Section 8.3], problems 13, 14
c. [Chung, Section 4.4], exercises 1, 4, 6, 7, 9, 11
d. [Varadhan, Chapter 2], exercises 9, 10, 11
e. Prove [Breiman, Propositions 8.12, 8.15, 8.17, 8.19]
Remark 0.1. The assertion of the recommended exercise [a] will be used several
times in the next lectures.
Suggested exercises.
a. [Chung, Section 4.4], exercises 2, 3, 8, 10, 12
10
Summary. This lecture is based on [Breiman, Sections 8.6, 8.7, 8.9 and 8.11] and
on [Varadhan, Section 3.6].
Content and Comments.
0:00 Lévy continuity theorem. [Breiman, Theorem 8.28].
7:22 Expansion of the characteristic function. [Breiman, Proposition 8.44].
27:30 The characteristic function of the sum of two independent random variables.
[Breiman, Proposition 8.33].
30:32 Application: The central limit theorem for i.i.d. random variables with
finite second moments. This is [Varadhan, Theorem 3.17]. See [Breiman,
Theorem 8.20] for another proof of this result. We use here the expansion
of log(1 + z) for z ∈ C. See [Breiman, Proposition 8.46].
Further readings.
A. Read [Breiman, Sections 8.12] and [Varadhan, Section 2.2]. The starting
question is: does there exist two distributions with the same moments or
does the convergence of moments entail the convergence in distribution.
B. Read [Breiman, Sections 8.13]. It is proved there that the Laplace trans-
forms characterize the distribution of positive random variables.
Recommended exercises.
*a. Let (Xn : n ≥ 1) be a sequence of random variables which converges in
distribution to X. Denote by ϕn (t), ϕ(t) the characteristic functions of
Xn , X, respectively. Show that ϕn converges to ϕ uniformly on bounded
intervals. This is [Breiman, Proposition 8.31] and will be used many times
below.
b. Prove [Breiman, Propositions 8.30, 8.33, 8.37]
c. [Breiman, Chapter 8], exercises 16, 17 (see [Varadhan, Theorem 2.7]), 21
d. Prove [Varadhan, Theorem 2.6]
e. [Chung, Section 6.4], exercises 4, 7, 11, 24
Suggested exercises.
a. [Chung, Section 6.3], exercises 6, 8.
b. Prove [Chung, Theorem 6.4.6]
c. [Chung, Section 6.4], exercise 6.
12
Summary. This lecture is based on [Chung, Section and 5.5] and [Durrett-4th,
Sections 2.4].
Content and Comments.
0:00 Shannon’s entropy. This is [Durrett-4th, Example 2.4.3]
7:41 Renewal process. I followed [Chung, Section and 5.5]. This is also [Durrett-4th,
Example 2.4.1].
18
if and only if
lim sup max ϕn,j (t) − 1 = 0 for all T > 0.
n→∞ |t|≤T 1≤j≤n
Further readings.
A. A different approach to the one-dimensional central limit problem is pre-
sented in [Breiman, Section 9.5 – 9.7].
23
obtained at time [37:00] holds for all |t| ≤ t0 and all n large, say n ≥ n0 .
By changing the value of the constant C0 this inequality can be extended
to 1 ≤ n ≤ n0 . This explains why (0.1) is in force for all n ≥ 1.
38:31 Step 2: Condition (0.1) implies (a) and (b).
48:34 Actually, we proved the following result. Let (Zn,j : 1 ≤ j ≤ kn ) be an
array of independent random variables such that kn → ∞. Assume that
for all T > 0, there exists a finite constant CT such that
Xkn h i
E { 1 − cos(tZn,j ) } ≤ CT (0.2)
j=1
Recommended exercise.
(*a) Assume that for all T > 0, there exists a finite constant CT such that
kn
X h i
E { 1 − cos(tX
en,j ) } ≤ CT
j=1
for all |t| ≤ T and n ≥ 1. Show that for all δ > 0, there exists a finite
P
constant C0 such that j P [ | X
en,j | ≥ δ ] ≤ Cδ for all n ≥ 1.
25
Claim: Condition (0.1) implies that for every m > 0 there exists a finite constant
P e2 χ e
Cm such that j E[ X n,j |Xn,j |≤m ] ≤ Cm . Note that 1 has been replaced by m in
the indicator function.
Proof. Since 1 − cos(x) ≥ 0 for all x ∈ R, it follows from (0.1) that
X h i
E { 1 − cos(tX en,j ) }χ e
|tXn,j |≤π/4 ≤ CT
j
for all |t| ≤ T . Since there exists a > 0 such that 1 − cos(x) ≥ ax2 for |x| ≤ π/4,
X h i
a t2 E X 2
en,j χ|tXen,j |≤π/4 ≤ CT
j
as claimed
26
Recommended exercises.
a. [Varadhan, Section 3.8], exercises 21, 22, 23, 24, 25.
28
for all |t| ≤ T and n ≥ 1. This has been proved in Lecture 20A, see equation
(0.1). It follows from this bound (see Lecture 20A, time [38:31] and exercise
(*a)) that there exists a finite constant C0 such that
Z Z
(a) Mn (dx) ≤ C0 and (b) x2 dMn ≤ C0
|x|≥δ |x|≤1
for all n ≥ 1,
16:04 Step A3: For all > 0, there exists A > 0 such that Mn ([−A, A]c ) ≤
for all n ≥ 1, A ≥ A . I refer here to the result proved at time [44:28] in
Lecture 8.
2 2
23:01 Step A4:R Let ω(x) = x /(1 + x ). Then, there exists C0 < ∞ such that
αn = ω(x) Mn (dx) ≤ C0 for all n ≥ 1.
26:42 Step B: Claim: Given a subsequence (nk : k ≥ 1), there exist a sub-
subsequence (nkj : j ≥ 1) and a Lévy measure M such that
Z Z Z x0 Z x0
2
f (x) Mnkj (dx) → f (x) M (dx) and x Mnkj (dx) → x2 M (dx)
−x0 −x0
for all f ∈ Cb∗ (R) and x0 > 0 such that M ({−x0 , x0 }) = 0. Here, Cb∗ (R) is
the set of bounded, continuous functions f : R → R for which there exists
δ > 0 such that f (x) = 0 for all |x| ≤ δ.
29:10 Step B1: Assume that αn → 0. Then, the claim formulated in Step B holds
and M = 0.
37:00 Step B2: Assume that αn → α > 0. Then, the claim formulated in Step B
holds.
52:57 Step C: Claim: Given a subsequence (nk : k ≥ 1), there exist a sub-
subsequence (nkj : j ≥ 1) and σ 2 ≥ 0 such that σn2 k → σ 2 . In particular,
j
given a subsequence (nk : k ≥ 1), there exist a sub-subsequence (nkj : j ≥
1) and a triple (M, σ 2 , 0) such that (Mnkj , σn2 k , 0) → (M, σ 2 , 0).
j
56:56 Step D: Lemma: Assume that Xn → X and that Xn + an → Y in distri-
bution. Then an → a for some a and Y = X + a in distribution.
1:01:32 Step E: The sequence ankj converges to some a ∈ R.
1:03:58 Conclusion of the tightness part of the proof: Given a subsequence (nk :
k ≥ 1), there exist a sub-subsequence (nkj : j ≥ 1) and a triple (M, σ 2 , a)
such that (Mnkj , σn2 k , ankj ) → (M, σ 2 , a).
j
1:05:15 Uniqueness of limits.
30
Further readings.
A. The introduction of the measures dνn = [x2 /(1 + x2 )] dMn in the previous
proof is taken from [Breiman, Theorem 9.17].
31
0 Under the previous hypotheses, the mean µ and variances σ 2 of S are given
by
kn
X h i
σ 2 = lim E (Xn,j − an,j )2 χ|Xn,j −an,j |≤x0 (0.6)
n→∞
j=1
kn n
X o
µ = lim an,j + E θ(Xn,j − an,j )
n→∞
j=1
Note that in the first line the limit does not depend on x0 because PMn2→ 0.
f
2 2 2
0 Assume, further, that E[Xn,j ] = 0, σn,j = E[Xn,j ] < ∞, σn = j σn,j →
σ 2 and S ∼ N (0, σ 2 ). Then, Lindeberg’s condition holds. This (re)proves
the assertion that Lindeberg’s condition are not only sufficient for conver-
gence to aP Gaussian random variable, but also necessary.
0 Claim 1: j an,j → 0. This claim is not necessary for the argument. For
this reason
P skipped
I its proof in the lecture. It is presented below.
2
0 Claim 2: a
j n,j → 0.
0 Conclusion: Lindeberg’s condition holds. For all δ > 0,
kn
X 2
lim E Xn,j χ|Xn,j |>δ = 0
n→∞
j=1
P
Proof that j an,j → 0. Since µ = 0,
kn n
X o
lim an,j + E θ(Xn,j − an,j ) = 0
n→∞
j=1
32
Fix 0 < δ < 1/2. By (0.5) and since θ is a bounded function, we may introduce the
indicator of |Xn,j | ≤ δ inside the expectation to get that
kn n
X o
lim an,j + E θ(Xn,j − an,j ) χ|Xn,j |≤δ = 0 (0.7)
n→∞
j=1
where the value of the constant C0 may change from line to line. The last sum is
bounded by the sum appearing in (0.6) provided we choose x0 sufficiently large.
This proves the claim.
It follows from this claim and (0.7) that there exists a finite constant C0 such
that
Xkn n
o
lim sup an,j + E (Xn,j − an,j ) χ|Xn,j |≤δ ≤ C0 δ .
n→∞
j=1
Note that the function θ disappeared. Since maxj an,j is bounded, (0.5) and the
previous equation yield that
Xkn
lim sup E Xn,j χ|Xn,j |≤δ ≤ C0 δ .
n→∞
j=1
By (0.5) again,
Xkn
lim sup E Xn,j χδ<|Xn,j |≤1 = 0 .
n→∞
j=1
Hence,
Xkn
lim sup E Xn,j χ|Xn,j |≤1 ≤ C0 δ .
n→∞
j=1
To complete the proof, it remains to recall that an,j = E[ Xn,j χ|Xn,j |≤1 ] and to let
δ → 0.
33
17:11 Let G = σ(Aj : j ≥ 1). Exercise: show that B belongs to G if and only if
B = ∪j∈M Aj for some M ⊂ N.
18:31 A function h : Ω → R is measurable with respect to G if and only if it is
constant on each set Aj . Proof that it is measurable if it is constant.
26:18 Proof that it is constant
P if it is measurable.
33:44 Let E[ f | G ] = n≥1 E[ f | An ] χAn . Then, E[ f | G ] is G-measurable.
37:35 For every bounded function h, measurable with respect to G,
Z Z
E[ f | G ] h dP = f h dP
B
43:27 Step 1: For every set B measurable with respect to G,
Z Z
E[ f | G ] dP = f dP
B B
48:17 Step 2: Extension to bounded functions, measurable with respect to G.
51:07 Summary of the properties of E[ f | G ] in the case where G = σ(Aj : j ≥ 1).
56:44 Definition of the conditional expectation E[ f | G ] for an integrable function
f ∈ F and a σ-algebra G ⊂ F.
58:30 Existence and uniqueness of the conditional expectation.
1:04:06 Uniquenes of the conditional expectation.
1:06:20 The conditional expectation E[ f | G ] is integrable.
1:10:56 If f is G-measurable, then E[ f | G ] = f .
1:12:32 Let f be an integrable function. Then, f can be decomposed as f = f1 +f2 ,
where f1 is G-measurable and f2 ⊥ G in the sense that E[ f h ] = 0 for all
bounded functions h which are G-measurable.
1:17:10 Definition of E[ f | X ] and of P [ A | G ].
34
Further readings.
A. The proof of the Radon-Nykodim theorem can be found in [Taylor, Section
6.4] or in my lectures on measure theory.
B. [Varadhan, Section 4.1] provides a short proof of Radon-Nykodim theorem,
which relies on the Hahn-Jordan decomposition of a measure.
C. [Durrett-4th, Section 5.1], examples 1 – 6
D. [Breiman, Section 4.1] proposes a slightly different approach to conditional
expectation.
Recommended exercises.
a. Show that E E[ Z | G ] = E[Z] if Z is integrable.
b. [Varadhan, Section 4.1], exercises 1 – 8 recall important facts from mea-
sure theory which are used in the proofs of the properties of conditional
expectation.
c. [Chung, Section 9.1], exercises 1 – 6
Suggested exercises.
a. [Breiman, Section 4.1], problems 2, 3, 5, 7, 8.
35
Further readings.
A. [Varadhan, Section 4.2], specially Remark 4.5.
B. [Breiman, Section 4.2]
C. [Durrett-4th, Section 5-1]
Exercises a, b, c and d below are strongly recommended.
Recommended exercises.
*a. [Varadhan, Section 4.2], exercise 9
*b. Prove [Durrett-4th, Section 5-1], Theorems 4 and 8
*c. [Durrett-4th, Section 5-1], exercises 8, 9
*d. Prove [Breiman, Section 4.2] Proposition 4.20.(4)
e. [Chung, Section 9.1], exercises 7, 8, 9, 12
f. [Durrett-4th, Section 5-1], exercises 3, 4, 6
Suggested exercises.
36
Suggested exercises.
a. [Varadhan, Section 4.3], exercise 11
b. [Durrett-4th, Section 5-1], exercises 13, 15.
c. [Breiman, Section 4.2], problem 17
38
References
[Breiman] L. Breiman, Probability. Corrected reprint of the 1968 original. Classics in Applied
Mathematics, 7. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA,
1992. xiv+421 pp. ISBN: 0-89871-296-3
[Chung] K. L. Chung, A course in Probability Theory. Third edition. Academic Press, Inc., San
Diego, CA, 2001. xviii+419 pp. ISBN: 0-12-174151-6
[Dembo-Zeitouni] A. Dembo, O. Zeitouni, Large deviations techniques and applications. Cor-
rected reprint of the second (1998) edition. Stochastic Modelling and Applied Probability,
38. Springer-Verlag, Berlin, 2010. ISBN: 978-3-642-03310-0
[Deuschel-Stroock] J-D. Deuschel, D. W. Stroock, Large deviations. Pure and Applied Mathemat-
ics, 137. Academic Press, Inc., Boston, MA, 1989. ISBN: 0-12-213150-9
[Durrett] R. Durrett, Probability: Theory and Examples. Second edition. Duxbury Press, Bel-
mont, CA, 1996. xiii+503 pp. ISBN: 0-534-24318-5
[Durrett-4th] R. Durrett, Probability: Theory and Examples. Fourth edition. Cambridge Series in
Statistical and Probabilistic Mathematics, 31. Cambridge University Press, Cambridge, 2010.
ISBN: 978-0-521-76539-8
[Rudin] W. Rudin, Principles of Mathematical Analysis. Third edition. McGraw-Hill, Inc. New
York (1964).
[Taylor] S. J. Taylor, Introduction to Measure and Integration, Cambridge University Press, 1973.
ISBN 978-0-521-09804-5
[Varadhan-LD] Large deviations and applications. CBMS-NSF Regional Conference Series in Ap-
plied Mathematics, 46. Society for Industrial and Applied Mathematics (SIAM), Philadelphia,
PA, 1984. ISBN: 0-89871-189-4
[Varadhan] S. R. S. Varadhan, Probability Theory. Courant Lecture Notes in Mathematics, 7.
New York University, Courant Institute of Mathematical Sciences, New York; American
Mathematical Society, Providence, RI, 2001. ISBN: 0-8218-2852-5