probasim
probasim
probasim
• Binomial theorem
𝑛
𝑛
(𝑥 + 𝑦)𝑛 = ∑( )𝑥𝑛−𝑗 𝑦𝑗 (3)
𝑗=0
𝑗
1.1. Counting
• Number of permutations of 𝑘 elements:
‣ Number of ways to order 𝑘 elements
‣ Only order matters
𝑃𝑘 = 𝑘! (4)
1
Exercises
Exercise 1.1 (Handshakes and kisses)
There are 𝑓 girls and 𝑔 boys in a room. Boys exchange handshakes, girs exchange kisses, boys
and girls exchange kisses. How many kisses in total?
The number of kisses exchanegs among girls is the number of subsets of cardinality 2 of a set of
𝑓(𝑓−1)
cardinality 𝑓, that is ( 𝑓2 ) = 2 . Or, think that each girl gives 𝑓 − 1 kisses, and one needs a factor
of one half to avoid double counting.
For the number of kisses exchanged between boys and girls: the first girl gives 𝑔 kisses, the second
girl gives 𝑔 kisses, and so on, so we have 𝑓𝑔 in total.
𝑓(𝑓 − 1)
number of kisses = + 𝑓𝑔 (7)
2
Exercise 1.2 (Throwing a dice) Throw a fair dice with 𝑓 faces 𝑛 times. What’s the prob to never
get the same result twice?
with cardinality
𝑓!
𝑃 (𝐴) = (12)
𝑓 𝑛 (𝑓− 𝑛)!
Exercise 1.3 (Birthday paradox) What is the probability that at least 2 people out of 𝑛 have the
same birthday? (Assume: uniform birth probability and year with 𝑦 number of days).
Quick solution
𝑃 (𝐴) = 1 − 𝑃 ⎛
⎜no two people have the same birthday⎞
⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟ ⎟
⎝ Ex. 2 ⎠ (13)
𝑦!
=1− 𝑛
𝑦 (𝑦 − 𝑛)!
Formal solution Let 𝒩 = {1, …, 𝑛} and 𝒴 = {1, …, 𝑦} with 𝑛 ≤ 𝑦. The jet set is
2
Ω = distributions of possible birthdays of 𝑛 people
(14)
= {𝜔 = (𝜔1 , …, 𝜔𝑛 ) : 𝜔𝑖 ∈ 𝒴 for all 𝑖 ∈ 𝒩} = 𝒴𝑛
where 𝜔𝑖 is the birthday of the 𝑖-th person. The cardinality of the jet set is
𝐴 = {𝜔 ∈ Ω : ∃𝑖 ≠ 𝑗 ∈ 𝒩 : 𝜔𝑖 = 𝜔𝑗 } (16)
Note that this is the complementary event to the event defined in Equation 10 of Exercise 2. Thus we
can compute its probability as
𝑃 (𝐴) = 1 − 𝑃 (𝐴𝑐 ) (17)
Exercise 1.4 (Same birthday as the prof) What is the probability that at least 1 student out of 𝑛
has the same birthday of the prof? (Assume: uniform birth probability and year with 𝑦 number
of days).
Quick solution
Formal solution 1 As above 𝒩 = {1, …, 𝑛} and 𝒴 = {1, …, 𝑦} with 𝑛 ≤ 𝑦. The jet set is Ω = 𝒴𝑛+1 ,
that is the set of possible birthdays of 𝑛 + 1 people, the (𝑛 + 1)-th being the prof. Its cardinality is
|Ω| = 𝑦𝑛+1 . The event we’re looking at is
𝐴 = {𝜔 ∈ Ω : ∃𝑖 ∈ 𝒩 : 𝜔𝑖 = 𝜔𝑛+1 } (19)
3
𝐴𝑐 = {𝜔 ∈ Ω : 𝜔𝑖 ≠ 𝜔𝑛+1 ∀𝑖 ∈ 𝒩} (20)
𝑐
|𝐴 |
As usual 𝑃 (𝐴) = 1 − 𝑃 (𝐴𝑐 ) = 1 − |Ω|
, with
(𝑦 − 1)𝑛
|𝐴𝑐 | = ⏟𝑦 ⋅ ⏟
(21)
prof students
𝑦(𝑦−1)𝑛 𝑛
So, 𝑃 (𝐴) = 1 − 𝑦𝑛+1
= 1 − ( 𝑦−1
𝑦
) , in agreement with Equation 18.
Formal solution 2 Using the probability of the complementary event is often the smartest way to
proceed, but for the sake of completeness let’s see how to get the same result directly. Let 𝐴𝑗 be the
event “exactly 𝑗 students out of 𝑛 have the same birthday as the prof”. The event we look at then is
𝐴 = ⊔𝑗∈𝒩 𝐴𝑗 (22)
The cardinality of 𝐴𝑗 is
𝑛
|𝐴𝑗 | = ⏟
1…1 ⋅ (𝑦
⏟⏟ −⏟⏟⏟
1)…(𝑦 −
⏟⏟1) ⋅ ⏟𝑦 ⋅ ( )
⏟𝑗
𝑗 times 𝑛−𝑗 times prof
number of ways to choose 𝑗 elements out of 𝑛 (24)
𝑛
= 𝑦(𝑦 − 1)𝑛−𝑗 ( )
𝑗
4
𝑃 (𝐴 ∩ 𝐵) = 𝑃 (𝐵 ∩ 𝐴) ⇒ 𝑃 (𝐴 | 𝐵)𝑃 (𝐵) = 𝑃 (𝐵 | 𝐴)𝑃 (𝐴) (30)
Exercises
Exercise 2.1 (Pile ou Face) Jet de 2 pieces, Ω = {PP, PF, FP, FF}. Cet espace est muni de la
probabilite uniforme. Soient les evenements:
• 𝐴 = 1ere piece donne P
• 𝐵 = 2eme piece donne F
• 𝐶 = le deux pieces donnent le meme resultat
Questions:
• 𝐴 et 𝐵 sont indépendantes?
• 𝐴, 𝐵 et 𝐶 sont indépendants?
Solution 1 Let 𝐴𝑖 be the event the i-th drawn piece is good, with 𝑖 ∈ {1, 2}. We need the probability
of the event 𝐴2 ∩ 𝐴1 . By definition of conditional probavility,
1
𝑃 (𝐴2 ∩ 𝐴1 ) = 𝑃
⏟⏟(𝐴⏟
2 |⏟
𝐴⏟ 𝑃 (𝐴1 ) = .
1 )⏟
3 (33)
5 6
9 10
Solution 2 The jet set is the set of subsets of cardinality 2 of a set of cardinality 10, so |Ω| = ( 10
2
).
6
The event we consider is the set of subsets of cardinality 2 of a set of cardinality 6, so |𝐴| = ( 2 ).
Then
5
( 62 ) 6⋅5 1
𝑃 (𝐴) = = = . (34)
( 10
2
) 10 ⋅ 9 3
Intuitive solution The cards are 𝑅𝑅, 𝑅𝑊 , 𝑊 𝑊 with 𝑊 for white and 𝑅 for red. Call 𝑅𝑅 the
“red” card, 𝑊 𝑊 the “white” card, and 𝑊 𝑅 the “mixed” card. Since we observe a red face, the white
card cannot be on the table. There are three possibilities left: 1. we’re observing a face of the red card
(in which case the hidden face is red); 2. we are observing the other face of the red card (in which
case the hidden face is red); 3. we are observing the red face of the mixed card (in which case the
hidden face is white). So the hidden face is red 2 out of 3 times.
Formal solution The jet set contains the possible outcomes of a sequence of two events: 1. draw a
card (out of three), and 2. observe a face (out of two). Denote by 𝑅 a red face and by 𝑊 a white face,
and denote by a subscript 𝑜 the observed face, and by a subscript ℎ the hidden face. The possible
outcomes then are
Ω = {𝑅ℎ ∩ 𝑅𝑜 , 𝑅ℎ ∩ 𝑊𝑜 , 𝑊ℎ ∩ 𝑅𝑜 , 𝑊ℎ ∩ 𝑊𝑜 } (35)
where the first entry indicates the hidden face, and the second entry indicates the observed face. For
example, 𝑊ℎ ∩ 𝑅𝑜 is the event “the hidden face is white and the observed face is red”, and similarly for
the others.
In this formulation, every element in the jet set is the intersection of two (dependent) events of the
type 1. a face is hidden, and 2. a face is observed. Note that the event 𝑊ℎ ∩ 𝑅𝑜 is equivalent to the
event “the mixed card is drawn, and the red face is observed.” Under this second point of view, each
outcome in Ω is the intersection of two (dependent) events of the type 1. a card is drawn, and 2. a
face is observed. Denoting the event “draw the red card” by 𝑟, the event “draw the white card” by 𝑤,
and the event “draw the mixed card” by 𝑚, the jet set is equivalently
Ω = {𝑟 ∩ 𝑅𝑜 , 𝑚 ∩ 𝑊𝑜 , 𝑚 ∩ 𝑅𝑜 , 𝑤 ∩ 𝑊𝑜 } (36)
This formulation helps to understand that the probability on Ω is not uniform. The probabilities of
the events in Ω are computed by Equation 27:
𝑃 (𝑟 | 𝑅𝑜 )
𝑃 (𝑅ℎ ∩ 𝑅𝑜 ) = 𝑃 (𝑟 ∩ 𝑅𝑜 ) = (37)
𝑅𝑜
However, we do not know the probabilities on the right hand side. As a simple trick, remember that
𝑃 (𝐴 ∩ 𝐵) = 𝑃 (𝐵 ∩ 𝐴), so we can turn this around:
𝑃 (𝑅ℎ ∩ 𝑅𝑜 ) = 𝑃 (𝑅𝑜 ∩ 𝑟)
2
=𝑃 (𝑅 |⏟ 𝑃 (𝑟) =
𝑟)⏟ (38)
⏟⏟ ⏟ 𝑜 ⏟
6
1 1
3
6
𝑃 (𝑅ℎ ∩ 𝑊𝑜 ) = 𝑃 (𝑊𝑜 ∩ 𝑚)
1
=𝑃 (𝑊𝑜 |⏟
𝑚) 𝑃 (𝑚) = (39)
⏟⏟ ⏟ ⏟⏟ 6
1 1
2 3
𝑃 (𝑊ℎ ∩ 𝑅𝑜 ) = 𝑃 (𝑅𝑜 ∩ 𝑚)
1
=𝑃 (𝑅 𝑜 |⏟
𝑚) 𝑃 (𝑚) = (40)
⏟⏟ ⏟ ⏟⏟ 6
1 1
2 3
𝑃 (𝑊ℎ ∩ 𝑊𝑜 ) = 𝑃 (𝑊𝑜 ∩ 𝑤)
2
=𝑃 (𝑊 𝑜 |⏟𝑤) 𝑃 (𝑤) = (41)
⏟⏟ ⏟ ⏟⏟ 6
1 1
3
𝑃 (𝑅ℎ ∩ 𝑅𝑜 )
𝑃 (𝑅ℎ | 𝑅𝑜 ) =
𝑃 (𝑅𝑜 )
𝑃 (𝑅ℎ ∩ 𝑅𝑜 ) 2 (43)
= =
𝑃 (𝑅ℎ ∩ 𝑅𝑜 ) + 𝑃 (𝑊ℎ ∩ 𝑅𝑜 ) 3
= 1 − 𝑃 (𝑊ℎ | 𝑅𝑜 )
where the last line follows from Equation 28 and gives directly the answer. So in conclusion, given
the fact that we observe a red face, the hidden face is also red with probability 2/3.
Exercise 2.4 (Russian roulette) You are playing two-person Russian roulette with a revolver
featuring a rotating cylinder with six bullet slots. Each time the gun is triggered, the cylinder
rotates by one slot. Two bullets are inserted one next to the other into the cylinder, which is then
randomly positioned. Your opponent is the first to place the revolver against her temple. She
presses the trigger and… she stays alive. With great display of honor, she offers you to rotate the
barrel again at random before firing in turn. What do you decide?
Say the revolver shots from position 1. The event “the first player dies” is
die1 = {12, 61} (45)
1
so 𝑃 (die1 ) = 3
and 𝑃 (live1 ) = 23 . We need to compute
𝑃 (die2 ∩ live1 )
𝑃 (die2 | live1 ) = (46)
𝑃 (live1 )
7
Since the cylynder rotates after being triggered we have die2 = {56, 61} and die2 ∩ live1 = {56}, so
𝑃 (die2 | live1 ) = 16 / 23 = 14 < 𝑃 (die1 ). So you don’t shuffle the barrel.
such that
1. 𝜇(∅) = 0
2. countable additivity (also called 𝜎-additivity) is fulfilled, namely the measure of a disjoint
countable union of sets in 𝒜 is the sum of their measures:
∞ ∞
𝜇(⨆ 𝐴𝑖 ) = ∑ 𝜇(𝐴𝑖 ). (48)
𝑖=1 𝑖=1
Definition 3.1.3 (Probability measure) Given a set Ω and a 𝜎-algebra 𝒜 on Ω, a probaility measure
𝑃 is a measure (in the sense above) with the additional requirement that
𝑃 (Ω) = 1. (49)
• Note that this implies that 𝑃 (𝐴) ≤ 1 for all events 𝐴 ∈ 𝒜.
• A triple (Ω, 𝒜, 𝑃 ) where 𝒜 is a 𝜎-algebra on Ω and 𝑃 is a probability measure is called probability
space.
¹In french, this set is called tribu on Ω. The term 𝜎-algebra is also used – and is more common in the context of pure
analysis, c.f. [3] – whereas the term 𝜎-field is more common in the context of probabioloty theory, c.f. [1].
8
• the second property guarantees that 𝑃 {𝑋 = 𝑥} is well-defined for all 𝑥 ∈ 𝐹 , which allows for the
following definition:
Definition 3.2.2 (Distribution of a discrete random variable) The distribution (or law) of a random
variable 𝑋 is the function 𝜇 : 𝐹 → [0, 1] defined by
𝜇(𝑥) = 𝑃 {𝑋 = 𝑥} for all 𝑥 ∈ 𝐹 . (51)
• two discrete random variables 𝑋 and 𝑌 taking values resp. in 𝐹 and 𝐺 are independent if
𝑃 {𝑋 = 𝑥, 𝑌 = 𝑦} = 𝑃 {𝑋 = 𝑥}𝑃 {𝑌 = 𝑦} for all 𝑥 ∈ 𝐹 , 𝑦 ∈ 𝐺 (52)
• it is understood that {𝑋 = 𝑥, 𝑌 = 𝑦} is a shorthand for the event
{𝜔 ∈ Ω : 𝑋(𝜔) = 𝑥} ∩ {𝜔 ∈ Ω : 𝑌 (𝜔) = 𝑦} ∈ 𝒜. (53)
• the definition generalises to collections of DRVs, see Section 2.2.3 in [2].
Take-away
• (Ω, 𝒜, 𝑃 ) with 𝑃 : 𝒜 → [0, 1] and 𝑃 (Ω) = 1
• 𝑋 : Ω → 𝐹 countable, with {𝑋 = 𝑥} ∈ 𝒜 for all 𝑥 ∈ 𝐹
• 𝜇 : 𝐹 → [0, 1] such that 𝜇(𝑥) = 𝑃 {𝑋 = 𝑥}
9
𝜇 : 𝐹 → [0, 1]
𝜆𝑛 (56)
𝑛 ⟼ 𝑒−𝜆
𝑛!
Exercises
Exercise 3.1 (Sum of independent binomial distributions) Let 𝑋𝑖 ∼ ℬ(𝑛𝑖 , 𝑝) with 𝑖 ∈ {1, 2} be
independent discrete random variables following the Bernoulli law. Find the law of 𝑋1 + 𝑋2 .
⎛
⎜ ⎞
⎟
= 𝑃⎜
⎜ ⨆ {𝑋1 = 𝑘1 , 𝑋2 = 𝑘2 }⎟
⎟
⎜ 𝑘 ∈𝐹 ⎟
𝑖 𝑖
⎝ 1 2
𝑘 +𝑘 =𝑘 ⎠
= ∑ 𝑃 (𝑋1 = 𝑘1 )𝑃 (𝑋2 = 𝑘2 ) by c. add and indep.
𝑘𝑖 ∈𝐹𝑖
𝑘1 +𝑘2 =𝑘
𝑛1 𝑛
= ∑ ( )( 2 )𝑝𝑘1 +𝑘2 (1 − 𝑝)𝑛1 +𝑛2 −𝑘1 −𝑘2
𝑘𝑖 ∈𝐹𝑖
𝑘1 𝑘2
𝑘1 +𝑘2 =𝑘
𝑛1 𝑛
= 𝑝𝑘 (1 − 𝑝)𝑛1 +𝑛2 −𝑘 ∑ ( )( 2 )
𝑘𝑖 ∈𝐹𝑖
𝑘1 𝑘2
𝑘1 +𝑘2 =𝑘
Let’s focus on the sum. For each fixed 𝑘1 ∈ 𝐹1 , 𝑘2 is constrained to be 𝑘 − 𝑘1 . Furthermore, in order
for 𝑘2 to be ≥ 0, 𝑘1 can be at most equal to 𝑘. So the constraints
10
𝑘1 ∈ {0, …, 𝑛1 }
𝑘2 ∈ {0, …, 𝑛2 } (61)
𝑘1 + 𝑘2 = 𝑘
namely
𝑘
𝑛1 𝑛 𝑛 𝑛2 𝑛 + 𝑛2
∑ ( )( 2 ) = ∑ ( 1 )( )=( 1 )
𝑘1 𝑘2 𝑘1 𝑘 − 𝑘1 𝑘 (63)
𝑘𝑖 ∈𝐹𝑖 𝑘 =01
𝑘1 +𝑘2 =𝑘
since each summand in the the second sum is zero², and we get again Equation 63.
So in conclusion
𝑛1 + 𝑛2 𝑘
𝜇(𝑘) = ( )𝑝 (1 − 𝑝)𝑛1 +𝑛2 −𝑘 (65)
𝑘
namely 𝑋1 + 𝑋2 ∼ ℬ(𝑛1 + 𝑛2 , 𝑝).
Exercise 3.2 (Sum of independent Poisson distributions) Let 𝑋𝑖 ∼ 𝒫(𝜆𝑖 ) with 𝑖 ∈ {1, 2} be
independent discrete random variables following the Poisson law. Find the law of 𝑋1 + 𝑋2 .
11
𝑛 𝑛 𝑛−𝑛
𝜆1 1 𝜆2 1 𝑛!
𝜇(𝑛) = 𝑒−(𝜆1 +𝜆2 ) ∑ ⋅
𝑛1
𝑛 ! 𝑛 − 𝑛1 ! 𝑛!
=0 1
(𝜆1 + 𝜆2 )𝑛
= 𝑒−(𝜆1 +𝜆2 )
𝑛!
So 𝑋1 + 𝑋2 ∼ 𝒫(𝜆1 + 𝜆2 ).
Exercise 3.3 (Min of independent geometric distributions) Let 𝑋𝑖 ∼ 𝒢(𝑝𝑖 ) with 𝑖 ∈ {1, 2} be
independent DRVs following the geometric law. Find the law of min{𝑋1 + 𝑋2 }.
Let’s drop the subscript for a moment. For a DRV 𝑋 ∼ 𝒢(𝑝) and for 𝑘 ∈ ℕ∗ we need
𝑃 (𝑋 ≥ 𝑘) = 𝑃 ( ⨆ (𝑋 = 𝑖))
𝑥≥𝑘
= ∑ 𝑃 (𝑋 = 𝑖)
𝑖≥𝑘
= ∑ 𝑝(1 − 𝑝)𝑖−1
𝑖≥𝑘
(69)
= 𝑝(1 − 𝑝)𝑘−1 + 𝑝(1 − 𝑝)𝑘 + 𝑝(1 − 𝑝)𝑘+1 + …
= 𝑝(1 − 𝑝)𝑘−1 (1 + (1 − 𝑝) + (1 − 𝑝)2 + …)
∞
= 𝑝(1 − 𝑝)𝑘−1 ∑ (1 − 𝑝)𝑗
𝑗=0
= (1 − 𝑝)𝑘−1
12
4. Theory recap 2.10.24
Recall from last week that
Take-away
• (Ω, 𝒜, 𝑃 ) with 𝑃 : 𝒜 → [0, 1] and 𝑃 (Ω) = 1
• 𝑋 : Ω → 𝐹 countable, with {𝑋 = 𝑥} ∈ 𝒜 for all 𝑥 ∈ 𝐹
• 𝜇 : 𝐹 → [0, 1] such that 𝜇(𝑥) = 𝑃 {𝑋 = 𝑥}
13
• The variance is a measure of the spreading, dispersion, of a random variable around its expected
value
The two following properties of the variance are very useful for concrete calculations:
Lemma 4.2.2 (Variance as difference of expectations)
Some examples:
• 𝑋 ∼ ℬ(𝑝) ⇒ Var(𝑋) = 𝑝(1 − 𝑝), see Exercise 4.3 and proof.
• 𝑋 ∼ ℬ(𝑛, 𝑝) ⇒ Var(𝑋) = 𝑛𝑝(1 − 𝑝) (immediate by Proposition 4.2.4)
• 𝑋 ∼ 𝒫(𝜆) ⇒ Var(𝑋) = 𝜆, see proof.
Exercises
Exercise 4.1 (Lost messages) On a telecommunication channel, it has been estimated that in 𝑇
time units there arrives a number of messages that can be estimated by a DRV ∼ 𝒫(𝜆𝑇 ). Each
message has a loss probability equal to 𝑝, independent of the other messages. Find the
probability that the number of lost message in 𝑇 units of time is equal to 𝑙.
Without loss of generality rescale 𝜆 ← 𝜆𝑇 . We need to find the discrete random variable 𝐿 whose
range {0, 1, 2, …} ∋ 𝑙 contains the possible numbers 𝑙 of lost messages in one time unit. The
probability 𝑃 (𝐿 = 𝑙) to lose 𝑙 message is then by definition by the law of 𝐿.
Let 𝑋𝑖 be the DRV for the event “the 𝑖-th message is lost”. Since each message is lost with probability
𝑝, 𝑋𝑖 ∼ ℬ(𝑝) for all 𝑖 ∈ {1, 2, …}.
Let 𝐿𝑎 = ∑𝑎𝑖=1 𝑋𝑖 be the DRV whose range Im(𝐿𝑎 ) = {0, 1, …, 𝑎} ∋ 𝑙 contains the numbers 𝑙 of
possible lost messages out of 𝑎 arrived. Since 𝐿𝑎 is the sum of 𝑎 independent 𝑝-Bernoulli DRVs, 𝐿𝑎
follows the binomial distribution:
𝐿𝑎 ∼ ℬ(𝑎, 𝑝). (79)
Finally, let 𝐴 be the DRV estimating the number of arrived messages 𝑎 ∈ {0, 1, …, } in one time unit;
we are given that 𝐴 ∼ 𝒫(𝜆).
The law of 𝐿 is given by
∞
𝑃 (𝐿 = 𝑙) = 𝑃 [⨆ {𝐿𝑎 = 𝑙 ∩ 𝐴 = 𝑎}], (80)
𝑎=𝑙
14
that is, we look at the disjoint union of all the events in which, given 𝑎 arrived messages, 𝑙 are lost.
By countable additivity and independence,
∞
𝑃 (𝐿 = 𝑙) = ∑ 𝑃 (𝐿𝑎 = 𝑙)𝑃 (𝐴 = 𝑎) (81)
𝑎=𝑙
1 ∞
(𝜆 − 𝜆𝑝)𝑗 (82)
= 𝜆𝑙 𝑝𝑙 𝑒−𝜆 ∑
𝑙! 𝑗=0
𝑗!
1
= 𝜆𝑙 𝑝𝑙 𝑒−𝜆 𝑒𝜆−𝜆𝑝
𝑙!
(𝜆𝑝)𝑙 −𝜆𝑝
= 𝑒 .
𝑙!
So 𝑉 ∼ 𝒫(𝜆𝑝).
1
Exercise 4.2 (Poisson expectation) Let 𝑁 ∼ 𝒫(𝜆). Find 𝔼(𝑋 ≔ 𝑁+1
)
.
By Theorem 4.1.3 we have
∞
1 −𝜆 𝜆𝑛
𝔼(𝑋) = ∑ 𝑒 (83)
𝑛=0
𝑛+1 𝑛!
1−𝑒−𝜆
Multiply and divide by 𝜆 and shift the running index to get 𝔼(𝑋) = 𝜆
.
Exercise 4.3 (Archery) An archer shots 𝑛 arrows at a target. The shots are independent, and
each shot hits the target with probability 𝑝. Let 𝑋 be the random variable “number of times the
target is hit”.
1. What is the law of 𝑋?
2. What is the expectation of 𝑋?
3. What is the value of 𝑝 that maximises the variance of 𝑋?
The archer bets on his result. He gets 𝑔 euros when he hits the target, and loses 𝑙 euros when he
misses it. Let 𝑌 be the random variable that represent the net gain of the archer at the end of the
𝑛 shots.
4. What is the expectation of 𝑌 ?
5. What is the relation between 𝑔 and 𝑙 that guarantees the archer an expected gain of zero?
15
𝔼(𝑋1 + … + 𝑋𝑛 ) = 𝔼(𝑋1 ) + … + 𝔼(𝑋𝑛 ) = 𝑛𝑝 (85)
For example, if 𝑝 = 0.5 and 𝑛 = 10, this means that the archer expects to hit the target 5 times.
3. Let’s compute the variance of a Bernoulli and a binomial variable by Equation 76:
𝔼(𝑋𝑖2 ) = 12 ⋅ 𝑝 + 02 ⋅ (1 − 𝑝) = 𝑝 (86)
For example if 𝑛 = 10, 𝑔 = 1, 𝑙 = 2, we have 𝔼(𝑌 ) = 30𝑝 − 20; and if furthermore 𝑝 = 0.5 then
𝔼(𝑌 ) = −5.
5. To find the value of relation between 𝑔 and 𝑙 required to have an expected gain of zero solve the
equation 𝔼(𝑌 ) = 0 to get
𝑔 1−𝑝
= . (89)
𝑙 𝑝
Thus as the probability 𝑝 to hit the target goes to zero, a very big 𝑔𝑙 is required to guarantee an
expected gain of zero; viceversa 𝑔𝑙 becomes infinitely small as 𝑝 → 1. At 𝑝 = 0.5, as one would
expect, 𝑔 = 𝑙.
16
∞ ∞
1 2𝑛 − 1
𝔼(𝑀 ) = ∑ 𝐵𝑛 ⋅ = ∑ = +∞. (91)
𝑛=1
2𝑛 𝑛=1 2𝑛
The idea is that the expected bet amount diverges even if one stops betting at some point.
2. This is for the standard doubling scenario, without the constraint bet iff all previous outcomes are the
same.
After 𝑛 bets the bet amount is 𝐵𝑛 = ∑𝑛𝑘=1 𝑏𝑘 = 2𝑛 − 1. If I win at the (𝑛 + 1)-th round the net win
equals the initial bet:
2𝑏𝑛 − 𝐵𝑛 = 2𝑛 − (2𝑛 − 1) = +1. (92)
Now the probability to lose 𝑛 times and win at the (𝑛 + 1)-th time is (1 − 𝑝)𝑛 𝑝, so the expected net
gain is also equal to the initial bet:
∞
𝑝
𝔼(Net gain) = ∑ (1 − 𝑝)𝑛 𝑝 ⋅ (+1) = = 1. (93)
𝑛=0
𝑝
Bibliography
[1] P. Billingsley, Probability and Measure. John Wiley & Sons, 2012.
[2] B. Jourdain, Probabilités et statistique pour l'ingénieur. 2018.
[3] E. H. Lieb and M. Loss, Analysis, 2nd ed. in Graduate Studies in Mathematics 14. American
Mathematical Society, 2001.
17