HW 1 Sol
HW 1 Sol
HW 1 Sol
Homework Set #1
Properties of Entropy and Mutual Information
1. Entropy of functions of a random variable.
Let X be a discrete random variable. Show that the entropy of a
function of X is less than or equal to the entropy of X by justifying
the following steps:
(a)
= H(X).
(c)
H(g(X)).
X
0
1
1
1
3
1
3
1
3
Find
(a) H(X), H(Y ).
(b) H(X|Y ), H(Y |X).
(c) H(X, Y ).
(d) H(Y ) H(Y |X).
(e) I(X; Y ).
3. True or False questions
Copy each relation and write true or false. Then, if its true, prove it.
If it is false give a counterexample or prove that the opposite is true.
(a) H(X) H(X|Y )
(b) H(X) + H(Y ) H(X, Y )
(c) Let X, Y be two independent random variables. Then
H(X Y ) H(X).
4. Solution to True or False questions e.
(a) H(X) H(X|Y ) is true. Proof: In the class we showed that
I(X; Y ) > 0, hence H(X) H(X|Y ) > 0.
(b) H(X) + H(Y ) H(X, Y ) is false. Actually the opposite is true,
i.e., H(X) + H(Y ) H(X, Y ) since I(X; Y ) = H(X) + H(Y )
H(X, Y ) 0.
(c) Let X, Y be two independent random variables. Then
H(X Y ) H(X).
True
(a)
(b)
5. Bytes.
P
The entropy, Ha (X) = p(x) loga p(x) is expressed in bits if the
logarithm is to the base 2 and in bytes if the logarithm is to the base
256. What is the relationship of H2 (X) to H256 (X)?
Solution: Bytes.
H2 (X) =
p(x)
where (a) comes from the property of logarithms and (b) follows from
the definition of H256 (X). Hence we get
H2 (X) = 8H256 (X).
Solution: Example of joint entropy
(a) H(X) =
2
3
6. Two looks.
Here is a statement about pairwise independence and joint independence. Let X, Y1 , and Y2 be binary random variables. If I(X; Y1) = 0
and I(X; Y2 ) = 0, does it follow that I(X; Y1, Y2 ) = 0?
3
7. A measure of correlation.
Let X1 and X2 be identically distributed, but not necessarily independent. Let
H(X2 |X1 )
=1
.
H(X1 )
(a) Show =
I(X1 ;X2 )
.
H(X1 )
(b) Show 0 1.
(c) When is = 0?
(d) When is = 1?
Solution: A measure of correlation.
X1 and X2 are identically distributed and
=1
H(X2 |X1 )
H(X1 )
(a)
H(X1 ) H(X2 |X1 )
H(X1 )
H(X2 ) H(X2 |X1 )
=
(since H(X1 ) = H(X2 ))
H(X1 )
I(X1 ; X2 )
=
.
H(X1)
H(X2 |X1 )
1
H(X1 )
0 1.