The University of Nottingham
The University of Nottingham
The University of Nottingham
FUNDAMENTALS OF STATISTICS
Candidates may complete the front cover of their answer book and sign their desk card but must
NOT write anything else until the start of the examination period is announced.
An indication is given of the approximate weighting of each section of a question by means of a figure
enclosed by square brackets, eg [12], immediately following that section.
Dictionaries are not allowed with one exception. Those whose first language is not
English may use a standard translation dictionary to translate between that language and
English provided that neither language is the subject of this examination. Subject specific
translation dictionaries are not permitted.
No electronic devices capable of storing and retrieving text, including electronic
dictionaries, may be used.
(i) Write down the name of the class of distribution of type (1.1).
(ii) State the Neyman factorisation theorem.
(iii) Suppose that X = (X1 ,..,Xn ) is an independent and identically distributed (IID)
∑
n
sample from (1.1). Show that T (X) = n−1 B(Xi ) is sufficient for θ.
i=1
(iv) By considering the ratio
pX (x|θ) ∏ n ∏ n
(b) Suppose X has distribution N (θ, c2 θ) where c > 0 is a known constant, θ > 0 is an un-
known parameter and N (µ, σ 2 ) is the normal distribution with mean µ and variance σ 2 .
G14FOS-E1
∼ 3 ∼ G14FOS-E1
2 (a) (i) Explain what is meant by a randomised hypothesis test. When might we wish to use
such a test?
(ii) Define (I) a simple hypothesis and (II) a composite hypothesis.
(iii) State the Neyman-Pearson Lemma.
(iv) Define what is meant by a Uniformly Most Powerful (UMP) test of size α for testing
H0 : θ = θ0 against a composite hypothesis H1 . [10]
(b) Consider a sample of size 1 (i.e. n = 1) from the distribution with probability mass function
pX (x|θ) = (1 − θ)θx , x = 0, 1, 2, ..., θ ∈ (0, 1).
(i) Consider testing the null hypothesis H0 : θ = 12 against the alternative hypothesis
H1 : θ > 12 .
Determine the UMP test of H0 against H1 of size α = 0.05.
(ii) Now consider the alternative hypothesis H2 : θ < 21 .
Determine the UMP test of H0 against H2 of size α = 0.05
(iii) Discuss, giving justification, whether or not a UMP test of H0 : θ = θ0 against
H3 : θ ̸= θ0 exists, where θ ̸= θ0 means θ < θ0 or θ > θ0 . [10]
(c) Suppose that θ̂ = (θ̂1 , .., θ̂p )T is an estimator of θ = (θ1 , .., θp )T such that, for large n,
√
n(θ̂ − θ) ≈ Np (0p , V ),
where 0p is the p-vector of zeros and V is a p × p covariance matrix.
( ) 12
∑p
(i) Consider the function d(θ) = ∥θ∥ = θj2 .
j=1
3 (a) (i) Explain what is meant by a family of conjugate prior distributions for a given probability
model pX (x|θ).
(ii) Consider a single observation from a distribution with probability density function
{ 1 3 −4 −β/x
β x e x>0
f (x|β) = 2
0 x ≤ 0.
Determine a conjugate family of prior distributions.
(iii) Calculate the Bayes factor for comparing the following models:
1
M0 : f (x|β = 1), π(M0 ) = ;
2
−β 1
M1 : f (x|β), π(β) = e (β > 0), π(M1 ) = .
2
(iv) Show that model M0 is always preferred when x = 31 and explain why this is a sensible
finding. [14]
(b) (i) Let x = (x1 , .., xn ) denote a random sample from the Uniform(0, θ) distribution, i.e.
the uniform distribution on the interval (0, θ), where θ > 0. Show, providing careful
justification, that the likelihood L(θ|x) is given by
1
L(θ|x) = I{θ: m<θ} (θ)
θn
where m = max1≤i≤n xi .
[Note: For a set A, the indicator function IA (θ) = 1 if θ ∈ A and IA (θ) = 0 otherwise.]
(ii) Suppose that θ in (b)(i) has prior density
{
(1 − c)θ−c θ ∈ (0, 1)
π(θ) =
0 θ∈/ (0, 1),
where c ∈ (0, 1) is a constant. Determine the posterior density π(θ|x).
(iii) For each α ∈ (0, 1), the posterior α−quantile of θ is defined by
∫ θα
π(θ|x)dθ = α.
m
Show that
m
θα = .
{1 − c + cmn+c−1 }1/(n+c−1)
(iv) Derive the 100(1 − α)% Highest Posterior Credible Region for θ. [16]
G14FOS-E1
∼ 5 ∼ G14FOS-E1
where µ ∈ (−∞, ∞), σ 2 > 0, a > 0, d > 0, γ > 0 and m ∈ (−∞, ∞).
(a) (i) The inverse gamma distribution IG(a, d) is the marginal distribution of σ 2 in the
N IG(a, d, m, v) distribution. Determine the pdf of IG(a, d).
(ii) Find an expression for p(µ), the marginal pdf of µ in the N IG(a, d, m, v) distribution.
[10]
(b) Let x1 , .., xn denote a random sample from the N (µ, σ 2 ) distribution, and suppose that
(µ, σ 2 ) has prior distribution N IG(a, d, m, v). Show that the posterior distribution of
(µ, σ 2 ) is of the form N IG(a1 , d1 , m1 , v1 ), where the quantities d1 , m1 , and v1 , should be
expressed in terms of a, d, m, v and the sample x1 , .., xn .
[Note: You are not required to find a1 .] [10]
2 2
(c) Now suppose y = (y1 , .., yn ) is a random sample from the N (µ, σ0 ) distribution where σ0
is a known quantity. Suppose that µ has an N (δ, τ 2 ) prior where δ and τ 2 > 0 are known
constants. If z is a new observation from the N (µ, σ02 ) distribution, find the predictive
distribution f (z|y). [10]
5 Let pX (x|θ) and L(d, θ) denote, respectively, a probability model and a loss function, where
x ∈ X , the sample space, θ ∈ Θ, the parameter space, and d ∈ D, the decision space. Let
δ : X → D denote a decision rule.
(a) (i) Define the risk R(δ, θ) of the decision rule δ.
(ii) Define what is meant by an admissible decision rule.
(iii) Given a prior distribution π(θ), define the Bayes risk R̄(δ, π).
(iv) Prove that the Bayes risk R̄(δ, π) is minimised over the set of decision rules by
δπ (x) = d∗ (x) where, for each x ∈ X , d∗ (x) minimises the expected posterior loss
∫
L(d, θ)π(θ|x)dθ. [14]
θ∈Θ
(b) (i) Let L(d, θ) = (d − θ)2 denote the quadratic loss function. Show that in this case the
posterior expected loss is minimised by the posterior mean of θ.
(ii) Let x = (x1 , .., xn ) denote a random sample from the Poisson (θ) distribution given by
θx e−θ
P (X = x|θ) = , x = 0, 1, 2, ... , θ > 0.
x!
Find the posterior distribution π(θ|x) when the Jeffreys prior is used for θ.
(iii) Find the estimator of θ that minimises the posterior expected quadratic loss.
(iv) Discuss whether the estimator of θ you found in (b)(iii) admissable. [16]
G14FOS-E1 End