MNPM
MNPM
MNPM
p(yi = j) = Fj (ij ) = PM
i=1 exp(ij )
where assuming ij = x0ij leads to the conditional and ij = x0ij to the multinomial Logit model.
yij
= ij + i
where the s are independent distributed according to a type I (Gumbel) extreme value distribution
and where the observable dependent variable yi is linked with its latent counterpart yi via:
(
j if yij
= max(yi1
, yi2
, . . . , yiM
)
yi =
0 otherwise
implies that the probability for choosing category j is given by:
exp(ij )
p(yi = j) = Fj (ij ) = PM
i=1 exp(ij )
In this sense these two model formulations are equivalent and lead to the same choice probabilities.
2. Multinomial Probit Model
Assume again the following model for the latent variable yij
:
yij
= x0ij + ij .
In the multinomial Probit model it is assumed that the is follows a multivariate normal distribution
with covariance matrix where now is not restricted to be a diagonal matrix.
11 . . . 1M
...
.
M N D(0, ), with = IN and = E(i0i) =
M 1 . . . M M
j if yij
= max(yi1
, yi2
, . . . , yiM
)
yi = j if yij =
0 otherwise.
x0i((j+1)
j ), . . . , (ij iM ) >
x0i(M
j ) .
Looking at this probability one can see that only the differences between the yij
s are identified and
hence a reference category has to be assigned as it was the case for the Logit model. As a consequence
the covariance matrix also reduced in its dimension from (M M ) to (M 1) (M 1).
...
i1
i(j1) i(j+1)
Z
...
(
1i, . . . , i(j1), i(j+1), . . . , iM )d
i1 . . . d
i(j1)d
i(j+1) . . . d
iM
iM
And here lies the practical obstacle with multinomial Probit models. There are no closed form
expressions for such high dimensional integrals and hence for M 3 one has to simulate them
using monte carlo simulation techniques.
I=
f (x)dx.
a
b
a
1
f (x)dx = E(f (x)).
ba
. . . f (x) can be treated as a random variable which is uniformly distributed in the interval [a, b]
and hence the right hand side of the above expression represents the expected value of f (x).
1 X
f (xi).
E(f (x))
D i=1
Multiplying this expression with (b a) finally we get an approximation for the integral I:
D
1 X
I (b a)
f (xi).
D i=1
7. Multinomial Probit Model
exdx.
The exact value for this integral is 4.67. In the following we approximated this integral for
D = 10, 20, 50 and 100 draws, respectively. The following table contains the so obtained values
and the absolute deviations from the true value. One can see that with more draws the approximation
gets better.
Draws
I approx.
I exact
Absolute deviation
10
4.58
4.67
0.09
20
4.38
4.67
0.29
50
4.79
4.67
0.12
100
4.75
4.67
0.08
Univ. Prof. Dr. Reinhard Hujer, University Frankfurt/M.
...
1
(j1) (j+1)
Z
...
(
1, . . . , (j1), (j+1), . . . , M )d
1 . . . d
(j1)d
(j+1) . . . d
M
Note that we have skipped the individual index for notational convenience. Such integrals reflect the
probabilities for choosing a certain category. The above integral e.g. is the probability for choosing
category j, i.e. p(yi = j|xi).
There are number of different simulators which can be used in this context. Examples include
the Accept-Reject Procedure, Importance Sampling, Gibbs Sampling and the Metropolis-Hastings
Algorithm. One algorithm which has been found to be fast and accurate is the Geweeke-HajivassiliouKeane smooth recursive simulator (GHK-simulator) which will be presented in the following.
9. Multinomial Probit Model
0
p(yi = j|xi) = p 1 >
11
2 121
0
p 2 >
>
1 11
22
0
3 131 232
2 121
>
p 3 >
,
>
.
1
2
33
11
22
The advantage of this expression is the fact that the s are independent normal distributed random
variables and hence the probability p(yi = j|xi) which we want to evaluate can be equivalently
expressed as a product of independent but conditioned univariate cumulative density functions.
0
p(yi = j|xi) = p 1 >
11
2 121
p 2 >
22
3 131 232
p 3 >
.
33
N
d
d
d
X
1
0
2 + 121
3 + 131 + 232
p(yi = j|xi) =
.
D i=1
11
22
33
All other probabilities can be calculated in an analogous way. Plugging them into the likelihood
function standard maximization procedures can be applied to get estimates for the parameters.