Max Likelihood
Max Likelihood
Max Likelihood
To find the decision surface between two different classes, the nature of
probability density function is very important, which is specified by the
parameter vector.
For ex., Let us take Gaussian distribution : P(x/wj) ~ N(, )
It is described by two parameters mean vector and covariance matrix
Depending upon the nature of this parameter vectors, we can have different
types of decision boundary between two different classes. ( linear or non linear)
That is, we shall assume that the parameters for the different classes are
functionally independent. This permits us to work with each class separately,
and thus we have C no.of separate problems.
Suppose that D contains n samples, x1, x2 ..., xn then, since the samples were
drawn independently, we have
The maximum likelihood estimate of P(D|) is the value that maximizes P(D|)
Instead of taking likelihood P(D|), we can take the logarithmic of P(D|) for
analysis
= (1 , 2, p )t
Where
Thus, a set of necessary conditions for the maximum likelihood estimate for can be
obtained from the set of p- equations
The above graph shows several training points in one dimension, assumed to be
drawn from a Gaussian of a particular variance, but unknown mean.
From the graph, we can observe that, If we have more number of training points, the
likelihood P(D|) will be very narrow and the confidence in estimating the
maximum is high. Hence our confidence in estimating the value that maximizes the
likelihood is marked depends on the no. of samples. If we have more no. of
samples, the estimation of is more accurate.