Maximum Likelihood (ML) Parameter Estimation
The maximum likelihood estimate finds parameters that maximize the likelihood of observing the training data:
In practice, we maximize the log-likelihood
To find the parameters that maximize the likelihood of observing the training data, one must find the maximum of the log-likelihood function, . This is typically done by taking the partial derivative with respect to each parameter, setting it to zero, and solving the resulting system of equations.
The formula is:
- : The parameter vector that maximizes the likelihood.
- : The log-likelihood function.
- : The training data vectors.
- : The individual parameters to be estimated.
- : Total training samples in the dataset
- : Total parameters in the model (that need to be estimated)
Note: This notation means “set equal to zero”. We set the derivative to zero to find the critical points (maxima, minima, or saddle points) of the log-likelihood function. In this context, we are looking for the maximum.
Why Parameter Estimation is Needed
In pattern recognition, we need parameter estimation to train classifiers by determining the unknown parameters of probability distributions from training data. They are:
- Prior probabilities
- Likelihood probability density functions
Assuming that the Likelihood can by represented by a Gaussian, we find mean and covariance such that it closely represents the Likelihood. The ML estimates for the parameters are estimated as follows:
-
Sample Mean (unbiased): This is the average of all the training vectors.
-
Sample Covariance (unbiased): This is the average of the outer products of the centered data vectors.
Note: Using a denominator of T-1 provides an unbiased estimate.