Decision Surface

For a classifier where the likelihood for each class is a multivariate Gaussian distribution, the decision surfaces are generally hyperquadrics. A hyperquadric is a generalization of conic sections (like ellipses, parabolas, and hyperbolas) to higher dimensions. This means the boundaries separating the classes can be hyperplanes, pairs of hyperplanes, hyperspheres, hyperellipsoids, or hyperparaboloids.

Discriminant Function

Assuming a minimum-error-rate classification and Gaussian likelihoods , the logarithmic discriminant function is:

Expanding the Gaussian term, we get:

The term is a quadratic form in , which is why the resulting decision boundaries are hyperquadrics. The term is a constant across all classes and can be ignored.

What happens to the decision surface of a Gaussian classifier in the special case where all covariance matrices are equal and diagonal, i.e., ?

Special Case: Covariance matrices are equal and diagonal

When the covariance matrix for every class is assumed to be , it means the features are statistically independent and have the same variance for all classes. In this case, the quadratic term in the discriminant function simplifies, and the decision surface becomes a hyperplane. The discriminant function simplifies to:

where is the squared Euclidean distance. The quadratic terms cancel out when comparing and , leaving a linear equation in , which defines a hyperplane.

Special Case: Covariance matrices are equal and diagonal AND a flat prior?

If we have the special case of equal diagonal covariances () and a flat prior ( for all classes), the discriminant function simplifies even further to:

Maximizing this is equivalent to minimizing the Euclidean distance . Therefore, the classifier simply assigns a feature vector to the class with the nearest mean . This is known as a minimum Euclidean distance classifier.