# Linear discriminant analysis

## Introduction

Linear discriminant analysis is a linear classification approach. The development of linear discriminant analysis follows along the same intuition as the naive Bayes classifier. It results in a different formulation from the use of multivariate Gaussian distribution for modeling conditional distributions.

## Prerequisites

To understand linear discriminant analysis, we recommend familiarity with the concepts in

• Probability: A sound understanding of conditional and marginal probabilities and Bayes Theorem is desirable.
• Gaussian distribution: The underlying class-conditional distribution for this model.
• Introduction to machine learning: An introduction to basic concepts in machine learning such as classification, training instances, features, and feature types.
• Naive Bayes classifier: For some understanding of the similarity of the motivation behind the model.

Follow the above links to first get acquainted with the corresponding concepts.

## Problem setting

In classification, the goal of the predictive model is to identify the class that generated a particular instance.

Consider such an instance $\vx \in \real^N$, a vector consisting of $N$ features, $\vx = [x_1, x_2, \ldots, x_N]$.

We need to assign it to one of the $M$ classes $C_1, C_2, \ldots, C_M$ depending on the values of the $N$ features .

In the probabilistic sense, we need to discover the probability of the instance belonging to one of these classes. That is, if we can calculate $P(C_m | \vx)$ for all the classes, we can assign the instance to the class with the highest probability. Thus, the predicted class will be

$$\hat{y} = \argmax_{m \in \set{1,\ldots,M}} P(C_m | \vx) \label{eqn:class-pred}$$

The conditional probability $P(C_m|\vx)$ for each class is computed using the Bayes rule.

$$P(C_m | \vx) = \frac{P(\vx | C_m) P(C_m)}{P(\vx)} \label{eq:class-conditional-prob}$$

In this equation, $P(C_m)$ is the class-marginal probability.

In Equation \eqref{eq:class-conditional-prob}, the term $P(\vx)$ is the marginal probability of the instance $\vx$. Since this will be the same across all the classes, we can ignore this term.

Now, they key quantity remaining is $P(\vx|C_m)$, the class-conditional density of $\vx$. Up until here, the motivation is similar to that of the naive Bayes classifier. In the case of the naive Bayes classifier, we make the naive assumption of feature-wise splitting the class-conditional density of $\vx$.

In the case of linear discriminant analysis, we do it a bit differently.

## Multivariate Gaussian as class-conditional density

In the case of linear discriminant analysis, we model the class-conditional density $P(\vx | C_m)$ as a multivariate Gaussian.

$$P(\vx|C_m) = \frac{1}{\sqrt{2\pi |\mSigma_m|}} \expe{-\frac{1}{2}(\vx - \vmu_m)^T \mSigma_m^{-1} (\vx - \vmu_m)}$$

Here, $\vmu_m$ is the mean of the training examples for the class $m$ and $\mSigma_m$ is the covariance for those training examples.

In the case of linear discriminant analysis, the covariance is assumed to be the same for all the classes. This means, $\mSigma_m = \mSigma, \forall m$.

In comparing two classes, say $C_p$ and $C_q$, it suffices to check the log-ratio

$$\log \frac{P(C_p | \vx}{P(C_q | \vx)}$$

Let's look at this log-ratio in further detail by expanding it with appropriate substitutions.

\begin{align} \log \frac{P(C_p | \vx)}{P(C_q | \vx)} &= \log \frac{P(C_p)}{P(C_q)} + \log \frac{P(\vx|C_p)}{P(\vx|C_q)} \\\\ &= \log\frac{P(C_p)}{P(C_q)} - \frac{1}{2}(\vmu_p + \vmu_q)^T \mSigma^{-1} (\vmu_p - \vmu_q) + \vx^T \mSigma^{-1}(\vmu_p - \vmu_q) \label{eqn:log-ratio-expand} \end{align}

This equation is linear in $\vx$, hence the name linear discriminant analysis.

The normalizing factors in both probabilities cancelled in the division since they were both $\sqrt{2\pi |\mSigma|}$. Also, the square-term in both was $\vx^T\mSigma\vx$ and got cancelled, resulting in the linear term based classifier. Both these cancellation will not happen if $\mSigma_p \ne \mSigma_q$, an extension known as quadtratic discriminant analysis. Of course, quadratic discriminant analysis is not a linear classifier then, due to the presence of square terms $\vx^T(\mSigma_p + \mSigma_q)\vx$.

## Predictive model

The prediction follows from the following three conditions on the log-ratio in Equation \eqref{eqn:log-ratio-expand}.

• If the ratio is greater than 0, then the prediction is class $C_p$.
• If less than 0, it is class $C_q$.
• If the log-ratio is zero, then the instance lies on the decision-boundary between the two classes.

From Equation \eqref{eqn:log-ratio-expand}, we see that each class $m$ contributes the following term to the equaiton.

$$\delta_m(\vx) = \vx^T\mSigma^{-1}\vmu_m - \frac{1}{2}\vmu_m^T\mSigma^{-1}\vmu_m + \log P(C_m)$$

This linear formula is known as the linear discriminant function for class $m$. Equipped with this, the prediction can be further summarized as

$$\yhat = \argmax_{m} \delta_m(\vx)$$

## Training linear discriminant analysis

Training a linear discriminant analysis model requires the inference of three parameter types — class priors $P(C_m)$, class conditional means, $\vmu_m$, and the common covariance $\mSigma$.

The priors $P(C_m)$ is estimated as the fraction of training instances that belong to the class $C_m$.

$$P(C_m) = \frac{\text{Number of training instances belonging to } C_m}{\text{Total number of training examples}}$$

The mean of the class-conditional density for class $m$, that is $\vmu_m$, is computed as

$$\vmu_m = \frac{1}{L_m} \sum_{y_i = C_m} \vx_i$$

where, $L_m$ is the number of labeled examples of class $C_m$ in the training set.

The common covariance, $\mSigma$, is computed as

$$\mSigma = \frac{1}{L-M} \sum_{m=1}^{M} \sum_{y_i = C_m} \sum_{i} (\vx_i - \vmu_m)(\vx_i - \vmu_m)^T$$

## Training demo: binary classification

Let us fit a linear discriminant analysis model to some training data. In the first plot we show the learned classification boundary. In the other, we show the probability contours of the two Gaussian distributions.

Being a linear classifier, the trained model comfortably separates linearly separable classes. However, it is ineffective at non-linearly separable scenarios.

Note that by design the covariance is exactly the same for both classes, but the means differ.

## Dealing with feature types

Note that the predictive model involves the calculations of class-conditional means and the common covariance matrix. This is easy for binary and continuous features since both can be treated as real-valued features.

In the case of categorical features a direct metric score calculation is not possible. Therefore, we need to first preprocess the categorical variables using one-hot encoding to arrive at a binary feature representation.

## Multiclass classification

Dealing with multiclass problems with linear discriminant analysis is straightforward. In the development of the model, we never made any simplifying assumption that necessitates a binary classification scenario. As we explained in the section on predictive model, the unlabeled instance gets assigned to the class $C_m$ with the maximum value of the linear disriminant function $\delta_m(\vx)$.

## Training demo: Multiclass classification

Let us fit a linear discriminant analysis model to some multiclass training data. In the first plot we show the learned classification boundary. In the other, we show the probability contours of the two Gaussian distributions.

Thus, just inferring the mean for each class is sufficient to extend LDA to the multiclass setting.

Again, note that by design the covariance is exactly the same for all the classes, but the means differ.

## Number of parameters

For linear discriminant analysis, altogether, there are $M$ class priors, $M$ class-conditional means, and 1 shared covariance matrix. For the $N$-dimensional feature space, each mean is $N$-dimensional and the covariance matrix is $N \times N$ in size. This results in $M + M\times N + N\times N$ total parameters, or $\BigOsymbol( M \times (N+1) )$, if $M > N$.

In the case of quadratic discriminant analysis, there will be many more parameters, $(M-1) \times \left(N (N+3)/2 + 1\right)$.