Map > Data Mining > Predicting the Future > Modeling > Classification > Linear Discriminant Analysis |

## Linear Discriminant Analysis |

Linear Discriminant Analysis (LDA) is a classification method originally developed in 1936 by R. A. Fisher. It is simple, mathematically robust and often produces models whose accuracy is as good as more complex methods. |

Algorithm |

LDA is based upon the concept of searching for a linear combination of variables (predictors) that best separates two classes (targets). To capture the notion of separability, Fisher defined the following score function. |

Given the score function, the problem is to estimate the linear coefficients that maximize the score which can be solved by the following equations. |

One way of assessing the effectiveness of the discrimination is to calculate the
Mahalanobis distance between two groups. A distance greater than 3 means that in two
averages differ by more than 3 standard deviations. It means that the overlap (probability of misclassification) is quite small. |

Finally, a new point is classified by projecting it onto the maximally separating
direction and classifying it as C1 if: |

Example: |

Suppose we received a dataset from a bank regarding its small business clients who defaulted (red square) and those that did not (blue circle) separated by delinquent days (DAYSDELQ) and number of months in business (BUSAGE). We use LDA to find an optimal linear model that best separates two classes (default and non-default). |

The first step is to calculate the mean (average) vectors, covariance matrices and class probabilities. |

Then, we calculate pooled covariance matrix and finally the coefficients of the linear model. |

A Mahalanobis distance of 2.32
shows a small overlap between two groups which means a good separation
between classes by the linear model. |

Predictors Contribution |

A simple linear correlation between the model scores and predictors can be used to test which predictors contribute significantly to the discriminant function. Correlation varies from -1 to 1, with -1 and 1 meaning the highest contribution but in different directions and 0 means no contribution at all. |

## Quadratic Discriminant Analysis (QDA) |

QDA is a general discriminant function with a quadratic decision boundaries which can be used to classify datasets with two or more classes. QDA has more predictability power than LDA but it needs to estimate the covariance matrix for each classes. |

where Cis
the covariance matrix for the class _{k }k
(-1 means inverse matrix), |C|
is the determinant of the covariance matrix _{k}C,
and _{k}P(c) is the
prior probability of the class _{k}k. The
classification rule is simply to find the class with highest Z
value. |

Try to invent a real time LDA classifier. You should be able to add or remove data and variables (predictors and classes) on the fly. |

LDA Interactive |