classification discriminant...

Post on 17-Jun-2020

8 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Jeff Howbert Introduction to Machine Learning Winter 2014 1

Classification

Discriminant Analysis

slides thanks to Greg Shakhnarovich (CS195-5, Brown Univ., 2006)

Jeff Howbert Introduction to Machine Learning Winter 2014 2

Jeff Howbert Introduction to Machine Learning Winter 2014 3

Jeff Howbert Introduction to Machine Learning Winter 2014 4

Jeff Howbert Introduction to Machine Learning Winter 2014 5

Jeff Howbert Introduction to Machine Learning Winter 2014 6

We want to minimize “overlap” between projections of the two classes.One approach: make the class projections a) compact, b) far apart.An obvious idea: maximize separation between the projected means.

Jeff Howbert Introduction to Machine Learning Winter 2014 7

Jeff Howbert Introduction to Machine Learning Winter 2014 8

Jeff Howbert Introduction to Machine Learning Winter 2014 9

Jeff Howbert Introduction to Machine Learning Winter 2014 10

Jeff Howbert Introduction to Machine Learning Winter 2014 11

Example of applying Fisher’s LDA

maximize separation of means maximize Fisher’s LDA criterion→ better class separation

Jeff Howbert Introduction to Machine Learning Winter 2014 12

Fisher’s LDA gives an optimal choice of w, the vector for projection down to one dimension.For classification, we still need to select a threshold to compare projected values to. Two possibilities:– No explicit probabilistic assumptions. Find threshold

which minimizes empirical classification error.– Make assumptions about data distributions of the

classes, and derive theoretically optimal decision boundary.

Usual choice for class distributions is multivariate Gaussian.We also will need a bit of decision theory.

Using LDA for classification in one dimension

Jeff Howbert Introduction to Machine Learning Winter 2014 13

To minimize classification error:

Decision theory

≈= )|(maxargˆ xCpyC

At a given point x in feature space, choose as the predicted class the class that has the greatest probability at x.

Jeff Howbert Introduction to Machine Learning Winter 2014 14

Decision theory

≈= )|(maxargˆ xCpyC

At a given point x in feature space, choose as the predicted class the class that has the greatest probability at x.

probability densities for classes C1 and C2 relative probabilities for classes C1 and C2

Jeff Howbert Introduction to Machine Learning Winter 2014 15

Classification via discriminant analysis,using the classify() function.

Data for each class modeled as multivariate Gaussian.

matlab_demo_06.m

class = classify( sample, training, group, ‘type’ )

MATLAB interlude

test datapredicted testlabels

traininglabels

model for classcovariances

training data

Jeff Howbert Introduction to Machine Learning Winter 2014 16

Models for class covariances

MATLAB classify() function

‘linear’:all classes have same covariance matrix

→ linear decision boundary

‘diaglinear’:all classes have same diagonal covariance matrix

→ linear decision boundary

‘quadratic’:classes have different covariance matrices

→ quadratic decision boundary

‘diagquadratic’:classes have different diagonal covariance matrices

→ quadratic decision boundary

Jeff Howbert Introduction to Machine Learning Winter 2014 17

Example with ‘quadratic’ model of class covariances

MATLAB classify() function

Jeff Howbert Introduction to Machine Learning Winter 2014 18

Relative class probabilities for LDA

‘linear’:all classes have same covariance matrix

→ linear decision boundaryrelative class probabilities have

exactly same sigmoidal formas in logistic regression

top related