feature selection

19
Feature Selection Benjamin Biesinger - Manuel Maly - Patrick Zwickl

Upload: fidelina-catalin

Post on 01-Jan-2016

32 views

Category:

Documents


0 download

DESCRIPTION

Feature Selection. Benjamin Biesinger - Manuel Maly - Patrick Zwickl. Agenda. Introduction : What is feature selection? What is our contribution? Phases : What is the sequence of actions in our solution? Solution : How does it work in particular? Results : What is returned? - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Feature Selection

Feature SelectionBenjamin Biesinger - Manuel Maly - Patrick Zwickl

Page 2: Feature Selection

Agenda

• Introduction: What is feature selection? What is our contribution?

• Phases: What is the sequence of actions in our solution?

• Solution: How does it work in particular?

• Results: What is returned?

• Analysis: What to do with it? What can we conclude from it?

Page 3: Feature Selection

Introduction• Not all features of a data set are useful for

classification

• A large number of attributes negatively influences the computation time

• The most essential features should be used for classification

• Feature selection is an approach

• Different search strategies and evaluations are available, but which is the best?

• Automatic feature selection: Several algorithms are run, compared and analyzed for trends → Implemented by us

Page 4: Feature Selection

Phases

• Phases: (I) Meta-classification - (II) Classification

• Before: File loading & preparation

• Afterwards: Comparison + output generation

Page 5: Feature Selection

Solution

• Java command-line application utilizing the WEKA toolkit

• Command-line arguments: Filename (of dataset), Classifier algorithm name, Split (feature selection <-> classification percentage)

• Example: „winequality-red.csv M5Rules 20“

• Computation of results and display in system output of console

Page 6: Feature Selection

Solution (Flow 1)

1.Parsing of dataset and creation of WEKA-specific „Instances“ object.

2.Split of Instances object in two parts, depending on percentage entered by user.

3.Combining all evaluation and search algorithms given in properties-files, and applying on 1. Instances object, finally storing results in dedicated objects (SData).

4.Classifying all combinations from step 3 with classifier entered by user on 2. Instances object. Again storing results in SData objects.

Page 7: Feature Selection

Solution (Flow 2)

5.Gaining aggregate information on all results by iterating over SData objects.

6.Print trend analysis and information on combined evaluation and search algorithms, plus the corresponding classification results (time + mean absolute error).

Page 8: Feature Selection

Solution (Output Excerpt)

@TREND of selected features

Attribute: bottom-right-square has Count: 8

=============== Evaluation: ConsistencySubsetEval ===============

--- Search: GreedyStepwise ---

# of selected features: 1, selection time: 34, classification time: 36, mean abs. error:47,07%

# of selected features: 2, selection time: 35, classification time: 34, mean abs. error:43,16% …

--- Search: RandomSearch ---

Automatic feature number (no influence by user): 5, selection time: 74, classification time: 118, mean abs. error:44,46%

Page 9: Feature Selection

Results• Tested on 3 different datasets

• Tic Tac Toe

• Wine Quality (red)

• Balance Scale

• 2 comparisons per dataset were made

• For each feature selection individually

• Between different feature selection techniques

• Is there a trend which features are selected by most techniques?

Page 10: Feature Selection

1st Comparison

• Influence of number of selected features on

• Runtime

• Classification accuracy (measured in MAE)

Page 11: Feature Selection

1st Comparison Result

• Only those search algorithms used that implement RankedOutputSearch interface

• Capable to influence the number of features to select

• Number of features selected and MAE behave to each other directly proportional – to runtime inversely proportional

Page 12: Feature Selection

2nd Comparison

• Feature Selection Technique consists of

• Search algorithm

• Evaluation algorithm

• Not all combinations possible!

• Different feature selection techniques compared to each other concerning:

• Runtime

• Performance (measured in MAE)

Page 13: Feature Selection

2nd Comparison Result

• Different techniques select different amount of attributes

• In some extent, different attributes, too

• Some techniques are slower than others

• Huge runtime differences between search algorithms

• Some techniques select insufficient attributes to give acceptable results

Page 14: Feature Selection

Trend

• In all tested datasets there was a trend on which features were selected

• Higher count of selection implies bigger influence to the output

Page 15: Feature Selection

Analysis

• Different feature selection techniques – different characteristics

• ClassifierSubsetEval / RaceSearch very good classification results

• Less attributes – faster classification

• Algorithms that select less features are faster

• e.g. GeneticSearch

Page 16: Feature Selection

Lowest error rate

DatasetFeature Selection

TechniqueRuntime

Mean absolute error

Tic Tac ToeClassifierSubsetEval /

RaceSearch642 15,25

Wine Quality (red)

ClassifierSubsetEval / RaceSearch

3594 50,8

Balance Scale

many 9-34 21,96

Page 17: Feature Selection

Lowest runtime

DatasetFeature Selection

TechniqueRuntime

Mean absolute error

Tic Tac Toe x / RankSearch 17 50,85

Wine Quality (red)

WrapperSubsetEval / GeneticSearch

1732 63,57

Balance Scale

many 5-34 -

Page 18: Feature Selection

Trend

Dataset First Second Third

Tic Tac Toe Top-left-squareTop-right-

squareTop-middle-square

Wine Quality (red)

Volatile acidity Fixed acidity Chlorides

Balance Scale

Right-weight Right-distance Left-distance

Page 19: Feature Selection

Feature SelectionBenjamin Biesinger - Manuel Maly - Patrick Zwickl

Any Any questions?questions? The essential features

;)

hääh?

AnythinAnything g

missed?missed?

thxthx