constructing associative classifiers from decision tables

16
5/14-5/17, Toronto RSFDGrC - 2007 1 Constructing Constructing Associative Associative Classifiers from Classifiers from Decision Tables Decision Tables Jianchao Han California State University Dominguez Hills, USA T. Y. Lin San Jose State University, USA Jiye Li University of Waterloo, Canada Nick Cercone York University, Canada

Upload: charla

Post on 05-Jan-2016

16 views

Category:

Documents


1 download

DESCRIPTION

Constructing Associative Classifiers from Decision Tables. Jianchao Han California State University Dominguez Hills, USA T. Y. Lin San Jose State University, USA Jiye Li University of Waterloo, Canada Nick Cercone York University, Canada. Agenda. Introduction Related Work - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 1

Constructing Associative Constructing Associative Classifiers from Decision TablesClassifiers from Decision Tables

Jianchao HanCalifornia State University Dominguez Hills, USA

T. Y. LinSan Jose State University, USA

Jiye LiUniversity of Waterloo, Canada

Nick CerconeYork University, Canada

Page 2: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 2

AgendaAgenda

• Introduction

• Related Work

• Our Approach

• Algorithm Description

• An Example Demonstration

• Conclusion

Page 3: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 3

IntroductionIntroduction

• Associative classifier– A set of classification rules– Classification rules as a special form of

association rules– Classifier formed by finding constrained

association rules

• Rough set theory used to reduce data set

Page 4: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 4

Related WorkRelated Work

• Rough set theory– Attribute reduct

• Association rules– Rule template– Constrained rules

• Significant rules– Rule importance measurement

• Michalski’s coverage method

Page 5: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 5

Our ApproachOur Approach• Combining three strategies

– Association rules – Rough set theory to find attribute reducts– Coverage method to form a classifier

• Four steps– Finding attribute reducts– Finding constrained association rules– Measuring association rules – Selecting important rules to cover instances

Page 6: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 6

Generate All Attribute ReductsGenerate All Attribute Reducts

• Use existing reduct finding algorithms such as Genetic Reduct generation algorithm in ROSETTA

Page 7: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 7

Find Classification RulesFind Classification Rules

• For each attribute reduct– Use adapted Apriori algorithm to find

constrained association rules, where right side of rules is constrained to a class label

• Carefully determine the thresholds of support and confident for association rules

Page 8: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 8

Measure Importance of RulesMeasure Importance of Rules

• Rule importance definition

• Properties of the rule importance– 0 < Importance(Rule) ≤ 1

– If Rule only contains core attributes, its importance is 1.

reducts ofnumber total

Rule thegenerates that reducts ofnumber (Rule)Importance

Page 9: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 9

Rule PrecedenceRule Precedence• Given two rules R1 and R2 generated, R1 precedes

R2 (R1 has a higher precedence than R2), denoted Precedence(R1)>Precedence(R1), if – Importance(R1) > Importance(R2); or– Importance(R1) = Importance(R2), and

Confidence(R1) > Confidence(R2); or– Importance(R1) = Importance(R2), and

Confidence(R1) = Confidence(R2), and Support(R1) > Support (R2).

• Otherwise R1 and R2 are considered having the same precedence and denoted Precedence(R1)=Precedence(R2).

Page 10: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 10

Property of Rule PrecedenceProperty of Rule Precedence

• The precedence relationship is a total order relation

• Thus all rules can be sorted based on their precedence

Page 11: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 11

Find Associative ClassifierFind Associative Classifier

• Sort all rules in terms of their precedence consisting of importance, confidence and support

• Select next rule in the sorted sequence• If this rule covers some rows

– Delete all rows covered by this rule– Put this rule in the classifier

• Repeat until all rules are exhausted

Page 12: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 12

An ExampleAn ExampleMakemodel cyl door displace compress power trans weight mileage

USA 6 2 Medium High High Auto Medium Medium

USA 6 4 Medium Medium Medium Manual Medium Medium

USA 4 2 Small High Medium Auto Medium Medium

USA 4 2 Medium Medium Medium Manual Medium Medium

USA 4 2 Medium Medium High Manual Medium Medium

USA 6 4 Medium Medium High Auto Medium Medium

USA 4 2 Medium Medium High Auto Medium Medium

USA 4 2 Medium High High Manual Light High

Japan 4 2 Small High Low Manual Light High

Japan 4 2 Medium Medium Medium Manual Medium High

Japan 4 2 Small High High Manual Medium High

Japan 4 2 Small Medium Low Manual Medium High

Japan 4 2 Small High Medium Manual Medium High

USA 4 2 Small High Medium Manual Medium High

Page 13: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 13

Find All Attribute ReductsFind All Attribute Reducts

• Genetic reducer in ROSETTA • Four attribute reducts

Reduct # Reduct Attributes

1 Make, compress, power, trans

2 make, cyl, compress, trans

3 make, displace, compress, trans

4 make, cyl, door, displace, trans, weight

Page 14: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 14

Constrained Associated RulesConstrained Associated Rules• support threshold = 1% • confidence threshold = 100% • applying the adapted Apriori algorithm

# Rules Rule Preced

12345678910111213

(make, Japan) → (Mileage, High)(Trans, Auto) → (Mileage, Medium)(Compress, High), (Trans, Manual) → (Mileage, High)(make, USA), (Compress, Medium) → (Mileage, Medium)(Displace, Small), (Trans, Manual) → (Mileage, High)(Cyl, 6) → (Mileage, Medium)(USA, Car), (Displace, Medium), (Weight, Medium) → (Mileage, Medium) (make, USA), (Power, High) → (Mileage, Medium)(Compress, Medium), (Power, High) → (Mileage, Medium) (Power, Low) → (Mileage, High) (Door, 4) → (Mileage, Medium)(Weight, Light) → (Mileage, High)(Displace, Small), (Compress, Medium) → (Mileage, High)

4/4, 5 4/4, 43/4 53/4, 52/4, 52/4, 31/4, 61/4, 41/4, 31/4, 21/4, 21/4, 21/4, 1

Page 15: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 15

Construct ClassifierConstruct Classifier

• Covering method• Rule 1: Covers 5 rows 9 through 13• Rule 2: Covers 2 rows 8 and 14• Rule 3: Covers 4 rows 1, 3, 6, and 7 • Rule 4: Covers 3 rows 2, 4 and 5 • Since all rows in the original decision table have

been covered by Rules 1 through 4, the final associative classifier contains only these four class association rules

Page 16: Constructing Associative Classifiers from Decision Tables

5/14-5/17, Toronto RSFDGrC - 2007 16

ConclusionConclusion

• Introduce an approach to constructing associative classifiers based on– Rough set theory to find attribute reducts– Association rule mining algorithm– Covering method to build classifiers

• Present the rule importance and precedence measurement used in the proposed approach

• Demonstrate an example