data mining - department of computer science

43
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 3 of Data Mining by I. H. Witten, E. Frank and M. A. Hall

Upload: others

Post on 04-Feb-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Data Mining - Department of Computer Science

Data MiningPractical Machine Learning Tools and Techniques

Slides for Chapter 3 of Data Mining by I. H. Witten, E. Frank and M. A. Hall

Page 2: Data Mining - Department of Computer Science

2Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Output: Knowledge representation

● Tables● Linear models● Trees● Rules

●Classification rules●Association rules●Rules with exceptions●More expressive rules

● Instance-based representation● Clusters

Page 3: Data Mining - Department of Computer Science

3Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Output: representing structural patterns

● Many different ways of representing patterns♦ Decision trees, rules, instance-based, …

● Also called “knowledge” representation● Representation determines inference method● Understanding the output is the key to understanding the

underlying learning methods● Different types of output for different learning problems

(e.g. classification, regression, …)

Page 4: Data Mining - Department of Computer Science

4Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Tables● Simplest way of representing output:

● Use the same format as input!● Decision table for the weather problem:

● Main problem: selecting the right attributes

NoNormalRainy

NoHighRainy

YesNormalOvercast

YesHighOvercast

YesNormalSunny

NoHighSunny

PlayHumidityOutlook

Page 5: Data Mining - Department of Computer Science

5Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Linear models

● Another simple representation● Regression model

♦ Inputs (attribute values) and output are all numeric● Output is the sum of weighted attribute values

♦ The trick is to find good values for the weights

Page 6: Data Mining - Department of Computer Science

6Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

A linear regression function for the CPU performance data

PRP = 37.06 + 2.47CACH

Page 7: Data Mining - Department of Computer Science

7Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

● Binary classification● Line separates the two classes

♦ Decision boundary - defines where the decision changes from one class value to the other

● Prediction is made by plugging in observed values of the attributes into the expression

♦ Predict one class if output ≥ 0, and the other class if output < 0

● Boundary becomes a high-dimensional plane (hyperplane) when there are multiple attributes

Linear models for classification

Page 8: Data Mining - Department of Computer Science

8Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Separating setosas from versicolors

2.0 – 0.5PETAL-LENGTH – 0.8PETAL-WIDTH = 0

Page 9: Data Mining - Department of Computer Science

9Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Trees

● “Divide-and-conquer” approach produces tree● Nodes involve testing a particular attribute● Usually, attribute value is compared to constant● Other possibilities:

● Comparing values of two attributes● Using a function of one or more attributes

● Leaves assign classification, set of classifications, or probability distribution to instances

● Unknown instance is routed down the tree

Page 10: Data Mining - Department of Computer Science

10Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Nominal and numeric attributes

● Nominal:number of children usually equal to number values⇒ attribute won’t get tested more than once● Other possibility: division into two subsets

● Numeric:test whether value is greater or less than constant⇒ attribute may get tested several times● Other possibility: three-way split (or multi-way split)

● Integer: less than, equal to, greater than● Real: below, within, above

Page 11: Data Mining - Department of Computer Science

11Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Missing values

● Does absence of value have some significance?● Yes ⇒ “missing” is a separate value● No ⇒ “missing” must be treated in a special way

♦ Solution A: assign instance to most popular branch♦ Solution B: split instance into pieces

● Pieces receive weight according to fraction of training instances that go down each branch

● Classifications from leave nodes are combined using the weights that have percolated to them

Page 12: Data Mining - Department of Computer Science

12Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Trees for numeric prediction

● Regression: the process of computing an expression that predicts a numeric quantity

● Regression tree: “decision tree” where each leaf predicts a numeric quantity

♦ Predicted value is average value of training instances that reach the leaf

● Model tree: “regression tree” with linear regression models at the leaf nodes

♦ Linear patches approximate continuous function

Page 13: Data Mining - Department of Computer Science

13Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Linear regression for the CPU data

PRP = - 56.1 + 0.049 MYCT + 0.015 MMIN + 0.006 MMAX + 0.630 CACH - 0.270 CHMIN + 1.46 CHMAX

Page 14: Data Mining - Department of Computer Science

14Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Regression tree for the CPU data

Page 15: Data Mining - Department of Computer Science

15Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Model tree for the CPU data

Page 16: Data Mining - Department of Computer Science

16Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Classification rules

● Popular alternative to decision trees● Antecedent (pre-condition): a series of tests (just like the

tests at the nodes of a decision tree)● Tests are usually logically ANDed together (but may also

be general logical expressions)● Consequent (conclusion): classes, set of classes, or

probability distribution assigned by rule● Individual rules are often logically ORed together

♦ Conflicts arise if different conclusions apply

Page 17: Data Mining - Department of Computer Science

17Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

From trees to rules

● Easy: converting a tree into a set of rules♦ One rule for each leaf:

● Antecedent contains a condition for every node on the path from the root to the leaf

● Consequent is class assigned by the leaf● Produces rules that are unambiguous

♦ Doesn’t matter in which order they are executed● But: resulting rules are unnecessarily complex

♦ Pruning to remove redundant tests/rules

Page 18: Data Mining - Department of Computer Science

18Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

From rules to trees

● More difficult: transforming a rule set into a tree● Tree cannot easily express disjunction between rules

● Example: rules which test different attributes

● Symmetry needs to be broken● Corresponding tree contains identical subtrees

(⇒ “replicated subtree problem”)

If a and b then xIf c and d then x

Page 19: Data Mining - Department of Computer Science

19Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

A tree for a simple disjunction

Page 20: Data Mining - Department of Computer Science

20Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

The exclusive-or problem

If x = 1 and y = 0then class = a

If x = 0 and y = 1then class = a

If x = 0 and y = 0then class = b

If x = 1 and y = 1then class = b

Page 21: Data Mining - Department of Computer Science

21Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

A tree with a replicated subtree

If x = 1 and y = 1then class = a

If z = 1 and w = 1then class = a

Otherwise class = b

Page 22: Data Mining - Department of Computer Science

22Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

“Nuggets” of knowledge

● Are rules independent pieces of knowledge? (It seems easy to add a rule to an existing rule base.)

● Problem: ignores how rules are executed● Two ways of executing a rule set:

♦ Ordered set of rules (“decision list”)● Order is important for interpretation

♦ Unordered set of rules● Rules may overlap and lead to different conclusions for the same

instance

Page 23: Data Mining - Department of Computer Science

23Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Interpreting rules

● What if two or more rules conflict?♦ Give no conclusion at all?♦ Go with rule that is most popular on training data?♦ …

● What if no rule applies to a test instance?♦ Give no conclusion at all?♦ Go with class that is most frequent in training data?♦ …

Page 24: Data Mining - Department of Computer Science

24Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Special case: boolean class

● Assumption: if instance does not belong to class “yes”, it belongs to class “no”

● Trick: only learn rules for class “yes” and use default rule for “no”

● Order of rules is not important. No conflicts!● Rule can be written in disjunctive normal form

If x = 1 and y = 1 then class = aIf z = 1 and w = 1 then class = aOtherwise class = b

Page 25: Data Mining - Department of Computer Science

25Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Association rules

● Association rules…♦ … can predict any attribute and combinations of

attributes♦ … are not intended to be used together as a set

● Problem: immense number of possible associations♦ Output needs to be restricted to show only the most

predictive associations ⇒ only those with high support and high confidence

Page 26: Data Mining - Department of Computer Science

26Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Support and confidence of a rule● Support: number of instances predicted correctly ● Confidence: number of correct predictions, as proportion of

all instances that rule applies to● Example: 4 cool days with normal humidity

⇒ Support = 4, confidence = 100%● Normally: minimum support and confidence pre-specified

(e.g. 58 rules with support ≥ 2 and confidence ≥ 95% for weather data)

If temperature = cool then humidity = normal

Page 27: Data Mining - Department of Computer Science

27Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Interpreting association rules

● Interpretation is not obvious:

is not the same as

● It means that the following also holds:

If windy = false and play = no then outlook = sunny and humidity = high

If windy = false and play = no then outlook = sunny If windy = false and play = no then humidity = high

If humidity = high and windy = false and play = no then outlook = sunny

Page 28: Data Mining - Department of Computer Science

28Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Rules with exceptions

● Idea: allow rules to have exceptions● Example: rule for iris data

● New instance:

● Modified rule:0.2

Petalwidth

2.6

Petallength

Iris-setosa3.55.1

TypeSepalwidth

Sepallength

If petal-length ≥ 2.45 and petal-length < 4.45 then Iris-versicolor

If petal-length ≥ 2.45 and petal-length < 4.45 then Iris-versicolor EXCEPT if petal-width < 1.0 then Iris-setosa

Page 29: Data Mining - Department of Computer Science

29Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

A more complex example

● Exceptions to exceptions to exceptions …

default: Iris-setosa

except if petal-length ≥ 2.45 and petal-length < 5.355 and petal-width < 1.75

then Iris-versicolor

except if petal-length ≥ 4.95 and petal-width < 1.55 then Iris-virginica

else if sepal-length < 4.95 and sepal-width ≥ 2.45 then Iris-virginica

else if petal-length ≥ 3.35 then Iris-virginica

except if petal-length < 4.85 and sepal-length < 5.95

then Iris-versicolor

Page 30: Data Mining - Department of Computer Science

30Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Advantages of using exceptions

● Rules can be updated incrementally♦ Easy to incorporate new data♦ Easy to incorporate domain knowledge

● People often think in terms of exceptions● Each conclusion can be considered just in the

context of rules and exceptions that lead to it♦ Locality property is important for understanding large

rule sets♦ “Normal” rule sets don’t offer this advantage

Page 31: Data Mining - Department of Computer Science

31Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

More on exceptions

● Default...except if...then...

is logically equivalent toif...then...else

(where the else specifies what the default did)● But: exceptions offer a psychological advantage

♦ Assumption: defaults and tests early on apply more widely than exceptions further down

♦ Exceptions reflect special cases

Page 32: Data Mining - Department of Computer Science

32Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Rules involving relations

● So far: all rules involved comparing an attribute-value to a constant (e.g. temperature < 45)

● These rules are called “propositional” because they have the same expressive power as propositional logic

● What if problem involves relationships between examples (e.g. family tree problem from above)?

♦ Can’t be expressed with propositional rules♦ More expressive representation required

Page 33: Data Mining - Department of Computer Science

33Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

The shapes problem

● Target concept: standing up● Shaded: standing

Unshaded: lying

Page 34: Data Mining - Department of Computer Science

34Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

A propositional solution

Lying3210

Lying419

Standing

Lying

Standing

Lying

Standing

Standing

Class

492

367

387

434

463

442

SidesHeightWidth

If width ≥ 3.5 and height < 7.0then lying

If height ≥ 3.5 then standing

Page 35: Data Mining - Department of Computer Science

35Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

A relational solution

● Comparing attributes with each other

● Generalizes better to new data● Standard relations: =, <, >● But: learning relational rules is costly● Simple solution: add extra attributes

(e.g. a binary attribute is width < height?)

If width > height then lyingIf height > width then standing

Page 36: Data Mining - Department of Computer Science

36Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Rules with variables● Using variables and multiple relations:

● The top of a tower of blocks is standing:

● The whole tower is standing:

● Recursive definition!

If height_and_width_of(x,h,w) and h > wthen standing(x)

If is_top_of(x,z) and height_and_width_of(z,h,w) and h > wand is_rest_of(x,y)and standing(y)then standing(x)

If empty(x) then standing(x)

If height_and_width_of(x,h,w) and h > w and is_top_of(y,x)then standing(x)

Page 37: Data Mining - Department of Computer Science

37Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Inductive logic programming

● Recursive definition can be seen as logic program● Techniques for learning logic programs stem from the

area of “inductive logic programming” (ILP)● But: recursive definitions are hard to learn

♦ Also: few practical problems require recursion♦ Thus: many ILP techniques are restricted to non-recursive

definitions to make learning easier

Page 38: Data Mining - Department of Computer Science

38Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Instance-based representation

● Simplest form of learning: rote learning♦ Training instances are searched for instance that most

closely resembles new instance♦ The instances themselves represent the knowledge♦ Also called instance-based learning

● Similarity function defines what’s “learned”● Instance-based learning is lazy learning● Methods: nearest-neighbor, k-nearest-neighbor, …

Page 39: Data Mining - Department of Computer Science

39Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

The distance function● Simplest case: one numeric attribute

♦ Distance is the difference between the two attribute values involved (or a function thereof)

● Several numeric attributes: normally, Euclidean distance is used and attributes are normalized

● Nominal attributes: distance is set to 1 if values are different, 0 if they are equal

● Are all attributes equally important?♦ Weighting the attributes might be necessary

Page 40: Data Mining - Department of Computer Science

40Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Learning prototypes

● Only those instances involved in a decision need to be stored

● Noisy instances should be filtered out● Idea: only use prototypical examples

Page 41: Data Mining - Department of Computer Science

41Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Rectangular generalizations

● Nearest-neighbor rule is used outside rectangles● Rectangles are rules! (But they can be more

conservative than “normal” rules.) ● Nested rectangles are rules with exceptions

Page 42: Data Mining - Department of Computer Science

42Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Representing clusters I

Simple 2-D representation Venn diagram

Overlapping clusters

Page 43: Data Mining - Department of Computer Science

43Data Mining: Practical Machine Learning Tools and Techniques (Chapter 3)

Representing clusters II

1 2 3

a 0.4 0.1 0.5b 0.1 0.8 0.1c 0.3 0.3 0.4d 0.1 0.1 0.8e 0.4 0.2 0.4f 0.1 0.4 0.5g 0.7 0.2 0.1h 0.5 0.4 0.1…

Probabilistic assignment Dendrogram

NB: dendron is the Greek word for tree