multiple instance real boosting with aggregation functions

30
Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International Conference on Pattern Recognition November 14, 2012

Upload: rue

Post on 23-Feb-2016

60 views

Category:

Documents


0 download

DESCRIPTION

Multiple Instance Real Boosting with Aggregation Functions. Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International Conference on Pattern Recognition November 14, 2012. Multiple Instance Learning. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Multiple Instance Real Boosting with Aggregation Functions

Multiple Instance Real Boosting with Aggregation Functions

Hossein Hajimirsadeghi and Greg MoriSchool of Computing Science

Simon Fraser University

International Conference on Pattern RecognitionNovember 14, 2012

Page 2: Multiple Instance Real Boosting with Aggregation Functions

2

Multiple Instance Learning

• Traditional supervised learning gets Instance/label pairs

• A kind of weak learning to handle ambiguity in training data

• Standard Definitions:– Positive Bag: At least one of the instances is positive– Negative Bag: All the instances are negative

2x1x

5x4x3x

7x9x

8x6x

1x10x

2x

7x9x

8x6x

4x3x

11x12x

• Multiple Instance Learning (MIL) gets bag of instances/label pairs

Page 3: Multiple Instance Real Boosting with Aggregation Functions

3

Applications of MIL

• Image Categorization– [e.g., chen et al., IEEE-TPAMI 2006]

• Content-Based Image Retrieval– [e.g., Li et al., ICCV11]

Page 4: Multiple Instance Real Boosting with Aggregation Functions

4

Applications of MIL

• Text Categorization– [e.g., Andrews et al., NIPS02]

• Object Tracking– [e.g., Babenko et al., IEEE-TPAMI 2011]

Page 5: Multiple Instance Real Boosting with Aggregation Functions

5

Problem & Objective

• The information “At least one of the instances is positive” is very weak and ambiguous.– There are examples of MIL datasets where most

instances in the positive bags are positive. • We aim to mine through different levels of

ambiguity in the data:– For example: a few instances are positive, some

instances are positive, many instances are positive, most instances are positive, …

Page 6: Multiple Instance Real Boosting with Aggregation Functions

6

Approach

• Using the ideas in Boosting:– Finding a bag-level classifier by maximizing the expected

log-likelihood of the training bags– Finding an instance-level strong classifier as a combination

of weak classifiers like RealBoost Algorithm (Friedman et al. 2000), modified by the information from the bag-level classifier

• Using aggregation functions with different degrees of or-ness:– Aggregate the probability of instances to define probability

of a bag be positive

Page 7: Multiple Instance Real Boosting with Aggregation Functions

7

Ordered Weighted Averaging (OWA)

• OWA is an aggregation function:]1,0[]1,0[:owa n

n

iiin bwaaa

121 ),...,,owa(

1],1,0[1

n

iii ww

ji aib oflargest th theis

Yager et al. IEEE-TSMC, 1988

Page 8: Multiple Instance Real Boosting with Aggregation Functions

8

OWA: Example

?)6.0,1.0,9.0,5.0owa(

1) Sort the values:0.9, 0.6, 0.5, 0.1

1.05.06.09.0 4321 wwww

Ex: uniform aggregation (mean):

1.0415.0

416.0

419.0

41

2) Compute the weighted sum:

Page 9: Multiple Instance Real Boosting with Aggregation Functions

9

OWA: Linguistic Quantifiers

• Regular Increasing Monotonic (RIM) Quantifiers– All, Many, Half, Some, At Least One, …

)1()(niQ

niQwi

ppQ )(

Page 10: Multiple Instance Real Boosting with Aggregation Functions

10

OWA: RIM Quantifiers

• RIM Quantifier : All ppQ )(

1

1Q

nini

wi 10

)6.0,1.0,9.0,5.0owa(

1.01.0*15.0*06.0*09.0*0

)1()(niQ

niQwi

Ex:

Page 11: Multiple Instance Real Boosting with Aggregation Functions

11

OWA: RIM Quantifiers

• RIM Quantifier : At Least One ppQ )(0

1011

ii

wi

)6.0,1.0,9.0,5.0owa(

9.01.0*05.0*06.0*09.0*1

)1()(niQ

niQwi

Ex:

1

1Q

Page 12: Multiple Instance Real Boosting with Aggregation Functions

12

OWA: RIM Quantifiers

• RIM Quantifier : At Least Some ppQ )(5.0

Gives higher weights to the largest arguments

So, some high values are enough to make the result high

1

1Q

n1

n2n3

1w

2w3w

Page 13: Multiple Instance Real Boosting with Aggregation Functions

13

OWA: RIM Quantifiers

• RIM Quantifier : Many ppQ )(2

Gives lower weights to the largest arguments

So, many arguments should have high values to make the result high

1

1 Q

n1

n2n31w

3w2w

Page 14: Multiple Instance Real Boosting with Aggregation Functions

14

OWA: Linguistic Quantifiers Linguistic Quantifier Degree of orness

At least one of them (Max function)

0.999

Few of them 0.1 0.909

Some of them 0.5 0.667

Half of them 1 0.5

Many of them 2 0.333

Most of them 10 0.091

All of them (Min Function)

0.001

0

Page 15: Multiple Instance Real Boosting with Aggregation Functions

15

MIRealBoost

4x2x

3x1x

5x

6x

)( 1xp

)( 3xp

)( 2xp

)( 4xp

)( 5xp

)( 6xp

)( 2Xpk

)( 3Xpk

)( 1Xpk

)( 2XF b

)( 3XF b

)( 1XF b

)(xF

)( 1xF

)( 3xF

)( 2xF

)( 4xF

)( 5xF

)( 6xF OWA

InstanceProbabilities

BagProbabilities

TrainingBags

InstanceClassifier

)(XF b BagClassifier

Page 16: Multiple Instance Real Boosting with Aggregation Functions

16

MIRealBoost

• MIL training input:

• Objective to find the bag classifier:

]1,1[:

)(sign)(

b

bb

H

XFXH

bags possible all ofset

NN YXYXYX ,,...,,,, 212111

imiii xxxX ,...,, 21

]1,1[iY

Page 17: Multiple Instance Real Boosting with Aggregation Functions

17

MIRealBoost: Learning Bag Classifier

• Objective: Maximize the Expected Binomial Log-Likelihood:

)()(

)(

)(XFXF

XF

bb

b

eeeXp

)(1log)|1()(log)|1(max)(

XpXYPXpXYPXF b

)|1(1)|1(log

21)(

XYPXYPXF b

• Proved: ?

Page 18: Multiple Instance Real Boosting with Aggregation Functions

18

MIRealBoost: Estimate Bag Prob.?)|1( XYP

4x1x

3x2x )|( 2xyP

)|( 4xyP

)|( 1xyP

)|( 3xyP)|( XYP

Estimate probability of each instance

Aggregate

• Aggregation functions:• Noisy-OR

• OWA

X

j

jXx xyPxyP1

)|1(11|1(NOR

?

Page 19: Multiple Instance Real Boosting with Aggregation Functions

19

MIRealBoost: Estimate Instance Prob.

• Estimate Instance Probabilities by training the standard RealBoost classifier:

)(sign)( xFxH

M

mm XfxF

1

)()(

)()(

)(

)|1( xFxF

xF

eeexyP

?)|1( xyP

• Then:

Page 20: Multiple Instance Real Boosting with Aggregation Functions

20

MIRealBoost: Learning Instance Classifier• RealBoost classifier :

)()(1E min xfxFy

fmm

m

e

)(1),(by weighted

,given of PDF: xyF

w

meyxw

yxP

)1|()1|(log

21)(

yxPyxPxf

w

wm

• Proved:

?

Page 21: Multiple Instance Real Boosting with Aggregation Functions

MIRealBoost: Estimate Weak Classifiers

?)|( yxPw

yxPyxPijijij wyxw |)|( ,,

)1|( yxPw)1|( yxPw

Page 22: Multiple Instance Real Boosting with Aggregation Functions

22

MIRealBoost: Estimate Weak Classifiers

ipij Yy

)( ib

i XFYpij ew

)1|(

)1|(log

21)(

,,

,,

yxP

yxPxf

pijw

pijyijx

pij

pijij wyx

m

• We do not know true instance labels.

• Estimate the instance label by the bag label, weighted by the bag confidence

yxPijijij wyx |,,

?ijy

?

Page 23: Multiple Instance Real Boosting with Aggregation Functions

23

MIRealBoost Algorithm

4x2x

3x1x

5x6x

)( 1xf km)( 3xf km

)( 2xf km

)( 4xf km

)( 6xf km )( 5xf km

)( 1xpk

)( 3xpk

)( 2xpk

)( 4xpk)( 5xpk)( 6xpk

)( 2Xpk )( 3Xpk)( 1Xpk

kkk argmax*

)( 2XF b )( 3XF b)( 1XF b

For each feature k=1:K, compute the weak classifier

Compute the instance probabilities

Aggregate the instance probabilities to find bag probabilities

Compute the experimental log likelihood

)(1)(log

21)( *

*

XpXpXFk

kb

Page 24: Multiple Instance Real Boosting with Aggregation Functions

24

Experiments

• Popular MIL datasets:– Image categorization: Elephant, Fox, and Tiger– Drug activity prediction: Musk1 and Musk2

Page 25: Multiple Instance Real Boosting with Aggregation Functions

25

Results

• MIRealBoost classification accuracy with Different Aggregation functions

agg Elephant Fox Tiger Musk1 Musk2

NOR 83 63 72 85 74

Max 77 58 68 85 74

Few 75 58 70 83 72

Some 75 57 73 85 75

Half 72 54 70 90 77

Many 67 52 67 91 75

Most 54 50 51 83 69

All 50 50 50 84 69

Page 26: Multiple Instance Real Boosting with Aggregation Functions

26

Results

• Comparison with MILBoost Algorithm

Method Elephant Fox Tiger Musk1 Musk2

MIRealBoost 83 63 73 91 77

MILBoost 73 58 56 71 61

MILBoost results are reported from Leistner et al. ECCV10

Page 27: Multiple Instance Real Boosting with Aggregation Functions

27

Results

• Comparison between state-of-the-art MIL methodsMethod Elephant Fox Tiger Musk1 Musk2MIRealBoost 83 63 73 91 77MIForest 84 64 82 85 82MI-Kernel 84 60 84 88 89MI-SVM 81 59 84 78 84mi-SVM 82 58 79 87 84MILES 81 62 80 88 83AW-SVM 82 64 83 86 84AL-SVM 79 63 78 86 83EM-DD 78 56 72 85 85MIGraph 85 61 82 90 90miGraph 87 62 86 90 90

Page 28: Multiple Instance Real Boosting with Aggregation Functions

28

Conclusion

• Proposed MIRealBoost algorithm

• Modeling different levels of ambiguity in data– Using OWA aggregation functions which can realize a wide

range of orness in aggregation

• Experimental results showed:– encoding degree of ambiguity can improve the accuracy– MIRealBoost outperforms MILBoost and comparable with

state-of-the art methds

Page 29: Multiple Instance Real Boosting with Aggregation Functions

29

Thanks!

• supported by grants from the Natural Sciences and Engineering Research Council of Canada (NSERC).

Page 30: Multiple Instance Real Boosting with Aggregation Functions

30

MIRealBoost: Learning Instance Classifier

• Implementation details:– Each weak classifier is a stump (i.e., built from only one

feature).

– At each step, the best feature is selected as the feature which leads to the bag probabilities, which maximize the empirical log-likelihood of the bags.

)1|][(

)1|][(log

21)(

,],[

,],[

ykxP

ykxPxf

pij

pijij

pij

pijij

wykx

wykxkm