1 learning with local and global consistency presented by qiuhua liu duke university machine...

12
1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet, Tomas Navin Lal, Jason Weston and Bernhard Schölkopf at NIPS 2003

Upload: spencer-mccoy

Post on 31-Dec-2015

213 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

1

Learning with Local and Global Consistency

Presented by Qiuhua Liu

Duke University Machine Learning Group

March 23, 2007

By Dengyong Zhou, Olivier Bousquet, Tomas Navin Lal, Jason Weston and Bernhard Schölkopf at NIPS 2003

Page 2: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

2

Outline

The consistency assumption for semi-supervised learning: Why unlabeled data could help the classification?

The Consistency algorithm: a very simple algorithm based on the above assumption.

The Relation to Xuejun’s Label Iteration algorithm

Experimental results

Conclusions

Page 3: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

3

Semi-supervised Learning Problem

We all know that Semi-supervised Learning is very important, but why could we do that?

The key to semi-supervised learning problems is the prior assumption of consistency:

(1)Local Consistency: nearby points are likely to have the same label;

(2)Global Consistency: Points on the same structure (cluster or manifold) are likely to have the same label;

Page 4: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

4

Local and Global Consistency

The key to the consistency algorithm is to let every point iteratively spread its label information to its neighbors until a global stable state is achieved.

Page 5: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

5

Some Terms

x, data point set: L, Label set: F, a classification on x:

Y, initial classification on x, which is a special case of F with:

},...,,,...{ 11 nll xxxxx },...1{ cL

ncnn

c

cn

FFF

FFF

F

...

..............

...

21

11211

ijcjii Fyx argmax as labelingby

otherwise ,0

aslabeled is if ,1 jy xY ii

ij

Page 6: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

6

The Consistency Algorithm

1.Construct the affinity matrix W defined by a Gaussian kernel:

2. Normalize W symmetrically by

3. Iterate until converge.

4. Let denote the limit of the sequence {F(t)}. The classification results is :

if ,0

if ),2/exp( 22

ki

kixxw ki

ik

, 2/12/1 WDDS

where D is a diagonal matrix with .k ikii WD

YtSFtF )1()()1( *F

*argmax as Labeling ijFyx cjii

Page 7: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

7

The first two steps are the same as Spectral Clustering.

The third step:

• First term: each point receive information from its neighbors.

• Second term: retains the initial information.

From the iteration equation, it is very easy to show that:

So we could compute F* directly without iterations.

YtSFtF )1()()1(

YSItFFt

1* )()(lim

The Consistency Algorithm (Cont.)

Page 8: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

8

The convergence process

The initial label information are diffused along the two moons.

Page 9: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

9

Another view of the algorithm from regularization framework

It could be easily shown that iteration result F* is equivalent to minimize the following cost function:

n

ki

n

iiik

kki

iiik YFF

DF

DwFQ

1, 1

22

11

2

1)(

With ],...,,[ 11211 ci FFFF The first term is the Smoothing Constraint: nearby points are likely to have the same label;

The second term is the Fitting Constraint: the classification results does not change too much from the initial assignment.

Page 10: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

10

Some variations of the consistency algorithm

YPIF 1* )(

2/ 1 2/ 1 1 SD D W D P

Let:

Variation (1):

Variation (2): YPIF 1* )'(

Variation (3) is Xuejun’s label iteration algorithm, where Go

is our graph: YGIF t 1)(0

* )(

Page 11: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

11

Experiment Results

(a)Digit recognition: digit 1-4 from the USPS data set

(b)Text classification: topics including autos, motorcycles, baseball and hockey from the 20-newsgroups

Page 12: 1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,

12

Conclusions

The key to semi-supervised learning problem is the consistency assumption.

The consistency algorithm proposed was demonstrated

effective on the data set considered.