learning semantics with less supervision. agenda beyond fixed keypoints beyond keypoints open...

51
Learning Semantics with Less Supervision

Upload: christopher-mansell

Post on 01-Apr-2015

219 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Learning Semantics with Less Supervision

Page 2: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Agenda

• Beyond Fixed Keypoints• Beyond Keypoints• Open discussion

Page 3: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Part Discovery from Partial Correspondence

[Subhransu Maji and Gregory Shakhnarovich, CVPR 2013]

Page 4: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion
Page 5: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Keypoints in diverse categories

Where are the keypoints?Can you name them?

Page 6: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Does the name of a keypoint matter?

Maji and Shakhnarovich HCOMP’12

We can mark correspondences without naming parts

Page 7: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Annotation interface on MTurk

Example landmarks are provided:

Page 8: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Example annotations

Annotators mark 5 landmark pairs on average

Page 9: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Are the landmarks consistent across annotators?

Yes

Page 10: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Semantic part discovery

Given a window in the first image we can find the corresponding window in the second image

Page 11: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

propagate correspondence in the “semantic graph”

Page 12: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Semantic part discovery

Discover parts using breadth-first traversal

Iter 0 Iter 1 Iter 2

Page 13: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

The semantic graph alone is not good enough

Trained using latent LDAscale, translation, membership

Graph only Graph + appearance

Page 14: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Semantic part discovery

Graph only Graph + Appearance

Page 15: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Examples of learned parts

Page 16: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Part-based representation

other activations on the training setimage

Page 17: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Part-based representation

other activations on the training setimage

Page 18: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Detecting church buildings: individual parts

better seeds

graph mining

Page 19: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Detecting church buildings: collection of parts

• Detection is challenging due to structural variability• Latent LDA parts + voting AP=39.9%, DPM AP=34.7%

Page 20: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Label Transfer

Ask users to label parts where it makes sense:

-> arch

-> tower

-> window

Transfer labels on test images:

Page 21: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Agenda

• Beyond Fixed Keypoints• Beyond Keypoints• Open Discussion

Page 22: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Unsupervised Discovery of Mid-Level Discriminative Patches

Sarubh Singh, Abhinav Gupta and Alexei Efros, ECCV12

Page 23: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Can we get nice parts without supervision?

• Idea 0: K-means clustering in HOG space

Page 24: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Still not good enough

• The SVM memorizes bad examples and still scores them highly

• However, the space of bad examples is much more diverse

• So we can avoid overfitting if we train on a training subset but look for patches on a validation subset

Page 25: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Why K-means on HOG fails?

• Chicken & Egg Problem– If we know that a set of patches are visually

similar we can easily learn a distance metric for them

– If we know the distance metric, we can easily find other members

Page 26: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Idea 1: Discriminative Clustering

• Start with K-Means• Train a discriminative classifier for the distance

function, using all other classes as negative examples

• Re-assign patches to clusters whose classifier gives highest score

• Repeat

Page 27: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Idea 2: Discriminative Clustering+

• Start with K-Means or kNN• Train a discriminative classifier for the distance

function, using Detection• Detect the patches and assign to top k clusters• Repeat

Page 28: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Can we get good parts without supervision?

• What makes a good part?– Must occur frequently in one class

(representative)– Must not occur frequently in all classes

(discriminative)

Page 29: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Discriminative Clustering+

Page 30: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Discriminative Clustering+

Page 31: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Idea 3: Discriminative Clustering++

• Split the discovery dataset into two equal parts (training and validation)

• Train on the training subset• Run the trained classifier on the validation set

to collect examples• Exchange training and validation sets• Repeat

Page 32: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Discriminative Clustering++

Page 33: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Doublets: Discover second-order relationships

• Start with high-scoring patches• Find spatial correlations to other (weaker patches)• Rank the potential doublets on validation set

Page 34: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Doublets

Page 35: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

AP on MIT Indoor-67 scene recognition dataset

Page 36: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Blocks that shout: Distinctive Parts for Scene Classification

Juneja, Vedaldi, Jawahar and Zisserman, CVPR13

bookstore

buffet

computer room

closet

Page 37: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Three steps

• Seeding (proposing initial parts)

• Expansion (learning part detectors)

• Selection (identifying good parts)

Page 38: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Step 1: Seeding

• Segment the image• Find proposal regions based on “objectness”• Compute HOG features for each

Page 39: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Step 2: Expansion

• Train Exemplar SVM for each seed region [Malisiewitz et al]

• Apply it on validation set to collect more examples

• Retrain and repeat

Page 40: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Step 3: Selection

• Good parts should occur frequently in small number of classes but infrequently in the rest

• Collect top 5 parts from each validation image, sort occurrences of each part by score and keep the top r

• Compute the entropy for each part over the class distribution. Retain lowest-entropy parts

• Filter out any parts too similar to others (based on cosine similarity of their SVM weights)

Page 41: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Features and learning

• Features: Explored Dense RootSIFT, BoW, LLS, Improved Fisher Vectors

• Non-linear SVM (sqrt kernel)

Page 42: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Results on MIT Indoor-67

Singh et al Juneja et al

Seeding K-means on HOG Exemplar SVM

Feature space HOG IFV

SVM Linear Non-linear

Selection Purity & discriminativeness(penalizes parts that perform well for multiple clusters)

Entropy rank(allows for parts that work for multiple clusters)

AP on MIT 67 49.4 61.1

Page 43: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Learning Collections of Parts for Object Recognition

[Endres, Shih, Jiaa and Hoiem, CVPR13]

Page 44: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Overview of the method

• Seeding: Random samples including full bounding box and sub-window boxes

• Expanding: Exemplar SVM, fast training (using LDA)• Selection:

• Greedy method, pick parts that require each training example to be explained by a part

• Appearance Consistency: Include parts that have high SVM score• Spatial Consistency: Prefer parts that come from the same location

within bounding box• Training and Detection:

• Boosting over Category Independent Object Proposals [Endres & Hoiem]

Page 45: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Results on PASCAL 2010 detection

Averages of patches on the top 15 detections on the validation set for a set of parts

Page 46: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Agenda

• Beyond Fixed Keypoints• Beyond Keypoints• Open Discussion

Page 47: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Gender Recognition on Labeled Faces in the Wild

Method Gender AP

Kumar et al, ICCV 2009 95.52

Frontal Face poselet 96.43

Much easier dataset – no occlusion, high resolution, centered frontal faces

[Zhang et al, arXiv:1311.5591]

Page 48: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Gender Recognition on Labeled Faces in the Wild

Method Gender AP

Kumar et al, ICCV 2009 95.52

Frontal Face poselet 96.43

Poselets + Deep Learning 99.54

Much easier dataset – no occlusion, high resolution, centered frontal faces

Male of female?

[Zhang et al, arXiv:1311.5591]

Page 49: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Poselets vs DPMs vs Discriminative Patches

DPMs Poselets Discriminative Patches

Approach Parametric Non-parametric Non-parametric

Speed Faster (fewer types) Slower Slower (many types)

Redundancy Little A lot (improves accuracy)

A lot

Spatial model Sophisticated Primitive (threshold) Primitive

Supervision requirements

Needs 2 keypoints Needs more keypoints (10+)

No supervision

Uses multi-scale signal?

Two scale levels Yes, multiple scales yes

Jointly trained Yes No No

Attached semantics

Primitive Sophisticated Medium

Page 50: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Supervision in parts

unsupervised stronglysupervised

ISMSIFT

POSELETSDPMs

DISCRIMINATIVE PATCHES

Page 51: Learning Semantics with Less Supervision. Agenda Beyond Fixed Keypoints Beyond Keypoints Open discussion

Questions for open discussion• What is the future for mid-level parts?• More supervision vs less supervision?• Should low-level parts be hard-coded or

jointly trained?• Parametric vs non-parametric approaches?• Parts with/without associated semantics