annotation-efficient learningyaliu/files/annotation... · 2020. 9. 13. · base-learner initializer...

36
Annotation-Efficient Learning: Class-Incremental Learning and Few-Shot Learning Yaoyao Liu 11 Sept. 2020

Upload: others

Post on 18-Mar-2021

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Annotation-Efficient Learning: Class-Incremental Learning and Few-Shot Learning

Yaoyao Liu

11 Sept. 2020

Page 2: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Outline of today’s talk

1. Class-Incremental Learning Mnemonics Training: Multi-Class Incremental Learning without Forgetting CVPR 2020

2. Few-Shot Learning An Ensemble of Epoch-wise Empirical Bayes for Few-shot Learning ECCV 2020

2

Page 3: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Background: class-incremental learning

Phase 1

Data 1

Classifier

train

Test for Data 1

3

Page 4: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Background: class-incremental learning

Phase 2

Classifier

train

Test for Data 1

Classifier

Data 2

Test for Data 1+2

4

Page 5: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Background: class-incremental learning

Phase 3

Classifier

Test for Data 1

Classifier

Test for Data 1+2

Classifier

train

Data 3

Test for Data 1+2+3

5

Page 6: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Background: Class-Incremental Learning

Phase 3

Classifier

Test for Data 1

Classifier

Test for Data 1+2

Classifier

train

Data 3

Test for Data 1+2+3

overfit

Challenge: catastrophic forgetting

6

Page 7: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Literature review

Technique 1: Replay samples for the old classes:

iCaRL[1], IL2M[2], ...

References[1] Rebuffi, Sylvestre-Alvise, et al. "icarl: Incremental classifier and representation learning." CVPR 2017;[2] Belouadah, Eden, and Adrian Popescu. "Il2m: Class incremental learning with dual memory." CVPR 2019;[3] Li, Zhizhong, and Derek Hoiem. "Learning without forgetting." TPAMI 2017;[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019;[5] Douillard, Arthur, et al. "PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning." ECCV 2020.

Technique 2: Preserve the knowledge for the old model:

LwF[3], LUCIR[4], PODNet[5]...

Mnemonics exemplars

Weight transfer operations

7

Page 8: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Replay samples for the old classes

Phase 1

Data 1

Classifier

train

Test for Data 1

Exemplar 1

8

Page 9: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Replay samples for the old classes

Phase 2

train

Test for Data 1

Data 2

Test for Data 1+2

Exemplar 1 Memory limitation

Question: How to extract the exemplars?

ClassifierClassifier

9

Page 10: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Question: how to extract the exemplars?

Existing methods:

E.g., herding[1, 4, 6] : select the samples near the average embedding

References[1] Rebuffi, Sylvestre-Alvise, et al. "icarl: Incremental classifier and representation learning." CVPR 2017;[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019;[6] Wu, Yue, et al. “Large scale incremental learning.” CVPR 2019.

Limitations for existing methods:- Heuristic selection, not performance-based - Select from finite sets (real images)

Our method:

Question: Can we generate the optimal exemplars?

GenerationAlgorithm

Benefits for our method:+ Optimal selection by end-to-end training+ Select from continuous (infinite) synthetic data

Mnemonics exemplars

10

Page 11: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Question: how to formulate the optimization of the exemplars?

11

Page 12: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Literature review

Technique 1: Replay samples for the old classes:

iCaRL[1], IL2M[2], ...

References[1] Rebuffi, Sylvestre-Alvise, et al. "icarl: Incremental classifier and representation learning." CVPR 2017;[2] Belouadah, Eden, and Adrian Popescu. "Il2m: Class incremental learning with dual memory." CVPR 2019;[3] Li, Zhizhong, and Derek Hoiem. "Learning without forgetting." TPAMI 2017;[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019;[5] Douillard, Arthur, et al. "PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning." ECCV 2020.

Technique 2: Preserve the knowledge for the old model:

LwF[3], LUCIR[4], PODNet[5]...

Mnemonics exemplars

Weight transfer operations

12

Page 13: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Preserve the knowledge for the old model

Phase 1

Data 1

train

Test for Data 1

Exemplar 1

Classifier

13

Page 14: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Preserve the knowledge for the old model

Phase 2

New model

Data 2

Exemplar 1

Old model (frozen)

0.1 0.2 0.7 0.0 0.2 0.8 1.0 0.0 Predictions for old classes Groundtruth for new classes

Distillation loss Classification loss

Motivation: Force the new model and the old model to be similar

14

Page 15: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Preserve the knowledge for the old model

iCaRL[1] (CVPR 2017) Distillation on predictions

References[1] Rebuffi, Sylvestre-Alvise, et al. "icarl: Incremental classifier and representation learning." CVPR 2017;[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019;[5] Douillard, Arthur, et al. "PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning." ECCV 2020;[7] Taylor, Matthew E., and Peter Stone. "An introduction to intertask transfer for reinforcement learning." Ai Magazine 32.1 (2011): 15-15.

LUCIR[4] (CVPR 2019) Distillation on the final feature maps

PODNet[5] (ECCV 2020) Distillation on the feature maps from all layers

Distillation: preserve high-level knowledge for the old model

It is better to transfer low-level knowledge among tasks…[7]

Q: Can we preserve the low-level knowledge for the old model?

15

Page 16: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

How to transfer low-level knowledge for class-incremental learning?

Weight transfer operations:Channel-wise masks

16

Page 17: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

The data sequence from different classes

BO

P

BO

P

……

Global computing glow

Weight transfer operations

17

Our method: Technique 1 + Technique 2

BOP = Bilevel Optimization Program

Page 18: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Our method boosts the performance

Ours

LUCIR[4], BiC[6]

References[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019;[6] Wu, Yue, et al. “Large scale incremental learning.” CVPR 2019.

Dataset: ImageNet-Subset

18

Joint training(Upper bound)

Page 19: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Our method boosts the performance

References[1] Rebuffi, Sylvestre-Alvise, et al. "icarl: Incremental classifier and representation learning." CVPR 2017;[3] Li, Zhizhong, and Derek Hoiem. "Learning without forgetting." TPAMI 2017;[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019;[6] Wu, Yue, et al. “Large scale incremental learning.” CVPR 2019.

[1]

[2]

[6]

[4]

● Generic● Boost the performance for FOUR different baselines

19

Page 20: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Ablation study

References[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019.

20

Page 21: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

t-SNE results: clearer separation in the data

References[1] Rebuffi, Sylvestre-Alvise, et al. "icarl: Incremental classifier and representation learning." CVPR 2017;[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019.

[1, 4]

Dataset: ImageNet

21

Page 22: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Outline of today’s talk

1. Class-Incremental Learning Mnemonics Training: Multi-Class Incremental Learning without Forgetting CVPR 2020

2. Few-Shot Learning An Ensemble of Epoch-wise Empirical Bayes for Few-shot Learning ECCV 2020

22

Page 23: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Research background

23

• Limitation: most algorithms are based on supervised learning,

                        so we need lots of labeled samples to train the model

(Image from Yao Lu)

Medical images:expensive to label the data

Mitosis detection有丝分裂检测

Page 24: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Few-shot learning: learning with limited data

24

Question: how to learn a model with limited labeled data?Task: few-shot image classification

Seen classesMany-shot

Unseen classesFew-shot

(Images from Ravi and Larochelle)

Page 25: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Review: meta-learning

Seen classesMany-shot

(Images from Ravi and Larochelle)

Seen classes

Unseen classesFew-shot

reorganize

Same class number, same shot number

25

Page 26: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Review: meta-learning

(Images from Ravi, Larochelle)

Seen classes

Unseen classes

Training tasks

Test task

Meta-train Meta-test

26

Page 27: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Existing methods vs. our E3BM

27

Existing methods:

• A single base-learner• Arbitrary base-learning hyperparameters- Unstable

Our E3BM:

• An ensemble of multiple base-learners• Task-specific base-learning hyperparameters+ Stable and robust

(a) MAML [8]

θSGD

SIB

θ(c) SIB [10] (d) E3BM (ours)

θE3 BM

(b) MTL [9]

SGD

References[8] Finn, Chelsea, Pieter Abbeel, and Sergey Levine. “Model-agnostic meta-learning for fast adaptation of deep networks.” ICML 2017;[9] Sun, Qianru, et al. “Meta-transfer learning for few-shot learning.” CVPR 2019;[10] Hu, Shell Xu, et al. “Empirical Bayes Transductive Meta-Learning with Synthetic Gradients.” ICLR 2020.

Page 28: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Existing method: MAML[8]

28

meta update

Learning rate Combination weightBase-learner initializer

...

...

Epoch-wise base-learner

Predictions from a single base-learnerFor one training task:

References[8] Finn, Chelsea, Pieter Abbeel, and Sergey Levine. “Model-agnostic meta-learning for fast adaptation of deep networks.” ICML 2017.

Page 29: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Existing method: MAML[8]

29

meta update

Learning rate Combination weightBase-learner initializer

...

...

Epoch-wise base-learner

Predictions from a single base-learner

Arbitrary base-learning hyperparameters

For one training task:

References[8] Finn, Chelsea, Pieter Abbeel, and Sergey Levine. “Model-agnostic meta-learning for fast adaptation of deep networks.” ICML 2017.

Page 30: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Our method: E3BM framework

30

deploy

...

met

a up

date

Learning rate Combination weightBase-learner initializer

...

...

...deploy

...

Epoch-wise base-learner

Hyperprior Learner

...

Predictions from multiple base-learnersFor one training task:

Page 31: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Our method: E3BM framework

31

deploy

...

met

a up

date

Learning rate Combination weightBase-learner initializer

...

...

...deploy

...

Epoch-wise base-learner

Hyperprior Learner

...

Predictions from multiple base-learners

Task-specific base-learning hyperparameters

For one training task:

Page 32: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

The architecture of the hyperprior learner

mea

n

conc

at

mea

n

FC

FC

(a) Epoch-independent

For the m-th base epoch:

Page 33: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Boost the performance on THREE baselines

33

The 5-class few-shot classification results (%).

References[8] Finn, Chelsea, Pieter Abbeel, and Sergey Levine. “Model-agnostic meta-learning for fast adaptation of deep networks.” ICML 2017;[9] Sun, Qianru, et al. “Meta-transfer learning for few-shot learning.” CVPR 2019;[10] Hu, Shell Xu, et al. “Empirical Bayes Transductive Meta-Learning with Synthetic Gradients.” ICLR 2020.

Page 34: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Open-source resources

1. Class-Incremental Learning Mnemonics Training: Multi-Class Incremental Learning without Forgetting GitHub: https://github.com/yaoyao-liu/mnemonics-training

2. Few-Shot Learning An Ensemble of Epoch-wise Empirical Bayes for Few-shot Learning GitHub: https://github.com/yaoyao-liu/e3bm

34

Page 35: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

Thanks!Any questions?

Yaoyao [email protected]

Page 36: Annotation-Efficient Learningyaliu/files/annotation... · 2020. 9. 13. · Base-learner initializer Learning rate Combination weight..... Epoch-wise base-learner Predictions from

References[1] Rebuffi, Sylvestre-Alvise, et al. "iCaRL: Incremental classifier and representation learning." CVPR 2017;[2] Belouadah, Eden, and Adrian Popescu. "Il2m: Class incremental learning with dual memory." CVPR 2019;[3] Li, Zhizhong, and Derek Hoiem. "Learning without forgetting." TPAMI 2017;[4] Hou, Saihui, et al. "Learning a unified classifier incrementally via rebalancing." CVPR 2019;[5] Douillard, Arthur, et al. "PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning." ECCV 2020;[6] Wu, Yue, et al. “Large scale incremental learning.” CVPR 2019;[7] Taylor, Matthew E., and Peter Stone. "An introduction to intertask transfer for reinforcement learning." Ai Magazine 32.1 (2011): 15-15;[8] Finn, Chelsea, Pieter Abbeel, and Sergey Levine. “Model-agnostic meta-learning for fast adaptation of deep networks.” ICML 2017;[9] Sun, Qianru, et al. “Meta-transfer learning for few-shot learning.” CVPR 2019;[10] Hu, Shell Xu, et al. “Empirical Bayes Transductive Meta-Learning with Synthetic Gradients.” ICLR 2020;[11] Sun, Qianru, et al. "Meta-transfer learning for few-shot learning." CVPR 2019;[12] Liu, Yaoyao, et al. "Mnemonics Training: Multi-Class Incremental Learning without Forgetting." CVPR 2020;[13] Sun, Qianru, et al. "Meta-Transfer Learning through Hard Tasks." TPAMI 2020;[14] Liu, Yaoyao, Bernt Schiele, and Qianru Sun. "An Ensemble of Epoch-wise Empirical Bayes for Few-shot Learning." ECCV 2020.