deep face detector adaptation without ... - 腾讯 ai lab · deep face detector adaptation without...

8
Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad Abdullah Jamal University of Central Florida Orlando, FL 32816 a [email protected] Haoxiang Li Adobe Research San Jose, CA 95110 [email protected] Boqing Gong Tencent AI Lab Bellevue, WA 98004 [email protected] Outline Training details for the competing methods Continuous Scores on FDDB Results on COFW Qualitative Results A. Training details of the competing methods In this section, we give the details of training the com- peting baselines. The validation sets are used to determine all the free parameters. Fine-tuning [1]: For Faster-RCNN, we run the exper- iments for 7,000 iterations in total. The base learning rate is 1e-4 in the first 4,000 iterations and then re- duced to 1e-5 for the remaining 3,000 iterations. For CascadeCNN, we fine-tune the model with the learn- ing rate of 1e-4 for 10,000 iterations and another 5,000 iterations with the learning rate of 1e-5. LWF [2]: For Faster-RCNN, we train the model for 8,000 iterations. The base learning rate is set to 1e-4 for the first 6,000 iterations and then reduced to 1e- 5 for the next 2000 iterations. For CascadeCNN, we train the model for 15,000 iterations with the base learning rate of 1e-4 and reduce it to 1e-5 at the 10,000th iteration. In both the experiments, the learn- ing rate for the last layer is 10 times the base learning rate of the other layers. GDSDA [3]: We train the Faster-RCNN, and Cas- cadeCNN with a learning rate of 1e-4 for 8,000 iter- ations, and 12,000 iterations respectively. HTL [4]: We train the CascadeCNN using the learn- ing rate 1e-4 for 10,000 iterations and another 3,000 iterations using the learning rate 1e-5. Gradient Reversal [5]: We train the Faster-RCNN us- ing the base learning rate of 1e-4 for 20,000 iterations and then reduce it to 1e-5 for the next 10,000 iterations. Since the training sets of the source domain and the target domain are highly unbalanced, we alternatively take one image from either set to train the Gradient Reversal. B. Continuous scores on FDDB The FDDB dataset [6] defines both discrete and contin- uous scores to evaluate the face detection results. We have shown the ROC curves of the discrete scores of different methods in the main text. Figure 1 and Figure 2 show the ROC curves of the con- tinuous scores on FDDB for the Faster-RCNN and Cas- cadeCNN. The left panels exhibit the curves of the Faster- RCNN and the right panels show the curves for the Cas- cadeCNN. For CasacadeCNN, our approach outperforms all the competing methods in all the settings. For Faster- RCNN, our method can still boost the performance under the supervised setting and under semi-supervised setting when N =5. More importantly, our approach does not incur negative transfer, i.e., the results are either better than or about the same as the source detectors. It is actually worth pointing out that the annotations are inconsistent between the FDDB dataset and the source where the detectors are trained. As a result, the FDDB under-evaluates the adaptation methods. C. Results on COFW In addition to the FDDB dataset used in the main text, we additionally consider COFW [7] as the target domain here. COFW provides 1,345 training faces and 507 testing faces and includes heavy occlusion and large shape variations. Figure 3 and Figure 4 show the ROC curves of both dis- crete and continuous scores on COFW for both the Faster- 1

Upload: others

Post on 22-Mar-2021

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad

Deep Face Detector Adaptationwithout Negative Transfer or Catastrophic Forgetting

- Supplementary Materials -

Muhammad Abdullah JamalUniversity of Central Florida

Orlando, FL 32816a [email protected]

Haoxiang LiAdobe Research

San Jose, CA [email protected]

Boqing GongTencent AI Lab

Bellevue, WA [email protected]

Outline• Training details for the competing methods

• Continuous Scores on FDDB

• Results on COFW

• Qualitative Results

A. Training details of the competing methodsIn this section, we give the details of training the com-

peting baselines. The validation sets are used to determineall the free parameters.

• Fine-tuning [1]: For Faster-RCNN, we run the exper-iments for 7,000 iterations in total. The base learningrate is 1e-4 in the first 4,000 iterations and then re-duced to 1e-5 for the remaining 3,000 iterations. ForCascadeCNN, we fine-tune the model with the learn-ing rate of 1e-4 for 10,000 iterations and another 5,000iterations with the learning rate of 1e-5.

• LWF [2]: For Faster-RCNN, we train the model for8,000 iterations. The base learning rate is set to 1e-4for the first 6,000 iterations and then reduced to 1e-5 for the next 2000 iterations. For CascadeCNN, wetrain the model for 15,000 iterations with the baselearning rate of 1e-4 and reduce it to 1e-5 at the10,000th iteration. In both the experiments, the learn-ing rate for the last layer is 10 times the base learningrate of the other layers.

• GDSDA [3]: We train the Faster-RCNN, and Cas-cadeCNN with a learning rate of 1e-4 for 8,000 iter-ations, and 12,000 iterations respectively.

• HTL [4]: We train the CascadeCNN using the learn-ing rate 1e-4 for 10,000 iterations and another 3,000iterations using the learning rate 1e-5.

• Gradient Reversal [5]: We train the Faster-RCNN us-ing the base learning rate of 1e-4 for 20,000 iterationsand then reduce it to 1e-5 for the next 10,000 iterations.Since the training sets of the source domain and thetarget domain are highly unbalanced, we alternativelytake one image from either set to train the GradientReversal.

B. Continuous scores on FDDBThe FDDB dataset [6] defines both discrete and contin-

uous scores to evaluate the face detection results. We haveshown the ROC curves of the discrete scores of differentmethods in the main text.

Figure 1 and Figure 2 show the ROC curves of the con-tinuous scores on FDDB for the Faster-RCNN and Cas-cadeCNN. The left panels exhibit the curves of the Faster-RCNN and the right panels show the curves for the Cas-cadeCNN. For CasacadeCNN, our approach outperformsall the competing methods in all the settings. For Faster-RCNN, our method can still boost the performance underthe supervised setting and under semi-supervised settingwhen N = 5. More importantly, our approach does notincur negative transfer, i.e., the results are either better thanor about the same as the source detectors.

It is actually worth pointing out that the annotationsare inconsistent between the FDDB dataset and the sourcewhere the detectors are trained. As a result, the FDDBunder-evaluates the adaptation methods.

C. Results on COFWIn addition to the FDDB dataset used in the main text, we

additionally consider COFW [7] as the target domain here.COFW provides 1,345 training faces and 507 testing facesand includes heavy occlusion and large shape variations.

Figure 3 and Figure 4 show the ROC curves of both dis-crete and continuous scores on COFW for both the Faster-

1

Page 2: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad

RCNN and CascadeCNN face detectors under the super-vised setting. We can draw the same observation as onthe FDDB dataset, that our method shows no negativetransfer as compared to other competing methods for boththe Faster-RCNN and CascadeCNN detectors. GDSDA isan exception among the competing methods and leads tono negative transfer for CascadeCNN (but not for Faster-RCNN).

We also evaluate the catastrophic forgetting when the de-tectors are adapted to COFW. Figure 5 shows the perfor-mance on the validation set of the WIDER Face for Faster-RCNN. Our approach maintains a good performance in thesource domain compared with the original source detector.

D. More qualitative resultsWe show more qualitative results in Figure 6 and Fig-

ure 7. Our method is able to discard some of the false pos-itives from both the source detectors. We can also observethat our method is able to detect the true positives that havenot been detected by the source detectors.

References[1] Ruslan Salakhutdinov and Geoffrey Hinton. Deep

boltzmann machines. In Artificial Intelligence andStatistics, pages 448–455, 2009. 1

[2] Zhizhong Li and Derek Hoiem. Learning without for-getting. CoRR, abs/1606.09282, 2016. 1

[3] Shuang Ao, Xiang Li, and Charles X Ling. Fast gen-eralized distillation for semi-supervised domain adap-tation. In AAAI, 2017. 1

[4] Ilja Kuzborskij and Francesco Orabona. Stability andhypothesis transfer learning. In ICML (3), pages 942–950, 2013. 1

[5] Yaroslav Ganin and Victor Lempitsky. Unsuperviseddomain adaptation by backpropagation. In Interna-tional Conference on Machine Learning, pages 1180–1189, 2015. 1

[6] Vidit Jain and Erik Learned-Miller. Fddb: A benchmarkfor face detection in unconstrained settings. Technicalreport, 2010. 1

[7] Xavier P. Burgos-Artizzu, Pietro Perona, and PiotrDollar. Robust face landmark estimation under occlu-sion. In Proceedings of the 2013 IEEE InternationalConference on Computer Vision, ICCV ’13, pages1513–1520, Washington, DC, USA, 2013. IEEE Com-puter Society. 1

Page 3: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad

0.4

0.5

0.6

0.7

0.8

0.9

1

0 50 100 150 200 250 300 350 400

Tru

e p

ositiv

e r

ate

Total False positives

0 Folds (Unsupervised Adaptation)

OursSource

GP 0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 10 20 30 40 50 60 70 80

Tru

e p

ositiv

e r

ate

Total False positives

0 fold (Unsupervised Adaptation)

OursSource

GP

0.4

0.5

0.6

0.7

0.8

0.9

1

0 50 100 150 200 250 300 350 400

Tru

e p

ositiv

e r

ate

Total False positives

6 Folds (Supervised Adaptation)

OursSource

LWFFine-tuning

GDSDAGradient Reversal

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 10 20 30 40 50 60 70 80

Tru

e p

ositiv

e r

ate

Total False positives

6 Folds (Supervised Adaptation)

OursSource

LWFFine-tuning

GDSDAHTL

0.4

0.5

0.6

0.7

0.8

0.9

1

0 50 100 150 200 250 300 350 400

Tru

e p

ositiv

e r

ate

Total False positives

3 Folds (Semi-supervised Adaptation)

OursSource

Fine-tuneFine-tune+GP

GDSDA 0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 10 20 30 40 50 60 70 80

Tru

e p

ositiv

e r

ate

Total False positives

3 Folds (Semi-supervised Adaptation)

OursSource

Fine-tuneFine-tune+GP

GDSDA

Figure 1. Continuous score results on the FDDB under unsupervised, supervised, and semi-supervised settings (3 out of 6 folds of trainingimages annotated). (Left: Faster-RCNN, Right: CascadeCNN)

Page 4: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad

0.4

0.5

0.6

0.7

0.8

0.9

1

0 50 100 150 200 250 300 350 400

Tru

e p

ositiv

e r

ate

Total False positives

1 Fold (Semi-supervised Adaptation)

OursSource

Fine-tuneFine-tune+GP

GDSDA 0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 50 100 150 200 250 300 350 400

Tru

e p

ositiv

e r

ate

Total False positives

1 Fold (Semi-supervised Adaptation)

OursSource

Fine-tuneFine-tune+GP

GDSDA

0.4

0.5

0.6

0.7

0.8

0.9

1

0 50 100 150 200 250 300 350 400

Tru

e p

ositiv

e r

ate

Total False positives

5 Folds (Semi-supervised Adaptation)

OursSource

Fine-tuneFine-tune+GP

GDSDA 0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 10 20 30 40 50 60 70 80

Tru

e p

ositiv

e r

ate

Total False positives

5 Folds (Semi-supervised Adaptation)

OursSource

Fine-tuneFine-tune+GP

GDSDA

Figure 2. Continuous score results under the semi-supervised settings with N = {1, 5} out of 6 folds training images annotated on theFDDB dataset. (Left: Faster-RCNN, Right: CascadeCNN)

Page 5: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 10 20 30 40 50 60 70 80

Tru

e p

ositiv

e r

ate

Total False positives

Discrete Score

OursSource

Fine-tuningLWF

GDSDA 0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 10 20 30 40 50 60 70 80

Tru

e p

ositiv

e r

ate

Total False positives

Continuous Score

OursSource

Fine-tuningLWF

GDSDA

Figure 3. ROC Curves on COFW using FasterRCNN (supervised adaptation).

0.1 0.15

0.2 0.25

0.3 0.35

0.4 0.45

0.5 0.55

0.6 0.65

0.7 0.75

0.8 0.85

0.9 0.95

1

0 10 20 30 40 50 60 70 80 90 100

Tru

e p

ositiv

e r

ate

Total False positives

Discrete Score

OursSource

Fine-tuningLWF

GDSDAHTL

0.1 0.15

0.2 0.25

0.3 0.35

0.4 0.45

0.5 0.55

0.6 0.65

0.7 0.75

0.8 0.85

0.9 0.95

1

0 10 20 30 40 50 60 70 80 90 100

Tru

e p

ositiv

e r

ate

Total False positives

Continuous Score

OursSource

Fine-tuningLWF

GDSDAHTL

Figure 4. ROC Curves on COFW using CascadeCNN (supervised adaptation).

(a) Easy Set (b) Medium Set (c) Hard Set

Figure 5. Evaluation of catastrophic forgetting on source domain after supervised adaptation to target domain (COFW): detection resultson the validation set of WIDER FACE (Easy, Medium and Hard sets).

Page 6: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad

Figure 6. Qualitative results of adapting Faster-RCNN. The image on the left of each pair shows the detection results by the source modeland the right image shows our method in the supervised adaptation setting.

Page 7: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad
Page 8: Deep Face Detector Adaptation without ... - 腾讯 AI Lab · Deep Face Detector Adaptation without Negative Transfer or Catastrophic Forgetting - Supplementary Materials - Muhammad

Figure 7. More qualitative results, from left to right are face detection results on FDDB dataset with a CascadeCNN (1), the same detectorbut adapted by our method to the target domain (FDDB) with no data annotation (2), and with some data annotation (3).