exploring context and visual pattern of relationship for...

5
Exploring Context and Visual Pattern of Relationship for Scene Graph Generation Wenbin Wang 1,2 , Ruiping Wang 1,2 , Shiguang Shan 1,2,3 , Xilin Chen 1,2 1 Key Laboratory of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing, 100190, China 2 University of Chinese Academy of Sciences, Beijing, 100049, China 3 Peng Cheng Laboratory, Shenzhen, 518055, China [email protected], {wangruiping, sgshan, xlchen}@ict.ac.cn In the following parts, we will firstly provide more de- tails of the models and training process. Then we provide qualitative results on VG dataset for comparing the method IMP**+Isc with IMP** under the setting of PREDCLS metric to further demonstrate the superiority of our intersec- tion region. Moreover, more qualitative samples for com- paring Mem with IMP** under the setting of PREDCLS metric and SGGEN metric respectively are also provided which show the effectiveness of relationship context. For convenience, here we briefly recall models men- tioned in the main paper again. IMP** [5](i.e. Ref. [42] in the main paper): Our base- line which uses union region to extract relationship features. We reimplement this model and re-train it us- ing our object detector. IMP** is our reimplemented version. IMP**+Isc: Based on the reimplemented model IMP**, it replaces the union region with intersection region. Mem: Our context-utilized model. It uses union re- gion to extract relationship features. 1. Models and Training Details Usage of Intersection Region. In the experiments, we compare the results between union region and intersection region. Besides, in order to explore for a better perfor- mance, we further try to combine these two types of fea- tures. The combination version of relationship features is computed as: f R comb = Conv(Conv([f Runi , f Risc ])), (1) where f Runi and f Risc represent relationship feature maps extracted by union regions and intersection regions respec- tively. [·] denotes channel-wise concatenation. The consec- utive convolutions have filters of size 3 × 3 and remain the spatial size of feature maps unchanged. Besides, to avoid that the area of intersection region is too small, we enlarge its area with factor σ. We empirically set σ to 1.2. Detector. We use Faster-RCNN [3] with VGG-16 [4] backbone as our front-end object detector. The VGG-16 is pretrained on ImageNet [1]. The detector is further fine- tuned on Visual Genome objects of 150 categories and opti- mized with SGD algorithm on a single Titan Xp GPU. Dur- ing the whole 90k iterations, the learning rate is initialized as 1.0×10 -3 and divided by 10 at 35k and 70k iterations re- spectively. The detector gets 25% mAP on Visual Genome. Once the detector is trained, we freeze its layers. End-to-End Training and Attention Assembling. With our pretrained detector, the whole framework is then trained on ground truth scene graph annotations. For each image, we firstly sample 128 RoIs (region of interest), of which 25% are foreground. The foreground RoIs are se- quentially sampled according to their degrees in the scene graph, which means that if an RoI is related with more other RoIs, it is more probable to be sampled. Next we sam- ple relationships among the sampled RoIs, of which 75% are positive. The loss is the sum of cross entropy for pred- icate classification, cross entropy for object classification, and smooth L1 loss for object location regression in each round of predictions. We optimize the model with SGD on a single Titan Xp GPU with an initial learning rate 1.0×10 -3 . Furthermore, we also try to assemble the predictions from each iteration with attention mechanism [2]. Therefore, when making a prediction in each iteration, an extra atten- tion weight is predicted at the same time. 2. Qualitative Results of Intersection Region In Fig.1, we show qualitative samples for comparing IMP** and IMP**+Isc. The results are generated under the setting of PREDCLS metric. In each pair of images, the left one is generated by IMP** while the right one is 1

Upload: others

Post on 30-May-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Exploring Context and Visual Pattern of Relationship for ...openaccess.thecvf.com/content_CVPR_2019/supplemental/Wang_Ex… · tioned in the main paper again. IMP**[5] (i.e. Ref

Exploring Context and Visual Pattern of Relationshipfor Scene Graph Generation

Wenbin Wang1,2, Ruiping Wang1,2, Shiguang Shan1,2,3, Xilin Chen1,2

1Key Laboratory of Intelligent Information Processing of Chinese Academy of Sciences (CAS),Institute of Computing Technology, CAS, Beijing, 100190, China

2University of Chinese Academy of Sciences, Beijing, 100049, China3Peng Cheng Laboratory, Shenzhen, 518055, China

[email protected], {wangruiping, sgshan, xlchen}@ict.ac.cn

In the following parts, we will firstly provide more de-tails of the models and training process. Then we providequalitative results on VG dataset for comparing the methodIMP**+Isc with IMP** under the setting of PREDCLSmetric to further demonstrate the superiority of our intersec-tion region. Moreover, more qualitative samples for com-paring Mem with IMP** under the setting of PREDCLSmetric and SGGEN metric respectively are also providedwhich show the effectiveness of relationship context.

For convenience, here we briefly recall models men-tioned in the main paper again.

• IMP** [5] (i.e. Ref. [42] in the main paper): Our base-line which uses union region to extract relationshipfeatures. We reimplement this model and re-train it us-ing our object detector. IMP** is our reimplementedversion.

• IMP**+Isc: Based on the reimplemented modelIMP**, it replaces the union region with intersectionregion.

• Mem: Our context-utilized model. It uses union re-gion to extract relationship features.

1. Models and Training DetailsUsage of Intersection Region. In the experiments, we

compare the results between union region and intersectionregion. Besides, in order to explore for a better perfor-mance, we further try to combine these two types of fea-tures. The combination version of relationship features iscomputed as:

fRcomb = Conv(Conv([fRuni , fRisc ])), (1)where fRuni and fRisc represent relationship feature mapsextracted by union regions and intersection regions respec-tively. [·] denotes channel-wise concatenation. The consec-utive convolutions have filters of size 3 × 3 and remain the

spatial size of feature maps unchanged. Besides, to avoidthat the area of intersection region is too small, we enlargeits area with factor σ. We empirically set σ to 1.2.

Detector. We use Faster-RCNN [3] with VGG-16 [4]backbone as our front-end object detector. The VGG-16 ispretrained on ImageNet [1]. The detector is further fine-tuned on Visual Genome objects of 150 categories and opti-mized with SGD algorithm on a single Titan Xp GPU. Dur-ing the whole 90k iterations, the learning rate is initializedas 1.0×10−3 and divided by 10 at 35k and 70k iterations re-spectively. The detector gets 25% mAP on Visual Genome.Once the detector is trained, we freeze its layers.

End-to-End Training and Attention Assembling.With our pretrained detector, the whole framework is thentrained on ground truth scene graph annotations. For eachimage, we firstly sample 128 RoIs (region of interest), ofwhich 25% are foreground. The foreground RoIs are se-quentially sampled according to their degrees in the scenegraph, which means that if an RoI is related with more otherRoIs, it is more probable to be sampled. Next we sam-ple relationships among the sampled RoIs, of which 75%are positive. The loss is the sum of cross entropy for pred-icate classification, cross entropy for object classification,and smooth L1 loss for object location regression in eachround of predictions. We optimize the model with SGD on asingle Titan Xp GPU with an initial learning rate 1.0×10−3.Furthermore, we also try to assemble the predictions fromeach iteration with attention mechanism [2]. Therefore,when making a prediction in each iteration, an extra atten-tion weight is predicted at the same time.

2. Qualitative Results of Intersection RegionIn Fig.1, we show qualitative samples for comparing

IMP** and IMP**+Isc. The results are generated underthe setting of PREDCLS metric. In each pair of images,the left one is generated by IMP** while the right one is

1

Page 2: Exploring Context and Visual Pattern of Relationship for ...openaccess.thecvf.com/content_CVPR_2019/supplemental/Wang_Ex… · tioned in the main paper again. IMP**[5] (i.e. Ref

by IMP**+Isc. The orange and yellow triplet items beloweach pair are those missed and wrongly detected by IMP**respectively but detected correctly by IMP**+Isc.

With the intersection region, the model is able to learnbetter representations which are closer to their real visualpatterns. Interestingly, we find some triplets relevant toanimals shown in last several samples, such as dog-on-skateboard, cat-on-towel, which do not frequently appear.Our intersection region still helps detect them successfully.It proves that our intersection region reduces distractionfrom object information and pays attention to the visual pat-tern of relationship itself.

3. Qualitative Results of Relationship Context

In Fig.2 we demonstrate qualitative results for compar-ing IMP** and Mem under the setting of PREDCLS metric.Similarly, the orange triplets are those missed by IMP** butdetected correctly by Mem. With relationship context, themodel obtains higher recall especially for images in whichlots of predicates repeat.

Finally, we give scene graph generation samples in Fig.3for comparing IMP** and Mem under the setting of SGGENmetric. Our Mem model uses both object and relationshipcontext, which makes great contributions to object and re-lationship inference.

References[1] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei.

Imagenet: A large-scale hierarchical image database. In Pro-ceedings of the IEEE Conference on Computer Vision and Pat-tern Recognition (CVPR), pages 248–255, 2009. 1

[2] V. Mnih, N. Heess, A. Graves, et al. Recurrent models of vi-sual attention. In Advances in Neural Information ProcessingSystems (NIPS), pages 2204–2212, 2014. 1

[3] S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towardsreal-time object detection with region proposal networks. InAdvances in Neural Information Processing Systems (NIPS),pages 91–99, 2015. 1

[4] K. Simonyan and A. Zisserman. Very deep convolutionalnetworks for large-scale image recognition. arXiv preprintarXiv:1409.1556, 2014. 1

[5] D. Xu, Y. Zhu, C. B. Choy, and L. Fei-Fei. Scene graph gen-eration by iterative message passing. In Proceedings of theIEEE Conference on Computer Vision and Pattern Recogni-tion (CVPR), pages 5410–5419, 2017. 1

2

Page 3: Exploring Context and Visual Pattern of Relationship for ...openaccess.thecvf.com/content_CVPR_2019/supplemental/Wang_Ex… · tioned in the main paper again. IMP**[5] (i.e. Ref

man holding surfboardperson holding glass

arm of person

woman wearing shirt, pant man wearing shoe

man on board

nose of airplane

man holding surfboardsink on counter

screen on laptop

fence behind dog dog has face, earman carrying surfboard

dog on skateboard cat on towel

Figure 1. Qualitative results of IMP** vs. IMP**+Isc. The results are generated under the setting of PREDCLS metric. Ineach pair of results, the left one is generated by IMP** while the right one is by IMP**+Isc. The orange and yellow tripletitems below each pair are those missed and wrongly detected by IMP** respectively but detected correctly by IMP**+Isc.

3

Page 4: Exploring Context and Visual Pattern of Relationship for ...openaccess.thecvf.com/content_CVPR_2019/supplemental/Wang_Ex… · tioned in the main paper again. IMP**[5] (i.e. Ref

dog-1 has ear-1

dog-1 has face

dog-2 has ear-2

woman on boat

dog has head, ear, paw1, paw-2, leg, tail

tail of plane

wing of plane

man-1 wearing shirt

man-2 has hair

bus haswindshield, door-1,door-2,door-3,

window-1, window-2, window-3

bike has tire

man wearing shirt

man wearing tie

man holding cup

man wearing shirt

bear has

eye,

nose,

mouth

Figure 2. More qualitative results of IMP** vs. Mem. The results are generated under the setting of PREDCLS metric. Ineach pair of results, the left one is generated by IMP** while the right one is by Mem. The orange triplet items are thosemissed by IMP** but detected correctly by Mem.

4

Page 5: Exploring Context and Visual Pattern of Relationship for ...openaccess.thecvf.com/content_CVPR_2019/supplemental/Wang_Ex… · tioned in the main paper again. IMP**[5] (i.e. Ref

elephant-1

trunk-1

ear

has

has

has

elephant-2 trunk-2

elephant-1

trunk-1

ear

has

has

has

elephant-2 trunk-2

ondog beach

ondog beach

man-1

man-2

surfboard

on

riding

man

wave

surfboard

on

on

man-1 short

man-3

wearing

man-2

onon

man short

wave

wearing

surfboard

onon

cow

head

nose

has

has

cow

head

nose

has

has

woman

shoe-1

shoe-2wearing

wearing

woman

shoe-1

shoe-2wearing

wearing

logo

beachkite

on

on

umbrella

polepersonon

logo beachon

onumbrella

polebagon

man

beach

surfboardholding

on

man

beach

surfboardholding

on

Figure 3. More scene graph generation samples of IMP** vs. Mem. The results are generated under the setting of SGGENmetric. In each row, the left image and scene graph are generated by IMP** while the right ones are generated by Mem. Inimages and scene graphs, red boxes are predicted and overlap with the ground truth (but the classes of red boxes in imagesmay be wrong), yellow boxes are ground truth with no match. In scene graphs, red edges are true positives, orange edges arefalse negatives, purple boxes and edges are false positives. Some yellow boxes in scene graphs which do not exist in imagesmean that they are detected correctly but the model fails in detecting their relationships with any other objects.

5