scnet: learning semantic correspondence · scnet: learning semantic correspondence kai han1 rafael...

10
SCNet: Learning Semantic Correspondence Kai Han 1 Rafael S. Rezende 4,5 Bumsub Ham 2 Kwan-Yee K. Wong 1 Minsu Cho 3 Cordelia Schmid 4, * Jean Ponce 5,4 1 The University of Hong Kong 2 Yonsei Univ. 3 POSTECH 4 Inria 5 Department of Computer Science, ENS / CNRS / PSL Research University Abstract This paper addresses the problem of establishing seman- tic correspondences between images depicting different in- stances of the same object or scene category. Previous approaches focus on either combining a spatial regular- izer with hand-crafted features, or learning a correspon- dence model for appearance only. We propose instead a convolutional neural network architecture, called SCNet, for learning a geometrically plausible model for semantic correspondence. SCNet uses region proposals as match- ing primitives, and explicitly incorporates geometric con- sistency in its loss function. It is trained on image pairs ob- tained from the PASCAL VOC 2007 keypoint dataset, and a comparative evaluation on several standard benchmarks demonstrates that the proposed approach substantially out- performs both recent deep learning architectures and previ- ous methods based on hand-crafted features. 1. Introduction Our goal in this paper is to establish semantic corre- spondences across images that contain different instances of the same object or scene category, and thus feature much larger changes in appearance and spatial layout than the pic- tures of the same scene used in stereo vision, which we take here to include broadly not only classical (narrow-baseline) stereo fusion (e.g., [31, 34]), but also optical flow com- putation (e.g., [15, 33, 42]) and wide-baseline matching (e.g., [30, 43]). Due to such a large degree of variations, the problem of semantic correspondence remains very chal- lenging. Most previous approaches to semantic correspon- dence [2, 17, 20, 26, 37, 43] focus on combining an effec- tive spatial regularizer with hand-crafted features such as SIFT [28], DAISY [39] or HOG [6]. With the remarkable success of deep learning approaches in visual recognition, several learning-based methods have also been proposed for * Univ. Grenoble Alpes, Inria, CNRS, Grenoble INP, LJK, 38000 Grenoble, France. Figure 1: Learning semantic correspondence. We propose a convolutional neural network, SCNet, to learn semantic correspondence using both appearance and geometry. This allows us to handle a large degree of intra-class and scene variations. This figure shows a pair of input images (top) and a warped image (bottom) using its semantic correspon- dence by our method. (Best viewed in color.) both stereo vision [9, 12, 47, 48] and semantic correspon- dence [5, 21, 50]. Yet, none of these methods exploits the geometric consistency constraints that have proven to be a key factor to the success of their hand-crafted counter- parts. Geometric regularization, if any, occurs during post- processing but not during learning (e.g., [47, 48]). In this paper we propose a convolutional neural net- work (CNN) architecture, called SCNet, for learning ge- ometrically plausible semantic correspondence (Figure 1). Following the proposal flow approach to semantic cor- respondence of Ham et al.[11], we use object propos- als [29, 40, 51] as matching primitives, and explicitly in- corporate the geometric consistency of these proposals in our loss function. Unlike [11] with its hand-crafted fea- tures, however, we train our system in an end-to-end manner using image pairs extracted from the PASCAL VOC 2007 keypoint dataset [7]. A comparative evaluation on several 1 arXiv:1705.04043v3 [cs.CV] 17 Aug 2017

Upload: others

Post on 11-Oct-2020

10 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

SCNet: Learning Semantic Correspondence

Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1

Minsu Cho3 Cordelia Schmid4,∗ Jean Ponce5,4

1The University of Hong Kong 2Yonsei Univ. 3POSTECH 4Inria5Department of Computer Science, ENS / CNRS / PSL Research University

Abstract

This paper addresses the problem of establishing seman-tic correspondences between images depicting different in-stances of the same object or scene category. Previousapproaches focus on either combining a spatial regular-izer with hand-crafted features, or learning a correspon-dence model for appearance only. We propose instead aconvolutional neural network architecture, called SCNet,for learning a geometrically plausible model for semanticcorrespondence. SCNet uses region proposals as match-ing primitives, and explicitly incorporates geometric con-sistency in its loss function. It is trained on image pairs ob-tained from the PASCAL VOC 2007 keypoint dataset, anda comparative evaluation on several standard benchmarksdemonstrates that the proposed approach substantially out-performs both recent deep learning architectures and previ-ous methods based on hand-crafted features.

1. IntroductionOur goal in this paper is to establish semantic corre-

spondences across images that contain different instancesof the same object or scene category, and thus feature muchlarger changes in appearance and spatial layout than the pic-tures of the same scene used in stereo vision, which we takehere to include broadly not only classical (narrow-baseline)stereo fusion (e.g., [31, 34]), but also optical flow com-putation (e.g., [15, 33, 42]) and wide-baseline matching(e.g., [30, 43]). Due to such a large degree of variations,the problem of semantic correspondence remains very chal-lenging. Most previous approaches to semantic correspon-dence [2, 17, 20, 26, 37, 43] focus on combining an effec-tive spatial regularizer with hand-crafted features such asSIFT [28], DAISY [39] or HOG [6]. With the remarkablesuccess of deep learning approaches in visual recognition,several learning-based methods have also been proposed for

∗Univ. Grenoble Alpes, Inria, CNRS, Grenoble INP, LJK, 38000Grenoble, France.

Figure 1: Learning semantic correspondence. We proposea convolutional neural network, SCNet, to learn semanticcorrespondence using both appearance and geometry. Thisallows us to handle a large degree of intra-class and scenevariations. This figure shows a pair of input images (top)and a warped image (bottom) using its semantic correspon-dence by our method. (Best viewed in color.)

both stereo vision [9, 12, 47, 48] and semantic correspon-dence [5, 21, 50]. Yet, none of these methods exploits thegeometric consistency constraints that have proven to bea key factor to the success of their hand-crafted counter-parts. Geometric regularization, if any, occurs during post-processing but not during learning (e.g., [47, 48]).

In this paper we propose a convolutional neural net-work (CNN) architecture, called SCNet, for learning ge-ometrically plausible semantic correspondence (Figure 1).Following the proposal flow approach to semantic cor-respondence of Ham et al. [11], we use object propos-als [29, 40, 51] as matching primitives, and explicitly in-corporate the geometric consistency of these proposals inour loss function. Unlike [11] with its hand-crafted fea-tures, however, we train our system in an end-to-end mannerusing image pairs extracted from the PASCAL VOC 2007keypoint dataset [7]. A comparative evaluation on several

1

arX

iv:1

705.

0404

3v3

[cs

.CV

] 1

7 A

ug 2

017

Page 2: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

standard benchmarks demonstrates that the proposed ap-proach substantially outperforms both recent deep architec-tures and previous methods based on hand-crafted features.

Our main contributions can be summarized as follows:

• We introduce a simple and efficient model for learningto match regions using both appearance and geometry.

• We propose a convolutional neural network, SCNet, tolearn semantic correspondence with region proposals.

• We achieve state-of-the-art results on several bench-marks, clearly demonstrating the advantage of learningboth appearance and geometric terms.

2. Related workHere we briefly describe representative approaches re-

lated to semantic correspondence.

Semantic correspondence. SIFT Flow [26] extends clas-sical optical flow to establish correspondences across sim-ilar but different scenes. It uses dense SIFT descriptorsto capture semantic information beyond naive color val-ues, and leverages a hierarchical optimization technique ina coarse-to-fine pipeline for efficiency. Kim et al. [20]and Hur et al. [17] propose more efficient generalizationsof SIFT Flow. Instead of using SIFT features, Yang etal. [43] use DAISY [39] for an efficient descriptor extrac-tion. Inspired by an exemplar-LDA approach [13], Bris-tow et al. [2] use whitened SIFT descriptors, making se-mantic correspondence robust to background clutter. Re-cently, Ham et al. [11] introduces proposal flow that usesobject proposals as matching elements for semantic corre-spondence robust to scale and clutter. This work shows thatthe HOG descriptor gives better matching performance thandeep learning features [23, 35]. Taniai et al. [37] also useHOG descriptors, and show that jointly performing coseg-mentation and establishing dense correspondence are help-ful in both tasks. Despite differences in feature descriptorsand optimization schemes, these semantic correspondenceapproaches use a spatial regularizer to ensure flow smooth-ness on top of hand-crafted or pre-trained features.

Deep learning for correspondence. Recently, CNNshave been applied to classical dense correspondence prob-lems such as optical flow and stereo matching to learn fea-ture descriptors [46, 47, 48] or similarity functions [12, 46,47]. FlowNet [9] uses an end-to-end scheme to learn op-tical flow with a synthetic dataset, and several recent ap-proaches also use supervision from reconstructed 3D scenesand stereo pairs [12, 46, 47, 48]. MC-CNN [47] and its effi-cient extension [48] train CNN models to predict how welltwo image patches match and use this information to com-pute the stereo matching cost. DeepCompare [46] learnsa similarity function for patches directly from images ofa 3D scene, which allows for various types of geometric

and photometric transformations (e.g., rotation and illumi-nation changes). These approaches are inherently limitedto matching images of the same physical object/scene. Incontrast, Long et al. [27] use CNN features pre-trained forImageNet classification tasks (due to a lack of availabledatasets for learning semantic correspondence) with per-formance comparable to SIFT flow. To overcome the dif-ficulty in obtaining ground truth for semantic correspon-dence, Zhou et al. [50] leverage 3D models, and uses flowconsistency between 3D models and 2D images as a super-visory signal to train a CNN. Another approach to generat-ing ground truth is to directly augment the data by densi-fying sparse keypoint annotations using warping [11, 18].The universal correspondence network (UCN) of Choy etal. [5] learns semantic correspondence using an architec-ture similar to [48], but adds a convolutional spatial trans-former networks for improved robustness to rotation andscale changes. Kim et al. [21] introduce a convolutionaldescriptor using self-similarity, called fully convolutionalself-similarity (FCSS), and combine the learned semanticdescriptors with the proposal flow [11] framework. Theseapproaches to learning semantic correspondence [5, 50] orsemantic descriptors [21] typically perform better than tra-ditional hand-crafted ones. Unlike our method, however,they do not incorporate geometric consistency between re-gions or object parts in the learning process.

3. Our approachWe consider the problem of learning to match regions

with arbitrary positions and sizes in pairs of images. Thissetting is general enough to cover all cases of region sam-pling used in semantic correspondence: sampling a denseset of regular local regions as in typical dense correspon-dence [2, 20, 26, 38] as well as employing multi-scale objectproposals [1, 16, 29, 40, 51]. In this work, following pro-posal flow [11], we focus on establishing correspondencesbetween object proposal boxes.

3.1. Model

Our basic model for matching starts from the probabilis-tic Hough matching (PHM) approach of [4, 11]. In a nut-shell, given some potential match m between two regions,and the supporting data D (a set of potential matches), thePHM model can be written as

P (m|D) =∑x

P (m|x,D)P (x|D)

= Pa(m)∑x

Pg(m|x)P (x|D), (1)

where x is the offset (e.g., position and scale change) be-tween all potential matches m = [r, s] of two regions rand s. Pa(m) is the probability that the match between two

Page 3: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

regions is correct based on appearance only, and Pg(m|x)is the probability based on geometry only, computed usingthe offset x1. PHM computes a matching score by replacinggeometry prior P (x|D) with the Hough voting h(x|D) [4]:

h(x|D) =∑

m′∈DPa(m′)Pg(m′|x). (2)

This turns out to be an effective spatial matching model thatcombines appearance similarity with global geometric con-sistency measured by letting all matches vote on the poten-tial offset x [4, 11].

In our learning framework, we consider similaritiesrather than probabilities, and rewrite the PHM score for thematch m as

z(m,w) = f(m,w)∑x

g(m,x)∑

m′∈Df(m′, w)g(m′, x)

= f(m,w)∑

m′∈D[∑x

g(m,x)g(m′, x)]f(m′, w),

(3)where f(m,w) is a parameterized appearance similarityfunction between the two regions in the potential match m,x is as before an offset variable (position plus scale), andg(m,x) measures the geometric compatibility between thematch m and the offset x.

Now assuming that we have a total number of n potentialmatches, and identifying matches with their indices, we canrewrite this score as

z(m,w) = f(m,w)∑m′

Kmm′f(m′, w),

where Kmm′ =∑x

g(m,x)g(m′, x), (4)

and the n × n matrix K is the kernel matrix associatedwith the feature vector ϕ(m) = [g(m,x1), . . . , g(m,xs)]

T ,where x1 to xs form the finite set of values that the offsetvariable x runs over: indeed Kmm′ = ϕ(m) · ϕ(m′).2

Given training pairs of images with associated true andfalse matches, we can learn our similarity function by min-imizing with respect to w

E(w) =

n∑m=1

l[ym, z(m,w)] + λΩ(w), (5)

where l is a loss function, ym is the the ground-truth la-bel (either 1 [true] or 0 [false]) for the match m, and Ωis a regularizer (e.g., Ω(w) = ||w||2). We use the hingeloss and L2 regularizer in this work. Finally, at test time,we associate any region r with the region s maximizingz([r, s], w∗), where w∗ is the set of learned parameters.

1We suppose that appearance matching is independent of geometrymatching and the offset.

2Putting it all together in an n-vector of scores, this can also be rewrit-ten as z(w) = f(w)Kf(w), where z(w) = (z(1, w), . . . , z(n,w))T ,“” stands for the elementwise product between vectors, and f(w) =(f(1, w), . . . , f(n,w))T .

3.2. Similarity function and geometry kernel

There are many possible choices for the function f thatcomputes the appearance similarity of the two regions r ands making up match number m. Here we assume a trainableembedding function c (as will be shown later, c will be theoutput of a CNN in our case) that outputs a L2 normalizedfeature vector. For the appearance similarity between tworegions r and s, we then use a rectified cosine similarity:

f(m,w) = max(0, c(r, w) · c(s, w)), (6)

that sets all negative similarity values to zero, thus makingthe similarity function sparser as well as insensitive to neg-ative matches during training, with the additional benefit ofgiving nonnegative weights in Eq. (3).

Our geometry kernel Kmm′ records the fact that twomatches (roughly) correspond to the same offset: Con-cretely, we discretize the set of all possible offsets into bins.Let us denote by h the function mapping a match m ontothe corresponding bin x, we now define g by

g(m,x) =

1, if h(m) = x

0, otherwise.(7)

Thus, the kernel Kmm′ simply measures whether twomatches share the same offset bin or not:

Kmm′ =

1, if h(m) = h(m′)

0, otherwise.(8)

In practice, x runs over a grid of predefined offset values,and h(m) assigns match m to the nearest offset point. Ourkernel is sparse, which greatly simplifies the computationof the score function in Eq. (4): Indeed, let Bx denote theset of matches associated with the bin x, the score functionz reduces to

z(m,w) = f(m,w)∑

m′∈Bh(m)

f(m′, w). (9)

This trainable form of the PHM model from [4, 11] can beused within Eq. (5).

Note that since our simple geometry kernel is only de-pendent on matches’ offsets, we obtain the same geometryterm value of

∑m′∈Bh(m)

f(m′, w) for any match m thatfalls into the same bin h(m). This allows us to compute thisgeometry term value only once for each non-empty bin xand then share it for multiple matches in the same bin. Thissharing makes computing z several times faster in practice.3

3.3. Gradient-based learning

The feature embedding function c(m,w) in the modelabove can be implemented by any differentiable architec-ture, for example a CNN-based one, and the score function

3If the geometry kernel is dependent on something other than offsets,e.g., matches’ absolute position or their neighborhood structure, this shar-ing is not possible.

Page 4: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

Conv layers

(r1, …, rp)

IAFA

ROIpooling

c (r1,w)…

c (rp,w)

Conv layers

(s1, …, sp)

IBFB

ROIpooling

FC L2Norm

c (s1,w)…

c (sp,w)

z (m1,w)…

z (mn,w) ×K

ReLU

f (m1,w)…

f (mn,w) FCg L2Norm

FC L2Norm

FCg L2Norm

ReLUfg (m1,w)

…fg (mn,w)

cg (r1,w)…

cg (rp,w)

cg (s1,w)…

cg (sp,w)

Figure 2: The SCNet architectures. Three variants are proposed: SCNet-AG, SCNet-A, and SCNet-AG+. The basic architec-ture, SCNet-AG, is drawn in solid lines. Colored boxes represent layers with learning parameters and the boxes with the samecolor share the same parameters. “×K” denotes the voting layer for geometric scoring. A simplified variant, SCNet-A, learnsappearance information only by making the voting layer an identity function. An extended variant, SCNet-AG+, contains anadditional stream drawn in dashed lines. SCNet-AG learns a single embedding c for both appearance and geometry, whereasSCNet-AG+ learns an additional and separate embedding cg for geometry. See text for details. (Best viewed in color.)

z can be learned using stochastic gradient descent. Let usnow consider the problem of minimizing the objective func-tion E(w) defined by Eq. (5).4 This requires computing thegradient with respect to w of the function z:

∇z(m,w) = [∑

m′∈DKmm′f(m′, w)]∇f(m,w)

+f(m,w)∑

m′∈DKmm′∇f(m′, w). (10)

Denoting by n the size of D, this involves n evaluationsof both f and ∇f . Computing the full gradient of E thusrequires at most n2 evaluations of both f and∇f , which be-comes computationally intractable when n is large enough.The score function of Eq. (9) with the sparse kernel ofEq. (8), however, greatly reduces the gradient computation:∇z(m,w) = [

∑m′∈Bh(m)

f(m′, w)]∇f(m,w)

+f(m,w)∑

m′∈Bh(m)

∇f(m′, w). (11)

Note that computing the gradient for matchm involves onlya small set of matches falling into the same offset bin h(m).

4. SCNet architectureAmong many possible architectures implementing the

proposed model, we propose using a convolutional neuralnetwork (CNN), dubbed SCNet, that efficiently processes

4We take Ω(w) = 0 for simplicity in this section, but tackling anonzero regularizer is easy.

regions and learns our matching model. Three variants,SCNet-AG, SCNet-A, SCNet-AG+, are illustrated in Fig. 2.

In each case, SCNet takes as input two images IAand IB , and maps them onto feature maps FA and FB

by CNN layers. Given region proposals (r1, . . . , rp) and(s1, . . . , sp) for the two images, parallel ROI pooling lay-ers [10, 14] extract feature maps of the same size for eachproposal. This is an efficient architecture that shares convo-lutional layers over all region proposals.

SCNet-AG. The proposal features are fed into a fully-connected layer, mapped onto feature embedding vec-tors, and normalized into unit feature vectors c(ri, w) andc(sj , w), associated with the regions ri and sj of IA andIB , respectively. The value of f(m,w) for the match massociated with regions ri and sj is computed as the rec-tified dot product of c(ri) and c(sj) (Eq. (6)), which de-fines the appearance similarity f(m,w) for match m. Ge-ometric consistency is enforced with the kernel describedin Sec. 3.2, using a voting layer, denoted as “×K”, thatcomputes score z(m,w) from the appearance similarity andgeometric consensus of proposals. Finally, matching is per-formed by identifying the maximal z(m,w) scores, usingboth appearance and geometric similarities.

SCNet-A. We also evaluate a similar architecture withoutthe geometry term. This architecture drops the voting layer(denoted by ×K in Fig. 2) from SCNet-AG, directly usingf(m,w) as a score function. This is similar to the universalcorrespondence network (UCN) [5]. The main differencesare the use of object proposals and the use of a different lossfunction.

Page 5: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

SCNet-AG+. Unlike SCNet-AG, which learns a singleembedding c for both appearance and geometry, SCNet-AG+ learns an additional and separate embedding cg forgeometry that is implemented by an additional stream inthe SCNet architecture (dashed lines in Fig. 2). This corre-sponds to a variant of Eq. (9), as follows:

z+(m,w) = f(m,w)∑

m′∈Bh(m)

fg(m′, w), (12)

where fg is the rectified cosine similarity computed by cg .Compared to the original score function, this variant allowsthe geometry term to learn a separate embedding functionfor geometric scoring. This may be beneficial particularlywhen a match’s contribution to geometric scoring needs tobe different from the appearance score. For example, amatch of rigid object parts (wheel of cars) may contributemore to geometric scoring than that of deformable objectparts (leg of horses). The separate similarity function fgallows more flexibility in learning the geometric term.

Implementation details. We use the VGG16 [36] modelthat consists of a set of convolutional layers with 3 × 3 fil-ters, a ReLU layer and a pooling layer. We find that tak-ing the first 4 convolutional layers is a good trade-off forour semantic feature extraction purpose without loosing lo-calization accuracy. These layers output features with 512channels. For example, if the net takes input of 224×224×3images, the convolutional layers produce features with thesize of 14×14×512. For the ROI pooling layer, we choosea 7 × 7 filter following the fast R-CNN architecture [10],which produces a feature map with size of 7 × 7 × 512for each proposal. To transform the feature map for eachproposal into a feature vector, we use the FC layer with asize of 7× 7× 512× 2048. The 2048 dimensional featurevector associated with each proposal are then fed into theL2 normalization layer, followed by the dot product layer,ReLU, our geometric voting layer, and loss layer. The con-volutional layers are initialized by the pretrained weightsof VGG16 and the fully connected layers have random ini-tialization. We train our SCNet by mini-batch SGD, withlearning rate 0.001, and weight decay 0.0005. During train-ing, each mini-batch arises from a pair of images associ-ated with a number of proposals. In our implementation,we generated 500 proposals for each image, which leads to500× 500 potential matches.

For each mini-batch, we sample matches for training asfollows. (1) Positive sampling: For a proposal ri in IA, weare given its ground truth match r′i in IB . We pick all theproposals sj in IB with IoU(sj , r

′i) > Tpos to be positive

matches for ri. (2) Negative sampling: Assume we obtaink positive pairs w.r.t ri. We also need to have k negativepairs w.r.t ri. To achieve this, we first find the proposalsst in IB with IoU(st, r

′i) < Tneg . Assuming p proposals

satisfying the IoU constraint, we find the proposals with topk appearance similarity with ri among those p proposals. Inour experiment, we set Tpos = 0.6, and Tneg = 0.4.

5. Experimental evaluation

In this section we present experimental results and anal-ysis. Our code and models will be made available online:http://www.di.ens.fr/willow/research/scnet/.

5.1. Experimental details

Dataset. We use the PF-PASCAL dataset that consists of1300 image pairs selected from PASCAL-Berkeley key-point annotations5 of 20 object classes. Each pair of im-ages in PF-PASCAL share the same set of non-occludedkeypoints. We divide the dataset into 700 training pairs,300 validation pairs, and 300 testing pairs. The image pairsfor training/validation/testing are distributed proportionallyto the number of image pairs of each object class. In train-ing, we augment the data into a total of 1400 pairs by hor-izontal mirroring. We also test our trained models with thePF-WILLOW dataset [11], Caltech-101 [8] and PASCALParts [49] to further validate a generalization of the models.

Region proposal. Unless stated otherwise, we choose touse the method of Manen et al. (RP) [29]. The use of RPproposals is motivated by the superior result reported in[11], which is verified once more by our evaluation. In test-ing we use 1000 proposals for each image as in [11], whilein training we use 500 proposals for efficiency.

Evaluation metric. We use three metrics to compare theresults of SCNet to other methods. First, we use the prob-ability of correct keypoint (PCK) [44], which measures theprecision of dense flow at sparse keypoints of semantic rele-vance. It is calculated on the Euclidean distance d(φ(p), p∗)between a warped keypoint φ(p) and ground-truth one p∗6.Second, we use the probability of correct regions (PCR) in-troduced in [11] as an equivalent of the the PCK for regionbased correspondence. PCR measures the precision of a re-gion matching between region r and its correspondent r∗ onthe intersection over union (IoU) score 1 − IoU(φ(r), r∗).Both metrics are computed against a threshold τ in [0, 1]and we measure PCK@τ and PCR@τ as the percentagecorrect below τ . Third, we capture the quality of matchingproposals by the mean IoU of the top k matches (mIoU@k).Note that these metrics are used to evaluate two differenttypes of correspondence. Indeed, PCK is an evaluation met-ric for dense flow field, whereas PCR and mIoU@k are usedto evaluate region-based correspondences [11].

5http://www.di.ens.fr/willow/research/proposalflow/6To better take into account the different sizes of images, we normalize

the distance by dividing by the diagonal of the warped image, as in [5]

Page 6: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

IoU threshold0 0.2 0.4 0.6 0.8 1

PC

R

0

0.2

0.4

0.6

0.8

1PCR for SCNet architecture vs Proposal Flow

SCNet-A [0.47]SCNet-AG [0.49]SCNet-AG+ [0.50]NAM

HOG [0.32]

PHMHOG

[0.39]

LOMHOG

[0.45]

IoU threshold0 0.2 0.4 0.6 0.8 1

PC

R

0

0.2

0.4

0.6

0.8

1PCR for SCNet descriptor vs HOG

NAMSCNet

[0.47]

PHMSCNet

[0.47]

LOMSCNet

[0.52]

NAMHOG

[0.32]

PHMHOG

[0.39]

LOMHOG

[0.45]

IoU threshold0 0.2 0.4 0.6 0.8 1

PC

R

0

0.2

0.4

0.6

0.8

1PCR for SCNet vs VGG

SCNet-A [0.47]SCNet-AG [0.49]SCNet-AG+ [0.50]VGG [0.18]VGG-L2 [0.44]VGG-L2-FC [0.43]

IoU threshold0 0.2 0.4 0.6 0.8 1

PC

R

0

0.2

0.4

0.6

0.8

1PCR for object proposals (SCNet-AG+)

US [0.26]SW [0.38]SS [0.42]RP [0.50]

Number of top matches, k20 40 60 80 100

mIo

U@

k

0.2

0.3

0.4

0.5

0.6

0.7mIoU@k for SCNet architecture vs Proposal Flow

SCNet-A [61.7]SCNet-AG [64.2]SCNet-AG+ [64.4]NAM

HOG [43.9]

PHMHOG

[52.9]

LOMHOG

[58.0]

Number of top matches, k20 40 60 80 100

mIo

U@

k

0.2

0.3

0.4

0.5

0.6

0.7mIoU@k for SCNet descriptor vs HOG

NAMSCNet

[61.7]

PHMSCNet

[60.9]

LOMSCNet

[65.6]

NAMHOG

[43.9]

PHMHOG

[52.9]

LOMHOG

[58.0]

Number of top matches, k20 40 60 80 100

mIo

U@

k

00.10.20.30.40.50.60.7

mIoU@k for SCNet vs VGG

SCNet-A [61.7]SCNet-AG [64.2]SCNet-AG+ [64.4]VGG [10.1]VGG-L2 [54.7]VGG-L2-FC [52.8]

Number of top matches, k20 40 60 80 100

mIo

U@

k

0.20.30.40.50.60.7

mIoU@k for object proposals (SCNet-AG+)

US [40.1]SW [40.5]SS [57.4]RP [64.4]

(a) (b) (c) (d)

Figure 3: (a) Performance of SCNet on PF-PASCAL, compared to Proposal Flow methods [11]. (b) Performance of SCNetand HOG descriptors on PF-PASCAL, evaluated using Proposal Flow methods [11]. (c) Comparison to ImageNet-trainedbaselines. (d) Comparison of different proposals. PCR and mIoU@k plots are shown at the top and bottom, respectively.AuC is shown in the legend. (Best viewed in pdf.)

5.2. Proposal flow components

We use the PF-PASCAL dataset to evaluate regionmatching performance. This setting allows our method tobe tested against three other methods in [11]: NAM, PHMand LOM. NAM finds correspondences using handcraftedfeatures only. PHM and LOM additionally consider globaland local geometric consistency, respectively, between re-gion matchings. We also compare our SCNet-learned fea-ture against whitened HOG [6], the best performing hand-craft feature of [11]. Experiments on the PF-WILLOWdataset [11] showed similar results with the ones on the PF-PASCAL dataset. For details, refer to our project webpage.

Quantitative comparison. Figure 3(a) compares SCNetmethods with the proposal flow methods [11] on the PF-PASCAL dataset. Our SCNet models outperform theother methods that use the HOG feature. Our geomet-ric models (SCNet-AG, SCNet-AG+) substantially outper-form the appearance-only model (SCNet-A), and SCNet-AG+ slightly outperform SCNet-AG. This can also be seenfrom the area under curve (AuC) presented in the legend.This clearly show the effectiveness of deep learned featuresas well as geometric matching. In this comparison, we fixthe VGG16 layer and only learn the FC layers. In our ex-periment, we also learned all layers including VGG 16 andthe FC layers in our model (fully finetuned), but the im-provement over the partially learned model was marginal.Figure 3(b) shows the performance of NAM, PHM, LOMmatching when replacing HOG feature with our learned fea-ture in SCNet-A. We see that SCNet features greatly im-proves all the matching methods. Interestingly, LOM us-

ing SCNet feature outperforms our best performing SC-Net model, SCNet-AG+. However, the LOM method ismore than 10 times slower than SCNet-AG+: on averagethe method takes 0.21s for SCNet-A feature extraction and3.15s for the actual matching process, whereas our SCNet-AG+ only takes 0.33s in total. Most of the time in LOM isspent in computing its geometric consistency term. We fur-ther evaluated three additional baselines using ImageNet-trained VGG (see Figure 3(c)). Namely (i) VGG: We di-rectly use the features from ImageNet-trained VGG, fol-lowed by ROI-pooling to make the features for each pro-posal of the same size (7 × 7 × 512). We then flatten thefeatures into vectors of dimension 175616. (ii) VGG-L2:We l2-normalize the flattened feature of (i). (iii) VGG-L2-FC: We perform a random projection from (ii) to a featureof dimension 2048 (the same dimension with SCNet, 12.25times smaller than (i) and (ii)) by adding a randomly initial-ized FC layer on top of (ii). Note that this is exactly equiv-alent to SCNet-A without training on the target dataset.

Results with different object proposals. SCNet can becombined with any region proposal methods. In this exper-iment, we train and evaluate SCNet-AG+ on PF-PASCALwith four region proposal methods: randomized prim (RP)[29], selective search (SS) [41], random uniform sam-pling (US), and sliding window (SW). US and SW are ex-tracted using the work of [16], and SW is similar to regulargrid sampling used in other popular methods [20, 26, 33].Figure 3(d) compares matching performance in PCR andmIoU@k when using the different proposals. RP performsbest, and US performs worst with a large margin. This

Page 7: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

bike image pair NAMHOG [37] SCNet-A [104] SCNet-AG+ [107]

wine bottle image pair NAMHOG [88] SCNet-A [177] SCNet-AG+ [180]

Figure 4: Region matching examples. Numbers beside methods stand for numbers of correct matches.

Table 1: Per-class PCK on PF-PASCAL at τ = 0.1. For all methods using object proposals, we use 1000 RP proposals [29].

Method aero bike bird boat bottle bus car cat chair cow d.table dog horse moto person plant sheep sofa train tv meanNAMHOG [11] 72.9 73.6 31.5 52.2 37.9 71.7 71.6 34.7 26.7 48.7 28.3 34.0 50.5 61.9 26.7 51.7 66.9 48.2 47.8 59.0 52.5PHMHOG [11] 78.3 76.8 48.5 46.7 45.9 72.5 72.1 47.9 49.0 84.0 37.2 46.5 51.3 72.7 38.4 53.6 67.2 50.9 60.0 63.4 60.3LOMHOG [11] 73.3 74.4 54.4 50.9 49.6 73.8 72.9 63.6 46.1 79.8 42.5 48.0 68.3 66.3 42.1 62.1 65.2 57.1 64.4 58.0 62.5

UCN [5] 64.8 58.7 42.8 59.6 47.0 42.2 61.0 45.6 49.9 52.0 48.5 49.5 53.2 72.7 53.0 41.4 83.3 49.0 73.0 66.0 55.6SCNet-A 67.6 72.9 69.3 59.7 74.5 72.7 73.2 59.5 51.4 78.2 39.4 50.1 67.0 62.1 69.3 68.5 78.2 63.3 57.7 59.8 66.3

SCNet-AG 83.9 81.4 70.6 62.5 60.6 81.3 81.2 59.5 53.1 81.2 62.0 58.7 65.5 73.3 51.2 58.3 60.0 69.3 61.5 80.0 69.7SCNet-AG+ 85.5 84.4 66.3 70.8 57.4 82.7 82.3 71.6 54.3 95.8 55.2 59.5 68.6 75.0 56.3 60.4 60.0 73.7 66.5 76.7 72.2

shows that the region proposal process is an important fac-tor for matching performance.

Qualitative comparison. Region matching results forNAM, SCNet-A, and SCNet-AG+ are shown in Figure 4.In this example, at the IoU threshold 0.5, the numbers ofcorrect matches are shown for all methods. We can see thatSCNet models perform significantly better than NAM withHOG feature, and SCNet-A is outperformed by SCNet-AG+ that learns a geometric consistency term.

5.3. Flow field

Given a sparse region matching result and its corre-sponding scores, we generate dense semantic flow using adensifying technique in [11]. In brief, we select out a regionmatch with the highest score, and assign dense correspon-dences to the pixels within the matched regions by linearinterpolation. This process is repeated without replacementof the region match until we assign dense correspondencesto all pixels in the source image. The results are evaluatedon PF-PASCAL dataset. To evaluate transferability perfor-mance of the models, we also test them on other datasetssuch as PF-WILLOW [11], Caltech-101 [8] and PASCALParts [49] datasets, and compare with state-of-the-art resultson these datasets. In these cases direct comparison betweenlearning-based methods may not be fair in the sense thatthey are trained on different datasets.

Results on PF-PASCAL. We compare SCNet with Pro-posal Flow [11] and UCN [5] on the PF-PASCAL dataset,and summarize the result in Table 1. The UCN is retrainedusing the code provided by the authors on the PF-PASCALdataset for fair comparison. Using the raw network of [5]trained on a different subset of PASCAL yields as expectedlower performance, with a mean PCK of 36.0 as opposed tothe 55.6 obtained for the retrained network. The three vari-ants of SCNet do consistently better than UCN as well asall methods in [11], with a PCK of 66.3 or above. Amongall the methods, SCNet-AG+ performs best with a PCK of72.2. Figure 5 presents two examples of dense matchingfor PF-PASCAL. Ground truth are presented as circles andpredicted keypoints are presented as crosses. We observe abetter performance of SCNet-AG and SCNet-AG+.

Results on PF-WILLOW. For evaluating transferability,we test (PF-PASCAL trained) SCNet and UCN on the PF-WILLOW dataset [11] and compare the results with recentmethods in Table 2 where PCK is averaged over all classes.The postfix ‘w/SF’ and ‘w/PF’ represent that matching isperformed by SIFT Flow [26] and Proposal Flow [11], re-spectively. On this dataset where the data has a different dis-tribution, SCNet-AG slightly outperforms the A and AG+variants ([email protected]). We observe that all SCNet modelssignificantly outperform UCN, which is trained on the samedataset with the SCNet models, as well as other methods

Page 8: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

Source Target HOGNAM HOGLOM SCNet-A SCNet-AG+

Source Target HOG NAM HOG LOM SCNet-A SCNet-AG+

Source Target NAMHOG LOMHOG SCNet-A SCNet-AG+

Figure 5: Quantitative comparison of dense correspondence. We show the keypoints of the target image in circles and thepredicted keypoints of the source in crosses, with a vector that depicts the matching error. (Best viewed in pdf.)

Table 2: Fixed-threshold PCK on PF-WILLOW.

Method [email protected] [email protected] [email protected] Flow [26] 0.247 0.380 0.504DAISY w/SF [43] 0.324 0.456 0.555DeepC w/SF [46] 0.212 0.364 0.518LIFT w/SF [45] 0.224 0.346 0.489VGG w/SF [36] 0.224 0.388 0.555FCSS w/SF [21] 0.354 0.532 0.681FCSS w/PF [21] 0.295 0.584 0.715LOMHOG [11] 0.284 0.568 0.682UCN[5] 0.291 0.417 0.513SCNet-A 0.390 0.725 0.873SCNet-AG 0.394 0.721 0.871SCNet-AG+ 0.386 0.704 0.853

Table 3: Results on Caltech-101.

Methods LT-ACC IoU LOC-ERRNAMHOG [11] 0.70 0.44 0.39PHMHOG [11] 0.75 0.48 0.31LOMHOG [11] 0.78 0.50 0.26DeepFlow [33] 0.74 0.40 0.34SIFT Flow [26] 0.75 0.48 0.32DSP [20] 0.77 0.47 0.35FCSS w/SF [21] 0.80 0.50 0.21FCSS w/PF [21] 0.83 0.52 0.22SCNet-A 0.78 0.50 0.28SCNet-AG 0.78 0.50 0.27SCNet-AG+ 0.79 0.51 0.25

Table 4: Results on PASCAL Parts.

Methods IoU PCKNAMHOG [11] 0.35 0.13PHMHOG [11] 0.39 0.17LOMHOG [11] 0.41 0.17Congealing [24] 0.38 0.11RASL [32] 0.39 0.16CollectionFlow [19] 0.38 0.12DSP [20] 0.39 0.17FCSS w/SF [21] 0.44 0.28FCSS w/PF [21] 0.46 0.29SCNet-A 0.47 0.17SCNet-AG 0.47 0.17SCNet-AG+ 0.48 0.18

using hand-crafted features [26, 43, 22] and learned fea-tures [35, 46, 12, 45, 36, 21].

Results on Caltech-101. We also evaluate our approachon the Caltech-101 dataset [8]. Following the experimen-tal protocol in [20], we randomly select 15 pairs of im-ages for each object class, and evaluate matching accuracywith three metrics: Label transfer accuracy (LT-ACC) [25],the IoU metric, and the localization error (LOC-ERR) ofcorresponding pixel positions. Table 3 shows that SCNetachieves comparable results with the state of the art. Thebest performer, FCSS [21], is trained on images from thesame Caltech-101 dataset, while SCNet models are not.

Results on PASCAL Parts. Following [11], we use thedataset provided by [49] where the images are sampled fromthe PASCAL part dataset [3]. For this experiment, we mea-sure the weighted IoU score between transferred segmentsand the ground truth, with weights determined by the pixelarea of each part. To evaluate alignment accuracy, we mea-sure the PCK metric (α = 0.05) using keypoint annotations

for the PASCAL classes. Following [11] once again, weuse selective search (SS) to generate proposals for SCNetin this experiment. The results are summarized in Table 4.SCNet models outperform all other results on the datasetin IoU, and SCNet-AG+ performs best among them. FCSSw/PF [21] performs better in PCK on this dataset.

These results verify that SCNet models have successfullylearned semantic correspondence.

6. Conclusion

We have introduced a novel model for learning semanticcorrespondence, and proposed the corresponding CNN ar-chitecture that uses object proposals as matching primitivesand learns matching in terms of appearance and geometry.The proposed method substantially outperforms both recentdeep learning architectures and previous methods based onhand-crafted features. The result clearly demonstrates theeffectiveness of learning geometric matching for seman-tic correspondence. In future work, we will explore bettermodels and architectures to leverage geometric information.

Page 9: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

Acknowledgments. This work was supported by the ERCgrants VideoWorld and Allegro, the Institut Universitairede France, the National Research Foundation of Korea(NRF) grant funded by the Korea government (MSIP) (No.2017R1C1B2005584) as well as the MSIT (Ministry of Sci-ence and ICT), Korea, under the ICT Consilience Creativeprogram (IITP-2017-R0346-16-1007). We gratefully ac-knowledge the support of NVIDIA Corporation with thedonation of a Titan X Pascal GPU used for this research.We also thank JunYoung Gwak and Christopher B. Choyfor their help in comparing with UCN.

References[1] P. Arbelaez, J. Pont-Tuset, J. Barron, F. Marques, and J. Ma-

lik. Multiscale combinatorial grouping. In Proc. IEEE Conf.Comp. Vision Patt. Recog., 2014. 2

[2] H. Bristow, J. Valmadre, and S. Lucey. Dense semantic cor-respondence where every pixel is a classifier. In Proc. Int.Conf. Comp. Vision, 2015. 1, 2

[3] X. Chen, R. Mottaghi, X. Liu, S. Fidler, R. Urtasun, et al. De-tect what you can: Detecting and representing objects usingholistic models and body parts. In Proc. IEEE Conf. Comp.Vision Patt. Recog., 2014. 8

[4] M. Cho, S. Kwak, C. Schmid, and J. Ponce. Unsupervisedobject discovery and localization in the wild: Part-basedmatching with bottom-up region proposals. In Proc. IEEEConf. Comp. Vision Patt. Recog., 2015. 2, 3

[5] C. Choy, J. Gwak, S. Savarese, and M. Chandraker. Uni-versal correspondence network. In Proc. Neural Info. Proc.Systems, 2016. 1, 2, 4, 5, 7, 8

[6] N. Dalal and B. Triggs. Histograms of oriented gradients forhuman detection. In Proc. IEEE Conf. Comp. Vision Patt.Recog., 2005. 1, 6

[7] M. Everingham, L. Van Gool, C. Williams, J. Winn, andA. Zisserman. The pascal visual object classes challenge2007 (voc2007) results. 1

[8] L. Fei-Fei, R. Fergus, and P. Perona. One-shot learning ofobject categories. IEEE Trans. Patt. Anal. Mach. Intell.,28(4):594–611, 2006. 5, 7, 8

[9] P. Fischer, A. Dosovitskiy, E. Ilg, P. Hausser, C. Hazırbas,V. Golkov, P. van der Smagt, D. Cremers, and T. Brox.Flownet: Learning optical flow with convolutional networks.In Proc. IEEE Conf. Comp. Vision Patt. Recog., 2015. 1, 2

[10] R. Girshick. Fast r-cnn. In Proc. Int. Conf. Comp. Vision,2015. 4, 5

[11] B. Ham, M. Cho, C. Schmid, and J. Ponce. Proposal flow. InProc. IEEE Conf. Comp. Vision Patt. Recog., 2016. 1, 2, 3,5, 6, 7, 8

[12] X. Han, T. Leung, Y. Jia, R. Sukthankar, and A. C. Berg.MatchNet: Unifying feature and metric learning for patch-based matching. In Proc. IEEE Conf. Comp. Vision Patt.Recog., 2015. 1, 2, 8

[13] B. Hariharan, J. Malik, and D. Ramanan. Discriminativedecorrelation for clustering and classification. In Proc. Eu-ropean Conf. Comp. Vision, pages 459–472. Springer, 2012.2

[14] K. He, X. Zhang, S. Ren, and J. Sun. Spatial pyramid poolingin deep convolutional networks for visual recognition. InProc. European Conf. Comp. Vision, 2014. 4

[15] B. K. Horn and B. G. Schunck. Determining optical flow: Aretrospective. Artificial Intelligence, 59(1):81–87, 1993. 1

[16] J. Hosang, R. Benenson, P. Dollar, and B. Schiele. Whatmakes for effective detection proposals? IEEE Trans. Patt.Anal. Mach. Intell., 2015. 2, 6

[17] J. Hur, H. Lim, C. Park, and S. C. Ahn. Generalized de-formable spatial pyramid: Geometry-preserving dense cor-respondence estimation. In Proc. IEEE Conf. Comp. VisionPatt. Recog., 2015. 1, 2

[18] A. Kanazawa, D. W. Jacobs, and M. Chandraker. WarpNet:Weakly supervised matching for single-view reconstruction.In Proc. IEEE Conf. Comp. Vision Patt. Recog., 2016. 2

[19] I. Kemelmacher-Shlizerman and S. M. Seitz. Collectionflow. In Proc. IEEE Conf. Comp. Vision Patt. Recog., 2012.8

[20] J. Kim, C. Liu, F. Sha, and K. Grauman. Deformable spatialpyramid matching for fast dense correspondences. In Proc.IEEE Conf. Comp. Vision Patt. Recog., 2013. 1, 2, 6, 8

[21] S. Kim, D. Min, B. Ham, S. Jeon, S. Lin, and K. Sohn. Fcss:Fully convolutional self-similarity for dense semantic corre-spondence. In Proc. IEEE Conf. Comp. Vision Patt. Recog.,2017. 1, 2, 8

[22] S. W. Kim, D. Min, B. Ham, and K. Sohn. Dasc: Dense adap-tative self-correlation descriptor for multi-modal and multi-spectral correspondence. In Proc. IEEE Conf. Comp. VisionPatt. Recog., 2015. 8

[23] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenetclassification with deep convolutional neural networks. InProc. Neural Info. Proc. Systems, 2012. 2

[24] E. G. Learned-Miller. Data driven image models throughcontinuous joint alignment. IEEE Trans. Patt. Anal. Mach.Intell., 28(2):236–250, 2006. 8

[25] C. Liu, J. Yuen, and A. Torralba. Nonparametric scene pars-ing via label transfer. IEEE Trans. Patt. Anal. Mach. Intell.,33(12):2368–2382, 2011. 8

[26] C. Liu, J. Yuen, and A. Torralba. SIFT flow: Dense corre-spondence across scenes and its applications. IEEE Trans.Patt. Anal. Mach. Intell., 33(5):978–994, 2011. 1, 2, 6, 7, 8

[27] J. L. Long, N. Zhang, and T. Darrell. Do convnets learncorrespondence? In Proc. Neural Info. Proc. Systems, 2014.2

[28] D. G. Lowe. Distinctive image features from scale-invariantkeypoints. Int. J. of Comp. Vision, 60(2):91–110, 2004. 1

[29] S. Manen, M. Guillaumin, and L. Van Gool. Prime objectproposals with randomized Prim’s algorithm. In Proc. Int.Conf. Comp. Vision, 2013. 1, 2, 5, 6, 7

[30] J. Matas, O. Chum, M. Urban, and T. Pajdla. Robust wide-baseline stereo from maximally stable extremal regions. Im-age and vision computing, 22(10):761–767, 2004. 1

[31] M. Okutomi and T. Kanade. A multiple-baseline stereo.IEEE Trans. Patt. Anal. Mach. Intell., 15(4):353–363, 1993.1

[32] Y. Peng, A. Ganesh, J. Wright, W. Xu, and Y. Ma. Rasl:Robust alignment by sparse and low-rank decomposition for

Page 10: SCNet: Learning Semantic Correspondence · SCNet: Learning Semantic Correspondence Kai Han1 Rafael S. Rezende4,5 Bumsub Ham2 Kwan-Yee K. Wong1 Minsu Cho3 Cordelia Schmid4, Jean Ponce5,4

linearly correlated images. IEEE Trans. Patt. Anal. Mach.Intell., 34(11):2233–2246, 2012. 8

[33] J. Revaud, P. Weinzaepfel, Z. Harchaoui, and C. Schmid.Deepmatching: Hierarchical deformable dense matching.ArXiv e-prints, 2015. 1, 6, 8

[34] C. Rhemann, A. Hosni, M. Bleyer, C. Rother, andM. Gelautz. Fast cost-volume filtering for visual correspon-dence and beyond. In Proc. IEEE Conf. Comp. Vision Patt.Recog., 2011. 1

[35] E. Simo-Serra, E. Trulls, L. Ferraz, I. Kokkinos, P. Fua, andF. Moreno-Noguer. Discriminative learning of deep convo-lutional feature point descriptors. In Proc. Int. Conf. Comp.Vision, 2015. 2, 8

[36] K. Simonyan and andrew Zisserman. Very deep convolu-tional networks for large-scale visual recognition. In Proc.IEEE Conf. Comp. Vision Patt. Recog., 2014. 5, 8

[37] T. Taniai, S. N. Sinha, and Y. Sato. Joint recovery of densecorrespondence and cosegmentation in two images. In Proc.IEEE Conf. Comp. Vision Patt. Recog., 2016. 1, 2

[38] M. Tau and T. Hassner. Dense correspondences across scenesand scales. IEEE Trans. Patt. Anal. Mach. Intell., 2015. 2

[39] E. Tola, V. Lepetit, and P. Fua. Daisy: An efficient densedescriptor applied to wide-baseline stereo. IEEE Trans. Patt.Anal. Mach. Intell., 32(5):815–830, 2010. 1, 2

[40] J. R. Uijlings, K. E. van de Sande, T. Gevers, and A. W.Smeulders. Selective search for object recognition. Int. J. ofComp. Vision, 104(2):154–171, 2013. 1, 2

[41] J. R. R. Uijlings, K. E. A. van de Sande, T. Gevers, andA. W. M. Smeulders. Selective search for object recognition.Int. J. of Comp. Vision, 104(2):154–171, 2013. 6

[42] P. Weinzaepfel, J. Revaud, Z. Harchaoui, and C. Schmid.Deepflow: Large displacement optical flow with deep match-ing. In Proc. Int. Conf. Comp. Vision, 2013. 1

[43] H. Yang, W.-Y. Lin, and J. Lu. Daisy filter flow: A gener-alized discrete approach to dense correspondences. In Proc.IEEE Conf. Comp. Vision Patt. Recog., 2014. 1, 2, 8

[44] Y. Yang and D. Ramanan. Articulated human detection withflexible mixtures of parts. IEEE Trans. Patt. Anal. Mach.Intell., 35(12):2878–2890, 2013. 5

[45] K. M. Yi, E. Trulls, V. Lepetit, and P. Fua. Lift: Learnedinvariant feature transform. In Proc. European Conf. Comp.Vision, 2016. 8

[46] S. Zagoruyko and N. Komodakis. Learning to compare im-age patches via convolutional neural networks. In Proc.IEEE Conf. Comp. Vision Patt. Recog., 2015. 2, 8

[47] J. Zbontar and Y. LeCun. Computing the stereo matchingcost with a convolutional neural network. In Proc. IEEEConf. Comp. Vision Patt. Recog., 2015. 1, 2

[48] J. Zbontar and Y. LeCun. Stereo matching by training a con-volutional neural network to compare image patches. Jour-nal of Machine Learning Research, 17(1-32):2, 2016. 1, 2

[49] T. Zhou, Y. Jae Lee, S. X. Yu, and A. A. Efros. FlowWeb:Joint image set alignment by weaving consistent, pixel-wisecorrespondences. In Proc. IEEE Conf. Comp. Vision Patt.Recog., 2015. 5, 7, 8

[50] T. Zhou, P. Krahenbuhl, M. Aubry, Q. Huang, and A. A.Efros. Learning dense correspondence via 3d-guided cycle

consistency. In Proc. IEEE Conf. Comp. Vision Patt. Recog.,2016. 1, 2

[51] C. L. Zitnick and P. Dollar. Edge boxes: Locating object pro-posals from edges. In Proc. European Conf. Comp. Vision,2014. 1, 2