robust view interpolation with mesh cutting...middlebury 2.0 and 3d-hevc test sequences with ground...

4
Robust View Interpolation with Mesh Cutting Xiubao Jiang, Ronggang Wang, Jiajia Luo, Zhenyu Wang, Wen Gao School of Electronic and Computer Engineering, Shenzhen Graduate School, Peking University [email protected], {rgwang, wangzhen}@pkusz.edu.cn, [email protected], [email protected] Abstract—Existing view interpolation methods like DIBR require accurate disparity map which greatly limits their applications. To this aim, we propose a novel mesh-based view interpolation algorithm capable of synthesizing visually coherent virtual views with rough disparity map estimated by stereo matching algorithms. We adopt an edge-aware mesh cutting method to explicitly handle occlusion and preserve sharp depth discontinuities. Experiments on Middlebury dataset and 3D-HEVC test sequences demonstrate that proposed method outperforms DIBR and state-of-the-art mesh-based view interpolation algorithm in terms of visual quality and PSNR. Index Terms—view interpolation, mesh cutting, image-based rendering I. I NTRODUCTION View interpolation is a special case of image-based rendering that renders virtual viewpoints on the line linking the optical centers of known viewpoints [1]. This problem is addressed by two main stream approaches, i.e. depth-image based rendering (DIBR) and warping-based rendering. Provided a set of images and their corresponding disparity maps, DIBR [2] renders a novel viewpoint by re-projecting the pixels of existing images onto it. When accurate camera calibration and depth information are available, DIBR can explicitly handle occlusion caused by camera motion and render high quality virtual view. Tian et al. [2] implemented a reference software VSRS based on DIBR and proposed post- processing procedures such as ’boundary-aware splatting’ and ’smart blending’. Sensitive to error in disparity map is the main drawback of these point-based DIBR methods. Typically, erroneous disparities around depth discontinuities would cause significant artifacts in rendered views. Unfortunately, eliminating such errors is hard even for state-of-the-art stereo matching methods ([3], [4], [5]). Recent works on DIBR ([3], [6], [7], [8]) used multiple triangle meshes generated from disparity map as rendering proxies. Adopting a more compact model, mesh-based methods can preserve texture structure in the rendered view better. However, these methods have limited robustness as they either enforce hard segmentation constraint [9] or entirely rely on accurate disparity map to locate depth discontinuities. In contrast to DIBR, warp-based rendering is performed in image space. It does not require dense correspondence between input views. As long as a sparse matching of image features is available, it can render visually plausible virtual views. Stefanoski et al. [10] proposed a view synthesis system for multi-view autostereoscopic content creation. Sparse disparities and image saliency information of images were used to compute a discrete warp function. Warp based rendering cannot handle occlusion by depth comparison. Furthermore, 2-D correspondence estimation algorithms were applied in most warp based methods, such as SIFT, SURF and ORB, which does not exploit epipolar constraint. In this paper, we propose a novel view interpolation algorithm aiming to cope with the specific limitations as mentioned above. The main contributions of our work are: i) A disparity refinement algorithm is proposed to detect errors in depth estimation and correct them for plausible view interpolation. This procedure endows our algorithm with robustness against inaccuracy of stereo matching. ii) Quadrilateral mesh is splitted (if necessary) along depth discontinuities for explicit occlusion handling. Both depth and color clues are exploited for accurately locating depth discontinuities. II. PROPOSED METHOD Fig. 1 shows the workflow of proposed view interpolation algorithm. A. Stereo Matching We use stereo matching algorithms proposed in [4], [11] to compute disparity map for left and right input view. Conventional left-right consistency and ratio check is adopted to label each position in the disparity map as either ’valid’, ’invalid’ or ’occlusion’ [5]. The disparities of occlusion positions are replaced with correct disparity from the background [5]. B. Disparity Refinement In this step, quasi-dense disparities on regular grid positions are extracted based on the incomplete disparity map produced 978-1-5090-5316-2/16/$31.00 c 2016 IEEE VCIP 2016, Nov. 27 – 30, 2016, Chengdu, China

Upload: others

Post on 05-Sep-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Robust View Interpolation with Mesh Cutting...Middlebury 2.0 and 3D-HEVC test sequences with ground truth intermediate views for evaluation. We set the parameters as follows ([:] denotes

Robust View Interpolation with Mesh Cutting

Xiubao Jiang, Ronggang Wang, Jiajia Luo, Zhenyu Wang, Wen Gao

School of Electronic and Computer Engineering, Shenzhen Graduate School, Peking University

[email protected], {rgwang, wangzhen}@pkusz.edu.cn, [email protected], [email protected]

Abstract—Existing view interpolation methods like DIBRrequire accurate disparity map which greatly limits theirapplications. To this aim, we propose a novel mesh-basedview interpolation algorithm capable of synthesizing visuallycoherent virtual views with rough disparity map estimated bystereo matching algorithms. We adopt an edge-aware meshcutting method to explicitly handle occlusion and preservesharp depth discontinuities. Experiments on Middlebury datasetand 3D-HEVC test sequences demonstrate that proposedmethod outperforms DIBR and state-of-the-art mesh-based viewinterpolation algorithm in terms of visual quality and PSNR.

Index Terms—view interpolation, mesh cutting, image-basedrendering

I. INTRODUCTION

View interpolation is a special case of image-basedrendering that renders virtual viewpoints on the line linkingthe optical centers of known viewpoints [1]. This problem isaddressed by two main stream approaches, i.e. depth-imagebased rendering (DIBR) and warping-based rendering.

Provided a set of images and their corresponding disparitymaps, DIBR [2] renders a novel viewpoint by re-projectingthe pixels of existing images onto it. When accurate cameracalibration and depth information are available, DIBR canexplicitly handle occlusion caused by camera motion andrender high quality virtual view. Tian et al. [2] implemented areference software VSRS based on DIBR and proposed post-processing procedures such as ’boundary-aware splatting’ and’smart blending’. Sensitive to error in disparity map is the maindrawback of these point-based DIBR methods. Typically,erroneous disparities around depth discontinuities wouldcause significant artifacts in rendered views. Unfortunately,eliminating such errors is hard even for state-of-the-art stereomatching methods ([3], [4], [5]). Recent works on DIBR ([3],[6], [7], [8]) used multiple triangle meshes generated fromdisparity map as rendering proxies. Adopting a more compactmodel, mesh-based methods can preserve texture structurein the rendered view better. However, these methods havelimited robustness as they either enforce hard segmentationconstraint [9] or entirely rely on accurate disparity map tolocate depth discontinuities.

In contrast to DIBR, warp-based rendering is performedin image space. It does not require dense correspondencebetween input views. As long as a sparse matching of imagefeatures is available, it can render visually plausible virtualviews. Stefanoski et al. [10] proposed a view synthesissystem for multi-view autostereoscopic content creation.Sparse disparities and image saliency information of imageswere used to compute a discrete warp function. Warp basedrendering cannot handle occlusion by depth comparison.Furthermore, 2-D correspondence estimation algorithms wereapplied in most warp based methods, such as SIFT, SURFand ORB, which does not exploit epipolar constraint.

In this paper, we propose a novel view interpolationalgorithm aiming to cope with the specific limitations asmentioned above. The main contributions of our work are:

i) A disparity refinement algorithm is proposed to detecterrors in depth estimation and correct them for plausibleview interpolation. This procedure endows our algorithm withrobustness against inaccuracy of stereo matching.

ii) Quadrilateral mesh is splitted (if necessary) along depthdiscontinuities for explicit occlusion handling. Both depthand color clues are exploited for accurately locating depthdiscontinuities.

II. PROPOSED METHOD

Fig. 1 shows the workflow of proposed view interpolationalgorithm.

A. Stereo Matching

We use stereo matching algorithms proposed in [4], [11] tocompute disparity map for left and right input view.

Conventional left-right consistency and ratio check isadopted to label each position in the disparity map aseither ’valid’, ’invalid’ or ’occlusion’ [5]. The disparities ofocclusion positions are replaced with correct disparity fromthe background [5].

B. Disparity Refinement

In this step, quasi-dense disparities on regular grid positionsare extracted based on the incomplete disparity map produced

978-1-5090-5316-2/16/$31.00 c© 2016 IEEE VCIP 2016, Nov. 27 – 30, 2016, Chengdu, China

Page 2: Robust View Interpolation with Mesh Cutting...Middlebury 2.0 and 3D-HEVC test sequences with ground truth intermediate views for evaluation. We set the parameters as follows ([:] denotes

Stereo MatchingDisparity

RefinementMesh Cutting

Fig. 1. Overview of the proposed method.

in the previous step. This is achieved with a two-pass disparityrefinement algorithm as shown in Fig. 2.

In the first pass, a regular grid (stride (∆x,∆y)) is overlaidon the disparity map. Grid node N is labeled as ‘valid’ ifthe ratio of valid disparities in a (∆x,∆y) window surpassesthreshold value τ . A disparity plane is then fitted to the validdisparities and used to compute the disparity of node N .

(a)

(b)

(c)

(d) (e)

Fig. 2. (a) Incomplete disparity map. (b) Searching window in the first pass,gray pixels indicate valid disparities. (c) Graph used in the second pass. (d)Refined disparity map. (e) Original color image.

In the second pass, we assign disparities for grid nodesmarked as ‘invalid’ in the first pass. For an invalid grid nodeNS . A graph is created for all grid nodes in the W/6×H/6window centered at NS , where W and H are image width andheight respectively.A shortest walk algorithm proposed in [12]is adopted to find a target node of similar visual appearanceand assign its disparity to NS .

C. Mesh Cutting

We apply mesh cutting to quadrilaterals containing depthdiscontinuities and cut them into different depth layer.

For each quadrilateral, we first cluster the disparities of itsvertices into N disparity layers such that the absolute disparitydifference within the same layer is less than 2. N = 2suggests that the quadrilateral contains depth discontinuitiesand should be cut into two parts. If N ≥ 3, we simply

corner cut

horizontal cut

vertical cut

source node

target nodes

(a) (b)

Fig. 3. (a) Illustration of mesh cutting, where node color indicates the depthlayer it belongs to. (b) Illustration of boundary extraction.

discard corresponding quadrilateral as it contains contradicteddisparities. We observe in our experiments that only less than1% of total quadrilaterals are discarded.

For N = 2, we further categorize the mesh cutting into’vertical-cut’, ’horizontal-cut’ and ’corner-cut’ according tothe layer distribution of quadrilateral vertices, as shown inFig. 3.

Take ‘horizontal cut’ as an example to illustrate theboundary extraction procedure. First, a pixel on the twovertical sides with the largest edge strength is selected assource node. All pixels on the opposite side of source pixelare marked as target nodes. The boundary extraction isformulated as finding optimal path from source node to targetnodes with minimum path cost by dynamic programming.Specifically, The cost of the directed link(or edge) betweentwo adjacent pixels p and q are:

EdgeC(p,q) = ES(q) + λDED(p,q) + λCEC(q). (1)

ES corresponds to the edge strength of pixel q generatedby state-of-the-art edge detection algorithm [13].ED is the gradient direction term that associates high cost

for sharp changes in boundary orientation. Specifically, letG(p) be a unit vector of gradient direction at pixel p, denoteG′(p) as the unit vector perpendicular to G(p) (as shown inFig. 3) and Ang(v1,v2) denote the angle between two vectorsv1, v2.The formulation of ED is:

ED(p,q) = Ang(G′(p),p− q) +Ang(G′(q),p− q). (2)

EC is the color modeling term. We use GMMs(Gaussian Mixture Model) with K components (K=3 inour implementation) to model ’Inside’ and ’Outside’ color

Page 3: Robust View Interpolation with Mesh Cutting...Middlebury 2.0 and 3D-HEVC test sequences with ground truth intermediate views for evaluation. We set the parameters as follows ([:] denotes

properties respectively. More specifically, for a pixel p, aninside pixel p1 is sampled a distance h from p in the gradientdirection ( e.g. p1 = p + h ·G(p)) and an outside pixel p2 issampled an equal distance in the opposite direction. EC isformulated as:

EC = GMM Inside(p1) +GMM Outside(p2), (3)

where

GMM Inside(p1) =

K∑i=1

−log(p(p1|θi))− log(πi), (4)

where p(·) corresponds to Gaussian probability distributionand πi and θi are parameters of ’Inside’ GMM color model.As shown in Fig. 3, given the source node and its gradientdirection, we can identify vertex B as an inside pixel.Pixels of the same disparity layer with B in a W/6 × H/6window are sampled to train the GMM parameters as in [14].GMM Outside is formulated in the similar way.

Compared to color segmentation based method ([3], [6],[7]), the proposed method incorporates both color and depthclues to accurately locate depth discontinuities.

D. Rendering

Given a stereo pair IL, IR and their corresponding renderingproxies, we render the interpolated view as follows:

1) Texture Mapping: For quadrilaterals without depthcontinuities, we cut them into two triangles and performprojective texture mapping on the proxy meshes. For thequadrilaterals cut into two irregular patches in ’Mesh Cutting’procedure, we fit a 3D plane to each patch corresponding tothe valid disparities within it and render them respectively.

2) Blending, Hole filling: The interpolated view is renderedby blending the two synthesized views together as in [3]. Dueto the discarded quadrilaterals, there are small missing areasin the output view and we use hole filling algorithm proposedin [15] to fill them up.

III. EXPERIMENTAL RESULTS

In this section, extensive tests are conducted on theMiddlebury 2.0 and 3D-HEVC test sequences with groundtruth intermediate views for evaluation.

We set the parameters as follows ([.] denotes roundingfunction):

∆x ∆y τ λD λC h[0.04W ] [0.04H] 0.5 0.06 0.6 5

We use proposed method with disparity maps generated by[4], [11] to synthesize interpolated views, which are referredas Proposed-CSBP and Proposed-ADCensus. Following Fickelet al. [6], interpolated views synthesized by VSRS3.5 with thesame disparity maps are used for comparison. We refer to themas VSRS-ADCensus and VSRS-CSBP respectively. We alsoimplemented the warp-based rendering method proposed in[10], e.g. IDW. Finally, state-of-the-art algorithm MeshStereo[3] is included in our comparison.

Table I shows a comparison of all methods on nine sampledtest cases from Middlebury 2.0 in terms of PSNR.

Compared to DIBR method, the proposed method exhibitsrobustness w.r.t the errors of disparity map. For example,when using less accurate disparity map, the rendering qualityof ‘Cloth3’ drastically drops 4dB with DIBR compared to atolerable 1dB loss with proposed method. For most of the testcases, our method obtains leveled PSNR with state-of-the-artview interpolation algorithm [3].

We also evaluate proposed method on real world 3D-HEVC test sequences ‘Balloons’, ‘Kendo’, ‘Newspaper’,‘Poznan Street’. Average PSNR of the rendered sequences forProposed-ADCensus, DIBR-ADCensus, DIBR-GT and IDWare shown in Table II. DIBR-GT refers to using VSRS withinput of ground truth disparity maps. Synthesized views arepresented in Fig. 4 for comparison of visual effect.

As shown in Fig. 4, warped-based method IDW producesbending edges on large baseline (consequently large disparityrange) sequence newspaper. DIBR-ADCensus produces themost noticeable artifacts among all methods. These artifactsare mostly coincided with errors in the estimated disparitymaps. In comparison, our method demonstrates great tolerancefor such disparity error and produces visually convincingresults nonetheless. Generally, DIBR-GT produces authenticand high quality intermediate view on ‘Balloons’ and ‘Kendo’.Minor artifacts, however, are still noticeable on the books in’Newspaper’ and car boundary in ‘Poznan Street’. The artifacton the books is caused by rounded disparity values in groundtruth disparity map.

TABLE IICOMPARISON OF AVERAGE PSNR (dB) VALUES OF THE RENDERINGSEQUENCE. S1-S4 CORRESPOND TO SEQUENCE ‘Balloons’, ‘Kendo’,

‘Newspaper’, ‘Poznan Street’.

S1 S2 S3 S4IDW 27.74 28.55 24.63 30.48

VSRS-ADCeunsus 29.31 30.57 25.32 31.45VSRS-GT 31.78 34.19 28.71 33.03Proposed 32.33 33.77 29.45 34.99

IV. CONCLUSION

In this paper, we present a novel algorithm for viewinterpolation. Compared with existing point-based DIBRmethods which require accurate disparity map, roughdisparity map can be refined with our approach to obtainvisually plausible results. Additionally, we perform meshcutting to handle occlusion and preserve sharp depthdiscontinuities. Evaluations on the Middlebury 2.0 benchmarkand 3D-HEVC test sequences demonstrated its excellentperformance.

ACKNOWLEDGMENT

Thanks to National Natural Science Foundation ofChina 61370115, China 863 project of 2015AA015905,Shenzhen Peacock Plan, Shenzhen Research Projects ofJCYJ20150331100658943 and JCYJ20160506172227337,

Page 4: Robust View Interpolation with Mesh Cutting...Middlebury 2.0 and 3D-HEVC test sequences with ground truth intermediate views for evaluation. We set the parameters as follows ([:] denotes

TABLE ICOMPARISON OF PSNR VALUES OF THE RENDERING RESULTS FROM SAMPLED TEST CASES ON MIDDLEBURY 2.0.

Aloe Baby1 Books Bowling2 Cloth3 Laundry Reindeer Rocks Wood2IDW 24.88 25.12 24.21 25.23 23.65 21.67 22.65 28.11 27.33VSRS-CSBP 26.13 27.33 24.34 26.55 25.78 24.65 22.34 28.43 28.88VSRS-ADCeunsus 26.91 28.62 29.42 28.68 29.63 26.39 23.46 30.49 31.53Mesh Stereo 29.34 35.84 30.07 33.16 35.05 28.66 31.97 35.95 37.72Proposed-CSBP 28.12 34.14 33.45 31.22 33.59 27.32 28.23 34.13 37.78Proposed-ADCensus 29.99 36.02 34.61 32.83 34.42 30.33 30.86 36.64 39.57

(a)IDW (b)DIBR-ADCensus (c)DIBR-GT (d)Proposed (e)Ground Truth

Fig. 4. Comparisons of view interpolation results of ’Balloons’, ’Kendo’, ’Newspaper’, ’Poznan Street’ and zooming in regions .

and Guangdong Province Projects of 2014B010117007 and2014B090910001 for funding.

REFERENCES

[1] H. Shum and S. B. Kang, “Review of image-based renderingtechniques,” in Visual Communications and Image Processing 2000.International Society for Optics and Photonics, 2000, pp. 2–13.

[2] D. Tian, P.-L. Lai, P. Lopez, and C. Gomila, “View synthesis techniquesfor 3d video,” in SPIE Optical Engineering+ Applications. InternationalSociety for Optics and Photonics, 2009, pp. 74 430T–74 430T–11.

[3] C. Zhang, Z. Li, Y. Cheng, R. Cai, H. Chao, and Y. Rui, “Meshstereo:A global stereo model with mesh alignment regularization for viewinterpolation,” in Proceedings of the IEEE International Conference onComputer Vision, 2015, pp. 2057–2065.

[4] X. Mei, X. Sun, M. Zhou, S. Jiao, H. Wang, and X. Zhang, “Onbuilding an accurate stereo matching system on graphics hardware,”in Computer Vision Workshops (ICCV Workshops), 2011 IEEEInternational Conference on. IEEE, 2011, pp. 467–474.

[5] J. Zbontar and Y. LeCun, “Computing the stereo matching cost with aconvolutional neural network,” arXiv preprint arXiv:1409.4326, 2014.

[6] G. P. Fickel, C. R. Jung, T. Malzbender, R. Samadani, and B. Culbertson,“Stereo matching and view interpolation based on image domaintriangulation,” Image Processing, IEEE Transactions on, vol. 22, no. 9,pp. 3353–3365, 2013.

[7] C. L. Zitnick, S. B. Kang, M. Uyttendaele, S. Winder, and R. Szeliski,“High-quality video view interpolation using a layered representation,”in ACM Transactions on Graphics (TOG), vol. 23, no. 3. ACM, 2004,pp. 600–608.

[8] P. Merkle, Y. Morvan, A. Smolic, D. Farin, K. Mueller, P. de With,and T. Wiegand, “The effects of multiview depth video compressionon multiview rendering,” Signal Processing: Image Communication,vol. 24, no. 1, pp. 73–88, 2009.

[9] M. Bleyer, C. Rother, and P. Kohli, “Surface stereo with softsegmentation,” in Computer Vision and Pattern Recognition (CVPR),2010 IEEE Conference on. IEEE, 2010, pp. 1570–1577.

[10] N. Stefanoski, O. Wang, M. Lang, P. Greisen, S. Heinzle, andA. Smolic, “Automatic view synthesis by image-domain-warping,”Image Processing, IEEE Transactions on, vol. 22, no. 9, pp. 3329–3341,2013.

[11] Q. Yang, L. Wang, and N. Ahuja, “A constant-space belief propagationalgorithm for stereo matching,” in Computer vision and patternrecognition (CVPR), 2010 IEEE Conference on. IEEE, 2010, pp. 1458–1465.

[12] G. Chaurasia, S. Duchene, O. Sorkine-Hornung, and G. Drettakis,“Depth synthesis and local warps for plausible image-based navigation,”ACM Transactions on Graphics (TOG), vol. 32, no. 3, p. 30, 2013.

[13] P. Dollar and C. L. Zitnick, “Fast edge detection using structuredforests,” IEEE Transactions on Pattern Analysis & Machine Intelligence,no. 1, pp. 1–1.

[14] C. Rother, V. Kolmogorov, and A. Blake, “Grabcut: Interactiveforeground extraction using iterated graph cuts,” in ACM transactionson graphics (TOG), vol. 23, no. 3. ACM, 2004, pp. 309–314.

[15] M. Bertalmio, L. Vese, G. Sapiro, and S. Osher, “Simultaneous structureand texture image inpainting,” Image Processing, IEEE Transactions on,vol. 12, no. 8, pp. 882–889, 2003.