1084 ieee transactions on image processing, vol. 23, …...art image interpolation algorithms. index...

13
1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014 Image Interpolation via Graph-Based Bayesian Label Propagation Xianming Liu, Member, IEEE, Debin Zhao, Member, IEEE, Jiantao Zhou, Member, IEEE, Wen Gao, Fellow, IEEE , and Huifang Sun, Fellow, IEEE Abstract—In this paper, we propose a novel image interpola- tion algorithm via graph-based Bayesian label propagation. The basic idea is to first create a graph with known and unknown pixels as vertices and with edge weights encoding the similarity between vertices, then the problem of interpolation converts to how to effectively propagate the label information from known points to unknown ones. This process can be posed as a Bayesian inference, in which we try to combine the principles of local adaptation and global consistency to obtain accurate and robust estimation. Specially, our algorithm first constructs a set of local interpolation models, which predict the intensity labels of all image samples, and a loss term will be minimized to keep the predicted labels of the available low-resolution (LR) samples sufficiently close to the original ones. Then, all of the losses evaluated in local neighborhoods are accumulated together to measure the global consistency on all samples. Moreover, a graph- Laplacian-based manifold regularization term is incorporated to penalize the global smoothness of intensity labels, such smoothing can alleviate the insufficient training of the local models and make them more robust. Finally, we construct a unified objective function to combine together the global loss of the locally linear regression, square error of prediction bias on the available LR samples, and the manifold regularization term. It can be solved with a closed-form solution as a convex optimization problem. Experimental results demonstrate that the proposed method achieves competitive performance with the state-of-the- art image interpolation algorithms. Index Terms— Image interpolation, graph, label propagation, local adaptation, global consistency, regression. I. I NTRODUCTION I MAGE interpolation, which is the art of rescaling a low- resolution (LR) image to a high-resolution (HR) version, has become a very active area of research in image processing. Manuscript received November 22, 2012; revised October 11, 2013; accepted November 3, 2013. Date of publication December 11, 2013; date of current version January 27, 2014. This work was supported in part by the National Science Foundation of China under Grants 61300110 and 61272386, in part by the Macau Science and Technology Development Fund under Grant FDCT/009/2013/A1, and in part by the Research Committee at University of Macau under grant SRG023-FST13-ZJT and MRG021/ZJT/2013/FST X. Liu and D. Zhao are with the School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China (e-mail: [email protected]; [email protected]). J. Zhou is with the Department of Computer and Information Science, Faculty of Science and Technology, University of Macau, Taipa, Macau (e-mail: [email protected]). W. Gao is with the National Engineering Laboratory for Video Technology, and Key Laboratory of Machine Perception, School of Electrical Engineering and Computer Science, Peking University, Beijing 100871, China (e-mail: [email protected]). H. Sun is with the Mitsubishi Electric Research Laboratories, Cambridge, MA 02139 USA (e-mail: [email protected]). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TIP.2013.2294543 The interest in image interpolation is born not only in the great practical importance of enhancing resolution of images, such as in the fields of digital photography, computer vision, computer graphics, medical imaging and consumer electron- ics, but also the important theoretical value of using image interpolation to understand the validity of different image models in inverse problems. In the last several years, there have been a great deal of works on image interpolation. In general, image interpolation techniques can be catego- rized into three families: upscaling-based methods [1]–[13], reconstruction-based methods [14], [15], and learning-based methods [16]–[18]. Considering the underlying image model during interpolation, most of image interpolation algorithms could be categorized as globally nonadaptive and locally adaptive ones. A globally nonadaptive algorithm trains the interpolation model using the whole image sample set, while a locally adaptive algorithm trains the model by using only useful local information. The representative globally nonadaptive methods are those based on classical data-invariant linear filters, such as bilinear, bicubic [2], and cubic spline algorithms [3]. These methods have a relatively low computation complexity but suffer from the inability to adapt to varying pixel structures, which results in blurred edges and annoying artifacts. The locally adaptive algorithms usually demonstrate better empirical results than globally nonadaptive ones since it is much easier to seek some functions that are capable of pro- ducing good predictions on some specified regions of images. In the literature, many locally adaptive learning methods have been proposed with great success. Li and Orchard [4] propose to estimate local covariance coefficients from a LR image, and then project the estimated covariance to the HR image to adapt the interpolation. The improved new edge-directed interpolation (INEDI) method [5] modifies NEDI by varying the size of the training window according to the edge size and achieves better performance. In [6], a method named ICBI is proposed to use local second order information to adapt the interpolation and an iterative refinement is further exploited to remove artifacts while preserving image features and texture. In [7], Zhang and Wu propose the named SAI algorithm, which learns and adapts varying scene structures using a locally linear regression model, and interpolates the missing pixels in a group by a soft-decision manner. Liu et al. in [8] propose an effective image interpolation algorithm based on regularized local linear regression (RLLR), in which the ordinary least squares error norm is replaced with the moving least squares error norm leading to a robust estimator of local image 1057-7149 © 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Upload: others

Post on 08-Jul-2020

21 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014

Image Interpolation via Graph-Based BayesianLabel Propagation

Xianming Liu, Member, IEEE, Debin Zhao, Member, IEEE, Jiantao Zhou, Member, IEEE,Wen Gao, Fellow, IEEE, and Huifang Sun, Fellow, IEEE

Abstract— In this paper, we propose a novel image interpola-tion algorithm via graph-based Bayesian label propagation. Thebasic idea is to first create a graph with known and unknownpixels as vertices and with edge weights encoding the similaritybetween vertices, then the problem of interpolation converts tohow to effectively propagate the label information from knownpoints to unknown ones. This process can be posed as a Bayesianinference, in which we try to combine the principles of localadaptation and global consistency to obtain accurate and robustestimation. Specially, our algorithm first constructs a set of localinterpolation models, which predict the intensity labels of allimage samples, and a loss term will be minimized to keep thepredicted labels of the available low-resolution (LR) samplessufficiently close to the original ones. Then, all of the lossesevaluated in local neighborhoods are accumulated together tomeasure the global consistency on all samples. Moreover, a graph-Laplacian-based manifold regularization term is incorporated topenalize the global smoothness of intensity labels, such smoothingcan alleviate the insufficient training of the local models andmake them more robust. Finally, we construct a unified objectivefunction to combine together the global loss of the locally linearregression, square error of prediction bias on the availableLR samples, and the manifold regularization term. It can besolved with a closed-form solution as a convex optimizationproblem. Experimental results demonstrate that the proposedmethod achieves competitive performance with the state-of-the-art image interpolation algorithms.

Index Terms— Image interpolation, graph, label propagation,local adaptation, global consistency, regression.

I. INTRODUCTION

IMAGE interpolation, which is the art of rescaling a low-resolution (LR) image to a high-resolution (HR) version,

has become a very active area of research in image processing.

Manuscript received November 22, 2012; revised October 11, 2013;accepted November 3, 2013. Date of publication December 11, 2013; dateof current version January 27, 2014. This work was supported in part by theNational Science Foundation of China under Grants 61300110 and 61272386,in part by the Macau Science and Technology Development Fund under GrantFDCT/009/2013/A1, and in part by the Research Committee at University ofMacau under grant SRG023-FST13-ZJT and MRG021/ZJT/2013/FST

X. Liu and D. Zhao are with the School of Computer Science andTechnology, Harbin Institute of Technology, Harbin 150001, China (e-mail:[email protected]; [email protected]).

J. Zhou is with the Department of Computer and Information Science,Faculty of Science and Technology, University of Macau, Taipa, Macau(e-mail: [email protected]).

W. Gao is with the National Engineering Laboratory for Video Technology,and Key Laboratory of Machine Perception, School of Electrical Engineeringand Computer Science, Peking University, Beijing 100871, China (e-mail:[email protected]).

H. Sun is with the Mitsubishi Electric Research Laboratories, Cambridge,MA 02139 USA (e-mail: [email protected]).

Color versions of one or more of the figures in this paper are availableonline at http://ieeexplore.ieee.org.

Digital Object Identifier 10.1109/TIP.2013.2294543

The interest in image interpolation is born not only in thegreat practical importance of enhancing resolution of images,such as in the fields of digital photography, computer vision,computer graphics, medical imaging and consumer electron-ics, but also the important theoretical value of using imageinterpolation to understand the validity of different imagemodels in inverse problems. In the last several years, therehave been a great deal of works on image interpolation.In general, image interpolation techniques can be catego-rized into three families: upscaling-based methods [1]–[13],reconstruction-based methods [14], [15], and learning-basedmethods [16]–[18].

Considering the underlying image model duringinterpolation, most of image interpolation algorithms could becategorized as globally nonadaptive and locally adaptive ones.A globally nonadaptive algorithm trains the interpolationmodel using the whole image sample set, while a locallyadaptive algorithm trains the model by using only useful localinformation. The representative globally nonadaptive methodsare those based on classical data-invariant linear filters, suchas bilinear, bicubic [2], and cubic spline algorithms [3]. Thesemethods have a relatively low computation complexity butsuffer from the inability to adapt to varying pixel structures,which results in blurred edges and annoying artifacts.

The locally adaptive algorithms usually demonstrate betterempirical results than globally nonadaptive ones since it ismuch easier to seek some functions that are capable of pro-ducing good predictions on some specified regions of images.In the literature, many locally adaptive learning methods havebeen proposed with great success. Li and Orchard [4] proposeto estimate local covariance coefficients from a LR image,and then project the estimated covariance to the HR imageto adapt the interpolation. The improved new edge-directedinterpolation (INEDI) method [5] modifies NEDI by varyingthe size of the training window according to the edge size andachieves better performance. In [6], a method named ICBI isproposed to use local second order information to adapt theinterpolation and an iterative refinement is further exploited toremove artifacts while preserving image features and texture.In [7], Zhang and Wu propose the named SAI algorithm, whichlearns and adapts varying scene structures using a locallylinear regression model, and interpolates the missing pixels ina group by a soft-decision manner. Liu et al. in [8] propose aneffective image interpolation algorithm based on regularizedlocal linear regression (RLLR), in which the ordinary leastsquares error norm is replaced with the moving least squareserror norm leading to a robust estimator of local image

1057-7149 © 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Page 2: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

LIU et al.: IMAGE INTERPOLATION VIA GRAPH-BASED BAYESIAN LABEL PROPAGATION 1085

structure. Similar to RLLR, Hung and Siu in [9] extend theSAI algorithm in a weighted manner to get more robust results.Although the local model based methods achieve wonderfulperformance, the description ability of this family of modelsis limited. It is difficult to guarantee that the interpolatedimage is best under the global view. Another problem forlocal learning is that in some cases the number of samples isnot enough to sufficiently train the interpolation model, whichwill result in over-fitting the data. Therefore, how to improvethe performances of the local learning algorithms is still achallenging problem.

After reviewing the global and local methods, a naturalquestion is if combining the robustness of global models withthe flexibility of local models can lead to a more effectivesolution. From a statistical perspective, image interpolationis an ill-posed problem. The key to this task is the priorassumption about the properties that the intensity labels shouldhave over the image sample set. One common assumption fornatural images is intensity consistency [4], [19], which means:1) nearby points are likely to have the same or similar intensitylabels; and 2) points on the same structure (manifold) are likelyto have the same or similar intensity labels. Note that the firstassumption means natural images are locally smooth, whichdefines the local consistency; while the second one meansnatural images possess the non-local self-similarity property,which defines the global consistency. Accordingly, it is areasonable idea to consider both local and global informationcontained in images during model learning.

Alternatively, from a machine learning perspective, theavailable LR image pixels can be regarded as labeled sampleswhile the missing HR pixels as unlabeled ones. In the fieldof machine learning, the success of semi-supervised learn-ing [20]–[25] is plausibly due to effective utilization of thelarge amounts of unlabeled data to extract information thatis useful for generalization. Therefore, it is reasonable toleverage both labeled and unlabeled data to achieve betterpredictions. This is especially useful for interpolation onsome bridge regions (i.e., regions connecting two differentobjects), where the number of labeled points is usually notenough to train a robust predictor. We model the wholeimage sample set as a undirected graph with the vertex setcorresponding to the labeled and unlabeled samples, and theedge set representing the relationships between vertices. In thisway, we can model the geometric relationships between alldata points through the graph. Moreover, the reference [20]states that predicting the labels of the unlabeled data on agraph is equivalent to propagating the labels of the labeleddata along the edges to the unlabeled ones. Therefore, thetask of image interpolation converts to the problem of graph-based label propagation, that is, how to effectively propagatethe label information from the known LR samples throughthe graph to the missing HR ones to get accurate and robustestimation.

According to the intuition stated above, in this paper, wepropose a novel image Interpolation algorithm via Graph-based Bayesian Label Propagation (IGBLP). We formulate theprocess of label propagation as a Bayesian inference. The aimis to obtain a labeling of the vertices that is both smooth over

the graph and compatible with the labeled data. Specially,the proposed method predicts the intensity label of a datapoint according to its local neighborhood in a linear way, andthen uses a global optimization to ensure robust predictions.In each neighborhood, an optimal model is estimated viaregularized locally linear regression. With this model, theintensity labels of all samples in the neighborhood can bepredicted. A loss term will be minimized to keep the predictedlabels of available LR samples sufficiently close to the originalones. Then, all of the losses evaluated in local neighborhoodsare accumulated together to measure the global consistencyon the label and unlabeled data. Moreover, a graph-Laplacianbased manifold regularization term is incorporated to penalizethe global smoothness of intensity labels, such smoothingcan alleviate the insufficient training of the local modelsand make them more robust. Finally, we propose a unifiedloss function to combine together the global loss of thelocally linear regression, square error of intensity labels ofthe available LR samples and the manifold regularizationterm, which could be solved with a closed-form solution asa convex optimization problem. In this way, a graph-basedlabel propagation algorithm with local and global consistencyis developed.

The rest of this paper is organized as follows. Section IIintroduces the general framework of Bayesian label propa-gation for interpolation. Section III presents how transductiveregression can be performed with local and global consistency.Section IV details the convex optimization solution of theproposed algorithm. Experimental results are presented inSection V. Section VI concludes the paper.

II. GRAPH-BASED BAYESIAN LABEL PROPAGATION

FOR INTERPOLATION

The problem of image interpolation could be definedas follows: given an image sample set including n pixelsX = {x1, x2, . . . , xl , xl+1, . . . , xn} ∈ �2, of which XL ={x1, x2, . . . , xl} are the available LR samples with intensityvalues labeled by yL = {yi }l

i=1; and XU = {xl+1, . . . , xn} arethe missing HR samples, their intensity values are unlabeled.Given the dataset D = [X = XL ∪ XU , yL ], the task ofimage interpolation converts to infer the posteriori probabilityp(yU |D), where yU = {yi }n

i=l+1 is the labels of the remainingn − l missing samples in XU . By the fact that unlabeledsamples are given beforehand and no other test samples willever be considered, this is a transductive regression problem.

We model the whole image sample set as a undirected graphG = (V, E) with the vertex set V = X corresponding to thelabeled and unlabeled pixels, and the edge set E ⊆ V × Vrepresenting the relationships between vertices. Each edge isassigned a weight Wi j which reflects the affinity of xi and x j .In this way, we can model the geometric relationships of allsamples in the form of a graph. Through the defined graph,predicting the labels of the unlabeled data is equivalent topropagating the labels of the labeled data along the edges tothe unlabeled ones.

To leverage both labeled and unlabeled samples, we assumethat the hard intensity label yi depends upon the hidden soft

Page 3: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

1086 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014

label fi for all i , then the posterior can be written as:

p(yU |D) =∫

fp(yU |f)p(f |D)df, (1)

where f = { f1, f2, . . . , fn} are the soft labels, which aredefined as outputs of the prediction functions, i.e., the esti-mates of true labels of all samples in the graph. In this way, allsamples in the graph including labeled and unlabeled ones areparticipated into the inference process. We first approximatethe posterior p(f |D) and then use the result of Eq. (1) as theestimated intensity labels of missing HR samples. Using theBayes rule we can write p(f |D) as:

p(f |D) = p(f |X, yL) ∝ p(f |X)p(yL|f). (2)

The term p(f |X) is the probability of the soft labels giventhe sample set. It can be regarded as the prior. One commonlyused prior assumption for natural images is insistency consis-tency [4], [19], which states: 1) nearby points are more likelyto have the same or similar intensity labels; and 2) pointson the same manifold are more likely to have the same orsimilar intensity labels. This prior term means that it shouldgive higher probability to the labeling that respects the localsmoothness and global consistency of the data graph. Equiv-alently, we can interpret this probability in an exponentialform [25]:

p(f |X) ∝ exp

(−1

2F(f, X)

), (3)

where F(·, ·) is a general form of the regularization function.The second term p(yL |f) is the likelihood that incorporates

the information provided by the LR intensity labels. Assumingconditional independence of the observed labels given thehidden soft labels, the likelihood can be written as:

p(yL|f) =∏l

i=1p(yi | fi ). (4)

The likelihood models the probabilistic relation between theobserved label yi and the hidden label fi , which can also beinterpreted in an exponential form:

p(yi | fi ) ∝ exp

(−1

2L(yi , fi )

), (5)

where L(·, ·) is a general form of the loss function.With the above definition, we can easily derive the following

result:

p(f |D) ∝ exp{−

(∑li=1 L(yi , fi ) + λF(f, X)

)}. (6)

III. TRANSDUCTIVE REGRESSION WITH LOCAL

AND GLOBAL CONSISTENCY

As stated in the above section, maximum a posterioriprobability (MAP) estimate is equivalent to minimizing anaugmented optimization objective function. As a consequence,the problem of posterior inference converts to how to appro-priately define and combine the loss and regularization terms,which should thoroughly respect the statistical property of thetarget image. In the following, we will show how to definethe loss and regularization terms from the principle of localadaptation and global consistency.

A. The Principle of Global Consistency

Keep in mind that the desired soft labels should be bothsmooth over the graph and compatible with the labeled data.First, let us consider the definition of loss function (i.e., thelikelihood) from the principle of global consistency. Givenlabeled samples {XL , yL} = {(x1, y1), . . . , (xl , yl)}, we tryto select a good interpolation model f by minimizing thefollowing global loss function:

Jg =l∑

i=1

L(yi , f (xi , w)) + λ|| f ||2F , (7)

where L(·, ·) measures the bias of estimated soft labelsfrom the original ones, w is the parameter vector of theinterpolation model, || f ||F is the induced norm of f in thefunctional space F (e.g., F can be a reproducing kernel Hilbertspace (RKHS) induced by some kernel k), which reflectsthe complexity of the predictor f . Clearly, Eq. (7) is in asupervised and global manner, which only utilizes labeledsamples to train one interpolation function for the whole imagesamples.

Then, let us consider the definition of regularization term(i.e., the prior term). From a geometric perspective, thereis a probability distribution p on X × � to generate imagesamples. The available LR samples are (x, y) pairs generatedaccording to p(x, y), the rest missing HR samples are simplydrawn according to the marginal distribution p(x) of p. Onemight hope that knowledge of the marginal p(x) can beexploited for better model learning. In the field of machinelearning, Belkin et al. [24] propose a geometric frameworkfor learning from labeled and unlabeled examples. They statethat there is a specific relationship between the marginaldistribution p(x) and the conditional distribution p(y|x). Itassumes that if two points x1 and x2 are close in the intrinsicgeometry of p(x), the conditional distribution p(y|x1) andp(y|x2) should be similar. In another word, p(y|x) should varysmoothly along the geodesics in the intrinsic geometry of p(x).According to the geometric intuition stated above, we extendthe framework of model learning from supervised to semi-supervised by incorporating additional information about thegeometric structure of the marginal p(x), which seeks a globaloptimal interpolation function f by minimizing the followingobjective function:

Rg =l∑

i=1

L(yi , f (xi , w)) + γA|| f ||2F + γI || f ||2I, (8)

where the additional penalty term || f ||2I reflects the intrin-sic geometric information of the marginal distribution p(x).In most cases, the marginal distribution p(x) is unknown.Therefore, we have to attempt to get empirical estimatesof p(x) and || f ||2I . We assume the support of p(x) is acompact submanifold M ∈ �n , then naturally || f ||2I can beapproximated by

‖ f ‖2I =

x∈M‖∇M f ‖2 dp(x), (9)

Page 4: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

LIU et al.: IMAGE INTERPOLATION VIA GRAPH-BASED BAYESIAN LABEL PROPAGATION 1087

where ∇M f denotes the gradient of f with respect to M.Intuitively, || f ||2I measures how f varies on M, i.e., thesmoothness of f .

In the case of digital image, we can derive the discretizedform for gradient as follows:

∇M f = f (xi ) − f (x j )

d(xi , x j ), (10)

where d(xi , x j ) represents the distance between xi and x j .Then we can rewrite || f ||2I as:

|| f ||2I =n∑

i, j=1

( f (xi ) − f (x j ))2Wi j = fTLf, (11)

where Wi j is in inverse proportion to d2(xi , x j ). Wi j isthe edge weight in the data adjacency graph which reflectsthe affinity between xi and x j . In graph construction, edgeweights plays a crucial role. In this paper, we combine theedge-preserving property of bilateral filter [27] and the robustproperty of non-local-means weight [19] to design the edgeweights, which are defined as follows:

Wi j = 1

Cexp

{−||x j − xi ||2

ε2

}

× exp

{− G · ||SW (x j ) − SW (xi )||2

ε2

}, ε > 0, (12)

where the first exponential term considers the geometricalnearby, and the second one considers the structural similarity.The structural similarity of xi and x j is computed by com-paring the similarity of windows SW (xi ) and SW (x j ), whichare the local patches centered on xi and x j . G is a Gaussiankernel used to further take into account the distance betweenthe central pixel and other pixels in the local patch. We defineL = D − W ∈ �n×n as the graph Laplacian where D isa diagonal matrix with D(i, i) =∑

j Wi j . Through the graphLaplacian, the label information is propagated from labeledsamples to unlabeled ones.

B. The Principle of Local Adaptation

The objective function defined above provides us an excel-lent framework to learn from both labeled and unlabeledsamples. However, the structural loss is defined in a globalway, i.e., for the whole image, we only need to pursue oneinterpolation model f . Actually, as pointed out by [26], itis usually not easy to find a unique function which holdsgood predictability in the entire data space. But it is mucheasier to seek some models that are capable of producinggood predictions on some specified regions of the input space.Accordingly, we resort to the local learning strategy to improvethe accuracy of predictions. More specially, for each data pointxi ∈ X, we consider the linear affine transformation modelf (·; wi , bi ) defined as follows:

f (xi ; wi , bi ) = wTi �(xi ) + bi , (13)

where wi and bi are the weight vector and bias of thelinear estimator, �(xi ) ∈ �d×1 is the intensity label vec-tor of 8-connected neighboring samples of xi [7]; f (xi )

is the estimated intensity label of xi , which is obtainedby a weighted aggregation of the pixels in its localneighborhood.

To compute the model parameters, we split the whole inputspace N into c overlapped local neighborhoods and formulatemodel learning as a set of c optimization problems. It is usuallymore effective to minimize the following local cost functionfor each neighborhood Ni (1 ≤ i ≤ c):

J il =

l∑j=1

K(x j −ci , ε)L(y j , f (x j ; wi , bi ))+γA|| fwi ,bi ||2F ,

(14)

where K(x j − ci , ε) is the similarity kernel centered at ci

with width ε, which is defined in the same form as the edgeweight Wi j :

K(x j − ci , ε) = 1

Ciexp

{−||x j − ci ||2

ε2

}

× exp

{− G · ||SW (x j ) − SW (ci )||2

ε2

}, ε > 0. (15)

As shown in the above equation, K includes two parts, whichplay dual role in model learning: the first exponential functionis to choose samples located in the local neighborhood Ni

centered on ci ; the second exponential function is actuallythe form of non-local-means weight [19], which is to choosesamples with similar structure and therefore can reduce theinfluence of outlier in regression.

In this way, we define a function fwi ,bi with parameter(wi , bi ) for each local region Ni , that is, we define c localinterpolation functions { fwi ,bi }c

i=1. Intuitively, similar localneighborhoods should share similar model parameters. There-fore, it is natural to add together the losses estimated on allof the c neighborhoods, and the total local structural loss isdefined as

Jl =c∑

i=1

J il =

c∑i=1

l∑j=1

K(x j − ci , ε)L(y j , f (x j ; wi , bi ))

+γA|| fwi ,bi ||2F . (16)

By minimizing the above objective function, a collaborativemodel learning mechanism is achieved.

Now let us return to the semi-supervised learning scenario,which aims to learn from both labeled and unlabeled datasamples. As shown in Eq. (1), we introduce a set of soft labels{ f1, f2, . . . , fn} into the loss function such that fi directlydetermines the final estimated label of xi . Then we canredefine the total local loss as

Jl =c∑

i=1

n∑j=1

K(x j − ci , ε)L( f j , f (x j ; wi , bi ))

+γA|| fwi ,bi ||2F . (17)

[3pt] In this way, interpolation models are trained locally usingall samples in the neighborhood. Note that by minimizingJl we can obtain the optimal { fi }n

i=1 and {wi , bi }ci=1. The

advantage of this approach is that while it may be nontrivial

Page 5: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

1088 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014

to produce an image with the desired property everywhere, itis often easier to obtain the property locally.

C. Local and Global Consistency

Recalling the graph-Laplacian regularization frameworkintroduced in the global principle, we also expect fi tohave some geometrical properties. More concretely, we hope{ fi }n

i=1 to be sufficiently smooth with respect to the intrinsicdata graph. Therefore, we finally combine the local and globalprinciple and construct a unified objective function which usesboth labeled and unlabeled data and achieves local and globalconsistency:

R =l∑

i=1

L(yi , f (xi , w)) + λ

⎛⎝ c∑

i=1

n∑j=1

K(x j − ci , ε)L( f j , f (x j ; wi , bi ))+γA|| fwi ,bi ||2F

⎞⎠

+γI fT Lf, (18)

where the first term is called the prediction loss, whichmeasures the inconsistency between the predicted and initiallabels on known LR samples XL . The second term is called thetotal local structural loss, which is defined to reflect the ideathat we split the input space into c local neighborhoods andperform model learning collaboratively. This term punishes thepredictability and complexity of the local prediction functions,which is therefore called the local regularization. The thirdterm is called manifold regularization, which creates thedependency between different local models, and therefore, cantie the individual local model learning together. It penalizes thesmoothness of the intensity labels over the entire data graph,thus is referred to as the global regularization.

IV. CONVEX OPTIMIZATION

In the previous section, we introduce the global and localprincipal for image interpolation, and construct a unifiedframework to perform graph-based label propagation withlocal and global consistency. To derive a practical imageinterpolation algorithm, we should derive an efficient solutionfor the objective function defined in Eq. (18). In the following,let us take this issue into account.

A. The Derivation of Analytical Solution

Now return to the total local loss defined in Eq. (17). Withthe loss function L(·, ·) defined as square loss like that inthe least square regression, the total local loss can be furtherformulated as

Jl =c∑

i=1

∑x j ∈N (xi)

θ(xi , x j )(

wTi �(x j ) + bi − f j

)2

+γA|| fwi ,bi ||2, (19)

where

θ(xi , x j ) = exp

{− G · ||SW (x j ) − SW (xi )||2

ε2

}(20)

is the non-local-means part of K, which plays the same roleas moving weights in moving least squares [28]; and N (xi )represents the local neighborhood centered at xi . Similarly, thelocal structural loss in each neighborhood define in Eq. (14)can be rewritten as

J il =

∑x j ∈N (xi)

θ(xi , x j )(

wTi �(x j ) + bi − f j

)2

+γA|| fwi ,bi ||2. (21)

Let

Gi =[

�Ti 1√

γA · Id 0

],�i = [�(xi1),�(xi2 ), . . . ,�(xini

)],fi = [ fi1 , fi2 , . . . , fini

, 0T ]T , (22)

where xi j is the j -th neighbor of xi , ni is the cardinality ofN (xi ), 1 = [1, 1, . . . , 1]T ∈ �ni×1 and 0 is a d × 1 zerovector, Eq. (21) can be formulated in the matrix form as

J il =

{Gi

[wi

bi

]− fi

}T

· V ·{

Gi

[wi

bi

]− fi

}, (23)

where V = diag(θ(xi , xi1 ), . . . , θ(xi , xini

), 1, . . . , 1)

∈�(ni+d)×(ni+d). V is a diagonal matrix, so the above equationcan be further formulated as

J il =

{V

12 Gi

[wi

bi

]− V

12 fi

}T {V

12 Gi

[wi

bi

]− V

12 fi

}. (24)

Let Gi = V12 Gi and fi = V

12 fi , the above equation becomes

J il =

{Gi

[wi

bi

]− fi

}T {Gi

[wi

bi

]− fi

}. (25)

To derive the optimal transformation parameters [wi bi ]T , wetake the derivative of the loss function J i

l with respect to[wi bi ]T and set the derivative to 0, then the optimal solutioncan be represented by

[wi

bi

]∗= (Gi

TGi )

−1GiT

fi . (26)

With this solution, the total structural loss defined in Eq. (19)becomes

Jl =∑

i

J il =

∑i

fiT

GiT

Gi fi , (27)

where Gi = I − Gi (GiT

Gi )−1Gi

T. It can be easily demon-

strated that Gi is a orthogonal projection matrix. With theproperty of orthogonal projection matrix, Jl can be rewrittenas:

Jl =∑

i

J il =

∑i

fiT

Gi fi . (28)

We split the matrix Gi into four blocks after the ni -th rowand column:

Gi =[

Ai Bi

Ci Di

], (29)

Page 6: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

LIU et al.: IMAGE INTERPOLATION VIA GRAPH-BASED BAYESIAN LABEL PROPAGATION 1089

where Ai ∈ �ni ×ni . Let fi = [ fi1 , fi2 , . . . , fini]T and K =

diag(θ(xi, xi1 ), θ(xi , xi2), . . . , θ(xi , xini)) ∈ �ni ×ni , then

fi = V12 fi =

[K

12

Id

] [fi

0

]=

[K

12 fi

0

]. (30)

According to Eq. (29) and Eq. (30), we can derive

fiT

Gi fi = [fTi K

12 0T ]

[Ai Bi

Ci Di

] [K

12 fi

0

]

= fTi K

12 Ai K

12 fi = fT

i Ai fi . (31)

For the derivation of Ai , please refer to Appendix.Define the vector f = [fT

1 , fT2 , . . . , fT

c ] as the concatenatedlabel vector in all of c local neighborhoods, and define theselection matrix S as a 0-1 matrix with Si j = 1 if x j ∈ N (xi ),so f = Sf. We further define the block-diagonal matrix

G =⎛⎜⎝

A1 0. . .

0 Ac

⎞⎟⎠, (32)

then Eq. (28) can be rewritten as

Jl = fT ST GSf. (33)

Let

M = ST GS, (34)

so finally Eq. (19) can be rewritten as

Jl = fT Mf. (35)

Finally, the objective function formulated in Eq. (18) canbe rewritten as the following matrix form:

R = (f − y)T J(f − y) + λfT Mf + γI fT Lf, (36)

where y = [yL , 0] and yL records the intensity labels of thelabeled image samples, J ∈ �n×n is a diagonal matrix whosediagonal elements are one for labeled samples and zero forunlabeled data.

Theorem 1. Minimization of the objective function inEq. (36) is a convex optimization problem.

Proof: First, we compute the first-order derivative of Rwith respect to f:

∂ R

∂f= 2J(f − y) + λ(M + MT )f + 2γI Lf = 0, (37)

Then, we compute the second order derivative of R withrespect to f:

∂ R

∂f2 = 2J + λ(M + MT ) + 2γI L � 0 (38)

Due to the second-order derivative of R with respect to f is apositive definite matrix, the objective function in Eq. (36) isproved to be a convex optimization problem [32].

Setting the first-order derivative to zero:

∂ R

∂f= 2J(f − y) + λ(M + MT )f + 2γI Lf = 0, (39)

the optimal f can be finally represented as:

f = 2(

2J + λ(M + MT ) + 2γI L)−1

Jy. (40)

Since f in Eq. (40) is linear with respect to all knownvalues, the final solution is actually in an analytical form,and can be computed efficiently without any iterative process.After deriving f, we choose the values corresponding to theunlabeled samples in f as the estimated intensity values ofthe missing HR samples, and keep the original values of LRsamples unchanged.

B. Complexity Analysis and Discussion

The result in Eq. (40) gives a closed-form solution based onthe inversion of a matrix in �n×n . Let T (n) be the complexityof computing the inverse of a matrix in �n×n , and T (n) =O(n3) using standard method or T (n) = O(n2.376) with themethod of Coppersmith and Winogard, where n is the numberof labeled and unlabeled samples in the input space. To achievegood tradeoff between effectiveness and efficiency, we use thesame strategy as non-local-means [19], i.e., not use the wholeimage but a large neighborhood as the input space N .

To gain some insight into our method, we investigate theproposed method from a co-regularized perspective. The lasttwo terms in Eq. (36) can both be reviewed as regularizationterms contained with different regularization matrices, one isderived from the view of local adaptation and the other isfrom global consistency. Benefiting from the co-regularizationterms, our method achieves good tradeoff between adaptationand robustness. The interpolation models induced from thelocal view enjoy the flexibility to adapt to the local statistics ofimages. However, local adaptation alone is not enough. It mayrun into the risk of over-fitting the data. To alleviate this issue,the global regularization term is introduced to enforce that agood labeling should not change too much between similaritypoints (similarity is measured by the edge weights in thegraph). Consequently, the global term ensures the interpolationfunctions vary smoothly with respect to the intrinsic structurecollectively revealed by both measured and unmeasured pix-els. In a nut shell, different types of regularization matricescan better reveal complementary information and thus couldprovide a more accurate and robust predictor. For clarity, thealgorithm flow is shown in Algorithm 1.

An intuitive understanding about why our method workswell is shown as follows. In NEDI and SAI, only one localneighborhood is used as the training window to learn modelparameters. As illustrated in Fig. 1, the red window representssuch a single local neighborhood. However, it usually runs intothe risk of over-fitting the data, which reduces the accuracyof model and further degrades the quality of reconstructedHR images. To alleviate this issue, we split the whole inputspace into a set of overlapped local neighborhoods to carryout models learning collaboratively. As illustrated in Fig. 1,the blue window denotes the input space, where multiple localneighborhoods are utilized for model learning including notonly the red window but also many black ones. More specially,similar local neighborhoods are enforced to share similarmodel parameters. Such a collaborative learning mechanismis achieved by the global regularization term, in which theedge weight (defined in Eq. (12)) is computed to reflect theaffinity of any vertices even that they are far away from

Page 7: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

1090 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014

Algorithm 1 Algorithm of Image Interpolation via Graph-Based Bayesian Label Propagation

Fig. 1. Illustration of collaborative model learning.

each other. Through creating the dependency between differentlocal models, the individual local model learning problems aretied together. In this way, we can effectively reduce the riskof overfitting and enhance the accuracy and robustness of thelearned interpolation model.

V. EXPERIMENTAL RESULTS

In this section, experimental results are presented to demon-strate the performance of the proposed image interpolationalgorithm. Given the fact that numerous image interpola-tion algorithms have been developed during the last twodecades, it would be virtually impossible for us to perform a

thorough comparative study of the proposed algorithm. Here,the proposed IGBLP algorithm is compared with linear inter-polator Bicubic [2], as well as recent state-of-the-art methodsincluding new edge directed interpolation (NEDI) [4], direc-tional filtering and data fusion (DFDF) [12], improve NEDI(INEDI) [5], iterative curvature-based interpolation (ICBI)[6], sparse coding based super-resolution (ScSR) [18], soft-decision adaptive interpolation (SAI) [7], regularized locallinear regression (RLLR) [8]. In our experiments, Bicubicinterpolation is performed with the MATLAB built-in function,and the source codes of other compared methods are kindlyprovided by their authors. For thoroughness and fairness ofour comparison study, we test seven widely used imagesin the literature and one texture image, as illustrated inFig. 2. The source code of this paper can be available fromhttp://homepage.hit.edu.cn/pages/xmliu

A. Comparison With State-of-the-Arts

We first report the comparison results when the scalingfactor S is two. Following the same setting as NEDI andSAI, we downsample these HR images by a factor of twoin both row and column dimensions without antialiasingfiltering to get the corresponding LR images, from which theoriginal HR images are reconstructed by the algorithms underconsideration.

Since the original HR images are known in the simulation,we can compare the interpolated results with the true images,and measure the objective quality of those interpolated images.Table I tabulates the objective quality comparison of the ninedifferent methods when applied to the eight test images ofFig. 2. Since PSNR is an average quality measurement overthe whole image, we also use edge PSNR (EPSNR) as themeasurement to test the reconstruction fidelity of image edges.In our study, the Sobel edge filter is used to locate the edge inthe original image, and the PSNR of the pixels on the edge areused to generate the EPSNR. From Table I, it can be observedthat for most instances the proposed algorithm works better onPSNR than other eight methods. Compared with global meth-ods, such as Bicubic, the proposed method can significantlyimprove the objective quality of generated HR images. Theaverage PSNR gain is 0.88dB. Our method also outperformsthe edge detection based local methods, such as DFDF, forwhich the average PSNR gain is 1.02dB. Compared with NEDIand its variance INEDI, our method achieves higher objectiveperformance, the average PSNR gains are 1.96dB and 1.02dBrespectively. ScSR exploits an example training set to learn thecoupled low- and high-resolution dictionaries, and enforce thatsparse representations between the low- and high-resolutionimage patch pair with respect to their own dictionaries shouldbe the same. The assumption of “the same sparse repre-sentation” is too strong. When the approximation atoms arenot correctly selected, it produces poor results. By exploitinglabeled and unlabeled samples together and keep local andglobal consistency in transductive regression, our method leadsto a significant performance benefits compared with SAI andRLLR, both of which are state-of-the-art locally adaptivemodel based methods. The average PSNR gains are 0.7dB and0.45dB, respectively. The proposed method also achieves the

Page 8: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

LIU et al.: IMAGE INTERPOLATION VIA GRAPH-BASED BAYESIAN LABEL PROPAGATION 1091

Fig. 2. Eight sample images in the test set.

TABLE I

OBJECTIVE QUALITY COMPARISON OF NINE INTERPOLATION ALGORITHMS (IN dB)

TABLE II

PERFORMANCE COMPARISON OF NINE INTERPOLATION ALGORITHMS THROUGH SSIM AND FSIM

highest average EPSNR value. It demonstrates the proposedmethod can effectively preserve local edge structure in images.

Furthermore, in our study, we use SSIM [29] and FSIM [30]as metrics to measure the performance of these interpolationalgorithms. From Table II, it can be observed that our methodachieves the highest average SSIM and FSIM scores amongall of the competing methods. Given the fact that human visualsystem (HVS) is the ultimate receiver of the enlarged images,we also show the subjective comparison results. The test imageAirplane exhibits strong and sharp edges in varying directionsand the test image Monarch exhibits edges with differentscales. Such characteristics make them as prime images totest the fidelity of edge reconstruction. Figs. 3 and 5 illustratethe subjective quality and reconstruction error comparison onthese two images. It can be clearly observed that the images

reconstructed by the Bicubic interpolator suffer from blurrededges, jaggies, and annoying ringing artifacts. ScSR producesartifacts along contours. The reconstruction quality can beimproved to some extent by NEDI, DFDF, INEDI and ICBI,but the image quality is still lower than SAI, RLLR and theproposed IGBLP. SAI and RLLR show improvements overthese methods in the regions of edges and textures, reducingthe visual defects of these methods. Thanks to the combina-tion of local and global information, IGBLP achieves morewonderful visual quality compared with all other methods.The produced edges in our method are clean and sharp. Theoutstanding performance of the proposed method is morevivid by observing the error image of Airplane. Our algorithmproduces smaller interpolation error than other methods. Suchresults clearly demonstrate the superiority of the proposed

Page 9: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

1092 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014

Fig. 3. Reconstruction images and error comparison for Monarch. (A) Bicubic, (B) NEDI, (C) DFDF, (D) ICBI, (E) INEDI, (F) ScSR, (G) SAI, (H) RLLR,(I) IGBLP.

Fig. 4. The influence of the dimension of the input space N on performanceand complexity for Airplane and Monarch.

method in reconstructing the high frequency, such as edgesand textures.

Note that the proposed method can be easily generalizedto enlarge images with the scaling factor S greater than two.If S = 2z with z being a positive integer, one can just applyour method z times. If 2z < S < 2z+1, one can first apply the

proposed method z times then apply Bicubic interpolation onthe enlarged image s times such that 2zs = S. In the Table IV,we give the comparison results with respect to PSNR andSSIM when the scaling factor is three. The downsamplingprocess is similar with that of scaling factor two. From theresults, we can find in most cases our method outperformsother methods with respect to PSNR and SSIM, and achievesthe best average performance on the eight test images.

B. Parameters Setting

There are a few parameters involved in the proposedalgorithm. The input space N includes a set of local neighbor-hoods Ni . The size of N is set to 17×17 and the size of localneighborhood Ni is set to 7×7. The size of similarity windowis set to 5×5 in computing the edge weight Wi j in Eq. (12) andθ(xi , x j ) in Eq. (20), ε2 is fixed to 0.5. Since any analyticalresult is still elusive to obtain, the regularization parametersλ and γI are evaluated by cross validation [31] from anoffline training set including twenty images (not including theeight test images). We divide the offline training set into anestimation subset including ten images and a validation subsetincluding the other ten images. The former subset is used toobtain a parameter estimate and the latter is used to validatethe performance under the estimate. According to the training

Page 10: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

LIU et al.: IMAGE INTERPOLATION VIA GRAPH-BASED BAYESIAN LABEL PROPAGATION 1093

Fig. 5. Reconstruction images and error comparison for Airplane. (A) Bicubic, (B) NEDI, (C) DFDF, (D) ICBI, (E) INEDI, (F) ScSR, (G) SAI, (H) RLLR,(I) IGBLP.

results, in practical implementation we set λ and γI as 0.5 and0.01 respectively.

As mentioned in Section III-B, to achieve a good tradeoffbetween effectiveness and efficiency, we do not use the wholeimage but a relative large neighborhood as the input space N .Airplane and Monarch are used as examples to test how thedimension of N influences the results in terms of performanceand computational complexity. As depicted in Fig. 4, we canfind both PSNR and running time increase progressively withthe dimension of the input space. The PSNR value improvesat a rate of gradually slow, while the complexity increases at

a rate of gradually fast. According to such a trend, in practicalimplementation, we set the dimension of N as 17 × 17 toachieve good tradeoff between effectiveness and efficiency.

C. Complexity Comparison

Finally, we show the results of complexity comparison. Formore clear comparison, Table III gives the PSNR versus aver-age processing times results on a typical computer (Intel XeonCPU 2.83GHz, 3G Memory) of the nine algorithms. All meth-ods are running on Matlab. The proposed method achieved

Page 11: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

1094 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014

TABLE III

OBJECTIVE QUALITY VERSUS AVERAGE PROCESSING TIMES (dB/SECONDS) RESULTS

TABLE IV

QUALITY COMPARISON OF NINE INTERPOLATION ALGORITHMS (IN DB) WHEN THE SCALING FACTOR IS THREE

the best interpolation performance at expense of relative highcomputational complexity. In the above test, we apply theproposed method on all missing samples. We can manage thecomputational complexity by reducing the number of samplesto estimate. One way is to apply the proposed method only foredge and texture regions, and apply simple bilinear or bicubicinterpolation for smooth regions.

VI. CONCLUSION

In this paper, we presented an effective image interpolationalgorithm through graph-based label propagation to achievelocal and global consistency. Our method is novel in twoaspects: (1) both labeled and unlabeled data are explored inthe process of model learning. Such a transductive manner canimprove the accuracy of predictor for some bridge regions;(2) local and global consistency is achieved during regression,which can make the predictor more robust. These two aspectscan be cast into an unified optimization framework, whichcan be efficiently solved with a closed-form solution. Exper-imental results on benchmark test images demonstrate thatthe proposed method achieves very competitive interpolationperformance with the state-of-the-art interpolation algorithms.In future work, we will study how to find a good trade-offbetween performance and computational complexity.

APPENDIX

First, we can derive Gi = V12 Gi =

[K

12

Id

][

�Ti 1√

γA · Id 0

]=

[K

12 �T

i K12 1√

γA · Id 0

].

Let �T = K12 �T

i and ϒ = K12 1, and so Gi =[

�T ϒ√γA · Id 0

]. Then

(Gi

TGi

)−1 =[

��T + γAId �ϒ

ϒT �T 1

]−1

(41)

According to the inverse computation principle of blockmatrix, we can obtain

(Gi

TGi

)−1 =⎡⎣ F−1

i + F−1i �ϒϒT �T F−1

i1−e − F−1

i �ϒ

1−e

−ϒT �T F−1i

1−e1

1−e

⎤⎦. (42)

where

Fi = ��T + γAId , e = ϒT �T F−1i �ϒ. (43)

We can further obtain

Gi

(Gi

TGi

)−1Gi

T

=[

�T ϒ√γA · Id 0

] (Gi

TGi

)−1[

�√

γA · Id

ϒT 0T

]. (44)

Finally, we can derive that

Ai = Ini −(�T F−1

i �

+ �T F−1i �ϒϒT �T F−1

i �−�T F−1i �ϒϒT −ϒϒT �T F−1

i �+ϒϒT

1−e

).

(45)

ACKNOWLEDGMENT

The authors would like to thank the anonymous reviewersfor their constructive suggestions which helped us improvingour manuscript.

Page 12: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

LIU et al.: IMAGE INTERPOLATION VIA GRAPH-BASED BAYESIAN LABEL PROPAGATION 1095

REFERENCES

[1] X. Liu, D. Zhao, R. Xiong, S. Ma, W. Gao, and H. Sun, “Transductiveregression with local and global consistency for image super-resolution,”in Proc. IEEE Data Compress. Conf., DCC, Mar. 2011, pp. 173–182.

[2] R. G. Keys, “Cubic convolution interpolation for digital image process-ing,” IEEE Trans. Acoust., Speech, Signal Process., vol. 29, no. 6,pp. 1153–1160, Dec. 1981.

[3] H. Hou and H. Andrews, “Cubic splines for image interpolation anddigital filtering,” IEEE Trans. Acoust., Speech, Signal Process., vol. 26,no. 6, pp. 508–517, Dec. 1978.

[4] X. Li and M. T. Orchard, “New edge-directed interpolation,” IEEE.Trans. Image Process., vol. 10, no. 10, pp. 1521–1527, Oct. 2001.

[5] N. Asuni and A. Giachetti, “Accuracy improvements and artifactsremoval in edge based image interpolation,” in Proc. 3rd Int. Conf.Comput. Vis. Theory Appl., 2008, pp. 58–65.

[6] A. Giachetti and N. Asuni, “Real-time artifact-free image upscaling,”IEEE Trans. Image Process., vol. 20, no. 10, pp. 2760–2768, Oct. 2011.

[7] X. Zhang and X. Wu, “Image interpolation by adaptive 2-D autore-gressive modeling and soft-decision estimation,” IEEE Trans. ImageProcess., vol. 17, no. 6, pp. 887–896, Jun. 2008.

[8] X. Liu, D. Zhao, R. Xiong, S. Ma, W. Gao, and H. Sun, “Imageinterpolation via regularized local linear regression,” IEEE. Trans. ImageProcess., vol. 20, no. 12, pp. 3455–3469, Dec. 2011.

[9] K.-W. Hung and W.-C. Siu, “Robust soft-decision interpolation usingweighted least squares,” IEEE. Trans. Image Process., vol. 21, no. 3,pp. 1061–1069, Mar. 2012.

[10] R. Xiong and W. Gao, “Tanner graph based image interpolation,” inProc. IEEE Data Compress. Conf., Mar. 2010, pp. 376–385.

[11] S. Battiato, G. Gallo, and F. Stanco, “A locally-adaptive zoomingalgorithm for digital images,” Image Vis. Comput., vol. 20, no. 11,pp. 805–812, 2002.

[12] L. Zhang and X. Wu, “Image interpolation via directional filtering anddata fusion,” IEEE. Trans. Image Process., vol. 15, no. 8, pp. 2226–2238,Aug. 2006.

[13] H. Takeda, S. Farsiu, and P. Milanfar, “Kernel regression for imageprocessing and reconstruction,” IEEE. Trans. Image Process., vol. 16,no. 2, pp. 349–366, Feb. 2007.

[14] S. Dai, M. Han, W. Xu, Y. Wu, Y. Gong, and A. K. Katsaggelos, “Soft-Cuts: A soft edge smoothness prior for color image super-resolution,”IEEE. Trans. Image Process., vol. 18, no. 4, pp. 969–981, May 2009.

[15] W. Dong, L. Zhang, G. Shi, and X. Wu, “Nonlocal back-projection foradaptive image enlargement,” in Proc. IEEE Int. Conf. Image Process.,Oct. 2009, pp. 349–352.

[16] W. Freeman, E. Pasztor, and O. Carmichael, “Learning low-level vision,”Int. J. Comput. Vis., vol. 40, no. 1, pp. 25–47, 2000.

[17] K. Kim and Y. Kwon, “Single-image super-resolution using sparseregression and natural image prior,” IEEE Trans. Pattern Anal. Mach.Intell., vol. 32, no. 6, pp. 1127–1133, Jun. 2010.

[18] J. Yang, J. Wright, T. Huang, and Y. Ma, “Image super-resolution viasparse representation,” IEEE. Trans. Image Process., vol. 19, no. 11,pp. 2861–2873, Nov. 2010.

[19] A. Buades, B. Coll, and J. Morel, “Nonlocal image and movie denois-ing,” Int. J. Comput. Vis., vol. 76, no. 2, pp. 123–139, 2008.

[20] X. Zhu and Z. Ghahramani, “Learning from labeled and unlabeled datawith label propagation,” School Comput. Sci., Carnegie Mellon Univ.,Pittsburgh, PA, USA, Tech. Rep. CMU-CALD-02-107, 2002.

[21] F. Wang, “A general learning framework based on local and globalregularizations,” Pattern Recognit., vol. 43, no. 9, pp. 3120–3129,Sep. 2010.

[22] C. Cortes and M. Mohri, “On transductive regression,” in Proc. NIPS,Jun. 2008, pp. 305–312.

[23] D. Zhou, O. Bousquet, T. N. Lal, J. Weston, and B. Scholkopf,“Learning with local and global consistency,” in Proc. Adv. NIPS, 2004,pp. 321–328.

[24] M. Belkin, P. Niyogi, and V. Sindhwani, “Manifold regularization:A geometric framework for learning from examples,” J. Mach. Learn.Res., vol. 7, pp. 2399–2434, Dec. 2006.

[25] A. Kapoor, Y. Qi, H. Ahn, and R. W. Picard, “ Hyperparameter andkernel learning for graph based semi-supervised classification,” in Proc.Adv. NIPS, 2005, pp. 627–634.

[26] V. Vapnik, The Nature of Statistical Learning Theory. New York, NY,USA: Springer-Verlag, 1995.

[27] C. Tomasi and R. Manduchi, “Bilateral filtering for gray and colorimages,” in Proc. 6th Int. Conf. Comput. Vis., Jan. 1998, pp. 839–846.

[28] D. Levin, “The approximation power of moving least squares,” Math.Comput., vol. 67, no. 224, pp. 1517–1531, 1998.

[29] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Imagequality assessment: From error visibility to structural similarity,” IEEE.Trans. Image Process., vol. 13, no. 4, pp. 600–612, Apr. 2004.

[30] L. Zhang, L. Zhang, X. Mou, and D. Zhang, “FSIM: A feature similarityindex for image quality assessment,” IEEE. Trans. Image Process.,vol. 20, no. 8, pp. 2378–2386, Aug. 2011.

[31] S. Arlot and A. Celisse, “A survey of cross-validation procedures formodel selection,” Statist. Surveys, vol. 4, pp. 40–79, Feb. 2010.

[32] G. H. Golub and C. F. Van Loan, Matrix Computations, 3rd ed.Baltimore, MD, USA: Johns Hopkins Univ. Press, 1983.

Xianming Liu is currently an Assistant Professorwith the Department of Computer Science, HarbinInstitute of Technology (HIT), Harbin, China. Hereceived the B.S., M.S., and Ph.D. degrees in com-puter science from HIT in 2006, 2008, and 2012,respectively. From 2009 to 2012, he was withthe National Engineering L aboratory for VideoTechnology, Peking University, Beijing, China, asa Research Assistant. In 2011, he was with theDepartment of Electrical and Computer Engineering,McMaster University, Canada, as a Visiting Student.

From 2012 to 2013, he was a Post-Doctoral Fellow with McMaster University.His research interests include image/video coding, image/video processing,and machine learning.

Debin Zhao received the B.S., M.S., and Ph.D.degrees in computer science from the Harbin Insti-tute of Technology, China, in 1985, 1988, and 1998,respectively. He is currently a Professor with theDepartment of Computer Science, Harbin Instituteof Technology. He has published over 200 technicalarticles in refereed journals and conference proceed-ings in the areas of image and video coding, videoprocessing, video streaming and transmission, andpattern recognition.

Jiantao Zhou (M’11) is currently an AssistantProfessor with the Department of Computer andInformation Science, Faculty of Science and Tech-nology, University of Macau. He received the B.Eng.degree from the Department of Electronic Engineer-ing, Dalian University of Technology, Dalian, China,in 2002, the M.Phil. degree from the Department ofRadio Engineering, Southeast University, Nanjing,China, in 2005, and the Ph.D. degree from theDepartment of Electronic and Computer Engineer-ing, Hong Kong University of Science and Technol-

ogy, Hong Kong, in 2009. He held various research positions at the Universityof Illinois at Urbana-Champaign, the Hong Kong University of Scienceand Technology, and McMaster University. His research interests includemultimedia security and forensics, and high-fidelity image compression. Hewas a co-author of a paper that received the Best Paper Award in the IEEEPacific-Rim Conference on Multimedia in 2007.

Page 13: 1084 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, …...art image interpolation algorithms. Index Terms—Image interpolation, graph, label propagation, local adaptation, global

1096 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 23, NO. 3, MARCH 2014

Wen Gao (M’92–SM’05–F’09) received the Ph.D.degree in electronics engineering from the Univer-sity of Tokyo, Japan, in 1991.

He is a Professor of computer science with PekingUniversity, China. Before joining Peking University,he was a Professor of computer science with theHarbin Institute of Technology from 1991 to 1995,and a Professor with the Institute of ComputingTechnology of Chinese Academy of Sciences. Hehas published extensively including five books andover 600 technical articles in refereed journals and

conference proceedings in the areas of image processing, video coding andcommunication, pattern recognition, multimedia information retrieval, multi-modal interface, and bioinformatics. He served on the editorial board for sev-eral journals, such as the IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS

FOR VIDEO TECHNOLOGY, the IEEE TRANSACTIONS ON MULTIMEDIA,the IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, theEURASIP Journal of Image Communications, the Journal of Visual Com-munication and Image Representation. He chaired a number of prestigiousinternational conferences on multimedia and video signal processing, such asthe IEEE ICME and ACM Multimedia, and also served on the advisory andtechnical committees of numerous professional organizations.

Huifang Sun (M’85–SM’93–F’00) received the B.S.degree in electrical engineering from the HarbinEngineering Institute, Harbin, China, and the Ph.D.degree in electrical engineering from the Universityof Ottawa, Canada. In 1986, he joined FairleighDickinson University, Teaneck, NJ, USA, as anAssistant Professor, and promoted to an AssociateProfessor. From 1990 to 1995, he was with the DavidSarnoff Research Center (Sarnoff Corp), Princeton,NJ, USA, as a member of technical staff and laterpromoted to Technology Leader of Digital Video

Technology where his activities were MPEG video coding, HDTV, and GrandAlliance HDTV development. He joined the Mitsubishi Electric ResearchLaboratories (MERL), in 1995, as a Senior Principal Technical Staff andwas promoted as the Deputy Director in 1997, Vice President, and MERLfellow in 2003 and is currently a MERL fellow. He has co-authored twobooks and published more than 150 journal and conferences papers. He has 60granted U.S. patents. He was an Associate Editor of the IEEE TRANSACTION

ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, and TechnicalCommittee Chair of the Visual Communications and Image Processing ofthe IEEE Circuits and Systems.