hybrid inter- and intra-wavelet scale image restoration

10
Pattern Recognition 36 (2003) 1737 – 1746 www.elsevier.com/locate/patcog Hybrid inter- and intra-wavelet scale image restoration Lei Zhang a , Paul Bao b; , Xiaolin Wu c a Department of Computing, The Hong Kong Polytechnic University, Hung Hum, Kowloon, Hong Kong b Department of Information Engineering, The Chinese University of Hong Kong, Shatin, New Territory, Hong Kong c Department of Computer Science, University of Western Ontario, London, Ont., Canada Received 15 February 2002; received in revised form 18 November 2002; accepted 18 November 2002 Abstract This paper exploits both the inter- and intra-scale interdependencies that exist in wavelet coecients to improve image restoration from noise-corrupted data. Using an over-complete wavelet expansion, we group the wavelet coecients with the same spatial orientation at several scales. We then apply the linear minimum mean squared-error estimation to smooth noise. This scheme exploits the inter-scale correlation information of wavelet coecients. To exploit the intra-scale dependencies, we calculate the co-variance matrix of each vector locally using a centered square-shaped window. Experiments show that the proposed hybrid scheme signicantly outperforms methods exploiting only the intra- or inter-scale dependencies. The perfor- mance of noise removal also depends on wavelet lters. In our experiments a biorthogonal wavelet, which best characterizes the image inter-scale dependencies, achieves the best results. ? 2003 Pattern Recognition Society. Published by Elsevier Science Ltd. All rights reserved. Keywords: Image restoration; Overcomplete wavelet expansion; Inter- and intra-scale dependency; LMMSE 1. Introduction Wavelet transform (WT) [14] can decorrelate images into nearly independent coecients and is a powerful tool in adaptive signal processing such as coding and noise re- moval. In the wavelet domain, image energy is compacted into a small number of signicant coecients that repre- sent instantaneous features. The small valued coecients tend to be localized due to image-independent additive noise, which will increase the image power in general. Wavelet thresholding has proved to be an eective de- noising technique [511]. In such methods, a threshold is preset to eliminate noise and insignicant structures. Those coecients less than the threshold will be set to 0 while coecients above the threshold will be preserved (hard thresholding) or reduced (soft thresholding). After thresh- olding, the energy of the noisy image may decrease. Corresponding author. Tel.: +852-260-98452; fax: +852-260- 35032. E-mail address: [email protected] (P. Bao). Generally, the wavelet coecients within each subband can be modeled as independent identically distributed (i.i.d) random variables with generalized Gaussian distribution (GGD) [1,5,6,12]. With the GGD modeling, Chang et al. [5] proposed the BayesShrink that employs a subband adaptive soft threshold. Chang’s approach signicantly out- performs Donoho’s classical Wavelet Shrinkage [7], which is a global nonparametric estimation method. Although WT well decorrelates images, there still exist strong dependencies between wavelet coecients. Exploita- tion of such dependency information with proper statistical models could further improve the performance of coding and denoising algorithms. As noted by Liu and Moulin [13], the statistical wavelet models can be classied into intra- scale models, inter-scale models and composite dependency models. A number of intra-scale models have been proposed in Refs. [6,14,15]. Chang et al. [6] dened each wavelet coef- cient as a mixture of GGD variables with unknown slowly spatially varying parameters. The estimation of these parameters is conditioned on a function of its neighboring coecients. The morphological coder [15] also exploits the 0031-3203/03/$30.00 ? 2003 Pattern Recognition Society. Published by Elsevier Science Ltd. All rights reserved. doi:10.1016/S0031-3203(02)00350-3

Upload: lei-zhang

Post on 02-Jul-2016

218 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Hybrid inter- and intra-wavelet scale image restoration

Pattern Recognition 36 (2003) 1737–1746www.elsevier.com/locate/patcog

Hybrid inter- and intra-wavelet scale image restoration

Lei Zhanga, Paul Baob;∗, Xiaolin WucaDepartment of Computing, The Hong Kong Polytechnic University, Hung Hum, Kowloon, Hong Kong

bDepartment of Information Engineering, The Chinese University of Hong Kong, Shatin, New Territory, Hong KongcDepartment of Computer Science, University of Western Ontario, London, Ont., Canada

Received 15 February 2002; received in revised form 18 November 2002; accepted 18 November 2002

Abstract

This paper exploits both the inter- and intra-scale interdependencies that exist in wavelet coe3cients to improve imagerestoration from noise-corrupted data. Using an over-complete wavelet expansion, we group the wavelet coe3cients with thesame spatial orientation at several scales. We then apply the linear minimum mean squared-error estimation to smooth noise.This scheme exploits the inter-scale correlation information of wavelet coe3cients. To exploit the intra-scale dependencies,we calculate the co-variance matrix of each vector locally using a centered square-shaped window. Experiments show that theproposed hybrid scheme signi7cantly outperforms methods exploiting only the intra- or inter-scale dependencies. The perfor-mance of noise removal also depends on wavelet 7lters. In our experiments a biorthogonal wavelet, which best characterizesthe image inter-scale dependencies, achieves the best results.? 2003 Pattern Recognition Society. Published by Elsevier Science Ltd. All rights reserved.

Keywords: Image restoration; Overcomplete wavelet expansion; Inter- and intra-scale dependency; LMMSE

1. Introduction

Wavelet transform (WT) [1–4] can decorrelate imagesinto nearly independent coe3cients and is a powerful toolin adaptive signal processing such as coding and noise re-moval. In the wavelet domain, image energy is compactedinto a small number of signi7cant coe3cients that repre-sent instantaneous features. The small valued coe3cientstend to be localized due to image-independent additivenoise, which will increase the image power in general.Wavelet thresholding has proved to be an e@ective de-noising technique [5–11]. In such methods, a threshold ispreset to eliminate noise and insigni7cant structures. Thosecoe3cients less than the threshold will be set to 0 whilecoe3cients above the threshold will be preserved (hardthresholding) or reduced (soft thresholding). After thresh-olding, the energy of the noisy image may decrease.

∗ Corresponding author. Tel.: +852-260-98452; fax: +852-260-35032.E-mail address: [email protected] (P. Bao).

Generally, the wavelet coe3cients within each subband canbe modeled as independent identically distributed (i.i.d)random variables with generalized Gaussian distribution(GGD) [1,5,6,12]. With the GGD modeling, Chang et al.[5] proposed the BayesShrink that employs a subbandadaptive soft threshold. Chang’s approach signi7cantly out-performs Donoho’s classicalWavelet Shrinkage [7], whichis a global nonparametric estimation method.

Although WT well decorrelates images, there still existstrong dependencies between wavelet coe3cients. Exploita-tion of such dependency information with proper statisticalmodels could further improve the performance of codingand denoising algorithms. As noted by Liu and Moulin [13],the statistical wavelet models can be classi7ed into intra-scale models, inter-scale models and composite dependencymodels.

A number of intra-scale models have been proposed inRefs. [6,14,15]. Chang et al. [6] de7ned each wavelet coef-7cient as a mixture of GGD variables with unknown slowlyspatially varying parameters. The estimation of theseparameters is conditioned on a function of its neighboringcoe3cients. The morphological coder [15] also exploits the

0031-3203/03/$30.00 ? 2003 Pattern Recognition Society. Published by Elsevier Science Ltd. All rights reserved.doi:10.1016/S0031-3203(02)00350-3

Page 2: Hybrid inter- and intra-wavelet scale image restoration

1738 L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746

spatial clustering of wavelet coe3cients. MKhLcak et al. [14]estimated the two-order local statistics of each coe3cientwith a centered square-shaped window and developed alinear minimum mean squared-error estimation (LMMSE)like method.

The wavelet dependency information is distributed notonly in the intra-scale manner but also in the inter-scalemanner. If a coe3cient at a coarser scale has small magni-tude, its descendant coe3cients at 7ner scales are also likelyto be small. Shapiro exploited this property to develop thewell-known embedded zerotree wavelet coder [16]. Con-versely, if a wavelet coe3cient produced by a true signalis of large magnitude at 7ner scale, its parents at coarserscales will also be large. However for those coe3cients pro-duced by noise, the magnitude will decay rapidly along thescales. With this observation, Xu et al. [17] multiplied theadjacent wavelet scales to sharpen the important structureswhile reducing noise. They then distinguished edge pointsfrom noise in the multiscale products. The inter-scale depen-dencies in the wavelet domain have also been representedby autoregressive models and hidden Markov models in lit-erature [18–21].

Wavelet intra- and inter-scale dependency informationcan be combined to improve noisy coe3cients estimationresults. In Ref. [22], Portilla et al. modeled each coe3cientas the product of a Gaussian random vector and a hid-den multiplier variable that included adjacent scales in theconditioning local neighborhood. Liu and Moulin [23] alsoemployed a composite denoising model. Furthermore theyanalyzed the dependency that existed in wavelet coe3cientswith a measurement called mutual information [13]. Vari-ous wavelet models were also compared with respect to thecapability of capturing the dependency information.

In this paper, we will present an adaptive inter- andintra-scale hybrid noise removal scheme with an over-complete wavelet expansion (OWE). This expansion hasbeen observed to perform much better in denoising than a(bi)orthogonal wavelet transform (OWT) that is a lack oftranslation invariant [6,9,24]. No downsampling occurs inOWE. Thus at each scale the image samples are the same.We combined the wavelet coe3cients at the same positionacross adjacent scales as a vector, and then applied theLMMSE to the vector variable. This incorporates naturallythe inter-scale dependencies of wavelet coe3cients to theestimation. To exploit the intra-scale dependency, we es-timated the second order statistics of the vector variablelocally by a centered square-shaped window within eachsubband. An LMMSE algorithm with a two-scale com-posite dependency model is 7nally proposed and validatedwith benchmark images. The performance of the schemealso depends on the selection of the wavelet 7lters.

The paper is organized as follows. Section 2 introducesthe OWT and OWE. In Section 3, a hybrid wavelet inter-and intra-scale model is proposed and an e3cient denoisingalgorithm is developed. The e@ectiveness of several wavelet7lters for the algorithm is also discussed. Experiments on

. . .

. . .

H F2S

S

1S

1W 2W

F

K K

2S

2W 1W

H

G G

S

1S

Fig. 1. The decomposition and reconstruction structures of (bi)OWT. The downward and upward arrows indicate downsamplingand upsampling.

benchmark images by intra-scale, inter-scale and the pro-posed hybrid schemes are shown in Section 4. The conclu-sion is given in Section 5.

2. Wavelet transform and overcomplete expansion

A WT represents a signal f as a linear combination ofelementary atoms or building blocks. More details about thetheory of wavelets and their application in signal processingcan be found in Daubechies [3], Mallat [1,2] and Vetterli[4]. Denoted by m;n the dyadic dilation and translation of amother wavelet can be described as

m;n(t) = 2−m=2 (2−mt − n) (2.1)

with m; n∈ Z . Then f can be written as

f =∑

cm;n(f) m;n: (2.2)

For orthogonal wavelets,

cm;n(f) = 〈f; m;n〉=∫

f(t) m;n(t) dt; (2.3)

where 〈·; ·〉 is the inner product in L2(R). For bi-orthogonalwavelets,

cm;n(f) = 〈f; m;n〉; (2.4)

where m;n is the dual wavelet of m;n [3].The (bi) OWT structure is shown in Fig. 1. The down-

ward and upward arrows indicate downsampling andupsampling. H and G are the low- and high-pass analytic7lters while F and K are the corresponding synthetic 7lters.When the wavelet bases are orthogonal, F and K are theconjugated 7lters of H and G. Besides the Haar wavelet,there is no other compactly supported orthogonal waveletthat is (anti-)symmetrical. This is a very important propertyin signal processing. Compactly supported bi-orthogonalwavelets discard the orthogonality to preserve the symmetricproperty.

OWT is translation variant due to the downsampling indecomposition. This will cause some visual artifacts (suchas Gibbs phenomena) in threshold-based denoising [24]. Ithas been observed that the overcomplete wavelet expansion(OWE, undecimated WT or translation invariant transform)achieves better results in noise removal [6,9,24]. The de-noised image by OWE is an average of several circularlyshifted denoised versions of the same image by OWT. Thenoise is smoothed more and the artifacts are weakened.

Page 3: Hybrid inter- and intra-wavelet scale image restoration

L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746 1739

. . .

. . .

H0

G0

H1

G1

021 F

021 K

121 F

121 K

S S

1S 1S2S 2S

1W 1W2W 2W

Fig. 2. The illustration of overcomplete wavelet expansion (OWE).Filter Hj is obtained by placing (2j−1 − 1) zeros between thecoe3cients of H0.

The denoising scheme presented in this paper is based onOWE whose structure is shown in Fig. 2. No downsamplingor upsampling occurs but the analytic and synthetic 7ltersare changed in each decomposition or reconstruction stage.Hj is obtained by placing (2j−1 − 1) zeros between thecoe3cients of H0. Similarly for Gj , Fj and Kj . The zeropadding of wavelet 7lters, instead of downsampling waveletcoe3cients, is used to decrease the frequency bandwidth.

The 2-D OWT and OWE can be extended from 1-D byseparable 7ltering. Wavelet coe3cients in the detail bandscan be obtained in three directions: horizontal, vertical anddiagonal. Fig. 3 shows the one-stage decomposition struc-tures of 2-D OWT and OWE. Filter F ′ is the transition ofF . At each scale the wavelet coe3cients of OWE have thesame number of samples as the input image.

3. Hybrid intra- and inter-scale model and denoisingalgorithm

3.1. The LMMSE of wavelet coe8cients

Suppose the original image f is corrupted with additiveGaussian white noise �

g= f + �; (3.1)

where �∈N (0; �2). By implementing OWE to the noisy im-age, we can denote the wavelet coe3cient at scale j as

Zj = Xj + Vj; (3.2)

1+jS

jH

jG

'jH

'jG

'jG

'jH

DjW 1+

HjW 1+

VjW 1+

1+jS

H

jS

G

'H

'G

'G

'H

DjW 1+

HjW 1+

VjW 1+

(a) (b)

jS

Fig. 3. One stage decomposition of the 2-D (a) OWT; (b) OWE. F ′ is the transition of F . WHj , WV

j and WDj are the wavelet coe3cients

at horizontal, vertical and diagonal directions.

where Xj is the OWE of f and Vj is the OWE of � for thelinearity of the WT.

Many threshold-based methods have been proposed torecover the image from noisy observation. The WaveletShrinkage �t(z)= sgn(z)max(|z|− t; 0) was originally pro-posed by Donoho [7] with a soft threshold t = �

√2 logN

based on orthonormal wavelet bases. N is the samplelength of input signal. Although Donoho demonstrated it isasymptotically optimal in the minimax sense, it is knownthat the universal threshold would over-smooth signals.Donoho improved his work by the SureShrink scheme inRef. [8], which adaptively computes the threshold by mini-mizing Stein’s unbiased risk estimate. Bayesian estimationbased soft threshold was also proposed in some publications[5,10].

It is generally accepted that the wavelet coe3cientsconform to the generalized GGD within each subband[1,5,6,12]. With the GGD modeling, Chang et al. [5]presented a subband-dependent soft threshold t = �2=�Xj

(assuming the wavelet base is orthonormal), where �Xj

is the standard deviation of wavelet coe3cients Xj . Theso-called BayesShink scheme outperformed many otherpopular thresholds.

In this paper, we apply the LMMSE, instead of soft thresh-olding, to wavelet coe3cients. Suppose the variances of Vj

and Xj are �2j and �2

Xj , respectively, the LMMSE of variablexj is

xj = czj (3.3)

with

c =�2Xj

�2Xj+ �2

j: (3.4)

Noise Vj is Gaussian distributed and independent of Xj .If Xj is also of Gaussian distribution, it is well knownthat LMMSE will evolve as the optimal minimum meansquare-error estimation (MMSE) [25]. Unfortunately Xj

subjects to GGD, which degenerates to the Gaussian distri-bution only in a very special case [1,5].

The standard deviation of noise Vj is computed as

�j = ‖ j‖�; (3.5)

Page 4: Hybrid inter- and intra-wavelet scale image restoration

1740 L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746

where j is the dilation of 2-D mother wavelet : j(x; y)=2−j (2−jx; 2−jy), and ‖ • ‖ is the normal operator: ‖ ‖=√∫∫

(x; y) dx dy.If the noiseless image Xj is unknown, �Xj can be estimated

as follows:

�2Xj = �2

Zj − �2j (3.6)

with

�2Zj =

1MN

M∑m=1

N∑n=1

Z2j (m; n): (3.7)

This is the variance of noisy image Zj whose size is M ×N .LMMSE is similar to the soft thresholding strategy to

some extent. Unlike shrinking noisy wavelet coe3cient zwith a threshold t: x=sgn(z)max(|z|− t; 0), LMMSE modi-7es the coe3cient with a factor c: x=cz. It could be observedfrom Eq. (3.4) that c is less than 1, and then the magnitudeof estimated wavelet coe3cient x will be less than that of z.This would lead to a power reduction of the restored image,similar to the threshold-based schemes.

3.2. The intra-scale dependencies exploited model

The LMMSE scheme presented in Section 3.1 is subbandadaptive, but the spatial dependencies within each subbandare not exploited. It is usually true that a wavelet coe3-cient is better correlated with its local neighborhood. Esti-mating the data locally is widely used technique in imagecompression and denoising algorithms. In Ref. [6], Changet al. proposed a spatially adaptive wavelet thresholdingscheme based on context modeling. Each wavelet coe3cientis modeled as a mixture of GGD variables with unknownslowly spatially varying parameters. The estimation of theseparameters is conditioned on a function of its neighboringcoe3cients.

MKhLcak et al. [14] developed a spatially adaptive schemewith LMMSE strategy. For each wavelet coe3cient, thevariance �2

Xj is estimated locally with a centered square win-dow �j rather than the whole subband. All the points within�j are assumed to have the same variance �2

�j , which iscalculated as

�2�j =max

0;

1N�j

∑(m;n)∈�j

Z2j (m; n)− �2

j

; (3.8)

where N�j is number of points in window �j . The modi7-cation factor c of the data in window �j is then computedas

c =�2�j

�2�j

+ �2j: (3.9)

MKhLcak et al. [14] also imposed an exponential prior dis-tribution f�(�2) = �e−��2 on the variance �2

�j . Thus an ap-proximate maximum a posteriori (MAP) estimator of �2

�j

is

�2�j =max

0;

N�j

4�

−1 +

√√√√1 +8�N 2

�j

∑(m;n)∈�j

Z2j (m; n)

− �2

j

: (3.10)

The estimation described in Eq. (3.10) was achieved us-ing a more satisfying denoising performance than that ofEq. (3.8). When the size of window �j increases, the twoestimations are identical asymptotically. MKhLcak’s schemeexploits the wavelet intra-scale dependencies and its noiseremoval performance was superior to that of Chang [6].

3.3. The inter-scale dependencies exploited model

Wavelet coe3cient dependencies exist not only withinbut also across subbands. The adjacent wavelet scales arestrongly correlated in general. Observing that small magni-tude coe3cients at coarser scales are more likely to yieldinsigni7cant descendants at a 7ner scale, Shapiro developedthe well-known embedded zerotree wavelet coder [16]. Incontrast, large magnitude wavelet coe3cients produced bytrue signal at 7ner scales are more likely to have signi7cantparents at coarser scales. However, the coe3cients causedby noise would decay rapidly along scales. With this obser-vation, Xu et al. [17] multiplied the adjacent wavelet scalesto sharpen edge structures while weakening noise. Sadler etal. [26] analyzed the multiscale wavelet products and ap-plied them to step detection and estimation.

Another popular wavelet inter-scale dependency rep-resentation is through autoregressive models or hiddenMarkov models. By structuring the image wavelet coe3-cients as a quadtree along scales, Banham [21] imposed anautoregressive model that evolves from coarse scales to 7nescales on the quadtree nodes. He then applied a multiscaleKalman smoothing 7lter after some pre7ltering. The hiddenMarkov models (HMMs), especially the hidden Markovtree model (HMT) proposed by Crouse [18], characterizewell the joint statistics of wavelet coe3cients across scales.Each coe3cient is assigned a hidden state, conditioned onwhich the coe3cients are i.i.d Gaussian. HMT was fur-ther improved by Fan [19] with a new four-state modelcalled HMT-2. Romberg [20] also simpli7ed the HMTby exploiting the inherent self-similarity of real-worldimages.

In this section we present a new inter-scale dependen-cies exploited model based on the LMMSE algorithm. Inthe next section the model will be extended to inter-scaleand intra-scale combined form. By implementing OWE on

Page 5: Hybrid inter- and intra-wavelet scale image restoration

L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746 1741

the noisy image, we group the points with the same spatialposition at J scales as a vector:*Z(m; n) = col{Z1(m; n); : : : ; ZJ (m; n)}; (3.11)

where col denotes column vector. Also*Z =

*X +

*V (3.12)

with*X(m; n) = col{X1(m; n); : : : ; XJ (m; n)} and

*V(m; n) = col{V1(m; n); : : : ; VJ (m; n)}; (3.13)

where*v is the Gaussian vector noise and independent of

*x

The LMMSE of*x can be computed as

*x = P(P + R)−1*z ; (3.14)

where P and R are the autocovariance matrices of*x and

*v ,

respectively,

P = E[*x

*xT] = E

x21 · · · x1xJ

.... . .

...

xJ x1 · · · x2J

and

R= E[*v*vT] = E

v21 · · · v1vJ

.... . .

...

vJ v1 · · · v2J

: (3.15)

The diagonal element of R, E[v2j ], is equal to �2j , which

can be computed by Eq. (3.5). Noise variables vi and vj ,where i = j, are the projections of v on the wavelet bases i and j . They are correlated due to the similarity between i and j . The correlation coe3cient is

'i; j =

∫∫ i(x; y) j(x; y) dx dy√∫∫

2i (x; y) dx dy

√∫∫ 2j (x; y) dx dy

; (3.16)

where vi and vj are jointly Gaussian and their density is

p(vi; vj)

=1

2)�i�j

√1− '2

i; j

e1=2(1−'2i; j)[v2i =�

2i −2'i; jvivj =�i�j+v2j =�

2j ]: (3.17)

Thus the expectation E[vivj] is

Ri; j = E[vivj] =∫ ∫

vivjp(vi; vj) dvi dvj = 'i; j�i�j: (3.18)

Each of the components of matrix P can be computed by

Pi; j = E[xixj] = E[zizj]− Ri; j ; (3.19)

where E[zizj] is estimated as

E[zizj] ≈ 1MN

M∑m=1

N∑n=1

Zi(m; n) · Zj(m; n): (3.20)

3.4. The intra- and inter-scale dependencies combinedscheme

The inter-scale model proposed in Section 3.3 can beimproved to merge intra-scale dependencies. This can beaccomplished by estimating the covariance matrix P locallywith a proper window.

A centered square window � is set for image*Z . All the

vector data in � is assumed to have the same statistics. Theelements of local autocovariance matrix P� are estimated by

P�i; j ≈ max

0;

1N�

∑(m;n)∈�

Zi(m; n)Zj(m; n)− Ri; j

; (3.21)

where N� is the number of vector data in window �. Thenthe image coe3cients within local window � are estimatedby*x = P�(P� + R)−1*z : (3.22)

The above LMMSE (3.22) incorporates the wavelet inter-and intra-scale dependency information together to smooththe noisy image coe3cients.

3.5. The LMMSE and thresholding hybrid algorithm

Suppose the input image is decomposed into J wavelet

scales, and it is found that setting*Z as a J -D vector by Eq.

(3.11) will not yield satisfying results. This could be ex-plained in two ways. Firstly, scale j is strongly correlatedwith scale j+1 but its correlation with scale j+2; j+3; : : : ; Jdecreases rapidly. These coarser scales will not conveymuchuseful information to improve the estimation of scale j. Sec-ondly, a signi7cant structure has larger spatial supports atcoarser scales than 7ner scales. One point at a coarse scalemay appear as an edge, but at 7ner scales the points with thesame spatial orientation may be noise predominated. Thus7ne scales may impose negative e@ects on coarse scales.

With these considerations, we will not employ 7ne scalemeasurements to update coarse scale image estimations. Werecover Xj only from the measurements at adjacent scales j

and j + 1. De7ne the vector image*Z j as

*Z j(m; n) = col{Zj(m; n); Zj+1(m; n)}: (3.23)

So*Z j =

*X j +

*V j (3.24)

with*X j(m; n) = col{Xj(m; n); Xj+1(m; n)} and

*V j(m; n) = col{Vj(m; n); Vj+1(m; n)}: (3.25)

Applying the hybrid intra- and inter-scale LMMSE scheme

in 3.4 to*Z j the estimated image

*xj is obtained. We preserve

only the component X j of it. Similarly, the estimation resultX j+1 would be obtained by applying the LMMSE to data*Z j+1.

Page 6: Hybrid inter- and intra-wavelet scale image restoration

1742 L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746

In the threshold-based denoising schemes, wavelet coef-7cients whose magnitude is below the preset threshold willbe set to 0. This insigni7cant data is predominantly noiseand should be discarded. To merge the merits of threshold-ing into the LMMSE scheme, we apply LMMSE only to thecoe3cients above a threshold and shrink those coe3cientsbelow the threshold to 0. The scale adaptive threshold em-ployed in this paper is

tj = c�j (3.26)

with c being a constant between 3 and 3.5. Most of the Gaus-sian noise will fall into three times its standard deviation.This threshold was proposed by Pan et al. [19] and producesgood results.

The LMMSE and thresholding hybrid algorithm is sum-marized by pseudocode as follows.

First implement the overcomplete wavelet expansionto get Zj; j = 1; 2; : : : ; J .Loop for each scale j{Form the vector image

*Z j and compute the noise

autocovariance matrix R.Set threshold tj = c�j and the centered square window�j with a proper size.Loop for each image block shaped by �j

{Estimate the autocovariance matrix P�j .Loop for each coe8cient within the block{If |Zj(m; n)|¡tjX j(m; n) = 0

Else*xj(m; n) = P�j (P�j + R)−1Z j(m; n)

End

The @rst element of*xj(m; n); X j(m; n), is preserved.

} End of loop} End of loop

} End of loopReconstruct the estimated image by X j ; j = 1; 2; : : : ; J .

3.6. Wavelet bases selection

In Ref. [13], Liu and Moulin measured the wavelet coef-7cient dependencies by mutual information, which relatesfundamentally to data compression, estimation and classi7-cation performance. They found that themutual informationdepends on not only the used statistical models but also theemployed wavelet 7lters. The denoising performance of theproposed scheme is also wavelet dependent. In this paper, wefocus on the widely used compactly supported orthogonaland biorthogonal wavelets constructed by Daubechies [3].

We denote Daubechies’ orthogonal wavelet as Dau(N ),where N = 1; 2; : : : ;∞ is the vanishing moment of thewavelet whose 7lter length is 2N . The biorthogonal wavelet

is denoted by CDF(N; N ′), where N is the vanishing mo-ment of the analytic wavelet 7lter and N ′ is that of thesynthetic wavelet 7lter. Orthogonal wavelets Dau(N ), withthe exception of the Haar wavelet, lack (anti-)symmetry.This is an important property in signal and image pro-cessing. Biorthogonal wavelets CDF(N; N ′) discard theorthogonality to preserve the (anti-)symmetric property.

The proposed inter-scale model exploits the high corre-lation of adjacent wavelet scales. The denoising e3ciencyrelies on how much positive information the coarser scalecan convey to its adjacent 7ner scale. If a signi7cant edgepoint occurs at a 7ner scale then it is expected that a cor-responding edge point will appear with the same sign atcoarser scale. Otherwise, a coarser scale may be detrimentalto a 7ner scale estimation.

De7ne variable yj as the product of two adjacent scalewavelet coe3cients

yj = xjxj+1: (3.27)

Di@erent wavelet 7lters result in di@erent distributions of yj .The histograms of y1 for image Lena obtained from fourwavelets: Dau (2), Dau (4), CDF (1; 3) and CDF (2; 4)are illustrated in Fig. 4. (The histograms of yj , j¿ 1, aresimilar to those of y1 in shape.)

The expectation is that the histograms of yj are asym-metrical and have heavy tails in the positive X -axis, im-plying high correlation of adjacent scales. From Fig. 4 wecan see that the histograms of y1 become more symmetri-cal when the vanishing moments (as well as the length) ofthe wavelet 7lters increase. For example, the histograms forDau (4) are more symmetrical than those for Dau (2). Thesame to CDF (1; 3) and CDF (2; 4). It is observed that thehistograms’ shape for biorthogonal wavelet CDF (1; 3) isbetter than other wavelets. The positive tails are rather con-siderable. CDF (1; 3) has only 1 vanishing moment for an-alytic 7lter while having 3 vanishing moments for synthetic7lter to smooth noise. It is likely to be the most suitablewavelet for our algorithm.

4. Experiments

We call the proposed algorithm in Section 3.6, M3. It iscompared with MKhLcak’s [14] intra-scale dependencies ex-ploited scheme (M1) and the inter-scale dependencies ex-ploited scheme (M2) presented in Section 3.3. It shouldbe noted that we also hybridize scheme M2 with a sim-ilar thresholding strategy as in scheme M3. The waveletsDau (2), Dau (4), CDF (1; 3) and CDF (2; 4) are employedin the experiments.

Several benchmark images have been used to test the threeschemes and the conclusions are identical. The experimentalresults on images Lena and Peppers are reported here. Thecentered square window used in M1 and M3 is of 9×9 size.Three levels of Gaussian white noise with � = 20; 25; 30are added to Lena and Peppers, respectively. The denoising

Page 7: Hybrid inter- and intra-wavelet scale image restoration

L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746 1743

-100 -50 0 50 1000

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

Horizontal

-100 -50 0 50 1000

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

Horizontal

-100 -50 0 50 100 -100 -50 0 50 1000

0.02

0.04

0.06

0.08

0.1

0.12

Vertical

0

0.02

0.04

0.06

0.08

0.1

0.12

0.14

0.16

Diagonal

-100 -50 0 50 1000

0.02

0.04

0.06

0.08

0.1

0.12

0.14

0.16

Diagonal

(a) Dau (2)

(b) Dau (4)

-100 -50 0 50 1000

0.02

0.04

0.06

0.08

0.1

0.12

0.14

Vertical

-100 -50 0 50 100Horizontal

-100 -50 0 50 100Diagonal

(c) CDF (1,3)

-100 -50 0 50 100Vertical

0

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0

0.02

0.04

0.06

0.08

0.1

0

0.05

0.1

0.15

0.2

-100 -50 0 50 100Horizontal

-100 -50 0 50 100Diagonal

(d) CDF (2,4)

-100 -50 0 50 100Vertical

0

0.05

0.1

0.15

0.2

0

0.02

0.04

0.06

0.08

0.1

0

0.02

0.04

0.06

0.08

0.1 0.12

Fig. 4. Histograms of the products of the 7rst two scale wavelet coe3cients for di@erent wavelets. (a) Dau (2); (b) Dau (4); (c) CDF (1; 3);(d) CDF (2; 4).

results by the three methods and four wavelets are listed inTables 1 and 2.

It is observed that algorithm M3 always performs bet-ter than M1 and M2 for a 7xed wavelet base. Furthermore,

scheme M3 by wavelet CDF (1; 3) yields the best denoisingperformance. This is highlighted in Tables 1 and 2. Whilebeing a good bridge to exploit wavelet inter-scale dependen-cies, CDF (1; 3) is not so good at characterizing the wavelet

Page 8: Hybrid inter- and intra-wavelet scale image restoration

1744 L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746

Table 1SNR (dB) results for the three methods for image Lena in di@erent noise levels

Lena Dau (2) Dau (4) CDF (1,3) CDF (2,4)

� = 20 M1 25.74 25.69 25.40 25.66M2 25.76 25.70 25.83 25.35M3 25.95 25.82 26.09 25.67

� = 25 M1 24.56 24.45 24.26 24.44M2 24.53 24.46 24.69 24.29M3 24.98 24.84 25.11 24.69

� = 30 M1 23.53 23.39 23.28 23.37M2 23.55 23.46 23.71 23.27M3 24.09 23.94 24.21 23.81

Table 2SNR (dB) results for the three methods with image Peppers for di@erent noise levels

Peppers Dau (2) Dau (4) CDF (1,3) CDF (2,4)

� = 20 M1 24.49 24.33 24.20 24.39M2 24.70 24.51 24.75 24.27M3 24.94 24.63 25.20 24.52

� = 25 M1 23.42 23.24 23.17 23.31M2 23.74 23.55 23.82 23.29M3 24.09 23.76 24.32 23.62

� = 30 M1 22.44 22.21 22.19 22.27M2 22.75 22.54 22.89 22.31M3 23.31 22.98 23.55 22.82

intra-scale dependencies and it is not an appropriate can-didate for scheme M1. From Tables 1 and 2 we see thatscheme M1 by CDF (1; 3) always reports the poorest re-sults. In contrast methods M2 and M3 take advantage of thewavelet inter-scale dependencies. So wavelets CDF (1; 3)and Dau (2), which have stronger correlation between ad-jacent scales, are more suitable for the two schemes. In Fig.5, the noisy Lena (�= 30) and the three denoised versionsby biorthogonal wavelet CDF (1; 3) are illustrated. Fig. 6shows the noisy Peppers (�=30) and the denoised imagesby CDF (1; 3).

5. Conclusion

In this paper, we presented a wavelet based noisy im-age restoration approach that exploits both the intra- andinter-scale dependencies. The overcomplete wavelet expan-sion, which is more e3cient in denoising than the orthog-onal wavelet expansion, is employed in the scheme. Thewavelet coe3cients with the same spatial orientation at ad-jacent scales are combined as a 2-D vector. By applying the

LMMSE to the vector variable, the inter-scale correlationinformation is utilized to update the estimation of one scale.To exploit the wavelet intra-scale dependencies, we useda centered square window to estimate the autocovariancematrices of wavelet coe3cients locally. Finally, we com-bined the composite dependencies exploited algorithm witha thresholding strategy. The experimental results on bench-mark images illustrate the proposed scheme outperformsthe methods that only exploit the intrascale or interscaledependencies.

The performance of the proposed hybrid intrascale andinterscale denoising scheme also depends on the selection ofthe wavelet 7lters. Four orthogonal and biorthogonal com-pactly supported wavelets were considered in this paper. Thebiorthogonal wavelet CDF (1; 3), which is good at exploit-ing inter-scale dependencies, produces the best performance.The inter-scale correlation decreases with the increasing ofthe vanishing moment of the wavelet 7lter. The wavelet 7l-ter CDF (1; 3) has the least vanishing moment for analytic7lter (1-order) while having 3-order vanishing moments forthe synthetic 7lter to smooth noise. The proposed schemeby CDF (1; 3) produced good results in our experiments.

Page 9: Hybrid inter- and intra-wavelet scale image restoration

L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746 1745

(a) (b)

(d)(c)

Fig. 5. Denoising results of Lena by wavelet CDF (1; 3). (a) Noisy Lena (� = 30). (b) Restored Lena by scheme M1 (SNR = 23:28 dB).(c) Restored Lena by scheme M2 (SNR = 23:71 dB). (d) Restored Lena by scheme M3 (SNR = 24:21 dB).

(a) (b)

(c) (d)

Fig. 6. Denoising results of Peppers by wavelet CDF (1; 3). (a) Noisy Peppers (� = 30). (b) Restored Peppers by scheme M1(SNR = 22:19 dB). (c) Restored Peppers by scheme M2 (SNR = 22:89 dB). (d) Restored Peppers by scheme M3 (SNR = 23:55 dB).

Page 10: Hybrid inter- and intra-wavelet scale image restoration

1746 L. Zhang et al. / Pattern Recognition 36 (2003) 1737–1746

References

[1] S.G. Mallat, A theory for multiresolution signal decomposi-tion: the wavelet representation, IEEE Trans. Pattern Anal.Mach. Intell. 11 (7) (1989) 674–693.

[2] S. Mallat, S. Zhong, Characterization of signals frommultiscale edges, IEEE Trans. Pattern Anal. Mach. Intell.14 (7) (1992) 710–732.

[3] I. Daubechies, Ten lectures on Wavelets, SIAM, Philadelphia,PA, 1992.

[4] M. Vetterli, C. Herley, Wavelet and 7lter banks: theoryand design, IEEE Trans. Signal Process. 40 (9) (1992)2207–2232.

[5] S.G. Chang, B. Yu, M. Vetterli, Adaptive wavelet thresholdingfor image denoising and compression, IEEE Trans. ImageProcess. 9 (9) (2000) 1532–1546.

[6] S.G. Chang, B. Yu, M. Vetterli, Spatially adaptive waveletthresholding with context modeling for image denoising, IEEETrans. Image Process. 9 (9) (2000) 1522–1531.

[7] D.L. Donoho, De-noising by soft thresholding, IEEE Trans.Inf. Theory 41 (5) (1995) 613–627.

[8] D.L. Donoho, I.M. Johnstone, Adapting to unknownsmoothness via wavelet shrinkage, J. Am. Stat. Assoc. 90 (12)(1995) 1200–1224.

[9] Q. Pan, L. Zhang, et al., Two denoising methods by wavelettransform, IEEE Trans. Signal Process. 47 (12) (1999)3401–3406.

[10] F. Abramovich, T. Sapatinas, et al., Wavelet thresholding via aBayesian approach, J. R. Stat. Soc. Ser. B 60 (1998) 725–749.

[11] A. Chambotle, R.A. Devore, et al., Nonlinear wavelet imageprocessing: variational problems, compression, and noiseremoval through wavelet shrinkage, IEEE Trans. ImageProcess. 7 (7) (1998) 319–335.

[12] M. Antonini, M. Barlaud, P. Mathieu, I. Daubechies, Imagecoding using wavelet transform, IEEE Trans. Image Process.1 (4) (1992) 205–220.

[13] J. Liu, P. Moulin, Information-theoretic analysis ofinterscale and intrascale dependencies between image waveletcoe3cients, IEEE Trans. Image Process. 10 (11) (2001)1647–1658.

[14] M.K. MKhLcak, I. Kozintsev, K. Ramchandran, P. Moulin,Low-complexity image denoising based on statistical

modeling of wavelet coe3cients, IEEE Signal Process. Lett.6 (12) (1999) 300–303.

[15] S.D. Servetto, K. Ramchandran, M.T. Orchard, Imagecoding based on a morphological representation ofwavelet data, IEEE Trans. Image Process. 8 (9) (1999)1161–1174.

[16] J.M. Shapiro, Embedded image coding using zerostrees ofwavelet coe3cients, IEEE Trans. Signal Process. 41 (12)(1993) 3445–3462.

[17] Y. Xu, et al., Wavelet transform domain 7lters: a spatiallyselective noise 7ltration technique, IEEE Trans. ImageProcess. 3 (11) (1994) 747–758.

[18] M. Crouse, R. Nowak, R. Baraniuk, Wavelet-based statisticalsignal processing using hidden Markov models, IEEE Trans.Signal Process. 42 (4) (1998) 886–902.

[19] G. Fan, X.G. Xia, Improved hidden Markov models in thewavelet-domain, IEEE Trans. Signal Process. 49 (1) (2001)115–120.

[20] J.K. Romberg, H. Choi, R.G. Baraniuk, Bayesiantree-structured image modeling using wavelet-domain hiddenMarkov models, IEEE Trans. Image Process. 10 (7) (2001)1056–1068.

[21] M.R. Banham, A.K. Katsaggelos, Spatially adaptivewavelet-based multiscale image restoration, IEEE Trans.Image Process. 5 (4) (1996) 619–634.

[22] J. Portilla, V. Strela, M.J. Wainwright, E.P. Simoncelli,Adaptive wiener denoising using a Gaussian scale mixturemodel in the wavelet domain, Proceedings of the EightInternational Conference on Image Processing, Thessaloniki,Greece, 2001, pp. II. 37–40.

[23] J. Liu, P. Moulin, Image denoising based on scale-spacemixture modeling for wavelet coe3cients, Proceedings ofthe International Conference on Image Processing (ICIP’99)Kobe, Japan, 1999, pp. I. 386–390.

[24] R.R. Coifman, D.L. Donoho, Translation-invariant de-noising,in: A. Antoniadis, G. Oppenheim (Eds.), Wavelet andStatistics, Springer, Berlin, Germany, 1995.

[25] E.W. Karmen, J.K. Su, Introduction to Optimal Estimation,Springer, London, 1999.

[26] B.M. Sadler, A. Swami, Analysis of multiscale products forstep detection and estimation, IEEE Trans. Inf. Theory 45 (4)(1999) 1043–1051.

About the Author—LEI ZHANG was born in 1974 in P.R. China. He received the M.Sc and Ph.D Degree in Electrical and Engineering fromNorthwestern Polytechnical University, Xi’an, P.R. China, respectively, in 1998 and 2001. He is currently working as a research associate inthe Dept. of Computing, The Hong Kong Polytechnic University. His research interests include digital signal and image processing, wavelettransform and optimal estimation theory.

About the Author—PAUL BAO received the B.Sc (82) and M.Sc (84) Degrees in Mathematics and Computer Science from ChongqingUniversity and Jiao Tong University respectively and Ph.D. (88) in Computer Science from University of Calgary, Canada for the work oncomputer graphics, computational geometry and algorithm design. He served on the faculty of Computer Science Department, Universityof Calgary from 1988–1990 and then proceeded to work in IBM Canada as a sta@ analyst in the areas of Object-Oriented Technologies andDistributed Computing Systems from 1990–1995. Since July 1995, he has been with Computing Department, the Hong Kong PolytechnicUniversity where he is now an associate professor. Dr. Bao’s research interests included Computer Graphics, Image-based Rendering, ImageCompression, Neural Network for Image Processing, Compression and fuzzy-based Financial Engineering.

About the Author—XIAOLIN WU received his B.Sc. in computer science from Wuhan Cehui Technical University, Wuhan, China, in1982 and Ph.D. in computer science from University of Calgary, Calgary, Canada, in 1988. He is currently a professor with Departmentof Computer Science, University of Western Ontario, London, Canada. His research interests include algorithms for visual computing andcommunication, data compression, image processing and computer graphics. He has published more than 70 research papers in those areas.Professor Wu is an active JPEG member and the 1998 UWO Distinguished Research Professor.