coherent multiscale image processing using quaternion waveletswailam/research/mythesis.pdf ·...

61
RICE UNIVERSITY Coherent Multiscale Image Processing using Quaternion Wavelets by Wai Lam Chan A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree Master of Science Approved, Thesis Committee: Richard Baraniuk, Chair Professor of Electrical and Computer Engineering, Rice University Michael Orchard Professor of Electrical and Computer Engineering, Rice University Deceased Hyeokho Choi Adjunct Assistant Professor of Electrical and Computer Engineering, Rice University Houston, Texas May, 2006

Upload: others

Post on 12-Jul-2020

14 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

RICE UNIVERSITY

Coherent Multiscale Image Processing

using Quaternion Wavelets

by

Wai Lam Chan

A Thesis Submitted

in Partial Fulfillment of the

Requirements for the Degree

Master of Science

Approved, Thesis Committee:

Richard Baraniuk, ChairProfessor of Electrical and ComputerEngineering, Rice University

Michael OrchardProfessor of Electrical and ComputerEngineering, Rice University

Deceased

Hyeokho ChoiAdjunct Assistant Professor of Electricaland Computer Engineering,Rice University

Houston, Texas

May, 2006

Page 2: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

ABSTRACT

Coherent Multiscale Image Processing

using Quaternion Wavelets

by

Wai Lam Chan

This thesis develops a quaternion wavelet transform (QWT) as a new multiscale

analysis tool for geometric image features. The QWT is a near shift-invariant tight

frame representation whose coefficients sport a magnitude and three phases: two

phases encode local image shifts while the third contains textural information. The

QWT can be computed using a dual-tree filter bank with linear computational com-

plexity. We develop the QWT by applying an alternative theory of 2-D Hilbert trans-

forms and analytic signals to the 1-D complex wavelet transform. To demonstrate

the properties of the QWT’s coherent magnitude/phase representation, we develop a

efficient and accurate algorithm for estimating the local geometrical structure of im-

ages and a multiscale algorithm for estimating the disparity between a pair of images.

The latter algorithm has potentials for various applications in image registration and

flow estimation. It uses an interesting multiscale phase unwrapping procedure and

features linear complexity and sub-pixel accuracy.

Page 3: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

Acknowledgments

My sincere thanks to Dr. Hyeokho Choi for his valuable insights and guidance; to

Dr. Richard Baraniuk for his encouragement and research support; to Dr. Michael

Orchard for being part of my thesis committee; to my friends, Vinay Ribeiro and

Gang Hua, for their help in preparation for my defense; and to many others for their

constant prayer, support and friendship; and last but not the least, to my family for

their love and patience.

Page 4: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

Contents

Abstract ii

Acknowledgments iii

List of Illustrations vi

1 Introduction 1

2 Quaternion Wavelet Transform 8

2.1 Real and Complex Wavelet Transforms . . . . . . . . . . . . . . . . . 8

2.2 Quaternion Wavelet Transform (QWT) . . . . . . . . . . . . . . . . . 14

2.2.1 Quaternion Hilbert Transform . . . . . . . . . . . . . . . . . . 14

2.2.2 QWT Construction . . . . . . . . . . . . . . . . . . . . . . . . 16

3 QWT Properties 18

3.1 Tight Frame . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18

3.2 Relationship to the 2-D CWT . . . . . . . . . . . . . . . . . . . . . . 19

3.3 QWT of Real Signals . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

3.4 QWT Phase Properties . . . . . . . . . . . . . . . . . . . . . . . . . . 22

4 QWT Applications 24

4.1 Edge Geometry Estimation . . . . . . . . . . . . . . . . . . . . . . . . 24

Page 5: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

v

4.2 QWT-based Image Disparity Estimation . . . . . . . . . . . . . . . . 30

5 Conclusions 39

A QFT Plancharel Theorem 41

B Continuous QWT of 2-D edge signal 42

B.1 Proof of Theorem 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42

B.2 Calculation of (4.2) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44

B.3 Calculation of QWT Phase Angles for a Step Edge . . . . . . . . . . 44

C Theoretical Analysis of Leakage Effect 46

Bibliography 49

Page 6: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

Illustrations

1.1 Three real wavelets (from the vertical, horizontal, and diagonal subbands,

respectively) from the 2-D DWT basis generated from the length-14

Daubechies filter. The color map sets large negative values to blue, zero to

green, and large positive values to red. . . . . . . . . . . . . . . . . . . . 2

1.2 Six complex wavelets from the 2-D dual-tree CWT frame generated from

the orthogonal near-symmetric filters [1] in the first stage and the Q-filters

[2] in subsequent stages. (a) Real parts; (b) imaginary parts. . . . . . . . 4

1.3 Three quaternion wavelets from the 2-D dual-tree QWT frame. Each

quaternion wavelet comprises four components that are 90◦ phase shifts of

each other in the vertical, horizontal, and both directions. (a) Diagonal

subband, from left to right: ψh(x)ψh(y) (a usual, real DWT tensor

wavelet), ψh(x)ψg(y), ψg(x)ψh(y), ψg(x)ψg(y). The image on the far right

is the quaternion wavelet magnitude |ψq(x, y)|, a non-oscillating function.

(b) Horizontal subband, from left to right: φh(x)ψh(y) (a usual, real DWT

tensor wavelet), φh(x)ψg(y), φg(x)ψh(y), φg(x)ψg(y). (c) Vertical subband,

from left to right: ψh(x)φh(y) (a usual, real DWT tensor wavelet),

ψh(x)φg(y), ψg(x)φh(y), ψg(x)φg(y). . . . . . . . . . . . . . . . . . . . 6

Page 7: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

vii

2.1 (a) Magnitude of the 1-D step response of length-8 Daubechies DWT

coefficients d`,p(t0) at four consecutive scales. The horizontal axis

represents the location of the step edge, and the vertical axis represents

the magnitude of the corresponding wavelet coefficient. The vertical line

passes through wavelet coefficient values for one particular location of an

edge. Wavelet magnitudes vary significantly across both scales and various

step locations, indicating the shift-variance of the DWT. (b) Step response

of the CWT coefficients with the same locales. The non-oscillatory

response indicates the near shift-invariance of the CWT. . . . . . . . . . 10

2.2 The 1-D dual-tree CWT is implemented using a pair of filter banks

operating on the same data simultaneously. Outputs of the filter banks are

the dual-tree scaling coefficients, chpand cgp , and the wavelet coefficients,

dh`,pand dg`,p

, at scale ` and shift p. The CWT coefficients are then

obtained as dh`,p+ j dg`,p

. . . . . . . . . . . . . . . . . . . . . . . . . . 11

2.3 (a) The CWT coefficient’s single phase angle responds linearly to image

shifts d in a direction orthogonal to the wavelet’s orientation. (b) Two of

the QWT coefficient’s three phase angles respond linearly to image shifts d

in an absolute horizontal/vertical coordinate system. . . . . . . . . . . . 13

2.4 Complex Fourier domain relationships among the four quadrature

components of a quaternion wavelet ψq(x, y) in the diagonal subband. . . 17

Page 8: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

viii

3.1 Effect of varying θ3 on the structure of the corresponding weighted

quaternion wavelet from the diagonal subband (left to right):

θ3 = −π4 ,−π

8 , 0,π8 ,

π4 respectively. The corresponding wavelet changes from

textured (θ3 = 0) to oriented (θ3 = ±π4 ). . . . . . . . . . . . . . . . . . 23

4.1 (a) Parameterization of a single edge in a dyadic image block

(wedgelet [3]). (b) QFT spectrum of the edge; shaded squares represent

the quaternion wavelets in the vertical, horizontal, and diagonal subbands.

The energy of the edge is concentrated along the two dark lines crossing at

the origin and is captured by the horizontal subband with spectral center

at quaternion frequency (u, v). The region bounded by the dashed line

demonstrates the spectral support of the QWT basis “leaking” into the

neighboring quadrant. . . . . . . . . . . . . . . . . . . . . . . . . . . . 25

4.2 Magnitude of a QWT coefficient as a function of edge orientation β

and offset d for the (a) horizontal φ(x)ψ(y) subband and (b) diagonal

ψ(x)ψ(y) subband. (c), (d) Absolute error for estimating d by the

phase angles (θ1,θ2) from the same two subbands, relative to the

normalized edge length of the dyadic block (= 1 unit). Estimates are

accurate in the regions where the coefficient magnitudes are large. . . 26

Page 9: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

ix

4.3 Local edge geometry estimation using the QWT. (a) Several edgy regions

from the “cameraman” image are shown on the left; (b)–(e) on the right

are edge estimates from the corresponding QWT coefficients. The upper

row shows the original image region, the lower row shows a wedgelet (see

Figure 4.1(a)) having the edge parameter estimates (β, d). (No attempt is

made to capture the texture within the block.) . . . . . . . . . . . . . . 30

4.4 Multiscale QWT phase-based disparity estimation results. (a), (b)

Reference and target images from the “Rubik’s cube” image sequence [4].

(c) Disparity estimates between two images in the sequence, shown as

arrows overlaid on top of the reference image (zoomed in for better

visualization). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

4.5 Comparison of multiscale QWT phase-based disparity estimation with two

motion estimation algorithms, Gradient Correlation (GC) [5] and

exhaustive search (ES). Performance measure is the PSNR (in dB) between

the motion-compensated image and the target image of three test image

sequences (“Rubik”, “Taxi” and “Mobcal”). (a) Frame-by-frame PSNR

performance comparison in the “Rubik” sequence. (b) Table of average

PSNR performance (over all frames) for each test sequence. Multiscale

QWT phase-based method demonstrates the best performance among the

three test algorithms for the “Rubik” sequence, and shows comparable

performance to the other algorithms for the “Taxi” and “Mobcal” sequences. 36

Page 10: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

1

Chapter 1

Introduction

The encoding and estimation of the relative locations of image features play an im-

portant role in many image processing applications, ranging from feature detection

and target recognition to image compression. In edge detection, for example, the

goal is to locate object boundaries in an image. In image denoising or compression,

state-of-the-art techniques achieve significant performance improvements by exploit-

ing information on the relative locations of large transform coefficients [6, 7, 8, 9].

An efficient way to compute and represent relative location information in signals

is through the phase of the Fourier transform. The Fourier shift theorem provides a

simple linear relationship between the signal shift and the Fourier phase. When only

a local region of the signal is of interest, the short-time Fourier transform (STFT)

provides a local Fourier phase for each windowed portion of the signal. The use of the

Fourier phase to decipher the relative locations of image features is well established in

the image processing and computer vision communities for applications such as stereo

matching and image registration [10, 11, 12]. Indeed, the classic experiment of Lim

and Oppenheim [13] demonstrated that for natural images the Fourier phase contains

a wealth of information, often more than the magnitude. Estimating location infor-

mation using local phase provides more robust estimates with sub-pixel accuracy and

requires less computational effort compared to purely amplitude-based approaches.

Page 11: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

2

Figure 1.1 : Three real wavelets (from the vertical, horizontal, and diagonal subbands,respectively) from the 2-D DWT basis generated from the length-14 Daubechies filter. Thecolor map sets large negative values to blue, zero to green, and large positive values to red.

Another, more far-flung application of Fourier phase is in radar interferometry, which

infers a topographic map from the phase differences between signals at two (or more)

receivers [14].

For signals containing isolated singularities, such as piecewise smooth functions,

the discrete wavelet transform (DWT) has proven to be more efficient than the STFT.

The locality and zooming properties of the wavelet basis functions leads to a sparse

representation of such signals that compacts the signal energy into a small number

of coefficients. Wavelet coefficient sparsity is the key enabler of algorithms such as

wavelet-based denoising by shrinkage [15]. Many natural images consist of smooth

or textured regions separated by edges and are well-suited to wavelet analysis and

representation. Other advantages of wavelet analysis include its multiscale structure,

invertibility, and efficient filter-bank implementation. 2-D DWT basis functions are

easily formed as the tensor products of 1-D DWT basis function along the vertical

and horizontal directions; see Figure 1.1.

The conventional, real-valued DWT, however, suffers from two major drawbacks.

The first drawback is shift-variance: a small shift of the signal causes significant

fluctuations in wavelet coefficient energy, making it difficult to extract or model signal

Page 12: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

3

information from the coefficient values. The second drawback is the lack of a notion

of phase to encode signal location information as in the Fourier case.

Complex wavelet transforms (CWTs) provide an avenue to remedy these two

drawbacks of the DWT. It is interesting to note that the earliest modern wavelets,

those of Grossmann and Morlet [16], where in fact complex, and Grossman continu-

ally emphasized the power of the CWT phase for signal analysis and representation.

Subsequent researchers have developed orthogonal or biorthogonal CWTs; see, for

example [17, 18, 19, 20, 21, 22].

A productive line of research has developed over the past decade on the dual-tree

CWT, which in 1-D combines two orthogonal or biorthogonal wavelet bases using

complex algebra into a single system, with one basis corresponding to the “real part”

of the complex wavelet and the other to the “imaginary part” [23]. Ideally, the real

and imaginary wavelets are a Hilbert transform pair (90◦ out of phase) and form an

analytic wavelet supported on only the positive frequencies in the Fourier domain,

just like the cosine and sine components of a complex sinusoid. The 1-D dual-tree

CWT is a slightly (2×) redundant tight frame, and the magnitudes of its coefficients

are nearly shift invariant [23]. There also exists an approximate linear relationship

between the dual-tree CWT phase and the locations of 1-D signal singularities [24]

as in the Fourier shift theorem.

The 2-D dual-tree CWT for images is based on the theory of the 2-D Hilbert

transform (HT) and 2-D analytic signal first suggested by Hahn [25]. In particular,

a 2-D dual-tree complex wavelet is formed by the 1-D HT of a usual 2-D real DWT

wavelet in either or both of the horizontal or/and vertical directions. The result is

Page 13: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

4

(a)

(b)

Figure 1.2 : Six complex wavelets from the 2-D dual-tree CWT frame generated from theorthogonal near-symmetric filters [1] in the first stage and the Q-filters [2] in subsequentstages. (a) Real parts; (b) imaginary parts.

a 4× redundant tight frame with six directional subbands; see Figure 1.2 [23, 26].

The 2-D CWT is near shift-invariant, and its magnitude-phase representation has a

complex phase component that encodes shifts of local 1-D structures in images such

as edges and ridges [27]. As a result, the 2-D dual-tree CWT has proved useful for a

variety of tasks in image processing [8, 27, 28, 29].

Each 2-D dual-tree CWT basis function has one phase angle, which encodes the

1-D shift of image features perpendicular to its orientation. This may be sufficient

for analyzing local 1-D structures such as edges. However, when the feature under

analysis is intrinsically 2-D [4] — for example an image T-junction — then its rel-

ative location is defined in both the horizontal and vertical directions. This causes

ambiguity in the CWT phase shift; we cannot resolve the image shifts in both the

horizontal and vertical directions from the change of only one CWT coefficient phase.

To overcome this ambiguity, we must conduct a joint analysis with two CWT phases

from differently oriented subbands, which can complicate image analysis and model-

ing considerably.

Page 14: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

5

In this thesis, we explore an alternative theory of the 2-D HT and analytic signal

due to Bulow [4, 30] and show that it leads to an alternative to the dual-tree CWT. In

Bulow’s HT, the 2-D analytic signal is defined by limiting the 2-D Fourier spectrum to

a single quadrant. Applying this theory within the dual tree framework, we develop

and study in this thesis a new dual-tree quaternion wavelet transform (QWT), where

each quaternion wavelet consists of a real part (a usual real DWT wavelet) and three

imaginary parts that are organized by quaternion algebra; see Figure 1.3. The QWT,

is a 4× redundant tight frame with three subbands (horizontal, vertical and diagonal).

It is also near shift-invariant.

The QWT inherits a quaternion magnitude-phase representation from the quater-

nion Fourier transform (QFT). The first two QWT phases (θ1,θ2) encode the shifts of

image features in the absolute horizontal/vertical coordinate system, while the third

phase θ3 encodes edge orientation mixtures and textural information.

To illustrate the power of coherent processing using both the QWT magnitude

and phase, we consider two image processing applications. In the first application,

edge estimation, we demonstrate the efficiency of the QWT magnitude-phase repre-

sentation for encoding the orientation and offset of edges in local image blocks. Our

algorithm is entirely based on the QWT shift theorem and the interpretation of the

QWT as a local QFT analysis. In the second application, image registration, we design

a new multiscale image disparity estimation algorithm. The QWT provides a natu-

ral multiscale framework for measuring and adjusting local disparities and performing

phase unwrapping from coarse to fine scales with linear computational efficiency. The

convenient QWT encoding of location information in the absolute horizontal/vertical

Page 15: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

6

(a)

(b)

(c)

Figure 1.3 : Three quaternion wavelets from the 2-D dual-tree QWT frame. Each quater-nion wavelet comprises four components that are 90◦ phase shifts of each other in the ver-tical, horizontal, and both directions. (a) Diagonal subband, from left to right: ψh(x)ψh(y)(a usual, real DWT tensor wavelet), ψh(x)ψg(y), ψg(x)ψh(y), ψg(x)ψg(y). The image onthe far right is the quaternion wavelet magnitude |ψq(x, y)|, a non-oscillating function. (b)Horizontal subband, from left to right: φh(x)ψh(y) (a usual, real DWT tensor wavelet),φh(x)ψg(y), φg(x)ψh(y), φg(x)ψg(y). (c) Vertical subband, from left to right: ψh(x)φh(y)(a usual, real DWT tensor wavelet), ψh(x)φg(y), ψg(x)φh(y), ψg(x)φg(y).

coordinate system facilitates averaging across subband estimates for more robust per-

formance. Our algorithm offers sub-pixel estimation accuracy and runs faster than

existing disparity estimation algorithms like block matching and phase correlation [5].

When the underlying image disparity field is smooth, our method also has superior

performance over these existing techniques.

Previous work in quaternions and the theory of 2-D HT and analytic signals for

image processing includes Bulow’s extension of the Fourier transform and complex

Gabor filters to a quaternion Fourier transform (QFT) [4]. Our QWT can be inter-

Page 16: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

7

preted as a local QFT and thus inherits many of its interesting and useful theoretical

properties such as the quaternion phase representation, symmetry properties, and

shift theorem. In addition, our QWT sports a linear-time and invertible computa-

tional algorithm based on filter banks. There are also interesting connections between

our QWT and the (non-redundant) quaternion wavelet representations of Ates and

Orchard [31] and Gang and Orchard [32]. It is also worthwhile to note a third al-

ternative 2-D HT called the Riesz transform and its associated analytic signal called

the monogenic signal [33]. The monogenic signal, generated by spherical quadrature

filters, has a vector-valued phase that encodes both the orientations of intrinsically

1-D (edge-like) image features and their shifts normal to the edge orientation. While

useful for image processing applications such as edge detection and stereo correspon-

dence [33, 34], this transform is not invertible and thus can only be used as an analysis

tool.

This thesis is organized as follows. We start by briefly reviewing the DWT and

dual-tree CWT in Chapter 2, which then develops the QWT. Chapter 3 discusses

some important properties of the new transform, in particular, its phase response to

singularities. We develop and demonstrate the QWT phase-based image registration

algorithm in Chapter 4. Chapter 5 concludes the thesis and discusses the potential

of the QWT for future applications. The appendix contains detailed derivations and

proofs of some QWT properties and theorems from Chapter 3 and 4.

Page 17: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

8

Chapter 2

Quaternion Wavelet Transform

2.1 Real and Complex Wavelet Transforms

This section overviews the real DWT and the dual-tree CWT. We also develop a new

formulation for the 2-D CWT [23, 26] using the theory of 2-D HTs.

The DWT represents a 1-D real signal f(t) in terms of shifted versions of a scaling

function φ(t) and shifted and scaled versions of a wavelet function ψ(t) [35]. The

functions φL,p(t) = 2Lφ(2Lt − p) and ψ`,p(t) = 2`ψ(2`t − p), ` ≥ L, p ∈ Z form an

orthonormal basis, and we can represent any f(t) ∈ L2(R) as

f(t) =∑

p∈Z

cL,pφL,p(t) +∑

`≥L,p∈Z

d`,pψ`,p(t), (2.1)

where cL,p =∫

f(t)φL,p(t)dt and d`,p =∫

f(t)ψ`,p(t)dt are the scaling and wavelet co-

efficients, respectively. The parameter L sets the coarsest scale space that is spanned

by φL,p(t). Behind each wavelet transform is a filterbank based on lowpass and high-

pass filters.

The standard 2-D DWT is obtained using tensor products of 1-D DWTs over the

horizontal and vertical dimensions. The result is the scaling function φ(x)φ(y) and

three subband wavelets ψ(x)ψ(y), φ(x)ψ(y), and ψ(x)φ(y) that are oriented in the

diagonal, horizontal, and vertical directions, respectively [35] (see Figure 1.1).

The DWT suffers from shift variance; that is, a small shift in the signal can

Page 18: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

9

greatly perturb the magnitude of wavelet coefficients around singularities. Consider

a simple 1-D example: a piecewise smooth signal f(t − t0) such as a step function

u(t) analyzed by a DWT basis having a sufficient number of vanishing moments. Its

wavelet coefficients are samples of the step response of the wavelet

d`,p(t0) = ∆

∫ 2`t0−p

−∞

ψ(t) dt (2.2)

where ∆ is the height of the jump. Since ψ(t) is a bandpass function that oscillates

around zero, so does its step response d`,p as a function of p. Moreover, the factor 2`

in the upper limit (` ≥ 0) amplifies the sensitivity of d`,p to the time shift t0, leading

to strong shift variance. Figure 2.1(a) demonstrates this shift-variance of the DWT.

The wavelet coefficients representing the edge oscillate significantly both across scales

and across shifts within the same scale. Besides shift-variance, the DWT also lacks

a notion of phase to encode signal location information. These problems complicate

modeling and information extraction in the wavelet domain.

The 1-D dual-tree CWT expands a real signal in terms of two sets of wavelet

and scaling functions obtained from two independent filterbanks [23], as shown in

Figure 2.2. We will use the notation φh(t) and ψh(t) to denote the scaling and

wavelet functions and chL,pand dh`,p

to denote their corresponding coefficients, where

h specifies a particular set of wavelet filters. The wavelet functions ψh(t) and ψg(t)

from the two trees play the role of the real and imaginary parts of a complex analytic

wavelet ψc(t) = ψh(t)+jψg(t). The imaginary wavelet ψg(t) is the 1-D HT of the real

wavelet ψh(t). The combined system is a 2× redundant tight frame that, by virtue

Page 19: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

10

(a) D8 magnitude step response (b) complex magnitude step response

Figure 2.1 : (a) Magnitude of the 1-D step response of length-8 Daubechies DWT coeffi-cients d`,p(t0) at four consecutive scales. The horizontal axis represents the location of thestep edge, and the vertical axis represents the magnitude of the corresponding wavelet coef-ficient. The vertical line passes through wavelet coefficient values for one particular locationof an edge. Wavelet magnitudes vary significantly across both scales and various step loca-tions, indicating the shift-variance of the DWT. (b) Step response of the CWT coefficientswith the same locales. The non-oscillatory response indicates the near shift-invariance ofthe CWT.

of the fact that |ψc(t)| is non-oscillating, is shift-invariant.∗ Figure 2.1(b) shows the

smooth magnitude step responses of the CWT coefficients at four consecutive scales.

Compared to the oscillatory step response in the DWT, the CWT step response

magnitude varies monotonically with the distance of the edge from the location of the

wavelet basis function. Thus the CWT magnitude is always large around singularities,

which is beneficial in many applications [36].

It is useful to recall that the Fourier transform of the imaginary wavelet Ψg(ω)

equals −jΨh(ω) when ω > 0 and jΨh(ω) when ω < 0. Thus, the Fourier transform of

the complex wavelet function Ψh(ω) + jΨg(ω) = Ψc(ω) has no (or little in practice)

∗In practice, in order to have finite-length wavelets, the HT is only approximately satisfied, ψc(t)

is only approximately analytic, and the CWT is only approximately shift-invariant [23, 26].

Page 20: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

11

2

2g (n)

g (n)0

1

2

2

h (n)

h (n)

0

1

2

2

h (n)

h (n)

0

1

2

2g (n)

g (n)0

1

2

2

2

2g (n)

g (n)0

1

h (n)

h (n)

0

1 h1ψ

g1

ψ

g2

ψ

h2ψ

h3ψ

φh

g3

ψ

φg(n)

(n)

(n)

(n)

(n)

(n)

(n)

(n)

Figure 2.2 : The 1-D dual-tree CWT is implemented using a pair of filter banks operatingon the same data simultaneously. Outputs of the filter banks are the dual-tree scalingcoefficients, chp

and cgp , and the wavelet coefficients, dh`,pand dg`,p

, at scale ` and shift p.The CWT coefficients are then obtained as dh`,p

+ j dg`,p.

energy in the negative frequency region, making it an analytic signal [23].

Extending the 1-D CWT to 2-D requires an extension of the HT and analytic

signal. There exist not one but several different definitions that each zero out a

different portion of the 2-D frequency plane [30]. We will consider two definitions.

The first, proposed in [25], employs complex algebra and zeros out frequencies on all

but a single quadrant (kx, ky > 0, for example). In this formulation, the complete

2-D analytic signal consists of two parts: one having spectrum on the upper right

quadrant (kx, ky > 0) and the other on the upper left quadrant (kx < 0, ky > 0) [30].

Definition 1 [25] Let f be a real, 2-D function. The complete 2-D complex analytic

signal is defined in the space domain, x = (x, y), as the pair of complex signals

fA1(x) = (f(x)− fHi(x)) + j(fHi1(x) + fHi2(x)), (2.3)

fA2(x) = (f(x) + fHi(x)) + j(fHi1(x)− fHi2(x)), (2.4)

Page 21: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

12

where

fHi1(x) = f(x) ∗ ∗ δ(y)πx

, (2.5)

fHi2(x) = f(x) ∗ ∗ δ(x)πy

, (2.6)

fHi(x) = f(x) ∗ ∗ 1

π2xy. (2.7)

The function fHi is the total HT; the functions fHi1 and fHi2 are the partial HTs; δ(x)

and δ(y) are impulse sheets along the y-axis and x-axis respectively; and ∗∗ denotes

2-D convolution.

The 2-D complex analytic signal in (2.3)–(2.4) is the notion behind the 2-D dual-

tree CWT [23, 26]. Each 2-D CWT basis function is a 2-D complex analytic signal

consisting of a standard DWT tensor wavelet plus three additional real wavelets

obtained by 1-D HTs along either or both coordinates. For example, starting from

real DWT’s diagonal-subband tensor product wavelet f(x) = ψh(x)ψh(y) from above

we obtain from equations (2.5)-(2.7) its partial and total HTs

(fHi1 , fHi2 , fHi) = (ψg(x)ψh(y), ψh(x)ψg(y), ψg(x)ψg(y)).

¿From Definition 1, we then obtain the two complex wavelets

ψc1(x, y) = (ψh(x)ψh(y)− ψg(x)ψg(y)) + j(ψh(x)ψg(y) + ψg(x)ψh(y)), (2.8)

ψc2(x, y) = (ψh(x)ψh(y) + ψg(x)ψg(y)) + j(ψh(x)ψg(y)− ψg(x)ψh(y)), (2.9)

having orientations, 45◦ and −45◦, respectively. Similar expressions can be obtained

for the other two subbands (±15◦ and ±75◦) based on ψh(x)φh(y) and φh(x)ψh(y).

Each 2-D CWT coefficient has only one phase angle, which encodes the 1-D shift

of image features perpendicular to its subband direction. Figure 2.3(a) illustrates this

Page 22: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

13

(a) (b)

Figure 2.3 : (a) The CWT coefficient’s single phase angle responds linearly to image shiftsd in a direction orthogonal to the wavelet’s orientation. (b) Two of the QWT coefficient’sthree phase angles respond linearly to image shifts d in an absolute horizontal/verticalcoordinate system.

phase-shift property. This encoding may be sufficient for local 1-D structures such

as edges, since we can define shifts of an edge uniquely by a single value, say d, in

the direction perpendicular to the edge. However, even in this case, the analysis is

not so straightforward when the edge does not align with the six orientations of the

CWT subbands. And for intrinsically 2-D (non-edge-like) image features such as in

Figure 2.3(a), we must define its relative location using two values (d1,d2) in the x and

y directions, respectively. This creates ambiguity in the CWT phase shift. We can re-

solve this ambiguity with the orthogonal CWT subband, but this complicates the use

of the CWT for image analysis, modeling, and other image processing applications.

In contrast, Figure 2.3(b) illustrates a more convenient encoding of image shifts in

absolute x, y-coordinates (with two phase angles) using the quaternion phases of our

new QWT, to which we now turn our attention.

Page 23: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

14

2.2 Quaternion Wavelet Transform (QWT)

2.2.1 Quaternion Hilbert Transform

There are several alternatives to the 2-D analytic signal of Definition 1; we focus on

one here due to Bulow [30]. It combines the partial and total HTs from (2.5)–(2.7) to

form an analytic signal comprising a real part and three imaginary components that

are manipulated using quaternion algebra [4].

The set of quaternions H = {a+ j1b+ j2c+ j3d | a, b, c, d ∈ R} has multiplication

rules j1j2 = −j2j1 = j3 and j21 = j2

2 = −1 as well as component-wise addition and

multiplication by real numbers [37]. Additional multiplication rules include: j23 = −1,

j2j3 = −j3j2 = j1 and j3j1 = −j1j3 = j2. Note that the quaternionic multiplication is

not commutative. The conjugate q∗ of a quaternion q = a+ j1b+ j2c+ j3d is defined

by q∗ = a− j1b− j2c− j3d while the magnitude is defined as |q| = √qq∗.

An alternative representation for a quaternion is through its magnitude and three

phase angles: q = |q| ej1θ1ej3θ3ej2θ2 [4], where (θ1, θ2, θ3) are the quaternion phase

angles, with each angle uniquely defined within the range (θ1, θ2, θ3) ∈ [−π, π) ×

[−π2, π

2)× [−π

4, π

4].

The operation of conjugation in the set of complex numbers, C = a+ jb|a, b ∈ R

where j2 = −1, is a so-called algebra involution that fulfills the two following prop-

erties. Let z, w ∈ C → (z∗)∗ = z and (wz)∗ = w∗z∗. In H, there are three nontrivial

Page 24: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

15

algebra involutions:

α(q) = a+ j1b− j2c− j3d, (2.10)

β(q) = a− j1b+ j2c− j3d, (2.11)

γ(q) = a− j1b− j2c+ j3d. (2.12)

Using these involutions we can extend the definition of Hermitian symmetry. A

function f : R2 → H is called quaternionic Hermitian if, for each (x, y) ∈ R2,

f(−x, y) = β(f(x, y)) and f(x,−y) = α(f(x, y)). (2.13)

This leads directly to an alternative definition of 2-D analytic signal.

Definition 2 [30] Let f be a real 2-D signal. The 2-D quaternion analytic signal is

defined as

fqA(x) = f(x) + j1fHi1(x) + j2fHi2(x) + j3fHi(x), (2.14)

where the functions fHi1 , fHi2 and fHi are defined as in (2.5)–(2.7).

This 2-D quaternion analytic signal forms the core of the quaternion Fourier

transform (QFT) [4]. The QFT of a real 2-D signal f is given by

F q(u) =

R2

e−j12πuxf(x)e−j22πvydx, (2.15)

where u = (u, v) indexes the QFT domain, x = (x, y) indexes the space domain, and

the quaternion exponential

e−j12πuxe−j22πvy (2.16)

is the QFT basis function (and itself a 2-D quaternion analytic signal). We employ

the notation u = (u, v) for the QFT domain, as opposed to (kx, ky) for the usual

complex Fourier transform domain.

Page 25: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

16

2.2.2 QWT Construction

Our new 2-D dual-tree QWT rests on the quaternion definition of 2-D analytic signal.

By organizing the four quadrature components of a 2-D wavelet (the real wavelet and

its 2-D HTs) as a quaternion, we obtain a 2-D analytic wavelet and its associated

quaternion wavelet transform (QWT). For example, for the diagonal subband, with

(f, fHi1, fHi2 , fHi) = (ψh(x)ψh(y), ψg(x)ψh(y), ψh(x)ψg(y), ψg(x)ψg(y)), we obtain the

quaternion wavelet

ψq(x, y) = ψh(x)ψh(y)− j1ψg(x)ψh(y)− j2ψh(x)ψg(y) + j3ψg(x)ψg(y). (2.17)

Figure 1.3(a) illustrates the four components of a quaternion wavelet and its

quaternion magnitude for the diagonal subband. The partial and total HT com-

ponents resemble ψh(x)ψh(y) but are phase-shifted by 90◦ in the horizontal, vertical,

and both directions, respectively. The magnitude of each quaternion wavelet (square

root of the sum-of-squares of all four components) is a bell-shaped function. We can

also interpret the four components of ψq(x, y) in the Fourier domain as multiplying

the quadrants of the Fourier transform of ψh(x)ψh(y) by ±j and ±1, as shown in

Figure 2.4.

The construction and properties are similar for the other two subband quater-

nion wavelets based on φh(x)ψh(y) and ψh(x)φh(y) (see the horizontal and vertical

subbands in Figures 1.3(b) and (c), respectively). In summary, in contrast with the

six complex pairs of CWT wavelets (12 functions in total), the QWT sports three

quaternion sets of four QWT wavelets (12 functions in total).

Finally, note that the quaternion wavelet transform is approximately a windowed

Page 26: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

17

+1

+1 +1

+1

(a) Ψh(kx)Ψh(ky)

+j −j

+j −j

(b) Ψg(kx)Ψh(ky)

−j −j

+j+j

(c) Ψh(kx)Ψg(ky)

+1 −1

+1−1

(d) Ψg(kx)Ψg(ky)

Figure 2.4 : Complex Fourier domain relationships among the four quadrature componentsof a quaternion wavelet ψq(x, y) in the diagonal subband.

quaternion Fourier transform (QFT) [4]. In contrast to the QFT in (2.15), the basis

functions for the QWT are the scales and shifts of the quaternion wavelet (ψh(x) −

j1ψg(x))(ψh(y)− j2ψg(y)) plus the wavelets for the other two subbands.

Page 27: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

18

Chapter 3

QWT Properties

Since the QWT is based on combining 1-D CWT functions, it preserves many of

the attractive properties of the CWT. Furthermore, the quaternion organization and

manipulation provide new features not present in either the 2-D DWT or CWT. In

this section, we discuss some of the key properties of the QWT with special emphasis

on the QWT phase.

3.1 Tight Frame

The QWT contains four orthonormal basis sets and thus forms a 4× redundant tight

frame. The components of the QWT wavelets can be organized in matrix form as

G =

ψh(x)ψh(y) ψh(x)φh(y) φh(x)ψh(y)

ψg(x)ψh(y) ψg(x)φh(y) φg(x)ψh(y)

ψh(x)ψg(y) ψh(x)φg(y) φh(x)ψg(y)

ψg(x)ψg(y) ψg(x)φg(y) φg(x)ψg(y)

. (3.1)

More precisely, the frame contains shifted and scaled versions of the functions in G

plus the scaling function φh(x)φh(y) and its 2-D HTs. Each column of the matrix G

contains the four components of the quaternion wavelet corresponding to a subband of

the QWT. For example, the first column contains the quaternion wavelet components

in Figure 1.3(a), that is, the tensor product wavelet ψh(x)ψh(y) and its 2-D partial and

Page 28: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

19

total HTs from Section 2.2. Each row of G contains the wavelet functions necessary

to form one orthonormal basis set; that is the sum-of-squares of the inner product

between an arbitrary function f ∈ L2(R) and each function in one row of G across

all shifts and scales equals ‖f‖2. Since G has four rows, a similar summation for

all functions in G equals 4‖f‖2, thus satisfying the tight-frame property with 4×

redundancy. An important consequence is that the QWT is stably invertible. The

wavelet coefficients corresponding to the projections onto the functions in G can be

computed using a 2-D dual-tree filter bank with linear computational complexity.

3.2 Relationship to the 2-D CWT

There exists a unitary transformation linking the QWT coefficients and the 2-D CWT

coefficients. The components of the CWT can be written in matrix form as

C =1√2

ψh(x)ψh(y) + ψg(x)ψg(y) ψh(x)φh(y) + ψg(x)φg(y) φh(x)ψh(y) + φg(x)ψg(y)

ψg(x)ψh(y) + ψh(x)ψg(y) ψg(x)φh(y) + ψh(x)φg(y) φg(x)ψh(y) + φh(x)ψg(y)

ψg(x)ψh(y)− ψh(x)ψg(y) ψg(x)φh(y)− ψh(x)φg(y) φg(x)ψh(y)− φh(x)ψg(y)

ψh(x)ψh(y)− ψg(x)ψg(y) ψh(x)φh(y)− ψg(x)φg(y) φh(x)ψh(y)− φg(x)ψg(y)

.

(3.2)

The columns of C contain the complex wavelets oriented at ±45◦, ±15◦ and ±75◦,

respectively. We obtain the CWT wavelets by multiplying the matrix G in (3.1) by

Page 29: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

20

the unitary matrix

U =1√2

1 0 0 1

0 1 1 0

0 1 −1 0

1 0 0 −1

. (3.3)

Since G = UC, the CWT also satisfies the tight-frame property with the same 4×

redundancy factor. As we will see in the next section, both the CWT phase and the

QWT phases encode 2-D image feature shifts; however there exists no straightforward

relationship between the phase angles of the QWT and CWT coefficients.

3.3 QWT of Real Signals

For real signals and the DWT, we have the Plancharel theorem that allows us to

interpret wavelet analysis as local Fourier analysis with a multiscale frequency tiling

[38]. We seek a similar interpretation for the QWT as local QFT analysis in the first

quadrant of the QFT domain. This interpretation is not immediate, since quaternion

multiplication in the QFT domain is non-commutative, and the QFT convolution

theorem is not as simple as in the complex case. We first prove the QFT Plancharel

theorem and then provide an inner product formula in the QFT domain.

QFT Plancharel Theorem. Let f(x) and w(x) be two real 2-D signals, and

let F q(u) and W q(u) be their respective QFTs. Then the inner product of f(x) and

w(x) in the space domain can be stated equivalently in the QFT domain as

R2

f(x)w(x)dx =

R2

F q·e(u)γ(W q(u)) + F q

·o(u)α(W q(u))du, (3.4)

where γ(·) and α(·) are the algebra involutions defined in (2.10)–(2.12).

Page 30: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

21

The functions F q·e and F q

·o are respectively the even and odd components of F q

with respect to the spatial coordinate y, as defined in the QFT convolution theorem

[4]

F q·e(u) =

R2

e−j12πux f(x) cos(−2πvy)dx, (3.5)

F q·o(u) =

R2

e−j12πux f(x) j2 sin(−2πvy)dx. (3.6)

We call the right side of (3.4) the QFT inner product between F q(u) and W q(u).

Details of the proof are included in Appendix A.

The QFT Plancharel Theorem holds for two real functions. To better interpret

the QWT, we can compute a formula for the inner product between a real function

and a quaternion analytic function.

Theorem 1 Let f(x) be a real 2-D signal and let wqA(x) be a 2-D quaternion analytic

function whose real component is w(x). Let F q(u) and W q(u) be the QFTs of f(x)

and w(x), respectively. Then the inner product of f(x) and wqA(x) in the space domain

can be expressed as the QFT inner product in a single quadrant of the (u, v)-plane

R2

f(x)wqA(x)dx =

S1

F q·e(u)γ(W q(u)) + F q

·o(u)α(W q(u))du, (3.7)

where S1 = {(u, v) : u ≥ 0, v ≥ 0}.

Note the integration limit of the QFT inner product: S1 in (3.7), as opposed to

R2 in (3.4). Details of the proof for Theorem 1 are included in Appendix B.1.

Page 31: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

22

3.4 QWT Phase Properties

Recall from Section 2.2.1 that each QWT coefficient can be expressed in terms of its

magnitude and phase as q = |q| ej1θ1ej3θ3ej2θ2 . We seek a shift theorem for the QWT

phase that is analogous to that for the CWT. Since the QWT performs a local QFT

analysis, the shift theorem for the QFT [4] holds approximately for the QWT. When

we shift an image from f(x) to f(x − d), the QFT phase undergoes the following

transformation

(θ1(u), θ2(u), θ3(u))→ (θ1(u)− 2πud1, θ2(u)− 2πvd2, θ3(u)) (3.8)

where d = (d1, d2) denotes the shift in the horizontal and vertical directions respec-

tively.

To transfer the shift theorem from the QFT to the QWT, we exploit the fact

that the QWT is approximately a windowed QFT. Thus, each quaternion wavelet

coefficient can be interpreted as the inner product between the real image and a

windowed version of the quaternion exponential (2.16). The scale of analysis controls

the center frequency (u′, v′) of the windowed exponential (that is, the quaternion

wavelet) in the QFT plane. The magnitude and phase of the resulting coefficient are

determined by two factors: the spectral content of the image and the center frequency

(u′, v′) of the wavelet. These two factors determine the frequency parameters (u, v)

we should use in the shift theorem for the QFT (3.8). We call this (u, v) (which

is different from (u′, v′) in general due to the scale of analysis and the quaternion

wavelet) the effective center frequency for the corresponding wavelet coefficient. Note

that the effective center frequency (u, v) should always lie in the first quadrant of the

Page 32: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

23

Figure 3.1 : Effect of varying θ3 on the structure of the corresponding weighted quaternionwavelet from the diagonal subband (left to right): θ3 = −π

4 ,−π8 , 0,

π8 ,

π4 respectively. The

corresponding wavelet changes from textured (θ3 = 0) to oriented (θ3 = ±π4 ).

QFT domain (that is, u, v ≥ 0).

Thus, before we can apply the shift theorem for QWT image analysis, we must

estimate the effective center frequency for each QWT coefficient. Fortunately, for

images having a smooth spectrum over the support of the quaternion wavelet in the

QFT domain, the wavelet’s center frequency (u′, v′) is a close approximation to the

effective center frequency (u, v). Thus, using the relationship between the QWT phase

and image shifts, we can estimate the shift (d1, d2) of one image relative to a second

image from the phase change (∆θ1,∆θ2). Conversely, we can estimate the phase shift

once we know the image shift.

We can interpret the third QWT phase angle θ3 as the relative amplitude of image

energy along two orthogonal directions as in [4]. By adjusting only θ3 of the QWT

ψ(x)ψ(y) subband coefficients, we observe a gradual change in appearance of the

corresponding wavelets from oriented to texture-like and back (see Figure 3.1). This

property could prove useful for the analysis of images with rich textures [4]. As we

describe below in Section 4.1, this third phase also relates to the orientation of a

single edge.

Page 33: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

24

Chapter 4

QWT Applications

The QWT, with its shift-invariance and quaternion phase representation, opens up

new image processing capabilities unattainable with either the 2-D DWT or CWT. As

in the Fourier case, the shift theorem establishes the relationship between the QWT

phases and relative location information in images. In this section, we demonstrate

two applications of the QWT, including edge structure estimation in Section 4.1, and

registration in Section 4.2.

4.1 Edge Geometry Estimation

Edge structures (see Figure 4.1(a)) are the fundamental building blocks of many real-

world images. A typical image can be considered as a piecewise smooth signal in

2-D containing blocks of smooth regions and edge singularities. Wavelet transforms

are very amenable to processing these singularities. Previous efforts to incorporate

the behavior of wavelet coefficients around edge structures into wavelet-based image

models have achieved considerable success in applications such as edge extraction and

image compression [39, 27]. Therefore, understanding the QWT phase response of

edges is key to applying the new tool to real-world images.

First consider an image block containing a single step edge, as in Figure 4.1(a).

Its QWT is the QFT inner product between the quaternion wavelets and the QFT

Page 34: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

25

d

β

(a) single edge model (wedgelet)

β

β

( u , v )

u

v positive quadrant

leakage leakage quadrant

(b) edge QFT spectrum

Figure 4.1 : (a) Parameterization of a single edge in a dyadic image block (wedgelet [3]). (b)QFT spectrum of the edge; shaded squares represent the quaternion wavelets in the vertical,horizontal, and diagonal subbands. The energy of the edge is concentrated along the twodark lines crossing at the origin and is captured by the horizontal subband with spectralcenter at quaternion frequency (u, v). The region bounded by the dashed line demonstratesthe spectral support of the QWT basis “leaking” into the neighboring quadrant.

spectrum of the edge image. For an edge oriented at angle β, any shift (d1, d2) in the

(x, y) directions satisfying the constraint

d1 cosβ + d2 sin β = d (4.1)

corresponds to the same shift as moving the edge by d perpendicularly. The QFT

of this step edge f(x) with orientation and offset (β, d) (see the parameterization in

Figure 4.1(a)) is given by the following expression:

F q(u) = e−j12πud1

(

δ(u cosβ − v sin β)

2(u sinβ + v cos β)(−j1 − j2)

+δ(u cosβ + v sin β)

2(−u sin β + v cosβ)(j1 − j2)

)

e−j22πvd2 . (4.2)

The details of this calculation are included in Appendix B.2.

Thus, the QFT spectrum of the edge lies on two lines through the origin having

orientations 90◦ − β and β − 90◦ in the QFT domain, as shown in Figure 4.1(b).

Page 35: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

26

0 45 900

0.50

600

β

dM

agni

tude

(a) Magnitudes

0 45 900

0.50

600

Mag

nitu

de

βd

(b) Magnitudes

0 45 900

0.50

0.2

0.4

0.6

0.8

1

Est

imat

ion

erro

r

β

d

(c) Estimation errors

0 20 40 60 800

0.50

0.2

0.4

0.6

0.8

1

Est

imat

ion

erro

r

βd

(d) Estimation errors

Figure 4.2 : Magnitude of a QWT coefficient as a function of edge orientation β andoffset d for the (a) horizontal φ(x)ψ(y) subband and (b) diagonal ψ(x)ψ(y) subband.(c), (d) Absolute error for estimating d by the phase angles (θ1,θ2) from the sametwo subbands, relative to the normalized edge length of the dyadic block (= 1 unit).Estimates are accurate in the regions where the coefficient magnitudes are large.

Based on our theoretical analysis in Section 3.3, the effective center frequency of a

quaternion wavelet must lie in the first quadrant and, therefore, can be expressed as

(u, v) = c(| cosβ|, | sinβ|), where c is a positive constant that depends on β and the

location of the spectral tile corresponding to each QWT basis. Using the relations

(u, v) = c(| cosβ|, | sinβ|) in (4.1) and noting that 2πud1 = ∆θ1 and 2πvd2 = ∆θ2,

we obtain the expression

d =∆θ1 ±∆θ2

2πc, (4.3)

where we choose ∆θ1 + ∆θ2 when tan β > 0, and ∆θ1 −∆θ2 when tanβ < 0.

To verify this relationship experimentally, we apply the QWT to the edge image

Page 36: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

27

in Figure 4.1(a) and analyze the QWT magnitudes and phases corresponding to a

32 × 32 sub-block. Figures. 4.2(a) and (b) verify that the coefficient magnitudes are

almost invariant to signal shift d. Note that the magnitude is maximum when the

edge orientation matches the orientation of the basis function (0◦ for the φ(x)ψ(y)

subband, 45◦ for the ψ(x)ψ(y) subband, and 90◦ for the ψ(x)φ(y) subband). We

experimentally study the linear relationships between ∆θ1 ± ∆θ2 and d for various

β in all subbands to obtain the slope estimates 12πc

in (4.3). We obtain an estimate

of c ≈ 0.7 for the ψ(x)φ(y) subband using the vertical edge (β = 0) which gives

the largest QWT magnitude for this subband; similarly for the ψ(x)φ(y) subband.

Using the 45◦ edge, we obtain an estimate of c ≈ 1 for the ψ(x)ψ(y) subband. This

simplification of using the same c in each subband to estimate d is valid because c is

sensitive to the location of the spectral tile of the corresponding QWT basis but not

to β. Now, we can use equation (4.3) to estimate the edge offset d from the QWT

phase using d = 0 as the reference image. Figures 4.2(c) and (d) show the estimation

accuracy for two subbands across various β and d values. In small-magnitude regions,

the phase angles are very sensitive to noise, and thus estimation errors are large. Using

only the phase change of the QWT subband with large magnitude, we can estimate

d with a maximum error of approximately 0.02 relative to the normalized unit edge

length of the dyadic block under analysis.

We also analyze the behavior of the third phase angle θ3 for the same edge block

in Figure 4.1(a). Our theoretical analysis in Appendix B.3 yields the property that

the quaternion wavelet coefficient for a step edge with arbitrary (β, d) always has

θ3 = ±π4, if we assume that all quaternion wavelets have a perfect single-quadrant

Page 37: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

28

QFT spectrum (analytic signal). This special case corresponds to the “singular case”

in the quaternion phase calculation, under which we can uniquely define only the sum

or the difference of θ1 and θ2 [4]. This is consistent with (4.3), which always gives a

unique sum (or difference) of the first two phase angles for a given edge offset d.

However, practical quaternion wavelets will not have a perfect single-quadrant

QFT spectral support, but rather will leak somewhat into neighboring quadrants.

Then, we obtain an alternative formula for the phase angle

θ3 = −1

2arcsin

(

1− ε1 + ε

)

, (4.4)

where ε is a rough measure of the ratio of wavelet or signal energy in the main

quadrant to energy in the leakage quadrant (see Appendix C for the proof). The sum

(or difference) between θ1 and θ2 is the same as in the case without leakage. In other

words, the linear relationship between θ1±∆θ2 and edge offset d in (4.3) is unaffected

by leakage. These theoretical derivations are consistent with our experimental results.

Based on our experimental analysis, we propose a hybrid algorithm to estimate the

edge geometry (β, d) based on the QWT phase (θ1, θ2, θ3) and the magnitude ratios

between the three subbands. Our algorithm is reminiscent of the edge estimation

scheme in [27].

We first obtain the three QWT coefficients corresponding to a dyadic image block

of interest. To estimate the orientation β of the edge block, we choose the QWT sub-

band with the largest magnitude. This subband tells us the approximate orientation

of the edge (±45◦ for diagonal, ±15◦ for horizontal and ±75◦ for vertical), which the

sign of θ3 tells us whether the direction is positive or negative. We can then use either

Page 38: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

29

the magnitude ratios among the three subbands or θ3 to make a more accurate esti-

mation of the orientation. We experimentally analyzed the QWT magnitude ratios

and θ3 corresponding to changing edge orientations β by multiples of 5◦ and, using

standard curve-fitting techniques, developed a simple relationship between these pa-

rameters and the edge orientation. The resulting orientation estimation algorithm

achieves a maximum error of only β ± 3◦ in practice.

To estimate the offset d of the edge block, we only need (θ1, θ2) in the horizontal

and vertical subbands. We choose (θ1, θ2) from the subband with the larger magnitude

and apply the same offset estimation scheme as in [40]. This offset estimation has a

maximum error of approximately ±0.02 relative to the normalized unit edge length

of the dyadic block under analysis. This demonstrates the sub-pixel accuracy of this

estimation algorithm. Such accuracy in our orientation and offset estimation schemes

is obtained with the straight edge model in Figure 4.1(a), but we will see that our

algorithm also works well for real-world images.

According to (4.3), within a 2π-range of ∆θ1±∆θ2, the range of d is limited to an

interval of length 1c≈ 1.43, which ensures that the edge stays within the image block

under analysis. Therefore, in our offset estimation, we need only consider one 2π-

range of ∆θ1 ±∆θ2 and do not need to perform any “phase-unwrapping”. However,

we need to obtain an offset estimate for each polarity of the edge with orientation β

estimated above, say d+ and d−. Then, we use the inner product between the image

block and two wedgelets [3] with the estimated edge parameters (β, d+) and (β, d−)

to determine the correct polarity. We apply the algorithm described above to the

popular “cameraman” image in Figure 4.3.

Page 39: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

30

t

Figure 4.3 : Local edge geometry estimation using the QWT. (a) Several edgy regions fromthe “cameraman” image are shown on the left; (b)–(e) on the right are edge estimates fromthe corresponding QWT coefficients. The upper row shows the original image region, thelower row shows a wedgelet (see Figure 4.1(a)) having the edge parameter estimates (β, d).(No attempt is made to capture the texture within the block.)

Our results demonstrate the close relationship between edge geometry and QWT

magnitude and phases, in particular, the encoding of edge location in the QWT phases

(θ1,θ2) and the encoding of edge orientation in the QWT magnitude and third phase

θ3.

4.2 QWT-based Image Disparity Estimation

The relative location information encoded in the phase angles of the QWT is particu-

larly useful for image registration [11]. The image registration problem arises in many

applications such as in video processing to estimate motion between successive frames,

in time-lapse seismic imaging to study changes in a reservoir over time, in medical

imaging to monitor a patient’s body, in super-resolution, etc. In these applications,

we often need to estimate the local shifts of a target image (or several) compared

to a reference image. We can then use this disparity information to build a warp-

ing function to align the images. In this section, we describe an efficient multiscale

Page 40: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

31

scheme to estimate local shifts in an image based on the QWT phase. Compared to

other recent phase-based image registration approaches [4, 41, 42, 5], we demonstrate

the simplicity of our approach and the effectiveness of the QWT representation for

encoding location information.

Recall that the QWT phase property states that a shift (d1, d2) in an image changes

the QWT phase from (θ1, θ2, θ3) to (θ1− 2πud1, θ2− 2πvd2, θ3). Thus, for each QWT

coefficient, if we know (u, v), the effective center frequency of the local image region

analyzed by the corresponding QWT basis functions, we can estimate the local image

shifts (d1, d2) from the phase differences.

Suppose we are given a reference image A(x, y) and a target image B(x, y) for

which we wish to estimate the local shifts, First, by manually translating the reference

image A(x, y) by known small amounts both horizontally and vertically, we obtain

two images A(x, y) and A(x − d1, y − d2). After computing QWTs of A(x, y) and

A(x − d1, y − d2), we can use the phase differences (∆θ1,∆θ2) between the QWT

coefficients to obtain estimates for the effective spectral center (u, v) for each dyadic

block across all scales as u = ∆θ1

2πd1and v = ∆θ2

2πd2. The range of QWT phase angles

limits our estimates (u, v) to [− 12D, 1

2D) and [− 1

4D, 1

4D) for horizontal and vertical

shifts, respectively, where D is the length of one side of the dyadic block corresponding

to each coefficient.

Once we know the effective center frequency (u, v) for each QWT coefficient, we

can estimate the local image shifts by measuring the difference between the QWT

phase corresponding to the same local blocks in the images A(x, y) and B(x, y). The

main challenge here is the phase wrap-around due to the limited range of phase angles;

Page 41: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

32

that is, each observed phase difference can be mapped to more than one disparity

estimate. Specifically, for QWT phase differences (∆θ1,∆θ2) between the reference

and target images, we can express the possible image shifts of each dyadic block as

d1 =∆θ1 + π(2n+ k)

2πu, d2 =

∆θ2 +mπ

2πv, (4.5)

where n,m ∈ Z and k ∈ {0, 1}. Depending on m, k is chosen such that it equals 0

when m is even and equals 1 when m is odd. This special wrap-around effect in (4.5)

is due to the limited range in θ1 and θ2 (to [−π, π) and [−π2, π

2) respectively) and the

periodicity of intrinsically 2-D signals [4].

In our multiscale disparity estimation technique, we use coarse scale shift estimates

to unwrap the phase in the finer scales. If we assume the true image shift is small

compared to the size of dyadic squares at the coarsest scale L, then we can set

m = n = k = 0 in (4.5) at this scale (no phase wrap-around) and obtain correct

estimates for d1 and d2. Effectively, this assumption of no phase wrap-around at the

coarsest scale limits the maximum image shift that we can estimate correctly. Once

we have shift estimates at scale L, for each dyadic block at scale ` = L − 1, we

estimate the shifts in the following way:

1. bilinearly interpolate estimates from previous scale(s) to obtain predicted esti-

mates (dp1, d

p2),

2. substitute the phase differences in (4.5) and determine (n, k,m) so that (d1, d2)

are closest to (dp1, d

p2),

3. remove unreliable (d1, d2),

Page 42: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

33

4. repeat Steps 1–3 for the finer scales ` = L− 2, L− 3, . . .

In Step 3, we use a similar reliability measure as in the confidence mask [4] to

threshold unreliable phase and offset estimates. We also threshold based on the

magnitude of the QWT coefficients. We iterate the above process until a fine enough

scale (` = 2), since estimates become unreliable at this scale and below. The QWT

coefficients for the small dyadic blocks have small magnitudes and their phase angles

are very sensitive to noise.

We improve this basic iterative algorithm by combining estimates across subbands

and scales. First, with proper interpolation, we can average over estimates from all

scales containing the same image block. Second, we can average estimates from the

three QWT subbands for the same block to yield more accurate estimates, but we

need to discard some unreliable subband estimates (for example, horizontal disparity

d1 in the horizontal subband and d2 in the vertical subband). We incorporate these

subband/scale averaging steps into each iteration of Steps 1–4.∗

Figure 4.4 shows the result of our QWT phase-based disparity estimation scheme

for two selected frames in a rotating “Rubik’s cube” video sequence [4]. The arrows

indicate both the directions and magnitudes of the local shifts, with the magnitudes

stretched for better visibility. We can clearly see the rotation of the Rubik’s cube on

the circular stage. Our experiments have indicated that the algorithm that averages

over both subbands and scales gives the most robust shift estimates. This finding is

consistent with our hypothesis, since the averaging procedure effectively utilizes all

∗Matlab software is available at http://www.dsp.rice.edu/software/qwt.shtml.

Page 43: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

34

(a) reference image (b) target image (c) disparity estimates

Figure 4.4 : Multiscale QWT phase-based disparity estimation results. (a), (b) Referenceand target images from the “Rubik’s cube” image sequence [4]. (c) Disparity estimatesbetween two images in the sequence, shown as arrows overlaid on top of the reference image(zoomed in for better visualization).

shift information extracted by the QWT.

A performance measure for disparity/motion estimation algorithms, commonly

used in video compression, is the Peak Signal-to-Noise Ratio (PSNR) between the

motion-compensated image C(x) and the target image B(x), given by

10 log10

(

(255)2N∑

x(B(x)− C(x))2)

)

, (4.6)

where N is the number of image pixels. We first compute the motion vectors as in

Figure 4.4(c). Then, we obtain the motion-compensated image C(x) by shifting and

interpolating every image block in the reference image A(x) and calculate the PSNR

between B(x) and C(x).

For comparison, we choose a block-matching technique known as exhaustive search

(ES), which is very computationally demanding but has the best performance among

all general block-matching techniques. We also choose a phase correlation technique

known as Gradient Correlation (GC), which has been shown to have better PSNR

Page 44: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

35

performance than other recent phase correlation methods [5]. Figure 4.5 shows the

comparison results for three image sequences, the “Rubik” and “Taxi” sequences

commonly used in the optical flow literature and the MPEG test sequence “Mobcal”

used in [5]. Figure 4.5(a) demonstrates the superior performance of our QWT phase-

based algorithm over other algorithms for the “Rubik” sequence, which has piecewise-

smooth image frames and a smooth underlying disparity flow. For the two other

sequences, which contain some discontinuities in their underlying flows, the QWT

phase-based algorithm also has comparable performance (see table in Figure 4.5(b)).

Since the multiscale averaging step in our algorithm tends to smooth out the estimated

flow, our algorithm is not expected to perform as well for discontinuous motions fields

of rigid objects moving past each other.

Additional advantages of our QWT-based algorithm include linear computational

efficiency and sub-pixel estimation accuracy. Thus, for an N -pixel image, our O(N)

algorithm is more efficient than the O(N logN) FFT-based GC and is significantly

faster than ES, which can take up to O(N2) with the search parameter on the order of

N . General block-matching techniques such as ES can also only decipher disparities

in an integer number of pixels. However, as with other phase-based algorithms, our

QWT-based algorithm can achieve estimation accuracy up to a fractional number of

pixels. Our experiments compute motion vectors only for 8 × 8 blocks. Although it

is more common to use 32 × 32 blocks for video compression, we choose a smaller

block size because, for image registration, a high-density disparity map is needed to

register elaborate image features (for example, patterns on the Rubik’s cube and the

rotation stage; see Figure 4.4).

Page 45: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

36

0 5 10 15 2035

36

37

38

39

40

41

42

Frame Number

PS

NR

QWTGradCorrES

(a) PSNR vs. frame number

Rubik Taxi Mobcal

QWT 39.4 36.2 25.5

GC 36.7 36.6 26.9

ES 37.2 37.0 26.3

(b) average PSNR performance (in dB)

Figure 4.5 : Comparison of multiscale QWT phase-based disparity estimation with twomotion estimation algorithms, Gradient Correlation (GC) [5] and exhaustive search (ES).Performance measure is the PSNR (in dB) between the motion-compensated image and thetarget image of three test image sequences (“Rubik”, “Taxi” and “Mobcal”). (a) Frame-by-frame PSNR performance comparison in the “Rubik” sequence. (b) Table of averagePSNR performance (over all frames) for each test sequence. Multiscale QWT phase-basedmethod demonstrates the best performance among the three test algorithms for the “Rubik”sequence, and shows comparable performance to the other algorithms for the “Taxi” and“Mobcal” sequences.

Besides phase correlation, there exist other phase-based algorithms for disparity

estimation and image registration [4, 41, 42]. These approaches use phase as a feature

map Φ(x) where the phase function Φ maps 2-D x, y-coordinates to phase angles.

They assume the phase function to stay constant upon a shift from the reference

image to the target image; that is, Φ1(x) = Φ2(x+d) where Φ1 is the phase function

for the reference image and Φ2 for the target image. Then, the disparity estimation

problem is simplified to calculating the optical flow for these phase functions. In

contrast, our algorithm is entirely based on the dual-tree QWT and its shift theorem,

without using any optical flow assumptions [43].

Our approach is similar to the QFT disparity estimation algorithm of [4]in its

Page 46: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

37

use of quaternion phase angles. However, the QFT approach requires the design

of a special filter to compute the phase derivative function in advance, while our

approach only needs to first estimate the local frequencies (u, v). Our approach is

more computationally efficient (order O(N) as opposed to O(N logN) for an N -pixel

image.). Provided a continuous underlying disparity flow, our multiscale scheme

yields a denser and more accurate disparity map, even for smooth regions within an

image.

Kingsbury et al. have also developed a multiscale displacement estimation algo-

rithm using the 2-D CWT [41, 42]. Their approach combines information from all six

CWT subbands in an optimization framework based on the optical flow assumptions.

In addition to disparity estimation, it simultaneously registers the target image to the

reference image. In comparison, both the QWT and CWT methods are multiscale

and wavelet-based and are thus, in general, best for smooth underlying disparity flow.

However, our algorithm is much simpler and easy to use because it does not involve

the tuning of many input parameters for the iterative optimization procedures as

in the CWT algorithm. Although our method only estimates local disparities with-

out warping the image, we can apply another standard warping procedure to easily

register the two images from the estimated disparities.

Thanks to the ability of the QWT to encode location information in absolute

horizontal/vertical coordinates, we can easily combine the QWT subband estimates to

yield more accurate flow estimation results. Combining subband location information

in the 2-D CWT is more complicated, since each subband encodes the disparities by

complex phase angles in a reference frame different (rotated) from other subbands.

Page 47: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

38

Based on our experimental results and comparison of the design of our flow estimation

algorithm with previous approaches, the QWT demonstrates its ability to efficiently

represent, encode and process location information in images.

Page 48: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

39

Chapter 5

Conclusions

We have introduced a new 2-D multiscale wavelet representation, the QWT, that is

particularly efficient for coherent processing of relative location information in images.

This tight-frame transform generalizes complex wavelets to higher dimensions and

inspires new processing and analysis methods for wavelet phase.

Our development of the QWT is based on an alternative definition of the 2-D

HT and 2-D analytic signal and on quaternion algebra. The resulting quaternion

wavelets have three phase angles; two of them encode phase shifts in an absolute hor-

izontal/vertical coordinate system while the third encodes textual information. The

QWT’s approximate shift theorem enables efficient and easy-to-use analysis of the

phase behavior around edge regions. We have developed a novel multiscale phase-

based disparity estimation scheme. Through efficient combination of disparity esti-

mates across scale and wavelet subbands, our algorithm clearly demonstrates the ad-

vantages of coherent processing in this new QWT domain. Inherited from its complex

counterpart, the QWT also features near shift-invariance and linear computational

complexity through its dual-tree implementation.

Beyond 2-D, the generalization of the Hilbert transform to n-D signals using hy-

percomplex numbers can be used to develop higher dimensional wavelet transforms

suitable for signals containing low-dimensional manifold structures [44]. The QWT

Page 49: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

40

developed here could play an interesting role in the analysis of (n − 2)-D manifold

singularities in n-D space. This efficient hypercomplex wavelet representation could

bring us new ways to solve high-dimensional signal compression and processing prob-

lems.

Page 50: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

41

Appendix A

QFT Plancharel Theorem

Proof: Given the 2-D real signal w(x) and its QFT W q(u), the following QFT rela-

tionships hold:

w(x, y) ←→ W q(u, v) = γ(W q(−u,−v)), (A.1)

w(−x,−y) ←→ γ(W q(u, v)), (A.2)

w(x′ − x, y′ − y) ←→ e−j12πux′

γ(W q(u, v))e−j22πvy′

, (A.3)

The quaternion Hermitian symmetry of the QFT of w(x) gives the relationship in

(A.1) (see Theorem 2.2 and Theorem 2.11 in [4]), which also implies (A.2). By first

expressing w(x′ − x, y′ − y) as w(−(x− x′),−(y − y′)), (A.3) is obtained from (A.2)

by the QFT shift theorem (see Theorem 2.8 in [4]). Starting from the expression in

the space domain, we have∫

R2

f(x′)w(x′)d2x′ =

R2

f(x′)

(∫

R2

ej12πu′x′

γ(W q(−u′))ej22πv′y′

d2u′

)

d2x′

=

R2

f(x′)

(∫

R2

e−j12πux′

γ(W q(u))e−j22πvy′

d2u

)

d2x′

=

R2

f(x′)

R2

(∫

R2

e−j12πuxw(x′ − x)e−j22πvyd2x

)

d2ud2x′

=

R2

R2

e−j12πux

(∫

R2

f(x′)w(x′ − x)d2x′

)

e−j22πvyd2xd2u

=

R2

F q·e(u)γ(W q(u)) + F q

·o(u)γ(β(W q(u)))d2u

=

R2

F q·e(u)γ(W q(u)) + F q

·o(u)α(W q(u))d2u. (A.4)

Step 5 uses the QFT convolution theorem (Theorem 2.6) in [4].

Page 51: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

42

Appendix B

Continuous QWT of 2-D edge signal

B.1 Proof of Theorem 1

According to Definition 2, the 2-D quaternion analytic signal with real component

w(x) can be expressed in the space domain as:

wqA(x) = w(x) + j1wHi1(x) + j2wHi2(x) + j3wHi(x). (B.1)

For any real function g(x), we can express its CFT and QFT respectively as

Gq(u) = A(u) + j1B(u) + j2C(u) + j3D(u) and (B.2)

G(u) = (A(u)−D(u)) + j(B(u) + C(u)), (B.3)

where j =√−1. We can see this equivalence from the CFT and the QFT formula

which basically involve the same sine and cosine basis elements. Also, notice that

this proof uses u = (u, v) to denote the general frequency domain (both the CFT and

QFT domain) because both Fourier transforms are discussed simultaneously.

Let F (u) and W (u) be the CFT of f(x) and w(x) respectively. We can verify that

the integrand from the CFT inner product G(u) = F (u)W ∗(u) is equivalent to the

integrand from the QFT inner product Gq(u) = F q·e(u)γ(W q(u)) + F q

·o(u)α(W q(u)).

These two integrands from the CFT and the QFT inner products are equivalent

because they both come from the real 2-D signal g(x) = (f ∗w−)(x) described in the

proof of the QFT Plancharel Theorem.

Page 52: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

43

Denote S2 = {(u, v) : u ≤ 0, v ≥ 0} as the second quadrant, and S3 and S4

similarly in an anti-clockwise fashion. Using the CFT Plancharel theorem, the inner

products between f(x) and each of the four components in wqA(x) can be expressed in

terms of the CFT inner product between F (u) and W (u), i.e., G(u) = F (u)W ∗(u)),

multiplied by ±j or ±1 in the appropriate quadrant, as in (B.4)-(B.7):

p =

S1

G(u) +

S2

G(u) +

S3

G(u) +

S4

G(u), (B.4)

q =

S1

−jG(u) +

S2

jG(u) +

S3

jG(u) +

S4

−jG(u), (B.5)

r =

S1

−jG(u) +

S2

−jG(u) +

S3

jG(u) +

S4

jG(u), (B.6)

s =

S1

−G(u) +

S2

G(u) +

S3

−G(u) +

S4

G(u). (B.7)

As a result, the inner product between f(x) and the quaternion analytic function

wqA(x) can be expressed, using quaternion algebra, as p+ j1q+ j2r+ j3s. Substituting

(B.3) into (B.4)-(B.7), then using conjugate symmetry of G(u), yields an expression

involving only quaternions

p + j1q + j2r + j3s

= 2

S1

(A(u) + j1B(u) + j2C(u) + j3D(u))− (A(u)− j1B(u)− j2C(u)− j3D(u))j3

2

S2

(A(u)− j1B(u) + j2C(u) + j3D(u)) + (A(u)− j1B(u)− j2C(u) + j3D(u))j3

= 2

S1

(Gq(u)− α(Gq(u))j3) + 2

S2

(β(Gq(u)) + γ(Gq(u))j3), (B.8)

where α(·), β(·) and γ(·) are the algebra involutions defined as in (2.10)–(2.12). By

the quaternion Hermite symmetry of Gq(u), further simplification obtains

p+ j1q + j2r + j3s = 4

S1

Gq(u)d2u

= 4

S1

F q·e(u)γ(W q(u)) + F q

·o(u)α(W q(u))d2u, (B.9)

Page 53: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

44

which is an integral in the single quadrant S1. Equation (B.9) establishes the QFT

inner product of any quaternion basis function wqA(x) and a general real 2-D signal

f(x) as an integral in S1 alone.

B.2 Calculation of (4.2)

First, express the step edge as a 2-D separable function (a constant function along

the x-direction multiplied by a 1-D step function along the y-direction). The QFT

of such a function is −j2 δ(u)v

. Then, apply the QFT affine theorem (Theorem 2.12 in

[4]) with the transformation matrix involving rotation β and offset d.

B.3 Calculation of QWT Phase Angles for a Step Edge

This calculation combines the results from (3.7) and (4.2). Consider the special

case when the edge signal f(x) has zero offset (d = 0). From equation (4.2), its

QFT is F q(u) = (−j1 − j2)H(u) in S1 where H(u) is the component involving the

δ-sheet in (4.2). Let the QFT of the real component of a QWT basis, w(x), be

W q(u) = a(u) + j1b(u) + j2c(u) + j3d(u). From equation (3.7),

S1

Gq(u)d2u =

S1

F q·e(u)γ(W q(u)) + F q

·o(u)α(W q(u))d2u

=

S1

H(u)(b(u) + c(u))d2u + j1

S1

H(u)(a(u)− d(u)))d2u

+j2

S1

H(u)(a(u)− d(u)))d2u− j3∫

S1

H(u)(b(u) + c(u))d2u

= A1 + j1B1 + j2B1 − j3A1, (B.10)

Page 54: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

45

where A1 and B1 are the simplified expressions involving a(u), b(u), etc. After

normalizing (B.10) with its magnitude 2(A21 +B2

1), compute the third phase angle as

θ3 = −1

2arcsin

(

A21 +B2

1

2(A21 +B2

1)

)

= −π4. (B.11)

Depending on the orientation β of the 2-D edge signal f(x), its QFT F q(u) can be

(j1 − j2)H(u) in S1, which gives θ3 = π4.

Moreover, this special quaternion in (B.10) with θ3 = −π4

is in the singular case;

as described in [4], its other phase angles (θ1, θ2) are non-unique but the sum θ1 + θ2

is unique with the following expression

θ1 + θ2 =1

2arcsin

(

2A1B1

A21 +B2

1

)

, (B.12)

whose value largely depends on the QFT spectrum of the basis w(x) and on the edge

orientation and offset (β,d).

Page 55: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

46

Appendix C

Theoretical Analysis of Leakage Effect

In practice, in order to have finite-length basis functions, the 2-D Hilbert transforms

of the real QWT basis component w(x) is only approximately satisfied. Therefore,

the non-ideal quaternion basis function wqA(x) does not have a perfect single-quadrant

spectrum in S1, but has a certain amount of “leakage” into other quadrants.

Consider the inner product of the edge signal f(x) and wqA(x) in both the main

quadrant (S1) and the leakage quadrant (S2).∗ The inner product in the main quad-

rant is given by equation (B.10) while the inner product in the leakage quadrant can

be written as

S2

Gq(u)d2u =

S2

F q·e(u)γ(W q(u)) + F q

·o(u)α(W q(u))d2u

=

S2

H(u)(b(u)− c(u))d2u + j1

S2

H(u)(a(u) + d(u)))d2u

−j2∫

S2

H(u)(a(u) + d(u)))d2u + j3

S2

H(u)(b(u)− c(u))d2u

= A2(u) + j1B2(u)− j2B2(u) + j3A2(u). (C.1)

Again, A2 and B2 are the simplified expressions involving a(u), b(u), etc.

Therefore, combining equation (B.10) and (C.1) gives the inner product between

∗Leakage quadrant can be either S2 or S3 depending on the spectral support of the basis element

w(x).

Page 56: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

47

the non-ideal basis and the edge signal

S1

Gq(u)d2u +

S2

Gq(u)d2u = (A1(u) + A2(u)) + j1(B1(u) +B2(u))

+j2(B1(u)−B2(u)) + j3(−A1(u) + A2(u)),

whose magnitude is 2(A21 +A2

2 +B21 +B2

2). Its third phase angle can be expressed as

θ3 = −1

2arcsin

(

2[(A21 +B2

1)− (A22 +B2

2)]

2(A21 + A2

2 +B21 +B2

2)

)

= −1

2arcsin

(

1− ε1 + ε

)

, (C.2)

where ε =A2

2+B2

2

A2

1+B2

1

is an expression for the ratio of basis or signal energy in the main

quadrant to energy in the leakage quadrant. The ratio ε is always less than unity

because the basis function has higher energy in the main quadrant than in the leakage

quadrant. When ε is close to unity, the phase angle is largely affected by leakage.

When ε is close to zero, leakage has little effect and θ3 ≈ −π4

as in the ideal case.

This relationship between ε and θ3 agrees with our experimental observation of varying

edge orientation β. When β ≈ 0◦, the QFT of the edge lies on the u-axis; thus the

signal energy ratio ε ≈ 1 (same contribution from both main and leakage quadrant)

and θ3 ≈ 0. When β ≈ 45◦, there is little overlap between the QFT of the edge with

the basis spectrum in the leakage quadrant; thus ε ≈ 0 and θ3 ≈ −π4.

Let us turn our attention to the other two phase angles (θ1,θ2). Since θ3 is not

necessarily −π4

for the QFT of the edge signal, i.e., the singular case no longer holds,

there exists unique expressions for both θ1 and θ2 as follows:

θ1 =1

2arctan

(

A1B2 + A2B1

A1A2 − B1B2

)

,

θ2 =1

2arctan

(

A2B1 − A1B2

A1A2 +B1B2

)

.

Page 57: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

48

However, when combining the two expressions above by trigonometric identity

(sum of two arctangents), the sum of θ1 and θ2 is the same as in the ideal case (without

leakage) as in (B.12), which only involves terms A1 and B1 from the main quadrant.

This result has the important implication that, in spite of leakage, the QWT shift

theorem still holds, i.e., θ1 ± θ2 varies linearly with edge offset d. Equation (4.3),

relating the QWT phase of an edge θ1 ± θ2 to edge offset d, is still valid. Besides,

θ1 − θ2 =1

2arcsin

(−2A2B2

A22 +B2

2

)

(C.3)

only involves terms A2 and B2 from the leakage quadrant. Similar to its counterpart

in (B.12), θ1− θ2 largely depends on the edge orientation and offset (β,d) and on the

QFT spectral support of w(x), but only in the leakage quadrant. However, it is much

harder to estimate the behavior of θ1 − θ2 with varying offset d since w(x) only has

a small amount of energy in the leakage quadrant.

Page 58: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

49

Bibliography

[1] A. F. Abdelnour and I. W. Selesnick, “Design of 2-band orthogonal near-symmetric CQF,” IEEE Int. Conf. on Acoustics, Speech, and Signal Processing,pp. 3693–3696, May 2001.

[2] N. G. Kingsbury, “A dual-tree complex wavelet transform with improved or-thogonality and symmetry properties,” in Proc. of IEEE Int. Conf. on ImageProcessing, vol. 2, (Vancouver), pp. 375–378, September 2000.

[3] M. Wakin, J. Romberg, H. Choi, and R. Baraniuk, “Rate-Distortion OptimizedImage Compression using Wedgelets,” in IEEE International Conference on Im-age Processing, September 2002.

[4] T. Bulow, Hypercomplex Spectral Signal Representations for the Processing andAnalysis of Images. PhD thesis, Christian Albrechts University, Kiel, Germany,1999.

[5] V. Argyriou and t. Vlachos, “Using gradient correlation for sub-pixel motionestimation of video sequences,” IEEE Int. Conf. on Acoustics, Speech, and SignalProcessing, pp. 329–332, May 2004.

[6] J. Shapiro, “Embedded Image Coding using Zerotrees of Wavelet Coefficients,”IEEE Trans. Signal Processing, vol. 41, pp. 3445–3462, April 1993.

[7] M. Crouse, R. Nowak, and R. Baraniuk, “Wavelet-Based Statistical Signal Pro-cessing Using Hidden Markov Models,” IEEE Trans. Signal Processing, vol. 46,pp. 886–902, April 1998.

[8] H. Choi, J. Romberg, R. Baraniuk, and N. Kingsbury, “Hidden Markov tree mod-eling of Complex Wavelet Transforms,” IEEE Int. Conf. on Acoustics, Speech,and Signal Processing, pp. 133–136, June 2000.

[9] L. Sendur and I. Selesnick, “Bivariate shrinkage functions for wavelet-baseddenoising exploiting interscale dependency,” IEEE Trans. Signal Processing,vol. 50, pp. 2744–2756, Nov. 2002.

[10] J. Weng, “Image matching using the windowed Fourier phase,” Int. J. of Com-puer Vision, vol. 11, pp. 211–236, April 1993.

[11] B. Zitova and J. Flusser, “Image registration Methods: A Survey,” Image andVision Computing, vol. 21, pp. 977–1000, 2003.

Page 59: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

50

[12] D. J. Fleet and A. D. Jepson, “Computation of component image velocity fromlocal phase information,” Int. J. of Compuer Vision, vol. 5, no. 1, pp. 77–104,1990.

[13] A. Oppenheim and J. Lim, “The importance of phase in signals,” in Proceedingsof the IEEE, vol. 69, pp. 529–541, 1981.

[14] R. E. Blahut, Theory of Remote Image Formation, ch. 7. Cambridge UniversityPress, 2004.

[15] D. L. Donoho, “De-noising by soft-thresholding,” IEEE Trans. Info. Theory,vol. 41, no. 3, pp. 613–627, 1995.

[16] J. M. A. Grossmann, R. Kronland-Martinet, “Reading and understanding con-tinuous wavelet transforms,” in Wavelets: Time-Frequency methods and phasespace, pp. 2–20, Berlin: Springer-Verlag, 1989.

[17] J. M. Lina and M. Mayrand, “Complex Daubechies wavelets,” J. App. Harm.Analysis, vol. 2, pp. 219–229, 1995.

[18] B. Belzer, J.-M. Lina, and J. Villasenor, “Complex, linear-phase filters for effi-cient image coding,” IEEE Trans. on Signal Processing, vol. 43, pp. 2425–2427,Oct. 1995.

[19] H. F. Ates and M. T. Orchard, “A nonlinear image representation in waveletdomain using complex signals with single quadrant spectrum,” in Proc. AsilomarConf. Signals, Systems, and Computers, 2003.

[20] R. van Spaendonck, T. Blu, R. Baraniuk, and M. Vetterli, “Orthogonal Hilberttransform filter banks and wavelets,” in Proc. IEEE Int. Conf. Acoust., Speech,Signal Processing (ICASSP), vol. 6, pp. 505–508, April 6-10 2003.

[21] M. Wakin, M. Orchard, R. G. Baraniuk, and V. Chandrasekaran, “Phase andmagnitude perceptual sensitivities in nonredundant complex wavelet representa-tions,” in Proc. Asilomar Conf. Signals, Systems, and Computers, 2003.

[22] F. Fernandes, M. Wakin, and R. G. Baraniuk, “Non-redundant, linear-phase,semi-orthogonal, directional complex wavelets,” in Proc. IEEE Int. Conf.Acoust., Speech, Signal Processing (ICASSP), (Montreal), May 2004.

[23] N. G. Kingsbury, “Complex wavelets for shift invariant analysis and filtering ofsignals,” J. App. Harm. Analysis, vol. 10, pp. 234–253, May 2001.

[24] J. Romberg, H. Choi, and R. Baraniuk, “Multiscale edge grammars for complexwavelet transforms,” in Proc. of Intl. Conf. on Image Processing, (Thessaloniki,Greece), pp. 614–617, Oct. 2001.

[25] S. L. Hahn, “Multidimensional complex signals with single-orthant spectra,” inProceedings of the IEEE, vol. 80, pp. 1287–1300, August 1992.

Page 60: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

51

[26] I. W. Selensnick and K. Y. Li, “Video denoising using 2-D and 3-D dual-treecomplex wavelet transforms,” in Proc. of SPIE Wavelets X, vol. 76, (San Diego,CA), August 4-8 2003.

[27] J. Romberg, M. Wakin, H. Choi, and R. Baraniuk, “A geometric hidden Markovtree wavelet model,” in Proc. of SPIE Wavelets X, (San Diego, CA), August2003.

[28] J. Romberg, H. Choi, R. Baraniuk, and N. Kingsbury, “A hidden Markov treemodel for the Complex Wavelet Transform,” tech. rep., Rice University, ECE,September 2002.

[29] L. Sendur and I. W. Selesnick, “Bivariate shrinkage functions for wavelet-baseddenoising exploiting interscale dependency,” IEEE Trans. Signal Processing,vol. 50, pp. 2744–2756, November 2002.

[30] T. Bulow and G. Sommer, “A novel approach to the 2d analytic signal,” in CAIP,(Ljubljana, Slovenia), 1999.

[31] H. Ates, Modeling location information for wavelet image coding. PhD thesis,Princeton University, EE, 2003.

[32] G. Hua, “Noncoherent image denoising,” Master’s thesis, Rice University, ECE,Houston, Texas, 2005.

[33] M. Felsberg and G. Sommer, “The monogenic signal,” IEEE Trans. Signal Pro-cessing, vol. 49, pp. 3136–3144, December 2001.

[34] M. Felsberg, “Disparity from monogenic phase,” in 24. DAGM Symposium Mus-tererkennung, Zurich (L. v. Gool, ed.), vol. 2449 of Lecture Notes in ComputerScience, pp. 248–256, Springer, Heidelberg, 2002.

[35] M. Vetterli and J. Kovacevic, Wavelets and Subband Coding. Englewood Cliffs,NJ: Prentice Hall, 1995.

[36] I. Selesnick, “Signal Denoising using 1-D Double-Density Complex WaveletTransform.” http://taco.poly.edu/selesi/DoubleSoftware/index.html.

[37] I. L. Kantor and A. S. Solodovnikov, Hypercomplex Numbers. Springer-Verlag,1989.

[38] C. S. Burrus, R. A. Gopinath, and H. Guo, Introduction to Wavelets and WaveletTransforms, ch. 7. Prentice Hall, 1998.

[39] M. Wakin, J. Romberg, H. Choi, and R. Baraniuk, “Wavelet-domain approxi-mation and compression of piecewise smooth images,” to appear in IEEE Trans.on Image Processing, May 2005.

Page 61: Coherent Multiscale Image Processing using Quaternion Waveletswailam/research/myThesis.pdf · 2006-04-26 · ABSTRACT Coherent Multiscale Image Processing using Quaternion Wavelets

52

[40] W. L. Chan, H. Choi, and R. Baraniuk, “Quaternion wavelets for image anal-ysis and processing,” in Proc. of IEEE Int. Conf. on Image Processing, vol. 5,(Singapore), pp. 3057–3060, Oct. 2004.

[41] M. Hemmendroff, M. Anderson, T. Kronander, and H. Knutsson, “Phase-basedmultidimensional volume registration,” IEEE Trans. Medical Imaging, vol. 21,pp. 1536–1543, December 2002.

[42] N. Kingsbury, “Multi-scale displacement estimation and registration for 2-d and3-d datasets.” http://www.eng.cam.ac.uk/˜ngk, July 2005.

[43] B. K. Horn and B. G. Schunck, “Determining optical flow,” Artificial Intelligence,vol. 17, pp. 185–203, 1981.

[44] W. L. Chan, H. Choi, and R. Baraniuk, “Directional hypercomplex wavelets formultidimensional signal analysis and processing,” IEEE Int. Conf. on Acoustics,Speech, and Signal Processing, pp. 996–999, May 2004.