miis 2015 -...

87
Isao Yamada Tokyo Institute of Technology MIIS 2015 July 5-6, 2015, Xidian University, Xi'an China Two Spices of Convex Optimization For Certain Inverse Problems

Upload: others

Post on 09-Jul-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Isao YamadaTokyo Institute of Technology

MIIS 2015 July 5-6, 2015, Xidian University, Xi'an China

Two Spices of Convex Optimization For Certain Inverse Problems

A touch of spice brings miracle flavors to our great meal !

Wasabi Sushi

I am going to introduce Two Mathematical Spices :

1. Hybrid Steepest Descent Method

2. Algebraic Phase Unwrapping

which will bring Deeper & Wider applications of

Convex Optimization to Inverse Problems !

Good Luck !

Gauss(1777-1855)

Fourier (1768-1830)

Kalman(1930-)

Information Sciences, e.g. Signal Processing,

are growing with Mathematics !

Wiener(1894-1964)

Shannon(1916-2001)

P.L. Lions (1956-)

1994 Fields medal

Terence Tao (1975-)

2006 Fields medal

Newton(1643-1727)

von Neumann(1903-1955)

Euclid(Mid.4th century BC -Mid.3rd century BC)

Extraction of Valuable Information from Observed Data !

has been a most successful strategy !

A Goal of Information Sciences

MinimizeSubject to

Optimization

We like BEST Solution !

Measurement(Degraded image)

Signal Processing

is searching for BEST

Restored Image

All available knowledge

Physical Constraints, Statistical Property, Perceptual Characteristic

e.g.

e.t.c.

Signal Analysis

Fourier, Wavelet, Curvelet, etc

Most information can be expressed in terms of a Point ( Stone) in a Vector Space !Example of

Optimization is similar to Mountain Climbing !

Minimize

Maximize

Vector Space

Find a Stone on the Highest place !

Many Signal Processing Researchersare TRYING to

How can we find BEST solutions ?

Step 1. Design an Optimization Problem

Step 2. Translate (P1) into a computationally tractable problem (P2)

Minimize

Subject to(P1)

Use all available knowledge

Step 3. Apply a suitable algorithm to (P2)

Best in what sense ?

(P1)

Collection of Computationally Tracable Problems ?

Translation Based on Variety of

Mathematics

(P2)

Step 2

in in

Main Stage for Step 2

David Hilbert

: Real vector space equipped with Real Hilbert Space

Inner Product

Norm

J. von Neumann

NOTE: Completeness always holds if

converges to if

You can define flexibly !

is Closed in

Proper

Lower Semi-continuous

Convex

Convex Optimization defined in Real Hilbert Space

Minimize

where

Let's focus on

How can we find BEST solutions ?

Step 1. Design an Optimization Problem

Step 2. Translate (P1) into a computationally tractable problem (P2)

Minimize(P1)Use all available knowledge

Step 3. Apply a suitable algorithm to (P2)

Best in what sense ?

where is nonexpansive i.e.

A Common Strategy in Step 2 for Convex Optimization

Translation Based on

Convex Analysis &

Monotone Operator Theory

(P2)Step 2

Minimize Find

(P1)

: a certain Nonexpansive Operator defined with proximity operators /gradients

: a certain Nonlinear Operators

where

Step 2

More precisely, the solution sets of convex optimization problems

can often be expressed as

Example (Step 2) : Convex Feasibility Poblem

Metric projection onto a closed convex set

is

For its best approximation in a closed convex set exists uniquely.

NOTE: Computational issue depends on shape of set !

Nonexpansive

s.t.

A Common Principle in Step 3

Suppose isNonexpansive

Thenfor any

Note:This fact has been a key of many algorithms.

Theorem(Krasnosel’skii-Mann, e.g. [Groetsch'72])

POCS *Bregman, Soviet Math. Dokl (1965)*Gubin, Polyak, Raik, USSR Comp.Math. Math Physics (1967)

Example (Step 3) : Convex Feasibility Poblem

can be seen as an example of K-M Alg !This is because

Most successful convex optimization algorithms : Proximal Splitting Algorithms

(Forward backward splitting/Primal-dual splitting/ADMM etc)

can also be seen as K-M Alg (Step 3)applied to

: computable Nonexpansive Operator defined with proximity operators /gradients

: certain computable Nonlinear Operators

where

Step 2

Metric projection

Proximity operator

Forward-Backward splitting operator

Douglas-Rachford splitting operator (ADMM: Dual Variant [Fortin-Glowinski '83])

Primal-Dual splitting Operator

Applicability of Nonexpansive Operators is incleasing with advances in Convex Optimization

Great !

Proper Lower semicontinuous Convex function.

Proximity Operator (J.J.Moreau '62)

Good News (Proximable functions)

Example 1

Example 2

Example 3

Let

FACT (Convex envelope of )

Largest convex lower bound !

Example 1

Convex envelope of a function

All mappings reviewed so far belong to Averaged Nonexpansive Operators !

: nonexpansive,

is

: nonexpansive map s.t.

In this case, K-M Alg. guarantees :

Then

Let

def

29

Theorem ([Ogura-Yamada'02 / Combettes-Yamada'14])

- averaged - averaged

Let

Then

A Key for Broad Applications of K-M Alg.

behind the strategies in the Modern Convex Optimization algs,

(a ) Nonexpansive Operators have incredible power to describe, as their fixed point set, the closed convex sets of great value in Hilbert Space

(b) K-M alg allows us to access only one unspecial point in the fixed point set, but does not reveal the other points in the gold mine !

suggests

The fixed point expression

Digital Camerascan describe a CAT in 3 dimensional space in terms of a digital data !

Digital camera can focus on the face of cat by zooming in the image !

Mathematicians can describe a closed convex set in Hilbert space in terms of nonexpansive operators !

Mathematicians can focus on the intersection of multiple closed convex sets in Hilbert space by using nonexpansive operators !

K-M alg allows us to access only one unspecial point in the valuable closed convex set but does not reveal other points in the gold mine !

remain a mystery !

This situation of K-M algorithm is very similar to Fortune Cookie !

All look same but they are very different in many aspects !

Actually VERY DIFFERENT

K-M alg allows us to access only one unspecial point in the valuable closed convex set but does not reveal other points in the gold mine !

Results are VERY DIFFERENTin many aspects

Can we choose best one without crunching all cookies ?

The answer is YES ! Hybrid steepest descent method can accomplish this mission !

K-M algorithm

Hybrid Steepest Descent Method

The 1st Spice is here !

Hybrid Steepest Descent Method(Yamada et al,'96, Deutsch-Yamada'98, Yamada'01, Yamada-Ogura'04 etc)

can minimize over

where

Lipschitz ContinuousQuasi-Nonexpansive mapping

Slowly decreasing

Smooth Convex Function

1. This is a generalization of [P.L. Lions'77].2. This can select a very best solution among all fixed points !

More Spicy Principle in Step 3

If is nonexpansive,

is an extension of simplest case:

[Halpern'67/Reich'74/ Lions'77/Wittmann'92/Combettes'95/Bauschke'96]

can be seen as a remarkable extension of Projected Gradient Method [Goldstein '64].

Moreover, alternative form of HSDM

Hybrid Steepest Descent Method (HSDM) [Yamada'01]

Nonexpansive with bounded

Smooth Convex function, s.t.

Theorem (nonstrictly convex, [Ogura-Yamada'03]) Suppose

satisfies

where

Then

is desired to be minimized additionally.

Suppose

Minimize

Subject to

NOTE: is not in usually !

Hierarchical Optimization

Find a point inStandard Optimization

Find a Hardest Stone on the summit of a mountain !

(= Hierarchical Optimization)

: Bounded Linear Operator

Primal Dual Splitting

Primal Problem

Suppose

Dual Problem

(Condat '13/ Combettes, Vu, ... etc)

where

Under a certain mild condition,

Fenchel-Rockafellar Conjugate of

Example [Condat '13]

Define with

and by

where

Then

by

Krasnosel'skii-Mann Theorem guarantees :

SupposePrimal Dual Splitting is applicable to general case.

Masking Operatordiscarding 40%Pixels randomly

Gaussian Noise

Application to Image Restoration

Step 1

Total VariationData fidelity

[Yamada-Ono Eusipco'13]

Step 1a

Total VariationData fidelity

Step 1bPrimal Dual

splitting operator

Apply Krasnosel'skii-Mann algorithm orStep 2

Hybrid Steepest Descent Method to

Primal Dual splitting operator

+Krasnosel'skii-Mann

Primal Dual splitting operator

+Hybrid Steepest Descent Method

Promoting Sparsity of Curvelet Coefficients

Algebra helps Convex Optimization recover 2D Phase Surface from its Wrapped Samples

You do not have enough time ! Move on the 2nd Spice ! Otherwise …...

Smooth phase recovery from its wrapped samples (2D-Phase Unwrapping) has been challenging.

available only at simply connected

Smooth surfaceWrapped Samples

(Observable) (Unknown : must be estimated)

At each

Challenge in 2D Phase Unwrapping

is available.

Only

Find a smooth satisfying

2D-Phase Unwrapping

2D-Phase Unwrapping is a key for estimating Crucial physical information such as :

Degree of Magnetic Field Inhomogeneity in water / fat separation in MRI

Surface Topography as measured by - Interferometric Synthetic Aperture Radar (InSAR) - Interferometric Synthetic Aperture Sonar (InSAS)

Profile of Mechanical Parts by X-ray.

Simple Observation for Convex Optimization [Kitahara and Yamada'13]

where

is available at each

is available at each

is available at each

Theorem 1 [Kitahara-Yamada'13]

Let satisfy

satisfying

Then for an arbitrarily fixed

There exists a unique

NOTE

Poincaré (1854-1912)

This is scalar potential of a vector field !

Proposed approach to 2D-Phase Unwrapping

Step 2 (Numerical Integration)

where and

Step1 (Convex Optimization)

subject to

Compute

as an estimate of the unwrapped phase

Suppress Rapid Local Change of

On Step 1(Convex Optimization: Example) Find

All functions with continuous fourth derivatives differentiated no more than twice w.r.t. each variable [Pretlova '76].

Minimizes

Subject to

Bicubic SplineA Generalized Hermite-Birkhoff Interpolation Theorem

 

can be decomposed into a finite sum of integrals of the following type:

where

On Step 2 (Numerical Integration)

Polynomials

0 0.2 0.4 0.6 0.8 1-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

t

A(t

) [r

ad/]

Can we really accomplish the Numerical Integration ?

Indeed, we face difficulty as seen in a typical example of the unwrapped phase !

The answer is YES !Extend Sturm's idea on the use of my GCD algorithm !

Algebraic Phase Unwrapping [Yamada et al '98], [Yamada-Bose '02], [Yamada-Oguchi'11], [Kitahara-Yamada'15]

Sturm(1803-1855)

The 2nd Spice is here !

Theorem 2 (Algebraic Phase Unwrapping)

whereNumber of Sign Changes in Generated by

Algorithm 1

A Modification of the Euclidean Algorithm for Phase Unwrapping

A Modification of the Euclidean Algorithm for Phase Unwrapping

Polynomial Division may cause Certain Numerical instabilities !

Alternative Expression of Algebraic Phase Unwrapping

We need only

but do not need coefficients of for computation of

Key Observation

If the polynomial sequence is Regular, i.e,

Then

Note: More general relation is also given [Kitahara & Yamada '13].

Theorem 3 (Alternative Expression,[Kitahara & Yamada '13] )

We can express without using polynomial division.

Habitch '48

Collins '67

Brown-Traub'71

Algebraic Phase

Unwrapping

Thales(c. 624-546 BC) Euclid

(c. 323-283 BC) Fourier (1768-1830) Sturm

(1803-1855)

Poincaré (1854-1912)

A self-portrait of Hokusai (1760-1849)

at the age of 82

Numerical Example

2.5 km

30 m

14 m

Proposed method (STEP 1) Bivariate Spline Interpolation

Bicubic Interpolating Spline as a minimizer of

in some functional space

Estimate of Unwrapped Phase by Proposed Algebraic Recovery

Perfectly smooth phase surface is finally

obtained after the long journey !

Proposed method (STEP 2) Algebraic Phase Unwrapping

Topographic Mapping from Interferometric SAR

Elementary Geometry tells us

a way to translate phase into hight !

Topographic Mapping from Unwrapped Phase by the Proposed Method

Wow !Beautiful Mt.Fuji

Mathematical Developments of Hybrid Steepest Descent Method

● I. Yamada, The hybrid steepest descent method for the variational inequality problem over the intersection of fixed point sets of nonexpansive mappings, Studies in Computational Mathematics, vol. 8, 2001,pp. 473-504.

● I. Yamada, N. Ogura, N. Shirakawa, A numerically robust hybrid steepest descent method for the convexly constrained generalized inverse problems, Contemporary Mathematics 313, pp.269-305, AMS,2002.

● I. Yamada, N.Ogura, Hybrid steepest descent method for variational inequality problem over the fixed point set of certain quasi-nonexpansive mappings, Numer Funct Anal Optim, vol.25, pp. 619-655, 2004.

● C. Chidume, Geometric Properties of Banach Spaces and Nonlinear Iterations (Chapter 7:Hybrid steepest descent method for variational inequalities), Lecture Notes in Mathematics, vol.1965, Springer 2009.

● P.E. Mainge, Extension of the Hybrid steepest descent method to a class of variational inequalities and fixed point problems with nonself-mappings, Numer Funct Anal Optim, vol.29, pp.820-834, 2008.

Many more …..

Applications of Hybrid Steepest Descent Method

to Hierarchical Convex Optimizations in Signal Processing

1. I. Yamada, M. Yukawa, M. Yamagishi, “Minimizing the Moreau envelope of nonsmooth convex functions over the fixed point set of certain quasi-nonexpansive mappings," In: Fixed Point Algorithms for Inverse Problems in Science and Engineering (H.H.Bauschke, et al., eds.) , pp. 345-390, Springer-Verlag, 2011.

2. S. Ono, I. Yamada, “Hierarchical Convex Optimization with Primal-Dual Splitting," IEEE Trans. Signal Process., vol.63, no.2, pp.373-388, 2014.

3. I. Yamada and K. Oguchi, “High-resolution estimation of the directions-of-arrival distribution by algebraic phase unwrapping algorithms,” Multidimensional Systems and Signal Processing, vol.22, no.1-3, pp.191-211, 2011.

2. I. Yamada and N.K.Bose,“Algebraic phase unwrapping and zero distribution of polynomial for continuous-time systems,'' IEEE Trans. Circuits and Systems 1: Fundamental Theory and Applications, vol.49, no.3, pp.298-304, 2002.

1. I. Yamada, K.Kurosawa, H.Hasegawa and K.Sakaniwa, “Algebraic multidimensional phase unwrapping and zero distribution of complex polynomials – Characterization of multivariate stable polynomials," IEEE Trans. Signal Processing, vol.46, no.6, pp.1639-1664, 1998.

Algebraic Phase Unwrapping

4. D. Kitahara and I. Yamada, “Algebraic phase unwrapping along the real axis - Extensions and Stabilizations,” Multidimensional Systems and Signal Processing, DOI 10.1007/s11045-013-0234-7 (43 pages, April, 2013).

5. D.Kitahara and I. Yamada, Algebraic phase unwrapping over collection of triangles based on two-dimensional spline smoothing, ICASSP 2014.

6. D.Kitahara and I. Yamada, A virtual resampling technique for algebraic two-dimensional phase unwrapping, ICASSP 2015.

Minimize

Subject to

Generalized Hermite-Birkhoff Interpolation Problem

Why Piecewise Polynomial ?

is given as a Polynomial Spline of degree 2m-1 !

Multidimensional extensions & applications

([Atteia '68], [Laurent '69], [Ritter '69])

[Pretlova '76], [Wahba'90],  [Unser'99], [Lai & Schumaker'07] etc

On Step 1 (contd)

Example 1 (Expression of Unwrapped Phase )

Algorithm 1 generates

Example 1 (Contd. )

Functional Data Analysis Process

1. Select Basis Set (Functional space behind Data )

2. Select Smoothing Operator (To surppress rapid local change of model function ) – e.g., differential equation– equivalent to a Bayesian prior over coefficients to estimate

3. Estimate coefficients to optimize some objective function

4. Model criticism, residual plots, etc.

5. Hypothesis testing