better data assimilation through gradient descent

36
Better Data Assimilation through Gradient Descent Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of Economics London Mathematical Society - EPSRC Durham Symposium Mathematics of Data Assimilation

Upload: gallia

Post on 07-Feb-2016

36 views

Category:

Documents


0 download

DESCRIPTION

London Mathematical Society - EPSRC Durham Symposium Mathematics of Data Assimilation. Better Data Assimilation through Gradient Descent. Leonard A. Smith, Kevin Judd and Hailiang Du Centre for the Analysis of Time Series London School of Economics. Outline. Perfect model scenario (PMS) - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Better Data Assimilation through Gradient Descent

Better Data Assimilation through Gradient Descent

Leonard A. Smith, Kevin Judd and Hailiang Du

Centre for the Analysis of Time Series London School of Economics

London Mathematical Society - EPSRC Durham Symposium

Mathematics of Data Assimilation

Page 2: Better Data Assimilation through Gradient Descent

Outline

Perfect model scenario (PMS)

GD method GD is NOT 4DVAR Results compared with Ensemble KF

Imperfect model scenario (IPMS)

GD method with stopping criteria GD is NOT WC4DVAR Results compared with Ensemble KF

Conclusion & Further discussion

Page 3: Better Data Assimilation through Gradient Descent

Experiment Design (PMS)

Page 4: Better Data Assimilation through Gradient Descent

Ensemble techniques

Generate ensemble directly, e.g. Particle Filter, Ensemble Kalman Filter

Generate ensemble from perturbations of a reference trajectory, e.g. SVD on 4DVAR

Gradient Descent (GD) Method

K Judd & LA Smith (2001) Indistinguishable States I: The Perfect Model Scenario, Physica D 151: 125-141.

Page 5: Better Data Assimilation through Gradient Descent

Gradient Descent (Shadowing Filter)

Page 6: Better Data Assimilation through Gradient Descent

Gradient Descent (Shadowing Filter)

5s

0s

4s

)( 5sF

Page 7: Better Data Assimilation through Gradient Descent

Gradient Descent (Shadowing Filter)

Page 8: Better Data Assimilation through Gradient Descent

Gradient Descent (Shadowing Filter)

Page 9: Better Data Assimilation through Gradient Descent

Gradient Descent (Shadowing Filter)

Page 10: Better Data Assimilation through Gradient Descent

GD is NOT 4DVAR

Difference in cost function

Noise model assumption

Observational noise model 4DVAR cost function

GD cost function not depend on noise model

Assimilation window

4DVAR dilemma: difficulties of locating the global minima with long assimilation

window

losing information of model dynamics and observations without long window

Page 11: Better Data Assimilation through Gradient Descent

Methodology

Page 12: Better Data Assimilation through Gradient Descent

Form ensemble

Obs

t=0

Reference trajectory

GD result

Page 13: Better Data Assimilation through Gradient Descent

Form ensemble

t=0Candidate trajectories

Sample the local space

Perturb observations and run GD

Page 14: Better Data Assimilation through Gradient Descent

Form ensemble

t=0Ensemble trajectory

Draw ensemble members according to likelihood

Page 15: Better Data Assimilation through Gradient Descent

Form ensemble

Obs

t=0Ensemble trajectory

Page 16: Better Data Assimilation through Gradient Descent

Ensemble members in the state space

Compare ensemble members generated by Gradient Descent method and Ensemble Adjustment Kalman Filter method in the state space.

Low dimensional example to visualize, higher dimensional results later.

Page 17: Better Data Assimilation through Gradient Descent

Ikeda Map, Std of observational noise 0.05, 512 ensemble

members

Page 18: Better Data Assimilation through Gradient Descent

Evaluate ensemble via Ignorance

The Ignorance Score is defined by:

where Y is the verification.

Ikeda Map and Lorenz96 System, the noise model is N(0, 0.4) and

N(0, 0.05) respectively. Lower and Upper are the 90 percent

bootstrap resampling bounds of Ignorance score

Ensemble->p(.)

Page 19: Better Data Assimilation through Gradient Descent

Imperfect Model Scenario

Page 20: Better Data Assimilation through Gradient Descent

Toy model-system pairs

Ikeda system:

Imperfect model is obtained by using the truncated polynomial, i.e.

Page 21: Better Data Assimilation through Gradient Descent

Toy model-system pairs

Lorenz96 system:

Imperfect model:

Page 22: Better Data Assimilation through Gradient Descent

Insight of Gradient Descent

Define the implied noise to be

and the imperfection error to be

Page 23: Better Data Assimilation through Gradient Descent

Insight of Gradient Descent

5s0s

4s

)( 5sf

w0

Page 24: Better Data Assimilation through Gradient Descent

Insight of Gradient Descent

w

Page 25: Better Data Assimilation through Gradient Descent

Insight of Gradient Descent

0w

Page 26: Better Data Assimilation through Gradient Descent

Statistics of the pseudo-orbit as a function of the number of Gradient Descent iterations for both higher dimension Lorenz96 system-model pair experiment (left) and low dimension Ikeda system-model pair experiment (right).

Implied noise

Imperfection error

Distance from

the “truth”

Page 27: Better Data Assimilation through Gradient Descent

GD with stopping criteria

GD minimization with “intermediate” runs produces more consistent pseudo-orbits

Certain criteria need to be defined in advance to decide when to stop or how to tune the number of iterations.

The stopping criteria can be built by testing the consistency between implied noise and the noise model

or by minimizing other relevant utility function

Page 28: Better Data Assimilation through Gradient Descent

Imperfection error vs model error

Model error Imperfection error

Obs Noise level: 0.01

Not accessible!

Page 29: Better Data Assimilation through Gradient Descent

Imperfection error vs model error

Imperfection error

Obs Noise level: 0.002 Obs Noise level: 0.05

Page 30: Better Data Assimilation through Gradient Descent

GD vs WC4DVAR

WC4DVAR Model error

assumption

GDModel error

estimates

Page 31: Better Data Assimilation through Gradient Descent

Forming ensemble

Apply the GD method on perturbed observations.

Apply the GD method on perturbed pseudo-orbit.

Apply the GD method on the results of other data assimilation methods. Particle filter?

Page 32: Better Data Assimilation through Gradient Descent

Imperfect model experiment: Ikeda system-model pair, Std of

observational noise 0.05, 1024 EnKF ensemble members, 64 GD ensemble members

Page 33: Better Data Assimilation through Gradient Descent

Evaluate ensemble via Ignorance

The Ignorance Score is defined by:

where Y is the verification.

Ikeda system-model pair and Lorenz96 system-model pair, the noise model is N(0, 0.5) and N(0, 0.05) respectively. Lower and Upper are the 90 percent bootstrap resampling bounds of Ignorance score

Systems Ignorance Lower Upper

EnKF GD EnKF GD EnKF GD

Ikeda -2.67 -3.62 -2.77 -3.70 -2.52 -3.55

Lorenz96

-3.52 -4.13 -3.60 -4.18 -3.39 -4.08

Page 34: Better Data Assimilation through Gradient Descent

Conclusion Methodology of applying GD for data assimilation in

PMS is demonstrated outperforms the 4DVAR and Ensemble Kalman filter methods

Outside PMS, mmethodology of applying GD for data assimilation with a stopping criteria is introduced and shown to outperform the WC4DVAR and Ensemble Kalman filter methods.

Applying the GD method with a stopping criteria also produces informative estimation of model error.

No data assimilation without dynamics.

Page 35: Better Data Assimilation through Gradient Descent

Thank you!

[email protected]

Centre for the Analysis of Time Series:http://www2.lse.ac.uk/CATS/home.aspx

Page 36: Better Data Assimilation through Gradient Descent