The prototype carbon inverse problem:
Sponsored Links
This presentation is the property of its rightful owner.
1 / 47

David Baker NCAR / Terrestrial Sciences Section 23 July 2006 PowerPoint PPT Presentation

  • Uploaded on
  • Presentation posted in: General

The prototype carbon inverse problem: estimation of regional CO 2 sources and sinks from global atmospheric [CO 2 ] measurements. David Baker NCAR / Terrestrial Sciences Section 23 July 2006. Description of the source/sink problem CO 2 measurement availability thru time

Download Presentation

David Baker NCAR / Terrestrial Sciences Section 23 July 2006

An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -

Presentation Transcript

The prototype carbon inverse problem: estimation of regional CO2 sources and sinks from global atmospheric [CO2] measurements

David Baker

NCAR / Terrestrial Sciences Section

23 July 2006

Description of the source/sink problem

CO2 measurement availability thru time

Method determined by computational issues

The hierarchy of inverse methods used:

Batch least-squares

The (full) Kalman filter/smoother

The use of the adjoint: what it can (and can’t) do

The ensemble Kalman filter/smoother

Variational data assimilation


TransCom3 Regions (22) and Measurement Sites (78)

Discretize problem: solve for monthly fluxes for 22 regions

Measurement equations: Hx=z

x – regional CO2 fluxes, monthly avg

z – CO2 measurements, monthly avg

H – transport matrix; columns are Green’s functions of a 1-month pulse of unitized CO2 flux from each region calculated using the transport model

Application: CO2 flux estimation

Form of the Batch Measurement Equations




Transport basis functions


Optimal fluxes found by minimizing



Batch least-squares

Calculation of Interannual Variability (IAV) -- Europe (Region #11)


Monthly fluxes

13 transport models

Deseasonalized fluxes

IAV (subtract mean)

IAV w/ errors

Total flux IAV (land + ocean)


Land flux IAV



Flux IAV


Flux IAV [PgC/yr],

11 land regions

Flux IAV [PgC/yr],

11 ocean regions

Incorrect assumptions about space/time patterns across large space/time blocks can cause large biases in inversions

To fix this, solve at as fine a resolution possible, and transform final estimate to coarser scales post-inversion, if necessary

Sounds good, but poorly-known correlations must still be used in the fine-scale inversions to constrain the results

Resolution of robust results ultimately set by measurement density and precision

“Representativeness” or “Discretization” Error

Transport model runs to generate H:

22 regions x 16 years x 12 months x 36 months = 152 K tracer-months (if using real winds)

22 x 12 x 36 = 9.5 K tracer-months (climatological winds)

Matrix inversion computations: O(N3)

N= 22 regions x 16 years x 12 months = 4.4 K

Matrix storage: O (N*M) --- 66 MB

M = 78 sites x 16 years x 12 months = 15 K

Computational considerations

New measurements augmenting the weekly flasks:

More continuous analyzers

More aircraft flights

More towers (both tall and eddy-flux)

Ships/planes of opportunity

Low-cost in situ analyzers

CO2-sondes, tethered balloons, etc.

Satellite-based column-integrated CO2, CO2 profiles

Higher frequency, more sensitive to continental air

Most importantly: MORE DATA, better coverage

Will permit more detail to be estimated

Present  Future

Eddy-flux tower sites

~22 regions, monthly, ~60 sites, 1987-present: N

~50 regions, weekly, ~100 sites, c. 2000: 10·N

~200 regions, daily?, ~200 continuous sites, ~2005: 300·N

~60000 regions (1°x1°), hourly, ~250 meas/hour, from satellites launched ~2008: 2,000,000·N

Upshot: batch least squares not feasible for much longer

Growth of measurement density and likely solution resolution

1) Break span into smaller blocks (Bousquet, et al 1999) -- 1-2 year spin-up time limits this, due to required overlaps

2) Solve in the (smaller) measurement space -- i.e., O(M3) (Peylin, et al 2005)

3) Use the adjoint to the transport model to efficiently generate H basis functions in time-reverse mode (Kaminski, et al, 1999; Roedenbeck, et al 2005)

2) and 3) allow one to efficiently solve for many fluxes when the number of measurements is small(er), but both break down when both the number of measurements and fluxes are large

Three approaches to handling these computational limitations:

Transport model runs to generate H:

22 regions x 16 years x 12 months x 36 months = 152 K tracer-months (if using real winds)

22 x 12 x 36 = 9.5 K tracer-months (climatological winds)

Matrix inversion computations: O(N3)

N= 22 regions x 16 years x 12 months = 4.4 K

Matrix storage: O (N*M) --- 66 MB

M = 78 sites x 16 years x 12 months = 15 K

Three key ideas to get around these limitations:

Use sequential estimators instead of batch inversion

Use indirect (iterative) solvers instead of direct ones

Solve for only an approximate covariance, not the full-rank one

Computational considerations

Kalman Filter/Smoother



Kalman Filter/Smoother



Measurement update step at time k:

Dynamic propagation step from times k to k+1:

Put multiple months of flux in state vector xk, method becomes effectively a fixed-lag Kalman smoother

Kalman Filter Equations

For retrospective analyses, a 2-sided smoother

gives more accurate estimates than a 1-sided filter.

The 4-D Var method is 2-sided, like a smoother.

(Gelb, 1974)

Full KF only a partial fix due to still-large P matrices

Ensemble KF (EnKF): full covariance matrix replaced by approximation from an ensemble

EnKF can be thought of as a reduced-rank square root KF with each column of the reduced-rank P1/2 being given by a state deviation (dx) from an ensemble of independently propagated and updated states x

Matrix operations involving P in the basic KF equations replaced with dot products of state deviations (dx) summed over the ensemble

See Peters, et al (2005) for implementation

Traditional (full-rank) Kalman filter ensemble Kalman filter

Ensemble KF still a relatively new method -- details still being worked out:

Proper method for introducing measurement and dynamical errors into ensemble

Proper means for handling time/space correlations in measurements

Proper way to apply it to time-dependent CO2 fluxes

Variational data assimilation is more developed


Wide use in numerical weather prediction

Based on a minimization -- can always be made to converge

Con*: -- Requires an adjoint model

*Since ensemble smoothers need an adjoint, too, for true retrospective analyses (those for which a fixed-lag smoother won’t work well), this is not necessarily a con in those cases

Ensemble KF vs. Variational DA

4-D Var Data Assimilation Method

  • Find optimal fluxes u and initial CO2 field xo to minimize

  • subject to the dynamical constraint

  • where

    • x are state variables (CO2 concentrations),

    • h is a vector of measurement operators

    • z are the observations,

    • R is the covariance matrix for z,

    • uo is an a priori estimate of the fluxes,

    • Puo is the covariance matrix for uo,

    • xo is an a priori estimate of the initial concentrations,

    • Pxo is the covariance matrix for xo,

    • is the transition matrix (representing the transport model), and

      G distributes the current fluxes into the bottom layer of the model

4-D Var Data Assimilation Method

Adjoin the dynamical constraints to the cost function using Lagrange multipliers 

SettingF/xi = 0 gives an equation for i, the adjoint of xi:

The adjoints to the control variables are given byF/uiandF/xoas:

The gradient vector is given by F/uiandF/xoo, so the optimal u and x0 may then be found using one’s favorite descent method. I have been using the BFGS method, since it conveniently gives an estimate of the leading terms in the covariance matrix.

4-D Var Iterative Optimization Procedure














































= 

Minimum of cost function J

Variable metric method -- accumulates 2nd order information (approximate covariance matrix) as minimization proceeds

Search direction: sk= Hkk

Hk approximates [2J/u2]-1= Pu

Precondition with H0=Pu°

H never stored, but reconstructed from p and q as needed

BFGS Minimization Approach

Allows for dynamical errors

Add dynamical mismatches ito J as

Note that

Then solve for i from iand add these to the propagated state

“Soft” Dynamical Constraint

Parameterized Chemical Transport Model (PCTM; Kawa, et al, 2005)

Driven by reanalyzed met fields from NASA/Goddard’s GEOS-DAS3 scheme

Lin-Rood finite volume advection scheme

Vertical mixing: diffusion plus a simple cloud convection scheme

Exact adjoint for linear advection case

Basic resolution 2x 2.5, 25  layers, t30 min, with ability to reduce resolution to

4x 5, t60 min

6x 10, t120 min < we’ll use this in the example

12x 15, t180 min

Measurements binned at t resolution

Atmospheric Transport Model

Monday: Test performance of 4DVar method in a simulation framework, with dense data (6°x10°, lowest level of model, every t, 1 ppm (1) error)

Case 1 -- no data noise added, no prior

Case 2 -- w/ data noise added, no prior

Case 3 -- w/ data noise added, w/ prior

Case 4 -- no data noise added, w/ prior

Tuesday: with Case 3 above, do OSSEs for

Case 5 -- dense data, but OCO column average

Case 6 -- OCO ground track and column average

Case 7 -- extended version of current network

Case 8 -- current in situ network

Possibilities for projects: examine the importance of

Data coverage and accuracy vs. targeted flux resolution

Prior error pattern and correlation structures

Measurement correlations in time/space

Errors in the setup assumptions (“mistuning”)

Effect of biases in the measurements

Data Assimilation Experiments

Home directory: /project/projectdirs/m598/dfb/4DVar_Example/scripts/case1/BFGS

Work directories: /scratch/scratchdirs/dfb/case1/work_fwd & /scratch/scratchdirs/dfb/case1/work_adj

Submit batch job by typing ‘llsubmit runBFGS2_LL’ while in /project/projectdirs/m598/dfb/4DVar_Example/scripts/case1/BFGS/

This executes the main driver script, found in BFGSdriver4d.F, in same directory, which controls setting up all the files and running FWD and ADJ inside the minimization loop

The scripts that execute the FWD and ADJ runs of the model are found in …/4DVar_Example/scripts/case1/, named run.co2.fvdas_bf_fwd(adj)_trupri997_hourly

Check progress of job by typing ‘llqs’

Jobs currently set up to do a 1 year-long run (360 days), solving for the fluxes in 5-day long chunks, at 6.4x10 resolution, with t =2 hours

How to run the 4D-Var code

In /scratch/scratchdirs/dfb/case1/work_fwd/costfuncval_history/temp

Column 1 -- measurement part of cost function

Column 2 -- flux prior part of cost function

Column 4 -- total cost function value

Column 10 -- weighted mismatch from true flux

Column 11 -- unweighted mismatch from true flux

Columns 4, 10, and 11 ought to be decreasing as the run proceeds

Columns X and Y give the iteration count and 1-D search count

How to monitor job while running

A results file in netCDF format written to: /scratch/scratchdirs/dfb/case1/work_fwd/

sftp this to (rename it, so that you don’t overwrite another group’s file)

Pull up an X-window to davinci and ssh -X

On davinci, ‘module load ncview’

Then ‘ncview’

Click on a field to look at it

Hint: set ‘Range’ to +/- 2e-8 for most fields

How to view detailed results

The code for the FWD and ADJ model is in ../4DVar_Example/src_fwd_varres and src_adj_varres

Measurement files are located in /scratch/scratchdirs/dfb/case1/meas

Two files controlling the tightness of the prior and whether or not noise is added to the measurements are /scratch/scratchdirs/dfb/case1/work_fwd/ferror and /scratch/scratchdirs/dfb/case1/work_fwd/measnoise_on.dat

Other code details

2-hourly measurements in the lowest model level at 6.4x 10, 1 ppm error (1)

Iterate 30 descent steps, 1-year-long run, starts 1/1

4 cases

Case 1 -- No measurement noise added, no prior

Case 2 -- Add measurement noise added, no prior

Case 3 -- Add noise, and apply a prior

Case 4 -- No noise, but apply a prior

Designed to test the method and understand the impact of data errors and the usefulness of the prior

Case 3 is the most realistic and will be used to do OSSEs for several possible future networks for Tuesday’s problem set

Monday’s Experiment

Use Case 3 from above to test more-realistic measurement networks:

Current in situ network

Extended version of current network

OCO satellite

Hourly 6.4x 10 column measurements

Essentially an “OSSE” (observing system simulation experiment) -- tells you how well your instrument should do in constraining the fluxes. Only gives the random part of the error, not biases

Tuesday’s Experiment

Across 1 day

OCO Groundtrack,

Jan 1st

(Boxes at 6x 10)

2 days

5 days

  • Login