80
Craig H. Bishop Naval Research Laboratory, Monterey JCSDA Summer Colloquium July 2012 Santa Fe, NM Background Error Covariance Modeling 1

Background Error Covariance Modeling

  • Upload
    robbin

  • View
    56

  • Download
    1

Embed Size (px)

DESCRIPTION

Background Error Covariance Modeling. Craig H. Bishop Naval Research Laboratory, Monterey JCSDA Summer Colloquium July 2012 Santa Fe, NM. Overview. Strategies for flow dependent error covariance modeling Ensemble covariances and the need for localization - PowerPoint PPT Presentation

Citation preview

Page 1: Background Error Covariance Modeling

Craig H. BishopNaval Research Laboratory, Monterey

JCSDA Summer ColloquiumJuly 2012

Santa Fe, NM

Background Error Covariance Modeling

1

Page 2: Background Error Covariance Modeling

Overview• Strategies for flow dependent error covariance modeling

– Ensemble covariances and the need for localization– Non-adaptive ensemble covariance localization– Adaptive ensemble covariance localization– Scale dependent adaptive localization

• Flexibility of localization scheme allowed by DA algorithm.

• Hidden error covariances and the need for a linear combination of static-climatological error covariances and flow dependent covariances.

• New theory for deriving weights for Hybrid covariance model from (innovation, ensemble-variance) pairs

• Conclusions

Page 3: Background Error Covariance Modeling

Ensembles give flow dependent, but noisy correlations

Stable flow error correlations

km

km

Unstable flow error correlations

Small Ensembles and Spurious Correlations

Page 4: Background Error Covariance Modeling

Fixed localization functions limit adaptivity

Most ensemble DA techniquesreduce noise by multiplying ensemble correlation function by fixedlocalization function (green line).

Resulting correlations (blue line) are too thin when true correlation is broad and too noisy when true correlation is thin.

Stable flow error correlations

km

km

Unstable flow error correlations

Fixed localization

Fixed localization

Small Ensembles and Spurious Correlations

Page 5: Background Error Covariance Modeling

Fixed localization functions limit ensemble-based 4D DA

Stable flow error correlations

km

km

Unstable flow error correlations

• Current ensemble localization functions poorly represent propagating error correlations.

t = 0

t = 0

Small Ensembles and Spurious Correlations

Page 6: Background Error Covariance Modeling

Stable flow error correlations

km

km

Unstable flow error correlations

• Current ensemble localization functions poorly represent propagating error correlations.

t = 0

t = 0

t = 1

t = 1

Small Ensembles and Spurious Correlations

Fixed localization functions limit ensemble-based 4D DA

Page 7: Background Error Covariance Modeling

• Green line now gives an example of one of the adaptive localization functions that are the subject of this talk.

Small Ensembles and Spurious Correlations

Want localization to adapt to width and propagation of true correlation

Stable flow error correlations

Unstable flow error correlations

t = 0

t = 0

t = 1

t = 1

km

km

Page 8: Background Error Covariance Modeling

Ideally, localization would yield the mean or mode of the posterior based on informed guesses of the prior and likelihood distributions.

Bayesian perspective onlocalization

Given a true correlation , we can generate the likelihood pdf

| of -member sample correlations from theory.

Can do this for any where 1 1.

But to get the desired pdf

tij

K t KL ij ij ij

t tij ij

ij

c

c c K c

c c

c

1

1

| from the single realization of

obtained from an ensemble, we need Bayes' theorem

||

|

where gives pdf of true correlations base

tij

tij

t K Kij ij

K t tL ij ij ijt K

ij ij cK t t t

L ij ij ij ijc

tij

c c

c c cc c

c c c d c

c

d on information. prior

|t Kij ijc c

Page 9: Background Error Covariance Modeling

Prior correlation distribution might be a function of:

• distance from sample correlation of 1 • latitude, longitude, height• Rossby radius, Richardson number• estimated geostrophic coupling (for multi-

variate case)• estimated group velocity of errors (for

variables separated in time)• anisotropy of nearby features (e.g. fronts)

How about trying to extract this info from the ensemble?

Non

-ada

ptiv

e lo

caliz

atio

nA

dapt

ive

loca

lizat

ion

Page 10: Background Error Covariance Modeling

Ad-hoc adaptive error covariance localization

Page 11: Background Error Covariance Modeling

Smoothed ENsemble COrrelations Raised to a Power (SENCORP)

(Bishop and Hodyss, 2007, QJRMS)

1 128 256

1 128 256

(a) (b)

(c) (d)

64 128c i, 128tc i, 64 128c i, 64 128sc i,

264 128sc i, 128NAGc i,

64 128 128NAGc i, c i, 264 64128 128sc i, c i,

Bishop and Hodyss,2007, QJRMS.

In matrix form, f f

K s sP P C C

Page 12: Background Error Covariance Modeling

initialtfinalt (a) (b)

(c) (d)

K CEnsemble correlation matrix

n elementwise products of ensemble correlation

Non-adaptivelocalizationmatrix

K = 64 member ensemble

nK C

T E E

4 1/ 2 1/ 2 464 64

T C E E C

464 ,128iC

C i,128

C64 i,128

C i,128

4 1/ 2 1/ 2 464 64 64

T C C E E C

Ensemble COrrelations Raised to A Power (ECO-RAP)

Bishop and Hodyss, 2009ab, Tellus

Page 13: Background Error Covariance Modeling

• Nice features of ECO-RAP include:– Reduces to a propagating non-adaptive

localization in limit of high power– Can apply square root theorem and

separability assumption to reduce memory requirements

Bishop and Hodyss, 2009ab, Tellus

Page 14: Background Error Covariance Modeling

Square root theorem provides memory efficient representation (Bishop and Hodyss, 2009b, Tellus)

1 1

1 1 1 1 1 1

1

Raw covariance = , smooth ens correlation , hence

Hence,

K Kf T s s sT

K k k j jk j

K K K K K Kf s s s s s s s s s s

K mk nk mj nj mi ni mk mj mi nk nj nimnk j i k j i

Kf s s s s

K k j ii

z z z z z z z z z z z z

P z z C z z

P C C

P C C z z z

1 1

K K Ts s Tk j i D D

k j

z z z Z Z

Modulated ensemble memberModulated ensemble

Root is huge ensemble of modulated ensemble members!K2(K+1)/2 can be linearly independent (K=128 => a possible 1,056,768 linear independent members)

Page 15: Background Error Covariance Modeling

Raw ensemble member k Smooth ensemble member j

Smooth ensemble member i Modulated ensemble member

kz sjz

siz s s

k j iz z z

Example of a modulated ensemble member

Page 16: Background Error Covariance Modeling

Data Assimilation using Modulated EnsembleS

(DAMES) using Navy global atmospheric

model, NOGAPS

Page 17: Background Error Covariance Modeling

(a)

(c) (d)

(b)18 Z 12 Z

Sigma Level

Longitude Longitude

Sigma Level

Unlocalized ensemble covariance function of meridional wind at 18 UTC and 12 UTC with 18 UTC meridional wind variable at 90E, 40S sigma-level 15 (about 400 hPa).

No localization

Bishop, C.H. and D. Hodyss, 2011: Adaptive ensemble covariance localization in ensemble 4D-VAR state estimation, Mon. Wea . Rev. 139, 1241-1255.

Page 18: Background Error Covariance Modeling

(a) (b)18 Z 12 Z

Longitude Longitude

(c) (d)

Sigma Level Sigma Level

Ensemble covariance function localized with the partially adaptive ensemble covariance localization function (PAECL).

Adaptive localization

Bishop, C.H. and D. Hodyss, 2011: Adaptive ensemble covariance localization in ensemble 4D-VAR state estimation, Mon. Wea . Rev. 139, 1241-1255.

Page 19: Background Error Covariance Modeling

(a) (b)18 Z 12 Z

Longitude Longitude

(c) (d)

Sigma Level Sigma Level

Ensemble covariance function localized with the non-adaptive ensemble covariance localization (NECL).

Non-adaptive localization

Page 20: Background Error Covariance Modeling

(a)

(c)

(b)

(d)Sigma Level

Longitude Longitude

Sigma Level

Comparison of structure of optimally tuned adaptive (a) and non-adaptive (b) localization functions at 12 UTC. Fig’s (c) and (d) give the corresponding vertical structure of the adaptive and non-adaptive localization functions along the N latitude circle.

This study was only able to compare the performance of the two schemes over a long enough period to establish any significance between the performance of non-adaptive and adaptive localization. In simpler models, adaptive localization has been shown to beat or match the performance of non-adaptive localization – depending on the need for adaptive localization.

Page 21: Background Error Covariance Modeling

• Multi-scale issues

21

Page 22: Background Error Covariance Modeling

Synoptic scales need to be analyzed

1,000 km

Page 23: Background Error Covariance Modeling

so do mesoscales,

100 km

100 km

Page 24: Background Error Covariance Modeling

and convective scales need to be analyzed too!

10 km

Page 25: Background Error Covariance Modeling

Motivation: Our Multi-Scale World

Convection near a mid-latitude cycloneSimple Model Ensemble Perturbation

Page 26: Background Error Covariance Modeling

Motivation: Our Multi-Scale World

Convection near a mid-latitude cyclone

Red – True 1-point covarianceBlack – 32 member ensemble 1-point covariance

Page 27: Background Error Covariance Modeling

Will traditional localization help?

Localization with Broad Function Localization with Narrow Function

Green – Localized 1-point covariance

Page 28: Background Error Covariance Modeling

Now imagine the convection is moving …

Initial Time Final Time

Red – True 1-point covarianceBlack – 32 member ensemble 1-point covariance

Page 29: Background Error Covariance Modeling

Will traditional localization help?Localization with Broad Function

Green – Localized 1-point covariance

Initial Time Final Time

Page 30: Background Error Covariance Modeling

New Method: Multi-scale DAMES• Spatially smooth the ensemble members and call this the large-

scale ensemble– Use a step-function in wavenumber space

• Subtract the large-scale ensemble from the raw ensemble and call this the small-scale ensemble – This implies a partition like

• Apply the DAMES method to expand ensemble size

• to each ensemble separately– Create localization ensemble members, i.e. choose variable and smooth– Use modulation products to construct “modulation” ensemble members

• Add modulation ensemble members from the large-scale ensemble and the small-scale ensemble together to form one set of modulation ensemble members– This implies that the final ensemble looks like

L Si i ix x x

MSD LD SDi i ix x x

1 1 1

K K K Tf s s s s s sK k j i k j i

k j i

P C C z z z z z z

L SiMS D iD iDx x x

Page 31: Background Error Covariance Modeling

Decompose into large and small-scale portions …

Large-Scale Ensemble Perturbation Small-Scale Ensemble Perturbation

Page 32: Background Error Covariance Modeling

DAMES: Step 1 – Smooth the perturbations Smoothed Large-Scale Ensemble Perturbations

(Smooth Member 1) (Smooth Member 2)

Page 33: Background Error Covariance Modeling

DAMES: Step 2 – Modulate smooth members and normalize the resulting ensemble

Large-Scale Ensemble Perturbations

(Smooth Member 1) x (Smooth Member 2) Associated Localization Function

Page 34: Background Error Covariance Modeling

DAMES: Step 3 – Modulate raw member

(Raw Member 2) x (Smooth Member 1) x (Smooth Member 2)

This is a modulation ensemble member!

Large-Scale Ensemble Perturbations

Page 35: Background Error Covariance Modeling

Large-Scale and Small-Scale Perturbations

(Raw Member 2) x (Smooth Member 1) x (Smooth Member 2)

Large-Scale Ensemble Perturbation Small-Scale Ensemble Perturbation

Page 36: Background Error Covariance Modeling

The MS-DAMES Ensemble

MSD LD SDi i ix x x

Recall that the MS-DAMES modulation ensemble is the sum of the large-scale and small-scale members:

L SiMS D iD iDx x x

Subsequent research has shown that treating the small and large scale modulated members as individual members (rather than adding them together) actually works better.

Page 37: Background Error Covariance Modeling

1-point Covariance functions

Initial Time Final Time

Red – True 1-point covarianceGreen – Non-adaptively localized covarianceBlue – Multi-scale DAMES covariance

The Multi-scale DAMES algorithm gave a qualitatively good result!

Page 38: Background Error Covariance Modeling

Have we localized the covariance?

Effective Localization

Blue – MS-DAMES 1-point covarianceBlack – Raw 1-point covariance

Page 39: Background Error Covariance Modeling

Let’s Assimilate Some Obs …

Initial Time Analysis Error (R = 0)

Red – OptimalGreen – Non-adaptiveBlue – MS-DAMES

• Two cases: Observe every point at the final time with ob error = 0 or 1

• Obtain the initial state using the modulation ensemble to propagate the effect of the obs back in time

• 16 trials: Initial time RMS(Analysis Error) Ob error = 0 Ob error = 1 Optimal 0.27 0.61 MS-D 0.31 0.63 Non-adapt 1.4 0.92

The MS-DAMES algorithm gave a superior quantitative result

Page 40: Background Error Covariance Modeling

Summary

• Ensemble based error covariance models require some form of covariance localization that serves to (a) increase the effective ensemble size, and (b) attenuate spurious correlations.

• When errors (a) move a significant distance relative to their correlation length scale over the DA window and/or (b) exhibit differing scales at differing locations, adaptive localization can significantly improve the covariance model.

• When differing error scales move in differing directions multi-scale ensemble covariance localization is likely to be of use.

40

Page 41: Background Error Covariance Modeling

Current DA ill-suited to multi-scale problem(a) (b)

x x

Covariance with (x=128, t=0) at t=0 Covariance with (x=128, t=0) at t=12Red gives true covarianceBlue gives sample covariance from 32 member ensembleGreen gives localization functionBlack gives localized ensemble covariance

Red gives true covarianceBlue gives sample covariance from 32 member ensembleGreen gives localization functionBlack gives localized ensemble covariance

(a) (b)

x x

Covariance with (x=128, t=0) at t=0 Covariance with (x=128, t=0) at t=12Red gives true covarianceBlue gives sample covariance from 32 member ensembleGreen gives localization functionBlack gives localized ensemble covariance

Red gives true covarianceBlue gives sample covariance from 32 member ensembleGreen gives localization functionBlack gives localized ensemble covariance

Multiscale correlation functions. In (a) and (b) red lines give the true covariance of variables in 256 variable model at t=0 and t=12, respectively, with the 128th variable at t=0. Blue lines give the corresponding raw ensemble covariance from a 32 member ensemble. Black lines give the corresponding localized ensemble covariance. Green lines give the non-adaptive localization function used to localize the ensemble covariances.

Current physical space ensemble covariance localization techniques inadequate for multi-scale problem

Page 42: Background Error Covariance Modeling

Adaptive Localization needed because:

• True error correlation length scale is a function of time and location

• The location of correlated errors propagates through time

• Multiple error correlation length scales may exist simultaneously

Naval Research Laboratory Marine Meteorology Division Monterey, California

Page 43: Background Error Covariance Modeling

What is the true error distribution ?

• Imagine an unimaginably large number of quasi-identical Earths.

Each Earth has one true state and one prediction but these differ from one Earth to another.

Collect all Earths having the same true state but differing forecasts of this state to

defin . T|e

t f

f t x x

x x

he error distribution is the distribution of differences

between individual forecasts and single truth within this set.

Collect all Earths having the same historical observati

fixed-tr

ons b

u

u

th

f

t diy

1 2

fering true atmospheric

states to define . The differences between individual truths and the

mean truth within this

| , ,...

fixed-obs set define the forecast error distribution. i

t ti x x y y

(Slartibartfast – Magrathean designer of planets, D. Adams, Hitchhikers …)

43

Page 44: Background Error Covariance Modeling

44

What is an error covariance?

12 1

2

12 1 2 1 21

Covariance between forecast error of forecast of variable 1 with

forecast error of forecast of variable 2 is defined by1lim

where indexes the replicate Earth t

f

f

nf f f f

i ii

P

P nn

i

12 12

12 1 12 2

2 2

1 1 2 2

o which the errors pertain. Oftenwe represent in terms of its correlation ; i.e.

where a d

n

f f

f f f f

P C

P C

Page 45: Background Error Covariance Modeling

1. The temperature forecast for Albuquerque is much colder than the verifying observation by 5K. Does this mean that the forecast for Santa Fe was also to cold?

2. What if the forecast error was associated with an approaching cold front?

3. How would the orientation of the cold front change your answer to question 1?

45

Why do error covariances matter?

Page 46: Background Error Covariance Modeling

• We don’t know the true state and hence cannot produce error samples.

• We can attempt to – infer forecast error covariances from

innovations (y-Hxf) (e.g. Hollingsworth and Lohnberg, 1986, Tellus)

• And/or – create proxies of error from first principals

(e.g. ensemble perturbations)46

Problem

Page 47: Background Error Covariance Modeling

47

Static forecast error covariances from innovations

Assume that the jth observation where is the true

value of the ob. and is the true atmospheric state. Assume that the

forecast of the observation then the innovati

t o t tj j j j j

t

f t fj j j

y H H y

H H

x x

x

x x

on

associated with the jth observation is given by

;

in other words, the innovation is just the difference between the observation errorand the forecast error (for unb

f t o t f o fj j j j j j j j jv y H H H x x x

iased obs/fcsts).

Assume that 0.

Now consider the covariance of innovations at nearby observation sites and .

o fj j

f f o f f o o oi j i j i j i j i j

i j

v v

Page 48: Background Error Covariance Modeling

48

Static forecast error covariances from innovations

2 2

Would observation errors be correlated with forecast errors?

Assuming that 0,

= when . However,

, when .

In other words, for

o o f o o f o f o fj i j i j i i i j j

f fi j i j ij

f oi i i i ii ii

v v p i j

v v P R i j

uncorrelated observation errors,innovation covariance equals forecast error covariance, while innovation variance equals the forecast error variance plus the observation error variance.

Page 49: Background Error Covariance Modeling

49

Hollingsworth-Lönnberg Method(Hollingsworth and Lönnberg, 1986)

Innovation covariances binned by separation distance

Extrapolate green curve to zero separation, and compare with innovation variance

Fcst error variance Pii

Static forecast error covariances from innovations

Ob error variance Rii

Includes uncorrelated error of representation

Desroziers’ Method(Desroziers et al 2005) TO O

FAE

d d R

TO O TF FE

d d R HBH

TA O TF FE

d d HBHFrom O-F, O-A, and A-F statistics, the observation error covariance matrix R, the representer HBHT, and their sum can be diagnosed

An attractive property of the HL method is that its estimates are entirely independent of the estimates of P and R that are used in the data assimilation scheme.

Desrozier’s method depends on differences between analyses and observations. These differences are entirely dependent on the assumptions made in the DA scheme.

Page 50: Background Error Covariance Modeling

50

Bauer et al., 2006, QJRMS

Static forecast error covariances from innovations

Page 51: Background Error Covariance Modeling

51

C/O Mike Fisher

Why does the correlation function for the AIREP data look qualitatively different to that from the SSMI radiances?

Page 52: Background Error Covariance Modeling

• Pros– Ultimately, observations are our only means to

perceive forecast error.– Innovation based approaches enable both forecast

error covariances and observation error variances to be simultaneously estimated.

• Cons– Only gives error estimates where there are

observations (what about the deep ocean, upper atmosphere, cloud species, etc)

– Provides extremely limited information about multi-variate balance.

– Limited flow dependent error covariance information.52

Pros and cons of error covariances from binned innovations

Page 53: Background Error Covariance Modeling

1. Parish and Derber’s (1992, MWR)“very crude 1st step” using the difference between a 48 hr and 24 hr fcsts valid at the same time as a proxy for 6 hr fcst error has been widely used.

2. Oke et al. (2008, Ocean Modelling) use deviations of state about 3 month running average as a proxy for forecast error.

3. Both 1 and 2 can be made to be somewhat consistent with innovations 53

Covariances of proxies of forecast error

How could we produce better proxies of forecast error?

Page 54: Background Error Covariance Modeling

• Forecast error distributions depend on analysis error distributions and model error distributions.

• Analysis error distributions depend on the data assimilation scheme used and the location and accuracy of the observations assimilated.

• Estimation of these distributions is difficult in practice but there is theory for it.

54

Covariances of proxies of forecast error

Page 55: Background Error Covariance Modeling

55

pdf of truth given prior information (the prior)

| likelihood density of observations given a particular

| posterior distribution of truth given prior information and

Ideal DA would

t t

t t

t

L

x x

y x y x

x y y

use

||

|

(Bayes' theorem) to turn a prior pdf of truth into an observation informed posterior pdf.

t tt

t t

V

L

L dV

y x xx y

y x x

The effect of observations on errors,Bayes’ theorem

Page 56: Background Error Covariance Modeling

Prior pdf of truth

prior tx

Ensemble forecasts are used to estimate this distribution. They are a collection of weather forecasts started from differing but equally plausible initial conditions and propagated forward using a collection of equally plausible dynamical or stochastic-dynamical models.

Probability density

Value of truth

Page 57: Background Error Covariance Modeling

Likelihood density function

prior tx

22

2 0

0

1 1Likelihood | exp , 22

,

~ 0,

1 and 1/ 4in this example.

t

t

t

y xL y x

RR

y x

N R

y R

Value of truth

In interpreting the likelihood function (red curve) note that y is fixed at y=1.

The red curve describes how the probability density of obtaining an error prone observation of y=1 varies with the true value xt.

Page 58: Background Error Covariance Modeling

Posterior pdf

prior tx

2 0

0

Likelihood | ,

,

~ 0,1 in this example

t

t

L y x

y x

N Ry

Posterior

||

|

t tt

t t t

L y x xx y

L y x x dx

Value of truth

Probability density No operational or near operational data assimilation schemes are capable of accurately representing such multi-modal posterior distributions.

Page 59: Background Error Covariance Modeling

59

Page 60: Background Error Covariance Modeling

60

Works for Gaussian forecast and observation errors.

Page 61: Background Error Covariance Modeling

61

Ensemble of perturbed obs 4DVARS does not solve Bayes’ theorem

Green line is pdf of ensemble of converged perturbed obs 4DVARs having the correct prior and correct observation error variance.

Blue line is the pdf of ensemble of 4DVARS after 1st inner loop (not converged)

Black line is the true posterior pdf.

Page 62: Background Error Covariance Modeling

62

EnKF doesn’t solve Bayes’ theorem either

Cyan line is posterior pdf from EnKF

Black line is the true posterior pdf.

Page 63: Background Error Covariance Modeling

1. Ensembles of 4DVARs and/or EnKFs provide accurate flow dependent analysis and forecast error distributions provided all error distributions are Gaussian and accurately specified.

2. In the presence of non-linearities and non-Gaussianity, the 4DVAR/EnKF proxies are inaccurate but probably not as inaccurate as proxies for which 1 does not hold.

3. One can use an archive of past flow dependent error proxies to define a static or quasi-climatological error covariance. (Examples follow) 63

Recapitulation on proxy error distributions

Page 64: Background Error Covariance Modeling

64

Page 65: Background Error Covariance Modeling

65

Page 66: Background Error Covariance Modeling

Computationally Efficient Quasi-Static Error Covariance Models

66

Page 67: Background Error Covariance Modeling

67

Page 68: Background Error Covariance Modeling

68

Boer, G. J., 1983: Homogeneous and Isotropic Turbulence on the Sphere. J. Atmos. Sci., 40, 154–163.

Pointed out that isotropic correlation functions on the sphere are obtained from EDE^T where E is a matrix listing spherical harmonics and D is a diagonal matrix whose values (variances) only depend on the total wave number.

Page 69: Background Error Covariance Modeling

69

1

1

1

1

0 0...

0 0

0...

,0

.

0 0...

0 0

so and define the

0 0 0

0 0 00 0

0 0

0 0 0

0 0 0

n

n

n

n

h

h

h

h

VH V

V

V V

vertical correlation matrix of total wavenumber = 1and total wavenumber modes.n

Page 70: Background Error Covariance Modeling

70

Page 71: Background Error Covariance Modeling

71

Page 72: Background Error Covariance Modeling

72

Wavelet transforms permit a compromise between these two extremes. ECMWF currently has a wavelet transform based background error covariance model.May have time to touch on this tomorrow.

Page 73: Background Error Covariance Modeling

73

Page 74: Background Error Covariance Modeling

74

Page 75: Background Error Covariance Modeling

75

Page 76: Background Error Covariance Modeling

76

Page 77: Background Error Covariance Modeling

77

Page 78: Background Error Covariance Modeling

78

Divergence without omega equation

Divergence with omega equation

Page 79: Background Error Covariance Modeling

79

Sophisticated balance operators impart a degree of flow dependence to both the error correlations

and the error variances!

Page 80: Background Error Covariance Modeling

Recapitulation on today’s lecture• Differences between forecasts and observations can be used to

infer aspects of spatio-temporal averages of – Observation error variance– Forecast error variance– Quasi-isotropic error correlations

• Monte Carlo approaches (Perturbed obs 3D/4D VAR, EnKF) and deterministic EnKFs (ETKF, EAKF, MLEF) provide compelling error proxies for both flow-dependent error covariance models and flow-dependent error covariance models.

• In variational schemes, the need for cost-efficient matrix multiplies has led to elegant idealizations of the forecast error covariance matrix– sophisticated balance constraints can be built into these models.

• There were many approaches I did not cover (Recursive filters, Wavelet Transforms, etc).

• Tomorrow: Ensemble based flow dependent error covariance models

80