Compressibility equation: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
en>Saehry
mNo edit summary
 
en>Addbot
m Bot: Migrating 2 interwiki links, now provided by Wikidata on d:q3730852
 
Line 1: Line 1:
Hi there, I am Sophia. Invoicing is my occupation. Mississippi is the only location I've been residing in but I will have to transfer in a year or two. What me and my family members adore is doing ballet but I've been taking on new things recently.<br><br>Here is my homepage; tarot card readings ([http://Www.Zavodpm.ru/blogs/glennmusserrvji/14565-great-hobby-advice-assist-allow-you-get-going recommended you read])
The '''ensemble Kalman filter''' (EnKF) is a [[recursive filter]] suitable for problems with a large number of variables, such as [[discretization]]s of [[partial differential equation]]s in geophysical models. The EnKF originated as a version of the [[Kalman filter]] for large problems (essentially, the [[covariance matrix]] is replaced by the [[sample covariance matrix|sample covariance]]), and it is now an important [[data assimilation]] component of [[ensemble forecasting]]. EnKF is related to the [[particle filter]] (in this context, a particle is the same thing as ensemble member) but the EnKF makes the assumption that all probability distributions involved are [[Normal distribution|Gaussian]];  when it is applicable, it is much more efficient than the [[particle filter]].
<!-- The original version of this page was converted from LaTeX by [[User:Jmath666/latex2wiki.pl]] -->
<!-- first version written by [[User:Oleg Alexandrov]], now developed and maintained by [[User:Jmath666]] -->
<!-- NOTE: this page is using references by [[Wikipedia:Footnote]] -->
<!-- NOTE: please follow this reference scheme and leave these comments in. -->
 
==Introduction==
 
The Ensemble Kalman Filter (EnKF) is a [[Monte Carlo method|Monte Carlo]] implementation of the [[Bayesian inference|Bayesian update]] problem: given a [[probability density function]] (pdf) of the state of the modeled system (the ''[[Prior probability|prior]]'', called often the forecast in geosciences) and the data likelihood, the [[Bayes theorem]] is used to obtain the pdf after the data likelihood has been taken into account (the ''[[Posterior probability|posterior]]'', often called the analysis). This is called a Bayesian update. The Bayesian update is combined with advancing the model in time, incorporating new data from time to time. The original [[Kalman Filter]]<ref name="Kalman-1960-NAL">R. E. Kalman, ''A new approach to linear filtering and prediction problems'', Transactions of the ASME -- Journal of Basic Engineering, Series D, 82 (1960), pp. 35--45.
 
</ref> assumes that all pdfs are [[normal distribution|Gaussian]] (the Gaussian assumption) and provides algebraic formulas for the change of the [[mean]] and the [[covariance matrix]] by the Bayesian update, as well as a formula for advancing the covariance matrix in time provided the system is linear. However, maintaining the covariance matrix is not feasible computationally for high-dimensional systems. For this reason, EnKFs were developed.<ref name="Evensen-1994-SDA">G. Evensen, ''Sequential data assimilation with nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics'', Journal of Geophysical Research, 99 (C5) (1994), pp. 143--162.
 
</ref><ref name="Houtekamer-1998-DAE">P. Houtekamer and H. L. Mitchell, ''Data assimilation using an ensemble Kalman filter technique'', Monthly Weather Review, 126 (1998), pp. 796--811.
 
</ref> EnKFs represent the distribution of the system state using a collection of state vectors, called an [[Numerical weather prediction#Ensembles|ensemble]], and replace the covariance matrix by the [[sample covariance]] computed from the ensemble. The ensemble is operated with as if it were a [[random sample]], but the ensemble members are really not [[Statistical independence|independent]] - the EnKF ties them together. One advantage of EnKFs is that advancing the pdf in time is achieved by simply advancing each member of the ensemble. For a survey of EnKF and related data assimilation techniques, see G. Evensen.<ref name="Evensen-2007-DAE">G. Evensen, ''Data assimilation : The ensemble Kalman filter, Springer, Berlin, 2007.</ref>
 
==A derivation of the EnKF==
 
===The Kalman Filter===
 
Let us review first the [[Kalman filter]]. Let <math>\mathbf{x}</math> denote the <math>n</math>-dimensional [[state vector]] of a model, and assume that it has [[normal distribution|Gaussian probability distribution]] with mean <math>\mathbf{\mu}</math> and covariance <math>Q</math>, i.e., its pdf is
 
:<math> p(\mathbf{x})\propto\exp\left(  -\frac{1}{2}(\mathbf{x}-\mathbf{\mu })^{\mathrm{T}}Q^{-1}(\mathbf{x}-\mathbf{\mu})\right)  . </math>
 
Here and below, <math>\propto</math> means proportional; a pdf is always scaled so that its integral over the whole space is one. This <math>p(\mathbf{x})</math>, called the ''[[prior probability|prior]]'', was evolved in time by running the model and now is to be updated to account for new data. It is natural to assume that the error distribution of the data is known; data have to come with an error estimate, otherwise they are meaningless. Here, the data <math>\mathbf{d}</math> is assumed to have Gaussian pdf with covariance <math>R</math> and mean <math>H\mathbf{x}</math>, where <math>H</math> is the so-called [[Hat matrix|observation matrix]]. The covariance matrix <math>R</math> describes the estimate of the error of the data; if the random errors in the entries of the data vector <math>\mathbf{d}</math> are independent, <math>R</math> is diagonal and its diagonal entries are the squares of the [[standard deviation]] (“error size”) of the error of the corresponding entries of the data vector <math>\mathbf{d}</math>. The value <math>H\mathbf{x}</math> is what the value of the data would be for the state <math>\mathbf{x}</math> in the absence of data errors. Then the probability density <math>p(\mathbf{d}|\mathbf{x})</math> of the data <math>\mathbf{d}</math> conditional of the system state <math>\mathbf{x}</math>, called the [[Likelihood function|data likelihood]], is
 
:<math> p\left(  \mathbf{d}|\mathbf{x}\right)  \propto\exp\left(  -\frac{1}{2}(\mathbf{d}-H\mathbf{x})^{\mathrm{T}}R^{-1}(\mathbf{d}-H\mathbf{x})\right) . </math>
 
The pdf of the state and the [[Likelihood function|data likelihood]] are combined to give the new probability density of the system state <math>\mathbf{x}</math> conditional on the value of the data <math>\mathbf{d}</math> (the ''[[posterior probability|posterior'']]) by the [[Bayes theorem#Bayes' theorem for probability densities|Bayes theorem]],
 
:<math> p\left(  \mathbf{x}|\mathbf{d}\right)  \propto p\left(  \mathbf{d}|\mathbf{x}\right)  p(\mathbf{x}). </math>
 
The data <math>\mathbf{d}</math> is fixed once it is received, so denote the posterior state by <math>\mathbf{\hat{x}}</math> instead of <math>\mathbf{x}|\mathbf{d}</math> and the posterior pdf by <math>p\left(  \mathbf{\hat{x}}\right)  </math>. It can be shown by algebraic manipulations<ref name="Anderson-1979-OF">B. D. O. Anderson and J. B. Moore, ''Optimal filtering'', Prentice-Hall, Englewood Cliffs, N.J., 1979.
 
</ref> that the posterior pdf is also Gaussian,
 
:<math> p\left(  \mathbf{\hat{x}}\right)  \propto\exp\left(  -\frac{1}{2}(\mathbf{\hat{x}}-\mathbf{\hat{\mu}})^{\mathrm{T}}\hat{Q}^{-1}(\mathbf{\hat{x}}-\mathbf{\hat{\mu}})\right)  , </math>
 
with the posterior mean <math>\mathbf{\hat{\mu}}</math> and covariance <math>\hat{Q}</math> given by the Kalman update formulas
 
:<math> \mathbf{\hat{\mu}}=\mathbf{\mu}+K\left(  \mathbf{d}-H\mathbf{\mu}\right) ,\quad\hat{Q}=\left(  I-KH\right)  Q, </math>
 
where
 
:<math> K=QH^{\mathrm{T}}\left(  HQH^{\mathrm{T}}+R\right)  ^{-1}</math>
 
is the so-called [[Kalman filter#Kalman gain derivation|Kalman gain]] matrix.
 
===The Ensemble Kalman Filter===
 
The EnKF is a Monte Carlo approximation of the Kalman filter, which avoids evolving the covariance matrix of the pdf of the state vector <math>\mathbf{x}</math>. Instead, the pdf is represented by an ensemble
 
:<math> X=\left[  \mathbf{x}_{1},\ldots,\mathbf{x}_{N}\right]  =\left[  \mathbf{x}_{i}\right]. </math>
 
<math>X</math> is an <math>n\times N</math> matrix whose columns are the ensemble members, and it is called the ''prior ensemble''. Ideally, ensemble members would form a [[Random sample|sample]] from the prior distribution. However, the ensemble members are not in general [[Statistical independence|independent]] except in the initial ensemble, since every EnKF  step ties them together. They are deemed to be approximately independent, and all calculations proceed as if they actually were independent.
 
Replicate the data <math>\mathbf{d}</math> into an <math>m\times N</math> matrix
 
:<math> D=\left[  \mathbf{d}_{1},\ldots,\mathbf{d}_{N}\right]  =\left[  \mathbf{d}_{i}\right], \quad \mathbf{d}_{i}=\mathbf{d}+\mathbf{\epsilon_{i}}, \quad \mathbf{\epsilon_{i}} =N(0,R), </math>
 
so that each column <math>\mathbf{d}_{i}</math> consists of the data vector <math>\mathbf{d}</math> plus a random vector from the <math>m</math>-dimensional normal distribution <math>N(0,R)</math>. If, in addition, the columns of <math>X</math> are a sample from the [[prior probability]] distribution, then the columns of
 
:<math> \hat{X}=X+K(D-HX) </math>
 
form a sample from the [[posterior probability]] distribution. [To see this in the scalar case with <math>H=1</math>: Let <math>x_i = \mu + \xi_i, \; \xi_i \sim N(0, \sigma_x^2)</math>, and <math>d_i = d + \epsilon_i, \; \epsilon_i \sim N(0, \sigma_d^2).</math>  Then <math>\hat{x}_i = \left(\frac{1/\sigma_x^2}{1/\sigma_x^2 + 1/\sigma_d^2} \mu + \frac{1/\sigma_d^2}{1/\sigma_x^2 + 1/\sigma_d^2} d \right)+ \left(\frac{1/\sigma_x^2}{1/\sigma_x^2 + 1/\sigma_d^2} \xi_i + \frac{1/\sigma_d^2}{1/\sigma_x^2 + 1/\sigma_d^2} \epsilon_i \right) </math>. The first sum is the posterior mean, and the second sum, in view of the independence, has a variance <math>\left(\frac{1/\sigma_x^2}{1/\sigma_x^2 + 1/\sigma_d^2}\right)^2 \sigma_x^2 + \left(\frac{1/\sigma_d^2}{1/\sigma_x^2 + 1/\sigma_d^2}\right)^2 \sigma_d^2 =  \frac{1}{1/\sigma_x^2 + 1/\sigma_d^2}</math>, which is the posterior variance. ]
 
The EnKF is now obtained<ref name="Johns-2005-CEK">C. J. Johns and J. Mandel, ''A two-stage ensemble Kalman filter for smooth data assimilation''. Environmental and Ecological Statistics, in print. Special issue, Conference on New Developments of Statistical Analysis in Wildlife, Fisheries, and Ecological Research, Oct 13-16, 2004, Columbia, MI. CCM Report 221, University of Colorado at Denver and Health Sciences Center, 2005. [http://www.math.ucdenver.edu/ccm/reports/rep221.pdf report]
 
</ref> simply by replacing the state covariance <math>Q</math> in Kalman gain matrix <math>K</math> by the sample covariance <math>C</math> computed from the ensemble members (called the ''ensemble covariance''). That is: <math>K=CH^{\mathrm{T}}\left(  HCH^{\mathrm{T}}+R\right)  ^{-1}</math>
 
==Implementation==
 
===Basic formulation===
 
Here we follow.<ref name="Burgers-1998-ASE">G. Burgers, P. J. van Leeuwen, and G. Evensen, ''Analysis scheme in the ensemble Kalman filter'', Monthly Weather Review, 126 (1998), pp. 1719--1724.
 
</ref><ref name="Evensen-2003-EKF">G. Evensen, ''The ensemble Kalman filter: Theoretical formulation and practical implementation'', Ocean Dynamics, 53 (2003), pp. 343--367.
 
</ref> Suppose the ensemble matrix <math>X</math> and the data matrix <math>D</math> are as above. The ensemble mean and the covariance are
 
:<math> E\left(  X\right)  =\frac{1}{N}\sum_{k=1}^{N}\mathbf{x}_{k},\quad C=\frac{AA^{T}}{N-1}, </math>
 
where
 
:<math> A=X-E\left(  X\right) \mathbf{e}_{1\times N}  =X-\frac{1}{N}\left(  X\mathbf{e}_{N\times1}\right) \mathbf{e}_{1\times N}, </math>
 
and <math>\mathbf{e}</math> denotes the matrix of all ones of the indicated size.
 
The posterior ensemble <math>X^{p}</math> is then given by
 
:<math> X^{p}=X+CH^{T}\left(  HCH^{T}+R\right)  ^{-1}(D-HX), </math>
 
where the perturbed data matrix <math>D</math> is as above.
 
Note that since <math>R</math> is a covariance matrix, it is always [[positive semidefinite matrix|positive semidefinite]] and usually [[positive semidefinite matrix|positive definite]], so the inverse above exists and the formula can be implemented by the  [[Cholesky decomposition]].<ref name="Mandel-2006-EIE">J. Mandel, ''Efficient implementation of the ensemble Kalman filter''. CCM Report 231, University of Colorado at Denver and Health Sciences Center. [http://www.math.ucdenver.edu/ccm/reports/rep231.pdf link], June 2006.
 
</ref> In,<ref name="Burgers-1998-ASE"/><ref name="Evensen-2003-EKF"/> <math>R</math> is replaced by the sample covariance <math>\tilde{D} \tilde{D}^{T}/\left(  N-1\right)  </math> where <math>\tilde{D} = D - \frac{1}{N} d \, \mathbf{e}_{1\times N}</math>and the inverse is replaced by a [[pseudoinverse]], computed using the [[Singular Value Decomposition]] (SVD) .
 
Since these formulas are matrix operations with dominant [[BLAS#Level 3|Level 3]] operations,<ref name="Golub-1989-MAC">G. H. Golub and C. F. V. Loan, ''Matrix Computations'', Johns Hopkins Univ. Press, 1989. Second Edition.
 
</ref> they are suitable for efficient implementation using software packages such as [[LAPACK]] (on serial and [[shared memory]] computers) and [[ScaLAPACK]] (on [[distributed memory]] computers).<ref name="Mandel-2006-EIE"/> Instead of computing the [[inverse matrix|inverse]] of a matrix and multiplying by it, it is much better (several times cheaper and also more accurate) to compute the [[Cholesky decomposition]] of the matrix and treat the multiplication by the inverse as solution of a linear system with many simultaneous right-hand sides.<ref name="Golub-1989-MAC"/>
 
===Observation matrix-free implementation===
 
Since we have replaced the covariance matrix with ensemble covariance, this leads to a simpler formula where ensemble observations are directly used without explicitly specifying the matrix <math>H</math>. More specifically, define a function <math>h(\mathbf{x})</math> of the form
 
:<math> h(\mathbf{x})=H\mathbf{x}. </math>
 
The function <math>h</math> is called the ''[[observation function]]'' or, in the [[inverse problem]]s context, the ''[[Inverse problem#Probabilistic formulation of inverse problems|forward operator]]''. The value of <math>h(\mathbf{x})</math> is what the value of the data would be for the state <math>\mathbf{x}</math> assuming the measurement is exact. Then the posterior ensemble can be rewritten as
 
:<math> X^{p}=X+\frac{1}{N-1}A\left(  HA\right)  ^{T}P^{-1}(D-HX) </math>
 
where
 
:<math> HA=HX-\frac{1}{N}\left(  \left(  HX\right)  \mathbf{e}_{N\times1}\right) \mathbf{e}_{1\times N}, </math>
 
and
 
:<math> P=\frac{1}{N-1}HA\left(  HA\right)  ^{T}+R, </math>
 
with
 
:<math>\begin{align} \left[  HA\right]  _{i}  &  =H\mathbf{x}_{i}-H\frac{1}{N}\sum_{j=1}^{N}\mathbf{x}_{j}\ &  =h\left(  \mathbf{x}_{i}\right)  -\frac{1}{N}\sum_{j=1}^{N}h\left( \mathbf{x}_{j}\right)  . \end{align}</math>
 
Consequently, the ensemble update can be computed by evaluating the observation function <math>h</math> on each ensemble member once and the matrix <math>H</math> does not need to be known explicitly. This formula holds also<ref name="Mandel-2006-EIE"/> for an observation function <math>h(\mathbf{x})=H\mathbf{x+f}</math> with a fixed offset <math>\mathbf{f}</math>, which also does not need to be known explicitly. The above formula has been commonly used for a nonlinear observation function <math>h</math>, such as the position of a [[hurricane]] [[vortex]].<ref name="Chen-2006-AVP">Y. Chen and C. Snyder, ''Assimilating vortex position with an ensemble Kalman filter''. Monthly Weather Review, to appear, 2006. [http://www.mmm.ucar.edu/people/snyder/papers/ChenSnyder2006_draft.pdf preprint].
 
</ref> In that case, the observation function is essentially approximated by a linear function from its values at ensemble members.
 
===Implementation for a large number of data points===
 
For a large number <math>m</math> of data points, the multiplication by <math>P^{-1}</math> becomes a bottleneck. The following alternative formula is advantageous when the number of data points <math>m</math> is large (such as when assimilating gridded or pixel data) and the data error [[covariance matrix]] <math>R</math> is diagonal (which is the case when the data errors are uncorrelated), or cheap to decompose (such as banded due to limited covariance distance). Using the [[Sherman–Morrison–Woodbury formula]]<ref name="Hager-1989-UIM">W. W. Hager, ''Updating the inverse of a matrix'', SIAM Rev., 31 (1989), pp. 221--239.
 
</ref>
 
:<math> (R+UV^{T})^{-1}=R^{-1}-R^{-1}U(I+V^{T}R^{-1}U)^{-1}V^{T}R^{-1}, </math>
 
with
 
:<math> U=\frac{1}{N-1}HA,\quad V=HA, </math>
 
gives
 
:<math>\begin{align} P^{-1}  &  =\left(  R+\frac{1}{N-1}HA\left(  HA\right)  ^{T}\right)  ^{-1}\ =  \\
&  =R^{-1}\left[  I-\frac{1}{N-1}\left(  HA\right)  \left(  I+\left( HA\right)  ^{T}R^{-1}\frac{1}{N-1}\left(  HA\right)  \right)  ^{-1}\left( HA\right)  ^{T}R^{-1}\right]  , \end{align}</math>
 
which requires only the solution of systems with the matrix <math>R</math> (assumed to be cheap) and of a system of size <math>N</math> with <math>m</math> right-hand sides. See<ref name="Mandel-2006-EIE"/> for operation counts.
 
==Further extensions==
 
The EnKF version described here involves randomization of data. For filters without randomization of data, see.<ref name="Anderson-2001-EAK">J. L. Anderson, ''An ensemble adjustment Kalman filter for data assimilation'', Monthly Weather Review, 129 (2001), pp. 2884--2903.
 
</ref><ref name="Evensen-2004-SSR">G. Evensen, ''Sampling strategies and square root analysis schemes for the EnKF'', Ocean Dynamics, 54 (2004), pp. 539--560.
 
</ref><ref name="Tippett-2003-ESR">M. K. Tippett, J. L. Anderson, C. H. Bishop, T. M. Hamill, and J. S. Whitaker, ''Ensemble square root filters'', Monthly Weather Review, 131 (2003), pp. 1485--1490.
 
</ref>
 
Since the ensemble covariance is [[rank deficient]] (there are many more state variables, typically millions, than the ensemble members, typically less than a hundred), it has large terms for pairs of points that are spatially distant. Since in reality the values of physical fields at distant locations are not that much [[correlated]], the covariance matrix is tapered off artificially based on the distance, which gives rise to [[Localized ensemble Kalman filters|localized EnKF]] algorithms.<ref name="Anderson-2003-LLS">J. L. Anderson, ''A local least squares framework for ensemble filtering'', Monthly Weather Review, 131 (2003), pp. 634--642.
</ref><ref name="Ott-2003-LEK">E. Ott, B. R. Hunt, I. Szunyogh, A. V. Zimin, E. J. Kostelich, M. Corazza, E. Kalnay, D. Patil, and J. A. Yorke, ''A local ensemble Kalman filter for atmospheric data assimilation'', Tellus A, 56 (2004), pp. 415--428.
</ref> These methods modify the covariance matrix used in the computations and, consequently, the posterior ensemble is no longer made only of linear combinations of the prior ensemble.
 
For nonlinear problems, EnKF can create posterior ensemble with non-physical states. This can be alleviated by [[Tikhonov regularization|regularization]], such as [[Penalty method|penalization]] of states with large spatial [[gradient]]s.<ref name="Johns-2005-CEK"/>
 
For problems with [[coherent feature]]s, such as [[hurricane]]s, [[thunderstorm]]s, [[fireline]]s, [[squall line]]s, and [[rain front]]s, there is a need to adjust the numerical model state by deforming the state in space (its grid) as well as by  correcting the state amplitudes additively. In Data Assimilation by Field Alignment,<ref name ="Ravela-2007">S. Ravela, K. Emanuel and D. McLaughlin, "Data Assimilation by Field Alignment". Physica(D), Volume 230, Issues 1–2, June 2007, Pages 127–145</ref> Ravela et al. introduce the joint position-amplitude adjustment model using ensembles, and systematically derive a sequential approximation which can be applied to both EnKF and other formulations. Their method does not make the assumption that amplitudes and position errors are independent or jointly Gaussian, as others do. The morphing EnKF<ref name="Beezley-2007-MEK">J. D. Beezley and J. Mandel, ''Morphing ensemble Kalman filters''. Tellus (2008) 60A, 131-140.  [http://www-math.ucdenver.edu/ccm/reports/rep240.pdf report].
 
</ref><ref name="Mandel-2006-PME">J. Mandel and J. D. Beezley, ''Predictor-corrector and morphing ensemble filters for the assimilation of sparse data into high dimensional nonlinear systems''. CCM Report 239, University of Colorado at Denver and Health Sciences Center. [http://www.math.ucdenver.edu/ccm/reports/rep239.pdf report], November 2006. 11th Symposium on Integrated Observing and Assimilation Systems for the Atmosphere, Oceans, and Land Surface (IOAS-AOLS), CD-ROM, Paper 4.12, 87th American Meteorological Society Annual Meeting, San Antonio, TX, January 2007, [http://www.ametsoc.org link].
 
</ref> employs intermediate states, obtained by techniques borrowed from [[image registration]] and [[morphing]], instead of linear combinations of states.
 
EnKFs rely on the Gaussian assumption, though they are of course used in practice for nonlinear problems, where the Gaussian assumption may not be satisfied. Related filters attempting to relax the Gaussian assumption in EnKF while preserving its advantages include filters that fit the state pdf with multiple Gaussian kernels,<ref name="Anderson-1999-MCI">J. L. Anderson and S. L. Anderson, ''A Monte Carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts'', Monthly Weather Review, 127 (1999), pp. 2741--2758.
 
</ref> filters that approximate the state pdf by [[Gaussian mixture]]s,<ref name="Bengtsson-2003-NFE">T. Bengtsson, C. Snyder, and D. Nychka, ''Toward a nonlinear ensemble filter for high dimensional systems'', Journal of Geophysical Research - Atmospheres, 108(D24) (2003), pp. STS 2--1--10. [http://www.image.ucar.edu/pub/nychka/manuscripts/bengtsson.pdf preprint].
 
</ref> a variant of the [[particle filter]] with computation of particle weights by [[density estimation]],<ref name="Mandel-2006-PME"/> and a variant of the particle filter with [[Cauchy distribution|thick tailed]] data pdf to alleviate [[Particle filter#Sampling Importance Resampling (SIR)|particle filter degeneracy]].<ref name="vanLeeuwen-2003-VMF">P. van Leeuwen, ''A variance-minimizing filter for large-scale applications'', Monthly Weather Review, 131 (2003), pp. 2071--2084.</ref>
 
== See also ==
* [[Data assimilation]]
* [[Kalman filter]]
* [[Numerical weather prediction#Ensembles]]
* [[Particle filter]]
* [[Recursive Bayesian estimation]]
 
==References==
{{reflist|2}}
 
==External links==
* [http://www.cmascenter.org/conference/2006/abstracts/zubrow_session7.pdf Ensemble Adjusted Kalman Filter applied to CO measurements] ('''EAKF''')
:<small>''(EAKF-CMAQ: DEVELOPMENT AND INITIAL EVALUATION OF AN ENSEMBLE ADJUSTMENT KALMAN FILTER BASED DATA ASSIMILATION FOR CO)''</small>
* [http://enkf.nersc.no EnKF webpage] ('''EnKF''')
* [http://topaz.nersc.no TOPAZ, real-time forecasting of the North Atlantic ocean and Arctic sea-ice with the EnKF] ('''TOPAZ''')
 
{{DEFAULTSORT:Ensemble Kalman Filter}}
[[Category:Linear filters]]
[[Category:Nonlinear filters]]
[[Category:Bayesian statistics]]
[[Category:Estimation theory]]
[[Category:Monte Carlo methods]]

Latest revision as of 19:46, 16 March 2013

The ensemble Kalman filter (EnKF) is a recursive filter suitable for problems with a large number of variables, such as discretizations of partial differential equations in geophysical models. The EnKF originated as a version of the Kalman filter for large problems (essentially, the covariance matrix is replaced by the sample covariance), and it is now an important data assimilation component of ensemble forecasting. EnKF is related to the particle filter (in this context, a particle is the same thing as ensemble member) but the EnKF makes the assumption that all probability distributions involved are Gaussian; when it is applicable, it is much more efficient than the particle filter.

Introduction

The Ensemble Kalman Filter (EnKF) is a Monte Carlo implementation of the Bayesian update problem: given a probability density function (pdf) of the state of the modeled system (the prior, called often the forecast in geosciences) and the data likelihood, the Bayes theorem is used to obtain the pdf after the data likelihood has been taken into account (the posterior, often called the analysis). This is called a Bayesian update. The Bayesian update is combined with advancing the model in time, incorporating new data from time to time. The original Kalman Filter[1] assumes that all pdfs are Gaussian (the Gaussian assumption) and provides algebraic formulas for the change of the mean and the covariance matrix by the Bayesian update, as well as a formula for advancing the covariance matrix in time provided the system is linear. However, maintaining the covariance matrix is not feasible computationally for high-dimensional systems. For this reason, EnKFs were developed.[2][3] EnKFs represent the distribution of the system state using a collection of state vectors, called an ensemble, and replace the covariance matrix by the sample covariance computed from the ensemble. The ensemble is operated with as if it were a random sample, but the ensemble members are really not independent - the EnKF ties them together. One advantage of EnKFs is that advancing the pdf in time is achieved by simply advancing each member of the ensemble. For a survey of EnKF and related data assimilation techniques, see G. Evensen.[4]

A derivation of the EnKF

The Kalman Filter

Let us review first the Kalman filter. Let denote the -dimensional state vector of a model, and assume that it has Gaussian probability distribution with mean and covariance , i.e., its pdf is

Here and below, means proportional; a pdf is always scaled so that its integral over the whole space is one. This , called the prior, was evolved in time by running the model and now is to be updated to account for new data. It is natural to assume that the error distribution of the data is known; data have to come with an error estimate, otherwise they are meaningless. Here, the data is assumed to have Gaussian pdf with covariance and mean , where is the so-called observation matrix. The covariance matrix describes the estimate of the error of the data; if the random errors in the entries of the data vector are independent, is diagonal and its diagonal entries are the squares of the standard deviation (“error size”) of the error of the corresponding entries of the data vector . The value is what the value of the data would be for the state in the absence of data errors. Then the probability density of the data conditional of the system state , called the data likelihood, is

The pdf of the state and the data likelihood are combined to give the new probability density of the system state conditional on the value of the data (the posterior) by the Bayes theorem,

The data is fixed once it is received, so denote the posterior state by instead of and the posterior pdf by . It can be shown by algebraic manipulations[5] that the posterior pdf is also Gaussian,

with the posterior mean and covariance given by the Kalman update formulas

where

is the so-called Kalman gain matrix.

The Ensemble Kalman Filter

The EnKF is a Monte Carlo approximation of the Kalman filter, which avoids evolving the covariance matrix of the pdf of the state vector . Instead, the pdf is represented by an ensemble

is an matrix whose columns are the ensemble members, and it is called the prior ensemble. Ideally, ensemble members would form a sample from the prior distribution. However, the ensemble members are not in general independent except in the initial ensemble, since every EnKF step ties them together. They are deemed to be approximately independent, and all calculations proceed as if they actually were independent.

Replicate the data into an matrix

so that each column consists of the data vector plus a random vector from the -dimensional normal distribution . If, in addition, the columns of are a sample from the prior probability distribution, then the columns of

form a sample from the posterior probability distribution. [To see this in the scalar case with : Let , and Then . The first sum is the posterior mean, and the second sum, in view of the independence, has a variance , which is the posterior variance. ]

The EnKF is now obtained[6] simply by replacing the state covariance in Kalman gain matrix by the sample covariance computed from the ensemble members (called the ensemble covariance). That is:

Implementation

Basic formulation

Here we follow.[7][8] Suppose the ensemble matrix and the data matrix are as above. The ensemble mean and the covariance are

where

and denotes the matrix of all ones of the indicated size.

The posterior ensemble is then given by

where the perturbed data matrix is as above.

Note that since is a covariance matrix, it is always positive semidefinite and usually positive definite, so the inverse above exists and the formula can be implemented by the Cholesky decomposition.[9] In,[7][8] is replaced by the sample covariance where and the inverse is replaced by a pseudoinverse, computed using the Singular Value Decomposition (SVD) .

Since these formulas are matrix operations with dominant Level 3 operations,[10] they are suitable for efficient implementation using software packages such as LAPACK (on serial and shared memory computers) and ScaLAPACK (on distributed memory computers).[9] Instead of computing the inverse of a matrix and multiplying by it, it is much better (several times cheaper and also more accurate) to compute the Cholesky decomposition of the matrix and treat the multiplication by the inverse as solution of a linear system with many simultaneous right-hand sides.[10]

Observation matrix-free implementation

Since we have replaced the covariance matrix with ensemble covariance, this leads to a simpler formula where ensemble observations are directly used without explicitly specifying the matrix . More specifically, define a function of the form

The function is called the observation function or, in the inverse problems context, the forward operator. The value of is what the value of the data would be for the state assuming the measurement is exact. Then the posterior ensemble can be rewritten as

where

and

with

Consequently, the ensemble update can be computed by evaluating the observation function on each ensemble member once and the matrix does not need to be known explicitly. This formula holds also[9] for an observation function with a fixed offset , which also does not need to be known explicitly. The above formula has been commonly used for a nonlinear observation function , such as the position of a hurricane vortex.[11] In that case, the observation function is essentially approximated by a linear function from its values at ensemble members.

Implementation for a large number of data points

For a large number of data points, the multiplication by becomes a bottleneck. The following alternative formula is advantageous when the number of data points is large (such as when assimilating gridded or pixel data) and the data error covariance matrix is diagonal (which is the case when the data errors are uncorrelated), or cheap to decompose (such as banded due to limited covariance distance). Using the Sherman–Morrison–Woodbury formula[12]

with

gives

which requires only the solution of systems with the matrix (assumed to be cheap) and of a system of size with right-hand sides. See[9] for operation counts.

Further extensions

The EnKF version described here involves randomization of data. For filters without randomization of data, see.[13][14][15]

Since the ensemble covariance is rank deficient (there are many more state variables, typically millions, than the ensemble members, typically less than a hundred), it has large terms for pairs of points that are spatially distant. Since in reality the values of physical fields at distant locations are not that much correlated, the covariance matrix is tapered off artificially based on the distance, which gives rise to localized EnKF algorithms.[16][17] These methods modify the covariance matrix used in the computations and, consequently, the posterior ensemble is no longer made only of linear combinations of the prior ensemble.

For nonlinear problems, EnKF can create posterior ensemble with non-physical states. This can be alleviated by regularization, such as penalization of states with large spatial gradients.[6]

For problems with coherent features, such as hurricanes, thunderstorms, firelines, squall lines, and rain fronts, there is a need to adjust the numerical model state by deforming the state in space (its grid) as well as by correcting the state amplitudes additively. In Data Assimilation by Field Alignment,[18] Ravela et al. introduce the joint position-amplitude adjustment model using ensembles, and systematically derive a sequential approximation which can be applied to both EnKF and other formulations. Their method does not make the assumption that amplitudes and position errors are independent or jointly Gaussian, as others do. The morphing EnKF[19][20] employs intermediate states, obtained by techniques borrowed from image registration and morphing, instead of linear combinations of states.

EnKFs rely on the Gaussian assumption, though they are of course used in practice for nonlinear problems, where the Gaussian assumption may not be satisfied. Related filters attempting to relax the Gaussian assumption in EnKF while preserving its advantages include filters that fit the state pdf with multiple Gaussian kernels,[21] filters that approximate the state pdf by Gaussian mixtures,[22] a variant of the particle filter with computation of particle weights by density estimation,[20] and a variant of the particle filter with thick tailed data pdf to alleviate particle filter degeneracy.[23]

See also

References

43 year old Petroleum Engineer Harry from Deep River, usually spends time with hobbies and interests like renting movies, property developers in singapore new condominium and vehicle racing. Constantly enjoys going to destinations like Camino Real de Tierra Adentro.

External links

(EAKF-CMAQ: DEVELOPMENT AND INITIAL EVALUATION OF AN ENSEMBLE ADJUSTMENT KALMAN FILTER BASED DATA ASSIMILATION FOR CO)
  1. R. E. Kalman, A new approach to linear filtering and prediction problems, Transactions of the ASME -- Journal of Basic Engineering, Series D, 82 (1960), pp. 35--45.
  2. G. Evensen, Sequential data assimilation with nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics, Journal of Geophysical Research, 99 (C5) (1994), pp. 143--162.
  3. P. Houtekamer and H. L. Mitchell, Data assimilation using an ensemble Kalman filter technique, Monthly Weather Review, 126 (1998), pp. 796--811.
  4. G. Evensen, Data assimilation : The ensemble Kalman filter, Springer, Berlin, 2007.
  5. B. D. O. Anderson and J. B. Moore, Optimal filtering, Prentice-Hall, Englewood Cliffs, N.J., 1979.
  6. 6.0 6.1 C. J. Johns and J. Mandel, A two-stage ensemble Kalman filter for smooth data assimilation. Environmental and Ecological Statistics, in print. Special issue, Conference on New Developments of Statistical Analysis in Wildlife, Fisheries, and Ecological Research, Oct 13-16, 2004, Columbia, MI. CCM Report 221, University of Colorado at Denver and Health Sciences Center, 2005. report
  7. 7.0 7.1 G. Burgers, P. J. van Leeuwen, and G. Evensen, Analysis scheme in the ensemble Kalman filter, Monthly Weather Review, 126 (1998), pp. 1719--1724.
  8. 8.0 8.1 G. Evensen, The ensemble Kalman filter: Theoretical formulation and practical implementation, Ocean Dynamics, 53 (2003), pp. 343--367.
  9. 9.0 9.1 9.2 9.3 J. Mandel, Efficient implementation of the ensemble Kalman filter. CCM Report 231, University of Colorado at Denver and Health Sciences Center. link, June 2006.
  10. 10.0 10.1 G. H. Golub and C. F. V. Loan, Matrix Computations, Johns Hopkins Univ. Press, 1989. Second Edition.
  11. Y. Chen and C. Snyder, Assimilating vortex position with an ensemble Kalman filter. Monthly Weather Review, to appear, 2006. preprint.
  12. W. W. Hager, Updating the inverse of a matrix, SIAM Rev., 31 (1989), pp. 221--239.
  13. J. L. Anderson, An ensemble adjustment Kalman filter for data assimilation, Monthly Weather Review, 129 (2001), pp. 2884--2903.
  14. G. Evensen, Sampling strategies and square root analysis schemes for the EnKF, Ocean Dynamics, 54 (2004), pp. 539--560.
  15. M. K. Tippett, J. L. Anderson, C. H. Bishop, T. M. Hamill, and J. S. Whitaker, Ensemble square root filters, Monthly Weather Review, 131 (2003), pp. 1485--1490.
  16. J. L. Anderson, A local least squares framework for ensemble filtering, Monthly Weather Review, 131 (2003), pp. 634--642.
  17. E. Ott, B. R. Hunt, I. Szunyogh, A. V. Zimin, E. J. Kostelich, M. Corazza, E. Kalnay, D. Patil, and J. A. Yorke, A local ensemble Kalman filter for atmospheric data assimilation, Tellus A, 56 (2004), pp. 415--428.
  18. S. Ravela, K. Emanuel and D. McLaughlin, "Data Assimilation by Field Alignment". Physica(D), Volume 230, Issues 1–2, June 2007, Pages 127–145
  19. J. D. Beezley and J. Mandel, Morphing ensemble Kalman filters. Tellus (2008) 60A, 131-140. report.
  20. 20.0 20.1 J. Mandel and J. D. Beezley, Predictor-corrector and morphing ensemble filters for the assimilation of sparse data into high dimensional nonlinear systems. CCM Report 239, University of Colorado at Denver and Health Sciences Center. report, November 2006. 11th Symposium on Integrated Observing and Assimilation Systems for the Atmosphere, Oceans, and Land Surface (IOAS-AOLS), CD-ROM, Paper 4.12, 87th American Meteorological Society Annual Meeting, San Antonio, TX, January 2007, link.
  21. J. L. Anderson and S. L. Anderson, A Monte Carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts, Monthly Weather Review, 127 (1999), pp. 2741--2758.
  22. T. Bengtsson, C. Snyder, and D. Nychka, Toward a nonlinear ensemble filter for high dimensional systems, Journal of Geophysical Research - Atmospheres, 108(D24) (2003), pp. STS 2--1--10. preprint.
  23. P. van Leeuwen, A variance-minimizing filter for large-scale applications, Monthly Weather Review, 131 (2003), pp. 2071--2084.