Workshop on Data Assimilation Oxford-Man Institute, September 24–28, 2012 Large sample central limit theorem for the ensemble Kalman filter (EnKF) Fran¸cois Le Gland INRIA Rennes + IRMAR Val´erie Monbet Universit´e de Rennes 1 + IRMAR supported by ANR project PREVASSEMBLE coordinated by Olivier Talagrand outline • EnKF as particle system with mean–field interaction • identification of the limit • large sample asymptotics of EnKF • toy example • central limit theorem for EnKF 1 (EnKF as particle system with mean–field interactions : 1) non–linear state–space model of a special form Xk = fk (Xk−1 ) + Wk Y k = Hk X k + V k with with Wk ∼ N(0, Qk ) Vk ∼ N(0, Rk ) with non–necessarily Gaussian initial condition X0 ∼ η0 observation noise covariance matrix Rk assumed invertible in such a model, conditional probability distribution (aka Bayesian filter) of hidden state Xk given past observations Y0:k = (Y0 , · · · , Yk ) is not Gaussian ensemble Kalman filter provides a Bayes–free Monte Carlo approach to numerically evaluate the Bayesian filter, as an alternative to particle filters 2 (EnKF as particle system with mean–field interactions : 2) I initialization : initial ensemble (X01,f , · · · , X0N,f ) is simulated as i.i.d. random vectors with probability distribution η0 , i.e. with same statistics as initial condition X0 1,a N,a I prediction (forecast) step : given analysis ensemble (Xk−1 , · · · , Xk−1 ) each member is propagated independently according to (set of decoupled equations) i,a ) + Wki Xki,f = fk (Xk−1 with Wki ∼ N(0, Qk ) notice that i.i.d. random vectors (Wk1 , · · · , WkN ) are simulated here, with same statistics as additive Gaussian noise Wk in original state equation : in particular 1,a N,a (Wk1 , · · · , WkN ) are independent of forecast ensemble (Xk−1 , · · · , Xk−1 ) define empirical covariance matrix PkN,f N 1 ∑ i,f i,f N,f ∗ = (Xk − mN,f ) (X − m k k k ) N i=1 of forecast ensemble (Xk1,f , · · · , XkN,f ) with mN,f k N 1 ∑ i,f = Xk N i=1 3 (EnKF as particle system with mean–field interactions : 3) I correction (analysis) step : given forecast ensemble (Xk1,f , · · · , XkN,f ) each member is updated independently according to (set of equations with mean–field interactions) Xki,a = Xki,f + Kk (PkN,f ) (Yk − Hk Xki,f − Vki ) with Vki ∼ N(0, Rk ) in terms of Kalman gain mapping defined by P 7−→ Kk (P ) = P Hk∗ (Hk P Hk∗ + Rk )−1 for any m × m covariance matrix P , and in terms of empirical covariance matrix PkN,f of forecast ensemble notice that i.i.d. random vectors (Vk1 , · · · , VkN ) are simulated here, with same statistics as additive Gaussian noise Vk in original observation equation : in particular (Vk1 , · · · , VkN ) are independent of forecast ensemble (Xk1,f , · · · , XkN,f ) 4 (EnKF as particle system with mean–field interactions : 4) mean–field interaction : given that Xki,a = Xki,f + Kk (PkN,f ) (Yk − Hk Xki,f − Vki ) each analysis member depends on whole forecast ensemble (Xk1,f , · · · , XkN,f ) through empirical probability distribution µN,f k N 1 ∑ δ i,f = N i=1 Xk actually only through empirical covariance matrix PkN,f of forecast ensemble −→ results in dependent analysis ensemble (Xk1,a , · · · , XkN,a ) 5 (EnKF as particle system with mean–field interactions : 5) question : does ensemble empirical probability distribution converge to Bayesian filter, defined as µ− k (dx) = P[Xk ∈ dx | Y0:k−1 ] and µk (dx) = P[Xk ∈ dx | Y0:k ] i.e. does µN,f k N 1 ∑ δ i,f −→ µ− = k N i=1 Xk and µN,a k N 1 ∑ δ i,a −→ µk = N i=1 Xk hold in some sense, as N ↑ ∞ ? answer in general is negative however, in the linear Gaussian case (i.e. if fk (x) depends linearly on x), then answer is positive and in particular ensemble empirical mean vector converges to Kalman filter, i.e. b N,f X k as N ↑ ∞ N 1 ∑ i,f b− = Xk −→ X k N i=1 and b N,a X k N 1 ∑ i,a bk = Xk −→ X N i=1 6 (EnKF as particle system with mean–field interactions : 6) decoupling approach : to study asymptotic behaviour of empirical probability distributions µN,f k N 1 ∑ = δ i,f N i=1 Xk µN,a k and N 1 ∑ = δ i,a N i=1 Xk of forecast ensemble and analysis ensemble, respectively, approximating i.i.d. random vectors are introduced ¯ i,f = X i,f , i.e. initial set of i.i.d. random vectors coincides exactly with initially X 0 0 initial forecast ensemble these vectors are propagated independently according to (set of fully decoupled equations) ¯ i,f = fk (X ¯ i,a ) + Wki X k k−1 with Wki ∼ N(0, Qk ) and ¯ i,a = X ¯ i,f + Kk (P¯ f ) (Yk − Hk X ¯ i,f − Vki ) X k k k k with Vki ∼ N(0, Rk ) ¯ i,f where P¯kf denotes (true) covariance matrix of i.i.d. random vectors X k 7 (EnKF as particle system with mean–field interactions : 7) heuristics : these i.i.d. random vectors are close (contiguous) to members in ensemble Kalman filter, since they • start from same initial values exactly • use same i.i.d. random vectors (Wk1 , · · · , WkN ) and (Vk1 , · · · , VkN ) exactly, already simulated and used in ensemble Kalman filter esentially a theoretical (not practical) concept • large sample asymptotics is simple to analyze, because of independance • true covariance matrix P¯kf is unknown, hence these i.i.d. random vectors are not computable in practice in contrast, members in ensemble Kalman filter are computable but dependent, because they all contribute to / use empirical covariance matrix PkN,f which results in mean–field interaction outline • EnKF as particle system with mean–field interaction • identification of the limit • large sample asymptotics of EnKF • toy example • central limit theorem for EnKF 8 (identification of the limit : 1) intuition : limiting probability distributions µ ¯fk and µ ¯ak are probability distributions ¯ i,f and X ¯ i,a respectively, and are completely of i.i.d. random vectors X k k characterized by integrals of arbitrary test functions ¯ i,f = X i,f and X i,f ∼ η0 , hence µ I initialization : recall that X ¯f = η0 0 0 0 0 I forecast (expression of µ ¯fk in terms of µ ¯ak−1 ) : recall that ¯ i,f = fk (X ¯ i,a ) + W i X k k k−1 with Wki ∼ N(0, Qk ) a ¯ i,a has probability distribution µ and since X ¯ k−1 (by definition), then k−1 ∫ ¯ i,f )] = E[ϕ(fk (X ¯ i,a ) + Wki ))] ϕ(x′ ) µ ¯fk (dx′ ) = E[ϕ(X k k−1 Rm ∫ = ∫ ϕ(fk (x) + w) pW ¯ak−1 (dx) k (dw) µ Rm Rm | {z } Tk ϕ(x) where pW k (dw) is Gaussian probability distribution with zero mean vector and covariance matrix Qk , i.e. probability distribution of random vector Wki 9 (identification of the limit : 2) I analysis (expression of µ ¯ak in terms of µ ¯fk ) : recall that ¯ i,a = X ¯ i,f + Kk (P¯ f ) (Yk − Hk X ¯ i,f − Vki ) X k k k k with Vki ∼ N(0, Rk ) sufficient conditions on drift function fk can be given, under which µ ¯fk has finite second order moments, in which case covariance matrix P¯kf is finite f ¯ i,f has probability distribution µ ¯ and since X k (by definition), then k ∫ ¯ i,a )] = E[ϕ(X ¯ i,f + Kk (P¯ f ) (Yk − Hk X ¯ i,f − V i ))] ϕ(x′ ) µ ¯ak (dx′ ) = E[ϕ(X k k k k k Rm ∫ = ∫ ϕ(x + Kk (P¯kf ) (Yk − Hk x − v)) qkV (v) dv µ ¯fk (dx) Rm Rd | {z } TkKF (¯ µfk ) ϕ(x) where qkV (v) is Gaussian density with zero mean vector and invertible covariance matrix Rk , i.e. probability density of random vector Vki 10 (identification of the limit : 3) on the other hand, Bayesian filter, defined as µ− k (dx) = P[Xk ∈ dx | Y0:k−1 ] satisfies recurrent relation ∫ ∫ ′ ϕ(x′ ) µ− (dx )= k and µk (dx) = P[Xk ∈ dx | Y0:k ] ∫ ϕ(fk (x) + w) pW k (dw) µk−1 (dx) Rm Rm | {z } Tk ϕ(x) Rm and (Bayes rule) ∫ Rm ∫ ϕ(x′ ) µk (dx′ ) = Rm ∫ ′ ϕ(x′ ) qkV (Yk − Hk x′ ) µ− (dx ) k Rm ′ qkV (Yk − Hk x′ ) µ− k (dx ) with initial condition µ− 0 = η0 clearly, limiting probability distributions of forecast / analysis ensemble do not a ¯fk ̸= µ− and µ ¯ coincide with Bayesian predictor / filter, i.e. µ k ̸= µk , except in the k linear Gaussian case outline • EnKF as particle system with mean–field interaction • identification of the limit • large sample asymptotics of EnKF • toy example • central limit theorem for EnKF 11 (large sample asymptotics of EnKF : 1) indeed, intuition is correct : ensemble empirical probability distributions µN,• k N 1 ∑ = δ i,• N i=1 Xk do converge (in some sense) as N ↑ ∞ to the probability distribution µ ¯•k of i.i.d. ¯ i,• (hence, not to the Bayesian filter) random vectors X k reference F. Le Gland, V. Monbet and Vu–Duc Tran Large sample asymptotics for the ensemble Kalman filter , chapter 22 in The Oxford Handbook of Nonlinear Filtering, 2011 12 (large sample asymptotics of EnKF : 2) Theorem (law of large numbers) under mild assumptions on drift function fk and on test function ϕ ∫ N ∑ 1 ϕ(Xki,• ) −→ ϕ(x) µ ¯•k (dx) N i=1 Rm in probability as N ↑ ∞ Theorem (Lp –convergence and rate of convergence) under mild assumptions on drift function fk and on test function ϕ, and provided initial condition X0 has finite moments of any order p sup N ≥1 for any order p √ ∫ N ∑ 1 N ( E| ϕ(Xki,• ) − ϕ(x) µ ¯•k (dx) |p )1/p < ∞ N i=1 Rm 13 (large sample asymptotics of EnKF : 3) to summarize : ensemble Kalman filter • gain matrix depends on empirical covariance matrix • ensemble empirical probability distribution converges to the wrong limit (different from Bayesian filter), except for linear Gaussian model √ • rate of convergence 1/ N vs. (any brand of) particle filter • weighted empirical probability distribution of particle system converges to the correct limit (Bayesian filter) √ • rate of convergence 1/ N , central limit theorem question : is there any advantage to use ensemble Kalman filter ? idea : prove central limit theorem (and compare asymptotic error variances) outline • EnKF as particle system with mean–field interaction • identification of the limit • large sample asymptotics of EnKF • toy example • central limit theorem for EnKF 14 (toy example : 1) toy example linear Gaussian system the target distribution, i.e. the Bayesian filter, is known explicitly as a Gaussian distribution, with mean and covariance provided by the Kalman filter hidden state √ Xk = a Xk−1 + 1 − a2 Wk with Wk ∼ N(0, σ 2 ) initial condition X0 ∼ N(0, σ 2 ) so that stationarity holds observations Yk = X k + Vk Vk ∼ N(0, s2 ) with numerical values a σ s 0.5 1 1 15 (toy example : 2) true state and Kalman filter 2.5 true state Kalman filter 2 1.5 1 0.5 0 −0.5 −1 −1.5 −2 −2.5 0 10 20 30 40 50 60 70 80 90 100 16 (toy example : 3) true state, Kalman filter and EnKF empirical mean : 10000 elements 2.5 true state Kalman filter EnKF 2 1.5 1 0.5 0 −0.5 −1 −1.5 −2 −2.5 0 10 20 30 40 50 60 70 80 90 b N,a with N = 10000 members EnKF empirical mean vector X k 100 17 (toy example : 4) EnKF histogram + Kalman filter density : time 100 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 −3 −2 −1 0 EnKF histogram with N = 10000 members 1 2 3 4 18 (toy example : 5) conclusion : not only does the EnKF empirical mean vector b N,a X k N 1 ∑ i,a = X N i=1 k bk , but more generally the EnKF empirical converge to the Kalman filter X probability distribution N ∑ 1 µN,a = δ i,a k N i=1 Xk converges to the Gaussian distribution with moments given by the Kalman filter next different question : how fast does the empirical mean vector converge to the Kalman filter, e.g. is the normalized difference √ N ∑ 1 i,a b N,a − X bk ) = √ b N (X (X k k − Xk ) N i=1 asymptotically normally distributed and how to compute the asymptotic variance ? 19 (toy example : 6) toy example (continued) numerical simulations : for EnKF / bootstrap particle filter / particle filter with optimal importance distribution • M Monte Carlo runs • each Monte Carlo run evaluates one ensemble / particle average, based on N members / particles and compares this average with the (known) limit • histogram of the M normalized differences is shown same toy example : stationary linear Gaussian system 20 (toy example : 7) histogram of normalized error of EnKF empirical mean : 10000 Monte Carlo runs 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 −3 −2 −1 0 1 √ 2 3 b N,a − X bk ) for N = 1000 histogram of EnKF normalized differences N (X k members and M = 10000 Monte Carlo runs 21 (toy example : 8) histogram of normalized error of PF empirical mean : 10000 Monte Carlo runs 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 −3 −2 −1 0 1 2 histogram of (bootstrap) particle filter normalized differences N = 1000 particles and M = 10000 Monte Carlo runs 3 √ ¯ N − mk ) for N (X k 22 (toy example : 9) histogram of normalized error of PF empirical mean : 10000 Monte Carlo runs 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 −3 −2 −1 0 1 2 histogram of (optimal) particle filter normalized differences N = 1000 particles and M = 10000 Monte Carlo runs 3 √ ¯ N − mk ) for N (X k 23 (toy example : 10) conclusion (as drawn from these simulations) : in terms of speed of convergence (a smaller asymptotic variance means a faster convergence) PF with optimal importance distribution ≫ bootstrap PF ≫ EnKF outline • EnKF as particle system with mean–field interaction • identification of the limit • large sample asymptotics of EnKF • toy example • central limit theorem for EnKF 24 (central limit theorem for EnKF : 1) Theorem (central limit theorem) under mild assumptions on drift function fk and on test function ϕ 1 √ N N ∑ ∫ [ϕ(Xki,• ) − i=1 Rm ϕ(x) µ ¯•k (dx) ] =⇒ N(0, v¯k• (ϕ)) in distribution as N ↑ ∞, with (more or less explicit) expression for asymptotic variance v¯k• (ϕ) beyond the qualitative statement • recurrence relations for the asymptotic variance ? • practical computations ? 25 (central limit theorem for EnKF : 2) because of the recursive nature of the ensemble Kalman filter, it seems natural to prove the CLT by induction, hence to rely on the following strategy, already used in K¨ unsch (Annals of Statistics, 2005) Lemma if ′ • conditionally w.r.t. FN , the r.v. ZN converges in distribution to a Gaussian r.v. with zero mean and variance V ′ , in the sense that for any fixed u ′ E[exp{j u ZN } | FN ] −→ exp{− 12 u2 V ′ } in probability, and in L1 by the Lebesgue dominated convergence thorem, ′′ • the r.v. ZN is measurable w.r.t. FN , and converges in distribution to a Gaussian r.v. with zero mean and variance V ′′ , i.e. for any fixed u ′′ E[exp{j u ZN }] −→ exp{− 12 u2 V ′′ } ′ ′′ then the r.v. ZN = ZN + ZN converges in distribution to a Gaussian r.v. with zero mean and variance V = V ′ + V ′′ , as N ↑ ∞ 26 (central limit theorem for EnKF : 3) I initialization : recall that X0i,f ∼ η0 and µ ¯f0 = η0 , hence 1 √ N N ∑ i=1 ∫ [ϕ(X0i,f ) − Rm ϕ(x) µ ¯f0 (dx) ] =⇒ N(0, v¯0f (ϕ)) in distribution as N ↑ ∞, with asymptotic variance ∫ ∫ v¯0f (ϕ) = var(ϕ, η0 ) = |ϕ(x)|2 η0 (dx) − | Rm Rm ϕ(x) η0 (dx) |2 27 (central limit theorem for EnKF : 4) I forecast step : recall that µ ¯fk = µ ¯ak−1 Tk , where ∫ Tk ϕ(x) = ϕ(fk (x) + w) pW k (dw) Rm Proposition asymptotic variance of forecast approximation a v¯kf (ϕ) = v¯k−1 (Tk ϕ) + σk2,f (ϕ) in terms of • asymptotic variance of analysis approximation at previous step, evaluated for a transformed test function • asymptotic Monte Carlo variance ∫ ∫ Tk |ϕ|2 (x) µ ¯ak−1 (dx) − σk2,f (ϕ) = Rm Rm |Tk ϕ(x)|2 µ ¯ak−1 (dx) 28 (central limit theorem for EnKF : 5) hint : ZN ∫ N 1 ∑ [ϕ(Xki,f ) − ϕ(x′ ) µ ¯fk (dx′ ) ] = √ N i=1 Rm 1 = √ N N ∑ ∫ i,a [ϕ(fk (Xk−1 ) + Wki ) − i=1 Rm ϕ(x′ ) µ ¯fk (dx′ ) ] N 1 ∑ i,a i,a [ ϕ(fk (Xk−1 ) + Wki ) − Tk ϕ(Xk−1 )] = √ N i=1 1 +√ N ′ ′′ = ZN + ZN N ∑ i=1 ∫ i,a [ Tk ϕ(Xk−1 )− Rm Tk ϕ(x) µ ¯ak−1 (dx) ] 29 (central limit theorem for EnKF : 6) conditionnally w.r.t. the analysis ensemble at previous stage, the random variables i,a i,a eik = ϕ(fk (Xk−1 ) + Wki ) − Tk ϕ(Xk−1 ) are independent (not identically distributed), with mean zero and variance ∫ ∫ i,a i,a 2 sik = |ϕ(fk (Xk−1 ) + w)|2 pW ϕ(fk (Xk−1 ) + w) pW k (dw) − | k (dw)| Rm Rm i,a i,a = Tk |ϕ|2 (Xk−1 ) − |Tk ϕ(Xk−1 )|2 respectively, for i = 1 · · · N , and it follows from our law of large numbers that N N 1 ∑ i 1 ∑ i,a i,a sk = [ Tk |ϕ|2 (Xk−1 ) − |Tk ϕ(Xk−1 )|2 ] N i=1 N i=1 ∫ −→ σk2,f (ϕ) = Rm ∫ Tk |ϕ|2 (x) µ ¯ak−1 (dx) − in probability and in L1 as N ↑ ∞ Rm |Tk ϕ(x)|2 µ ¯ak−1 (dx) 30 (central limit theorem for EnKF : 7) therefore N ∑ u N,a N,a ′ } | Fk−1 ] = E[exp{j √ eik } | Fk−1 ] −→ exp{− 12 u2 σk2,f (ϕ)} E[exp{j u ZN N i=1 in L1 as N ↑ ∞ moreover, it follows from the induction assumption that 1 ′′ ZN =√ N N ∑ i=1 ∫ i,a [ Tk ϕ(Xk−1 )− Rm a Tk ϕ(x) µ ¯ak−1 (dx) ] =⇒ N(0, v¯k−1 (Tk ϕ)) hence ′′ a E[exp{j u ZN }] −→ exp{− 12 u2 v¯k−1 (Tk ϕ)} as N ↑ ∞ . . . this concludes the proof 2 31 (central limit theorem for EnKF : 8) I analysis step : recall that µ ¯ak = TkKF (¯ µfk ) µ ¯fk , where ∫ TkKF (¯ µfk ) ϕ(x) = ϕ(x + Kk (P¯kf ) (Yk − Hk x − v)) qkV (v) dv Rd introduce new transform f f ∗ f KF f KF f , ϕ) (x − m ¯ ) M (¯ µ ) ϕ(x) + (x − m ¯ ) ϕ(x) = T (¯ µ QKF (¯ µ k k k k) k k k k where the matrix MkKF (¯ µfk , ϕ) is defined as f f ¯ MkKF (¯ µfk , ϕ) = Hk∗ (Hk P¯kf Hk∗ + Rk )−1 LKF (¯ µ , ϕ) (I − K ( P k k k k ) Hk ) f and where the matrix LKF (¯ µ k k , ϕ) is defined as ∫ ∫ f f ′ ¯ (Y − H x − v) ϕ (x + K ( P LKF (¯ µ , ϕ) = k k k k k ) (Yk − Hk x − v)) k Rm Rd qkV (v) dv µ ¯fk (dx) . . . motivation will come later 32 (central limit theorem for EnKF : 9) Proposition asymptotic variance of analysis approximation f 2,a v¯ka (ϕ) = v¯kf (QKF (¯ µ ) ϕ) + σ k k k (ϕ) in terms of • asymptotic variance of analysis approximation at previous step, evaluated for a transformed test function • asymptotic Monte Carlo variance ∫ ∫ σk2,a (ϕ) = TkKF (¯ µfk ) |ϕ|2 (x) µ ¯fk (dx) − Rm Rm |TkKF (¯ µfk ) ϕ(x)|2 µ ¯fk (dx) 33 (central limit theorem for EnKF : 10) hint : ZN ∫ N 1 ∑ ϕ(x′ ) µ ¯ak (dx′ ) ] = √ [ϕ(Xki,a ) − N i=1 Rm ∫ N 1 ∑ = √ [ϕ(Xki,f + Kk (PkN,f ) (Yk − Hk Xki,f − Vki )) − ϕ(x′ ) µ ¯ak (dx′ ) ] N i=1 Rm N 1 ∑ i,f ) ϕ(X [ ϕ(Xki,f + Kk (PkN,f ) (Yk − Hk Xki,f − Vki )) − TkKF (µN,f = √ k )] k N i=1 N 1 ∑ KF N,f +√ [ Tk (µk ) ϕ(Xki,f ) − TkKF (¯ µfk ) ϕ(Xki,f ) ] N i=1 1 +√ N N ∑ i=1 ′ ′′ ′′′ = ZN + ZN + ZN ∫ µfk ) ϕ(Xki,f ) − [ TkKF (¯ Rm TkKF (¯ µfk ) ϕ(x) µ ¯fk (dx) ] 34 (central limit theorem for EnKF : 11) conditionnally w.r.t. the forecast ensemble at current stage, the random variables eik = ϕ(Xki,f + Kk (PkN ) (Yk − Hk Xki,f − Vki )) are independent (not identically distributed), with mean zero and variance ∫ sik = |ϕ(Xki,f + Kk (PkN ) (Yk − Hk Xki,f − v)) |2 qkV (v) dv Rm ∫ −| Rm ϕ(Xki,f + Kk (PkN ) (Yk − Hk Xki,f − v)) qkV (v) dv |2 i,f i,f 2 2 KF N,f = TkKF (µN,f ) |ϕ| (X ) − |T (µ ) ϕ(X k k k k k )| respectively, for i = 1 · · · N , and it follows from our law of large numbers that N N 1 ∑ i 1 ∑ KF N,f i,f 2 sk = [ Tk (µk ) |ϕ|2 (Xki,f ) − |TkKF (µN,f ) ϕ(X k k )| ] N i=1 N i=1 ∫ ∫ −→ σk2,a (ϕ) = TkKF (¯ µfk ) |ϕ|2 (x) µ ¯fk (dx) − |TkKF (¯ µfk ) ϕ(x)|2 µ ¯fk (dx) Rm in probability and in L1 as N ↑ ∞ Rm 35 (central limit theorem for EnKF : 12) therefore N ∑ u ′ } | FkN,f ] = E[exp{j √ eik } | FkN,f ] −→ exp{− 12 u2 σk2,a (ϕ)} E[exp{j u ZN N i=1 in L1 as N ↑ ∞ ′′′ refrain from jumping directly to the last term ZN ... ′′ . . . indeed, more is still to come from the second term ZN 36 (central limit theorem for EnKF : 13) contribution of nonlinear transformation : Taylor expansion ′′ ZN N 1 ∑ KF N,f = √ [ Tk (µk ) ϕ(Xki,f ) − TkKF (¯ µfk ) ϕ(Xki,f ) ] N i=1 N ∫ 1 ∑ [ ϕ(x + Kk (PkN,f ) (Yk − Hk Xki,f − v)) = √ N i=1 Rd − ϕ(x + Kk (P¯kf ) (Yk − Hk Xki,f − v)) ] qkV (v) dv N ∫ 1 ∑ = √ ϕ′ (Xki,f + Kk (P¯kf ) (Yk − Hk Xki,f − v)) N i=1 Rd (I − Kk (P¯kf ) Hk ) (PkN,f − P¯kf ) Hk∗ (Hk P¯kf Hk∗ + Rk )−1 (Yk − Hk Xki,f − v) qkV (v) dv + εN k in terms of the derivative of the Kalman gain mapping P 7−→ Kk (P ) 37 (central limit theorem for EnKF : 14) further approximation : law of large numbers N ∫ ∑ 1 ′′ ZN = ϕ′ (Xki,f + Kk (P¯kf ) (Yk − Hk Xki,f − v)) N i=1 Rd (I − Kk (P¯kf ) Hk ) √ N (PkN,f − P¯kf ) Hk∗ (Hk P¯kf Hk∗ + Rk )−1 (Yk − Hk Xki,f − v) qkV (v) dv + εN k N ∫ ∑ √ 1 i,f urow (Xk , v) N (PkN,f − P¯kf ) ucol (Xki,f , v) qkV (v) dv + εN = k N i=1 Rd N 1 ∑ = √ [ (Xki,f − m ¯ fk )∗ MkKF (¯ µfk , ϕ) (Xki,f − m ¯ fk ) N i=1 ∫ (x − m ¯ fk )∗ MkKF (¯ µfk , ϕ) (x − m ¯ fk ) µ ¯fk (dx) ] + εN − k Rm 38 (central limit theorem for EnKF : 15) hint for the last identity : relies on the following argument N ∫ √ 1 ∑ i,f urow (Xk , v) N (PkN,f − P¯kf ) ucol (Xki,f , v) qkV (v) dv N i=1 Rd = trace[ √ N ∫ ∑ 1 ucol (Xki,f , v) urow (Xki,f , v) qkV (v) dv ] − P¯kf ) N i=1 Rd N (PkN,f √ ≈ trace[ N (PkN,f − P¯kf ) Mk ] ∫ with Mk = ∫ Rm Rd ucol (x, v) urow (x, v) qkV (v) dv µ ¯fk (dx) and on trace[ √ N (PkN,f N ∑ 1 − P¯kf ) Mk ] ≈ √ [ (Xki,f − m ¯ fk )∗ Mk (Xki,f − m ¯ fk ) N i=1 ∫ − (x − m ¯ fk )∗ Mk (x − m ¯ fk ) µ ¯fk (dx) ] Rm 39 (central limit theorem for EnKF : 16) upon identifying urow (x, v) = ϕ′ (x + Kk (P¯kf ) (Yk − Hk x − v)) (I − Kk (P¯kf ) Hk ) and ucol (x, v) = Hk∗ (Hk P¯kf Hk∗ + Rk )−1 (Yk − Hk x − v) this yields ∫ MkKF (¯ µfk , ϕ) = ∫ Rm Rd ucol (x, v) urow (x, v) qkV (v) dv µ ¯fk (dx) f f ¯ = Hk∗ (Hk P¯kf Hk∗ + Rk )−1 LKF (¯ µ , ϕ) (I − K ( P k k k k ) Hk ) where LKF µfk , ϕ) = k (¯ ∫ ∫ Rm Rd (Yk − Hk x − v) ϕ′ (x + Kk (P¯kf ) (Yk − Hk x − v)) qkV (v) dv µ ¯fk (dx) 40 (central limit theorem for EnKF : 17) therefore ′′ ′′′ + ZN ZN ∫ N 1 ∑ KF f f f N = √ [ Qk (¯ µk ) ϕ(Xki,f ) − QKF (¯ µ ) ϕ(x) µ ¯ (dx) ] + ε k k k k N i=1 Rm where new transform (now better motivated) is defined by f f ∗ f KF f KF f QKF (¯ µ ) ϕ(x) = T (¯ µ ) ϕ(x) + (x − m ¯ ) M (¯ µ , ϕ) (x − m ¯ k k k k k k k k) and where εN k converges to zero in probability as N ↑ ∞ 41 (central limit theorem for EnKF : 18) it follows from the induction assumption that ′′ ′′′ ZN + ZN ∫ N 1 ∑ KF f f f N QKF =√ [ Qk (¯ µk ) ϕ(Xki,f ) − (¯ µ ) ϕ(x) µ ¯ (dx) ] + ε k k k k N i=1 Rm f a =⇒ N(0, v¯k−1 (QKF (¯ µ k k ) ϕ)) hence f ′′ ′′′ a E[exp{j u (ZN + ZN )}] −→ exp{− 12 u2 v¯k−1 (QKF (¯ µ k k ) ϕ)} as N ↑ ∞ recall that N ∑ u ′ E[exp{j u ZN } | FkN,f ] = E[exp{j √ eik } | FkN,f ] −→ exp{− 12 u2 σk2,a (ϕ)} N i=1 in L1 as N ↑ ∞ . . . this concludes the proof 2 42 (central limit theorem for EnKF : 19) I practical computations : iterating the recurrence relations f 2,a v¯ka (ϕ) = v¯kf (QKF (¯ µ ) ϕ) + σ k k k (ϕ) and a v¯kf (ϕ) = v¯k−1 (Tk ϕ) + σk2,f (ϕ) yields f 2,a v¯ka (ϕ) = v¯kf (QKF (¯ µ ) ϕ) + σ k k k (ϕ) 2,f f 2,a a KF f = v¯k−1 (Tk QKF ϕ) + σ (Q (¯ µ ) (¯ µ ) ϕ) + σ k k k k k k (ϕ) | {z } | {z } σk2 (ϕ) RkKF with initialization f 2,a 2,a KF (¯ µ ) ϕ) + σ (ϕ) = var(Q (η ) ϕ, η ) + σ v¯0a (ϕ) = v¯0f (QKF 0 0 0 0 0 0 0 (ϕ) 43 (central limit theorem for EnKF : 20) a v¯ka (ϕ) = v¯k−1 (RkKF ϕ) + σk2 (ϕ) a a KF 2 v¯k−1 (RkKF ϕ) = v¯k−2 (Rk−1 RkKF ϕ) + σk−1 (RkKF ϕ) .. . KF a KF v¯la (Rl+1 · · · RkKF ϕ) = v¯l−1 (RlKF · · · RkKF ϕ) + σl2 (Rl+1 · · · RkKF ϕ) .. . v¯1a (R2KF · · · RkKF ϕ) = v¯0a (R1KF · · · RkKF ϕ) + σ12 (R2KF · · · RkKF ϕ) hence v¯ka (ϕ) = v¯0a (R1KF · · · RkKF ϕ) + k ∑ KF · · · RkKF ϕ) σl2 (Rl+1 l=1 in terms of backward–propagated functions KF KF Rl+1:k ϕ = Rl+1 · · · RkKF ϕ 44 (central limit theorem for EnKF : 21) further simplifications occur in the special case of • linear (and quadratic) test functions ϕ • linear drift function fk indeed • forward–propagated distributions µ ¯fk and µ ¯ak are Gaussian distributions with moments given by the Kalman filter KF • backward–propagated functions Rl+1 · · · RkKF ϕ remain quadratic at all steps and explicit calculations can be obtained forthcoming event International Conference on Ensemble Methods in Geophysical Sciences Toulouse, November 11-16, 2012 http://www.meteo.fr/cic/meetings/2012/ensemble.conference/ organized within the ANR project PREVASSEMBLE
© Copyright 2024