 Review
 Open Access
The Ensemble Kalman filter: a signal processing perspective
 Michael Roth^{1}Email authorView ORCID ID profile,
 Gustaf Hendeby^{1},
 Carsten Fritsche^{1} and
 Fredrik Gustafsson^{1}
https://doi.org/10.1186/s136340170492x
© The Author(s) 2017
 Received: 28 February 2017
 Accepted: 19 July 2017
 Published: 3 August 2017
Abstract
The ensemble Kalman filter (EnKF) is a Monte Carlobased implementation of the Kalman filter (KF) for extremely highdimensional, possibly nonlinear, and nonGaussian state estimation problems. Its ability to handle state dimensions in the order of millions has made the EnKF a popular algorithm in different geoscientific disciplines. Despite a similarly vital need for scalable algorithms in signal processing, e.g., to make sense of the ever increasing amount of sensor data, the EnKF is hardly discussed in our field.
This selfcontained review is aimed at signal processing researchers and provides all the knowledge to get started with the EnKF. The algorithm is derived in a KF framework, without the often encountered geoscientific terminology. Algorithmic challenges and required extensions of the EnKF are provided, as well as relations to sigma point KF and particle filters. The relevant EnKF literature is summarized in an extensive survey and unique simulation examples, including popular benchmark problems, complement the theory with practical insights. The signal processing perspective highlights new directions of research and facilitates the exchange of potentially beneficial ideas, both for the EnKF and highdimensional nonlinear and nonGaussian filtering in general.
1 Introduction
Numerical weather prediction [1] is an extremely highdimensional geoscientific state estimation problem. The state x comprises physical quantities (temperature, wind speed, air pressure, etc.) at many spatially distributed grid points, which often yields a state dimension n in the order of millions. Consequently, the Kalman filter (KF) [2, 3] or its nonlinear extensions [4, 5] that require the storage and processing of n×n covariance matrices cannot be applied directly. It is wellknown that the application of particle filters [6, 7] is not feasible either. In contrast, the ensemble Kalman filter [8, 9] (EnKF) was specifically developed as algorithm for highdimensional n.

is a randomsampling implementation of the KF;

reduces the computational complexity of the KF by propagating an ensemble of N<n state realizations;

can be applied to nonlinear statespace models without the need to compute Jacobian matrices;

can be applied to continuoustime as well as discretetime state transition functions;

can be applied to nonGaussian noise densities;

is simple to implement;

does not converge to the Bayesian filtering solution for N→∞ in general;

often requires extra measures to work in practice.
Also in the field of stochastic signal processing (SP) and Bayesian state estimation, highdimensional problems become more and more relevant. Examples include SLAM [10] where x contains an increasing number of landmark positions, or extended target tracking [11, 12] where x can contain many parameters to describe the shape of the target. Furthermore, scalable SP algorithms are required to make sense of the ever increasing amount of data from sensors in everyday devices.
EnKF approaches hardly appear in the relevant SP journals, though. In contrast, vivid theoretical development is documented in geoscientific journals under the umbrella term data assimilation (DA) [1]. Hence, a relevant SP problem is being addressed with only little participation from the SP community. Conversely, much of the DA literature makes little reference to relevant SP contributions. It is our intention to bridge this interesting gap.
There is further overlap that motivates for a closer investigation of the EnKF. First, the basic EnKF [9] can be applied to nonlinear and nonGaussian statespace models because it is entirely sampling based. In fact, the state evolution in geoscientific applications is typically governed by large nonlinear black box prediction models derived from partial differential equations. Furthermore, satellite measurements in weather applications are nonlinearly related to the states [1]. Hence, the EnKF has long been investigated as nonlinear filter. Second, the EnKF literature contains so called localization methods [13, 14] to systematically approach highdimensional problems by only acting on a part of the state vector in each measurement update. These ideas can be directly transferred to sigma point filters [5]. Third, the EnKF offers several interesting opportunities to apply SP techniques, e.g., via the application of bootstrap or regularization methods in the EnKF gain computation.
The contributions of this paper aim at making the EnKF more accessible to SP researchers. We provide a concise derivation of the EnKF based on the KF. A literature review highlights important EnKF papers with their respective contributions and facilitates easier access to the extensive and rapidly developing DA literature on the EnKF. Moreover, we put the EnKF in context with popular SP algorithms such as sigma point filters [4, 5] and the particle filter [6, 7]. Our presentation forms a solid basis for further developments and the transfer of beneficial ideas and techniques between the fields of SP and DA.
The structure of the paper is as follows. After an extensive literature review in Section 2, the EnKF is developed from the KF in Section 3. Algorithmic properties and challenges of the EnKF and the available approaches to face them are discussed in Sections 4 and 5, respectively. Relations to other filtering algorithms are discussed in Section 6. The theoretical considerations are followed by numerical simulations in Section 7 and some concluding remarks in Section 8.
2 Review
The following literature review provides important landmarks for the EnKF novice.
Statespace models and the filtering problem have been investigated since the 1960s. Early results include the Kalman filter (KF) [2] as algorithm for linear systems and the Bayesian filtering equations [15] as theoretical solution for nonlinear and nonGaussian systems. Because the latter approach cannot be implemented in general, approximate filtering algorithms are required. With a leap in computing capacity, the 1990s saw major developments. The samplingbased sigma point Kalman filters [4, 5] started to appear. Furthermore, particle filters [6, 7] were developed to approximately implement the Bayesian filtering equations through sequential importance sampling.
The first EnKF [8] was proposed in a geoscientific journal in 1994 and introduced the idea of propagating ensembles to mimic the KF. A systematic error that resulted in an underestimated uncertainty was later corrected by processing “perturbed measurements.” This randomization is well motivated in [9] but also used in [13].
Interestingly, [8] remains the most cited EnKF paper^{1}, followed by the overview article [16] and the monograph [17] by the same author. Other insightful overviews from a geoscientific perspective are [18, 19]. Many practical aspects of operational EnKF for weather prediction and reanalysis are described in [19–21]. Whereas the aforementioned papers were mostly published in geoscientific outlets, a special issue of the IEEE Control Systems Magazine appeared with review articles [22–24] and an EnKF case study [25]. Still, the above material was written by EnKF researchers with a geoscientific focus and in the applicationspecific terminology. Furthermore, references to the recent SP literature and other nonlinear KF variants [5] are scarce.
Some attention has been devoted to the EnKF also beyond the geosciences. Convergence properties for N→∞ have been established using different theoretical analyses of the EnKF [26–28]. Statistical perspectives are provided in the thesis [29] and the review [30]. A recommended SP view that connects the EnKF with Bayesian filtering and particle methods, including convergence results for nonlinear systems, is [31]. Examples of the EnKF as tool for tomographic imaging and target tracking are described in [32] and [33], respectively. Brief introductory papers that connect the EnKF with more established SP algorithms include [34] and [35]. The latter also served as basis for this article.
The majority of EnKF advances are still documented in geoscientific publications. Notable contributions include deterministic EnKF that avoid the randomization of [9] and propagate an ensemble of deviations from the ensemble mean [16, 36–38]. Their common basis as square root EnKF and the relation to square root KF [3] is discussed in [39]. The computational advantages in highdimensional EnKF with small ensembles (N≪n) come at the price of adverse effects, including the risk of filter divergence. The often encountered underestimation of uncertainty can be counteracted with covariance inflation [40]. A scheme with two EnKF in parallel that provide each other with gain matrices to reduce unwanted “inbreeding” has been suggested in [13]. The benefit of such a double EnKF is, however, debated [38, 41]. The lowrank approximation of covariance matrices can yield spurious correlations between supposedly uncorrelated state components and measurements. Localization techniques such as local measurement updates [13, 16, 42] or covariance tapering [14, 43] let the measurement only affect a part of the state vector. In other words, localization effectively reduces the dimension of each measurement update. Inflation and localization are essential components of operational EnKF [19]. Smoothing algorithms based on the EnKF are discussed in [17] and, more recently, [44]. Approaches that combine variational DA techniques [1] with the EnKF include [45, 46]. A list of further advances in the geoscientific literature is provided in the appendix of [17].
An interesting development for SP researchers is the reconsideration of particle filters (PF) for highdimensional geoscientific problems, with seemingly little reference to SP literature. An early example is [47]. The wellknown challenges, mostly related to the problem of importance sampling in high dimensions, are reviewed in [48, 49]. Several recent approaches [50–52] were successfully tested on a popular EnKF benchmark problem [53] that is also investigated in Section 7.4. Combinations of the EnKF with the deterministic sampling of sigma point filters [5] are given in [54] and [55]. However, the benefit of the unscented transformation [5, 56] in [55] is debated in [57]. Ideas to combine the EnKF with Gaussian mixture approaches are given in [58–60].
3 A signal processing introduction to the ensemble Kalman filter
with the ndimensional state x and the mdimensional measurement y. The initial state x _{0}, the process noise v _{ k }, and the measurement noise e _{ k } are assumed to be independent and described by \(\mathrm {E}(x_{0})=\hat {x}_{0}\), E(v _{ k })=0, E(e _{ k })=0, cov(x _{0})=P _{0}, cov(v _{ k })=Q, and cov(e _{ k })=R. In the Gaussian case, these moments completely characterize the distributions of x _{0}, v _{ k }, and e _{ k }.
More general noise and initial state distributions can, for example, be characterized by probability density functions p(x _{0}), p(v _{ k }), and p(e _{ k }).
Both (1) and (2) can be timevarying but the time indices for functions and matrices are omitted for convenience.
3.1 A brief Kalman filter review
The KF is an optimal linear filter [3] for (1) that propagates state estimates \(\hat {x}_{kk}\) and covariance matrices P _{ kk }.
with a gain matrix K _{ k }. Here, (5b) resembles a deterministic observer and only requires all eigenvalues of (I−K _{ k } H) inside the unit circle to obtain a stable filter.
where M _{ k } is the crosscovariance between the state and output predictions. Alternatives to the covariance update (5c) exist, but the shown Joseph form [3] will simplify the derivation of the EnKF. Furthermore, it is valid for all gain matrices K _{ k } beyond (6) and numerically wellbehaved. It should be noted that it is numerically advisable to obtain K _{ k } by solving K _{ k } S _{ k }=M _{ k } rather than explicitly computing \(S_{k}^{1}\) [61].
3.2 The ensemble idea
Reduced computational complexity is achieved because the explicit computation of \({\bar {P}}_{kk}\) is avoided in the EnKF recursion. Of course, this reduction comes at the price of a lowrank approximation in (7b) that entails some negative effects and requires extra measures.
is an ensemble of deviations from \({\bar {x}}_{kk}\), sometimes called ensemble anomalies [17]. The matrix multiplication in (9) provides a compact way to write the anomalies but is not the most efficient way to compute them.
3.3 The EnKF time update
where we generalized f to act on the columns of its input matrices. Apparently, the EnKF time update amounts to a onestepahead simulation of X _{ kk }. Consequently, also continuoustime dynamics can be considered by, for example, numerically solving partial differential equations to obtain X _{ k+1k }. Also nonGaussian initial state and process noise distributions with arbitrary densities p(x _{0}) and p(v _{ k }) can be employed as long as they allow sampling. Perhaps because of this flexibility, the time update is often omitted in the EnKF literature [9, 13].
3.4 The EnKF measurement update
The EnKF measurement update is referred to as analysis in the geoscientific literature. A prediction or forecast ensemble X _{ kk−1} is processed to obtain the filtering ensemble X _{ kk } that encodes the KF mean and covariance. We assume that a gain \({\bar {K}}_{k}=K_{k}\) is given and postpone its computation to the next section.
The resulting ensemble average (8a) is the KF mean \(\hat {x}_{kk}\) of (5b). However, with \(y_{k} \mathbbm {1}^{T}\) known, the sample covariance (8b) of X _{ kk } gives only the first term of (5c). Hence, X _{ kk } fails to carry the information in P _{ kk }.
where we generalized h to accept matrix inputs similar to (11).
In the EnKF literature, the prevailing view of inserting artificial noise is that perturbed measurements \(y_{k}\mathbbm {1}^{T}E_{k}\) are processed. This might appear unusual from an SP perspective since it suggests that information is distorted before processing. The introduction of output ensembles Y _{ kk−1}, in contrast, yields a direct connection to (4) and highlights the similarities between (15) and (5a).
An interesting point [60] is that the measurement y _{ k } enters linearly in (13) and (15) and merely shifts the ensemble locations. This highlights the EnKF roots in the linear KF in which P _{ kk } also remains unchanged by y _{ k }.
3.5 The EnKF gain
The optimal gain (6) in the KF is computed from the covariance matrices of the predicted state and output. In the EnKF, the required M _{ k } and S _{ k } are not available but must be approximated from the prediction ensembles (10) or (11), and (14) or (16).
The computations (17) are entirely sampling based, which is useful for the nonlinear case but introduces extra sampling errors. An obvious improvement for additive measurement noise e _{ k } with covariance R is given in Section 5.2, together with the square root EnKF that avoid the insertion of E _{ k } altogether.
4 Some properties and challenges of the EnKF
After a brief review of convergence results and the computational complexity of the EnKF, we discuss adverse effects that can occur in EnKF with finite ensemble size N.
4.1 Asymptotic convergence results
In linear Gaussian systems, the EnKF mean and covariance (7) converge to the KF results (5) as N→∞. This result has been established from different theoretical perspectives [26–28, 31].
For nonlinear systems, the convergence is not as tangible. An investigation of the EnKF as particle system is given in [31], with the outcome that the EnKF does not give the Bayesian filtering solution except for the linear Gaussian case. An illustration of this property is given in the example of Section 7.2.
4.2 Computational complexity
For the complexity analysis, we assume that we are only interested in the filtering results and that n>N>m, that is, the number of measurements is less than the ensemble size and state dimension.
The KF propagates the ndimensional mean vector \(\hat {x}_{kk}\) and the n×n covariance matrix P _{ kk } with n(n+1)/2 unique entries. These storage requirements of \(\mathcal {O}(n^{2}/2)\) dominate for large n>m. The EnKF requires the storage of only nN values. The space required to store the Kalman gain and other intermediate results is similar for the KF and EnKF. A reduction via sequential processing of measurements, as explained in Section 5.1, is possible for both.
For large n, the computational bottleneck of the KF is the covariance time update (3b). Without considering any potential structure in F, slightly less than \(\mathcal {O}(n^{3})\) floating point operations (flops) are required. Contemporary matrix multiplication routines [61] achieve a reduction to roughly \(\mathcal {O}(n^{2.4})\). The EnKF time update requires the propagation of N realizations. If each propagation costs \(\mathcal {O}(n^{2})\) flops, then time update is achieved in \(\mathcal {O}(n^{2}N)\) flops.
The computation of the KF gain requires \(\mathcal {O}(n^{2}m)\) flops for the computation of M _{ k } and S _{ k }. The solution of (6) for K _{ k } amounts to \(\mathcal {O}(m^{3})\). The actual measurement update (5) adds further \(\mathcal {O}(n^{2}m)\) flops. For large n, the total cost is \(\mathcal {O}(n^{2}m)\). In contrast, the EnKF parameters \(\bar M_{k}\) and \(\bar S_{k}\) can be computed in \(\mathcal {O}(nmN)\) flops which, again, dominates the total cost of the measurement update for large n. So, the EnKF flop count scales a factor \(\frac {N}{n}\) better.
4.3 Sampling and coupling effects for finite ensemble size
A serious issue in the EnKF is a commonly noted tendency to underestimate the state uncertainty when using N<n ensemble members [13, 18, 19]. In other words, the EnKF becomes overconfident and is likely to diverge [3] for too small N. A number of causes and related effects can be noted.
First, an ensemble X _{ kk−1} with too few members might not cover the relevant regions of the statespace well enough after the time update (10). The underestimated spread persists in the measurement update (13) or (15) and also X _{ kk } shows too little spread.
Second, the ensemble can only transport limited information and provide a sampling covariance \({\bar {P}}_{kk}\), (7b) or (8b), of at most rank N−1. Consequently, identically zero entries of P _{ kk } are difficult to reproduce and unwanted spurious correlations show up in \({\bar {P}}_{kk}\). An example would be an unreasonably large correlation between the temperature at two distant locations on the globe. Of course, these correlations also affect \({\bar {M}}_{k}\) and \({\bar {S}}_{k}\), and thus the EnKF gain \({\bar {K}}_{k}\) in (18). As a result, state components that are actually uncorrelated to y _{ k } are erroneously updated in (13) or (15). Again, this leads to a reduction in ensemble spread.
which clearly shows the nonlinear dependencies that impede Gaussianity of \(x_{kk}^{(i)}\). Although similar conclusions hold for the general case, concise effects on the ensemble spread are difficult to analyze. Some special cases (n=1 and n=m, H=I, R∝I) are investigated in [26] and shown to produce an underestimated \({\bar {P}}_{kk}\).
Finally, the random sampling in the measurement update by inserting measurement noise in (14) or (16) adds to the EnKF error budget. The inherent sampling errors can be reduced by using the square root EnKF of Section 5.2.
Experiments suggest that there is a threshold for N above which the EnKF works. A good example is given in [42]. Section 5 discusses methods such as inflation and localization that can reduce this minimum N.
5 Important extensions to the EnKF
The previous section highlighted some of the challenges of the EnKF. Here, we summarize the important extensions that are often essential to achieve a working EnKF with only few ensemble members.
5.1 Sequential updates
For the KF, it is algebraically equivalent to carry out m measurement updates (5) with the scalar components of y _{ k } instead of a batch update with the mdimensional y _{ k }, if the measurement noise covariance R is diagonal [3]. Although often treated as a side note only, this technique is very useful. It yields a more flexible algorithm with regard to the availability of measurements at each time step k and reduces the computational complexity. After all, the Kalman gain (6) merely requires a scalar division for each component of y _{ k }. An extension to blockdiagonal R is imminent.
Motivated by the large number of measurements in geoscientific problems, sequential updates have also been suggested for the EnKF [14]. Because of the randomness inherent to the EnKF, there is no algebraic equivalence between sequential and batch updates. Hence, the order in which measurements are processed has an effect on the filtering results.
Furthermore, an unusual alternative interpretation of sequential updates can be found in the EnKF literature. Namely, measurement updates are carried out “grid point by grid point” [13, 16, 42], that is, an iteration is carried out over state rather than measurement components. We will return to this aspect in Section 5.4.
5.2 Model knowledge in the EnKF and squareroot filters
The sampling based derivation of the EnKF in Eqs. (10) through (18) facilitates a compact presentation. However, the randomization through E _{ k } in (14) or (16) adds Monte Carlo sampling errors to the EnKF budget. This section discusses how these errors can be reduced for linear systems (1). Similar results for nonlinear systems with additive noise follow easily. The interpretation of ensembles as (rectangular) matrix square roots is a common theme in the following approaches. In (8b), for instance, \(\tfrac {1}{\sqrt {N1}}{\widetilde {X}}_{kk}\) can be seen as an n×N square root of \({\bar {P}}_{kk}\).
A QR decomposition [61] of the right matrix then yields a triangular m×m square root of \({\bar {S}}_{k}\), and the computation of \({\bar {K}}_{k}\) simplifies to forward and backward substitution. Such ideas have their origin in sigma point KF variants [62].
The KF permits offline computation of the covariance matrices P _{ kk } for all k because they do not depend on the measurements. In an EnKF for a linear system (1), we can mimic this behavior by propagating zeromean ensembles \({\widetilde {X}}_{kk}\) that only carry the information of P _{ kk }. This is the central idea of different square root EnKF [39] which were suggested in [36] (ensemble adjustment filter, EAKF) or [37, 38] (ensemble transform filter, ETKF). The name square root EnKF stems from a relation to square root KF [3] which propagate n×n matrix square roots \(P^{\frac {1}{2}}_{kk}\) with \(P^{\frac {1}{2}}_{kk} P^{\frac {\mathrm {T}}{2}}_{kk}=P_{kk}\). Most importantly, the artificial measurement noise and the inherent sampling error can be avoided.
which requires the left hand side to be positive definite. This property is easily established for the positive definite \({\bar {S}}_{k}\) of (21c) after realizing that the left hand side of (23b) is a Schur complement [61] of a positive definite matrix.
that correctly encodes P _{ kk } without using any random perturbations. Numerically efficient schemes to reduce the computational complexity of ETKF that work on N×N transform matrices can be found in the literature [39]. Other variants update the deviation ensemble via a multiplication from the left [36], which is more costly for large n. Some more conditions on \(\Pi _{k}^{\frac {1}{2}}\) must be met for \({\widetilde {X}}_{kk}\) to remain zeromean [63, 64].
where \({\bar {K}}_{k}\) is computed from the deviation ensembles.
There are indications that in nonlinear and nonGaussian systems the sampling based EnKF variants should be preferable over their square root counterparts: A lowdimensional example is studied in [65]; the impression is confirmed for a highdimensional problem in [66].
5.3 Covariance inflation
with a factor c>1. In the EnKF context, this heuristic has been proposed in [40]. Related concepts are dithering in the PF [7] and the “fudge factor” to increase P _{ kk−1} in the KF [67]. Extensions to adaptive inflation, where c is adjusted online, are discussed in [23].
5.4 Localization
Localization is a technique to address the issue of spurious correlations in the EnKF, and a crucial feature of operational EnKF [18, 19]. The underlying idea applies equally well to the EnKF and the KF, and can be used to systematically update only a part of the state vector with each measurement.
A KF measurement update (5) with the above K _{ k } does not affect the x _{3} estimate or covariance. Hence, there is a lower dimensional measurement update that only alters the statistics of x _{1} and x _{2}.
Localization in the EnKF enforces the above structure using two prevailing techniques, local updates [13, 16, 42] and covariance tapering [14, 43]. Both rely on prior knowledge of the covariance structure. For example, the state components are often connected to geographic locations in geoscientific applications. From the underlying physics, it is reasonable to assume zero correlation between distant states. Unfortunately, this viewpoint is not transferable to highdimensional problems in general.
Local updates were introduced for the samplingbased EnKF in [13] and for different square root EnKF in [16, 42]. Nonlinear measurement functions (2b) are linearized in the latter two. All of the above references update the state vector “grid point by grid point,” which appears unusual from a KF perspective [3]. In an iteration, local state vectors of small dimension (<N) are chosen and updated with a subset of supposedly relevant measurements. These “full rank” updates avoid some of the problems associated with small N and large n. However, discontinuities between state components are introduced [68]. Some heuristics to combine the local ensembles and further implementation details can be found in [42, 69].
Under the assumption of the structure in (27), a local analysis would amount to an EnKF update of the x _{1} and x _{2}components only, to avoid errors in x _{3}.
where we assumed a linear measurement relation (1b).
There are some technicalities associated with the tapering operation. Only positive semidefinite ρ guarantee that \((\rho \circ {\bar {P}}_{kk1})\) is a valid covariance [26]. Full rank ρ yield an increased rank in \((\rho \circ {\bar {P}}_{kk1})\) [14]. However, low rank ρ do not necessarily decrease the rank of \((\rho \circ {\bar {P}}_{kk1})\). A closely related problem to finding valid (positive semidefinite or definite) ρ is the creation of covariance functions and kernels in Gaussian processes [71]. Here, a methodology to create more complicated kernels from simpler ones could be used to create ρ.
instead of (29a) and to skip the tapering in S _{ k } altogether [43]. After all, for lowdimensional y _{ k } (small m) \({\bar {M}}_{k}\) has the strongest influence on the gain \({\bar {K}}_{k}\). Also, the matrix ρ _{ M } is constructed from prior knowledge about the correlation. In the geoscientific context, where the state components and measurements are associated with geographic locations, this is easy. In general, however, it might not be possible to devise an appropriate ρ _{ M }. Other variants [14, 26, 68] with tapering for \({\bar {S}}_{k}\) exist and have in common that they are only identical to (29) for H=I.
The multiplication with diag(r) merely removes the rows corresponding to x _{3}, which establishes an equivalence between local updates and covariance tapering. By picking a smoothly decaying r, we can furthermore avoid the discontinuities associated with local updates.
5.5 The EnKF gain and least squares
that are to be solved for each row of \({\bar {K}}_{k}\) and \({\widetilde {X}}_{kk1}\).
Hence, the EnKF iteration can be carried out without explicitly computing any sample covariance matrices if instead efficient solutions to the problem (32b) are employed. Furthermore, the problem (32b) could be modified using regularization [72] to enforce sparsity in \({\bar {K}}_{k}\). This would be an alternative approach to the localization methods discussed earlier. Related ideas to improve the Kalman gain using bootstrap methods [72] for computing \({\bar {M}}_{k}\) and \({\bar {S}}_{k}\) in (17) are discussed in [73, 74].
6 Relations to other algorithms
The EnKF for nonlinear systems (2) differs from other samplingbased nonlinear filters such as sigma point KF [5] or particle filters (PF) [7]. One reason for this is that the EnKF approximates the KF algorithm (with the side effect that it can be applied to (2)) rather than trying to solve the nonlinear filtering problem directly.
The biggest difference between the EnKF and sigma point filters [5] such as the unscented KF [4, 56] or divided difference KF [62] is the measurement update. Whereas the EnKF updates its ensembles, the latter carry out the KF measurement update (5) using approximately computed mean values and covariance matrices. That is, the samples or sigma points are condensed into a filtering estimate \(\hat {x}_{kk}\) and its covariance P _{ kk }, which entails a loss of information and can be seen as an inherent Gaussian assumption on the filtering density p(x _{ k }y _{1:k }). In contrast, the EnKF can preserve more information and deviations from Gaussianity in the ensemble. Similarities appear in the gain computations of the EnKF and sigma point KF. In both, the Kalman gain appears as a function of the sampling covariance matrices, although with the deterministic sigma points and weights in the latter. With their origin in the KF, both sigma point filters and the EnKF can be expected to share difficulties with multimodal posterior distributions.
Similar to the EnKF, the PF propagates N state realizations that are called particles. For the bootstrap particle filter [6], the prediction step corresponds to the EnKF time update (11). Apart from that, however, the differences dominate. First, the PF is designed as an approximate solution of the Bayesian filtering equations [15] using sequential importance sampling [7]. For N→∞, the PF solution recovers the true filtering density. Second, the samples in basic PF variants are generated from a proposal distribution only once every time instance and then left untouched. The measurement update amounts to updating the particle weights, which leads to a degeneracy problem for large n. In the EnKF, in contrast, the ensemble members are influenced by the time and the measurement update. Third, the PF relies on a crucial resampling step that is not present in the EnKF. An attempt to use the EnKF as proposal density in PF is described in [75]. A unifying interpretation of the EnKF and PF as ensemble transform filters can be found [76].
Still, the EnKF appears as a distinct algorithm besides sigma point KF and PF. Its properties and potential for nonlinear problems remain to be fully investigated. Existing results that the EnKF does not converge to the Bayesian filtering recursion [31] remain to be interpreted in a constructive manner.
7 Instructive simulation examples
Four examples are discussed in greater detail, among them one popular benchmark problem of the SP and DA literature each.
7.1 A scalar linear Gaussian model
A trajectory of (33) is simulated and a KF is used to compute the optimal variances P _{ kk }. Because the model is timeinvariant, the P _{ kk } quickly converge to a constant value. For k>3, P _{ kk }=0.0092 is obtained.
The average \({\bar {P}}_{kk}\) over the Monte Carlo realizations is close to the desired P _{ kk }. However, there is a large spread among the \({\bar {P}}_{kk}\) and the distribution is skewed toward zero with its median below P _{ kk }. Although N>n, there is a tendency to underestimate P _{ kk }.
7.2 The particle filter benchmark
with independent \(v_{k}\sim \mathcal {N}(0,10)\), \(e_{k}\sim \mathcal {N}(0,1)\), and \(x_{0}\sim \mathcal {N}(0,1)\). Because the model is scalar, the Bayesian filtering densities p(x _{ k }  y _{1:k }) can be computed numerically using point mass filters (PMF) [77]. A sampling based EnKF with N=500 is tested and kernel density estimates are used to obtain approximations of p(x _{ k }  y _{1:k }) from the ensembles. For comparison, we include a closely related sigma point KF variant that uses Monte Carlo integration with N=500 samples [5]. The only difference to the EnKF is that this Monte Carlo KF (MCKF) carries out the KF measurement update (5) to propagate a mean and a variance. We illustrate the results as Gaussian densities.
7.3 Batch smoothing using the EnKF
with dimension n=(L+1)n _{ x }. Second, we note that the measurements y _{ k }, k=1,…,L, have uncorrelated measurement noise and known relations to the components of ξ. For linear systems (1), the predicted mean and covariance of ξ can be easily derived, and smoothed estimates of all x _{ k }, k=0,…,L, can be obtained by sequentially processing all y _{ k } in KF measurement updates for ξ.
Also, other smoothing variants and the RauchTungStriebel (RTS) algorithm can be derived from state augmentation approaches [3]. Due to its sequential nature, however, the RTS smoother does not provide joint covariance matrices of x _{ k } and x _{ k+i } for i≠0. Except for this and the higher computational complexity of working with ξ, the batch and RTS smoothers are equivalent for (1).
An EnKF approach to batch smoothing mimics the above. A prediction ensemble for ξ is obtained by simulating N trajectories for random process noise and initial state realizations. This can also be carried out for nonlinear models (2). Then, sequential EnKF measurement updates are performed for all y _{ k }.
With n _{ x }=4 and L=49 we obtain n=200 as dimension of ξ. The RTS solution is compared to EnKF of ensemble size N={10,20,50}. Monte Carlo errors are reduced using (21) in the gain computations.
The ensemble is compactly gathered but does not cover the true trajectory well. The EnKF is overconfident. A last experiment explores how well an EnKF with N=20 captures the uncertainty of the state estimate. Furthermore, we discuss effects of the order in which the measurements are processed. Specifically, we compare the ensemble covariance of the positions x _{ k } to the exact cov(x _{ k },x _{ i }), i,k=0,…,L, obtained by KF updates for the augmented state ξ.
7.4 The 40dimensional Lorenz model
Our final example is a benchmark problem from the EnKF literature. We investigate the 40dimensional Lorenz96 model^{3} from [53] that is used in, e.g., [36, 38, 42, 50, 52, 63, 69].
where j=1,…,40 indexes the components of x, with the convention that x(0)=x(40) etc. Instead of the commonly used forcing term F(j)=8, we assume timedependent \(\mathsf {F}_{k}(j)\sim \mathcal {N}(8,1)\) that are constant for time intervals T=0.05 only and act as process noise. A RungeKutta method (RK4) is used to discretize (37) to obtain the nonlinear state difference Eq. (2a) with x _{ k }=x _{ k } and v _{ k }=F _{ k }. The step size T corresponds to about 6 h if x were an atmospheric quantity on a latitude circle of the earth [53]. Although the model (37) is said to be chaotic, the effects are only mild for short integration times T. In our experiments, all n=40 states are measured with additive Gaussian noise \(e_{k}\sim \mathcal {N}(0,I)\). The initial state is Gaussian with \(x_{0}\sim \mathcal {N}(0,P_{0})\), where P _{0} is drawn from a Wishart distribution with seed matrix I _{ n } and n degrees of freedom.
There is a tendency for peaks to move “westwards” as k increases. We note that there are also alternative approaches for estimating x, for example, by first linearizing and then discretizing (37). However, we adopt the RK4 discretization of the EnKF literature that yields a state transition that is easy to evaluate but difficult to linearize. Because of this, the EKF [3] cannot be applied easily and we obtain a challenging benchmark problem.
where \(\hat {x}_{kk}\) is the ensemble mean. We use the average ε _{ k } for k=100,…,L, denoted by \(\bar \varepsilon \), as quantitative performance measure for different EnKF. Useful EnKF must yield \(\bar \varepsilon <1\), which is the error when simply taking \(\hat {x}_{kk}=y_{k}\).
Averaged errors \(\bar \varepsilon \) for different EnKF
N  c  ρ  \(\bar \varepsilon \) 

1000  1  No  0.29 
40  1  No  0.44 
40  1.05  No  0.33 
40  1  Yes  0.29 
40  1.02  Yes  0.28 
20  >1  No  >1 
20  1.01  Yes  0.3 
10  1.05  Yes  0.34 
A number of lessons have been learned from related experiments. As alternative to the ρ in Fig. 17, a simpler taper that contains only ones and zeros to enforce the banded structure was used. Although this ρ was indefinite, a reduction in \(\bar \varepsilon \) was achieved without any numerical issues. Hence, the specific structure of ρ appears secondary. The smooth ρ of Fig. 17 remains preferable in terms of \(\bar \varepsilon \), though. Sequential processing of the measurements did not degrade the performance. Experiments without process noise give the lower errors \(\bar \varepsilon \) from, e.g., [38, 42].
8 Conclusions
With this paper, we have given a comprehensive and easy to understand introduction to the EnKF for signal processing researchers. The origin of the EnKF in the KF and its simple implementation have been demonstrated. The unique literature review provides quick access to the most relevant papers in the plethora of geoscientific EnKF publications. Furthermore, we have discussed the challenges related to small ensembles for highdimensional states, N<n, and the available solutions such as localization or inflation. Finally, we have tested the EnKF on signal processing and EnKF benchmark problems.
With its scalability and simple implementation, even for nonlinear and nonGaussian problems, the EnKF stands out as viable candidate for many state estimation problems. Furthermore, localization ideas and advanced concepts for estimating covariance matrices and the EnKF gain from the limited information in the ensembles provide new research directions for the EnKF and highdimensional filters in general, hopefully with an increased participation from the signal processing community.
9 Endnotes
^{1} With over 3000 citations between 1994 and 2016.
^{2} We assume that the components can be processed sequentially.
^{3} Also known as the Lorenz96, L95, L96, or L40 model.
Declarations
Acknowledgements
This work was supported by the project Scalable Kalman Filters granted by the Swedish Research Council (VR).
Authors’ contributions
MR wrote the majority of the text and performed the majority of the simulations. GH and CF contributed text to earlier versions of the manuscript and helped with the simulations. GH, CF, and FG commented on and approved the manuscript. FG initiated the research on ensemble Kalman filters. All authors read and approved the final manuscript.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
Authors’ Affiliations
References
 E Kalnay, Atmospheric modeling, data assimilation and predictability (Cambridge University Press, New York, 2002).View ArticleGoogle Scholar
 RE Kalman, A new approach to linear filtering and prediction problems. J. Basic Eng. 82(1), 35–45 (1960).View ArticleGoogle Scholar
 BD Anderson, JB Moore, Optimal filtering (Prentice Hall, Englewood Cliffs, 1979).MATHGoogle Scholar
 S Julier, J Uhlmann, H DurrantWhyte, in Proceedings of the American Control Conference 1995 vol.3.A new approach for filtering nonlinear systems (IEEESeattle, 1995), pp. 1628–1632.View ArticleGoogle Scholar
 M Roth, G Hendeby, F Gustafsson, Nonlinear Kalman filters explained: a tutorial on moment computations and sigma point methods. J Adv. Inf. Fusion. 11(1), 47–70 (2016).Google Scholar
 NJ Gordon, DJ Salmond, AF Smith, Novel approach to nonlinear/nonGaussian Bayesian state estimation. Radar Signal Process. IEE Proc. F. 140(2), 107–113 (1993).Google Scholar
 F Gustafsson, Particle filter theory and practice with positioning applications. IEEE Aerosp. Electron. Syst. Mag. 25(7), 53–82 (2010).View ArticleGoogle Scholar
 G Evensen, Sequential data assimilation with a nonlinear quasigeostrophic model using Monte Carlo methods to forecast error statistics. J. Geophys. Res. Oceans. 99(C5), 3–10162 (1014).Google Scholar
 G Burgers, JP van Leeuwen, G Evensen, Analysis scheme in the ensemble Kalman filter. Mon. Weather Rev. 126(6), 1719–1724 (1998).View ArticleGoogle Scholar
 H DurrantWhyte, T Bailey, Simultaneous localization and mapping: Part I. IEEE Robot. Autom. Mag. 13(2), 99–110 (2006).View ArticleGoogle Scholar
 M Baum, UD Hanebeck, Extended object tracking with random hypersurface models. IEEE Trans. Aerosp. Electron. Syst. 50(1), 149–159 (2014).View ArticleGoogle Scholar
 N Wahlström, E Özkan, Extended target tracking using Gaussian processes. IEEE Trans. Signal Proc. 63(16), 4165–4178 (2015).MathSciNetView ArticleGoogle Scholar
 PL Houtekamer, HL Mitchell, Data assimilation using an ensemble Kalman filter technique. Mon. Weather Rev. 126(3), 796–811 (1998).View ArticleGoogle Scholar
 PL Houtekamer, HL Mitchell, A sequential ensemble Kalman filter for atmospheric data assimilation. Mon. Weather Rev. 129(1), 123–137 (2001).View ArticleGoogle Scholar
 AH Jazwinski, Stochastic processes and filtering theory (Academic Press, New York, 1970).MATHGoogle Scholar
 G Evensen, The ensemble Kalman filter: theoretical formulation and practical implementation. Ocean Dyn. 53(4), 343–367 (2003).View ArticleGoogle Scholar
 G Evensen, Data assimilation: the ensemble Kalman filter, 2nd ed. (Springer, Dordrecht, New York, 2009).View ArticleMATHGoogle Scholar
 TM Hamill, in Predictability of Weather and Climate. Ensemblebased atmospheric data assimilation (Cambridge University PressCambridge, 2006).Google Scholar
 PL Houtekamer, HL Mitchell, Ensemble Kalman filtering. Q. J. R. Meteorol. Soc. 131(613), 3269–3289 (2005).View ArticleGoogle Scholar
 JS Whitaker, TM Hamill, X Wei, Y Song, Z Toth, Ensemble data assimilation with the NCEP global forecast system. Mon. Weather Rev. 136(2), 463–482 (2008).View ArticleGoogle Scholar
 GP Compo, JS Whitaker, PD Sardeshmukh, N Matsui, RJ Allan, X Yin, BE Gleason, RS Vose, G Rutledge, P Bessemoulin, S Brönnimann, M Brunet, RI Crouthamel, AN Grant, PY Groisman, PD Jones, MC Kruk, AC Kruger, GJ Marshall, M Maugeri, HY Mok, O Nordli, TF Ross, RM Trigo, XL Wang, SD Woodruff, SJ Worley, The twentieth century reanalysis project. Q. J. R. Meteorol. Soc. 137(654), 1–28 (2011).View ArticleGoogle Scholar
 S Lakshmivarahan, D Stensrud, Ensemble Kalman filter. IEEE Control. Syst. 29(3), 34–46 (2009).MathSciNetView ArticleGoogle Scholar
 J Anderson, Ensemble Kalman filters for large geophysical applications. IEEE Control. Syst. 29(3), 66–82 (2009).MathSciNetView ArticleGoogle Scholar
 G Evensen, The ensemble Kalman filter for combined state and parameter estimation. IEEE Control. Syst. 29(3), 83–104 (2009).MathSciNetView ArticleGoogle Scholar
 J Mandel, J Beezley, J Coen, M Kim, Data assimilation for wildland fires. IEEE Control. Syst. 29(3), 47–65 (2009).MathSciNetView ArticleGoogle Scholar
 R Furrer, T Bengtsson, Estimation of highdimensional prior and posterior covariance matrices in Kalman filter variants. J. Multivar. Anal. 98(2), 227–255 (2007).MathSciNetView ArticleMATHGoogle Scholar
 M Butala, J Yun, Y Chen, R Frazin, F Kamalabadi, in 15th IEEE International Conference on Image Processing. Asymptotic convergence of the ensemble Kalman filter (IEEESan Diego, 2008), pp. 825–828.Google Scholar
 J Mandel, L Cobb, JD Beezley, On the convergence of the ensemble Kalman filter. Appl. Math. 56(6), 533–541 (2011).MathSciNetView ArticleMATHGoogle Scholar
 M Frei, Ensemble Kalman Filtering and Generalizations (Dissertation, ETH, Zürich, 2013). nr. 21266.Google Scholar
 M Katzfuss, JR Stroud, CK Wikle, Understanding the ensemble Kalman filter. Am. Stat. 70(4), 350–357 (2016).MathSciNetView ArticleGoogle Scholar
 F Le Gland, V Monbet, V Tran, in The Oxford Handbook of Nonlinear Filtering, ed. by D Crisan, B Rozovskii. Large sample asymptotics for the ensemble Kalman filter (Oxford University PressOxford, 2011), pp. 598–634.Google Scholar
 M Butala, R Frazin, Y Chen, F Kamalabadi, Tomographic imaging of dynamic objects with the ensemble Kalman filter. IEEE Trans. Image Process. 18(7), 1573–1587 (2009).MathSciNetView ArticleGoogle Scholar
 J Dunik, O Straka, M Simandl, E Blasch, Randompointbased filters: analysis and comparison in target tracking. IEEE Trans. Aerosp. Electron. Syst. 51(2), 1403–1421 (2015).View ArticleGoogle Scholar
 S Gillijns, O Mendoza, J Chandrasekar, B De Moor, D Bernstein, A Ridley, in American Control Conference, 2006. What is the ensemble Kalman filter and how well does it work? (IEEEMinneapolis, 2006), pp. 4448–4453.Google Scholar
 M Roth, C Fritsche, G Hendeby, F Gustafsson, in European Signal Processing Conference 2015 (EUSIPCO 2015). The ensemble Kalman filter and its relations to other nonlinear filters (IEEEFrance, 2015).Google Scholar
 JL Anderson, An ensemble adjustment Kalman filter for data assimilation. Mon. Weather Rev. 129(12), 2884–2903 (2001).View ArticleGoogle Scholar
 CH Bishop, BJ Etherton, SJ Majumdar, Adaptive sampling with the ensemble transform Kalman filter. Part I: Theoretical aspects. Mon Weather Rev. 129(3), 420–436 (2001).Google Scholar
 JS Whitaker, TM Hamill, Ensemble data assimilation without perturbed observations. Mon. Weather Rev. 130(7), 1913–1924 (2002).View ArticleGoogle Scholar
 MK Tippett, JL Anderson, CH Bishop, TM Hamill, JS Whitaker, Ensemble square root filters. Mon. Weather Rev. 131(7), 1485–1490 (2003).View ArticleGoogle Scholar
 JL Anderson, SL Anderson, A Monte Carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts. Mon. Weather Rev. 127(12), 2741–2758 (1999).View ArticleGoogle Scholar
 PJ van Leeuwen, Comment on “data assimilation using an ensemble Kalman filter technique”. Mon. Weather Rev. 127(6), 1374–1377 (1999).View ArticleGoogle Scholar
 E Ott, BR Hunt, I Szunyogh, AV Zimin, EJ Kostelich, M Corazza, E Kalnay, DJ Patil, JA Yorke, A local ensemble Kalman filter for atmospheric data assimilation. Tellus A. 56(5), 415–428 (2004).View ArticleGoogle Scholar
 TM Hamill, JS Whitaker, C Snyder, Distancedependent filtering of background error covariance estimates in an ensemble Kalman filter. Mon. Weather Rev. 129(11), 2776–2790 (2001).View ArticleGoogle Scholar
 PN Raanes, On the ensemble RauchTungStriebel smoother and its equivalence to the ensemble Kalman smoother. Q. J. R. Meteorol. Soc. 142(696), 1259–1264 (2016).View ArticleGoogle Scholar
 M Zupanski, Maximum likelihood ensemble filter: theoretical aspects. Mon. Weather Rev. 133(6), 1710–1726 (2005).View ArticleGoogle Scholar
 TM Hamill, C Snyder, A hybrid ensemble Kalman filter–3d variational analysis scheme. Mon. Weather Rev. 128(8), 2905–2919 (2000).View ArticleGoogle Scholar
 PJ van Leeuwen, A varianceminimizing filter for largescale applications. Mon. Weather Rev. 131(9), 2071–2084 (2003).View ArticleGoogle Scholar
 C Snyder, T Bengtsson, P Bickel, J Anderson, Obstacles to highdimensional particle filtering. Mon. Weather Rev. 136(12), 4629–4640 (2008).View ArticleGoogle Scholar
 PJ van Leeuwen, Particle filtering in geophysical systems. Mon. Weather Rev. 137(12), 4089–4114 (2009).View ArticleGoogle Scholar
 PJ van Leeuwen, Nonlinear data assimilation in geosciences: an extremely efficient particle filter. Q. J. R. Meteorol. Soc. 136(653), 1991–1999 (2010).View ArticleGoogle Scholar
 M Frei, HR Künsch, Bridging the ensemble Kalman and particle filters. Biometrika. 100(4), 781–800 (2013).MathSciNetView ArticleMATHGoogle Scholar
 J Poterjoy, A localized particle filter for highdimensional nonlinear systems. Mon. Weather Rev. 144(1), 59–76 (2015).View ArticleGoogle Scholar
 EN Lorenz, in Predictability of Weather and Climate, ed. by T Palmer, R Hagedorn. Predictability—a problem partly solved (Cambridge University PressCambridge, 2006), pp. 40–58.View ArticleGoogle Scholar
 DT Pham, Stochastic methods for sequential data assimilation in strongly nonlinear systems. Mon. Weather Rev. 129(5), 1194–1207 (2001).View ArticleGoogle Scholar
 X Luo, I Moroz, Ensemble Kalman filter with the unscented transform. Phys. D Nonlinear Phenom. 238(5), 549–562 (2009).MathSciNetView ArticleMATHGoogle Scholar
 SJ Julier, JK Uhlmann, Unscented filtering and nonlinear estimation. Proc. IEEE. 92(3), 401–422 (2004).View ArticleGoogle Scholar
 P Sakov, Comment on “ensemble Kalman filter with the unscented transform”. Phys. D Nonlinear Phenom. 238(22), 2227–2228 (2009).MathSciNetView ArticleMATHGoogle Scholar
 AS Stordal, HA Karlsen, G Nævdal, HJ Skaug, B Vallès, Bridging the ensemble Kalman filter and particle filters: the adaptive Gaussian mixture filter. Comput. Geosci. 15(2), 293–305 (2011).View ArticleMATHGoogle Scholar
 I Hoteit, X Luo, DT Pham, Particle Kalman filtering: a nonlinear Bayesian framework for ensemble Kalman filters. Mon. Weather Rev. 140(2), 528–542 (2011).View ArticleGoogle Scholar
 M Frei, HR Künsch, Mixture ensemble Kalman filters. Comput. Stat. Data Anal. 58:, 127–138 (2013).View ArticleMATHGoogle Scholar
 LN Trefethen, D Bau III, Numerical linear algebra (SIAM, Philadelphia, 1997).View ArticleMATHGoogle Scholar
 M Nørgaard, NK Poulsen, O Ravn, New developments in state estimation for nonlinear systems. Automatica. 36(11), 1627–1638 (2000).MathSciNetView ArticleMATHGoogle Scholar
 P Sakov, PR Oke, Implications of the form of the ensemble transformation in the ensemble square root filters. Mon. Weather Rev. 136(3), 1042–1053 (2008).View ArticleGoogle Scholar
 DM Livings, SL Dance, NK Nichols, Unbiased ensemble square root filters. Phys D Nonlinear Phenom. 237(8), 1021–1028 (2008).MathSciNetView ArticleMATHGoogle Scholar
 WG Lawson, JA Hansen, Implications of stochastic and deterministic filters as ensemblebased data assimilation methods in varying regimes of error growth. Mon. Weather Rev. 132(8), 1966–1981 (2004).View ArticleGoogle Scholar
 O Leeuwenburgh, G Evensen, L Bertino, The impact of ensemble filter definition on the assimilation of temperature profiles in the tropical pacific. Q. J. R. Meteorol. Soc. 131(613), 3291–3300 (2005).View ArticleGoogle Scholar
 Y BarShalom, XR Li, T Kirubarajan, Estimation with applications to tracking and navigation: Theory Algorithms and Software (WileyInterscience, New York, 2001).View ArticleGoogle Scholar
 P Sakov, L Bertino, Relation between two common localisation methods for the EnKF. Comput. Geosci. 15(2), 225–237 (2010).View ArticleMATHGoogle Scholar
 BR Hunt, EJ Kostelich, I Szunyogh, Efficient data assimilation for spatiotemporal chaos: a local ensemble transform Kalman filter. Phys D Nonlinear Phenom. 230(1–2), 112–126 (2007).MathSciNetView ArticleMATHGoogle Scholar
 G Gaspari, SE Cohn, Construction of correlation functions in two and three dimensions. Q. J. R. Meteorol. Soc. 125(554), 723–757 (1999).View ArticleGoogle Scholar
 CE Rasmussen, CKI Williams, Gaussian processes for machine learning (The MIT Press, Cambridge, Mass, 2005).MATHGoogle Scholar
 T Hastie, R Tibshirani, J Friedman, The elements of statistical learning: data mining, inference, and prediction, 2nd ed. (Springer, New York, NY, 2011).MATHGoogle Scholar
 Y Zhang, DS Oliver, Improving the ensemble estimate of the Kalman gain by bootstrap sampling. Math. Geosci. 42(3), 327–345 (2010).View ArticleMATHGoogle Scholar
 I Myrseth, J Sætrom, H Omre, Resampling the ensemble Kalman filter. Comput Geosci. 55:, 44–53 (2013).View ArticleMATHGoogle Scholar
 N Papadakis, E Mémin, A Cuzol, N Gengembre, Data assimilation with the weighted ensemble Kalman filter. Tellus A. 62(5), 673–697 (2010).View ArticleGoogle Scholar
 S Reich, A Nonparametric ensemble transform method for Bayesian inference. SIAM J. Sci. Comput. 35(4), A2013–A2024 (2013).MathSciNetView ArticleMATHGoogle Scholar
 M Roth, F Gustafsson, in 42nd International Conference on Acoustics, Speech, and Signal Processing (ICASSP). Computation and visualization of posterior densities in scalar nonlinear and nonGaussian Bayesian filtering and smoothing problems (IEEENew Orleans, 2017).Google Scholar