• No results found

3 Ensemble Filter Algorithm

3.3 General case: Hidden Markov model

3.3.2 Drift towards Gaussianity

The EnKF algorithm is, as mentioned previously, asymptotically correct, whenne! 1, for Gauss-linear model HM model. Consider an initial ensemblee0generated from a non-Gaussian initial modelf(r0).

Assume further that the forward and likelihood functions are continuous.

The sequential updates in the EnKF will make the ensembleetmore and more Gaussian - the ensem-ble drifts towards Gaussianity. This drift is caused by the successive linearized updates when condition-ing on the data.

Several variants of the EnKF exist that address this issue:

1. Gaussian anamorphosis (GA) EnKF 2. Gaussian Mixture (GM) EnKF 3. Selection (S) EnKF

3.3.2.1 Gaussian anamorphosis (GA) EnKF The idea is to transform the ensemble to be marginally Gaussian before conditioning, the transformed ensemble will be marginally Gaussian, and to carry out the conditioning with the transformed ensemble. After conditioning, the ensemble is then back-transformed.

Applications have shown that Gaussian anamorphosis can successfully prevent the drift towards Gaus-sianity, see Zhou et al. (2012).

Consider a univariate random variableywith cdfFY(y) and a random sample (y1,...,yny) iid from FY(y). The cdf can be estimated as,

whereJ{·} is some semi parametric smoother of the empirical stepwise cdf estimator in the argument.

The smoother ensures that the back-transform is real valued.

The univariate Gaussian transform of one sampley0fromFY(y) is defined as,

˜

y01( ˆFY°1(y0);0,1). (34)

Note that ˜y0has an approximate standard Gaussian pdf. Similarly the back-transform of the univariate Gaussian sample ˜u0is,

u0=FˆY°11 ( ˜u0;0,1)). (35)

The smoothing of ˆFY(y) ensures thatu02Rand do not belong to the set {y1,...,yny} only.

The univariate transformation of an ensembleet: {(ru(i)t ,d(i)t ),i=1,...,ne} entails that for each en-semble member, each of the (n+m) dimensions must be transformed independently. Hence only ap-proximate univariate Gaussianity is ensured, the multivariate characteristics remaining unspecified. The

latter entails that the linearized conditioning is only approximately correct. Algorithm 3 presents the GA EnKF procedure.

Algorithm 3Gaussian anamorphosis EnKF et: { (ru(i)t ,d(i)t ),i=1,...,ne} Conditioning:

Assess all (n+m) dimensions byFY(y) fromet!FˆY(y) Univariate Gaussian transform ofetby ˆFY(y)!e˜t

Univariate Gaussian transform of observationsdtby ˆFY(y)!d˜t Estimateßr dfrome˜t!߈r d!Kˆt=°ˆr d[ ˆßd]°1

˜

rc(i)t =r˜u(i)t +Kˆt(d˜t°d˜(i)t ),i=1,...,ne

Univariate back-transform ofr˜c(i)t by ˆFY(y)!rc(i)t Forwarding:

ru(i)t+1=!t(rc(i)t )+rt;i=1,...,ne d(i)t+1=t+1(ru(i)t+1)+dt;i=1,...,ne

et+1: {(ru(i)t+1,d(i)t+1);i=1,...,ne}

3.3.2.2 Gaussian Mixture (GM) EnKF The basic idea is to specify the prior initial model as a Gaussian mixture (GM) model which can represent multimodal variables. Let the forward and likelihood models of the HM model be Gauss-linear. The posterior pdf will then also be a GM model and be analytically tractable. The conditioning step can be made independently for each component of the GM model, and the associated weight for each component can also be calcultated. For general forward and likelihood models, ensemble based filtering algorithms must be used. The difficulty is that each ensemble mem-ber must carry a mode indicator assigned to one of the components, and that this indicator may change during the forwarding step. These filtering algorithms have proven robust against the drift towards Gaus-sianity (Li et al., 2016; Ackerson and Fu, 1970; Chen and Liu, 2000; Smith, 2007; Dovera and Della Rossa, 2010; Bengtsson et al., 2003), at least for low-dimensional models.

Consider a set ofn-dimensional Gaussian pdfs,'n(r;µlrlr);l=1,...,L, denoted components, and a set of normalized mixture weightsº: {º1,...,ºL}. The prior initial model is specified to be a GM model :

f(r0)=XL

l=1

ºl£'n(r0lrlr). (36) Note that a particular realizationrswill belong to thek-th component of the mixture with probability:

k(rs)=

"L X

l=1

ºl£'n(rslr,ßlr)

#°1

£'n(rskrkr). (37)

3 ENSEMBLE FILTER ALGORITHM 17

Consider further the Gauss-linear likelihood model,

f(d|r)='m(d;Hrd|r). (38)

The associated posterior model will also be a Gaussian mixture model (Grana et al., 2017), f(r|d)=

XL l=1

ºl|d£'n(r;µlr|dlr|d), (39) where the conditional expectations and covariances are obtained by component wise conditioning on the observationsd. The posterior mixture weights are defined by

ºk|d=

"L X

l=1

ºl£'m(d;Hµlr,HßlrHT+ßd|r)

#°1

ºk£'m(d;kr,HßkrHT+ßd|r). (40) For general forward and likelihood models, the ensemble representation must contain a mode indicator associated with each ensemble member,

et: {(ru(i)t ,lt(i),d(i)t ,i=1,...,ne}, (41) with mode indicatorlt(i)2{1,...,L}. The dynamic updating of this mode indicator often appears as chal-lenging. The conditioning/forward steps in the GM-EnKF are detailed in Algorithm 4.

Algorithm 4Gaussian mixture EnKF e°tl: {(ru(i)t ,·,d(i)t ),i=1,...,ne} Conditioning:

Assessf(rut) frome°tl!fˆ(rut)=PL

l=1ºˆl£'n(r, ˆµlr, ˆßlr)

Assign ensemble memberito mode indicatorlt(i)2{1,...,L} with probability {∏l(ru(i))t ),l=1,...,L}

Define ensembleet: {(ru(i)t ,lt(i),d(i)t ,i=1,...,ne} Assessßlr d;l=1,...,Lfromet !߈lr d!Kˆlt=°ˆlr d[ ˆßld]°1 rc(i)t =ru(i)t +Kˆlt(i)t (dt°d(i)t ),i=1,...,ne

Forwarding:

ru(i)t+1=!t(rc(i)t )+rt,i=1,...,ne

d(i)t+1=t+1(ru(i)t+1)+dt,i=1,...,ne

e°lt+1: {(ru(i)t+1,d(i)t+1),i=1,...,ne}

The challenging part of the algorithm is to assess the GM modelf(rut). Other versions of the (GM) EnKF algorithm use the EM-algorithm, particle filters or clustering techniques. If the dimension ofris of some size, estimating a suitable GM model will be complicated.

3.3.2.3 Selection (S)EnKF The basic idea is to specify the prior initial model as a selection-Gaussian pdf, see Section 2.2, which can represent multimodal, skewed and/or peaked variables. For general for-ward and likelihood models, one must rely on ensemble based filtering algorithms. These algorithms are inspired by the analytically tractable model discussed in Section 2.2, and have proven to be robust with regard to drift towards Gaussianity (Conjard and Omre, 2020). Let the prior initial distributionf(r0) be a selection-Gaussian pdf with parameters£SG=(µr˜,ßr˜∫|r˜∫|˜r,A), see Equations 18 and 19. The auxiliary variables (˜r0,∫) are then jointly Gaussian and the variable of interest isr0=[˜r0|2A] which is selection-Gaussian.

The initial ensemblee0of the EnKF algorithm contains realizations of the auxiliary variables [r˜0,∫]

which are jointly Gaussian. The SEnKF algorithm is identical to the EnKF algorithm defined on these auxiliary variables. The forward model is given by

r˜t+1

while the likelihood model is given by

dt=!trt,≤dt). (43)

Based on these models, an algorithm identical to the traditional EnKF algorithm is activated, to obtain the ensembleeT+1={(˜ru(i)T+1,∫u(i)T+1),i=1,...,ne}. Note that a time index is added toto account for the data assimilation up to timeT. The expectation vectorµr˜and covariance matrixßr∫˜ are estimated from eT+1, and based on the jointly Gaussian'2n((r˜,∫);µˆr˜,߈r˜), the filter variable of interest [rT+1|d0:T]= [˜rT+1|2A,d0:T] is assessed by McMC simulation.

The conditioning and forwarding steps of the SEnKF algorithm are specified in Algorithm 5.

Algorithm 5Selection EnKF

3 ENSEMBLE FILTER ALGORITHM 19