\`x^2+y_1+z_12^34\`
Advanced Search
Article Contents
Article Contents

Decomposition of likelihoods and techniques for multi-scale data assimilation

  • *Corresponding author: John Maclean

    *Corresponding author: John Maclean 

The second author was supported by Office of Naval Research Grant N000142412218.

Abstract / Introduction Full Text(HTML) Figure(9) / Table(1) Related Papers Cited by
  • Approaches are developed for splitting of data assimilation (DA) problems into manageable pieces so that different DA methods may be applied. A block matrix generalized inverse result allows for the decomposition of Gaussian likelihoods based upon projected data models. A sophisticated implementation of a prototype method is developed and promising results are obtained for different model problems. Numerical techniques are focused on a combined particle filter/Kalman filter (PF-EnKf) using low-dimensional projected optimal proposal-particle filter (OP-PF) and ensemble Kalman filter (EnKf) after stratified resampling. The technique is evaluated using a modified "storm track" Lorenz '96 model in which the chaotic/laminar solution structure changes in space and time. Results obtained are shown to be competitive when compared with different EnKF methods and unprojected OP-PF.

    Mathematics Subject Classification: 93E11, 93C10, 93B17, 60G35.

    Citation:

    \begin{equation} \\ \end{equation}
  • 加载中
  • Figure 1.  Illustration of Theorem 3.3. Each image displays a matrix, and the shading of individual cells suggests the value of the corresponding matrix element. Our main result is as follows. Given an arbitrary observation covariance $ {\bf R} $ (top left) and projection $ {\bf \Pi}_1 $ (here, $ {\bf \Pi}_1 $ projects onto a 2d subspace), we compute the projected covariance $ {\bf R}_1 $ (top middle, shown in data space). Using only knowledge of $ {\bf R} $, $ {\bf \Pi}_1 $, and $ {\bf U}_1 $, we discover an $ {\bf R}_2 $ (top right) that can be expressed in model or data space. The unique property of $ {\bf R}_2 $ is that it allows one to reconstruct the original observation precision $ {\bf R}^{-1} $ as, essentially, the sum of $ {\bf R}_1^\dagger $ and $ {\bf R}_2^\dagger $. The visualization above is one way to understand our main result. Theorem 3.3 states the equivalent result $ (y - {\bf H} u)^T{\bf R}^{\dagger}(y - {\bf H} u) = (y_1 - {\bf H}_1 u)^T {\bf R}_1^{\dagger} (y_1 - {\bf H}_1 u) + (y_2 - {\bf H}_2 u)^T {\bf R}_2^{(g)} (y_2 - {\bf H}_2 u) $

    Figure 2.  Consider an ensemble of ten evenly weighted particles $ w^i = 1/10 $ for $ i = 1, ..., 10. $ The image plots the output of resampling methods after application to this ensemble. The number of resampled particles coming from the $ i $th original particle are shown by the number of markers in the $ i $th cell. The multinomial resampler (red crosses) returns two copies of the second, sixth, and eighth particles, and none of the first, seventh, or tenth. The stratified resampler (blue circles) returns one of each particle, which is clearly the optimal result

    Figure 3.  Error and condition number ratios for repeated experiments decomposing $ {\bf R} $ into $ {\bf R}_1, \, {\bf R}_2 $. Left: the reconstruction error, which is just the distance between $ (y - {\bf H} u)^T{\bf R}^{-1}(y - {\bf H} u) $ and $ \left[(y_1 - {\bf H}_1 u)^T {\bf R}_1^{\dagger} (y_1 - {\bf H}_1 u) + (y_2 - {\bf H}_2 u)^T {\bf R}_2^{(g)} (y_2 - {\bf H}_2 u)\right] $, is bounded by $ 10^{-13} $. Right: the condition number for $ {\bf R}_2 $ is typically equal to the condition number for $ {\bf R} $, but can be up to $ 100 $ times larger. The process of generating these results is fully described at the beginning of Section 5

    Figure 4.  Prior and data for the first two variables of the proof of concept in Section 5.1. The 20 particles comprising the prior ensemble are shown by the center of each circle. The radius of the circles indicates the prior uncertainty in each particle. The red line shows the data: we observe the horizontal coordinate, but do not observe the vertical coordinate. As the vertical coordinate $ u_2 $ is not correlated with any other variable, the two intersections between the red line and the prior ensemble suggest that the posterior would be bimodal

    Figure 5.  Posteriors for the first two variables of the proof of concept in Section 5.1. In the top row, we show the posteriors for a small ensemble. For both ensemble sizes $ N = 20 $ and $ N = 1000 $, the Pf-EnKF algorithm captures the bimodal posterior. It has an ESS around $ 0.22 N $, retains about $ 30\% $ of particles after resampling, and estimates the remaining variables accurately (shown in Table 1). In contrast, the Op-PF requires the larger ensemble size to capture the bimodal posterior. The EnKF (at either $ N $) approximately captures the posterior mean and variance, but concentrates the majority of its' mass in between the two modes of the posterior

    Figure 6.  Performance of DA algorithms for the standard L96 problem described in Section 5.2. To better distinguish between the three stable schemes, the RMSE for time 50-550 plots the moving average over 50 assimilation cycles. Mean RMSE and spread are: 0.19 and 0.21 for ETKF; 0.26 and 0.25 for the stochastic EnKF; 0.28 and 0.23 for Pf-EnKF. The mean ESS in the Pf-EnKF is $ 19.85 $ (out of twenty). The horizontal dotted line shows the expected observation error. As all three stable schemes have spread approximately equal to RMSE and significantly less than the observation standard deviation of 1, they appear well tuned. The Op-PF RMSE is similar to not performing assimilation at all

    Figure 7.  Space-time plot of typical solution to "storm track" Lorenz '96 model

    Figure 8.  RMSE plots for the 'storm track' model. The Pf-EnKF and Op-PF are stable and more accurate than ETKF or the stochastic EnKF. Mean RMSE and spread are: 0.37 and 0.37 for PF-EnKF; 0.40 and 0.31 for Op-PF; 1.36 and 0.13 for stochastic EnKF. The mean ESS in the Pf-EnKF is $ 18 $ (out of twenty). RMSE is plotted unaltered for the burn-in (time 0 to 50), and for clarity is plotted as a moving average over 50 points for time 50-550

    Figure 9.  Rank histograms for the plots in Figure 8. In optimal filtering scenarios, the rank histogram would be flat. The peaks at 0 and 1 indicate observations that are outside the span of the ensemble. With the model error (forcing $ F = 5 $ in the forecast model, while $ F = 8 $ generated data), all methods consistently under-estimate the true state, but Pf-EnKF does so the least

    Table 1.  Posterior error statistics for all $ 900 $ variables of the proof of concept in Section 5.1. The EnKF and Pf-EnKF obtain comparable RMSE and spread with $ N = 20 $ and $ N = 1000 $ ensemble members, while the Op-PF is degenerate (visible here as a spread of $ 0 $ in the posterior ensemble) with $ N = 20 $, and has a small ensemble size with $ N = 1000 $

    RMSE Spread ESS
    Pf-EnKF ($ N=20 $) 0.07 0.06 5.6
    Op-PF ($ N={20} $) 0.13 $ {\bf 0} $ $ {\bf 1} $
    EnKF ($ N=20 $) 0.08 0.05
    Pf-EnKF ($ N=1000 $) 0.09 0.07 225
    Op-PF ($ N={1000} $) 0.10 $ 0.06 $ $ 14.5 $
    EnKF ($ N=1000 $) 0.07 0.05
     | Show Table
    DownLoad: CSV
  • [1] A. AlbarakatiM. BudišićR. CrockerJ. Glass-KlaiberS. IamsJ. MacleanN. MarshallC. Roberts and E. S. Van Vleck, Model and data reduction for data assimilation: Particle filters employing projected forecasts and data with application to a shallow water model, Comput. Math. Appl., 116 (2022), 194-211. 
    [2] A. AlbarakatiM. Budišić and E. S. Van Vleck, Projected data assimilation using sliding window proper orthogonal decomposition, Journal of Computational Physics, 514 (2024), 113235.  doi: 10.1016/j.jcp.2024.113235.
    [3] J. Anderson, A method for producing and evaluating probabilistic forecasts from ensemble model integrations, Journal of Climate, 9 (1996), 1518-1530. 
    [4] P. BennerS. Gugercin and K. Willcox, A survey of projection-based model reduction methods for parametric dynamical systems, SIAM Review, 57 (2015), 483-531.  doi: 10.1137/130932715.
    [5] C. H. BishopJ. S. Whitaker and L. Lei, Gain form of the ensemble transform Kalman filter and its relevance to satellite data assimilation with model space ensemble covariance localization, Monthly Weather Review, 145 (2017), 4575-4592.  doi: 10.1175/MWR-D-17-0102.1.
    [6] J.-L. Blanco, Resampling Methods for Particle Filtering, 2023.
    [7] M. Bocquet and A. Carrassi, Four-dimensional ensemble variational data assimilation and the unstable subspace, Tellus A: Dynamic Meteorology and Oceanography, 69 (2017), 1304504.  doi: 10.1080/16000870.2017.1304504.
    [8] S. Boyd and  L. VandenbergheConvex Optimization, Cambridge University Press, 2004. 
    [9] S. L. BruntonM. BudišićE. Kaiser and J. N. Kutz, Modern Koopman theory for dynamical systems, SIAM Review, 64 (2022), 229-340.  doi: 10.1137/21M1401243.
    [10] A. CarrassiA. TrevisanL. DescampsO. Talagrand and F. Uboldi, Controlling instabilities along a 3DVar analysis cycle by assimilating in the unstable subspace: A comparison with the EnKF, Nonlinear Process. Geophys., 15 (2008), 503-521. 
    [11] A. Carrassi, M. Bocquet, J. Demaeyer, C. Grudzien, P. Raanes and S. Vannitsem, Data Assimilation for Chaotic Dynamics, Springer International Publishing, Cham, 2022. doi: 10.1007/978-3-030-77722-7_1.
    [12] D. Crisan and K. Li, Generalised particle filters with Gaussian mixtures, Stochastic Processes and their Applications, 125 (2015), 2643-2673.  doi: 10.1016/j.spa.2015.01.008.
    [13] B. de LeeuwS. DubinkinaJ. FrankA. SteyerX. Tu and E. Van Vleck, Projected shadowing-based data assimilation, SIAM J. Appl. Dyn. Syst., 17 (2018), 2446-2477.  doi: 10.1137/17M1141163.
    [14] A. DoucetS. Godsill and C. Andrieu, On sequential Monte Carlo sampling methods for Bayesian filtering, Statistics and Computing, 10 (2000), 197-208.  doi: 10.1023/A:1008935410038.
    [15] J. Gallier, The Schur complement and symmetric positive semidefinite (and definite) matrices, 2019, http:/www.cis.upenn.edu/~jean/schur-comp.pdf.
    [16] J. H. Kotecha and P. M. Djuric, Gaussian sum particle filtering, IEEE Transactions on Signal Processing, 51 (2003), 2602-2612.  doi: 10.1109/TSP.2003.816754.
    [17] S. KotsukiT. MiyoshiK. Kondo and R. Potthast, A local particle filter and its gaussian mixture extension implemented with minor modifications to the LETKF, Geoscientific Model Development, 15 (2022), 8325-8348. 
    [18] Y. Lee and A. J. Majda, Multiscale methods for data assimilation in turbulent systems, Multiscale Modeling and Simulation, 13 (2015), 691-713. 
    [19] T. LiM. Bolic and P. M. Djuric, Resampling Methods for Particle Filtering: Classification, implementation, and strategies, IEEE Signal Processing Magazine, 32 (2015), 70-86.  doi: 10.1109/MSP.2014.2330626.
    [20] E. N. Lorenz, Predictability - a problem partly solved, Proceedings of seminar on Predictability, 1 (1996), 1-18. 
    [21] J. Maclean and E. S. Van Vleck, Particle filters for data assimilation based on reduced-order data models, Quarterly Journal of the Royal Meteorological Society, 147 (2021), 1892-1907. doi: 10.1002/qj.4001.
    [22] A. J. MajdaD. Qi and T. P. Sapsis, Blended particle filters for large-dimensional chaotic dynamical systems, Proc. Natl. Acad. Sci. USA, 111 (2014), 7511-7516.  doi: 10.1073/pnas.1405675111.
    [23] L. PalatellaA. Carrassi and A. Trevisan, Lyapunov vectors and assimilation in the unstable subspace: Theory and applications, Journal of Physics A: Mathematical and Theoretical, 46 (2013), 254020. 
    [24] I. PasmansY. ChenA. Carrassi and C. K. R. T. Jones, Tailoring data assimilation to discontinuous Galerkin models, Quarterly Journal of the Royal Meteorological Society, 150 (2024), 2820-2847.  doi: 10.1002/qj.4737.
    [25] R. PotthastA. Walter and A. Rhodin, A localized adaptive particle filter within an operational NWP framework, Monthly Weather Review, 147 (2019), 345-362.  doi: 10.1175/MWR-D-18-0028.1.
    [26] D. Qi and A. J. Majda, Blended particle methods with adaptive subspaces for filtering turbulent dynamical systems, Phys. D, 298/299 (2015), 21-41.  doi: 10.1016/j.physd.2015.02.002.
    [27] A. RojahnN. SchenkP. J. V. Leeuwen and R. Potthast, Particle filtering and gaussian mixtures – on a localized mixture coefficients particle filter (lmcpf) for global nwp, Journal of the Meteorological Society of Japan, 101 (2023), 233-253.  doi: 10.2151/jmsj.2023-015.
    [28] L. SlivinskiE. SpillerA. Apte and B. Sandstede, A hybrid particle-ensemble kalman filter for lagrangian data assimilation, Monthly Weather Review, 143 (2015), 195-211.  doi: 10.1175/MWR-D-14-00051.1.
    [29] C. Snyder, Particle filters, the "optimal" proposal and high-dimensional systems, In Proceedings of the ECMWF Seminar on Data Assimilation for Atmosphere and Ocean, (2011), 1-10.
    [30] C. SnyderT. BengtssonP. Bickel and J. Anderson, Obstacles to high-dimensional particle filtering, Monthly Weather Review, 136 (2008), 4629-4640.  doi: 10.1175/2008MWR2529.1.
    [31] T. Sondergaard and P. F. Lermusiaux, Data assimilation with Gaussian mixture models using the dynamically orthogonal field equations. Part Ⅰ: Theory and scheme, Monthly Weather Review, 141 (2013), 1737-1760. 
    [32] A. S. StordalH. A. KarlsenG. NævdalH. J. Skaug and B. Vallès, Bridging the ensemble Kalman filter and particle filters: The adaptive Gaussian mixture filter, Computational Geosciences, 15 (2011), 293-305.  doi: 10.1007/s10596-010-9207-1.
    [33] A. Trevisan, M. D'Isidoro and O. Talagrand, Four-dimensional variational assimilation in the unstable subspace and the optimal subspace dimension, Q. J. R. Meteorol. Soc., 136 (2010), 487-496. doi: 10.1002/qj.571.
    [34] A. Trevisan and F. Uboldi, Assimilation of standard and targeted observations within the unstable subspace of the observation-analysis-forecast cycle system, Journal of the Atmospheric Sciences, 61 (2004), 103-113. 
    [35] P. J. V. Leeuwen, Particle filters for the geosciences, Advanced Data Assimilation for Geosciences: Lecture Notes of the Les Houches School of Physics: Special Issue, (2012), 291.  doi: 10.1093/acprof:oso/9780198723844.003.0013.
    [36] X. WangH. G. ChipilskiC. H. BishopE. SatterfieldN. Baker and J. S. Whitaker, A multiscale local gain form ensemble transform Kalman filter (MLGETKF), Monthly Weather Review, 149 (2021), 605-622.  doi: 10.1175/MWR-D-20-0290.1.
  • 加载中

Figures(9)

Tables(1)

SHARE

Article Metrics

HTML views(1529) PDF downloads(203) Cited by(0)

Access History

Other Articles By Authors

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return