Cerfacs Enter the world of high performance ...

🎓Mathis PEYRON thesis defense

  Tuesday 8 October 2024From 10h00 at 12h00

  Phd Thesis       JCA Room, CERFACS, Toulouse, France    

Latent space data assimilation by using deep learning

This thesis, which sits at the crossroads of data assimilation (DA) and deep learning (DL), introduces latent space data assimilation, a novel data-driven framework that significantly reduces computational costs and memory requirements, while also offering the potential for more accurate data assimilation results. Importantly, this method is agnostic to the specific data assimilation algorithm and neural network architecture used.

We extend the integration of deep learning by rethinking the assimilation process itself. Our approach aligns with reduced-space methods, which solve the assimilation problem by performing computations within a lower-dimensional space relative. These reduced-space methods have been developed primarily for operational use, as most data assimilation algorithms are prohibitively computationally costly, when implemented in their full theoretically form.

Our methodology is based on the joint training of an autoencoder and a surrogate neural network. The autoencoder iteratively learns how to accurately represent the physical dynamics of interest within a low-dimensional space, termed latent space. The surrogate is simultaneously trained to learn the time propagation of the latent variables. A chained loss function strategy is also proposed to ensure the stability of the surrogate network. Stability can also be achieved by implementing Lipschitz surrogate networks.

Reduced-space data assimilation is underpinned by Lyapunov stability theory, which mathematically demonstrates that, under specific hypotheses, the forecast and posterior error covariance matrices asymptotically conform to the unstable-neutral subspace, which is of much smaller in dimension than the full state space. While full-space data assimilation involves linear combinations within a high-dimensional, nonlinear, and possibly multi-scale dynamic environment, latent data assimilation, which operates on the core, potentially disentangled and simplified dynamics, is more likely to result in impactful corrections. Furthermore, classical data assimilation faces an inherent mathematical limitation in that it relies on linear computations. Latent data assimilation, however, offers a way to overcome this limitation by performing the assimilation directly within the meaningful underlying structures of the data, obtained through nonlinear transformations.

We test our methodology on a 400-dimensional dynamics – built upon a chaotic Lorenz96 system of dimension 40 -, and on the quasi-geostrophic model of the Object-Oriented Prediction System (OOPS) framework, developed by the ECMWF (European Centre for Medium-Range Weather Forecasts).

Jury

Arthur VIDARDResearch Fellow at INRIA GrenobleReviewer
Tijana JANJICProfessor at the Catholic University of Eichstätt-IngolstadtReviewer
Ronan FABLETProfessor at IMT AtlantiqueExaminer
Alban FARCHIResearcher at ECMWFExaminer
Selime GÜROLResearcher at CERFACSSupervisor
Serge GRATTON University Professor at Toulouse INPCo-Supervisor

CALENDAR

Monday

07

October

2024

Machine learning for data science

From Monday 7 October 2024 to Thursday 10 October 2024

  Training    

Tuesday

08

October

2024

🎓Mathis PEYRON thesis defense

Tuesday 8 October 2024From 10h00 at 12h00

  Phd Thesis       JCA Room, CERFACS, Toulouse, France    

Monday

14

October

2024

Fundamentals to understand and analyse high fidelity compressible Large Eddy Simulations

From Monday 14 October 2024 to Wednesday 16 October 2024

  Training       Cerfacs, Toulouse, France    

ALL EVENTS