演讲者:唐琪(佐治亚理工学院)
时间:2024-11-19 10:00-11:00
地点:腾讯会议 ID: 944-585-979
Abstract
In this talk I will discuss our recent effort to develop structure-preserving machine learning (ML) for time series data, focusing on both dissipative PDEs and singularly perturbed ODEs.
The first part presents a data-driven modeling method that accurately captures shocks and chaotic dynamics through a stabilized neural ODE framework. We learn the right-hand-side of an ODE by adding the outputs of two networks together, one learning a linear term and the other a nonlinear term. The architecture is inspired by the inertial manifold theorem. We apply this method to chaotic trajectories of the Kuramoto-Sivashinsky equation, where our model keeps long-term trajectories on the attractor and remains robust to noisy initial conditions.
The second part explores structure-preserving ML for singularly perturbed dynamical systems. A powerful tool to address these systems is the Fenichel normal form, which significantly simplifies fast dynamics near slow manifolds. I will discuss a novel realization of this concept using ML. Specifically, a fast-slow neural network (FSNN) is proposed, enforcing the existence of a trainable, attractive invariant slow manifold as a hard constraint. To illustrate the power of FSNN, I will show a fusion-motivated example where traditional numerical integrators all fail.