Home Research Teaching Publications Presentations  
My current research focuses on learning dynamics from data. One topic is nonparametric learning of the interaction laws in systems of interacting particles/agents, and another is data-driven model reduction for complex systems in computation such as fluid dynamics and molecular dynamics simulation. I view dynamical systems as a description of stochastic processes and take an inference approach to learn the dynamics from data, so I am also interested in closely related topics such as data assimilation, sequential Monte Carlo methods, deterministic and stochastic dynamical systems and PDEs, ergodicity theory, and learning theory.

1. Data-driven stochastic model reduction for complex dynamics

Space-time reduction: dimension reduction in space and time reduction with large time-stepping
Many problems in science and engineering involve nonlinear dynamical systems that are too complex or computationally expensive for full solutions, but one is interested mainly in a subset of the variables. Such problems appear, for example, in weather and climate modeling, in statistical mechanics, and in the mechanics of turbulent flow. In this setting, it is desirable to construct effective reduced models for the variables of interest using data that come either from observations or from fine numerical simulations. These reduced models are supposed to capture the key statistical and dynamical properties of the original systems, and therefore, stochastic reduced models are often preferred. Statistical inference and machine learning methods are natural tools for the construction of such reduced models for data. We focus on the mathematical understanding of such inference-based data-driven approach for model reduction of complex dynamics.

Discrete-time stochastic parametrization. In collaboration with Alexandre J. Chorin, I proposed a discrete-time parametrization framework to infer from discrete-time partial data a reduced model in the form of NARMAX (nonlinear autoregression moving average with exogenous input). This provides flexibility in the parametrization of memory effects as suggested by Mori-Zwanzig formalism, simplifies the inference from data and accounts for the discretization errors.
  • A. J. Chorin and F. Lu. Discrete approach to stochastic parametrization and dimension reduction in nonlinear dynamics. Proc. Natl. Acad. Sci. USA, 112 (2015), no. 32, 9804-9809.
  • F. Lu, K. K. Lin and A. J. Chorin. Comparison of continuous and discrete-time data-based modeling for hypoelliptic systems. Comm. App. Math. Com. Sc., 11 (2016), no. 2, 187–216.

  • Parametrization of approximate inertial manifolds. The major challenge in NARMAX (and general semi-parametric) inference is to derive a model structure. Together with Kevin K. Lin, we developed a method for deriving structures by parametrizing approximate inertial manifolds. This method applies to dissipative systems with inertial manifolds such as the Kuramoto-Sivashinsky equation.
  • F. Lu, K. K. Lin and A. J. Chorin. Data-based stochastic model reduction for the Kuramoto--Sivashinsky equation. Physica D, 340 (2017), 46-57.

  • Model reduction by statistic learning We view the model reduction as a problem of learning the forward map of the (stochastic) process of interest, which is often the large-scale variables. The reduced model approximates the forward map optimal in a suitable hypothesis function space. Thus, we do not require the existence of an inertial manifold. The major issues are: 1. to learn the forward map, which is often high-dimensional; 2. to quantify the approximation of the stochastic process in distribution. To address 1, we focus on finding structures to reduce the complexity of function so that it can be learned via parametric inference or nonparametric inference of low-dimensional functions. Our efforts aim for a better understanding of the modeling and provide guidance on the development of machine learning techniques.
  • K.K. Lin and F.Lu. Data-driven model reduction, Wiener projection, and the Mori-Zwanzig formalism. arXiv
  • F. Lu. Data-driven model reduction for stochastic Burgers equations. PDF   arXiv2010
  • Slides
  • SMR: from nonlinear Galerkin to parametric inference
  • Data assimilation with reduced models

  • 2. ISALT: inference-based schemes adaptive to large time-stepping

    Reduction in time: focus the computational efforts on the large time scale
    Efficient simulation of SDEs is essential in predictive modeling in multi-time scales. However, due to stiffness and accuracy requirements, most simulations have to forward with small time steps, thus being computationally costly. This is particularly true when the drift is non-global Lipschitz, and an implicit scheme is necessary to be stable and accurate. We introduce a discrete-time flow map approximation framework for inferring schemes adaptive to large time-stepping.
  • Xingjie Li, F. Lu and Felix X.F. Ye. ISALT: Inference-based schemes adaptive to large time-stepping for local Lipschitz ergodic systems.   arXiv2102   PDF   code

  • Related notes

  • Dealiasing of Fast Fourier Transform.
  • Open theoretical problems

  • Minimizer of quartic polynomials.