Thu, 01 Dec 2022
16:00
Virtual

Particle filters for Data Assimilation

Dan Crisan
(Imperial College London)

Note: we would recommend to join the meeting using the Teams client for best user experience.

Further Information
Abstract

Modern Data Assimilation (DA) can be traced back to the sixties and owes a lot to earlier developments in linear filtering theory. Since then, DA has evolved independently of Filtering Theory. To-date it is a massively important area of research due to its many applications in meteorology, ocean prediction, hydrology, oil reservoir exploration, etc. The field has been largely driven by practitioners, however in recent years an increasing body of theoretical work has been devoted to it. In this talk, In my talk, I will advocate the interpretation of DA through the language of stochastic filtering. This interpretation allows us to make use of advanced particle filters to produce rigorously validated DA methodologies. I will present a particle filter that incorporates three additional add-on procedures: nudging, tempering and jittering. The particle filter is tested on a two-layer quasi-geostrophic model with O(10^6) degrees of freedom out of which only a minute fraction are noisily observed.

A new mixed finite-element method for H2 elliptic problems
Farrell, P Hamdan, A MacLachlan, S Computers and Mathematics with Applications volume 128 300-319 (15 Dec 2022)
Analysis and Modeling of Client Order Flow in Limit Order Markets
CONT, R CUCURINGU, M Glukhov, V Prenzel, F Quantitative Finance
Theoretical study of the emergence of periodic solutions for the inhibitory NNLIF neuron model with synaptic delay
Ikeda, K Roux, P Salort, D Smets, D Mathematical Neuroscience and Applications volume 2 (26 Oct 2022)
Tue, 01 Nov 2022

12:30 - 13:00
C3

Asymptotic Analysis of Deep Residual Networks

Alain Rossier
Abstract

Residual networks (ResNets) have displayed impressive results in pattern recognition and, recently, have garnered considerable theoretical interest due to a perceived link with neural ordinary differential equations (neural ODEs). This link relies on the convergence of network weights to a smooth function as the number of layers increases. We investigate the properties of weights trained by stochastic gradient descent and their scaling with network depth through detailed numerical experiments. We observe the existence of scaling regimes markedly different from those assumed in neural ODE literature. Depending on certain features of the network architecture, such as the smoothness of the activation function, one may obtain an alternative ODE limit, a stochastic differential equation (SDE) or neither of these. Furthermore, we are able to formally prove the linear convergence of gradient descent to a global optimum for the training of deep residual networks with constant layer width and smooth activation function. We further prove that if the trained weights, as a function of the layer index, admit a scaling limit as the depth increases, then the limit has finite 2-variation.

An adaptive random bit multilevel algorithm for SDEs
Giles, M Hefter, M Mayer, L Ritter, K Multivariate Algorithms and Information-Based Complexity 15-31 (08 Jun 2020)
Dynamic calibration of order flow models with generative adversarial networks
Prenzel, F Cont, R Cucuringu, M Kochems, J ICAIF '22: 3rd ACM International Conference on AI in Finance (26 Oct 2022)
Multicore quantum computing
Jnane, H Undseth, B Cai, Z Benjamin, S Koczor, B Physical Review Applied volume 18 issue 4 (26 Oct 2022)
Relative defects in relative theories: trapped higher-form symmetries and irregular punctures in class S
Bhardwaj, L Giacomelli, S Hübner, M Schäfer-Nameki, S SciPost Physics volume 13 issue 4 (26 Oct 2022)
Thu, 01 Dec 2022

16:00 - 17:00
L3

Convergence of policy gradient methods for finite-horizon stochastic linear-quadratic control problems

Michael Giegrich
Abstract

We study the global linear convergence of policy gradient (PG) methods for finite-horizon exploratory linear-quadratic control (LQC) problems. The setting includes stochastic LQC problems with indefinite costs and allows additional entropy regularisers in the objective. We consider a continuous-time Gaussian policy whose mean is linear in the state variable and whose covariance is state-independent. Contrary to discrete-time problems, the cost is noncoercive in the policy and not all descent directions lead to bounded iterates. We propose geometry-aware gradient descents for the mean and covariance of the policy using the Fisher geometry and the Bures-Wasserstein geometry, respectively. The policy iterates are shown to obey an a-priori bound, and converge globally to the optimal policy with a linear rate. We further propose a novel PG method with discrete-time policies. The algorithm leverages the continuous-time analysis, and achieves a robust linear convergence across different action frequencies. A numerical experiment confirms the convergence and robustness of the proposed algorithm.

This is joint work with Yufei Zhang and Christoph Reisinger.

Subscribe to