Forthcoming Seminars

Please note that the list below only shows forthcoming events, which may not include regular events that have not yet been entered for the forthcoming term. Please see the past events page for a list of all seminar series that the department has on offer.

Past events in this series
1 March 2018
14:00
Prof Jan Hesthaven
Abstract

The development of reduced order models for complex applications, offering the promise for rapid and accurate evaluation of the output of complex models under parameterized variation, remains a very active research area. Applications are found in problems which require many evaluations, sampled over a potentially large parameter space, such as in optimization, control, uncertainty quantification and applications where near real-time response is needed.

However, many challenges remain to secure the flexibility, robustness, and efficiency needed for general large-scale applications, in particular for nonlinear and/or time-dependent problems.

After giving a brief general introduction to reduced order models, we discuss developments in two different directions. In the first part, we discuss recent developments of reduced methods that conserve chosen invariants for nonlinear time-dependent problems. We pay particular attention to the development of reduced models for Hamiltonian problems and propose a greedy approach to build the basis. As we shall demonstrate, attention to the construction of the basis must be paid not only to ensure accuracy but also to ensure stability of the reduced model. Time permitting, we shall also briefly discuss how to extend the approach to include more general dissipative problems through the notion of port-Hamiltonians, resulting in reduced models that remain stable even in the limit of vanishing viscosity and also touch on extensions to Euler and Navier-Stokes equations.

The second part of the talk discusses the combination of reduced order modeling for nonlinear problems with the use of neural networks to overcome known problems of on-line efficiency for general nonlinear problems. We discuss the general idea in which training of the neural network becomes part of the offline part and demonstrate its potential through a number of examples, including for the incompressible Navier-Stokes equations with geometric variations.

This work has been done with in collaboration with B.F. Afkram (EPFL, CH), N. Ripamonti EPFL, CH) and S. Ubbiali (USI, CH).

  • Computational Mathematics and Applications Seminar

Pages

Add to My Calendar