Thu, 23 Nov 2023
16:00
Lecture Room 4, Mathematical Institute

Mean-field Analysis of Generalization Errors

Dr Gholamali Aminian
(Alan Turing Institute)
Abstract

We propose a novel framework for exploring weak and $L_2$ generalization errors of algorithms through the lens of differential calculus on the space of probability measures. Specifically, we consider the KL-regularized empirical risk minimization problem and establish generic conditions under which the generalization error convergence rate, when training on a sample of size $n$ , is $\matcal{O}(1/n)$. In the context of supervised learning with a one-hidden layer neural network in the mean-field regime, these conditions are reflected in suitable integrability and regularity assumptions on the loss and activation functions.

Thu, 17 Jun 2021

16:00 - 17:00

Identifiability in inverse stochastic optimal control

HAOYANG CAO
(Alan Turing Institute)
Abstract

Abstract: In this work, we analyze a class of stochastic inverse optimal control problems with entropy regularization. We first characterize the set of solutions for the inverse control problem. This solution set exemplifies the issue of degeneracy in generic inverse control problems that there exist multiple reward or cost functions that can explain the displayed optimal behavior. Then we establish one resolution for the degeneracy issue by providing one additional optimal policy under a different discount factor. This resolution does not depend on any prior knowledge of the solution set. Through a simple numerical experiment with deterministic transition kernel, we demonstrate the ability of accurately extracting the cost function through our proposed resolution.

 

Joint work with Sam Cohen (Oxford) and Lukasz Szpruch (Edinburgh).

Fri, 06 Nov 2020

12:00 - 13:00

Bridging GANs and Stochastic Analysis

Haoyang Cao
(Alan Turing Institute)
Abstract

Generative adversarial networks (GANs) have enjoyed tremendous success in image generation and processing, and have recently attracted growing interests in other fields of applications. In this talk we will start from analyzing the connection between GANs and mean field games (MFGs) as well as optimal transport (OT). We will first show a conceptual connection between GANs and MFGs: MFGs have the structure of GANs, and GANs are MFGs under the Pareto Optimality criterion. Interpreting MFGs as GANs, on one hand, will enable a GANs-based algorithm (MFGANs) to solve MFGs: one neural network (NN) for the backward Hamilton-Jacobi-Bellman (HJB) equation and one NN for the Fokker-Planck (FP) equation, with the two NNs trained in an adversarial way. Viewing GANs as MFGs, on the other hand, will reveal a new and probabilistic aspect of GANs. This new perspective, moreover, will lead to an analytical connection between GANs and Optimal Transport (OT) problems, and sufficient conditions for the minimax games of GANs to be reformulated in the framework of OT. Building up from the probabilistic views of GANs, we will then establish the approximation of GANs training via stochastic differential equations and demonstrate the convergence of GANs training via invariant measures of SDEs under proper conditions. This stochastic analysis for GANs training can serve as an analytical tool to study its evolution and stability.

 
Tue, 25 Feb 2020

12:00 - 13:00
C1

A framework for constructing generative models of mesoscale structure in multilayer networks

Marya Bazzi
(Alan Turing Institute)
Abstract

Multilayer networks are a way to represent dependent connectivity patterns — e.g., time-dependence, multiple types of interactions, or both — that arise in many applications and which are difficult to incorporate into standard network representations. In the study of multilayer networks, it is important to investigate mesoscale (i.e., intermediate-scale) structures, such as communities, to discover features that lie between the microscale and the macroscale. We introduce a framework for the construction of generative models for mesoscale structure in multilayer networks.  We model dependency at the level of partitions rather than with respect to edges, and treat the process of generating a multilayer partition separately from the process of generating edges for a given multilayer partition. Our framework can admit many features of empirical multilayer networks and explicitly incorporates a user-specified interlayer dependency structure. We discuss the parameters and some properties of our framework, and illustrate an example of its use with benchmark models for multilayer community-detection tools. 

 

Subscribe to Alan Turing Institute