Thu, 24 Feb 2022
14:00
Virtual

Paving a Path for Polynomial Preconditioning in Parallel Computing

Jennifer Loe
(Sandia National Laboratories)
Abstract

Polynomial preconditioning for linear solvers is well-known but not frequently used in current scientific applications.  Furthermore, polynomial preconditioning has long been touted as well-suited for parallel computing; does this claim still hold in our new world of GPU-dominated machines?  We give details of the GMRES polynomial preconditioner and discuss its simple implementation, its properties such as eigenvalue remapping, and choices such as the starting vector and added roots.  We compare polynomial preconditioned GMRES to related methods such as FGMRES and full GMRES without restarting. We conclude with initial evaluations of the polynomial preconditioner for parallel and GPU computing, further discussing how polynomial preconditioning can become useful to real-word applications.

 

 

A link for this talk will be sent to our mailing list a day or two in advance.  If you are not on the list and wish to be sent a link, please contact @email.

Mon, 24 Jan 2022
12:45
Virtual

Factorization in Quantum Gravity and Supersymmetry

Murat Kologlu
(Oxford)
Abstract

One of the lasting puzzles in quantum gravity is whether the holographic description of a gravitational system is a single quantum mechanical theory or the disorder average of many. In the latter case, multiple copies of boundary observables do not factorize into a product, but rather have higher moments. These correlations are interpreted in the bulk as due to geometries involving spacetime wormholes which connect disjoint boundaries. 

 

I will talk about the question of factorization and the role of wormholes for supersymmetric observables, specifically the supersymmetric index. Working with the Euclidean gravitational path integral, I will start with a bulk prescription for computing the supersymmetric index, which agrees with the usual boundary definition. Concretely, I will focus on the setting of charged black holes in asymptotically flat four-dimensional N=2 ungauged supergravity. In this case, the gravitational index path integral has an infinite family of Kerr-Newman classical saddles with different angular velocities. However, fermionic zero-mode fluctuations annihilate the contribution of each saddle except for a single BPS one which yields the expected value of the index. I will then turn to non-perturbative corrections involving spacetime wormholes, and show that fermionic zero modes are present for all such geometries, making their contributions vanish. This mechanism works for both single- and multi-boundary path integrals. In particular, only disconnected geometries without wormholes contribute to the index path integral, and the factorization puzzle that plagues the black hole partition function is resolved for the supersymmetric index. I will also present all other single-centered geometries that yield non-perturbative contributions to the gravitational index of each boundary. Finally, I will discuss implications and expectations for factorization and the status of supersymmetric ensembles in AdS/CFT in further generality. Talk based on [2107.09062] with Luca Iliesiu and Joaquin Turiaci.

Mon, 14 Feb 2022

14:00 - 15:00
Virtual

The convex geometry of blind deconvolution

Felix Krahmer
(Technical University of Munich)
Abstract

Blind deconvolution problems are ubiquitous in many areas of imaging and technology and have been the object of study for several decades. Recently, motivated by the theory of compressed sensing, a new viewpoint has been introduced, motivated by applications in wireless application, where a signal is transmitted through an unknown channel. Namely, the idea is to randomly embed the signal into a higher dimensional space before transmission. Due to the resulting redundancy, one can hope to recover both the signal and the channel parameters. In this talk we analyze convex approaches based on lifting as they have first been studied by Ahmed et al. (2014). We show that one encounters a fundamentally different geometric behavior as compared to generic bilinear measurements. Namely, for very small levels of deterministic noise, the error bounds based on common paradigms no longer scale linearly in the noise level, but one encounters dimensional constants or a sublinear scaling. For larger - arguably more realistic - noise levels, in contrast, the scaling is again near-linear.

This is joint work with Yulia Kostina (TUM) and Dominik Stöger (KU Eichstätt-Ingolstadt).

Tue, 08 Mar 2022

15:30 - 16:30
Virtual

Learning Rates as a Function of Batch Size: A Random Matrix Theory Approach to Neural Network Training

Stefan Zohren
(University of Oxford)
Abstract

In this talk we cover recent work in collaboration with Diego Granziol and Steve Roberts where we study the effect of mini-batching on the loss landscape of deep neural networks using spiked, field-dependent random matrix theory. We demonstrate that the magnitude of the extremal values of the batch Hessian are larger than those of the empirical Hessian and derive an analytical expressions for the maximal learning rates as a function of batch size, informing practical training regimens for both stochastic gradient descent (linear scaling) and adaptive algorithms, such as Adam (square root scaling), for smooth, non-convex deep neural networks. Whilst the linear scaling for stochastic gradient descent has been derived under more restrictive conditions, which we generalise, the square root scaling rule for adaptive optimisers is, to our knowledge, completely novel. For stochastic second-order methods and adaptive methods, we derive that the minimal damping coefficient is proportional to the ratio of the learning rate to batch size. We validate our claims on the VGG/WideResNet architectures on the CIFAR-100 and ImageNet datasets. 

Thu, 03 Feb 2022
14:00
Virtual

Defect CFTs

Maria Nocchi
((Oxford University))
Abstract

Junior Strings is a seminar series where DPhil students present topics of common interest that do not necessarily overlap with their own research area. This is primarily aimed at PhD students and post-docs but everyone is welcome

Mon, 07 Feb 2022
12:45
Virtual

TBA

Michael Blake
(Bristol)
Tue, 01 Mar 2022

15:30 - 16:30
Virtual

CLTs for Pair Dependent Statistics of Circular Beta Ensembles

Ander Aguirre
(University of California Davis)
Abstract

In this talk, we give an overview of recent results on the fluctuation of the statistic $\sum_{i\neq j} f(L_N(\theta_i-\theta_j))$ for the Circular Beta Ensemble in the global, mesoscopic and local regimes. This work is morally related to Johansson's 1988 CLT for the linear statistic $\sum_i f(\theta_i)$ and Lambert's subsequent 2019 extension to the mesoscopic regime. The special case of the CUE ($\beta=2$) in the local regime $L_N=N$ is motivated by Montgomery's study of pair correlations of the rescaled zeros of the Riemann zeta function. Our techniques are of combinatorial nature for the CUE and analytical for $\beta\neq2$.

Mon, 07 Feb 2022

12:45 - 13:45
Virtual

On systems of maximal quantum chaos

Mike Blake
(University of Bristol)
Further Information

Note the unusual time and date

Abstract

A remarkable feature of chaos in many-body quantum systems is the existence of a bound on the quantum Lyapunov exponent. An important question is to understand what is special about maximally chaotic systems which saturate this bound. Here I will discuss a proposal for a `hydrodynamic' origin of chaos in such systems, and discuss hallmarks of maximally chaotic systems. In particular I will discuss how in maximally chaotic systems there is a suppression of exponential growth in commutator squares of generic few-body operators. This suppression appears to indicate that the nature of operator scrambling in maximally chaotic systems is fundamentally different to scrambling in non-maximally chaotic systems.

Mon, 07 Mar 2022

14:00 - 15:00
Virtual

Towards practical estimation of Brenier maps

Jonathan Niles-Weed
(New York University)
Abstract

Given two probability distributions in R^d, a transport map is a function which maps samples from one distribution into samples from the other. For absolutely continuous measures, Brenier proved a remarkable theorem identifying a unique canonical transport map, which is "monotone" in a suitable sense. We study the question of whether this map can be efficiently estimated from samples. The minimax rates for this problem were recently established by Hutter and Rigollet (2021), but the estimator they propose is computationally infeasible in dimensions greater than three. We propose two new estimators---one minimax optimal, one not---which are significantly more practical to compute and implement. The analysis of these estimators is based on new stability results for the optimal transport problem and its regularized variants. Based on joint work with Manole, Balakrishnan, and Wasserman and with Pooladian.

Mon, 21 Feb 2022

14:00 - 15:00
Virtual

Why things don’t work — On the extended Smale's 9th and 18th problems (the limits of AI) and methodological barriers

Anders Hansen
(University of Cambridge)
Abstract

The alchemists wanted to create gold, Hilbert wanted an algorithm to solve Diophantine equations, researchers want to make deep learning robust in AI, MATLAB wants (but fails) to detect when it provides wrong solutions to linear programs etc. Why does one not succeed in so many of these fundamental cases? The reason is typically methodological barriers. The history of  science is full of methodological barriers — reasons for why we never succeed in reaching certain goals. In many cases, this is due to the foundations of mathematics. We will present a new program on methodological barriers and foundations of mathematics,  where — in this talk — we will focus on two basic problems: (1) The instability problem in deep learning: Why do researchers fail to produce stable neural networks in basic classification and computer vision problems that can easily be handled by humans — when one can prove that there exist stable and accurate neural networks? Moreover, AI algorithms can typically not detect when they are wrong, which becomes a serious issue when striving to create trustworthy AI. The problem is more general, as for example MATLAB's linprog routine is incapable of certifying correct solutions of basic linear programs. Thus, we’ll address the following question: (2) Why are algorithms (in AI and computations in general) incapable of determining when they are wrong? These questions are deeply connected to the extended Smale’s 9th and 18th problems on the list of mathematical problems for the 21st century. 

Subscribe to Virtual