Tue, 10 Mar 2020

15:30 - 16:30
L6

Random matrices, random Young diagrams, and some random operators

Sasha Sodin
(Queen Mary University of London)
Abstract

The rows of a Young diagram chosen at random with respect to the Plancherel measure are known to share some features with the eigenvalues of the Gaussian Unitary Ensemble. We shall discuss several ideas, going back to the work of Kerov and developed by Biane and by Okounkov, which to some extent clarify this similarity. Partially based on joint work with Jeong and on joint works in progress with Feldheim and Jeong and with Täufer.

Tue, 18 Feb 2020

15:30 - 16:30
L6

Araç Kasko Değeri Sorgulama

Cosme Louart
(Univ. Grenoble Alpes)
Abstract

This presentation introduces a rigorous framework for the study of commonly used machine learning techniques (kernel methods, random feature maps, etc.) in the regime of large dimensional and numerous data. Exploiting the fact that very realistic data can be modeled by generative models (such as GANs), which are theoretically concentrated random vectors, we introduce a joint random matrix and concentration of measure theory for data processing. Specifically, we present fundamental random matrix results for concentrated random vectors, which we apply to the performance estimation of spectral clustering on real image datasets.

Tue, 25 Feb 2020

15:30 - 16:30
L6

Randomised algorithms for computing low rank approximations of matrices

Per-Gunnar Martinsson
(U.T. Austin)
Abstract

The talk will describe how ideas from random matrix theory can be leveraged to effectively, accurately, and reliably solve important problems that arise in data analytics and large scale matrix computations. We will focus in particular on accelerated techniques for computing low rank approximations to matrices. These techniques rely on randomised embeddings that reduce the effective dimensionality of intermediate steps in the computation. The resulting algorithms are particularly well suited for processing very large data sets.

The algorithms described are supported by rigorous analysis that depends on probabilistic bounds on the singular values of rectangular Gaussian matrices. The talk will briefly review some representative results.

Note: There is a related talk in the Computational Mathematics and Applications seminar on Thursday Feb 27, at 14:00 in L4. There, the ideas introduced in this talk will be extended to the problem of solving large systems of linear equations.

Thu, 06 Feb 2020

16:00 - 17:00
L4

Eigenvector overlaps for large random matrices and applications to financial data

Jean Philippe Bouchaud
(Capital Fund Management)
Abstract

Whereas the spectral properties of random matrices has been the subject of numerous studies and is well understood, the statistical properties of the corresponding eigenvectors has only been investigated in the last few years. We will review several recent results and emphasize their importance for cleaning empirical covariance matrices, a subject of great importance for financial applications.

Thu, 12 Mar 2020

14:00 - 15:00
L4

The Statistical Finite Element Method

Mark Girolami
(University of Cambridge)
Abstract

The finite element method (FEM) is one of the great triumphs of applied mathematics, numerical analysis and software development. Recent developments in sensor and signalling technologies enable the phenomenological study of systems. The connection between sensor data and FEM is restricted to solving inverse problems placing unwarranted faith in the fidelity of the mathematical description of the system. If one concedes mis-specification between generative reality and the FEM then a framework to systematically characterise this uncertainty is required. This talk will present a statistical construction of the FEM which systematically blends mathematical description with observations.

>

Thu, 27 Feb 2020

14:00 - 15:00
L4

Randomised algorithms for solving systems of linear equations

Gunnar Martinsson
(University of Texas at Austin)
Abstract

The task of solving large scale linear algebraic problems such as factorising matrices or solving linear systems is of central importance in many areas of scientific computing, as well as in data analysis and computational statistics. The talk will describe how randomisation can be used to design algorithms that in many environments have both better asymptotic complexities and better practical speed than standard deterministic methods.

The talk will in particular focus on randomised algorithms for solving large systems of linear equations. Both direct solution techniques based on fast factorisations of the coefficient matrix, and techniques based on randomised preconditioners, will be covered.

Note: There is a related talk in the Random Matrix Seminar on Tuesday Feb 25, at 15:30 in L4. That talk describes randomised methods for computing low rank approximations to matrices. The two talks are independent, but the Tuesday one introduces some of the analytical framework that supports the methods described here.

Thu, 20 Feb 2020

14:00 - 15:00
Rutherford Appleton Laboratory, nr Didcot

Learning with nonlinear Perron eigenvectors

Francesco Tudisco
(Gran Sasso Science Institute GSSI)
Abstract

In this talk I will present a Perron-Frobenius type result for nonlinear eigenvector problems which allows us to compute the global maximum of a class of constrained nonconvex optimization problems involving multihomogeneous functions.

I will structure the talk into three main parts:

First, I will motivate the optimization of homogeneous functions from a graph partitioning point of view, showing an intriguing generalization of the famous Cheeger inequality.

Second, I will define the concept of multihomogeneous function and I will state our main Perron-Frobenious theorem. This theorem exploits the connection between optimization of multihomogeneous functions and nonlinear eigenvectors to provide an optimization scheme that has global convergence guarantees.

Third, I will discuss a few example applications in network science and machine learning that require the optimization of multihomogeneous functions and that can be solved using nonlinear Perron eigenvectors.

 

 

Thu, 13 Feb 2020

14:00 - 15:00
L4

Numerical real algebraic geometry and applications

Jonathan Hauenstein
(University of Notre Dame)
Abstract

Systems of nonlinear polynomial equations arise in a variety of fields in mathematics, science, and engineering.  Many numerical techniques for solving and analyzing solution sets of polynomial equations over the complex numbers, collectively called numerical algebraic geometry, have been developed over the past several decades.  However, since real solutions are the only solutions of interest in many applications, there is a current emphasis on developing new methods for computing and analyzing real solution sets.  This talk will summarize some numerical real algebraic geometric approaches as well as recent successes of these methods for solving a variety of problems in science and engineering.

Thu, 06 Feb 2020

14:00 - 15:00
L4

Quantifying the Estimation Error of Principal Component

Raphael Hauser
(University of Oxford)
Abstract

(Joint work with: Jüri Lember, Heinrich Matzinger, Raul Kangro)

Principal component analysis is an important pattern recognition and dimensionality reduction tool in many applications and are computed as eigenvectors

of a maximum likelihood covariance that approximates a population covariance. The eigenvectors are often used to extract structural information about the variables (or attributes) of the studied population. Since PCA is based on the eigen-decomposition of the proxy covariance rather than the ground-truth, it is important to understand the approximation error in each individual eigenvector as a function of the number of available samples. The combination of recent results of Koltchinskii & Lounici [8] and Yu, Wang & Samworth [11] yields such bounds. In the presented work we sharpen these bounds and show that eigenvectors can often be reconstructed to a required accuracy from a sample of strictly smaller size order.

Subscribe to