Constrained and Multirate Training of Neural Networks
Abstract
I will describe algorithms for regularizing and training deep neural networks. Soft constraints, which add a penalty term to the loss, are typically used as a form ofexplicit regularization for neural network training. In this talk I describe a method for efficiently incorporating constraints into a stochastic gradient Langevin framework for the training of deep neural networks. In contrast to soft constraints, our constraints offer direct control of the parameter space, which allows us to study their effect on generalization. In the second part of the talk, I illustrate the presence of latent multiple time scales in deep learning applications.
Different features present in the data can be learned by training a neural network on different time scales simultaneously. By choosing appropriate partitionings of the network parameters into fast and slow parts I show that our multirate techniques can be used to train deep neural networks for transfer learning applications in vision and natural language processing in half the time, without reducing the generalization performance of the model.
Bounds for subsets of $\mathbb{F}_{p}^{n} \times \mathbb{F}_{p}^{n}$ without L-shaped configurations
Part of the Oxford Discrete Maths and Probability Seminar, held via Zoom. Please see the seminar website for details.
Abstract
I will discuss the difficult problem of proving reasonable bounds in the multidimensional generalization of Szemerédi's theorem and describe a proof of such bounds for sets lacking nontrivial configurations of the form (x,y), (x,y+z), (x,y+2z), (x+z,y) in the finite field model setting.
Recent progress on random graph matching problems
Part of the Oxford Discrete Maths and Probability Seminar, held via Zoom. Please see the seminar website for details.
Abstract
In this talk, I will review some recent progress on random graph matching problems, that is, to recover the vertex correspondence between a pair of correlated random graphs from the observation of two unlabelled graphs. In this talk, I will touch issues of information threshold, efficient algorithms as well as complexity theory. This is based on joint works with Hang Du, Shuyang Gong and Zhangsong Li.
There has been a lot of academic research in to sleep recently (notably by Oxford academic Russell Foster), but for those of us who struggle with the old shut-eye (and age doesn't help, we warn you) then how about Max Richter's 2015 Sleep, an eight and a half hour concept album based around the neuroscience of sleep. Its companion album, From Sleep, from which this track is taken, is only one hour long. The quick nap version, if you like.
13:00
Bounds on quantum evolution complexity via lattice cryptography
Abstract
I will present results from arXiv:2202.13924, where we studied the difference between integrable and chaotic motion in quantum theory as manifested by the complexity of the corresponding evolution operators. The notion of complexity of interest to us will be Nielsen’s complexity applied to the time-dependent evolution operator of the quantum systems. I will review Nielsen’s complexity, discuss the difficulties associated with this definition and introduce a simplified approach which appears to retain non-trivial information about the integrable properties of the dynamical systems.