A coordinate descent algorithm on the Stiefel manifold for deep neural network training
Abstract
We propose to use stochastic Riemannian coordinate descent on the Stiefel manifold for deep neural network training. The algorithm rotates successively two columns of the matrix, an operation that can be efficiently implemented as a multiplication by a Givens matrix. In the case when the coordinate is selected uniformly at random at each iteration, we prove the convergence of the proposed algorithm under standard assumptions on the loss function, stepsize and minibatch noise. Experiments on benchmark deep neural network training problems are presented to demonstrate the effectiveness of the proposed algorithm.
26 Years at Oxford
Abstract
I will reflect on my time as Professor of Numerical Analysis.
13:00
Knot Homologies from Landau Ginsburg Models
Abstract
In her recent work, Mina Aganagic proposed novel perspectives on computing knot homologies associated with any simple Lie algebra. One of her proposals relies on counting intersection points between Lagrangians in Landau-Ginsburg models on symmetric powers of Riemann surfaces. In my talk, I am going to present a concrete algebraic algorithm for finding such intersection points, turning the proposal into an actual calculational tool. I am going to illustrate the construction on the example of the sl_2 invariant for the Hopf link. I am also going to comment on the extension of the story to homological invariants associated to gl(m|n) super Lie algebras, solving this long-standing problem. The talk is based on our work in progress with Mina Aganagic and Elise LePage.