Fast Bayesian coresets via subsampling and quasi-Newton refinement
Naik, C Rousseau, J Campbell, T Advances in Neural Information Processing Systems 35 (NeurIPS 2022) volume 1 70-83 (01 Apr 2023)
Fast Bayesian Coresets via Subsampling and Quasi-Newton Refinement
Naik, C Rousseau, J Campbell, T Advances in Neural Information Processing Systems volume 35 (01 Jan 2022)
A new mixed finite-element method for H2 elliptic problems
Farrell, P Hamdan, A MacLachlan, S Computers and Mathematics with Applications volume 128 300-319 (15 Dec 2022)
Analysis and Modeling of Client Order Flow in Limit Order Markets
CONT, R CUCURINGU, M Glukhov, V Prenzel, F Quantitative Finance
Theoretical study of the emergence of periodic solutions for the inhibitory NNLIF neuron model with synaptic delay
Ikeda, K Roux, P Salort, D Smets, D Mathematical Neuroscience and Applications volume 2 (26 Oct 2022)
Tue, 01 Nov 2022

12:30 - 13:00
C3

Asymptotic Analysis of Deep Residual Networks

Alain Rossier
Abstract

Residual networks (ResNets) have displayed impressive results in pattern recognition and, recently, have garnered considerable theoretical interest due to a perceived link with neural ordinary differential equations (neural ODEs). This link relies on the convergence of network weights to a smooth function as the number of layers increases. We investigate the properties of weights trained by stochastic gradient descent and their scaling with network depth through detailed numerical experiments. We observe the existence of scaling regimes markedly different from those assumed in neural ODE literature. Depending on certain features of the network architecture, such as the smoothness of the activation function, one may obtain an alternative ODE limit, a stochastic differential equation (SDE) or neither of these. Furthermore, we are able to formally prove the linear convergence of gradient descent to a global optimum for the training of deep residual networks with constant layer width and smooth activation function. We further prove that if the trained weights, as a function of the layer index, admit a scaling limit as the depth increases, then the limit has finite 2-variation.

An adaptive random bit multilevel algorithm for SDEs
Giles, M Hefter, M Mayer, L Ritter, K Multivariate Algorithms and Information-Based Complexity 15-31 (08 Jun 2020)
Variant-specific symptoms of COVID-19 in a study of 1,542,510 adults in England
Whitaker, M Elliott, J Bodinier, B Barclay, W Ward, H Cooke, G Donnelly, C Chadeau-Hyam, M Elliott, P Nature Communications volume 13 (11 Nov 2022)
Dynamic calibration of order flow models with generative adversarial networks
Prenzel, F Cont, R Cucuringu, M Kochems, J ICAIF '22: 3rd ACM International Conference on AI in Finance (26 Oct 2022)
Multicore quantum computing
Jnane, H Undseth, B Cai, Z Benjamin, S Koczor, B Physical Review Applied volume 18 issue 4 (26 Oct 2022)
Subscribe to