Return to Colloquia & Seminar listing
Optimization Algorithms in the Large: Exact Dynamics, Average-case Analysis, and Stepsize Criticality
Mathematics of Data & DecisionsSpeaker: | Courtney Paquette, McGill U. |
Related Webpage: | https://cypaquette.github.io/ |
Location: | Zoom |
Start time: | Thu, Jan 13 2022, 2:10PM |
In this talk, I will present a framework, inspired by random matrix theory, for analyzing the dynamics of optimization algorithms (e.g., 1st-order methods, stochastic gradient descent (SGD), and momentum) when both the number of samples and dimensions are large. Using this new framework, we show that the dynamics of optimization algorithms on a least squares problem with random data become deterministic in the large sample and dimensional limit. In particular, the limiting dynamics for stochastic algorithms are governed by a Volterra integral equation. This model predicts that SGD undergoes a phase transition at an explicitly given critical stepsize that ultimately affects its convergence rate, which we also verify experimentally. Finally, when input data is isotropic, we provide explicit expressions for the dynamics and average-case convergence rates. These rates show significant improvement over the worst-case complexities.
https://ucdavis.zoom.us/j/99562940622
Passcode: 110853