Return to Colloquia & Seminar listing
Math/Stats Colloquium: The puzzle of dimensionality and feature learning in neural networks and kernel machines
Special EventsSpeaker: | Mikhail Belkin, UCSD |
Related Webpage: | http://misha.belkin-wang.org/bio.html |
Location: | Zoom |
Start time: | Thu, May 2 2024, 4:10PM |
Remarkable progress in AI has far surpassed expectations of just a few years ago. At their core, modern models, such as transformers, implement traditional statistical models--high order Markov chains. Nevertheless, it is not generally possible to estimate Markov models of that order given any possible amount of data. Therefore these methods must implicitly exploit low-dimensional structures present in data. Furthermore, these structures must be reflected in high-dimensional internal parameter spaces of the models. Thus, to build fundamental understanding of modern AI, it is necessary to identify and analyze these latent low-dimensional structures. In this talk, I will discuss how deep neural networks of various architectures learn low-dimensional features and how the lessons of deep learning can be incorporated in non-backpropagation-based algorithms that we call Recursive Feature Machines. I will provide a number of experimental results on different types of data, as well as some connections to classical sparse learning methods, such as Iteratively Reweighted Least Squares.