Printable PDF
Department of Mathematics,
University of California San Diego

****************************

Department of Mathematics Colloquim

Govind Menon

Brown University

Towards a geometric theory of deep learning

Abstract:

The mathematical core of deep learning is function approximation by neural networks trained on data using stochastic gradient descent. I will explain an emerging geometric framework for the analysis of this process. This includes a collection of rigorous results on training dynamics for the deep linear network (DLN) as well as general principles for arbitrary neural networks. The mathematics ranges over a surprisingly broad range, including geometric invariant theory, random matrix theory, and minimal surfaces. However, little background in these areas will be assumed and the talk will be accessible to a broad audience. The talk is based on joint work with several co-authors: Yotam Alexander, Nadav Cohen (Tel Aviv), Kathryn Lindsey (Boston College), Alan Chen, Zsolt Veraszto and Tianmin Yu (Brown).

Host: Federico Pasqualotto

February 4, 2026

4:00 PM

APM 6402

****************************