Aukosh Jagannath


Course: STAT 946 — Adv Topics: Mathematical Foundations of Deep Learning


Blurb: The goal of this course is to introduce and explore some of the recent theoretical advances that aim to understand modern deep learning methods and training regimes.

Topics may include: Universal approximations, Uniform convergence, Benign overparamterization, functional limit theory/scaling limits, NTK, comparison to kernel methods, Training dynamics and related phenomenology (Implicit regularization, Training regimes, sample complexity, Mean-Field/hydrodynamic limits, DMFT/Effective Dynamics), Loss landscapes and geometry, transformers, diffusion models.

There will be a heavy focus on the analysis of training dynamics (sample complexity and scaling limits). The course will be interspersed with primers on important tools and techniques from probability theory such as concentration of measure, random matrix theory, stochastic analysis, and related ideas inspired by statistical physics.

This will be a fast-paced, research-level, seminar-style course. We will be learning as a group. Students will play an active role in the course: as the course progresses students will pick up and explore these or other topics to catch us all up!

The following is tentative and comments/suggested reading are welcome!

Lectures

Week 1: Approximation theory
Week 2: Uniform convergence
Week 3: Implicit regularization and benign overparamterization
Week 4: RMT and double descent
Week 5: NTK and Lazy training
Week 6: Infinite Width limits
Week 7: Reading Week
Week 8: Sample complexity and scaling limits
Week 9: Spectral alignment and Transformers
Week 10: CANCELLED
Week 11: Diffusion Models + Student presentations
Week 12: Student Presentations
Week 13: Student Presentations
Week 14: Student Presentations