EECS Special Seminar: Alex Damian, "Foundations of Deep Learning: Optimization and Representation Learning"

Speaker

Alex Damian
Princeton University

Host

Piotr Indyk
CSAIL

Abstract: 
Deep learning's success stems from the ability of neural networks to automatically discover meaningful representations from raw data. In this talk, I will describe some recent insights into how optimization enables this learning process. First, I will explore how gradient descent enables neural networks to adapt to low-dimensional structure in the data, and how these ideas extend to  understanding in-context learning in transformers. I will then discuss my work toward a predictive theory of deep learning optimization that characterizes how different optimizers navigate deep learning loss landscapes and how these different behaviors affect training efficiency, stability, and generalization.
 

Bio: 
Alex Damian is a fifth-year Ph.D. student in the Program for Applied and Computational Mathematics (PACM) at Princeton University, advised by Jason Lee. His research is focused on deep learning theory with an emphasis on optimization and representation learning. His work has been supported by an NSF Graduate Research Fellowship and a Jane Street Graduate Research Fellowship.