Alex Damian – Foundations of Deep Learning: Optimization and Representation Learning

Thursday, March 20
11:00 am - 12:00 pm

34-401 Grier A

Abstract: 

Deep learning’s success stems from the ability of neural networks to  
automatically discover meaningful representations from raw data. In this  
talk, I will describe some recent insights into how optimization enables this  
learning process. First, I will explore how gradient descent enables neural  
networks to adapt to low-dimensional structure in the data, and how these  
ideas extend to  understanding in-context learning in transformers. I will  
then discuss my work toward a predictive theory of deep learning optimization  
that characterizes how different optimizers navigate deep learning loss  
landscapes and how these different behaviors affect training efficiency,  
stability, and generalization.
 
Bio: 

Alex Damian is a fifth-year Ph.D. student in the Program for Applied and  
Computational Mathematics (PACM) at Princeton University, advised by Jason  
Lee. His research is focused on deep learning theory with an emphasis on  
optimization and representation learning. His work has been supported by an  
NSF Graduate Research Fellowship and a Jane Street Graduate Research  
Fellowship.

Details

  • Date: Thursday, March 20
  • Time: 11:00 am - 12:00 pm
  • Category:
  • Location: 34-401 Grier A

Host

  • Piotr Indyk