Lei Wu (吴磊)

Lei Wu 


Assistant Professor
School of Mathematical Sciences
Center for Machine Learning Research
Peking University

Office: 静园6院 205
Email: leiwu (at) math (dot) pku (dot) edu (dot) cn

Google Scholar     Github      CV

About Me

I am currently an Assistant Professor in the School of Mathematical Sciences and Center for Machine Learning Research at Peking University.

Previously, I was a postdoc in PACM at Princeton University and in the Wharton Statistics and Data Science Department at the University of Pennsylvania. I completed my Ph.D. in computational mathematics at Peking University in 2018, advised by Prof. Weinan E. I received my B.S. degree in mathematics from Nankai University in 2012.

My research aims to understand the mechanisms behind the success of deep learning, with a particular focus on:

  • The approximation and representation power of neural networks

  • The dynamical behavior of popular optimization algorithms such as SGD and Adam

  • Emergent phenomena in the training of large language models (LLMs)

Recruiting

We are actively seeking self-motivated postdocs, PhD students, and undergraduate interns to join our team. If you are interested in collaborating with me, please email me your CV, transcript, and a brief description of your research interests.

Research Highlights

(See also: Full publication list)

Recent News

  • 2026-02: Constant-depth network with smooth activations released on arXiv.

    • Establishes that smooth activations (e.g., GELU, SiLU) enable smoothness adaptivity in constant-depth neural networks, achieving optimal approximation and statistical rates.

  • 2026-02: Fast catch-up, late switching accepted to ICLR 2026.

    • reveals a fast catch-up effect in batch-size scheduling, which holds across linear regression and LLM pre-training.

  • 2025-09: Functional Scaling Laws accepted to NeurIPS 2025 (Spotlight).

    • introduces the FSL framework, characterizing the entire loss trajectory rather than only the final-step scaling behavior, from linear regression to LLM pre-training.