Hi, my name is Marvin.
I’m an AI researcher straddling theory and empirics. Right now, I am working on the theory of generative models as well as ways to augment the capabilities and safety of language models. Earlier this year, I graduated summa cum laude from Harvard with a B.A. in Computer Science & Mathematics.
My most recent project was on a unified theoretical framework for feature emergence in generative models (ICML 2024; ICML 2025 oral). We explained where high-level features such as reasoning accuracy and toxicity emerge during the sampling trajectories of diffusion models and large language models. This work, also presented in my thesis, was awarded both the Hoopes Prize for outstanding undergraduate research and the Captain Jonathan Fay Prize, given to the top three theses across all disciplines at Harvard College.
I love meeting new people and chatting. Please reach out at marvin[dot]fangzhou[dot]li[at]gmail.com.
Selected Publications
* denotes equal contribution
Blink of an Eye: A Simple Theory for Feature Localization in Generative Models
Marvin Li, Aayush Karan, Sitan Chen.
ICML, 2025 (Oral, top 1% of submissions)
arXiv / code
A unifying theory showing why and when features suddenly “lock in” during generation in both diffusion and autoregressive models.
Critical Windows: Non-Asymptotic Theory for Feature Emergence in Diffusion Models
Marvin Li, Sitan Chen.
ICML, 2024
arXiv / code
Introduces tight, distribution-agnostic bounds pinpointing when image features appear along the diffusion trajectory.
MoPe: Model Perturbation-Based Privacy Attacks on Language Models
Marvin Li*, Jason Wang*, Jeffrey Wang*, Seth Neel.
EMNLP Main Conference, 2023
arXiv
Shows that second-order gradient information lets an attacker detect training-set membership far more reliably than loss-only baselines.