Research
I work broadly on reasoning in LLMs.
My previous research explores how to integrate LLMs into statistical/data science workflows in principled, statistically rigorous ways.
Before that, I worked on probabilistic methods (e.g., Sequential Monte Carlo).
My email is firstname.middle_initial.lastname@stanford.edu!
|
Automated Hypothesis Validation with Agentic Sequential Falsifications
Kexin Huang*, Ying Jin*, Ryan Li*, Michael Y. Li, Emmanuel Candès, Jure Leskovec
preprint, 2025
paper
LLMs + sequential hypothesis tests with Type-I error control
|
BoxingGym: Benchmarking Progress in Automated Experimental Design and Model Discovery
Kanishk Gandhi*, Michael Y. Li*, Lyle Goodyear,Louise Li, Aditi Bhaskar, Mohammed Zaman,Noah D. Goodman
preprint, 2025
paper
A benchmark for LLM driven experimental design and model discovery
|
Learning to Learn Functions
Michael Y. Li, Fred Callaway, William D. Thompson, Ryan P. Adams, Thomas L. Griffiths
Cognitive Science, 2023
paper
We propose hierarchical Bayesian models of how people learn to learn functions and validate our model in behavioral experiments.
|
|