Mehul Damani

Hello! I am a third year Ph.D. student at MIT advised by Jacob Andreas.
My research interests lie at the intersection of reinforcement learning (RL) and large language models (LLMs).
I am very excited by the potential of RL to improve reasoning, math, coding, and other capabilities in LLMs.Currently, I am thinking about how RL can be used to improve calibration and reduce hallucinations in LLMs. Finally, I have also been thinking about the paradigm of inference-time compute, and how optimally selecting inference-time techniques can significantly improve the efficiency of LLMs.
Previously, I worked with Lerrel Pinto at NYU on developing automatic curriculum learning methods for RL agents. Before that, I was a part of the MARMot Lab at NUS, where I worked with Guillaume Sartoretti on applying multi-agent reinforcement learning to traffic signal control and multi-agent pathfinding.
I’m always excited to explore new research directions and am open to collaborating or advising students. If you are interested in my research or simply want to chat, don’t hesitate to get in touch!
Selected Publications
- ICMLThe Surprising Effectiveness of Test-Time Training for Abstract ReasoningInternational Conference on Machine Learning 2025
- ICLRLearning How Hard to Think: Input-Adaptive Allocation of LM ComputationInternational Conference on Learning Representations 2025
- TMLROpen Problems and Fundamental Limitations of Reinforcement Learning from Human FeedbackTransactions on Machine Learning Research 2023