Vashisth Tiwari
PhD @ Carnegie Mellon
Hi! 👋
I’m a first-year PhD student at Carnegie Mellon University’s LTI, advised by Emma Strubell and Zico Kolter. Previously, I completed my master’s at CMU and a bachelor’s in Physics and Mathematics at the University of Rochester.
My research focuses on developing better, more efficient LLMs by improving training dynamics and data quality. I am particularly interested in knowledge distillation, synthetic data generation, and reinforcement learning for post-training.
I’ve been fortunate to learn from wonderful mentors. In summer 2026, I will be joining the Nemotron post-training team at Nvidia, where I’ll work with [Oleksii] (http://kuchaev.com) and Adi. In summer 2025, I was a Student Researcher with Giulia DeSalvo at DeepMind. At CMU, I’ve also worked with Beidi Chen, Lucio Dery, and Clara Na.
news
| Apr 01, 2026 | Research Intern @Nvidia (📍Santa Clara) w/ Oleksii and Adi on Nemotron Post-training |
|---|---|
| May 19, 2025 | Student Researcher @Google (📍Seattle) w/ Giulia DeSalvo on synthetic data for post-training |
| May 15, 2025 | ACL Main (Energy considerations on LLM inference) |
| Apr 14, 2025 | Will be joining CMU-LTI PhD Fall’25 cohort |
| Jan 23, 2025 | ICLR'25 (MagicDec) |
| Aug 19, 2024 | Released MagicDec: Breaking Throughput-Latency Trade-off for Long Context Generation with Speculative Decoding. Blog |