news

Jul 17, 2025 I’m attending ICML 2025, where I’ll be presenting some of my recent work on interpretability in large language models:
Jul 10, 2025 I’m participating in MARS 3.0 (Mentorship for Alignment Research Students), a research program run by the Cambridge AI Safety Hub. As part of this program, I will be working alongside a group of talented researchers on problems related to Chain-of-Thought reasoning in AI systems.
May 27, 2025 Workshop Announcement: Interpretability in LLMs using Geometric and Statistical Methods
I am organizing a workshop taking place on May 27-28 where we will explore recent developments in interpretability for large language models (LLMs) using geometric and statistical methods. For further details, check out the workshop page.