Hosted as part of the Machine Learning and AI Seminar Series in partnership with the DSI Foundations of Data Science Center; the Department of Statistics, Arts and Sciences; and Columbia Engineering


Speaker

Greg Durrett

Greg Durrett, Associate Professor, Computer Science Department and Center for Data Science, NYU Courant


Event Details

Friday, April 10, 2026 (11:00 AM – 12:00 PM ET)

Location: Hamilton Hall, Room 702

REGISTRATION DEADLINE: The Columbia Morningside campus is open to the Columbia community. If you do not have an active CUID, the deadline to register is at 12:00 PM the day before the event.

Register


Talk Information

LLM Reasoning Beyond Scaling

Abstract: Agentic large language models can write and debug complex code, solve competition-level math problems, and conduct in-depth literature review. These reasoning capabilities are enabled by scaling of data: pre-training data to learn vast knowledge, fine-tuning data to learn natural language reasoning, and RL environments to refine that reasoning. In this talk, I will investigate the current LLM reasoning paradigm, its boundaries, and the future of LLM reasoning beyond scaling. First, I will describe the state of reasoning models and where I think scaling will lead to additional successes. I will then shift to discussing issues which are not resolved by pure scaling. First, I will describe our work on calibrating models’ decisions through better understanding of their environments. We find that explicitly telling an LLM its likelihood to succeed or fail at tasks allows it to reason about cost-benefit tradeoffs in its action space. Then, I will describe our new benchmark CREATE, which tests LLMs’ capabilities for associative creativity. I will highlight limitations of LLMs applied to creative tasks like scientific ideation and where I see future work making progress in these areas.