About

This seminar series invites experts from across the country to come to Columbia and present the latest cutting-edge research in the field of Machine Learning and Artificial Intelligence. Running the gamut between theory and empirics, the seminar provides a single, unified space to bring together the ML/AI community at Columbia. Topics of interest include, but are not limited to, Language Models, Optimization for Deep Learning, Reinforcement and Imitation Learning,  Learning Theory, Interpretability and AI Alignment, AI for science, Probabilistic ML, and Bayesian methods.

Hosts & Co-Sponsors: DSI Foundations of Data Science Center; Department of Statistics, Arts and Sciences, Columbia Engineering

Registration

Registration for all CUID holders is preferred. If you do not have an active CUID, registration is required and is due at 12:00 PM the day prior to the seminar. Unfortunately, we cannot guarantee entrance to Columbia’s Morningside campus if you register following 12:00 PM the day prior to the seminar. Thank you for understanding!

Please contact Erin Elliott, DSI Events and Marketing Coordinator at ee2548@columbia.edu with any questions.

Next Seminar

Date: Friday, March 10, 2026 (11:00 AM – 12:00 PM)

Location: Hamilton Hall, Room 702

Greg Durrett

Greg Durrett, Associate Professor, Computer Science Department and Center for Data Science, NYU Courant

Title: LLM Reasoning Beyond Scaling

Abstract: Agentic large language models can write and debug complex code, solve competition-level math problems, and conduct in-depth literature review. These reasoning capabilities are enabled by scaling of data: pre-training data to learn vast knowledge, fine-tuning data to learn natural language reasoning, and RL environments to refine that reasoning. In this talk, I will investigate the current LLM reasoning paradigm, its boundaries, and the future of LLM reasoning beyond scaling. First, I will describe the state of reasoning models and where I think scaling will lead to additional successes. I will then shift to discussing issues which are not resolved by pure scaling. First, I will describe our work on calibrating models’ decisions through better understanding of their environments. We find that explicitly telling an LLM its likelihood to succeed or fail at tasks allows it to reason about cost-benefit tradeoffs in its action space. Then, I will describe our new benchmark CREATE, which tests LLMs’ capabilities for associative creativity. I will highlight limitations of LLMs applied to creative tasks like scientific ideation and where I see future work making progress in these areas.

Register

Upcoming Seminar Schedule (Spring 2026)

Please save the below dates and times to attend the seminar series.

Friday, April 17 (11:00 AM – 12:00 PM)

  • Location: Hamilton Hall, Room 702
  • Speaker: Danqi Chen, Associate Professor of Computer Science, Co-Leader of Princeton NLP Group, Associate Director of Princeton Language and Intelligence, Princeton University
  • Register