About

This seminar series invites experts from across the country to come to Columbia and present the latest cutting-edge research in the field of Machine Learning and Artificial Intelligence. Running the gamut between theory and empirics, the seminar provides a single, unified space to bring together the ML/AI community at Columbia. Topics of interest include, but are not limited to, Language Models, Optimization for Deep Learning, Reinforcement and Imitation Learning,  Learning Theory, Interpretability and AI Alignment, AI for science, Probabilistic ML, and Bayesian methods.

Hosts & Co-Sponsors: DSI Foundations of Data Science Center; Department of Statistics, Arts and Sciences, Columbia Engineering

Registration

Registration for all CUID holders is preferred. If you do not have an active CUID, registration is required and is due at 12:00 PM the day prior to the seminar. Unfortunately, we cannot guarantee entrance to Columbia’s Morningside campus if you register following 12:00 PM the day prior to the seminar. Thank you for understanding!

Please contact Erin Elliott, DSI Events and Marketing Coordinator at ee2548@columbia.edu with any questions.

Register

Next Seminar

Date: Friday, October 24, 2025 (11:00 AM – 12:00 PM)

Location: Columbia School of Social Work, Room C05

Furong Huang Headshot

Furong Huang, Associate Professor, Department of Computer Science at the University of Maryland

Title: Rethinking Test-Time Thinking: From Token-Level Rewards to Robust Generative Agents

Abstract: We present a unified perspective on test-time thinking as a lens for improving generative AI agents through finer-grained reward modeling, data-centric reasoning, and robust alignment. Beginning with GenARM, we introduce an inductive bias for denser, token-level reward modeling that guides generation during decoding, enabling token-level alignment without retraining. While GenARM targets reward design, ThinkLite-VL focuses on the data side of reasoning. It proposes a self-improvement framework that selects the most informative samples via MCTS-guided search, yielding stronger visual reasoning with fewer labels. Taking this a step further, MORSE-500 moves beyond selection to creation: it programmatically generates targeted, controllable multimodal data to systematically probe and stress-test models’ reasoning abilities. We then interrogate a central assumption in inference-time alignment: Does Thinking More Always Help? Our findings reveal that increased reasoning steps can degrade performance–not due to better or worse reasoning per se, but due to rising variance in outputs, challenging the naive scaling paradigm. Finally, AegisLLM applies test-time thinking in the service of security, using an agentic, multi-perspective framework to defend against jailbreaks, prompt injections, and unlearning attacks–all at inference time. Together, these works chart a path toward generative agents that are not only more capable, but more data-efficient, introspective, and robust in real-world deployment.


Upcoming Seminar Schedule (Fall 2025)

Please save the below dates, times, and locations to attend the seminar series.

Friday, November 7, 2025 (11:00 AM – 12:00 PM)

  • Location: School of Social Work, Room C05
  • Speaker: Florentin Guth, Faculty Fellow, Center for Data Science, NYU; and Research Fellow, Center for Computational Neuroscience, Flatiron Institute

Friday, November 21, 2025 (11:00 AM – 12:00 PM)

  • Location: School of Social Work, Room C05
  • Speaker: Andrej Risteski, Associate Professor, Machine Learning Department, Carnegie Mellon University

Friday, December 12, 2025 (11:00 AM – 12:00 PM)

  • Location: School of Social Work, Room 311/312
  • Speaker: Jason Weston, Research Scientist at Facebook, NY and a Visiting Research Professor at NYU