Thesis Defense: Scaling Cooperative Intelligence via Inverse Planning and Probabilistic Programming

Thesis Defense: Scaling Cooperative Intelligence via Inverse Planning and Probabilistic Programming
Presenter: Tan Zhi-Xuan
Please email xuan@mit.edu for Zoom link
How can we build cooperative machines that model and understand human minds — machines that assist us with our goals, coordinate on shared plans, infer the intentions behind our words, and even learn our norms and values? In this talk, I will introduce a scalable Bayesian approach to building such systems via inverse planning and probabilistic programming. By combining online model-based planners and sequential Monte Carlo inference into a single architecture, Sequential Inverse Plan Search (SIPS), we can infer human goals from actions in faster-than-real-time, while scaling to environments with hundreds of possible goals and long planning horizons that have proved intractable for earlier methods. SIPS can additionally make use of large language models (LLMs) as likelihood functions within probabilistic programs, allowing us to build AI assistants and copilots that reliably infer human goals from ambiguous instructions, then provide assistance under uncertainty with much higher success rates than LLMs can on their own. By applying this Bayesian approach in many-agent environments, we are also able to design agents that rapidly learn cooperative social norms from others' behavior, achieving mutually beneficial outcomes with orders of magnitude less data than model-free deep RL. I will conclude by charting out how this research program could deliver a new generation of cooperative AI systems grounded in rational AI engineering, while illuminating the computational foundations of human cooperation and addressing fundamental challenges in building human-aligned AI.
Thesis Committee: Vikash Mansinghka, Joshua Tenenbaum, Dylan Hadfield-Menell, Leslie Kaelbling