CS886 - Multiagent Systems - Winter 2025

Schedule & Reading List

Jan 6 Course Logistics / Introduction I
Course Logistics
Introduction to Multiagent Systems
How to read a research paper
Cooperative AI: Machines Must Learn to Find Common Ground
Open Problems in Cooperative AI
Paper bids due Jan 12
Jan 13 Introduction II / Presentation Advice / Background
Introduction to Game Theory
Multiagent Reinforcement Learning
Resources for Preparing Presentations: 1, 2, 3, 4, 5
Jan 20 Cooperation
Social Norm Complexity and Past Reputations in the Evolution of Cooperation
Santos et al. [Nature 2018]
Cooperate or Collapse: Emergence of Sustainable Cooperation in a Society of LLM Agents
Piatti et al. [NeurIPS 2024 version]
Jan 27 Learning and Social Dilemmas
Learning with Opponent Learning Awareness
Foerster et al. [AAMAS 2018 version]
Inequity aversion improves cooperation in intertemporal social dilemmas
Hughes et al. [NeurIPS 2018]
Feb 3 Complex Groups
Theory of Minds: Understanding Behavior in Groups through Inverse Planning
Shum et al. [AAAI 2019]
The AI Economist: Taxation policy design via two-level deep multiagent reinforcement learning
Zheng et al. [Science 2021]
Feb 10 Strategic Reasoning
Human-level play in the game of Diplomacy by combining language models with strategic reasoning
Bakhtin et al. [Science 2022]
STEER: Assessing the Economic Rationality of Large Language Models
Raman et al. [ICML 2024]
Project proposals due Feb 14
Feb 17 No class due to Reading Week
Feb 24 Mechanism Design
The Good Shepherd: An Oracle Agent for Mechanism Design
Balaguer et al [arXiv 2022]
Mechanism Design for Large Language Models
Duetting et al. [WWW 2024]
March 3 Social Choice and Voting
Generative Social Choice
Fish et al. [ACM EC 2024]
AI can Help Humans Find Common Ground in Democratic Deliberation
Tessler et al. [Science 2024]
March 10 Alignment
Constitutional AI: Harmlessness from AI Feedback
Bai et al. [2022]
Fine-Tuning Language Models from Human Preferences
Ziegler et al [2020]
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Rafailov et al [2023]]
March 17 RLHF
Distributional Preference Learning: Understanding and Accounting for Hidden Context in RLHF
Siththaranjan et al. [ICLR 2024]
Nash Learning from Human Feedback
Munos et al. [ICML 2024]
Looking Forward 24 Miscellaneous Topics
A Roadmap to Pluralistic Alignment
Sorensen et al. [ICML 2024]
Social Choice Should Guide AI Alignment in Dealing with Diverse Human Feedback
Conitzer et al. [ICML 2024]
March 31 Project Presentations
Final project report due April 11