| Module I: Foundations of Interactive Robotics |
| 1 |
Aug 31 |
1 |
Human-centered robotics 101 |
Lecture |
Recommended Movie: Elon Musk's Crash Course |
| 1 |
Sep 2 |
2 |
Foundations: Dynamical systems, probability, Bayesian inference, optimal control |
Lecture |
Required Reading: All models are wrong. George Box (1976) |
| 2 |
Sep 7 |
|
Labor Day 🏖️ |
|
|
| 2 |
Sep 9 |
3 |
Robotic motion planning (search, optimization, MDP/RL) |
Lecture |
Optional Reading 1: Planning Algorithms. LaValle (2006)
Optional Reading 2: A Tour of Reinforcement Learning: The View from Continuous Control. Recht (2019)
|
| 3 |
Sep 14 |
4 |
Robot safety (Operational design domain, Safety Filters, Hamilton-Jacobi reachability, control barrier functions, model predictive control and shielding) |
Lecture |
Required Reading 1: Operational Design Domain for Automated Driving Systems. Czarnecki (2018)
Required Reading 2: The Safety Filter: A Unified View of Safety-Critical Control in Autonomous Systems. Hsu et al. (2024)
|
| 3 |
Sep 16 |
5 |
Computing safe robot control policies (analytical, dynamic programming, learning-based) |
Lecture
Debate
HW release
Code demo
|
Debate Proposition: Pre-specifying an ODD for robots is a suitable way to characterize and enforce its safe operation around people.
Code Demo: Safety Filters
Optional Reading 1: DeepReach: A Deep Learning Approach to High-Dimensional Reachability. Bansal and Tomlin (2020)
Optional Reading 2: Learning Control Barrier Functions from Expert Demonstrations. Robey et al. (2020)
Homework 1: Trajectory optimization, MDP/RL, robot safety
|
| 4 |
Sep 21 |
6 |
Emerging robot safety methods |
Paper discussion |
Paper 1: Generalizing Safety Beyond Collision-Avoidance via Latent-Space Reachability Analysis. Nakamura et al. (2025)
Paper 2: Safe Exploration for Optimization with Gaussian Processes. Sui et al. (2015)
Paper 3: Real-Time Anomaly Detection and Reactive Planning with Large Language Models. Sinha et al. (2024)
|
| Module II: The Game Theory of Human–Robot Interaction |
| 4 |
Sep 23 |
7 |
Introduction to dynamic game theory |
Lecture
Code demo
|
Code Demo: ILQGames, KLGames
Required Reading: Efficient Iterative Linear-Quadratic Approximations for Nonlinear Multi-Player General-Sum Differential Games. Fridovich-Keil et al. (2020)
Optional Reading 1: Dynamic Noncooperative Game Theory. Başar and Olsder (1998)
Optional Reading 2: Smooth Game Theory. Fridovich-Keil et al. (2024)
Optional Reading 3: Blending Data-Driven Priors in Dynamic Games. Lidard et al. (2024)
|
| 5 |
Sep 28 |
8 |
Modeling HRI as a stochastic game |
Lecture |
Required Reading: Planning for Autonomous Cars that Leverage Effects on Human Actions. Sadigh et al. (2018)
Optional Reading 1: Individual Choice Behavior. Luce (1959)
Optional Reading 2: Understanding the Intentions of Others: Re-Enactment of Intended Acts by 18-Month-Old Children. Meltzoff (1995)
|
| 5 |
Sep 30 |
9 |
Predicting human intent and motion |
Lecture
HW due
|
Required Reading 1: Probabilistically Safe Robot Planning with Confidence-Based Human Predictions. Fisac et al. (2018)
Required Reading 2: LESS is More: Rethinking Probabilistic Models of Human Behavior. Bobu et al. (2020)
Optional Reading: Motion Transformer with Global Intention Localization and Local Movement Refinement. Shi et al. (2022)
|
| 6 |
Oct 5 |
10 |
Language-enabled HRI |
Paper discussion
Project proposal due
|
Paper 1: LILA: Language-Informed Latent Actions. Karamcheti et al. (2021)
Paper 2: Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners. Ren et al. (2023)
Paper 3: Risk-Calibrated Human-Robot Interaction via Set-Valued Intent Prediction. Lidard et al. (2024)
|
| 6 |
Oct 7 |
11 |
Safety filtering around humans |
Lecture
Debate
HW release
|
Debate Proposition: A robot that fails less often than humans performing the same type of task should be considered safe enough for deployment.
Optional Reading 1: On Infusing Reachability-Based Safety Assurance within Planning Frameworks for Human-Robot Vehicle Interactions. Leung et al. (2020)
Optional Reading 2: Active Uncertainty Reduction for Safe and Efficient Interaction Planning: A Shielding-Aware Dual Control Approach. Hu et al. (2023)
Homework 2: Dynamic games, human intent modeling, POMDP
|
| 7 |
Oct 12 |
12 |
Cooperative games and value alignment in HRI |
Lecture |
Required Reading 1: Cooperative Inverse Reinforcement Learning. Hadfield-Menell et al. (2016)
Required Reading 2: Pragmatic-Pedagogic Value Alignment. Fisac et al. (2017)
|
| 7 |
Oct 14 |
13 |
Planning with human internal states |
Paper discussion |
Paper 1: Contingency Games for Multi-Agent Interaction. Peters et al. (2024)
Paper 2: Improving Automated Driving through POMDP Planning with Human Internal States. Sunberg and Kochenderfer (2022)
Paper 3: Deception Game: Closing the Safety-Learning Loop in Interactive Robot Autonomy. Hu et al. (2023)
|
| Module III: Robots that Learn from Humans |
| 8 |
Oct 19 |
14 |
Robot learning basics: model-free/model-based RL, imitation learning, diffusion policy |
Lecture
HW due
|
Required Reading 1: Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor. Haarnoja et al. (2018)
Required Reading 2: Mastering Diverse Domains through World Models. Hafner et al. (2023)
Required Reading 3: Diffusion Policy: Visuomotor Policy Learning via Action Diffusion. Chi et al. (2024)
|
| 8 |
Oct 21 |
15 |
Midterm "mini-conference": 5-min lightning talks + 2-min Q&A |
Lightning talks |
|
| 9 |
Oct 26 |
16 |
Safe robot learning: multi-agent RL, safety RL, RL under safety filtering, verification of neural controllers |
Lecture |
Required Reading 1: Robust Adversarial Reinforcement Learning. Pinto et al. (2017)
Required Reading 2: Bridging Hamilton-Jacobi Safety Analysis and Reinforcement Learning. Fisac et al. (2019)
Required Reading 3: Safety Verification and Robustness Analysis of Neural Networks via Quadratic Constraints and Semidefinite Programming. Fazlyab et al. (2020)
Optional Reading: MAGICS: Adversarial RL with Minimax Actors Guided by Implicit Critic Stackelberg. Wang et al. (2024)
Optional Reading: Provably Optimal Reinforcement Learning under Safety Filtering. Oh et al. (2025)
Optional Reading: Verification of Neural Reachable Tubes via Scenario Optimization and Conformal Prediction. Lin and Bansal (2024)
|
| 9 |
Oct 28 |
17 |
Inverse RL and games |
Lecture
Debate
HW release
|
Debate Proposition: RL/IL is more useful than optimal control for human-centered robotics.
Required Reading: Maximum Entropy Inverse Reinforcement Learning. Ziebart et al. (2008)
Homework 3: Safe RL, inverse games
|
| 10 |
Nov 2 |
18 |
Learning rewards |
Paper discussion |
Paper 1: Asking Easy Questions: A User-Friendly Approach to Active Reward Learning. Biyik et al. (2019)
Paper 2: Reward-Rational (Implicit) Choice: A Unifying Formalism for Reward Learning. Jeon et al. (2020)
Paper 3: Inferring Objectives in Continuous Dynamic Games from Noise-Corrupted Partial State Observations. Peters et al. (2021)
|
| 10 |
Nov 4 |
19 |
RLHF, Human–AI safety (Guest lecture: Kaiqu Liang) |
Lecture
Debate
|
Debate Proposition: More data can enhance alignment and safety in human-centered robotics.
Required Reading: Deep Reinforcement Learning from Human Preferences. Christiano et al. (2017)
Required Reading: Human-AI Safety: A Descendant of Generative AI and Control Systems Safety. Bajcsy and Fisac (2024)
Optional Reading: From Refusal to Recovery: A Control-Theoretic Approach to Generative AI Guardrails. Pandya et al. (2025)
Optional Reading: Open Problems and Fundamental Limitations of Reinforcement Learning from Human Feedback. Casper et al. (2023)
|
| 11 |
Nov 9 |
20 |
Embodied AI safety |
Paper discussion |
Paper 1: Safety Guardrails for LLM-Enabled Robots. Ravichandran et al. (2025)
Paper 2: When Your AIs Deceive You: Challenges of Partial Observability in Reinforcement Learning from Human Feedback. Lang et al. (2024)
Paper 3: RLHS: Mitigating Misalignment in RLHF with Hindsight Simulation. Liang et al. (2025)
|
| Module IV: Robots that Help Humans Learn |
| 11 |
Nov 11 |
21 |
Legibility and predictability |
Lecture
HW due
|
Required Reading: Legibility and Predictability of Robot Motion. Dragan et al. (2013)
|
| 12 |
Nov 16 |
22 |
AI coaching |
Lecture
Debate
|
Debate Proposition: Explicit representations of human skill level are essential for a robot to effectively teach humans.
|
| 12 |
Nov 18 |
23 |
Robots that assist and teach |
Paper discussion |
Paper 1: Dreaming to Assist: Learning to Align with Human Objectives for Shared Control in High-Speed Racing. DeCastro et al. (2024)
Paper 2: Computational Teaching for Driving via Multi-Task Imitation Learning. Gopinath et al. (2024)
Paper 3: Safety with Agency: Human-Centered Safety Filter with Application to AI-Assisted Motorsports. Oh et al. (2025)
|
| 13 |
Nov 23 |
|
Fall Recess 🏖️ |
|
|
| 13 |
Nov 25 |
|
Fall Recess 🏖️ |
|
|
| 14 |
Nov 30 |
24 |
Guest lecture on AI coaching / Human–AI co-evolution |
Lecture |
|
| 14 |
Dec 2 |
25 |
Course wrap-up |
Lecture
Debate
|
Debate Proposition: AI systems should intentionally allow users to fail in order to promote long-term learning and independence when coaching humans.
Required Reading: Learning from Errors. Metcalfe (2017)
Required Reading: The Mundanity of Excellence. Chambliss (1989)
|
| 15 |
Dec 7 |
26 |
Project Final Presentations 1 |
Project presentation |
|
| 15 |
Dec 9 |
27 |
Project Final Presentations 2 |
Project presentation |
|