Johns Hopkins University · Department of Computer Science · Data Science and AI Institute · Laboratory for Computational Sensing and Robotics · Institute for Assured Autonomy

EN.601.497/697 Human-Centered Robotics: Models and Algorithms

Fall 2026 · 3 credits

Instructor: Prof. Haimin Hu

Level: Graduate & Undergraduate

Prerequisites: EN.601.220, 601.226, linear algebra, calculus, probability.

Course Description

In this course, we will study fundamental concepts in human-centered robotics, with an emphasis on mathematical models of human–robot interaction and decision-making algorithms for safely deploying robots in human-populated environments. We will ground these ideas in applications such as autonomous vehicles, aerial robots, and home robots.

We will start by introducing the main technical tools for robot planning and control in interactive, safety-critical settings through a game-theoretic lens. After that, we will turn our attention to two tightly coupled challenges: enabling robots to safely learn from humans and help humans learn. The course will combine seminar-style discussions of research papers and whiteboard-style lecture to introduce the key theoretical concepts, and the class project will give you an opportunity to explore the approaches covered in class and possibly combine them with your own research.

After this class, you will be familiar with the state of the art and open challenges in safe and performant human–robot interaction, and you will understand the guarantees and tradeoffs offered by different algorithmic frameworks for human-centered robotics.

Grading

Component Percentage
Homework assignments (3 × 10%, mix of theoretical and programming tasks) 30%
Paper discussion 5%
In-class debate 5%
Lightning talk 5%
Midterm project report 5%
Final project presentation 20%
Final project report 30%

Reference Textbooks

(There is no required textbook)

  • Dimitri Bertsekas, Reinforcement Learning and Optimal Control.
  • Tamer Basar, Geert Jan Olsder, Dynamic Noncooperative Game Theory, 2nd Edition.

Learning Objectives

  • You will learn to formulate human–robot interaction problems using tools from game theory, machine learning, and dynamical systems.

  • You will explore how different mathematical models can help inform robot decision-making in safety-critical, interactive settings.

  • You will learn principles and algorithms for ensuring safety when deploying robots in human-populated environments, including safety filters, dynamic games, planning under uncertainty, and learning-based approaches, and understand different guarantees and tradeoffs associated with various decision-making frameworks.

  • You will study how robots can support and accelerate human learning through strategic interactions, as well as the algorithmic principles behind it, including shared autonomy, assistance games, and active teaching methods that go beyond reactive assistance.

  • Through debate sessions, inspired by the ICRA Robotics Debates, you will develop critical thinking about emerging challenges in modern robotics, and learn to articulate, defend, and critique competing perspectives.

  • Through a 5-minute lightning talk, modeled after common practices at major robotics conferences, you will learn to communicate technical insights concisely and prepare for future conference engagement.

  • Through a semester-long project, you will learn the full pipeline of conducting research in human-centered robotics, including dissecting research papers, identifying limitations in existing approaches, formulating problems with suitable models, developing algorithmic solutions, and deploying and verifying them on hardware platforms. The Alliance AI Lab is committed to providing ongoing mentorship and equipment support for students wishing to advance their projects toward publication.

Tentative Syllabus

Week Date Lecture Topic Type Notes
Module I: Foundations of Interactive Robotics
1 Aug 31 1 Human-centered robotics 101 Lecture Recommended Movie: Elon Musk's Crash Course
1 Sep 2 2 Foundations: Dynamical systems, probability, Bayesian inference, optimal control Lecture Required Reading: All models are wrong. George Box (1976)
2 Sep 7 Labor Day 🏖️
2 Sep 9 3 Robotic motion planning (search, optimization, MDP/RL) Lecture Optional Reading 1: Planning Algorithms. LaValle (2006)
Optional Reading 2: A Tour of Reinforcement Learning: The View from Continuous Control. Recht (2019)
3 Sep 14 4 Robot safety (Operational design domain, Safety Filters, Hamilton-Jacobi reachability, control barrier functions, model predictive control and shielding) Lecture Required Reading 1: Operational Design Domain for Automated Driving Systems. Czarnecki (2018)
Required Reading 2: The Safety Filter: A Unified View of Safety-Critical Control in Autonomous Systems. Hsu et al. (2024)
3 Sep 16 5 Computing safe robot control policies (analytical, dynamic programming, learning-based) Lecture
Debate
HW release
Code demo
Debate Proposition: Pre-specifying an ODD for robots is a suitable way to characterize and enforce its safe operation around people.
Code Demo: Safety Filters
Optional Reading 1: DeepReach: A Deep Learning Approach to High-Dimensional Reachability. Bansal and Tomlin (2020)
Optional Reading 2: Learning Control Barrier Functions from Expert Demonstrations. Robey et al. (2020)
Homework 1: Trajectory optimization, MDP/RL, robot safety
4 Sep 21 6 Emerging robot safety methods Paper discussion Paper 1: Generalizing Safety Beyond Collision-Avoidance via Latent-Space Reachability Analysis. Nakamura et al. (2025)
Paper 2: Safe Exploration for Optimization with Gaussian Processes. Sui et al. (2015)
Paper 3: Real-Time Anomaly Detection and Reactive Planning with Large Language Models. Sinha et al. (2024)
Module II: The Game Theory of Human–Robot Interaction
4 Sep 23 7 Introduction to dynamic game theory Lecture
Code demo
Code Demo: ILQGames, KLGames
Required Reading: Efficient Iterative Linear-Quadratic Approximations for Nonlinear Multi-Player General-Sum Differential Games. Fridovich-Keil et al. (2020)
Optional Reading 1: Dynamic Noncooperative Game Theory. Başar and Olsder (1998)
Optional Reading 2: Smooth Game Theory. Fridovich-Keil et al. (2024)
Optional Reading 3: Blending Data-Driven Priors in Dynamic Games. Lidard et al. (2024)
5 Sep 28 8 Modeling HRI as a stochastic game Lecture Required Reading: Planning for Autonomous Cars that Leverage Effects on Human Actions. Sadigh et al. (2018)
Optional Reading 1: Individual Choice Behavior. Luce (1959)
Optional Reading 2: Understanding the Intentions of Others: Re-Enactment of Intended Acts by 18-Month-Old Children. Meltzoff (1995)
5 Sep 30 9 Predicting human intent and motion Lecture
HW due
Required Reading 1: Probabilistically Safe Robot Planning with Confidence-Based Human Predictions. Fisac et al. (2018)
Required Reading 2: LESS is More: Rethinking Probabilistic Models of Human Behavior. Bobu et al. (2020)
Optional Reading: Motion Transformer with Global Intention Localization and Local Movement Refinement. Shi et al. (2022)
6 Oct 5 10 Language-enabled HRI Paper discussion
Project proposal due
Paper 1: LILA: Language-Informed Latent Actions. Karamcheti et al. (2021)
Paper 2: Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners. Ren et al. (2023)
Paper 3: Risk-Calibrated Human-Robot Interaction via Set-Valued Intent Prediction. Lidard et al. (2024)
6 Oct 7 11 Safety filtering around humans Lecture
Debate
HW release
Debate Proposition: A robot that fails less often than humans performing the same type of task should be considered safe enough for deployment.
Optional Reading 1: On Infusing Reachability-Based Safety Assurance within Planning Frameworks for Human-Robot Vehicle Interactions. Leung et al. (2020)
Optional Reading 2: Active Uncertainty Reduction for Safe and Efficient Interaction Planning: A Shielding-Aware Dual Control Approach. Hu et al. (2023)
Homework 2: Dynamic games, human intent modeling, POMDP
7 Oct 12 12 Cooperative games and value alignment in HRI Lecture Required Reading 1: Cooperative Inverse Reinforcement Learning. Hadfield-Menell et al. (2016)
Required Reading 2: Pragmatic-Pedagogic Value Alignment. Fisac et al. (2017)
7 Oct 14 13 Planning with human internal states Paper discussion Paper 1: Contingency Games for Multi-Agent Interaction. Peters et al. (2024)
Paper 2: Improving Automated Driving through POMDP Planning with Human Internal States. Sunberg and Kochenderfer (2022)
Paper 3: Deception Game: Closing the Safety-Learning Loop in Interactive Robot Autonomy. Hu et al. (2023)
Module III: Robots that Learn from Humans
8 Oct 19 14 Robot learning basics: model-free/model-based RL, imitation learning, diffusion policy Lecture
HW due
Required Reading 1: Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor. Haarnoja et al. (2018)
Required Reading 2: Mastering Diverse Domains through World Models. Hafner et al. (2023)
Required Reading 3: Diffusion Policy: Visuomotor Policy Learning via Action Diffusion. Chi et al. (2024)
8 Oct 21 15 Midterm "mini-conference": 5-min lightning talks + 2-min Q&A Lightning talks
9 Oct 26 16 Safe robot learning: multi-agent RL, safety RL, RL under safety filtering, verification of neural controllers Lecture Required Reading 1: Robust Adversarial Reinforcement Learning. Pinto et al. (2017)
Required Reading 2: Bridging Hamilton-Jacobi Safety Analysis and Reinforcement Learning. Fisac et al. (2019)
Required Reading 3: Safety Verification and Robustness Analysis of Neural Networks via Quadratic Constraints and Semidefinite Programming. Fazlyab et al. (2020)
Optional Reading: MAGICS: Adversarial RL with Minimax Actors Guided by Implicit Critic Stackelberg. Wang et al. (2024)
Optional Reading: Provably Optimal Reinforcement Learning under Safety Filtering. Oh et al. (2025)
Optional Reading: Verification of Neural Reachable Tubes via Scenario Optimization and Conformal Prediction. Lin and Bansal (2024)
9 Oct 28 17 Inverse RL and games Lecture
Debate
HW release
Debate Proposition: RL/IL is more useful than optimal control for human-centered robotics.
Required Reading: Maximum Entropy Inverse Reinforcement Learning. Ziebart et al. (2008)
Homework 3: Safe RL, inverse games
10 Nov 2 18 Learning rewards Paper discussion Paper 1: Asking Easy Questions: A User-Friendly Approach to Active Reward Learning. Biyik et al. (2019)
Paper 2: Reward-Rational (Implicit) Choice: A Unifying Formalism for Reward Learning. Jeon et al. (2020)
Paper 3: Inferring Objectives in Continuous Dynamic Games from Noise-Corrupted Partial State Observations. Peters et al. (2021)
10 Nov 4 19 RLHF, Human–AI safety
(Guest lecture: Kaiqu Liang)
Lecture
Debate
Debate Proposition: More data can enhance alignment and safety in human-centered robotics.
Required Reading: Deep Reinforcement Learning from Human Preferences. Christiano et al. (2017)
Required Reading: Human-AI Safety: A Descendant of Generative AI and Control Systems Safety. Bajcsy and Fisac (2024)
Optional Reading: From Refusal to Recovery: A Control-Theoretic Approach to Generative AI Guardrails. Pandya et al. (2025)
Optional Reading: Open Problems and Fundamental Limitations of Reinforcement Learning from Human Feedback. Casper et al. (2023)
11 Nov 9 20 Embodied AI safety Paper discussion Paper 1: Safety Guardrails for LLM-Enabled Robots. Ravichandran et al. (2025)
Paper 2: When Your AIs Deceive You: Challenges of Partial Observability in Reinforcement Learning from Human Feedback. Lang et al. (2024)
Paper 3: RLHS: Mitigating Misalignment in RLHF with Hindsight Simulation. Liang et al. (2025)
Module IV: Robots that Help Humans Learn
11 Nov 11 21 Legibility and predictability Lecture
HW due
Required Reading: Legibility and Predictability of Robot Motion. Dragan et al. (2013)
12 Nov 16 22 AI coaching Lecture
Debate
Debate Proposition: Explicit representations of human skill level are essential for a robot to effectively teach humans.
12 Nov 18 23 Robots that assist and teach Paper discussion Paper 1: Dreaming to Assist: Learning to Align with Human Objectives for Shared Control in High-Speed Racing. DeCastro et al. (2024)
Paper 2: Computational Teaching for Driving via Multi-Task Imitation Learning. Gopinath et al. (2024)
Paper 3: Safety with Agency: Human-Centered Safety Filter with Application to AI-Assisted Motorsports. Oh et al. (2025)
13 Nov 23 Fall Recess 🏖️
13 Nov 25 Fall Recess 🏖️
14 Nov 30 24 Guest lecture on AI coaching / Human–AI co-evolution Lecture
14 Dec 2 25 Course wrap-up Lecture
Debate
Debate Proposition: AI systems should intentionally allow users to fail in order to promote long-term learning and independence when coaching humans.
Required Reading: Learning from Errors. Metcalfe (2017)
Required Reading: The Mundanity of Excellence. Chambliss (1989)
15 Dec 7 26 Project Final Presentations 1 Project presentation
15 Dec 9 27 Project Final Presentations 2 Project presentation