Robot Learning (M2608.002700) Spring 2025

Instructor: Prof. Songhwai Oh (오성회)
Email: songhwai (at) snu.ac.kr
Office Hours: Friday 3:00-4:00PM
Office: Building 133 Room 503
Course Number: M2608.002700
Time: M/W 3:30-4:45PM
Location: Building 301 Room 106
TA: Hojun Chung (정호준)
Email: hojun.chung (at) rllab.snu.ac.kr
Office: Building 133 Room 610
 

Course Description

An intelligent robotic agent is required to be able to adapt and learn from new environments, tasks, and situations through interactions with their surroundings. Robot learning is a field that combines robotics and artificial intelligence (AI) to develop technologies for robotic agents which can acquire new skills and knowledge through experience, similar to humans. Robot learning includes various techniques and approaches to make robots more autonomous and capable of performing tasks without explicit programming for all possible scenarios. In this course, we will review recent advances in robot learning, including imitation learning and deep reinforcement learning. We will first review Markov decision processes (MDP) and reinforcement learning. Then we will discuss recent developments in imitation learning, deep learning, and deep reinforcement learning, including topics such as behavior cloning, inverse reinforcement learning, policy gradient, deep Q-network (DQN), generative adversarial imitation learning, maximum entropy reinforcement learning, safe reinforcement learning, and offline reinforcement learning. This is an advanced graduate course and substantial reading and programming projects will be assigned. Students are expected to participate actively in class. Lectures will be in English.

Announcements

  • Project Schedule

  • [05/07] Project proposal (2 pages, 2 columns; no Abstract; IEEE double-column format)
  • [06/04] Project summary: submit (1) Title and (2) Abstract to TA
  • [06/09] Project video: submit one minute presentation video to TA
  • [06/11] Project presentation and poster session
  • [06/13] Project report (minimum 6 pages, 2 columns; IEEE double-column format)

Schedule

  • Week 1:

  • 03/05: Introduction

  • Week 2:

  • 03/10: Review on MDPs and POMDPs

  •  03/12: Reinforcement learning

  • Week 3:

  • 03/17: Review of probability theory; Gaussian process regression
  • 03/19: Behavior cloning (leveraged Gaussian process regression)

  • Week 4:

  • 03/24: DAgger, Behavior cloning applications

  • 03/26: Deep Q Learning

  • Week 5:

  • 03/31: Sparse MDPs
  • 04/02: Policy gradient

  • Week 6:

  • 04/07: GPS, TRPO, PPO

  • J. Schulman, S. Levine, P. Abbeel, M. I. Jordan, and P. Moritz, "Trust Region Policy Optimization," in Proc. of the International Conference on Machine Learning (ICML), Jul. 2015. [arXiv]
  • Sergey Levine, Vladlen Koltun, "Guided Policy Search," in Proc. of the International Conference on Machine Learning (ICML), Jun. 2013.
  • 04/09: Actor-critic

  • Mnih, Volodymyr, Adria Puigdomenech Badia, Mehdi Mirza, Alex Graves, Timothy Lillicrap, Tim Harley, David Silver, and Koray Kavukcuoglu. "Asynchronous methods for deep reinforcement learning." In International Conference on Machine Learning, pp. 1928-1937. 2016.
  • David Silver, Guy Lever, Nicolas Heess, Thomas Degris, Daan Wierstra, and Martin A. Riedmiller, "Deterministic Policy Gradient Algorithms," in Proc. of the International Conference on Machine Learning (ICML), Jun. 2014. 
  • Week 7 

  • 04/14: Maximum entropy RL

  • 04/16: Inverse reinforcement learning (IRL), GPIRL
  • Week 8 

  • 04/21: Maximum entropy IRL

  • Bagnell, J. Andrew, Nathan Ratliff, and Martin Zinkevich. "Maximum margin planning." In Proceedings of the International Conference on Machine Learning (ICML). 2006.
  • 04/23: GAN, GAIL, MCTEIL

  • Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, Yoshua Bengio, "Generative Adversarial Nets," Advances in Neural Information Processing Systems (NIPS), Dec. 2014.
  • Week 9:

  • 04/28: Safe Reinforcement Learning

  • 04/30: Offline Reinforcement Learning

  • Week 10: Paper Presentation

  • 05/05: (Holiday)
  • 05/07: Imitation Learning

  • Week 11: Paper Presentation

  • 05/12: Deep RL

  • 05/14: Deep RL

  • Week 12: Paper Presentation

  • 05/19: Diffusion Models / Flow Matching

  • Xixi Hu, Bo Liu, Xingchao Lui, Qiang Liu, "AdaFlow: Imitation Learning with Variance-Adaptive Flow-Based Policies",  Advances in Neural Information Processing Systems (NeurIPS), 2024.
  • 05/21: Distributional RL

  • Week 13: Paper Presentation

  • 05/26: Unsupervised RL

  • 05/28: Multi-Agent RL

  • Week 14: Paper Presentation

  • 06/02: Reinforcement Learning from Human Feedback (RLHF)

  • 06/04: Curriculum RL / Environment Design

  • Week 15: Paper & Poster Presentation

  • 06/09: Model-Based RL

  • Room 106 and 1st Floor, Building 301

References

  • Artificial Intelligence: A Modern Approach (4th edition), Stuart Russell and Peter Norvig, Prentice Hall, 2022. (AIMA Website)

Prerequisites

  • (430.457) Introduction to Intelligent Systems (지능시스템개론).
  • Also requires strong background in algorithms, linear algebra, probability, and programming.