Robot Learning (M2608.002700) Spring 2026

Instructor: Prof. Songhwai Oh (오성회)
Email: songhwai (at) snu.ac.kr
Office Hours: Friday 3:00-4:00PM
Office: Building 133 Room 503
Course Number: M2608.002700
Time: M/W 3:30-4:45PM
Location: Building 301 Room 106
TA: Geunje Cheon (천근제)
Email: geunje.cheon (at) rllab.snu.ac.kr
Office: Building 133 Room 610
 

Course Description

An intelligent robotic agent is required to be able to adapt and learn from new environments, tasks, and situations through interactions with its surroundings. Robot learning is a field that combines robotics and artificial intelligence (AI) to develop technologies for robotic agents that can acquire new skills and knowledge through experience, similar to humans. Robot learning encompasses various techniques and approaches that enable robots to become more autonomous and capable of performing tasks without explicit programming for all possible scenarios. In this course, we will review recent advances in robot learning, including imitation learning and deep reinforcement learning. We will first review Markov decision processes (MDP) and reinforcement learning. Then we will discuss recent developments in imitation learning, deep learning, and deep reinforcement learning, including topics such as behavior cloning, inverse reinforcement learning, policy gradient, deep Q-network (DQN), generative adversarial imitation learning, maximum entropy reinforcement learning, safe reinforcement learning, and offline reinforcement learning. Recent advances in robot foundation models will be discussed as well. This is an advanced graduate course and substantial reading and programming projects will be assigned. Students are expected to participate actively in class. Lectures will be in English.

Announcements

  • Project Schedule
  • [05/06] Project proposal (2 pages, 2 columns; no Abstract; IEEE double-column format)
  • [06/03] Project summary: submit (1) Title and (2) Abstract to TA
  • [06/08] Project video: submit one minute presentation video to TA
  • [06/10] Project presentation and poster session
  • [06/12] Project report (minimum 6 pages, 2 columns; IEEE double-column format)

Schedule

  • Week 1:
  • Week 2:
  • Week 3:
  • HW1: Deep learning tutorial (03/16 ~ 03/23 23:59KST)
  • Week 4:

  • HW2: Imitation learning (03/23 ~ 03/30 23:59KST)

  • Week 5:
  • HW3: Deep Q learning (03/30 ~ 04/06 23:59KST)
  • Week 6:
  • HW4: Proximal policy optimization (04/06 ~ 04/13 23:59KST)
  • J. Schulman, S. Levine, P. Abbeel, M. I. Jordan, and P. Moritz, "Trust Region Policy Optimization," in Proc. of the International Conference on Machine Learning (ICML), Jul. 2015. [arXiv]
  • Sergey Levine, Vladlen Koltun, "Guided Policy Search," in Proc. of the International Conference on Machine Learning (ICML), Jun. 2013.
  • Mnih, Volodymyr, Adria Puigdomenech Badia, Mehdi Mirza, Alex Graves, Timothy Lillicrap, Tim Harley, David Silver, and Koray Kavukcuoglu. "Asynchronous methods for deep reinforcement learning." In International Conference on Machine Learning, pp. 1928-1937. 2016.
  • David Silver, Guy Lever, Nicolas Heess, Thomas Degris, Daan Wierstra, and Martin A. Riedmiller, "Deterministic Policy Gradient Algorithms," in Proc. of the International Conference on Machine Learning (ICML), Jun. 2014.
  • Week 7:
  • HW5: Deep deterministic policy gradient (04/13 ~ 04/20 23:59KST)
  • Week 8
  • Bagnell, J. Andrew, Nathan Ratliff, and Martin Zinkevich. "Maximum margin planning." In Proceedings of the International Conference on Machine Learning (ICML). 2006.
  • Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, Yoshua Bengio, "Generative Adversarial Nets," Advances in Neural Information Processing Systems (NIPS), Dec. 2014.
  • Week 9
  • Week 10: Paper Presentation
  • 05/04: Imitation Learning
  • 05/06: RL Algorithm
  • Allen Z. Ren, Justin Lidard, Lars L. Ankile, Anthony Simeonov, Pulkit Agrawal, Anirudha Majumdar, Benjamin Burchfiel, Hongkai Dai, and Max Simchowitz, "Diffusion Policy Policy Optimization", in Proc. of the International Conference on Learning Representations (ICLR), 2025.
  • Week 11: Paper Presentation
  • 05/11: Offline RL
  • Jost Tobias Springenberg, Abbas Abdolmaleki, Jingwei Zhang, Oliver Groth, Michael Bloesch, Thomas Lampe, Philemon Brakel, Sarah Bechtle, Steven Kapturowski, Roland Hafner, Nicolas Heess, and Martin Riedmiller, "Offline Actor-Critic Reinforcement Learning Scales to Large Models", in Proc. of the International Conference on Machine Learning (ICML), 2024.
  • 05/13: Offline RL / Safe RL
  • Week 12: Paper Presentation
  • 05/18: Safe RL / Distributional RL
  • 05/20: Unsupervised RL
  • Week 13: Paper Presentation
  • 05/25: (Hoilday)
  • 05/27: Unsupervised RL / Multi-Agent RL
  • Week 14: Paper Presentation
  • 06/01: RLHF
  • 06/03: (Holiday)
  • Week 15
  • 06/08: Model-Based RL
  • 06/10: Poster Presentation

References

  • Artificial Intelligence: A Modern Approach (4th edition), Stuart Russell and Peter Norvig, Prentice Hall, 2022. (AIMA Website)

Prerequisites

  • Also requires strong background in algorithms, linear algebra, probability, and programming.