Our paper on constrained multi-objective reinforcement learning is accepted to ICLR 2025
[2025.01.23]
The following paper is accepted to the International Conference on Learning Representations (ICLR 2025):
Conflict-Averse Gradient Aggregation for Constrained Multi-Objective Reinforcement Learning by Dohyeong Kim, Mineui Hong, Jeongho Park, and Songhwai Oh
- Abstract: In real-world applications, a reinforcement learning (RL) agent should consider multiple objectives and adhere to safety guidelines. To address these considerations, we propose a constrained multi-objective RL algorithm named Constrained Multi-Objective Gradient Aggregator (CoMOGA). In the field of multi-objective optimization, managing conflicts between the gradients of the multiple objectives is crucial to prevent policies from converging to local optima. It is also essential to efficiently handle safety constraints for stable training and constraint satisfaction. We address these challenges straightforwardly by treating the maximization of multiple objectives as a constrained optimization problem (COP), where the constraints are defined to improve the original objectives. Existing safety constraints are then integrated into the COP, and the policy is updated by solving the COP, which ensures the avoidance of gradient conflicts. Despite its simplicity, CoMOGA guarantees convergence to global optima in a tabular setting. Through various experiments, we have confirmed that preventing gradient conflicts is critical, and the proposed method achieves constraint satisfaction across all tasks.
- Video
