Multi-Agent Reinforcement Learning (MARL) plays a crucial role in robotic coordination and control, yet existing simulation environments often lack the fidelity and scalability needed for real-world applications. In this work, we extend Isaac Lab to support efficient training of both homogeneous and heterogeneous multi-agent robotic policies in high-fidelity physics simulations. Our contributions include the development of diverse MARL environments tailored for robotic coordination, integration of Heterogeneous Agent Reinforcement Learning (HARL) algorithms, and a scalable GPU-accelerated framework optimized for large-scale training. We evaluate our framework using two state-of-the-art MARL algorithms Multi-Agent Reinforcement Learning with Proximal Policy Optimization (MAPPO) and Heterogeneous Agent Reinforcement Learning with Proximal Policy Optimization (HAPPO) across six challenging robotic tasks. Our results confirm the feasibility of training heterogeneous agents in high-fidelity environments while maintaining the scalability and performance benefits of Isaac Lab. By enabling realistic multi-agent learning at scale, our work lays a foundation for advancing MARL research in physics-driven robotics. The source code and demonstration videos are available at https://some45bucks.github.io/IsaacLab-HARL/.
Overview of the developed multi-agent RL environments in IsaacLab.
Comparison of agent coordination before and after training.
Learning curves comparing MAPPO and HAPPO algorithms.
@article{haight2024harl,
author = {Haight, Jacob and Peterson, Isaac and Allred, Christopher and Harper, Mario},
title = {Heterogeneous Multi-Agent Learning in Isaac Lab: Scalable Simulation for Robotic Collaboration},
journal = {TBA},
year = {2024},
}