MAPF_G2RL implements a graph-to-reinforcement learning pipeline to train centralized and decentralized agents that compute collision-free paths for multiple agents. It provides graph encoding, reward shaping, scenario generation, and performance evaluation modules. Users can configure graph topologies, agent counts, and training hyperparameters to adapt to varied environments.
MAPF_G2RL implements a graph-to-reinforcement learning pipeline to train centralized and decentralized agents that compute collision-free paths for multiple agents. It provides graph encoding, reward shaping, scenario generation, and performance evaluation modules. Users can configure graph topologies, agent counts, and training hyperparameters to adapt to varied environments.
MAPF_G2RL is an open-source research framework that bridges graph theory and deep reinforcement learning to tackle the multi-agent path finding (MAPF) problem. It encodes nodes and edges into vector representations, defines spatial and collision-aware reward functions, and supports various RL algorithms such as DQN, PPO, and A2C. The framework automates scenario creation by generating random graphs or importing real-world maps, and orchestrates training loops that optimize policies for multiple agents simultaneously. After learning, agents are evaluated in simulated environments to measure path optimality, makespan, and success rates. Its modular design allows researchers to extend core components, integrate new MARL techniques, and benchmark against classical solvers.
Who will use MAPF_G2RL?
AI researchers
Robotics engineers
Multi-agent systems developers
Graduate students in reinforcement learning
Warehouse automation teams
How to use the MAPF_G2RL?
Step1: Clone the MAPF_G2RL repository from GitHub
Step2: Install dependencies via pip using requirements.txt
Step3: Configure graph and training parameters in config files
Step4: Run the training script to train RL agents
Step5: Evaluate trained models on simulated environments
Step6: Analyze results and adjust hyperparameters as needed
Platform
mac
windows
linux
MAPF_G2RL's Core Features & Benefits
The Core Features
Graph encoding and preprocessing
Customizable reward shaping modules
Support for DQN, PPO, A2C algorithms
Scenario generator for random and real-world maps
Multi-agent training and evaluation pipelines
Performance logging and visualization tools
The Benefits
Accelerates MAPF research with ready-to-use RL pipelines
Improves path finding quality and scalability
Flexible configuration for diverse graph types
Easy extensibility for new algorithms
GPU acceleration for faster training
MAPF_G2RL's Main Use Cases & Applications
Robot fleet navigation in warehouses
Autonomous drone path planning in delivery networks