Ultimate 強化學習 Solutions for Everyone

Discover all-in-one 強化學習 tools that adapt to your needs. Reach new heights of productivity with ease.

強化學習

  • A Python Pygame environment for developing and testing reinforcement-learning autonomous driving agents on customizable tracks.
    0
    0
    What is SelfDrivingCarSimulator?
    SelfDrivingCarSimulator is a lightweight Python framework built on Pygame that offers a 2D driving environment for training autonomous vehicle agents using reinforcement learning. It supports customizable track layouts, configurable sensor models (like LiDAR and camera emulation), real-time visualization, and data logging for performance analysis. Developers can integrate their RL algorithms, adjust physics parameters, and monitor metrics such as speed, collision rate, and reward functions to iterate quickly on self-driving research and educational projects.
  • A Python framework enabling the design, simulation, and reinforcement learning of cooperative multi-agent systems.
    0
    0
    What is MultiAgentModel?
    MultiAgentModel provides a unified API to define custom environments and agent classes for multi-agent scenarios. Developers can specify observation and action spaces, reward structures, and communication channels. Built-in support for popular RL algorithms like PPO, DQN, and A2C allows training with minimal configuration. Real-time visualization tools help monitor agent interactions and performance metrics. The modular architecture ensures easy integration of new algorithms and custom modules. It also includes a flexible configuration system for hyperparameter tuning, logging utilities for experiment tracking, and compatibility with OpenAI Gym environments for seamless portability. Users can collaborate on shared environments and replay logged sessions for analysis.
  • An open-source Python framework featuring Pacman-based AI agents for implementing search, adversarial, and reinforcement learning algorithms.
    0
    0
    What is Berkeley Pacman Projects?
    The Berkeley Pacman Projects repository offers a modular Python codebase where users build and test AI agents in a Pacman maze. It guides learners through uninformed and informed search (DFS, BFS, A*), adversarial multi-agent search (minimax, alpha-beta pruning), and reinforcement learning (Q-learning with feature extraction). Integrated graphical interfaces visualize agent behavior in real time, while built-in test cases and an autograder verify correctness. By iterating on algorithm implementations, users gain practical experience in state space exploration, heuristic design, adversarial reasoning, and reward-based learning within a unified game framework.
  • BomberManAI is a Python-based AI agent that autonomously navigates and battles in Bomberman game environments using search algorithms.
    0
    0
    What is BomberManAI?
    BomberManAI is an AI agent designed to play the classic Bomberman game autonomously. Developed in Python, it interfaces with a game environment to perceive map states, available moves, and opponent positions in real time. The core algorithm combines A* pathfinding, breadth-first search for reachability analysis, and a heuristic evaluation function to determine optimal bomb placement and evasion strategies. The agent handles dynamic obstacles, power-ups, and multiple opponents on various map layouts. Its modular architecture enables developers to experiment with custom heuristics, reinforcement learning modules, or alternative decision-making strategies. Ideal for game AI researchers, students, and competitive bot developers, BomberManAI provides a flexible framework for testing and improving autonomous gaming agents.
  • Open-source PyTorch-based framework implementing CommNet architecture for multi-agent reinforcement learning with inter-agent communication enabling collaborative decision-making.
    0
    0
    What is CommNet?
    CommNet is a research-oriented library that implements the CommNet architecture, allowing multiple agents to share hidden states at each timestep and learn to coordinate actions in cooperative environments. It includes PyTorch model definitions, training and evaluation scripts, environment wrappers for OpenAI Gym, and utilities for customizing communication channels, agent counts, and network depths. Researchers and developers can use CommNet to prototype and benchmark inter-agent communication strategies on navigation, pursuit–evasion, and resource-collection tasks.
  • A customizable reinforcement learning environment library for benchmarking AI agents on data processing and analytics tasks.
    0
    0
    What is DataEnvGym?
    DataEnvGym delivers a collection of modular, customizable environments built on the Gym API to facilitate reinforcement learning research in data-driven domains. Researchers and engineers can select from built-in tasks like data cleaning, feature engineering, batch scheduling, and streaming analytics. The framework supports seamless integration with popular RL libraries, standardized benchmarking metrics, and logging tools to track agent performance. Users can extend or combine environments to model complex data pipelines and evaluate algorithms under realistic constraints.
  • A Java-based platform enabling development, simulation, and deployment of intelligent multi-agent systems with communication, negotiation, and learning capabilities.
    0
    0
    What is IntelligentMASPlatform?
    The IntelligentMASPlatform is built to accelerate development and deployment of multi-agent systems by offering a modular architecture with distinct agent, environment, and service layers. Agents communicate using FIPA-compliant ACL messaging, enabling dynamic negotiation and coordination. The platform includes a versatile environment simulator allowing developers to model complex scenarios, schedule agent tasks, and visualize agent interactions in real-time through a built-in dashboard. For advanced behaviors, it integrates reinforcement learning modules and supports custom behavior plugins. Deployment tools allow packaging agents into standalone applications or distributed networks. Additionally, the platform's API facilitates integration with databases, IoT devices, or third-party AI services, making it suitable for research, industrial automation, and smart city use cases.
  • A PyTorch framework enabling agents to learn emergent communication protocols in multi-agent reinforcement learning tasks.
    0
    0
    What is Learning-to-Communicate-PyTorch?
    This repository implements emergent communication in multi-agent reinforcement learning using PyTorch. Users can configure sender and receiver neural networks to play referential games or cooperative navigation, encouraging agents to develop a discrete or continuous communication channel. It offers scripts for training, evaluation, and visualization of learned protocols, along with utilities for environment creation, message encoding, and decoding. Researchers can extend it with custom tasks, modify network architectures, and analyze protocol efficiency, fostering rapid experimentation in emergent agent communication.
  • Simulates dynamic e-commerce negotiations using customizable buyer and seller AI agents with negotiation protocols and visualization.
    0
    0
    What is Multi-Agent-Seller?
    Multi-Agent-Seller provides a modular environment for simulating e-commerce negotiations using AI agents. It includes pre-built buyer and seller agents with customizable negotiation strategies, such as dynamic pricing, time-based concessions, and utility-based decision-making. Users can define custom protocols, message formats, and market conditions. The framework handles session management, offer tracking, and result logging with built-in visualization tools for analyzing agent interactions. It integrates easily with machine learning libraries for strategy development, enabling experimentation with reinforcement learning or rule-based agents. Its extensible architecture allows adding new agent types, negotiation rules, and visualization plugins. Multi-Agent-Seller is ideal for testing multi-agent algorithms, studying negotiation behaviors, and teaching concepts in AI and e-commerce domains.
  • NeuralABM trains neural-network-driven agents to simulate complex behaviors and environments in agent-based modeling scenarios.
    0
    0
    What is NeuralABM?
    NeuralABM is an open-source Python library that leverages PyTorch to integrate neural networks into agent-based modeling. Users can specify agent architectures as neural modules, define environment dynamics, and train agent behaviors using backpropagation across simulation steps. The framework supports custom reward signals, curriculum learning, and synchronous or asynchronous updates, enabling the study of emergent phenomena. With utilities for logging, visualization, and dataset export, researchers and developers can analyze agent performance, debug models, and iterate on simulation designs. NeuralABM simplifies combining reinforcement learning with ABM for applications in social science, economics, robotics, and AI-driven game NPC behaviors. It provides modular components for environment customization, supports multi-agent interactions, and offers hooks for integrating external datasets or APIs for real-world simulations. The open design fosters reproducibility and collaboration through clear experiment configuration and version control integration.
  • An open-source framework enabling training, deployment, and evaluation of multi-agent reinforcement learning models for cooperative and competitive tasks.
    0
    0
    What is NKC Multi-Agent Models?
    NKC Multi-Agent Models provides researchers and developers with a comprehensive toolkit for designing, training, and evaluating multi-agent reinforcement learning systems. It features a modular architecture where users define custom agent policies, environment dynamics, and reward structures. Seamless integration with OpenAI Gym allows for rapid prototyping, while support for TensorFlow and PyTorch enables flexibility in selecting learning backends. The framework includes utilities for experience replay, centralized training with decentralized execution, and distributed training across multiple GPUs. Extensive logging and visualization modules capture performance metrics, facilitating benchmarking and hyperparameter tuning. By simplifying the setup of cooperative, competitive, and mixed-motive scenarios, NKC Multi-Agent Models accelerates experimentation in domains such as autonomous vehicles, robotic swarms, and game AI.
  • An open-source reinforcement learning agent that learns to play Pacman, optimizing navigation and ghost avoidance strategies.
    0
    0
    What is Pacman AI?
    Pacman AI offers a fully functional Python-based environment and agent framework for the classic Pacman game. The project implements key reinforcement learning algorithms—Q-learning and value iteration—to allow the agent to learn optimal policies for pill collection, maze navigation, and ghost avoidance. Users can define custom reward functions and adjust hyperparameters such as learning rate, discount factor, and exploration strategy. The framework supports metric logging, performance visualization, and reproducible experiment setups. It is designed for easy extension, letting researchers and students integrate new algorithms or neural network-based learning approaches and benchmark them against baseline grid-based methods within the Pacman domain.
  • An RL framework offering PPO, DQN training and evaluation tools for developing competitive Pommerman game agents.
    0
    0
    What is PommerLearn?
    PommerLearn enables researchers and developers to train multi-agent RL bots in the Pommerman game environment. It includes ready-to-use implementations of popular algorithms (PPO, DQN), flexible configuration files for hyperparameters, automatic logging and visualization of training metrics, model checkpointing, and evaluation scripts. Its modular architecture makes it easy to extend with new algorithms, customize environments, and integrate with standard ML libraries such as PyTorch.
  • A lightweight Python library for creating customizable 2D grid environments to train and test reinforcement learning agents.
    0
    0
    What is Simple Playgrounds?
    Simple Playgrounds provides a modular platform for building interactive 2D grid environments where agents can navigate mazes, interact with objects, and complete tasks. Users define environment layouts, object behaviors, and reward functions via simple YAML or Python scripts. The integrated Pygame renderer delivers real-time visualization, while a step-based API ensures seamless integration with reinforcement learning libraries like Stable Baselines3. With support for multi-agent setups, collision detection, and customizable physics parameters, Simple Playgrounds streamlines the prototyping, benchmarking, and educational demonstration of AI algorithms.
  • Open-source PyTorch library providing modular implementations of reinforcement learning agents like DQN, PPO, SAC, and more.
    0
    0
    What is RL-Agents?
    RL-Agents is a research-grade reinforcement learning framework built on PyTorch that bundles popular RL algorithms across value-based, policy-based, and actor-critic methods. The library features a modular agent API, GPU acceleration, seamless integration with OpenAI Gym, and built-in logging and visualization tools. Users can configure hyperparameters, customize training loops, and benchmark performance with a few lines of code, making RL-Agents ideal for academic research, prototyping, and industrial experimentation.
  • A reinforcement learning framework enabling autonomous robots to navigate and avoid collisions in multi-agent environments.
    0
    0
    What is RL Collision Avoidance?
    RL Collision Avoidance provides a complete pipeline for developing, training, and deploying multi-robot collision avoidance policies. It offers a set of Gym-compatible simulation scenarios where agents learn collision-free navigation through reinforcement learning algorithms. Users can customize environment parameters, leverage GPU acceleration for faster training, and export learned policies. The framework also integrates with ROS for real-world testing, supports pre-trained models for immediate evaluation, and features tools for visualizing agent trajectories and performance metrics.
  • Dead-simple self-learning is a Python library providing simple APIs for building, training, and evaluating reinforcement learning agents.
    0
    0
    What is dead-simple-self-learning?
    Dead-simple self-learning offers developers a dead-simple approach to create and train reinforcement learning agents in Python. The framework abstracts core RL components, such as environment wrappers, policy modules, and experience buffers, into concise interfaces. Users can quickly initialize environments, define custom policies using familiar PyTorch or TensorFlow backends, and execute training loops with built-in logging and checkpointing. The library supports on-policy and off-policy algorithms, enabling flexible experimentation with Q-learning, policy gradients, and actor-critic methods. By reducing boilerplate code, dead-simple self-learning allows practitioners, educators, and researchers to prototype algorithms, test hypotheses, and visualize agent performance with minimal configuration. Its modular design also facilitates integration with existing ML stacks and custom environments.
  • SoccerAgent uses multi-agent reinforcement learning to train AI players for realistic soccer simulations and strategy optimization.
    0
    0
    What is SoccerAgent?
    SoccerAgent is a specialized AI framework designed for developing and training autonomous soccer agents using state-of-the-art multi-agent reinforcement learning (MARL) techniques. It simulates realistic soccer matches in 2D or 3D environments, offering tools to define reward functions, customize player attributes, and implement tactical strategies. Users can integrate popular RL algorithms (such as PPO, DDPG, and MADDPG) via built-in modules, monitor training progress through dashboards, and visualize agent behaviors in real time. The framework supports scenario-based training for offense, defense, and coordination protocols. With an extensible codebase and detailed documentation, SoccerAgent empowers researchers and developers to analyze team dynamics and refine AI-driven gameplay strategies for academic and commercial projects.
  • An open-source reinforcement learning agent using PPO to train and play StarCraft II via DeepMind's PySC2 environment.
    0
    0
    What is StarCraft II Reinforcement Learning Agent?
    This repository provides an end-to-end reinforcement learning framework for StarCraft II gameplay research. The core agent uses Proximal Policy Optimization (PPO) to learn policy networks that interpret observation data from the PySC2 environment and output precise in-game actions. Developers can configure neural network layers, reward shaping, and training schedules to optimize performance. The system supports multiprocessing for efficient sample collection, logging utilities for monitoring training curves, and evaluation scripts for running trained policies against scripted or built-in AI opponents. The codebase is written in Python and leverages TensorFlow for model definition and optimization. Users can extend components such as custom reward functions, state preprocessing, or network architectures to suit specific research objectives.
  • An RL-based AI agent that learns optimal betting strategies to play heads-up limit Texas Hold'em poker efficiently.
    0
    0
    What is TexasHoldemAgent?
    TexasHoldemAgent provides a modular environment built on Python to train, evaluate, and deploy an AI-powered poker player for heads-up limit Texas Hold’em. It integrates a custom simulation engine with deep reinforcement learning algorithms, including DQN, for iterative policy improvement. Key capabilities include hand state encoding, action space definition (fold, call, raise), reward shaping, and real-time decision evaluation. Users can customize learning parameters, leverage CPU/GPU acceleration, monitor training progress, and load or save trained models. The framework supports batch simulation to test various strategies, generate performance metrics, and visualize win rates, empowering researchers, developers, and poker enthusiasts to experiment with AI-driven gameplay strategies.
Featured