Comprehensive Q-Learning Tools for Every Need

Get access to Q-Learning solutions that address multiple requirements. One-stop resources for streamlined workflows.

Q-Learning

  • simple_rl is a lightweight Python library offering pre-built reinforcement learning agents and environments for rapid RL experimentation.
    0
    0
    What is simple_rl?
    simple_rl is a minimalistic Python library designed to streamline reinforcement learning research and education. It provides a consistent API for defining environments and agents, with built-in support for common RL paradigms including Q-learning, Monte Carlo methods, and dynamic programming algorithms like value and policy iteration. The framework includes sample environments such as GridWorld, MountainCar, and Multi-Armed Bandits, facilitating hands-on experimentation. Users can extend base classes to implement custom environments or agents, while utility functions handle logging, performance tracking, and policy evaluation. simple_rl's lightweight architecture and clear codebase make it ideal for rapid prototyping, teaching RL fundamentals, and benchmarking new algorithms in a reproducible, easy-to-understand environment.
  • An open-source Python framework featuring Pacman-based AI agents for implementing search, adversarial, and reinforcement learning algorithms.
    0
    0
    What is Berkeley Pacman Projects?
    The Berkeley Pacman Projects repository offers a modular Python codebase where users build and test AI agents in a Pacman maze. It guides learners through uninformed and informed search (DFS, BFS, A*), adversarial multi-agent search (minimax, alpha-beta pruning), and reinforcement learning (Q-learning with feature extraction). Integrated graphical interfaces visualize agent behavior in real time, while built-in test cases and an autograder verify correctness. By iterating on algorithm implementations, users gain practical experience in state space exploration, heuristic design, adversarial reasoning, and reward-based learning within a unified game framework.
  • An open-source reinforcement learning agent that learns to play Pacman, optimizing navigation and ghost avoidance strategies.
    0
    0
    What is Pacman AI?
    Pacman AI offers a fully functional Python-based environment and agent framework for the classic Pacman game. The project implements key reinforcement learning algorithms—Q-learning and value iteration—to allow the agent to learn optimal policies for pill collection, maze navigation, and ghost avoidance. Users can define custom reward functions and adjust hyperparameters such as learning rate, discount factor, and exploration strategy. The framework supports metric logging, performance visualization, and reproducible experiment setups. It is designed for easy extension, letting researchers and students integrate new algorithms or neural network-based learning approaches and benchmark them against baseline grid-based methods within the Pacman domain.
  • Dead-simple self-learning is a Python library providing simple APIs for building, training, and evaluating reinforcement learning agents.
    0
    0
    What is dead-simple-self-learning?
    Dead-simple self-learning offers developers a dead-simple approach to create and train reinforcement learning agents in Python. The framework abstracts core RL components, such as environment wrappers, policy modules, and experience buffers, into concise interfaces. Users can quickly initialize environments, define custom policies using familiar PyTorch or TensorFlow backends, and execute training loops with built-in logging and checkpointing. The library supports on-policy and off-policy algorithms, enabling flexible experimentation with Q-learning, policy gradients, and actor-critic methods. By reducing boilerplate code, dead-simple self-learning allows practitioners, educators, and researchers to prototype algorithms, test hypotheses, and visualize agent performance with minimal configuration. Its modular design also facilitates integration with existing ML stacks and custom environments.
  • Open-source Python toolkit offering random, rule-based pattern recognition, and reinforcement learning agents for Rock-Paper-Scissors.
    0
    0
    What is AI Agents for Rock Paper Scissors?
    AI Agents for Rock Paper Scissors is an open-source Python project that demonstrates how to build, train, and evaluate different AI strategies—random play, rule-based pattern recognition, and reinforcement learning (Q-learning)—in the classic Rock-Paper-Scissors game. It provides modular agent classes, a configurable game runner, performance logging, and visualization utilities. Users can easily swap agents, adjust learning parameters, and explore AI behavior in competitive scenarios.
  • HFO_DQN is a reinforcement learning framework that applies Deep Q-Network to train soccer agents in RoboCup Half Field Offense environment.
    0
    0
    What is HFO_DQN?
    HFO_DQN combines Python and TensorFlow to deliver a complete pipeline for training soccer agents using Deep Q-Networks. Users can clone the repository, install dependencies including the HFO simulator and Python libraries, and configure training parameters in YAML files. The framework implements experience replay, target network updates, epsilon-greedy exploration, and reward shaping tailored for the half field offense domain. It features scripts for agent training, performance logging, evaluation matches, and plotting results. Modular code structure allows integration of custom neural network architectures, alternative RL algorithms, and multi-agent coordination strategies. Outputs include trained models, performance metrics, and behavior visualizations, facilitating research in reinforcement learning and multi-agent systems.
  • Jason-RL equips Jason BDI agents with reinforcement learning, enabling Q-learning and SARSA-based adaptive decision making through reward experience.
    0
    0
    What is jason-RL?
    jason-RL adds a reinforcement learning layer to the Jason multi-agent framework, allowing AgentSpeak BDI agents to learn action-selection policies via reward feedback. It implements Q-learning and SARSA algorithms, supports configuration of learning parameters (learning rate, discount factor, exploration strategy), and logs training metrics. By defining reward functions in agent plans and running simulations, developers can observe agents improve decision making over time, adapting to changing environments without manual policy coding.
Featured