Comprehensive StarCraft II KI Tools for Every Need

Get access to StarCraft II KI solutions that address multiple requirements. One-stop resources for streamlined workflows.

StarCraft II KI

  • Simplified PyTorch implementation of AlphaStar, enabling StarCraft II RL agent training with modular network architecture and self-play.
    0
    0
    What is mini-AlphaStar?
    mini-AlphaStar demystifies the complex AlphaStar architecture by offering an accessible, open-source PyTorch framework for StarCraft II AI development. It features spatial feature encoders for screen and minimap inputs, non-spatial feature processing, LSTM memory modules, and separate policy and value networks for action selection and state evaluation. Using imitation learning to bootstrap and reinforcement learning with self-play for fine-tuning, it supports environment wrappers compatible with StarCraft II via pysc2, logging through TensorBoard, and configurable hyperparameters. Researchers and students can generate datasets from human gameplay, train models on custom scenarios, evaluate agent performance, and visualize learning curves. The modular codebase enables easy experimentation with network variants, training schedules, and multi-agent setups. Designed for education and prototyping rather than production deployment.
    mini-AlphaStar Core Features
    • Spatial and non-spatial feature encoding
    • LSTM-based memory modules
    • Separate policy and value networks
    • Imitation learning and reinforcement learning pipelines
    • Self-play environment wrappers via pysc2
    • TensorBoard logging and visualization
    • Configurable hyperparameters
    • Modular PyTorch codebase
  • An open-source multi-agent reinforcement learning framework enabling raw-level agent control and coordination in StarCraft II via PySC2.
    0
    0
    What is MultiAgent-Systems-StarCraft2-PySC2-Raw?
    MultiAgent-Systems-StarCraft2-PySC2-Raw offers a complete toolkit for developing, training, and evaluating multiple AI agents in StarCraft II. It exposes low-level controls for unit movement, targeting, and abilities, while allowing flexible reward design and scenario configuration. Users can easily plug in custom neural network architectures, define team-based coordination strategies, and record metrics. Built on top of PySC2, it supports parallel training, checkpointing, and visualization, making it ideal for advancing research in cooperative and adversarial multi-agent reinforcement learning.
Featured