Comprehensive Docker 배포 Tools for Every Need

Get access to Docker 배포 solutions that address multiple requirements. One-stop resources for streamlined workflows.

Docker 배포

  • Aladin is an open-source autonomous LLM agent enabling scripted workflows, memory-enabled decision-making, and plugin-based task orchestration.
    0
    0
    What is Aladin?
    Aladin provides a modular architecture that allows developers to define autonomous agents powered by large language models (LLMs). Each agent can load memory backends (e.g., SQLite, in-memory), utilize dynamic prompt templates, and integrate custom plugins for external API calls or local command execution. It features a task planner that breaks high-level goals into sequenced actions, executing them in order and iterating based on LLM feedback. Configuration is managed through YAML files and environment variables, making it adaptable to various use cases. Users can deploy Aladin via Docker Compose or pip installation. The CLI and FastAPI-based HTTP endpoints let users trigger agents, monitor execution, and inspect memory states, facilitating integration with CI/CD pipelines, chat interfaces, or custom dashboards.
  • A Docker-based framework to rapidly deploy and orchestrate autonomous GPT agents with built-in dependencies for reproducible development environments.
    0
    0
    What is Kurtosis AutoGPT Package?
    The Kurtosis AutoGPT Package is an AI Agent framework packaged as a Kurtosis module that delivers a fully configured AutoGPT environment with minimal effort. It provisions and wires up services such as PostgreSQL, Redis, and a vector store, then injects your API keys and agent scripts into the network. Using Docker and Kurtosis CLI, you can spin up isolated agent instances, view logs, adjust budgets, and manage network policies. This package removes infrastructure friction so teams can rapidly develop, test, and scale autonomous GPT-driven workflows in a reproducible manner.
  • Co-Sight is an open-source AI framework offering real-time video analytics for object detection, tracking, and distributed inference.
    0
    0
    What is Co-Sight?
    Co-Sight is an open-source AI framework that simplifies development and deployment of real-time video analytics solutions. It provides modules for video data ingestion, preprocessing, model training, and distributed inference on edge and cloud. With built-in support for object detection, classification, tracking, and pipeline orchestration, Co-Sight ensures low-latency processing and high throughput. Its modular design integrates with popular deep learning libraries and scales seamlessly using Kubernetes. Developers can define pipelines via YAML, deploy with Docker, and monitor performance through a web dashboard. Co-Sight empowers users to build advanced vision applications for smart city surveillance, intelligent transportation, and industrial quality inspection, reducing development time and operational complexity.
  • CrewAI Agent Generator quickly scaffolds customized AI agents with prebuilt templates, seamless API integration, and deployment tools.
    0
    0
    What is CrewAI Agent Generator?
    CrewAI Agent Generator leverages a command-line interface to let you initialize a new AI agent project with opinionated folder structures, sample prompt templates, tool definitions, and testing stubs. You can configure connections to OpenAI, Azure, or custom LLM endpoints; manage agent memory using vector stores; orchestrate multiple agents in collaborative workflows; view detailed conversation logs; and deploy your agents to Vercel, AWS Lambda, or Docker with built-in scripts. It accelerates development and ensures consistent architecture across AI agent projects.
  • An open-source framework enabling creation and orchestration of multiple AI agents that collaborate on complex tasks via JSON messaging.
    0
    0
    What is Multi AI Agent Systems?
    This framework allows users to design, configure, and deploy multiple AI agents that communicate via JSON messages through a central orchestrator. Each agent can have distinct roles, prompts, and memory modules, and you can plug in any LLM provider by implementing a provider interface. The system supports persistent conversation history, dynamic routing, and modular extensions. Ideal for simulating debates, automating customer support flows, or coordinating multi-step document generation, it runs on Python, with Docker support for containerized deployments.
  • OmniMind0 is an open-source Python framework enabling autonomous multi-agent workflows with built-in memory management and plugin integration.
    0
    0
    What is OmniMind0?
    OmniMind0 is a comprehensive agent-based AI framework written in Python that allows creation and orchestration of multiple autonomous agents. Each agent can be configured to handle specific tasks—such as data retrieval, summarization, or decision-making—while sharing state through pluggable memory backends like Redis or JSON files. The built-in plugin architecture lets you extend functionality with external APIs or custom commands. It supports OpenAI, Azure, and Hugging Face models, and offers deployment via CLI, REST API server, or Docker for flexible integration into your workflows.
  • Taiga is an open-source AI agent framework enabling creation of autonomous LLM agents with plugin extensibility, memory, and tool integration.
    0
    0
    What is Taiga?
    Taiga is a Python-based open-source AI agent framework designed to streamline the creation, orchestration, and deployment of autonomous large language model (LLM) agents. The framework includes a flexible plugin system for integrating custom tools and external APIs, a configurable memory module for managing long-term and short-term conversational context, and a task chaining mechanism to sequence multi-step workflows. Taiga also offers built-in logging, metrics, and error handling for production readiness. Developers can quickly scaffold agents with templates, extend functionality via SDK, and deploy across platforms. By abstracting complex orchestration logic, Taiga enables teams to focus on building intelligent assistants that can research, plan, and execute actions without manual intervention.
  • An extensible AI agent framework for designing, testing, and deploying multi-agent workflows with custom skills.
    0
    0
    What is ByteChef?
    ByteChef offers a modular architecture to build, test, and deploy AI agents. Developers define agent profiles, attach custom skill plugins, and orchestrate multi-agent workflows through a visual web IDE or SDK. It integrates with major LLM providers (OpenAI, Cohere, self-hosted models) and external APIs. Built-in debugging, logging, and observability tools streamline iteration. Projects can be deployed as Docker services or serverless functions, enabling scalable, production-ready AI agents for customer support, data analysis, and automation.
  • Pi Web Agent is an open-source web-based AI agent integrating LLMs for conversational tasks and knowledge retrieval.
    0
    0
    What is Pi Web Agent?
    Pi Web Agent is a lightweight, extensible framework for building AI chat agents on the web. It leverages Python FastAPI on the backend and a React frontend to deliver interactive conversations powered by OpenAI, Cohere, or local LLMs. Users can upload documents or connect external databases for semantic search via vector stores. A plugin architecture allows custom tools, function calls, and third-party API integrations locally, it offers full source code access, role-based prompt templates, and configurable memory storage to create customized AI assistants.
  • A web platform to discover, explore, and deploy diverse AI agents with searchable categories in one unified marketplace.
    0
    0
    What is AI Agent Marketplace?
    AI Agent Marketplace is built with Next.js and React to provide a centralized hub where users can browse, evaluate, and deploy a wide range of AI agents. The platform pulls agent metadata from community contributions, offering detailed descriptions, capability tags, and live in-browser demos. Users can filter agents by domain, function, or technology provider. For developers, the open-source repository includes a modular architecture with support for adding new agent entries, configuring API endpoints, and customizing UI components. Deployment options include hosting on Vercel or local Docker containers. By consolidating disparate AI agent projects into one searchable interface, the marketplace accelerates experimentation, collaboration, and integration into production workflows.
Featured