Comprehensive despliegue con Docker Tools for Every Need

Get access to despliegue con Docker solutions that address multiple requirements. One-stop resources for streamlined workflows.

despliegue con Docker

  • An open-source Python framework to build, orchestrate and deploy AI agents with memory, tools, and multi-model support.
    0
    0
    What is Agentfy?
    Agentfy provides a modular architecture for constructing AI agents by combining LLMs, memory backends, and tool integrations into a cohesive runtime. Developers declare agent behavior using Python classes, register tools (REST APIs, databases, utilities), and choose memory stores (local, Redis, SQL). The framework orchestrates prompts, actions, tool calls, and context management to automate tasks. Built-in CLI and Docker support enable one-step deployment to cloud, edge, or desktop environments.
  • Co-Sight is an open-source AI framework offering real-time video analytics for object detection, tracking, and distributed inference.
    0
    0
    What is Co-Sight?
    Co-Sight is an open-source AI framework that simplifies development and deployment of real-time video analytics solutions. It provides modules for video data ingestion, preprocessing, model training, and distributed inference on edge and cloud. With built-in support for object detection, classification, tracking, and pipeline orchestration, Co-Sight ensures low-latency processing and high throughput. Its modular design integrates with popular deep learning libraries and scales seamlessly using Kubernetes. Developers can define pipelines via YAML, deploy with Docker, and monitor performance through a web dashboard. Co-Sight empowers users to build advanced vision applications for smart city surveillance, intelligent transportation, and industrial quality inspection, reducing development time and operational complexity.
  • CrewAI Agent Generator quickly scaffolds customized AI agents with prebuilt templates, seamless API integration, and deployment tools.
    0
    0
    What is CrewAI Agent Generator?
    CrewAI Agent Generator leverages a command-line interface to let you initialize a new AI agent project with opinionated folder structures, sample prompt templates, tool definitions, and testing stubs. You can configure connections to OpenAI, Azure, or custom LLM endpoints; manage agent memory using vector stores; orchestrate multiple agents in collaborative workflows; view detailed conversation logs; and deploy your agents to Vercel, AWS Lambda, or Docker with built-in scripts. It accelerates development and ensures consistent architecture across AI agent projects.
  • An open-source framework enabling creation and orchestration of multiple AI agents that collaborate on complex tasks via JSON messaging.
    0
    0
    What is Multi AI Agent Systems?
    This framework allows users to design, configure, and deploy multiple AI agents that communicate via JSON messages through a central orchestrator. Each agent can have distinct roles, prompts, and memory modules, and you can plug in any LLM provider by implementing a provider interface. The system supports persistent conversation history, dynamic routing, and modular extensions. Ideal for simulating debates, automating customer support flows, or coordinating multi-step document generation, it runs on Python, with Docker support for containerized deployments.
  • A Python framework for building scalable multi-channel conversational AI agents with context management.
    0
    0
    What is Multiple MCP Server-based AI Agent BOT?
    This framework provides a server-based architecture supporting Multiple-MCP (Multi-Channel Processing) servers to handle concurrent conversations, maintain context across sessions, and integrate external services via plugins. Developers can configure connectors for messaging platforms, define custom function calls, and scale instances using Docker or native hosts. It includes logging, error handling, and a modular pipeline to extend capabilities without altering core code.
  • RAGApp simplifies building retrieval-augmented chatbots by integrating vector databases, LLMs, and toolchains in a low-code framework.
    0
    0
    What is RAGApp?
    RAGApp is designed to simplify the entire RAG pipeline by providing out-of-the-box integrations with popular vector databases (FAISS, Pinecone, Chroma, Qdrant) and large language models (OpenAI, Anthropic, Hugging Face). It includes data ingestion tools to convert documents into embeddings, context-aware retrieval mechanisms for precise knowledge selection, and a built-in chat UI or REST API server for deployment. Developers can easily extend or replace any component—add custom preprocessors, integrate external APIs as tools, or swap LLM providers—while leveraging Docker and CLI tooling for rapid prototyping and production deployment.
  • Taiga is an open-source AI agent framework enabling creation of autonomous LLM agents with plugin extensibility, memory, and tool integration.
    0
    0
    What is Taiga?
    Taiga is a Python-based open-source AI agent framework designed to streamline the creation, orchestration, and deployment of autonomous large language model (LLM) agents. The framework includes a flexible plugin system for integrating custom tools and external APIs, a configurable memory module for managing long-term and short-term conversational context, and a task chaining mechanism to sequence multi-step workflows. Taiga also offers built-in logging, metrics, and error handling for production readiness. Developers can quickly scaffold agents with templates, extend functionality via SDK, and deploy across platforms. By abstracting complex orchestration logic, Taiga enables teams to focus on building intelligent assistants that can research, plan, and execute actions without manual intervention.
  • An open-source Python AI agent framework enabling autonomous LLM-driven task execution with customizable tools and memory.
    0
    0
    What is OCO-Agent?
    OCO-Agent leverages OpenAI-compatible language models to transform plain-language prompts into actionable workflows. It provides a flexible plugin system for integrating external APIs, shell commands, and data-processing routines. The framework maintains conversation history and context in memory, enabling long-running, multi-step tasks. With a CLI interface and Docker support, OCO-Agent accelerates prototyping and deployment of intelligent assistants for operations, analytics, and developer productivity.
  • Pi Web Agent is an open-source web-based AI agent integrating LLMs for conversational tasks and knowledge retrieval.
    0
    0
    What is Pi Web Agent?
    Pi Web Agent is a lightweight, extensible framework for building AI chat agents on the web. It leverages Python FastAPI on the backend and a React frontend to deliver interactive conversations powered by OpenAI, Cohere, or local LLMs. Users can upload documents or connect external databases for semantic search via vector stores. A plugin architecture allows custom tools, function calls, and third-party API integrations locally, it offers full source code access, role-based prompt templates, and configurable memory storage to create customized AI assistants.
Featured