Comprehensive configuración YAML Tools for Every Need

Get access to configuración YAML solutions that address multiple requirements. One-stop resources for streamlined workflows.

configuración YAML

  • Eunomia is a config-driven AI agent framework enabling rapid assembly and deployment of multi-tool conversational agents via YAML.
    0
    0
    What is Eunomia?
    Eunomia leverages a configuration-first approach to orchestrate AI agents. Through YAML, users define agent roles, prompt templates, tool integrations, memory stores, and branching logic. The framework supports synchronous/asynchronous tools, retrieval-augmented generation, and chain-of-thought prompting. An extensible plugin system allows custom tools, memory backends, and logging integrations. Eunomia’s CLI scaffolds projects, validates configs, and runs agents locally or in cloud environments. This enables teams to quickly prototype, iterate on conversational workflows, and maintain agent solutions without heavy custom development.
  • An open-source AI agent framework enabling modular planning, memory management, and tool integration for automated, multi-step workflows.
    0
    0
    What is Pillar?
    Pillar is a comprehensive AI agent framework designed to simplify the development and deployment of intelligent multi-step workflows. It features a modular architecture with planners for task decomposition, memory stores for context retention, and executors that perform actions via external APIs or custom code. Developers can define agent pipelines in YAML or JSON, integrate any LLM provider, and extend functionality through custom plugins. Pillar handles asynchronous execution and context management out of the box, reducing boilerplate code and accelerating time-to-market for AI-driven applications such as chatbots, data analysis assistants, and automated business processes.
  • A lightweight Python library for creating customizable 2D grid environments to train and test reinforcement learning agents.
    0
    0
    What is Simple Playgrounds?
    Simple Playgrounds provides a modular platform for building interactive 2D grid environments where agents can navigate mazes, interact with objects, and complete tasks. Users define environment layouts, object behaviors, and reward functions via simple YAML or Python scripts. The integrated Pygame renderer delivers real-time visualization, while a step-based API ensures seamless integration with reinforcement learning libraries like Stable Baselines3. With support for multi-agent setups, collision detection, and customizable physics parameters, Simple Playgrounds streamlines the prototyping, benchmarking, and educational demonstration of AI algorithms.
  • Spellcaster is an open-source platform for defining, testing, and orchestrating GPT-powered AI agents through templated spells.
    0
    0
    What is Spellcaster?
    Spellcaster provides a structured approach to building AI Agents by using 'spells'—a combination of prompts, logic, and workflows. Developers write YAML configurations to define agents’ roles, inputs, outputs, and orchestration steps. The CLI tool executes spells, routes messages, and integrates seamlessly with OpenAI, Anthropic, and other LLM APIs. Spellcaster tracks execution logs, retains conversation context, and supports custom plugins for pre- and post-processing. Its debugging interface visualizes the sequence of calls and data flows, making it easier to identify prompt failures and performance issues. By abstracting complex orchestration patterns and standardizing prompt templates, Spellcaster reduces development overhead and ensures consistent agent behavior across environments.
  • SpongeCake is a Python framework that streamlines building custom AI agents with Langchain integrations and tool orchestration.
    0
    0
    What is SpongeCake?
    At its core, SpongeCake is a high-level abstraction layer over Langchain designed to accelerate AI agent development. It offers built-in support for registering tools—like web search, database connectors, or custom APIs—managing prompt templates, and persisting conversational memory. With both code-based and YAML-based configurations, teams can declaratively define agent behaviors, chain multi-step workflows, and enable dynamic tool selection. The included CLI facilitates local testing, debugging, and deployment, making SpongeCake ideal for building chatbots, task automators, and domain-specific assistants without repetitive boilerplate.
  • Agent Forge is a CLI framework for scaffolding, orchestrating, and deploying AI agents integrated with LLMs and external tools.
    0
    0
    What is Agent Forge?
    Agent Forge streamlines the entire lifecycle of AI agent development by offering CLI scaffold commands to generate boilerplate code, conversation templates, and configuration settings. Developers can define agent roles, attach LLM providers, and integrate external tools such as vector databases, REST APIs, and custom plugins using YAML or JSON descriptors. The framework enables local execution, interactive testing, and packaging agents as Docker images or serverless functions for easy deployment. Built-in logging, environment profiles, and VCS hooks simplify debugging, collaboration, and CI/CD pipelines. This flexible architecture supports creating chatbots, autonomous research assistants, customer support bots, and automated data processing workflows with minimal setup.
  • Agent Nexus is an open-source framework for building, orchestrating, and testing AI agents via customizable pipelines.
    0
    0
    What is Agent Nexus?
    Agent Nexus offers a modular architecture for designing, configuring, and running interconnected AI agents that collaborate to solve complex tasks. Developers can register agents dynamically, customize behavior through Python modules, and define communication pipelines via simple YAML configurations. The built-in message router ensures reliable inter-agent data flow, while integrated logging and monitoring tools help track performance and debug workflows. With support for popular AI libraries like OpenAI and Hugging Face, Agent Nexus simplifies the integration of diverse models. Whether prototyping research experiments, building automated customer service assistants, or simulating multi-agent environments, Agent Nexus streamlines development and testing of collaborative AI systems, from academic research to commercial deployments.
  • AgentIn is an open-source Python framework for building AI agents with customizable memory, tool integration, and auto-prompting.
    0
    0
    What is AgentIn?
    AgentIn is a Python-based AI agent framework designed to accelerate the development of conversational and task-driven agents. It offers built-in memory modules to persist context, dynamic tool integration to call external APIs or local functions, and a flexible prompt templating system for customized interactions. Multi-agent orchestration enables parallel workflows, while logging and caching improve reliability and auditability. Easily configurable via YAML or Python code, AgentIn supports major LLM providers and can be extended with custom plugins for domain-specific capabilities.
  • Agent-Baba enables developers to create autonomous AI agents with customizable plugins, conversational memory, and automated task workflows.
    0
    0
    What is Agent-Baba?
    Agent-Baba provides a comprehensive toolkit for creating and managing autonomous AI agents tailored to specific tasks. It offers a plugin architecture for extending capabilities, a memory system to retain conversational context, and workflow automation for sequential task execution. Developers can integrate tools like web scrapers, databases, and custom APIs into agents. The framework simplifies configuration through declarative YAML or JSON schemas, supports multi-agent collaboration, and provides monitoring dashboards to track agent performance and logs, enabling iterative improvement and seamless deployment across environments.
  • Agent of Code is an AI-powered coding agent that generates, debugs, and refactors code across multiple languages via OpenAI APIs.
    0
    0
    What is Agent of Code?
    Agent of Code is a versatile AI agent framework enabling developers to offload routine coding tasks to intelligent agents. It leverages large language models to translate natural language prompts into fully functional code, perform automated code reviews, debug existing code, and refactor legacy codebases. Users define agent goals and parameters through YAML or JSON configurations, select plugins for tasks like testing or CI integration, and execute agents via CLI. The framework orchestrates API calls, manages context windows, and assembles modular responses into cohesive code scripts. With an extensible architecture, developers can plug in custom modules, integrate with version control, and tailor the agent pipeline to project workflows.
  • Cognita is an open-source RAG framework that enables building modular AI assistants with document retrieval, vector search, and customizable pipelines.
    0
    0
    What is Cognita?
    Cognita offers a modular architecture for building RAG applications: ingest and index documents, select from OpenAI, TrueFoundry or third-party embeddings, and configure retrieval pipelines via YAML or Python DSL. Its integrated frontend UI lets you test queries, tune retrieval parameters, and visualize vector similarity. Once validated, Cognita provides deployment templates for Kubernetes and serverless environments, enabling you to scale knowledge-driven AI assistants in production with observability and security.
  • Dive is an open-source Python framework for building autonomous AI agents with pluggable tools and workflows.
    0
    0
    What is Dive?
    Dive is a Python-based open-source framework designed for creating and running autonomous AI agents that can perform multi-step tasks with minimal manual intervention. By defining agent profiles in simple YAML configuration files, developers can specify APIs, tools, and memory modules for tasks such as data retrieval, analysis, and pipeline orchestration. Dive manages context, state, and prompt engineering, allowing flexible workflows with built-in error handling and logging. Its pluggable architecture supports a wide range of language models and retrieval systems, making it easy to assemble agents for customer service automation, content generation, and DevOps processes. The framework scales from prototype to production, offering CLI commands and API endpoints to integrate agents seamlessly into existing systems.
  • Open-source Python framework for orchestrating dynamic multi-agent retrieval-augmented generation pipelines with flexible agent collaboration.
    0
    0
    What is Dynamic Multi-Agent RAG Pathway?
    Dynamic Multi-Agent RAG Pathway provides a modular architecture where each agent handles specific tasks—such as document retrieval, vector search, context summarization, or generation—while a central orchestrator dynamically routes inputs and outputs between them. Developers can define custom agents, assemble pipelines via simple configuration files, and leverage built-in logging, monitoring, and plugin support. This framework accelerates development of complex RAG-based solutions, enabling adaptive task decomposition and parallel processing to improve throughput and accuracy.
  • Collection of pre-built AI agent workflows for Ollama LLM, enabling automated summarization, translation, code generation and other tasks.
    0
    1
    What is Ollama Workflows?
    Ollama Workflows is an open-source library of configurable AI agent pipelines built on top of the Ollama LLM framework. It offers dozens of ready-made workflows—like summarization, translation, code review, data extraction, email drafting, and more—that can be chained together in YAML or JSON definitions. Users install Ollama, clone the repository, select or customize a workflow, and run it via CLI. All processing happens locally on your machine, preserving data privacy while allowing you to iterate quickly and maintain consistent output across projects.
  • Julep AI creates scalable, serverless AI workflows for data science teams.
    0
    0
    What is Julep AI?
    Julep AI is an open-source platform designed to help data science teams quickly build, iterate on, and deploy multi-step AI workflows. With Julep, you can create scalable, durable, and long-running AI pipelines using agents, tasks, and tools. The platform's YAML-based configuration simplifies complex AI processes and ensures production-ready workflows. It supports rapid prototyping, modular design, and seamless integration with existing systems, making it ideal for handling millions of concurrent users while providing full visibility into AI operations.
  • Framework to align large language model outputs with an organization's culture and values using customizable guidelines.
    0
    0
    What is LLM-Culture?
    LLM-Culture provides a structured approach to embed organizational culture into large language model interactions. You start by defining your brand’s values and style rules in a simple configuration file. The framework then offers a library of prompt templates designed to enforce these guidelines. After generating outputs, the built-in evaluation toolkit measures alignment against your cultural criteria and highlights any inconsistencies. Finally, you deploy the framework alongside your LLM pipeline—whether via API or on-premise—so that each response consistently adheres to your company’s tone, ethics, and brand personality.
  • A Python-based framework orchestrating dynamic AI agent interactions with customizable roles, message passing, and task coordination.
    0
    0
    What is Multi-Agent-AI-Dynamic-Interaction?
    Multi-Agent-AI-Dynamic-Interaction offers a flexible environment to design, configure, and run systems composed of multiple autonomous AI agents. Each agent can be assigned specific roles, objectives, and communication protocols. The framework manages message passing, conversation context, and sequential or parallel interactions. It supports integration with OpenAI GPT, other LLM APIs, and custom modules. Users define scenarios via YAML or Python scripts, specifying agent details, workflow steps, and stopping criteria. The system logs all interactions for debugging and analysis, allowing fine-grained control over agent behaviors for experiments in collaboration, negotiation, decision-making, and complex problem-solving.
  • Nexus Agents orchestrates LLM-powered agents with dynamic tool integration, enabling automated workflow management and task coordination.
    0
    0
    What is Nexus Agents?
    Nexus Agents is a modular framework for constructing AI-driven multi-agent systems with large language models at their core. Developers can define custom agents, integrate external tools, and orchestrate workflows through declarative YAML or Python configurations. It supports dynamic task routing, memory management, and inter-agent communication, ensuring scalable and reliable automation. With built-in logging, error handling, and CLI support, Nexus Agents streamlines building complex pipelines spanning data retrieval, analysis, content generation, and customer interactions. Its architecture allows easy extension with custom tools or LLM providers, empowering teams to automate business processes, research tasks, and operational workflows in a consistent and maintainable manner.
  • A Python framework for easily defining and executing AI agent workflows declaratively using YAML-like specifications.
    0
    0
    What is Noema Declarative AI?
    Noema Declarative AI allows developers and researchers to specify AI agents and their workflows in a high-level, declarative manner. By writing YAML or JSON configuration files, you define agents, prompts, tools, and memory modules. The Noema runtime then parses these definitions, loads language models, executes each step of your pipeline, handles state and context, and returns structured results. This approach reduces boilerplate, improves reproducibility, and separates logic from execution, making it ideal for prototyping chatbots, automation scripts, and research experiments.
  • AgentSmith is an open-source framework orchestrating autonomous multi-agent workflows using LLM-based assistants.
    0
    0
    What is AgentSmith?
    AgentSmith is a modular agent orchestration framework built in Python that enables developers to define, configure, and run multiple AI agents collaboratively. Each agent can be assigned specialized roles—such as researcher, planner, coder, or reviewer—and communicate via an internal message bus. AgentSmith supports memory management through vector stores like FAISS or Pinecone, task decomposition into subtasks, and automated supervision to ensure goal completion. Agents and pipelines are configured via human-readable YAML files, and the framework integrates seamlessly with OpenAI APIs and custom LLMs. It includes built-in logging, monitoring, and error handling, making it ideal for automating software development workflows, data analysis, and decision support systems.
Featured