Comprehensive YAML configuration Tools for Every Need

Get access to YAML configuration solutions that address multiple requirements. One-stop resources for streamlined workflows.

YAML configuration

  • Eunomia is a config-driven AI agent framework enabling rapid assembly and deployment of multi-tool conversational agents via YAML.
    0
    0
    What is Eunomia?
    Eunomia leverages a configuration-first approach to orchestrate AI agents. Through YAML, users define agent roles, prompt templates, tool integrations, memory stores, and branching logic. The framework supports synchronous/asynchronous tools, retrieval-augmented generation, and chain-of-thought prompting. An extensible plugin system allows custom tools, memory backends, and logging integrations. Eunomia’s CLI scaffolds projects, validates configs, and runs agents locally or in cloud environments. This enables teams to quickly prototype, iterate on conversational workflows, and maintain agent solutions without heavy custom development.
  • OpenMAS is an open-source multi-agent simulation platform providing customizable agent behaviors, dynamic environments, and decentralized communication protocols.
    0
    0
    What is OpenMAS?
    OpenMAS is designed to facilitate the development and evaluation of decentralized AI agents and multi-agent coordination strategies. It features a modular architecture that allows users to define custom agent behaviors, dynamic environment models, and inter-agent messaging protocols. The framework supports physics-based simulation, event-driven execution, and plugin integration for AI algorithms. Users can configure scenarios via YAML or Python, visualize agent interactions, and collect performance metrics through built-in analytics tools. OpenMAS streamlines prototyping in research areas such as swarm intelligence, cooperative robotics, and distributed decision-making.
  • An open-source AI agent framework enabling modular planning, memory management, and tool integration for automated, multi-step workflows.
    0
    0
    What is Pillar?
    Pillar is a comprehensive AI agent framework designed to simplify the development and deployment of intelligent multi-step workflows. It features a modular architecture with planners for task decomposition, memory stores for context retention, and executors that perform actions via external APIs or custom code. Developers can define agent pipelines in YAML or JSON, integrate any LLM provider, and extend functionality through custom plugins. Pillar handles asynchronous execution and context management out of the box, reducing boilerplate code and accelerating time-to-market for AI-driven applications such as chatbots, data analysis assistants, and automated business processes.
  • A lightweight Python library for creating customizable 2D grid environments to train and test reinforcement learning agents.
    0
    0
    What is Simple Playgrounds?
    Simple Playgrounds provides a modular platform for building interactive 2D grid environments where agents can navigate mazes, interact with objects, and complete tasks. Users define environment layouts, object behaviors, and reward functions via simple YAML or Python scripts. The integrated Pygame renderer delivers real-time visualization, while a step-based API ensures seamless integration with reinforcement learning libraries like Stable Baselines3. With support for multi-agent setups, collision detection, and customizable physics parameters, Simple Playgrounds streamlines the prototyping, benchmarking, and educational demonstration of AI algorithms.
  • Spellcaster is an open-source platform for defining, testing, and orchestrating GPT-powered AI agents through templated spells.
    0
    0
    What is Spellcaster?
    Spellcaster provides a structured approach to building AI Agents by using 'spells'—a combination of prompts, logic, and workflows. Developers write YAML configurations to define agents’ roles, inputs, outputs, and orchestration steps. The CLI tool executes spells, routes messages, and integrates seamlessly with OpenAI, Anthropic, and other LLM APIs. Spellcaster tracks execution logs, retains conversation context, and supports custom plugins for pre- and post-processing. Its debugging interface visualizes the sequence of calls and data flows, making it easier to identify prompt failures and performance issues. By abstracting complex orchestration patterns and standardizing prompt templates, Spellcaster reduces development overhead and ensures consistent agent behavior across environments.
  • SpongeCake is a Python framework that streamlines building custom AI agents with Langchain integrations and tool orchestration.
    0
    0
    What is SpongeCake?
    At its core, SpongeCake is a high-level abstraction layer over Langchain designed to accelerate AI agent development. It offers built-in support for registering tools—like web search, database connectors, or custom APIs—managing prompt templates, and persisting conversational memory. With both code-based and YAML-based configurations, teams can declaratively define agent behaviors, chain multi-step workflows, and enable dynamic tool selection. The included CLI facilitates local testing, debugging, and deployment, making SpongeCake ideal for building chatbots, task automators, and domain-specific assistants without repetitive boilerplate.
  • An AI agent automating test-driven development: it generates tests, implementation code, and runs iterations with GPT models.
    0
    0
    What is TDD-GPT-Agent?
    TDD-GPT-Agent integrates OpenAI’s GPT-4 or GPT-3.5 models in a Python-based CLI to drive a fully automated test-driven development cycle. Given a developer’s function specification, it generates pytest test files, runs tests locally, analyzes failures, and produces implementation code to satisfy assertions. It repeats the cycle until all tests pass. Configurable via a YAML file, the agent supports prompt customization, session logging, Git integration, and can be embedded in CI/CD pipelines for continuous quality assurance. This AI-driven workflow accelerates development, improves coverage, and enforces reliable code.
  • Agent Forge is a CLI framework for scaffolding, orchestrating, and deploying AI agents integrated with LLMs and external tools.
    0
    0
    What is Agent Forge?
    Agent Forge streamlines the entire lifecycle of AI agent development by offering CLI scaffold commands to generate boilerplate code, conversation templates, and configuration settings. Developers can define agent roles, attach LLM providers, and integrate external tools such as vector databases, REST APIs, and custom plugins using YAML or JSON descriptors. The framework enables local execution, interactive testing, and packaging agents as Docker images or serverless functions for easy deployment. Built-in logging, environment profiles, and VCS hooks simplify debugging, collaboration, and CI/CD pipelines. This flexible architecture supports creating chatbots, autonomous research assistants, customer support bots, and automated data processing workflows with minimal setup.
  • Agent Nexus is an open-source framework for building, orchestrating, and testing AI agents via customizable pipelines.
    0
    0
    What is Agent Nexus?
    Agent Nexus offers a modular architecture for designing, configuring, and running interconnected AI agents that collaborate to solve complex tasks. Developers can register agents dynamically, customize behavior through Python modules, and define communication pipelines via simple YAML configurations. The built-in message router ensures reliable inter-agent data flow, while integrated logging and monitoring tools help track performance and debug workflows. With support for popular AI libraries like OpenAI and Hugging Face, Agent Nexus simplifies the integration of diverse models. Whether prototyping research experiments, building automated customer service assistants, or simulating multi-agent environments, Agent Nexus streamlines development and testing of collaborative AI systems, from academic research to commercial deployments.
  • AgentIn is an open-source Python framework for building AI agents with customizable memory, tool integration, and auto-prompting.
    0
    0
    What is AgentIn?
    AgentIn is a Python-based AI agent framework designed to accelerate the development of conversational and task-driven agents. It offers built-in memory modules to persist context, dynamic tool integration to call external APIs or local functions, and a flexible prompt templating system for customized interactions. Multi-agent orchestration enables parallel workflows, while logging and caching improve reliability and auditability. Easily configurable via YAML or Python code, AgentIn supports major LLM providers and can be extended with custom plugins for domain-specific capabilities.
  • Agent-Baba enables developers to create autonomous AI agents with customizable plugins, conversational memory, and automated task workflows.
    0
    0
    What is Agent-Baba?
    Agent-Baba provides a comprehensive toolkit for creating and managing autonomous AI agents tailored to specific tasks. It offers a plugin architecture for extending capabilities, a memory system to retain conversational context, and workflow automation for sequential task execution. Developers can integrate tools like web scrapers, databases, and custom APIs into agents. The framework simplifies configuration through declarative YAML or JSON schemas, supports multi-agent collaboration, and provides monitoring dashboards to track agent performance and logs, enabling iterative improvement and seamless deployment across environments.
  • Agent of Code is an AI-powered coding agent that generates, debugs, and refactors code across multiple languages via OpenAI APIs.
    0
    0
    What is Agent of Code?
    Agent of Code is a versatile AI agent framework enabling developers to offload routine coding tasks to intelligent agents. It leverages large language models to translate natural language prompts into fully functional code, perform automated code reviews, debug existing code, and refactor legacy codebases. Users define agent goals and parameters through YAML or JSON configurations, select plugins for tasks like testing or CI integration, and execute agents via CLI. The framework orchestrates API calls, manages context windows, and assembles modular responses into cohesive code scripts. With an extensible architecture, developers can plug in custom modules, integrate with version control, and tailor the agent pipeline to project workflows.
  • Aladin is an open-source autonomous LLM agent enabling scripted workflows, memory-enabled decision-making, and plugin-based task orchestration.
    0
    0
    What is Aladin?
    Aladin provides a modular architecture that allows developers to define autonomous agents powered by large language models (LLMs). Each agent can load memory backends (e.g., SQLite, in-memory), utilize dynamic prompt templates, and integrate custom plugins for external API calls or local command execution. It features a task planner that breaks high-level goals into sequenced actions, executing them in order and iterating based on LLM feedback. Configuration is managed through YAML files and environment variables, making it adaptable to various use cases. Users can deploy Aladin via Docker Compose or pip installation. The CLI and FastAPI-based HTTP endpoints let users trigger agents, monitor execution, and inspect memory states, facilitating integration with CI/CD pipelines, chat interfaces, or custom dashboards.
  • Cognita is an open-source RAG framework that enables building modular AI assistants with document retrieval, vector search, and customizable pipelines.
    0
    0
    What is Cognita?
    Cognita offers a modular architecture for building RAG applications: ingest and index documents, select from OpenAI, TrueFoundry or third-party embeddings, and configure retrieval pipelines via YAML or Python DSL. Its integrated frontend UI lets you test queries, tune retrieval parameters, and visualize vector similarity. Once validated, Cognita provides deployment templates for Kubernetes and serverless environments, enabling you to scale knowledge-driven AI assistants in production with observability and security.
  • Dive is an open-source Python framework for building autonomous AI agents with pluggable tools and workflows.
    0
    0
    What is Dive?
    Dive is a Python-based open-source framework designed for creating and running autonomous AI agents that can perform multi-step tasks with minimal manual intervention. By defining agent profiles in simple YAML configuration files, developers can specify APIs, tools, and memory modules for tasks such as data retrieval, analysis, and pipeline orchestration. Dive manages context, state, and prompt engineering, allowing flexible workflows with built-in error handling and logging. Its pluggable architecture supports a wide range of language models and retrieval systems, making it easy to assemble agents for customer service automation, content generation, and DevOps processes. The framework scales from prototype to production, offering CLI commands and API endpoints to integrate agents seamlessly into existing systems.
  • Open-source Python framework for orchestrating dynamic multi-agent retrieval-augmented generation pipelines with flexible agent collaboration.
    0
    0
    What is Dynamic Multi-Agent RAG Pathway?
    Dynamic Multi-Agent RAG Pathway provides a modular architecture where each agent handles specific tasks—such as document retrieval, vector search, context summarization, or generation—while a central orchestrator dynamically routes inputs and outputs between them. Developers can define custom agents, assemble pipelines via simple configuration files, and leverage built-in logging, monitoring, and plugin support. This framework accelerates development of complex RAG-based solutions, enabling adaptive task decomposition and parallel processing to improve throughput and accuracy.
  • Collection of pre-built AI agent workflows for Ollama LLM, enabling automated summarization, translation, code generation and other tasks.
    0
    1
    What is Ollama Workflows?
    Ollama Workflows is an open-source library of configurable AI agent pipelines built on top of the Ollama LLM framework. It offers dozens of ready-made workflows—like summarization, translation, code review, data extraction, email drafting, and more—that can be chained together in YAML or JSON definitions. Users install Ollama, clone the repository, select or customize a workflow, and run it via CLI. All processing happens locally on your machine, preserving data privacy while allowing you to iterate quickly and maintain consistent output across projects.
  • Julep AI creates scalable, serverless AI workflows for data science teams.
    0
    0
    What is Julep AI?
    Julep AI is an open-source platform designed to help data science teams quickly build, iterate on, and deploy multi-step AI workflows. With Julep, you can create scalable, durable, and long-running AI pipelines using agents, tasks, and tools. The platform's YAML-based configuration simplifies complex AI processes and ensures production-ready workflows. It supports rapid prototyping, modular design, and seamless integration with existing systems, making it ideal for handling millions of concurrent users while providing full visibility into AI operations.
  • A Python framework for building modular AI agents with memory, planning, and tool integration.
    0
    0
    What is Linguistic Agent System?
    Linguistic Agent System is an open-source Python framework designed for constructing intelligent agents that leverage language models to plan and execute tasks. It includes components for memory management, tool registry, planner, and executor, allowing agents to maintain context, call external APIs, perform web searches, and automate workflows. Configurable via YAML, it supports multiple LLM providers, enabling rapid prototyping of chatbots, content summarizers, and autonomous assistants. Developers can extend functionality by creating custom tools and memory backends, deploying agents locally or on servers.
  • Framework to align large language model outputs with an organization's culture and values using customizable guidelines.
    0
    0
    What is LLM-Culture?
    LLM-Culture provides a structured approach to embed organizational culture into large language model interactions. You start by defining your brand’s values and style rules in a simple configuration file. The framework then offers a library of prompt templates designed to enforce these guidelines. After generating outputs, the built-in evaluation toolkit measures alignment against your cultural criteria and highlights any inconsistencies. Finally, you deploy the framework alongside your LLM pipeline—whether via API or on-premise—so that each response consistently adheres to your company’s tone, ethics, and brand personality.
Featured