Comprehensive コミュニティ主導の開発 Tools for Every Need

Get access to コミュニティ主導の開発 solutions that address multiple requirements. One-stop resources for streamlined workflows.

コミュニティ主導の開発

  • A reinforcement learning framework for training collision-free multi-robot navigation policies in simulated environments.
    0
    0
    What is NavGround Learning?
    NavGround Learning provides a comprehensive toolkit for developing and benchmarking reinforcement learning agents in navigation tasks. It supports multi-agent simulation, collision modeling, and customizable sensors and actuators. Users can select from predefined policy templates or implement custom architectures, train with state-of-the-art RL algorithms, and visualize performance metrics. Its integration with OpenAI Gym and Stable Baselines3 simplifies experiment management, while built-in logging and visualization tools allow in-depth analysis of agent behavior and training dynamics.
  • Swarms is an open-source platform to build, orchestrate, and deploy collaborative multi-agent AI systems with customizable workflows.
    0
    0
    What is Swarms?
    Swarms operates as a Python-first framework and web-based interface, empowering users to configure individual agents with specific roles, memory management, and custom prompts. Users define agent interactions through a visual flow builder or YAML configuration, orchestrating complex decision trees, debates, and collaborative tasks. The platform supports plugin integration for data querying, knowledge base access, and third-party API calls. Upon deployment, Swarms provides real-time monitoring of agent activities, performance metrics, and logs. It scales horizontally using container orchestration tools, enabling large-scale AI simulations, robotic control architectures, or intelligent workflow automations. The open-source architecture ensures extensibility, community-driven enhancements, and self-hosting options for full data control.
  • An extensible Python-based AI Agent for multi-turn conversation, memory, custom prompts, and Grok integration.
    0
    0
    What is Chatbot-Grok?
    Chatbot-Grok provides a modular AI Agent framework written in Python, designed to simplify development of conversational bots. It supports multi-turn dialogue management, retains chat memory across sessions, and allows users to define custom prompt templates. The architecture is extensible, letting developers integrate various LLMs including Grok, and connect to platforms such as Telegram or Slack. With clear code organization and plugin-friendly structure, Chatbot-Grok accelerates prototyping and deployment of production-ready chat assistants.
  • Doraemon-Agent is an open-source Python framework that orchestrates multi-step AI agents with plugin integration and memory management.
    0
    0
    What is Doraemon-Agent?
    Doraemon-Agent is an open-source Python platform and framework designed for developers to build sophisticated AI agents. It allows you to integrate custom plugins and external tools, maintain long-term memory across sessions, and execute chain-of-thought planning with multiple steps. Developers can configure agent roles, manage context, log interactions, and extend functionality through a plugin architecture. It simplifies the creation of autonomous assistants for tasks like data analysis, research support, or customer service automation.
  • Lila is an open-source AI agent framework that orchestrates LLMs, manages memory, integrates tools, and customizes workflows.
    0
    0
    What is Lila?
    Lila delivers a complete AI agent framework tailored for multi-step reasoning and autonomous task execution. Developers can define custom tools (APIs, databases, webhooks) and configure Lila to call them dynamically during runtime. It offers memory modules to store conversation history and facts, a planning component to sequence sub-tasks, and chain-of-thought prompting for transparent decision paths. Its plugin system allows seamless extension with new capabilities, while built-in monitoring tracks agent actions and outputs. Lila’s modular design makes it easy to integrate into existing Python projects or deploy as a hosted service for real-time agent workflows.
  • A Python sample demonstrating LLM-based AI agents with integrated tools like search, code execution, and QA.
    0
    0
    What is LLM Agents Example?
    LLM Agents Example provides a hands-on codebase for building AI agents in Python. It demonstrates registering custom tools (web search, math solver via WolframAlpha, CSV analyzer, Python REPL), creating chat and retrieval-based agents, and connecting to vector stores for document question answering. The repo illustrates patterns for maintaining conversational memory, dispatching tool calls dynamically, and chaining multiple LLM prompts to solve complex tasks. Users learn how to integrate third-party APIs, structure agent workflows, and extend the framework with new capabilities—serving as a practical guide for developer experimentation and prototyping.
  • Overeasy is an open-source AI agent framework enabling autonomous LLM-powered assistants with memory, tools integration, and multi-agent orchestration.
    0
    0
    What is Overeasy?
    Overeasy is a Python-based open-source framework for orchestrating LLM-driven AI agents across various domains. It provides a modular architecture to define agents, configure memory stores, and integrate external tools such as APIs, knowledge bases, and databases. Developers can connect to OpenAI, Azure, or self-hosted LLM endpoints and design dynamic workflows involving single or multiple agents. Overeasy’s orchestration engine handles task delegation, decision making, and fallback strategies, enabling robust digital workers for research, customer support, data analysis, scheduling, and more. Comprehensive documentation and example projects accelerate deployment on Linux, macOS, and Windows.
  • SmartRAG is an open-source Python framework for building RAG pipelines that enable LLM-driven Q&A over custom document collections.
    0
    0
    What is SmartRAG?
    SmartRAG is a modular Python library designed for retrieval-augmented generation (RAG) workflows with large language models. It combines document ingestion, vector indexing, and state-of-the-art LLM APIs to deliver accurate, context-rich responses. Users can import PDFs, text files, or web pages, index them using popular vector stores like FAISS or Chroma, and define custom prompt templates. SmartRAG orchestrates the retrieval, prompt assembly, and LLM inference, returning coherent answers grounded in source documents. By abstracting the complexity of RAG pipelines, it accelerates development of knowledge base Q&A systems, chatbots, and research assistants. Developers can extend connectors, swap LLM providers, and fine-tune retrieval strategies to fit specific knowledge domains.
  • A modular Python framework to build autonomous AI agents with LLM-driven planning, memory management, and tool integration.
    0
    0
    What is AI-Agents?
    AI-Agents provides a flexible agent architecture that orchestrates language model planners, persistent memory modules, and pluggable toolkits. Developers define tools for HTTP requests, file operations, and custom logic, then configure an LLM planner to decide which tool to invoke. Memory stores context and conversation history. The framework handles asynchronous execution, error recovery, and logging, enabling rapid prototyping of intelligent assistants, data analyzers, or automation bots without reinventing core orchestration logic.
  • Agent API by HackerGCLASS: a Python RESTful framework for deploying AI agents with custom tools, memory, and workflows.
    0
    0
    What is HackerGCLASS Agent API?
    HackerGCLASS Agent API is an open-source Python framework that exposes RESTful endpoints to run AI agents. Developers can define custom tool integrations, configure prompt templates, and maintain agent state and memory across sessions. The framework supports orchestrating multiple agents in parallel, handling complex conversational flows, and integrating external services. It simplifies deployment via Uvicorn or other ASGI servers and offers extensibility with plugin modules, enabling rapid creation of domain-specific AI agents for diverse use cases.
  • Arenas is an open-source framework enabling developers to prototype, orchestrate, and deploy customizable LLM-powered agents with tool integrations.
    0
    0
    What is Arenas?
    Arenas is designed to streamline the development lifecycle of LLM-powered agents. Developers can define agent personas, integrate external APIs and tools as plugins, and compose multi-step workflows using a flexible DSL. The framework manages conversation memory, error handling, and logging, enabling robust RAG pipelines and multi-agent collaboration. With a command-line interface and REST API, teams can prototype agents locally and deploy them as microservices or containerized applications. Arenas supports popular LLM providers, offers monitoring dashboards, and includes built-in templates for common use cases. This flexible architecture reduces boilerplate code and accelerates time-to-market for AI-driven solutions across domains like customer engagement, research, and data processing.
  • An open-source React-based chat UI framework enabling real-time LLM integration with customizable themes, streaming responses, and multi-agent support.
    0
    0
    What is Chipper?
    Chipper is a fully open-source React component library designed to simplify the creation of conversational interfaces powered by large language models. It offers real-time streaming of AI responses, built-in context and history management, support for multiple agents in a single chat, file attachments, and theme customization. Developers can integrate any LLM backend via simple props, extend with plugins, and style using CSS-in-JS for seamless branding and responsive layouts.
  • Co-Sight is an open-source AI framework offering real-time video analytics for object detection, tracking, and distributed inference.
    0
    0
    What is Co-Sight?
    Co-Sight is an open-source AI framework that simplifies development and deployment of real-time video analytics solutions. It provides modules for video data ingestion, preprocessing, model training, and distributed inference on edge and cloud. With built-in support for object detection, classification, tracking, and pipeline orchestration, Co-Sight ensures low-latency processing and high throughput. Its modular design integrates with popular deep learning libraries and scales seamlessly using Kubernetes. Developers can define pipelines via YAML, deploy with Docker, and monitor performance through a web dashboard. Co-Sight empowers users to build advanced vision applications for smart city surveillance, intelligent transportation, and industrial quality inspection, reducing development time and operational complexity.
  • JaCaMo is a multi-agent system platform integrating Jason, CArtAgO, and Moise for scalable, modular agent-based programming.
    0
    0
    What is JaCaMo?
    JaCaMo provides a unified environment for designing and running multi-agent systems (MAS) by integrating three core components: the Jason agent programming language for BDI-based agents, CArtAgO for artifact-based environmental modeling, and Moise for specifying organizational structures and roles. Developers can write agent plans, define artifacts with operations, and organize groups of agents under normative frameworks. The platform includes tooling for simulation, debugging, and visualization of MAS interactions. With support for distributed execution, artifact repositories, and flexible messaging, JaCaMo enables rapid prototyping and research in areas like swarm intelligence, collaborative robotics, and distributed decision-making. Its modular design ensures scalability and extensibility across academic and industrial projects.
  • ExampleAgent is a template framework for creating customizable AI agents that automate tasks via OpenAI API.
    0
    0
    What is ExampleAgent?
    ExampleAgent is a developer-focused toolkit designed to accelerate the creation of AI-driven assistants. It integrates directly with OpenAI’s GPT models to handle natural language understanding and generation, and offers a pluggable system for adding custom tools or APIs. The framework manages conversation context, memory, and error handling, enabling agents to perform information retrieval, task automation, and decision-making workflows. With clear code templates, documentation, and examples, teams can rapidly prototype domain-specific agents for chatbots, data extraction, scheduling, and more.
  • FMAS is a flexible multi-agent system framework enabling developers to define, simulate, and monitor autonomous AI agents with custom behaviors and messaging.
    0
    0
    What is FMAS?
    FMAS (Flexible Multi-Agent System) is an open-source Python library for building, running, and visualizing multi-agent simulations. You can define agents with custom decision logic, configure an environment model, set up messaging channels for communication, and execute scalable simulation runs. FMAS provides hooks for monitoring agent state, debugging interactions, and exporting results. Its modular architecture supports plugins for visualization, metrics collection, and integration with external data sources, making it ideal for research, education, and real-world prototypes of autonomous systems.
  • A modular open-source framework integrating large language models with messaging platforms for custom AI agents.
    0
    0
    What is LLM to MCP Integration Engine?
    LLM to MCP Integration Engine is an open-source framework designed to integrate large language models (LLMs) with various messaging communication platforms (MCPs). It provides adapters for LLM APIs like OpenAI and Anthropic, and connectors for chat platforms such as Slack, Discord, and Telegram. The engine manages session state, enriches context, and routes messages bi-directionally. Its plugin-based architecture enables developers to extend support to new providers and customize business logic, accelerating the deployment of AI agents in production environments.
  • Mina is a minimal Python-based AI agent framework enabling custom tool integration, memory management, LLM orchestration, and task automation.
    0
    0
    What is Mina?
    Mina provides a lightweight yet powerful foundation for constructing AI agents in Python. You can define custom tools (such as web scrapers, calculators, or database connectors), attach memory buffers to maintain conversational context, and orchestrate sequences of calls to language models for multi-step reasoning. Built on top of common LLM APIs, Mina handles asynchronous execution, error handling, and logging out of the box. Its modular design makes it easy to extend with new capabilities, while the CLI interface enables quick prototyping and deployment of agent-driven applications.
Featured