Comprehensive Forschungssupport Tools for Every Need

Get access to Forschungssupport solutions that address multiple requirements. One-stop resources for streamlined workflows.

Forschungssupport

  • Aurora coordinates multi-step planning, execution, and tool usage workflows for autonomous generative AI agents powered by LLMs.
    0
    0
    What is Aurora?
    Aurora provides a modular architecture for constructing generative AI agents that can autonomously tackle complex tasks through iterative planning and execution. It consists of a Planner component that breaks down high-level objectives into actionable steps, an Executor that invokes these steps using large language models, and a Tool integration layer for connecting APIs, databases, or custom functions. Aurora also includes memory management for context retention and dynamic re-planning capabilities to adjust to new information. With customizable prompts and plug-and-play modules, developers can rapidly prototype AI agents for tasks like content generation, research, customer support, or process automation, while maintaining full control over the agent’s workflows and decision logic.
    Aurora Core Features
    • LLM-driven planning
    • Executor module for task execution
    • Tool integration layer for APIs and functions
    • Memory management for context retention
    • Dynamic re-planning capabilities
    • Customizable prompt templates
  • GPA-LM is an open-source agent framework that decomposes tasks, manages tools, and orchestrates multi-step language model workflows.
    0
    0
    What is GPA-LM?
    GPA-LM is a Python-based framework designed to simplify the creation and orchestration of AI agents powered by large language models. It features a planner that breaks down high-level instructions into sub-tasks, an executor that manages tool calls and interactions, and a memory module that retains context across sessions. The plugin architecture allows developers to add custom tools, APIs, and decision logic. With multi-agent support, GPA-LM can coordinate roles, distribute tasks, and aggregate results. It integrates seamlessly with popular LLMs like OpenAI GPT and supports deployment on various environments. The framework accelerates the development of autonomous agents for research, automation, and application prototyping.
Featured