Skip to content

LLMling-Agent

PyPI License Package status Daily downloads Weekly downloads Monthly downloads Distribution format Wheel availability Python version Implementation Releases Github Contributors Github Discussions Github Forks Github Issues Github Issues Github Watchers Github Stars Github Repository size Github last commit Github release date Github language count Github commits this week Github commits this month Github commits this year Package status Code style: black PyUp

Read the documentation!

Getting Started

LLMling Agent is a framework for creating and managing LLM-powered agents. It integrates with LLMling's resource system and provides structured interactions with language models.

Unique features - Pydantic-AI based. Modern python (3.12+) - True async framework. Easy set-up of complex async agent flows. - Easy consistent APIs - Complete agent defintion via YAML files including extensive JSON schema to help with creating configurations. - Leveraging the complete pydantic-based type-safe stack (Pydantic- Pydantic-AI, Typer, SQLModel, etc.) - Complete integrated command sytem to control agents from prompt-based interfaces - Agent MCP server support, initialized when entering the async context. - Support for creating prompts for many common python type(s / instances). Your agent understands common datatypes. - Integration of Meta-Model system based on LLMling-models, also configurable via YAML. - Structured responses. With pydantic-AI at its core, the Agents allow injecting dependencies as well as defining a return type while keeping type safety. - Response type definition via YAML. Structured response Agents can be defined in the agent config. - Capabilites system allowing runtime modifications and "special" commands (on-the-fly agent generation, history lookups) - Complete database logging of Agent interactions - pytest-inspired way to create agents from YAML in a type-safe manner. "Auto-populated signatures." - Tasks, tools, and what else you can expect from an Agent framework. - Easy human-in-the-loop interactions on multiple levels (complete "providers" or model-based, see llmling-models) - A CLI application with extensive slash command support to build agent flows interactively. Set up message connections via commands.

Coming soon - Built-in event system for reactive agent behaviors (file changes, webhooks, timed events) - Real-time-monitoring via Textual app in truly async manner. Talk to your agents while they are working and monitor the progress!

Quick Start

The fastest way to start chatting with an AI:

# Start an ephemeral chat session (requires uv)
uvx llmling-agent quickstart openai:gpt-4o-mini

This creates a temporary agent ready for chat - no configuration needed! LLMling-Agent is Pydantic-ai based, so all pydantic-ai models can be used. The according API keys need to be set as environment variables.

For persistent agents, you can use:

# Create a basic agent configuration
llmling-agent init agents.yml

# Or use the interactive wizard (EXPERIMENTAL)
llmling-agent init agents.yml --interactive

This creates a basic agent configuration file that you can customize. The interactive mode will guide you through setting up your agents.

Basic Usage

The simplest way to use LLMling Agent is through its command-line interface:

# Start an interactive chat with an agent
llmling-agent chat my-agent

# Run an agent with a specific prompt
llmling-agent run my-agent "What is the current system status?"

Features

Dynamic Environment

LLMling Agent allows the AI to modify its own environment (when permitted): - Register new tools on the fly - Load and analyze resources - Install Python packages - Create new tools from code

These capabilities can be controlled via roles and permissions to ensure safe operation.

Interactive Chat Sessions

The chat interface provides rich features:

# Start a chat session
llmling-agent chat my-agent

# Available during chat:
/list-tools              # See available tools
/register-tool os.getcwd # Add new tools on the fly
/list-resources         # View accessible resources
/show-resource config   # Examine resource content
/enable-tool tool_name  # Enable/disable tools
/set-model gpt-4       # Switch models mid-conversation
# and many more!

Safe and Configurable

  • Fine-grained capability control (resource access, tool registration, etc.)
  • Role-based permissions (overseer, specialist, assistant)
  • Tool confirmation for sensitive operations
  • Command history and usage statistics

First Agent Configuration

Agents are defined in YAML configuration files. The environment (tools and resources) can be configured either inline or in a separate file: (see LLMling documentation for YAML details)

Option 1: Separate Environment File

# agents.yml - Agent configuration
agents:
  system_checker:
    model: openai:gpt-4o-mini
    environment: env_system.yml  # Reference to environment file
    system_prompts:
      - "You help users check their system status."

# env_system.yml - Environment configuration (LLMling format)
tools:
  get_system_info:
    import_path: platform.platform
    description: "Get system platform information"
  get_memory:
    import_path: psutil.virtual_memory
    description: "Get memory usage information"

Option 2: Inline Environment

# agents.yml - Complete configuration
agents:
  system_checker:
    model: openai:gpt-4o-mini
    environment:  # Inline environment configuration
      type: inline
      tools:
        get_system_info:
          import_path: platform.platform
          description: "Get system platform information"
        get_memory:
          import_path: psutil.virtual_memory
          description: "Get memory usage information"
    system_prompts:
      - "You help users check their system status."

Both approaches are equivalent - choose what works best for your use case: - Separate files: Better for reusing environments across agents or when configurations are large - Inline configuration: Simpler for small configurations or self-contained agents

Running Your First Agent

  1. Save both configuration files:
  2. agents.yml - Agent configuration
  3. env_system.yml - Environment configuration

  4. Add the agent configuration to LLMling Agent:

    llmling-agent add my-config agents.yml
    

  5. Start chatting with your agent:

    llmling-agent chat system_checker
    

  6. Or run it programmatically:

    from llmling_agent import Agent
    
    async with Agent.open_agent("agents.yml", "system_checker") as agent:
        result = await agent.run("How much memory is available?")
        print(result.data)
    

Agent Pool: Multi-Agent Coordination

The AgentPool allows multiple agents to work together on tasks. Here's a practical example of parallel file downloading:

# agents.yml
agents:
  file_getter_1:
    name: "File Downloader 1" # Agent name (can be anything unique)
    description: "Downloads files from URLs"
    model: openai:gpt-4o-mini  # Language model to use, takes pydantic-ai model names
    environment: # Environment configuration (can also be external YAML file)
      type: inline
      tools:
        download_file:  # Simple httpx-based download utility
        import_path: llmling_agent_tools.download_file
        description: "Download file from URL to local path"
    system_prompts:
      - |
        You are a download specialist. Just use the download_file tool
        and report its results. No explanations needed.

  file_getter_2:  # Same configuration as file_getter_1
    ... # ... (identical config to file_getter_1, omitting for brevity)

  overseer:
    name: "Download Coordinator"
    description: "Coordinates parallel downloads"
    model: openai:gpt-4o-mini
    system_prompts:
      - |
        You coordinate downloads by delegating to file_getter_1 and file_getter_2.
        Just delegate tasks and report results concisely. No explanations needed.
from llmling_agent.delegation import AgentPool

async def main():
    async with AgentPool.open("agents.yml") as pool:
        # Run downloads in parallel (sequential mode also available)
        team = pool.create_group(["file_getter_1", "file_getter_2"])
        responses = await team.run_parallel("Download https://example.com/file.zip")

        # Or let a coordinator orchestrate
        coordinator = pool.get_agent("coordinator")
        result = await overseer.run(
            "Download https://example.com/file.zip using both getters..."
        )


#### Features

- **Team Tasks**: Run tasks across multiple agents either sequentially or in parallel
- **Agent Cloning**: Create variations of agents with different configurations
- **Resource Sharing**: Agents in a pool can share resources and tools
- **Overseer Pattern**: Use overseer agents to coordinate specialist agents
- **Built-in Collaboration**: Tools for delegation, brainstorming, and debates

#### Configuration

The pool is configured through an agents manifest (YAML):

```yaml
agents:
  agent_1:
    model: openai:gpt-4
    # ... agent-specific config

  agent_2:
    model: openai:gpt-4
    # ... agent-specific config

  overseer:
    model: openai:gpt-4
    # ... overseer config

Each agent can have its own: - Model configuration - Role and capabilities - Tools and resources - System prompts and behavior settings

Message Forwarding

LLMling Agent supports message forwarding between agents, allowing creation of agent chains and networks. When an agent processes a message, it can forward it to other agents for further processing:

# Create two agents
async with Agent.open_agent("agents.yml", "analyzer") as agent_a, \
          Agent.open_agent("agents.yml", "reviewer") as agent_b:

    # Let agent_a pass its results to agent_b
    agent_a.pass_results_to(agent_b)

    # Start the chain - agent_b will process agent_a's output
    await agent_a.run("Analyze this code")
    await agent_b.complete_tasks()  # Wait for agent_b to finish
    agent_a.stop_passing_results_to(agent_b)

Each agent in the chain can: 1. Process the incoming message 2. Access the source agent as a dependency 3. Forward its own response to other agents

This enables simple creation of agent chains.

Conversation History and Analytics

LLMling Agent provides built-in conversation tracking and analysis:

# View recent conversations
llmling-agent history show
llmling-agent history show --period 24h  # Last 24 hours
llmling-agent history show --query "database"  # Search content

# View usage statistics
llmling-agent history stats  # Basic stats
llmling-agent history stats --group-by model  # Model usage
llmling-agent history stats --group-by day    # Daily breakdown

This diagram shows the main components of the LLMling Agent framework:

classDiagram
    %% Core relationships
    AgentsManifest --* AgentConfig : contains
    AgentsManifest --> AgentPool : creates
    AgentPool --* Agent : manages
    FileEnvironment --> Config : loads
    InlineEnvironment --* Config : contains
    Config --> RuntimeConfig : initialized as
    Agent --> RuntimeConfig : uses
    AgentConfig --> FileEnvironment : uses
    AgentConfig --> InlineEnvironment : uses
    Agent --* ToolManager : uses
    Agent --* ConversationManager : uses

    class Config ["[LLMling Core] Config"] {
        Base configuration format defining tools, resources, and settings
        +
        +tools: dict
        +resources: dict
        +prompts: dict
        +global_settings: GlobalSettings
        +from_file()
    }

    class RuntimeConfig ["[LLMling Core] RuntimeConfig"] {
        Runtime state of a config with instantiated components
        +
        +config: Config
        +tools: dict[str, LLMCallableTool]
        +resources: dict[str, Resource]
        +prompts: dict[str, BasePrompt]
        +register_tool()
        +load_resource()
    }

    class AgentsManifest {
        Complete agent configuration manifest defining all available agents
        +
        +responses: dict[str, ResponseDefinition]
        +agents: dict[str, AgentConfig]
        +create_pool()
        +open_agent()
    }

    class AgentConfig {
        Configuration for a single agent including model, environment and capabilities
        +
        +name: str
        +model: str | Model
        +environment: AgentEnvironment
        +capabilities: Capabilities
        +system_prompts: list[str]
        +get_config(): Config
    }

    class FileEnvironment {
        Environment loaded from external YAML file
        +
        +type: "file"
        +uri: str
    }

    class InlineEnvironment {
        Direct environment configuration without external files
        +
        +type: "inline"
        +tools: ...
        +resources: ...
        +prompts: ...
    }

    class AgentPool {
        Manager for multiple initialized agents
        +
        +manifest: AgentsManifest
        +agents: dict[str, Agent]
        +open()
    }

    class Agent {
        Main agent class handling LLM interactions and tool usage
        +
        +runtime: RuntimeConfig
        +tools: ToolManager
        +conversation: ConversationManager
        +run()
        +run_stream()
        +open()
    }

    class ToolManager {
        Manages tool registration, enabling/disabling and access
        +
        +register_tool()
        +enable_tool()
        +disable_tool()
        +get_tools()
        +list_tools()
    }

    class ConversationManager {
        Manages conversation state and system prompts
        +
        +get_history()
        +clear()
        +add_context_from_path()
        +add_context_from_resource()
    }