Helicone

Helicone is an LLM observability platform with built-in caching, rate limiting, and cost tracking.

Features

  • Request logging - Log all LLM requests and responses
  • Caching - Cache responses to reduce costs
  • Rate limiting - Control request rates per user
  • Cost tracking - Monitor spending across models
  • User analytics - Track usage by user or session
  • Prompt templates - Manage and version prompts

Configuration

Helicone works as a proxy. Configure agentgateway to route through Helicone:

# yaml-language-server: $schema=https://agentgateway.dev/schema/config
binds:
- port: 3000
  listeners:
  - routes:
    - policies:
        urlRewrite:
          authority:
            full: oai.helicone.ai
        backendTLS: {}
        requestHeaderModifier:
          add:
            Helicone-Auth: "Bearer $HELICONE_API_KEY"
      backends:
      - ai:
          name: openai
          hostOverride: oai.helicone.ai:443
          provider:
            openAI:
              model: gpt-4o-mini
      policies:
        backendAuth:
          key: "$OPENAI_API_KEY"

Benefits with agentgateway

Using agentgateway with Helicone provides:

FeatureagentgatewayHeliconeCombined
Request routingRoute to multiple LLMs via Helicone
CachingHelicone caches responses
Rate limitingLayered rate limiting
Cost trackingBasicDetailed cost analytics
MCP supportMCP with LLM monitoring

Learn more

Agentgateway assistant

Ask me anything about agentgateway configuration, features, or usage.

Note: AI-generated content might contain errors; please verify and test all returned information.

Tip: one topic per conversation gives the best results. Use the + button in the chat header to start a new conversation.

Switching topics? Starting a new conversation improves accuracy.
↑↓ navigate select esc dismiss

What could be improved?

Your feedback helps us improve assistant answers and identify docs gaps we should fix.

Need more help? Join us on Discord: https://discord.gg/y9efgEmppm

Want to use your own agent? Add the Solo MCP server to query our docs directly. Get started here: https://search.solo.io/.