LLM SPY records traffic between your AI Agent and LLMs, revealing exactly what it does - every prompt, tool call, and decision.
ANTHROPIC_BASE_URL="https://llms.klmn.sh" \ ANTHROPIC_CUSTOM_HEADERS="x-proxy-auth: KEY" \ claude > Build me a LangChain agent that teaches me how to cook sushi


Set up in minutes, monitor everything instantly
Point your AI agent to LLM SPY instead of the LLM provider directly. Just change one environment variable and you're monitoring.
View all LLM traffic in real-time. See prompts, responses, tool calls, token usage, and execution flow.
Share specific LLM conversations with your team via secure, secret links. Perfect for debugging, collaboration, or demonstrating AI behavior.
Monitor what Claude Code actually does by setting ANTHROPIC_BASE_URL
Complete transparency and control over your AI agents
Know exactly what your LLM does - every prompt, response, tool call, and decision. No more black box AI.
Just change one environment variable and you're monitoring. No SDK integration, no code changes required.
Use any LLM through any interface. Your app talks to 'Anthropic' but actually uses GPT-4.
Automatically groups related LLM interactions into coherent sessions, giving you the full conversation context.
LLM SPY intelligently detects and groups related requests into sessions. See the complete flow of multi-turn conversations, not just isolated API calls.
Identifies conversation patterns and links related messages together. No manual tagging or session IDs required.
View entire conversation threads with preserved context. Understand how your agent builds on previous exchanges.
Request 1 • 14:23:01
→ "Create a Python function to parse JSON"
Response 1 • 14:23:02
← "def parse_json(data): ..."
Request 2 • 14:23:15
→ "Add error handling to that function"
Response 2 • 14:23:16
← "try: ... except JSONDecodeError: ..."
Make any LLM speak any language. Your app talks to "Anthropic" but actually uses GPT-4, or vice versa.
LLM SPY acts as a translation layer between your application and any LLM provider. Switch between models without changing a single line of code.
Seamlessly converts between Anthropic, OpenAI, Google, and other LLM formats. Use any model with any interface.
See exactly how requests are translated between formats. Debug and understand the conversion process in real-time.
Everything you need to know about LLM SPY
No cost while hosting is manageable
It's free while it doesn't incur significant hosting or storage costs. We may start charging if it becomes an issue, but will keep it affordable.
Start NowJoin the alpha and get complete visibility into your AI agents
Get Started for Free