Skip to main content

What is Fallom?

Fallom is an LLM observability platform that gives you complete visibility into your AI applications. Understand what’s happening in production, debug issues fast, and optimize costs.

🔍 Observability

Every LLM call traced with tokens, latency, costs, prompts, and completions

📊 Session Analytics

Group calls by user or conversation. See the full picture.

🧪 A/B Testing

Test models and prompts in production with traffic splitting

📋 Evals

Run evaluations on your LLM outputs at scale

Get started in 2 minutes

Add 3 lines of code and start seeing your LLM calls.

How It Works

import fallom from "@fallom/trace";

await fallom.init({ apiKey: process.env.FALLOM_API_KEY });
const session = fallom.session({ configKey: "my-app", sessionId: "user-123" });

// Wrap your LLM client - that's it
const openai = session.wrapOpenAI(new OpenAI());
Every call through openai is now automatically traced to your dashboard.

Supported Providers

Works with any OpenAI-compatible API:
  • OpenAI - GPT-4o, GPT-4, GPT-3.5
  • Anthropic - Claude 3.5, Claude 3
  • Google - Gemini Pro, Gemini Flash
  • OpenRouter - Access 100+ models
  • Vercel AI SDK - Full framework support

Python

pip install fallom

TypeScript

npm install @fallom/trace