Helicone Observability
Helicone is an open-source LLM observability platform that helps you monitor, analyze, and optimize your AI applications. Built-in observability tracks every request automatically, providing comprehensive insights into performance, costs, user behavior, and model usage without requiring additional instrumentation.
Setup
The Helicone provider is available in the @helicone/ai-sdk-provider package. Install it with:
pnpm add @helicone/ai-sdk-provider
Setting up Helicone:
-
Create a Helicone account at helicone.ai
-
Get your API key from the Helicone Dashboard
-
Set your API key as an environment variable:
.envHELICONE_API_KEY=your-helicone-api-key -
Use Helicone in your application:
import { createHelicone } from '@helicone/ai-sdk-provider';import { generateText } from 'ai';const helicone = createHelicone({apiKey: process.env.HELICONE_API_KEY,});// Use the provider with any supported model: https://helicone.ai/modelsconst result = await generateText({model: helicone('claude-4.5-haiku'),prompt: 'Hello world',});console.log(result.text);
That's it! Your requests are now being logged and monitored through Helicone with automatic observability.
→ Learn more about Helicone AI Gateway
Key Observability Features
Helicone provides comprehensive observability for your AI applications with zero additional instrumentation:
Automatic Request Tracking
- Every request is logged automatically with full request/response data
- Track latency, tokens, costs, and model performance in real-time
- No OpenTelemetry setup or additional configuration required
Analytics Dashboard
- View metrics across all your AI requests: costs, latency, token usage, and error rates
- Filter by user, session, model, or custom properties
- Identify performance bottlenecks and optimize model selection
User & Session Analytics
- Track individual user behavior and usage patterns
- Monitor conversation flows with session tracking
- Analyze user engagement and feature adoption
Cost Monitoring
- Real-time cost tracking per request, user, feature, or model
- Budget alerts and cost optimization insights
- Compare costs across different models and providers
Debugging & Troubleshooting
- Full request/response logging for every call
- Error tracking with detailed context
- Search and filter requests to identify issues quickly
→ Learn more about Helicone Observability
Observability Configuration
User Tracking
Track individual user behavior and analyze usage patterns across your application. This helps you understand which users are most active, identify power users, and monitor per-user costs:
import { createHelicone } from '@helicone/ai-sdk-provider';import { generateText } from 'ai';
const helicone = createHelicone({ apiKey: process.env.HELICONE_API_KEY,});
const result = await generateText({ model: helicone('gpt-4o-mini', { extraBody: { helicone: { userId: 'user@example.com', }, }, }), prompt: 'Hello world',});What you can track:
- Total requests per user
- Cost per user
- Average latency per user
- Most common use cases by user segment
→ Learn more about User Metrics
Custom Properties
Add structured metadata to segment and analyze requests by feature, environment, or any custom dimension. This enables powerful filtering and insights in your analytics dashboard:
import { createHelicone } from '@helicone/ai-sdk-provider';import { generateText } from 'ai';
const helicone = createHelicone({ apiKey: process.env.HELICONE_API_KEY,});
const result = await generateText({ model: helicone('gpt-4o-mini', { extraBody: { helicone: { properties: { feature: 'translation', source: 'mobile-app', language: 'French', environment: 'production', }, }, }, }), prompt: 'Translate this text to French',});Use cases for custom properties:
- Compare performance across different features or environments
- Track costs by product area or customer tier
- Identify which features drive the most AI usage
- A/B test different prompts or models by tagging experiments
→ Learn more about Custom Properties
Session Tracking
Group related requests into sessions to analyze conversation flows and multi-turn interactions. This is essential for understanding user journeys and debugging complex conversations:
import { createHelicone } from '@helicone/ai-sdk-provider';import { generateText } from 'ai';
const helicone = createHelicone({ apiKey: process.env.HELICONE_API_KEY,});
const result = await generateText({ model: helicone('gpt-4o-mini', { extraBody: { helicone: { sessionId: 'convo-123', sessionName: 'Travel Planning', sessionPath: '/chats/travel', }, }, }), prompt: 'Tell me more about that',});Session tracking benefits:
- View complete conversation history in a single timeline
- Calculate total cost per session/conversation
- Measure session duration and message counts
- Identify where users drop off in multi-turn conversations
- Debug issues by replaying entire conversation flows
Advanced Observability Features
Tags and Organization
Add tags to organize and filter requests in your analytics dashboard:
import { createHelicone } from '@helicone/ai-sdk-provider';import { generateText } from 'ai';
const helicone = createHelicone({ apiKey: process.env.HELICONE_API_KEY,});
const result = await generateText({ model: helicone('gpt-4o-mini', { extraBody: { helicone: { tags: ['customer-support', 'urgent'], properties: { ticketId: 'TICKET-789', priority: 'high', department: 'support', }, }, }, }), prompt: 'Help resolve this customer issue',});Tags insights:
- Filter and group requests by tags
- Track performance across different categories
- Identify patterns in tagged requests
- Build custom dashboards around specific tags
→ Learn more about Helicone Features
Streaming Response Tracking
Monitor streaming responses with full observability, including time-to-first-token and total streaming duration:
import { createHelicone } from '@helicone/ai-sdk-provider';import { streamText } from 'ai';
const helicone = createHelicone({ apiKey: process.env.HELICONE_API_KEY,});
const result = await streamText({ model: helicone('gpt-4o-mini', { extraBody: { helicone: { userId: 'user@example.com', sessionId: 'stream-session-123', tags: ['streaming', 'content-generation'], }, }, }), prompt: 'Write a short story about AI',});
for await (const chunk of result.textStream) { process.stdout.write(chunk);}Streaming metrics tracked:
- Time to first token (TTFT)
- Total streaming duration
- Tokens per second
- Complete request/response logging even for streams
- User experience metrics for real-time applications
- All metadata (sessions, users, tags) tracked for streamed responses