Skip to main content
AgentMark’s observability platform helps you monitor, debug, and optimize your LLM applications in production. Built on OpenTelemetry, it automatically collects telemetry data from your prompts and provides actionable insights.
Developers set up observability in your application. See Development documentation for setup instructions.
Traces

What We Track

Inference Spans - Full lifecycle of prompt execution
  • Token usage and costs
  • Response times
  • Model information
  • Completion status
Tool Calls - When your prompts use tools
  • Tool name and parameters
  • Execution duration
  • Success/failure status
  • Return values
Streaming Metrics - For streaming responses
  • Time to first token
  • Tokens per second
  • Total streaming duration
Sessions - Group related traces together
  • Organize by user interaction
  • Track multi-step workflows
  • Monitor batch processing
  • Analyze performance patterns
Alerts - Monitor critical metrics
  • Cost thresholds
  • Latency monitoring
  • Error rate tracking
  • Evaluation score monitoring
  • Notifications via Slack or webhooks

Key Features

Real-Time Monitoring - View traces and metrics as they happen in your application. Graph Visualization - Interactive visual representation of complex workflows, showing component relationships, execution flow, and dependencies. Cost Tracking - Understand your LLM spending across models, users, and time periods. Performance Analysis - Identify bottlenecks and optimize response times. Quality Assurance - Track evaluation scores and manually annotate traces for quality assessment. Flexible Integration - Works with any AI framework through OpenTelemetry standard.

Next Steps

Have Questions?

We’re here to help! Choose the best way to reach us: