Skip to content

traccia-ai/traccia-py

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

51 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Traccia

Production-ready distributed tracing for AI agents and LLM applications

Traccia is a lightweight, high-performance Python SDK for observability and tracing of AI agents, LLM applications, and complex distributed systems. Built on OpenTelemetry standards with specialized instrumentation for AI workloads.

Traccia is available on PyPI.

✨ Features

  • πŸ” Automatic Instrumentation: Auto-patch OpenAI, Anthropic, requests, and HTTP libraries
  • πŸ€– Framework Integrations: Support for LangChain, CrewAI, and OpenAI Agents SDK
  • πŸ“Š LLM-Aware Tracing: Track tokens, costs, prompts, and completions automatically
  • πŸ“ˆ OpenTelemetry Metrics: Emit OTEL-compliant metrics for accurate cost/token tracking (independent of sampling)
  • ⚑ Zero-Config Start: Simple init() call with automatic config discovery
  • 🎯 Decorator-Based: Trace any function with @observe decorator
  • πŸ”§ Multiple Exporters: OTLP (compatible with Grafana Tempo, Jaeger, Zipkin), Console, or File
  • πŸ›‘οΈ Production-Ready: Rate limiting, error handling, config validation, robust flushing
  • πŸ›‘οΈ Guardrail Detection: Passive, zero-overhead detection of guardrails in traces β€” explicit, provider-native, and heuristic
  • πŸ“ Type-Safe: Full Pydantic validation for configuration
  • πŸš€ High Performance: Efficient batching, async support, minimal overhead
  • πŸ” Secure: No secrets in logs, configurable data truncation

πŸš€ Quick Start

Installation

pip install traccia

Basic Usage

from traccia import init, observe

# Initialize (auto-loads from traccia.toml if present)
init()

# Trace any function
@observe()
def my_function(x, y):
    return x + y

# That's it! Traces are automatically created and exported
result = my_function(2, 3)

With LLM Calls

from traccia import init, observe
from openai import OpenAI

init()  # Auto-patches OpenAI

client = OpenAI()

@observe(as_type="llm")
def generate_text(prompt: str) -> str:
    response = client.chat.completions.create(
        model="gpt-4",
        messages=[{"role": "user", "content": prompt}]
    )
    return response.choices[0].message.content

# Automatically tracks: model, tokens, cost, prompt, completion, latency
text = generate_text("Write a haiku about Python")

LangChain

Create a callback handler and pass it to config={"callbacks": [traccia_handler]}. Install the optional extra: pip install traccia[langchain].

from traccia import init
from traccia.integrations.langchain import CallbackHandler  # or TracciaCallbackHandler
from langchain_openai import ChatOpenAI

init()

# Create Traccia handler (no args)
traccia_handler = CallbackHandler()

# Use with any LangChain runnable
llm = ChatOpenAI(model="gpt-4o-mini")
result = llm.invoke(
    "Tell me a joke",
    config={"callbacks": [traccia_handler]}
)

Spans for LLM/chat model runs are created automatically with the same attributes as direct OpenAI instrumentation (model, prompt, usage, cost).

Note: pip install traccia[langchain] installs traccia plus langchain-core; you need this extra to use the callback handler. If you already have langchain-core (e.g. from langchain or langchain-openai), base pip install traccia may be enough at runtime, but traccia[langchain] is the supported way to get a compatible dependency.

OpenAI Agents SDK

Traccia automatically detects and instruments the OpenAI Agents SDK when installed. No extra code needed:

from traccia import init
from agents import Agent, Runner

init()  # Automatically enables Agents SDK tracing

agent = Agent(
    name="Assistant",
    instructions="You are a helpful assistant"
)
result = Runner.run_sync(agent, "Write a haiku about recursion")

Configuration: Auto-enabled by default when openai-agents is installed. To disable:

init(openai_agents=False)  # Explicit parameter
# OR set environment variable: TRACCIA_OPENAI_AGENTS=false
# OR in traccia.toml under [instrumentation]: openai_agents = false

Compatibility: If you have openai-agents installed but don't use it (e.g., using LangChain or pure OpenAI instead), the integration is registered but never invokedβ€”no overhead or extra spans.

CrewAI

Traccia automatically instruments CrewAI when it is installed in your environment.

from traccia import init
from crewai import Agent, Task, Crew, Process

init()  # Auto-enables CrewAI tracing when CrewAI is installed

researcher = Agent(role="Research Analyst", goal="Research a topic", llm="gpt-4o-mini")
task = Task(description="Research Shawn Michaels", agent=researcher)

crew = Crew(agents=[researcher], tasks=[task], process=Process.sequential, verbose=True)
result = crew.kickoff()

Traccia will create spans for the crew (crewai.crew.kickoff), each task (crewai.task.*), agents (crewai.agent.*), and underlying LLM calls, which nest under the existing OpenAI spans.

Configuration: Auto-enabled by default when crewai is installed. To disable:

init(crewai=False)  # Explicit parameter
# OR set environment variable: TRACCIA_CREWAI=false
# OR in traccia.toml under [instrumentation]: crewai = false

πŸ›‘οΈ Guardrail Detection

Traccia includes a passive guardrail detection engine that runs as a span processor β€” no runtime enforcement, no changes required to existing agent code. It inspects every span as it ends, classifies guardrail signals into structured findings, and writes results back onto spans so they appear in any configured exporter.

How it works

Detection is automatic once traccia.init() is called. The processor runs three tiers against every span's attributes:

Tier Source Confidence How
A β€” Explicit explicit high @observe(as_type="guardrail") or guardrail_span()
B β€” Provider-native provider_native high/medium LLM finish reason, stop reason, safety ratings
C β€” Heuristic heuristic always low Denial keywords in tool error messages

At the end of each trace (when the root span ends), the processor evaluates which guardrail categories should exist given the agent's observed capabilities (LLM calls, tool use, user-provided text) and reports which are missing.

Annotating guardrails explicitly (Tier A)

Option 1: guardrail_span context manager β€” recommended for inline checks

from traccia.guardrails import guardrail_span

with guardrail_span("pii_check", category="pii", enforcement_mode="warn") as span:
    result = run_pii_check(user_input)
    span.set_attribute("guardrail.triggered", result.found_pii)

Option 2: @observe(as_type="guardrail") decorator β€” for function-level guardrails

If your guardrail function returns a bool, guardrail.triggered is set automatically:

from traccia import observe

@observe(
    as_type="guardrail",
    attributes={
        "guardrail.name": "prompt_injection_check",
        "guardrail.category": "prompt_injection",
        "guardrail.enforcement_mode": "block",
    }
)
def check_injection(text: str) -> bool:
    return any(kw in text.lower() for kw in INJECTION_KEYWORDS)
    # True β†’ triggered (blocked), False β†’ not triggered

For non-bool returns, set guardrail.triggered manually on the current span.

Suppressing false positive missing-guardrail warnings

Batch pipelines or internal-only agents often get flagged for prompt_injection / input_validation because they make LLM calls with llm.prompt. Suppress specific categories for a run:

from traccia.guardrails import guardrail_span

# Convenience: suppress_missing on guardrail_span
with guardrail_span("root", category="unknown", suppress_missing=["prompt_injection", "input_validation"]):
    run_batch_pipeline()

# Or directly on any span
from traccia.guardrails import ATTR_GUARDRAIL_SUPPRESS_MISSING
span.set_attribute(ATTR_GUARDRAIL_SUPPRESS_MISSING, ["prompt_injection"])

What appears in traces

Per span (when a guardrail signal is found):

  • guardrail.finding.count β€” number of findings on this span
  • guardrail.findings β€” JSON array of GuardrailFinding objects

On the root span (aggregated summary of the entire run):

  • guardrail.summary β€” full GuardrailSummary JSON
  • guardrail.summary.coverage_confidence β€” "high", "medium", or "low"
  • guardrail.summary.missing_count β€” number of expected-but-missing guardrail categories
  • guardrail.summary.detected_categories β€” list of detected category strings

Provider-native detection (Tier B, automatic)

These signals are detected automatically from LLM span attributes β€” no annotation required:

Provider Signal Attribute
OpenAI finish_reason = "content_filter" llm.finish_reason
Azure OpenAI finish_reason = "content_filtered" llm.finish_reason
Google GenAI finish_reason = "SAFETY" llm.finish_reason
Anthropic stop_reason = "content_filter" llm.stop_reason
Anthropic Policy violation error message error.message + llm.vendor=anthropic
Google/LangChain "blocked": true or "probability": "HIGH" in safety ratings llm.safety_ratings

Disabling Tier C (heuristic) detection

Tier A/B stay on. Turn off tool-error keyword heuristics if you do not want Tier C at all:

init(guardrail_heuristics=False)

Or TRACCIA_GUARDRAIL_HEURISTICS=false, or in traccia.toml: [instrumentation] β†’ guardrail_heuristics = false.

Hard limits β€” what cannot be captured

  • Guardrails running outside the traced process (API gateways, proxies, external validators) are invisible unless they write span attributes.
  • A guardrail that exists but never fires cannot be distinguished from a missing guardrail. Only explicit annotation proves presence.
  • Prompt injection detection requires an explicit span β€” the model's output alone cannot reliably indicate whether a check ran.
  • Capability inference (handles_user_text) is inferred from llm.prompt being present; batch pipelines and user-facing agents look the same. Use suppression to opt out.

πŸ“– Configuration

Configuration Precedence

Traccia merges configuration from multiple sources with the following priority (highest to lowest):

  1. Explicit parameters β€” init(endpoint="...", agent_id="...") or start_tracing(...)
  2. Environment variables β€” TRACCIA_ENDPOINT, TRACCIA_AGENT_ID, etc.
  3. Config file β€” traccia.toml (current directory) or ~/.traccia/config.toml
  4. Defaults β€” Built-in SDK defaults

Example: If you set TRACCIA_ENDPOINT in your environment and pass endpoint=... to init(), the explicit parameter wins.


Configuration File

Create a traccia.toml file in your project root:

traccia config init

This creates a template config file:

[tracing]
# API key β€” required for the Traccia platform, not needed for local OTLP backends
api_key = ""

# Endpoint URL for OTLP trace ingestion (default: Traccia platform)
# For local OTLP backends use e.g. endpoint = "http://localhost:4318/v1/traces"
endpoint = "https://api.traccia.ai/v2/traces"

sample_rate = 1.0           # 0.0 to 1.0
auto_start_trace = true     # Auto-start root trace on init
auto_trace_name = "root"    # Name for auto-started trace
use_otlp = true             # Use OTLP exporter
# service_name = "my-app"   # Optional service name
# service_role has no env var β€” pass via init(service_role="orchestrator")

[exporters]
# Only enable ONE exporter at a time
enable_console = false        # Print traces to console
enable_file = false           # Write traces to file
file_exporter_path = "traces.jsonl"
reset_trace_file = false      # Reset file on initialization

[instrumentation]
enable_patching = true          # Auto-patch libraries (OpenAI, Anthropic, requests)
enable_token_counting = true    # Count tokens for LLM calls
enable_costs = true             # Calculate costs
openai_agents = true            # Auto-enable OpenAI Agents SDK integration
crewai = true                   # Auto-enable CrewAI integration
guardrail_heuristics = true     # Tier C heuristic guardrail detection (tool error keywords)
auto_instrument_tools = false   # Auto-instrument tool calls (experimental)
max_tool_spans = 100            # Max tool spans to create
max_span_depth = 10             # Max nested span depth

[rate_limiting]
# Optional: limit spans per second
# max_spans_per_second = 100.0
max_queue_size = 5000           # Max buffered spans
max_block_ms = 100              # Max ms to block before dropping
max_export_batch_size = 512     # Spans per export batch
schedule_delay_millis = 5000    # Delay between batches

[metrics]
enable_metrics = true           # Enable OpenTelemetry metrics
# metrics_endpoint = ""         # Defaults to {traces_base}/v2/metrics
metrics_sample_rate = 1.0       # Metrics sampling rate (1.0 = 100%)

[runtime]
# Optional runtime metadata (agent identity: prefer init(agent_id=..., agent_name=..., env=...) or TRACCIA_* env)
# session_id = ""
# user_id = ""
# tenant_id = ""
# project_id = ""
# agent_id = ""   # Single-agent: set in code or TRACCIA_AGENT_ID
# agent_name = ""
# env = ""        # e.g. production, staging, dev

[logging]
debug = false                   # Enable debug logging
enable_span_logging = false     # Enable span-level logging

[advanced]
# attr_truncation_limit = 1000  # Max attribute value length

Default endpoint

If you do not set endpoint (in config, environment, or when calling init() / start_tracing()), the SDK uses the Traccia platform by default (https://api.traccia.ai/v2/traces). You can override it to send traces to your own OTLP-compatible backend.

The default is defined in traccia.config: DEFAULT_OTLP_TRACE_ENDPOINT. The alias DEFAULT_ENDPOINT is kept for backward compatibility (same value).

OTLP Backend Compatibility

Traccia is fully OTLP-compatible and works with:

  • Grafana Tempo - http://tempo:4318/v1/traces
  • Jaeger - http://jaeger:4318/v1/traces
  • Zipkin - Configure via OTLP endpoint
  • SigNoz - Self-hosted observability platform
  • Traccia Platform - https://api.traccia.ai/v2/traces (requires API key)

Environment Variables

All config parameters can be set via environment variables with the TRACCIA_ prefix:

Tracing: TRACCIA_API_KEY, TRACCIA_ENDPOINT, TRACCIA_SAMPLE_RATE, TRACCIA_AUTO_START_TRACE, TRACCIA_AUTO_TRACE_NAME, TRACCIA_USE_OTLP, TRACCIA_SERVICE_NAME

Exporters: TRACCIA_ENABLE_CONSOLE, TRACCIA_ENABLE_FILE, TRACCIA_FILE_PATH, TRACCIA_RESET_TRACE_FILE

Instrumentation: TRACCIA_ENABLE_PATCHING, TRACCIA_ENABLE_TOKEN_COUNTING, TRACCIA_ENABLE_COSTS, TRACCIA_AUTO_INSTRUMENT_TOOLS, TRACCIA_MAX_TOOL_SPANS, TRACCIA_MAX_SPAN_DEPTH, TRACCIA_OPENAI_AGENTS, TRACCIA_CREWAI, TRACCIA_GUARDRAIL_HEURISTICS

Rate Limiting: TRACCIA_MAX_SPANS_PER_SECOND, TRACCIA_MAX_QUEUE_SIZE, TRACCIA_MAX_BLOCK_MS, TRACCIA_MAX_EXPORT_BATCH_SIZE, TRACCIA_SCHEDULE_DELAY_MILLIS

Runtime: TRACCIA_SESSION_ID, TRACCIA_USER_ID, TRACCIA_TENANT_ID, TRACCIA_PROJECT_ID, TRACCIA_AGENT_ID, TRACCIA_AGENT_NAME, TRACCIA_ENV

Legacy alias: TRACCIA_PROJECT (maps to project_id)

Logging: TRACCIA_DEBUG, TRACCIA_ENABLE_SPAN_LOGGING

Advanced: TRACCIA_ATTR_TRUNCATION_LIMIT

Programmatic Configuration

from traccia import init

# Override config programmatically (including agent identity for single-agent services)
init(
    endpoint="http://tempo:4318/v1/traces",
    sample_rate=0.5,
    enable_costs=True,
    max_spans_per_second=100.0,
    agent_id="my-agent",
    agent_name="My Agent",
    env="production",
)

Multi-Agent Orchestrator Services

For services that orchestrate many logical agents in one process, set a service role and scope per-run identity:

from traccia import init, runtime_config

init(
    service_name="my-multi-agent-api",
    service_role="orchestrator",
    auto_start_trace=False,
)

with runtime_config.run_identity(agent_id="billing-agent", agent_name="Billing Agent", env="production"):
    # run one logical agent task
    ...

This prevents the host service from being registered as a synthetic agent in the Traccia platform.

Safe Parallel Runs in One Process

If one process runs many agents concurrently (for example, an API server or orchestrator), use this pattern:

  • Call init() once per process (for example at startup), not per request.
  • Wrap each logical "run" in runtime_config.run_identity(...) to set agent id/name/env for that run.
  • Do not call stop_tracing() per request; use force_flush() to flush spans/metrics after a run without shutting down the provider.
from traccia import init, span, force_flush, runtime_config

init(service_name="multi-agent-service", auto_start_trace=False)

def run_agent(agent_id: str, env: str, payload: dict):
    # Scope identity to this run only
    with runtime_config.run_identity(agent_id=agent_id, agent_name=agent_id, env=env):
        with span("agent.run") as root:
            root.set_attribute("agent.id", agent_id)
            root.set_attribute("agent.run.mode", "api")
            # ... your agent logic here ...

    # Flush without tearing down the global provider
    force_flush(5.0)

🎯 Usage Guide

The @observe Decorator

The @observe decorator is the primary way to instrument your code:

from traccia import observe

# Basic usage
@observe()
def process_data(data):
    return transform(data)

# Custom span name
@observe(name="data_pipeline")
def process_data(data):
    return transform(data)

# Add custom attributes
@observe(attributes={"version": "2.0", "env": "prod"})
def process_data(data):
    return transform(data)

# Specify span type
@observe(as_type="llm")  # "span", "llm", "tool"
def call_llm():
    pass

# Skip capturing specific arguments
@observe(skip_args=["password", "secret"])
def authenticate(username, password):
    pass

# Skip capturing result (for large returns)
@observe(skip_result=True)
def fetch_large_dataset():
    return huge_data

Available Parameters:

  • name (str, optional): Custom span name (defaults to function name)
  • attributes (dict, optional): Initial span attributes
  • as_type (str): Span type - "span", "llm", "tool", or "guardrail"
  • skip_args (list, optional): List of argument names to skip capturing
  • skip_result (bool): Skip capturing the return value

Async Functions

@observe works seamlessly with async functions:

@observe()
async def async_task(x):
    await asyncio.sleep(1)
    return x * 2

result = await async_task(5)

Manual Span Creation

For more control, create spans manually:

from traccia import get_tracer, span

# Using convenience function
with span("operation_name") as s:
    s.set_attribute("key", "value")
    s.add_event("checkpoint_reached")
    do_work()

# Using tracer directly
tracer = get_tracer("my_service")
with tracer.start_as_current_span("operation") as s:
    s.set_attribute("user_id", 123)
    do_work()

Error Handling

Traccia automatically captures and records errors:

@observe()
def failing_function():
    raise ValueError("Something went wrong")

# Span will contain:
# - error.type: "ValueError"
# - error.message: "Something went wrong"
# - error.stack_trace: (truncated stack trace)
# - span status: ERROR

Nested Spans

Spans are automatically nested based on call hierarchy:

@observe()
def parent_operation():
    child_operation()
    return "done"

@observe()
def child_operation():
    grandchild_operation()

@observe()
def grandchild_operation():
    pass

# Creates nested span hierarchy:
# parent_operation
#   └── child_operation
#       └── grandchild_operation

πŸ› οΈ CLI Tools

Traccia includes a powerful CLI for configuration and diagnostics:

traccia config init

Create a new traccia.toml configuration file:

traccia config init
traccia config init --force  # Overwrite existing

traccia doctor

Validate configuration and diagnose issues:

traccia doctor

# Output:
# 🩺 Running Traccia configuration diagnostics...
# 
# βœ… Found config file: ./traccia.toml
# βœ… Configuration is valid
# 
# πŸ“Š Configuration summary:
#    β€’ API Key: ❌ Not set (optional)
#    β€’ Endpoint: https://api.traccia.ai/v2/traces
#    β€’ Sample Rate: 1.0
#    β€’ OTLP Exporter: βœ… Enabled

traccia check

Test connectivity to your exporter endpoint:

traccia check
traccia check --endpoint http://tempo:4318/v1/traces

🎨 Advanced Features

Rate Limiting

Protect your infrastructure with built-in rate limiting:

[rate_limiting]
max_spans_per_second = 100.0  # Limit to 100 spans/sec
max_queue_size = 5000         # Max buffered spans
max_block_ms = 100             # Block up to 100ms before dropping

Behavior:

  1. Try to acquire capacity immediately
  2. If unavailable, block for up to max_block_ms
  3. If still unavailable, drop span and log warning

When spans are dropped due to rate limiting, warnings are logged to help you monitor and adjust limits.

Sampling

Control trace volume with sampling:

# Sample 10% of traces
init(sample_rate=0.1)

# Sampling is applied at trace creation time
# Traces are either fully included or fully excluded

Token Counting & Cost Calculation

Automatic for supported LLM providers (OpenAI, Anthropic):

@observe(as_type="llm")
def call_openai(prompt):
    response = client.chat.completions.create(
        model="gpt-4",
        messages=[{"role": "user", "content": prompt}]
    )
    return response.choices[0].message.content

# Span automatically includes:
# - llm.token.prompt_tokens
# - llm.token.completion_tokens
# - llm.token.total_tokens
# - llm.cost.total (in USD)

Metrics

Traccia emits OTEL-compliant metrics for accurate cost and token tracking, independent of trace sampling.

Why Metrics?

With trace sampling (e.g., sample_rate=0.1), only 10% of traces are exported. Cost calculated from traces will be 10x underestimated. Metrics solve this by recording data for every LLM call, regardless of sampling.

Default Metrics

Traccia automatically emits these metrics:

Metric Type Unit Description
gen_ai.client.token.usage Histogram {token} Input/output tokens per call
gen_ai.client.operation.duration Histogram s LLM operation duration
gen_ai.client.operation.cost Histogram usd Cost per call (USD)
gen_ai.client.completions.exceptions Counter 1 Exception count
gen_ai.agent.runs Counter 1 Agent runs (CrewAI, OpenAI Agents)
gen_ai.agent.turns Counter 1 Agent turns
gen_ai.agent.execution_time Histogram s Agent execution time

Attributes: gen_ai.system (openai, anthropic), gen_ai.request.model, gen_ai.agent.id, gen_ai.agent.name

Configuration

from traccia import init

init(
    enable_metrics=True,  # Default: True
    metrics_endpoint="https://your-backend.com/v2/metrics",  
    metrics_sample_rate=1.0,  # Default: 1.0 (100%)
)

Or via traccia.toml:

[metrics]
enable_metrics = true
metrics_endpoint = "https://your-backend.com/v2/metrics"
metrics_sample_rate = 1.0

Or via environment variables:

export TRACCIA_ENABLE_METRICS=true
export TRACCIA_METRICS_ENDPOINT=https://your-backend.com/v2/metrics
export TRACCIA_METRICS_SAMPLE_RATE=1.0

Custom Metrics

Record your own metrics:

from traccia.metrics import record_counter, record_histogram

# Record a counter
record_counter("my_custom_events", 1, {"event_type": "user_action"})

# Record a histogram
record_histogram("my_custom_latency", 0.123, {"service": "api"}, unit="s")

Agent Metrics vs. Plain LLM Calls

Agent-level metrics (such as gen_ai.agent.runs and gen_ai.agent.execution_time) are only emitted when Traccia can see a real agent lifecycle (for example, CrewAI crews or OpenAI Agents SDK runs). For plain OpenAI/Anthropic calls and most simple LangChain usages, you will still get full LLM metrics (gen_ai.client.*), but no agent metrics unless you build an explicit agent abstraction on top.


πŸ”§ Troubleshooting

Enable Debug Logging

import logging
logging.basicConfig(level=logging.DEBUG)

# Or via config
init(debug=True)

# Or via env var
# TRACCIA_DEBUG=1 python your_script.py

Common Issues

Traces not appearing

  1. Check connectivity: traccia check
  2. Validate config: traccia doctor
  3. Enable debug logging
  4. Verify endpoint is correct and accessible

High memory usage

  • Reduce max_queue_size in rate limiting config
  • Lower sample_rate to reduce volume
  • Enable rate limiting with max_spans_per_second

Spans being dropped

  • Check rate limiter logs for warnings
  • Increase max_spans_per_second if set
  • Increase max_queue_size if spans are queued
  • Check traccia doctor output

πŸ“š API Reference

Core Functions

init(**kwargs) -> TracerProvider

Initialize the Traccia SDK. All parameters are optional; configuration is merged from traccia.toml β†’ env vars β†’ explicit parameters (highest wins).

Parameters:

Tracing

  • endpoint (str): OTLP endpoint URL (default: https://api.traccia.ai/v2/traces)
  • api_key (str): API key for the Traccia platform
  • sample_rate (float): Sampling rate 0.0–1.0 (default: 1.0)
  • auto_start_trace (bool): Auto-start a root trace on init (default: True)
  • auto_trace_name (str): Name for the auto-started trace (default: "root")
  • use_otlp (bool): Use OTLP exporter (default: True)
  • service_name (str): Service name (auto-detected if not set)
  • service_role (str): "orchestrator" to prevent this service being registered as an agent
  • config_file (str): Path to a custom traccia.toml

Exporters

  • enable_console_exporter (bool): Print spans to stdout (default: False)
  • enable_file_exporter (bool): Write spans to file (default: False)
  • file_exporter_path (str): Path for file exporter (default: "traces.jsonl")
  • reset_trace_file (bool): Clear file on init (default: False)

Instrumentation

  • enable_patching (bool): Auto-patch OpenAI, Anthropic, requests (default: True)
  • enable_token_counting (bool): Count tokens (default: True)
  • enable_costs (bool): Calculate costs (default: True)
  • openai_agents (bool): Auto-enable OpenAI Agents SDK integration (default: True)
  • crewai (bool): Auto-enable CrewAI integration (default: True)
  • guardrail_heuristics (bool): Enable Tier C heuristic guardrail detection (default: True)
  • auto_instrument_tools (bool): Experimental tool auto-instrumentation (default: False)
  • max_tool_spans (int): Max tool spans per trace (default: 100)
  • max_span_depth (int): Max nested span depth (default: 10)

Agent identity (single-agent services)

  • agent_id (str): Logical agent identifier
  • agent_name (str): Human-readable agent name
  • env (str): Deployment environment, e.g. "production", "staging"

Runtime metadata

  • session_id (str): Session identifier
  • user_id (str): User identifier
  • tenant_id (str): Tenant / org identifier
  • project_id (str): Project identifier

Metrics

  • enable_metrics (bool): Emit OTEL metrics (default: True)
  • metrics_endpoint (str): Metrics endpoint (derived from tracing endpoint if not set)
  • metrics_sample_rate (float): Metrics sampling rate (default: 1.0)

Rate limiting

  • max_spans_per_second (float): Rate limit spans/sec (default: None = unlimited)
  • max_block_ms (int): Max ms to block before dropping a span (default: 100)
  • max_queue_size (int): Max buffered spans (default: 5000)
  • max_export_batch_size (int): Spans per export batch (default: 512)
  • schedule_delay_millis (int): Batch export interval ms (default: 5000)

Misc

  • debug (bool): Enable debug logging (default: False)
  • attr_truncation_limit (int): Max attribute value length (default: None)

Returns: TracerProvider instance

stop_tracing(flush_timeout: float = 1.0) -> None

Stop tracing and flush pending spans.

Parameters:

  • flush_timeout (float): Max seconds to wait for flush

get_tracer(name: str = "default") -> Tracer

Get a tracer instance.

Parameters:

  • name (str): Tracer name (typically module/service name)

Returns: Tracer instance

span(name: str, attributes: dict = None) -> Span

Create a span context manager.

Parameters:

  • name (str): Span name
  • attributes (dict, optional): Initial attributes

Returns: Span context manager

Decorator

@observe(name=None, *, attributes=None, tags=None, as_type="span", skip_args=None, skip_result=False)

Decorate a function to create spans automatically.

Parameters:

  • name (str, optional): Span name (default: function name)
  • attributes (dict, optional): Initial attributes
  • tags (list[str], optional): User-defined identifiers for the observed method
  • as_type (str): Span type ("span", "llm", "tool", "guardrail")
  • skip_args (list, optional): Arguments to skip capturing
  • skip_result (bool): Skip capturing return value

Configuration

load_config(config_file=None, overrides=None) -> TracciaConfig

Load and validate configuration.

Parameters:

  • config_file (str, optional): Path to config file
  • overrides (dict, optional): Override values

Returns: Validated TracciaConfig instance

Raises: ConfigError if invalid

validate_config(config_file=None, overrides=None) -> tuple[bool, str, TracciaConfig | None]

Validate configuration without loading.

Returns: Tuple of (is_valid, message, config_or_none)


πŸ—οΈ Architecture

Data Flow

Application Code (@observe)
        ↓
   Span Creation
        ↓
   Processors (token counting, cost, enrichment)
        ↓
   Rate Limiter (optional)
        ↓
   Batch Processor (buffering)
        ↓
   Exporter (OTLP/Console/File)
        ↓
   Backend (Grafana Tempo / Jaeger / Zipkin / etc.)

Instrumentation vs Integrations

  • traccia.instrumentation.*: Infrastructure and vendor instrumentation.

    • HTTP client/server helpers (including FastAPI middleware).
    • Vendor SDK hooks and monkey patching (e.g., OpenAI, Anthropic, requests).
    • Decorators and utilities used for auto-instrumenting arbitrary functions.
  • traccia.integrations.*: AI/agent framework integrations.

    • Adapters that plug into higher-level frameworks via their official extension points (e.g., LangChain callbacks).
    • Work at the level of chains, tools, agents, and workflows rather than raw HTTP or SDK calls.

🀝 Contributing

Contributions are welcome! Whether it's bug fixes, new features, documentation improvements, or examples - we appreciate your help.

How to Contribute

  1. Fork the repository
  2. Create a feature branch: git checkout -b feature/amazing-feature
  3. Make your changes and add tests
  4. Run tests: pytest traccia/tests/
  5. Lint your code: ruff check traccia/
  6. Commit: git commit -m "Add amazing feature"
  7. Push: git push origin feature/amazing-feature
  8. Open a Pull Request

Development Setup

# Clone the repository (Python SDK)
git clone https://github.com/traccia-ai/traccia-py.git
cd traccia-py

# Create virtual environment
python -m venv venv
source venv/bin/activate  # On Windows: venv\Scripts\activate

# Install in editable mode with dev dependencies
pip install -e ".[dev]"

# Run tests
pytest traccia/tests/ -v

# Run with coverage
pytest traccia/tests/ --cov=traccia --cov-report=html

Code Style

  • Follow PEP 8
  • Use type hints where appropriate
  • Add docstrings for public APIs
  • Write tests for new features
  • Keep PRs focused and atomic

Areas We'd Love Help With

  • Integrations: Add support for more LLM providers (Cohere, AI21, local models)
  • Backends: Test and document setup with different OTLP backends
  • Examples: Real-world examples of agent instrumentation
  • Documentation: Tutorials, guides, video walkthroughs
  • Performance: Optimize hot paths, reduce overhead
  • Testing: Improve test coverage, add integration tests

πŸ“„ License

Apache License 2.0 - see LICENSE for full terms and conditions.


πŸ™ Acknowledgments

Built with:

Inspired by observability tools in the ecosystem and designed to work seamlessly with the OTLP standard.


πŸ“ž Support & Community


Made with ❀️ for the AI agent community

Packages

 
 
 

Contributors

Languages