Obiguard Tracing lets you track the full lifecycle of your LLM requests in a single, chronological view.
This is perfect for agentic workflows, chatbots, or multi-step LLM calls, by helping you understand and optimize your AI application’s performance.
Obiguard uses OpenTelemetry-compliant tracing to monitor your LLM requests. By attaching a trace ID to your requests, all related LLM calls are automatically grouped in the Traces View and displayed as “spans” within a trace.
A “span” represents a subgroup of LLM calls. Depending on your instrumentation, a span can refer to a group within a trace or a single LLM call.
Obiguard implements tracing using a tree-like structure, similar to OpenTelemetry.
Each node in this tree is a span, identified by a unique spanId
and an optional spanName
. Child spans reference their parent using parentSpanId
. Spans without a parent become root nodes.
Key | Node | Key - Python | Expected Value | Required? |
---|---|---|---|---|
traceId | trace_id | Unique string | YES | |
spanId | span_id | Unique string | NO | |
spanName | span_name | string | NO | |
parentSpanId | parent_span_id | Unique string | NO |