Invocation Log & Analytics
PromptShuttle provides detailed logging of every request and three levels of analytics aggregation for understanding cost, performance, and usage patterns.
Invocation log
Every request is logged with full request/response pairs, timing, cost, and metadata.
Browse the log
GET /api/v1/llm-logsFilters
cursor
string
Pagination cursor (from previous response)
pageSize
integer
Results per page (default 50)
from
datetime
Start time filter
to
datetime
End time filter
environments
string array
Filter by environment name
flows
string array
Filter by flow name
statuses
string array
Filter by status (Pending, Succeeded, Failed)
tags
string array
Filter by tags (any match)
customerId
string
Filter by end-customer external ID
rootOnly
boolean
Show only root-level requests (exclude sub-agents)
parentRequestId
string
Show children of a specific parent request
Log entry fields
Each log entry includes:
Request details — Messages, model, parameters, environment, tags
Response details — Text output, tool calls, citations, finish reason
Usage — Tokens in/out, reasoning tokens, cost in credits and USD
Timing — Duration in milliseconds
Hierarchy — Parent/root request IDs, agent depth, agent role
Customer — End-customer ID and external ID
Feedback — Feedback score if submitted
Request hierarchy
For multi-agent requests, view the full hierarchy:
Returns:
Breadcrumb — Ancestor path from root to current request
Parent — Direct parent summary
Children — Direct child agent summaries
Streaming events
View the SSE events emitted during a request:
Filter by event type: ?types=agentStarted,agentCompleted
Get event type counts: GET /api/v1/llm-logs/{id}/events/stats
Log entries (debug)
For requests with elevated log levels, view detailed internal log entries:
Analytics
PromptShuttle provides analytics at three levels of granularity, each answering different questions.
Level 1: Per-inference stats
What it answers: How are individual LLM calls performing?
Metrics per model:
Average tokens in/out
Average cost and latency
Percentiles (P50, P90, P95, P99) for tokens, latency, and cost
Token and latency histograms
Level 2: Per-request stats
What it answers: How are complete requests performing (including tool-calling loops)?
Metrics:
Total cost per request (across all LLM calls and tool invocations)
Total function calls per request
Request latency distribution
Broken down by primary model
Level 3: Per-tree stats (multi-agent)
What it answers: How are entire agent trees performing?
Metrics:
Total tree cost (root + all sub-agents)
Agent count per tree
Maximum depth reached
Depth distribution
Agent count distribution
All three levels at once
Returns all three levels plus time series data for invocations and cost.
Tenant-wide statistics
Query parameters:
period
TimeSpan
Time window (e.g. 7.00:00:00 for 7 days)
resolution
string
Time bucket size: Minutes5, Minutes15, Hour, Day, Week
Returns:
Invocation count (current vs. previous period)
Per-flow invocation time series
Cost time series
Token usage time series
Top flows by invocation count
Credit usage comparison
Account overview
Returns cost and request counts broken down by model and by flow, with percentage distribution.
Data export
Export your flows and tools for backup or migration:
Returns all flows (with all versions and templates) and tools in a single JSON document.
Import into another tenant:
Duplicate flows and tools are skipped (matched by name).
Last updated