Invocation Log & Analytics

PromptShuttle provides detailed logging of every request and three levels of analytics aggregation for understanding cost, performance, and usage patterns.

Invocation log

Every request is logged with full request/response pairs, timing, cost, and metadata.

Browse the log

GET /api/v1/llm-logs

Filters

Parameter
Type
Description

cursor

string

Pagination cursor (from previous response)

pageSize

integer

Results per page (default 50)

from

datetime

Start time filter

to

datetime

End time filter

environments

string array

Filter by environment name

flows

string array

Filter by flow name

statuses

string array

Filter by status (Pending, Succeeded, Failed)

tags

string array

Filter by tags (any match)

customerId

string

Filter by end-customer external ID

rootOnly

boolean

Show only root-level requests (exclude sub-agents)

parentRequestId

string

Show children of a specific parent request

Log entry fields

Each log entry includes:

  • Request details — Messages, model, parameters, environment, tags

  • Response details — Text output, tool calls, citations, finish reason

  • Usage — Tokens in/out, reasoning tokens, cost in credits and USD

  • Timing — Duration in milliseconds

  • Hierarchy — Parent/root request IDs, agent depth, agent role

  • Customer — End-customer ID and external ID

  • Feedback — Feedback score if submitted

Request hierarchy

For multi-agent requests, view the full hierarchy:

Returns:

  • Breadcrumb — Ancestor path from root to current request

  • Parent — Direct parent summary

  • Children — Direct child agent summaries

Streaming events

View the SSE events emitted during a request:

Filter by event type: ?types=agentStarted,agentCompleted

Get event type counts: GET /api/v1/llm-logs/{id}/events/stats

Log entries (debug)

For requests with elevated log levels, view detailed internal log entries:

Analytics

PromptShuttle provides analytics at three levels of granularity, each answering different questions.

Level 1: Per-inference stats

What it answers: How are individual LLM calls performing?

Metrics per model:

  • Average tokens in/out

  • Average cost and latency

  • Percentiles (P50, P90, P95, P99) for tokens, latency, and cost

  • Token and latency histograms

Level 2: Per-request stats

What it answers: How are complete requests performing (including tool-calling loops)?

Metrics:

  • Total cost per request (across all LLM calls and tool invocations)

  • Total function calls per request

  • Request latency distribution

  • Broken down by primary model

Level 3: Per-tree stats (multi-agent)

What it answers: How are entire agent trees performing?

Metrics:

  • Total tree cost (root + all sub-agents)

  • Agent count per tree

  • Maximum depth reached

  • Depth distribution

  • Agent count distribution

All three levels at once

Returns all three levels plus time series data for invocations and cost.

Tenant-wide statistics

Query parameters:

Parameter
Type
Description

period

TimeSpan

Time window (e.g. 7.00:00:00 for 7 days)

resolution

string

Time bucket size: Minutes5, Minutes15, Hour, Day, Week

Returns:

  • Invocation count (current vs. previous period)

  • Per-flow invocation time series

  • Cost time series

  • Token usage time series

  • Top flows by invocation count

  • Credit usage comparison

Account overview

Returns cost and request counts broken down by model and by flow, with percentage distribution.

Data export

Export your flows and tools for backup or migration:

Returns all flows (with all versions and templates) and tools in a single JSON document.

Import into another tenant:

Duplicate flows and tools are skipped (matched by name).

Last updated