Welcome
PromptShuttle is an LLM orchestration platform that sits between your application and the major LLM providers. It gives you a single API to manage prompts, route across models, run multi-step agentic workflows, and track costs — all with multi-tenant isolation out of the box.
Why PromptShuttle?
One API, every provider. Write your integration once and route to OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, xAI, or Perplexity. Switch models without changing client code.
Prompt versioning and environments. Manage prompt templates with version control. Promote versions across dev, staging, and production environments without redeploying your app.
Multi-agent orchestration. Build DAG-based workflows where templates can invoke other templates as sub-agents, with full execution tracing, cost tracking, and depth limits.
Drop-in OpenAI replacement. Use the OpenAI-compatible endpoint with any existing OpenAI SDK client — just change the base URL and API key.
Built-in cost control. Per-request cost ceilings, tenant-level budgets, detailed usage analytics at three levels of granularity (per-inference, per-request, per-agent-tree).
Function calling and tools. Define external HTTP tools, MCP server tools, or agent tools. PromptShuttle handles the tool-calling loop automatically.
Quick links
Make my first API call
Use my existing OpenAI SDK
Build a prompt template with versions
Execute a flow via API
Add function calling / tools
Connect an MCP server
Set up model aliases and fallbacks
Stream agent execution events
Track per-customer usage
Understand costs and credits
Browse request history
Understand the core model
Supported providers
OpenAI
GPT-4o, GPT-4o mini, o1, o3
Anthropic
Claude 4 Opus, Claude 4 Sonnet, Claude 3.5 Haiku
Gemini 2.5 Pro, Gemini 2.5 Flash
Groq
Llama, Mixtral (fast inference)
DeepSeek
DeepSeek Chat, DeepSeek Reasoner
xAI
Grok
Perplexity
Sonar (web-search integrated)
List all models and pricing programmatically via GET /api/v1/models/descriptors.
Base URL
Last updated