RouteShift sits between your code and the major LLM providers — OpenAI, Anthropic, Google, AWS Bedrock, Azure OpenAI, Together, Groq, and a growing roster of regional Chinese providers (Z.ai, Xiaomi, MiniMax, Moonshot, Alibaba Qwen). Every request is logged, categorized, costed, and routed against rules you control.Documentation Index
Fetch the complete documentation index at: https://docs.axiomancer.io/llms.txt
Use this file to discover all available pages before exploring further.
Get started
Sign up, mint your first key, and proxy your first request.
Marketing site
Pricing, providers, and the savings-share model.
What RouteShift covers
- Virtual API keys —
sk-proxy-keys with per-key allowed models, RPM/TPM limits, monthly budgets, expiry, metadata tags, rotation with grace periods, and a full audit trail. - Routing & fallbacks — rules that pick a target model (or a chain), with model aliases, weighted load balancing across N provider keys, latency-based and least-busy strategies, and automatic 5xx/429 cooldowns.
- Observability — every request is logged with tokens, cost, latency, and an inferred activity category. Sessions are stitched together for one-shot vs. retry-rate analysis and side-by-side model comparison.
- Optimize findings — a static-analysis pass over recent traffic that flags duplicate prompts, oversized system prompts, model-mismatch waste, and other cost leaks.
- Spend & savings — pricing pulled directly from the LiteLLM JSON catalog, monthly per-key budgets with projected-overrun alerts, per-key cost attribution in the billing breakdown, and a 3% savings-share monetization model.
How it fits the stack
RouteShift is OpenAI-API and Anthropic-API compatible, so any SDK that targets either upstream works without code changes — just point the base URL athttps://proxy.routeshift.io and use a sk-proxy- key. BYOK is the default: you bring provider keys, RouteShift never resells inference.