One Token. Every model. Drop-in OpenAI SDK.
Point your existing OpenAI SDK at TonBo's base URL and route every request to GPT-5, Claude, Gemini, Grok, Llama or Qwen — same key, same call signature, pay per token.
Drop-in SDK examples
curl https://api.tonboai.com/v1/chat/completions \
-H "Authorization: Bearer $TONBO_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-5",
"messages": [{"role": "user", "content": "Hello"}],
"stream": true
}'OpenAI SDK drop-in
Works with openai-python, openai-node, openai-go and any OpenAI-compatible client. Just change base_url — no rewrite.
One balance, all models
Pay-as-you-go tokens work across every model. No per-provider accounts, no per-vendor invoices — one dashboard, one bill.
Streaming out of the box
Server-Sent Events and chunked responses supported for every model. Build real-time agents without any protocol glue.
High availability
Automatic failover across upstream providers. If one vendor has an outage, TonBo re-routes so your product keeps responding.
Smart model routing
Pin a model explicitly, or let TonBo route by latency / cost / capability. Define routing policies per API key.
AES-256 tunnel
Every request traverses TonBo's encrypted AI tunnel with global smart routing — reliable access from any region, no data retention for training.
Start building
Ship multi-model apps without the plumbing
Sign up, get a Token, paste the base URL — and your app can call every frontier model in under 5 minutes.