We've built a never-goes-old system that powers 400B+ tokens and roughly $250,000 in LLM spend every day. We’re now open-sourcing the database produced by that system so you can use it and see exactly how pricing is handled under the hood.
→ portkey.ai/models — searchable pricing directory
→ https://lnkd.in/gKQXHxXq — raw JSON data
Free API, no auth needed
Supporting many AI models is not an integration problem.
It’s a correctness problem.
Pricing changes frequently. Token semantics differ per provider. Capabilities evolve independently (tools, reasoning, thinking, multimodal, embeddings, real-time). New models launch, others get deprecated, often without clear signals.
For an AI Gateway, this directly impacts cost accounting, prompt studio UX, analytics, and governance.
Another obvious challenge is logistics: how do you propagate model and pricing updates across SaaS, OSS, hybrid, and air-gapped deployments without forcing redeployments?
But logistics is the easier part.
What we built at Portkey:
- We decoupled model intelligence from deployments.
- We open-sourced a single repository that defines pricing and token calculation logic, normalized model capabilities per provider, and compatibility metadata required by the prompt engineering studio UI.
This repo is the canonical source of truth.
A GitHub workflow syncs it to an S3-backed registry:
- SaaS deployments consume cached S3 data
- Hybrid deployments fetch via a lightweight API (cached locally)
- Air-gapped deployments bake configs into images and support S3→S3 sync to avoid rebuilds
One commit updates everything, solving distribution.
Why this is only the foundation:
Correctness requires feedback loops.
Keeping this data accurate as the ecosystem moves needs more than config management. The next step is making model intelligence continuously correct and self-updating.
Stay tuned, something more exciting coming soon. 💯
#AIThatNeverBreaks #EnterpriseAIGateway