The world’s fastest AI Gateway with advanced routing & integrated Guardrails.
Use any of the supported models with a universal API (REST and SDKs)
Save costs and decrease latencies by using a cache
Connect to Remote MCP severs, allowing you to connect external tools and data sources.
Fallback between providers and models for resilience
Route to different targets based on custom conditional checks
Use vision, audio, image generation, and more models
Setup automatic retry strategies
Configure per-strategy circuit protection and failure handling
Load balance between various API Keys to counter rate-limits
Canary test new models in production
Easily handle unresponsive LLM requests
Set usage limits based on costs incurred or tokens used
Set hourly, daily, or per minute rate limits on requests or tokens sent
The various gateway strategies are implemented using Gateway configs. You can read more about configs below.
We’ve open sourced our battle-tested AI gateway to the community. You can run it locally with a single command:
While you’re here, why not give us a star? It helps us a lot!
You can also self-host the gateway and then connect it to Portkey. Please reach out on hello@portkey.ai and we’ll help you set this up!