Portkey provides a robust and secure gateway to seamlessly integrate open-source and fine-tuned LLMs from Predibase into your applications. With Portkey, you can leverage powerful features like fast AI gateway, caching, observability, prompt management, and more, while securely managing your LLM API keys through a virtual key system.
Provider Slug. predibase
Using Portkey, you can call your Predibase models in the familar OpenAI-spec and try out your existing pipelines on Predibase fine-tuned models with 2 LOC change.
Install the Portkey SDK in your project using npm or pip:
To use Predibase with Portkey, get your API key from here, then add it to Portkey to create the virtual key.
Predibase offers LLMs like Llama 3, Mistral, Gemma, etc. on its serverless infra that you can query instantly.
Predibase expects your account tenant ID along with the API key in each request. With Portkey, you can send your Tenand ID with the user
param while making your request.
With Portkey, you can send your fine-tune model & adapter details directly with the model
param while making a request.
The format is:
model = <base_model>:<adapter-repo-name/adapter-version-number>
For example, if your base model is llama-3-8b
and the adapter repo name is sentiment-analysis
, you can make a request like this:
Using Portkey, you can easily route to your dedicatedly deployed models as well. Just pass the dedicated deployment name in the model
param:
model = "my-dedicated-mistral-deployment-name"
You can enforce JSON schema for all Predibase models - just set the response_format
to json_object
and pass the relevant schema while making your request. Portkey logs will show your JSON output separately
The complete list of features supported in the SDK are available on the link below.
You’ll find more information in the relevant sections: