Skip to main content

Documentation Index

Fetch the complete documentation index at: https://portkey-docs-feat-update-docs-apr-22.mintlify.app/llms.txt

Use this file to discover all available pages before exploring further.

Portkey provides a robust and secure platform to observe, govern, and manage your locally or privately hosted custom models using Triton.
Hereโ€™s the official Triton Inference Server documentation for more details.

Integrating Custom Models with Portkey SDK

1

Expose your Triton Server

Expose your Triton server by using a tunneling service like ngrok or any other way you prefer. You can skip this step if youโ€™re self-hosting the Gateway.
ngrok http 11434 --host-header="localhost:8080"
2

Install the Portkey SDK

npm install --save portkey-ai
3

Initialize Portkey with Triton custom URL

  1. Pass your publicly-exposed Triton server URL to Portkey with customHost
  2. Set target provider as triton.
import Portkey from 'portkey-ai'

const portkey = new Portkey({
    apiKey: "PORTKEY_API_KEY",
    provider: "triton",
    customHost: "http://localhost:8000/v2/models/mymodel" // Your Triton Hosted URL
    Authorization: "AUTH_KEY", // If you need to pass auth
})
More on custom_host here.
4

Invoke Chat Completions

Use the Portkey SDK to invoke chat completions (generate) from your model, just as you would with any other provider:
const chatCompletion = await portkey.chat.completions.create({
    messages: [{ role: 'user', content: 'Say this is a test' }]
});

console.log(chatCompletion.choices);

Next Steps

Explore the complete list of features supported in the SDK:

SDK


Youโ€™ll find more information in the relevant sections:
  1. Add metadata to your requests
  2. Add gateway configs to your requests
  3. Tracing requests
  4. Setup a fallback from triton to your local LLM
Last modified on April 22, 2026