Here’s the official Triton Inference Server documentation for more details.
Integrating Custom Models with Obiguard SDK
1
Expose your Triton Server
Expose your Triton server by using a tunneling service like ngrok or any other way you prefer. You can skip this step if you’re self-hosting the Gateway.
2
Install the Obiguard SDK
3
Initialize Obiguard with Triton custom URL
- Pass your publicly-exposed Triton server URL to Obiguard with
customHost
- Set target
provider
astriton
.
custom_host
here.4
Invoke Chat Completions
Use the Obiguard SDK to invoke chat completions (generate) from your model, just as you would with any other provider: