Here’s a list of all model architectures supported on vLLM.
Integrating Custom Models with Obiguard SDK
Expose your vLLM Server
Expose your vLLM server by using a tunneling service like ngrok or any other way you prefer. You can skip this step if you’re self-hosting the Gateway.
Initialize Obiguard with vLLM custom URL
- Pass your publicly-exposed vLLM server URL to Obiguard with
customHost(by default, vLLM is onhttp://localhost:8000/v1) - Set target
providerasopenaisince the server follows OpenAI API schema.
- Python SDK
custom_host here.Using Virtual Keys
Virtual Keys serve as Obiguard’s unified authentication system for all LLM interactions, simplifying the use of multiple providers and Obiguard features within your application. For self-hosted LLMs, you can configure custom authentication requirements including authorization keys, bearer tokens, or any other headers needed to access your model.- Navigate to Virtual Keys in your Obiguard dashboard
- Click “Add Key” and enable the “Local/Privately hosted provider” toggle
- Configure your deployment:
- Select the matching provider API specification (typically
OpenAI) - Enter your model’s base URL in the
Custom Hostfield - Add required authentication headers and their values
- Select the matching provider API specification (typically
- Click “Create” to generate your virtual key
- Python SDK

