Provider Slug.
googleObiguard SDK Integration with Google Gemini Models
Obiguard provides a consistent API to interact with models from various providers. To integrate Google Gemini with Obiguard:1. Install the Obiguard SDK
Add the Obiguard SDK to your application to interact with Google Gemini’s API through Obiguard’s gateway.- Python SDK
2. Initialize Obiguard with the Virtual Key
To use Gemini with Obiguard, get your API key from here, then add it to Obiguard to create the virtual key.- Python SDK
3. Invoke Chat Completions with Google Gemini
Use the Obiguard instance to send requests to Google Gemini. You can also override the virtual key directly in the API call if needed.- Python SDK
Obiguard supports the
system_instructions parameter for Google Gemini 1.5 - allowing you to control the behavior and output of your Gemini-powered applications with ease.Simply include your Gemini system prompt as part of the {"role":"system"} message within the messages array of your request body.
Obiguard Gateway will automatically transform your message to ensure seamless compatibility with the Google Gemini API.Function Calling
Obiguard supports function calling mode on Google’s Gemini Models.Document, Video, Audio Processing with Gemini
Gemini supports attachingmp4, pdf, jpg, mp3, wav, etc. file types to your messages.
Gemini Docs:
- Python SDK
- cURL
Python
url field, like "url": "gs://cloud-samples-data/video/animals.mp4".
Your URL should have the file extension, this is used for inferring
MIME_TYPE which is a required parameter for prompting Gemini models with files.Sending base64 Image
Here, you can send thebase64 image data along with the url field too:
Grounding with Google Search
Vertex AI supports grounding with Google Search. This is a feature that allows you to ground your LLM responses with real-time search results. Grounding is invoked by passing thegoogle_search tool (for newer models like gemini-2.0-flash-001), and google_search_retrieval (for older models like gemini-1.5-flash) in the tools array.
If you mix regular tools with grounding tools, vertex might throw an error saying only one tool can be used at a time.
Extended Thinking (Reasoning Models) (Beta)
The assistants thinking response is returned in the
response_chunk.choices[0].delta.content_blocks array, not the response.choices[0].message.content string.gemini-2.5-flash-preview-04-17 gemini-2.5-flash-preview-04-17 support extended thinking.
This is similar to openai thinking, but you get the model’s reasoning as it processes the request as well.
Note that you will have to set strict_open_ai_compliance=False in the headers to use this feature.
Single turn conversation
- Python SDK
- OpenAI SDK
- cURL
Python
To disable thinking for gemini models like
gemini-2.5-flash-preview-04-17, you are required to explicitly set budget_tokens to 0.Gemini grounding mode may not work via Obiguard SDK. Contact [email protected] for assistance.

