Introduction
This documentation provides detailed information about the various ways you can access and interact with Obiguard - a robust AI gateway designed to simplify and enhance your experience with Large Language Models (LLMs) like OpenAI’s GPT models.
Whether you’re integrating directly with OpenAI, using a framework like Langchain or LlamaIndex, or building standalone applications, Obiguard offers a flexible, secure, and efficient way to manage and deploy AI-powered features.
3 Ways to Integrate Obiguard
Obiguard can be accessed through three primary methods, each catering to different use cases and integration requirements:
1. Obiguard SDKs (Python)
Ideal for: standalone applications or when you’re not already using the OpenAI integration. The SDKs are also highly recommended for seamless integration with frameworks like Langchain and LlamaIndex.
The Obiguard SDKs are available in Python and JavaScript, designed to provide a streamlined, code-first approach to integrating LLMs into your applications.
Installing the SDK
Choose the SDK that matches your development environment:
Usage
Once installed, you can use the SDK to make calls to LLMs, manage prompts, handle keys, and more, all through a simple and intuitive API.
2. OpenAI SDK through the Obiguard Gateway
Ideal for: if you’re currently utilizing OpenAI’s Python or Node.js SDKs. By changing the base URL and adding Obiguard-specific headers, you can quickly integrate Obiguard’s features into your existing setup.
Learn more here.
3. REST API
Ideal for: applications that prefer RESTful services. The base URL for all REST API requests is https://gatewawy.Obiguard.ai/v1
, with an
authentication header.
Learn more here.