Unified AI Gateway
One endpoint for all your AI models
Website | Documentation | Dashboard | Chat
VaultGate is a production-ready API gateway that unifies 40+ AI providers into a single OpenAI-compatible endpoint. Access models from OpenAI, Anthropic, Google, Mistral, DeepSeek, Meta, Groq, and more through one consistent API.
- OpenAI-Compatible API: Drop-in replacement for OpenAI's API format
- 100+ AI Models: Access text and image generation models from major providers
- Unified Endpoint: Single API for chat completions and image generation
- Real-time Streaming: Server-sent events for streaming responses
- API Key Management: Create keys with rate limits, token quotas, and model restrictions
- Interactive Playground: Test models directly in the browser
- Usage Analytics: Track requests, tokens, and costs per key
- Smart Failover: Automatic routing when primary providers are unavailable
- Visit endpoint.vaultgate.dev/auth
- Click "Register" to create a new account
- Enter your email, username, and password
- After registration, you will be redirected to the dashboard
- Navigate to the "Balance" section in the sidebar
- Add credits to your account using the available payment methods
- Credits are used to pay for API requests based on token usage
- You only pay for what you use, no subscriptions required
- Go to the "API Keys" page from the sidebar
- Click "Create API Key"
- Enter a name for your key (e.g., "My Application", "Development", "Production")
- Configure optional settings:
- Rate Limit: Maximum requests per minute (default: 60)
- Token Quota: Maximum tokens allowed for this key
- Model Restrictions: Limit which models this key can access
- Click "Create" and copy your API key immediately
- Your key starts with
vg_and will not be shown again after creation
- Visit the "Models" page to see all available AI models
- Filter models by type:
- Text: Chat and completion models (GPT-4, Claude, Gemini, etc.)
- Image: Image generation models (DALL-E, Flux, Midjourney, etc.)
- Hybrid: Models that support both text and image
- Filter by provider: OpenAI, Anthropic, Google, Mistral, Meta, and more
- Note the model ID for use in your API requests (e.g.,
gpt-4o-text,claude-3-5-sonnet-text)
Before integrating into your application, test models in the browser:
- Go to the "Playground" page
- Select a model from the dropdown
- Choose which API key to use for the request
- Adjust settings:
- Temperature: Controls randomness (0 = deterministic, 2 = creative)
- Max Tokens: Limit response length
- Streaming: Enable real-time response streaming
- Type your message and send
- View response time, token usage, and costs
- Export conversations or copy generated code
Use the unified endpoint with your API key:
Chat Completion:
curl https://api.vaultgate.dev/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer vg_your_api_key" \
-d '{
"model": "gpt-4o-text",
"messages": [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Hello!"}
],
"stream": true
}'Image Generation:
curl https://api.vaultgate.dev/v1/images/generations \
-H "Content-Type: application/json" \
-H "Authorization: Bearer vg_your_api_key" \
-d '{
"model": "dall-e-3-image",
"prompt": "A futuristic city at sunset",
"n": 1,
"size": "1024x1024"
}'- Visit the "Logs" page to see all API requests
- View details for each request:
- Model used
- Tokens consumed (prompt + completion)
- Response time
- Status (success/error)
- Track spending and usage patterns over time
- Set up alerts when approaching quota limits
1. Register Account --> endpoint.vaultgate.dev/auth
2. Add Credits --> Balance page
3. Create API Key --> API Keys page (save your vg_* key)
4. Browse Models --> Models page (find model IDs)
5. Test in Playground --> Playground page (optional but recommended)
6. Integrate API --> Use https://api.vaultgate.dev/v1
7. Monitor Usage --> Logs page
https://api.vaultgate.dev/v1
All requests require a Bearer token in the Authorization header:
Authorization: Bearer vg_your_api_key
API keys are created from the dashboard and start with vg_. Each key can have its own rate limits and permissions.
| Method | Endpoint | Description |
|---|---|---|
| GET | /v1/models | List all available models |
| POST | /v1/chat/completions | Generate chat completions (text) |
| POST | /v1/images/generations | Generate images from text prompts |
Model IDs include a type suffix to indicate their capabilities:
-textfor chat and completion models-imagefor image generation models-hybridfor models that support both text and image
Examples:
gpt-4o-text- OpenAI GPT-4o for chatclaude-3-5-sonnet-text- Anthropic Claude 3.5 Sonnetdall-e-3-image- OpenAI DALL-E 3 for imagesgemini-1.5-pro-hybrid- Google Gemini with multimodal support
{
"model": "gpt-4o-text",
"messages": [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Your message here"}
],
"temperature": 0.7,
"max_tokens": 1000,
"stream": true
}| Parameter | Type | Description |
|---|---|---|
| model | string | Required. Model ID from /v1/models |
| messages | array | Required. Array of message objects with role and content |
| temperature | number | Sampling temperature 0-2 (default: 1) |
| max_tokens | integer | Maximum tokens to generate |
| stream | boolean | Enable streaming responses (default: false) |
| top_p | number | Nucleus sampling parameter 0-1 |
{
"model": "dall-e-3-image",
"prompt": "A description of the image you want",
"n": 1,
"size": "1024x1024"
}| Parameter | Type | Description |
|---|---|---|
| model | string | Required. Image model ID |
| prompt | string | Required. Text description of the image |
| n | integer | Number of images to generate (default: 1) |
| size | string | Image dimensions (e.g., "1024x1024") |
| Status | Description |
|---|---|
| 400 | Bad Request - Invalid parameters |
| 401 | Unauthorized - Invalid or missing API key |
| 402 | Payment Required - Insufficient credits |
| 404 | Not Found - Model not available |
| 429 | Too Many Requests - Rate limit exceeded |
| 503 | Service Unavailable - Upstream provider error |
Rate limits are configured per API key. Default is 60 requests per minute. You can view and adjust limits in the API Keys page. When exceeded, you receive a 429 status code.
| Provider | Text Models | Image Models |
|---|---|---|
| OpenAI | GPT-4o, GPT-4, GPT-3.5 | DALL-E 3 |
| Anthropic | Claude 3.5, Claude 3 | - |
| Gemini Pro, Gemini Flash | Imagen | |
| Mistral | Mistral Large, Codestral | - |
| Meta | Llama 3.1, Llama 3 | - |
| DeepSeek | DeepSeek V3, DeepSeek Coder | - |
| Groq | Llama, Mixtral | - |
| Flux | - | Flux Pro, Flux Dev |
View the complete list at endpoint.vaultgate.dev/models
VaultGate works with any OpenAI-compatible SDK. Simply change the base URL:
from openai import OpenAI
client = OpenAI(
api_key="vg_your_api_key",
base_url="https://api.vaultgate.dev/v1"
)
response = client.chat.completions.create(
model="gpt-4o-text",
messages=[{"role": "user", "content": "Hello!"}]
)import OpenAI from 'openai';
const client = new OpenAI({
apiKey: 'vg_your_api_key',
baseURL: 'https://api.vaultgate.dev/v1'
});
const response = await client.chat.completions.create({
model: 'gpt-4o-text',
messages: [{ role: 'user', content: 'Hello!' }]
});curl https://api.vaultgate.dev/v1/chat/completions \
-H "Authorization: Bearer vg_your_api_key" \
-H "Content-Type: application/json" \
-d '{"model": "claude-3-5-sonnet-text", "messages": [{"role": "user", "content": "Hello!"}]}'VaultGate Chat is a full-featured chat interface for interacting with AI models:
- Access at chat.vaultgate.dev
- Connect using your VaultGate API key
- Chat with any available text model
- Generate images with image models
- Conversation history and management
- Model switching within conversations
- Markdown rendering and code syntax highlighting
- Export and share conversations
The landing page provides quick access to VaultGate Chat and the Endpoint dashboard.
The main dashboard landing page with overview and quick actions.
Browse and filter available AI models by provider and type.
Create and manage API keys with rate limits and permissions.
Add credits and track your spending.
Manage your account settings and preferences.
Test any model directly in your browser with real-time streaming.
View all API requests, token usage, and response times.
A full-featured chat interface for interacting with AI models.
Comprehensive API documentation and integration guides.
This repository contains four main components:
VaultGate/
├── VaultGate.portal/ # Landing page (React + Vite)
├── VaultGate.chat/ # Chat interface (React + Vite)
├── VaultGate.endpoint/ # API gateway dashboard and backend
│ ├── gateway/ # Dashboard frontend (Next.js)
│ └── gateway-server/ # API server (Deno)
└── VaultGate.docs/ # Documentation site (React + Vite)
- React 19
- Next.js 16
- TypeScript
- Tailwind CSS
- Framer Motion
- Deno 2.0
- Turso (LibSQL)
- Vercel (Frontend hosting)
- Deno Deploy / Railway (Backend)
- Website: vaultgate.dev
- Documentation: docs.vaultgate.dev
- Dashboard: endpoint.vaultgate.dev
- Chat: chat.vaultgate.dev
- API Base URL: api.vaultgate.dev/v1
How do I get started? Create an account at endpoint.vaultgate.dev, add credits, generate an API key, and start making requests.
Is VaultGate compatible with OpenAI SDKs?
Yes. VaultGate is fully OpenAI-compatible. Change the base URL to https://api.vaultgate.dev/v1 and use your VaultGate API key.
What happens if a provider goes down? VaultGate includes smart failover. If a primary provider is unavailable, requests are automatically routed to backup providers when possible.
How is pricing calculated? You pay per token used. Different models have different costs. Check the Models page for pricing details.
Can I restrict which models an API key can access? Yes. When creating an API key, you can specify which models or model types the key is allowed to use.
Is there a free tier? New accounts receive starter credits to test the platform. After that, you pay only for what you use.
MIT License
Built with care by the VaultGate Team










