AI-first, built in-house
Multi-provider AI Gateway with BYOK routing for 6 providers and 100+ models, plus Zy — our exclusive context-aware AI assistant that knows your project's schema, storage, functions, and auth.
// Chat completion — works with any provider
const response = await zm.ai.chat({
model: 'gpt-4o', // or claude-sonnet, gemini-pro, etc.
messages: [
{ role: 'system', content: 'You are a helpful assistant.' },
{ role: 'user', content: 'Explain quantum computing simply.' }
],
max_tokens: 500
})
// Streaming response
const stream = await zm.ai.chat({
model: 'claude-sonnet-4-20250514',
messages: [{ role: 'user', content: 'Write a poem.' }],
stream: true
})
for await (const chunk of stream) {
process.stdout.write(chunk.content)
}
// Prompt Arena — compare models side by side
const arena = await zm.ai.arena({
prompt: 'Summarize this article...',
models: ['gpt-4o', 'claude-sonnet', 'gemini-pro', 'deepseek-chat']
})
// Ask Zy — the context-aware assistant
const zy = await zm.ai.zy({
message: 'How do I query users from my database?'
})
// Returns project-specific guidance with actual endpointsMeet Zy — your project-aware AI
An in-house AI assistant that understands your database schema, storage buckets, auth config, and edge functions. No setup needed — it just works.
Context-Aware Intelligence
Zy auto-gathers your project's database schema, table structure, storage usage, auth config, and function status to give project-specific answers.
Schema-Aware Responses
Ask about your tables, columns, relationships — Zy knows your exact database structure including types, primary keys, and row counts.
Deep-Linked Guidance
Zy doesn't just explain — it generates clickable deep links directly to the relevant dashboard pages for your specific project.
Endpoint Documentation
Get real, ready-to-use API endpoint URLs with your actual project IDs and base URLs — not generic placeholders.
Cached & Fast
5-minute context cache + 1-hour response cache means instant answers for repeated queries. Smart caching, not redundant lookups.
Zero API Keys Needed
Zy runs on platform-level infrastructure. No BYOK required — every user gets access. Rate-limited to 10 req/min for fair usage.
Ask anything about your project
Zy reads your schema, buckets, functions, and auth state — then answers with card-based UI showing exact endpoints, file counts, and dashboard links.
One API, six providers, 100+ models
Bring your own API keys. Route to any model with a single unified endpoint. Full streaming, token tracking, and encrypted key storage.
Multi-LLM Routing
Switch between OpenAI, Anthropic, DeepSeek, Mistral, Llama, Qwen and 100+ models. Change one string to switch providers.
Prompt Arena
Compare up to 4 models side-by-side with a single request. Evaluate quality, latency, and cost across providers simultaneously.
Encrypted Key Vault
Provider API keys are encrypted at rest using AES (Fernet encryption) with rotation-capable master keys. Your keys, fully secured.
Token & Cost Tracking
Track token usage per project, per user, per model. Detailed logs for latency, token counts, and cost per request.
Spend Controls & Hard Caps
Set per-plan token limits. When exceeded, requests auto-block with an upgrade prompt — no surprise bills, no runaway costs.
Native Streaming (SSE)
Full Server-Sent Events support for all providers. Real-time response chunks for chat UIs and streaming workflows.
6 providers, one endpoint
Route to any provider through a single unified API. Bring your own keys.
Two AI systems, one platform
Use the AI Gateway for your app's AI features. Use Zy for developer guidance.
Add provider keys
Store your OpenAI, Anthropic, DeepSeek, Mistral, Llama, or Qwen API keys in the encrypted vault. AES-encrypted at rest.
Route to any model
Single API endpoint for all providers. Switch models by changing one string. Streaming and batch both supported.
Track usage & cost
Per-model token tracking, latency logging, and cost analytics. Set hard caps per project to prevent overruns.
Ask Zy for guidance
Built-in AI assistant that knows your schema, storage, and functions. Project-specific answers with ready-to-use endpoints.
Compare 4 models in one request
Send the same prompt to GPT-4o, Claude, Gemini, and DeepSeek simultaneously. Compare responses, token usage, latency, and cost — pick the best model for each use case.
Full visibility across all providers
Track token usage across all 6 providers in real-time. Per-model breakdowns, cost projections, and automatic hard caps when budgets are reached.
AI-first backend, built in-house
Free tier includes 50K AI tokens + unlimited Zy queries. Pro plan includes 1M tokens/month.