AI Gateway + Zy Assistant

AI-first, built in-house

Multi-provider AI Gateway with BYOK routing for 6 providers and 100+ models, plus Zy — our exclusive context-aware AI assistant that knows your project's schema, storage, functions, and auth.

Exclusive

Meet Zy — your project-aware AI

An in-house AI assistant that understands your database schema, storage buckets, auth config, and edge functions. No setup needed — it just works.

Context-Aware Intelligence

Zy auto-gathers your project's database schema, table structure, storage usage, auth config, and function status to give project-specific answers.

Schema-Aware Responses

Ask about your tables, columns, relationships — Zy knows your exact database structure including types, primary keys, and row counts.

Deep-Linked Guidance

Zy doesn't just explain — it generates clickable deep links directly to the relevant dashboard pages for your specific project.

Endpoint Documentation

Get real, ready-to-use API endpoint URLs with your actual project IDs and base URLs — not generic placeholders.

Cached & Fast

5-minute context cache + 1-hour response cache means instant answers for repeated queries. Smart caching, not redundant lookups.

Zero API Keys Needed

Zy runs on platform-level infrastructure. No BYOK required — every user gets access. Rate-limited to 10 req/min for fair usage.

Zy Assistant

Ask anything about your project

Zy reads your schema, buckets, functions, and auth state — then answers with card-based UI showing exact endpoints, file counts, and dashboard links.

Auto-detects schema, storage, auth, functions
Returns visual card-based responses
Generates project-specific endpoints
Deep links to dashboard pages
No API key needed — built into every project
Zy Assistant
CONTEXT-AWARE
Schema: 4 tables detectedAuth: 128 users registeredFunctions: 3 active
U
How do I query users from my database?
AI Gateway

One API, six providers, 100+ models

Bring your own API keys. Route to any model with a single unified endpoint. Full streaming, token tracking, and encrypted key storage.

Multi-LLM Routing

Switch between OpenAI, Anthropic, DeepSeek, Mistral, Llama, Qwen and 100+ models. Change one string to switch providers.

Prompt Arena

Compare up to 4 models side-by-side with a single request. Evaluate quality, latency, and cost across providers simultaneously.

Encrypted Key Vault

Provider API keys are encrypted at rest using AES (Fernet encryption) with rotation-capable master keys. Your keys, fully secured.

Token & Cost Tracking

Track token usage per project, per user, per model. Detailed logs for latency, token counts, and cost per request.

Spend Controls & Hard Caps

Set per-plan token limits. When exceeded, requests auto-block with an upgrade prompt — no surprise bills, no runaway costs.

Native Streaming (SSE)

Full Server-Sent Events support for all providers. Real-time response chunks for chat UIs and streaming workflows.

Providers

6 providers, one endpoint

Route to any provider through a single unified API. Bring your own keys.

OpenAI
Anthropic
DeepSeek
Mistral
Llama
Qwen
Models100+
Providers6
Free Tokens50K/mo
Key EncryptionAES-256
Zy Rate Limit10 req/min
How It Works

Two AI systems, one platform

Use the AI Gateway for your app's AI features. Use Zy for developer guidance.

Add provider keys

Store your OpenAI, Anthropic, DeepSeek, Mistral, Llama, or Qwen API keys in the encrypted vault. AES-encrypted at rest.

1

Route to any model

Single API endpoint for all providers. Switch models by changing one string. Streaming and batch both supported.

2

Track usage & cost

Per-model token tracking, latency logging, and cost analytics. Set hard caps per project to prevent overruns.

3

Ask Zy for guidance

Built-in AI assistant that knows your schema, storage, and functions. Project-specific answers with ready-to-use endpoints.

4
Prompt Arena

Compare 4 models in one request

Send the same prompt to GPT-4o, Claude, Gemini, and DeepSeek simultaneously. Compare responses, token usage, latency, and cost — pick the best model for each use case.

Side-by-side response comparison
Token & cost per model
Latency benchmarks
Quality vs cost evaluation
Prompt Arena — 4 Models
same prompt
PROMPTExplain quantum computing in simple terms.
Quantum computing uses qubits instead of classical bits. While a bit is either 0 or 1, a qubit can be in a superposition of both states simultaneously, enabling parallel processing of vast possibilities.
312
Tokens
1.4s
Latency
$0.004
Cost
Token Usage — All Providers
LIVE
452,300 / 1,000,000 tokens45%
Usage by Model — 6 Providers
OpenAI
158K
Anthropic
113K
DeepSeek
68K
Mistral
54K
Llama
36K
Qwen
23K
Total Spend (this month)$4.82
Budget limit$25.00
Usage Analytics

Full visibility across all providers

Track token usage across all 6 providers in real-time. Per-model breakdowns, cost projections, and automatic hard caps when budgets are reached.

Live token counter across 6 providers
Per-model spend breakdown
Budget limits & auto-cutoff
Latency & error rate logging
Monthly cost projections
FREE TIER INCLUDED

AI-first backend, built in-house

Free tier includes 50K AI tokens + unlimited Zy queries. Pro plan includes 1M tokens/month.