Edgee AI Gateway is still under active construction. We’re building fast, shipping incrementally, and writing docs in parallel.
Don’t hesitate to contact us if you have any questions or feedback.
What the AI Gateway focuses on
These are the core capabilities we’re designing the gateway around:Token Compression
Reduce input tokens by up to 50% with edge-native compression. Ideal for RAG, long contexts, and agent workflows.
Unified API
One integration that routes across 200+ models from OpenAI, Anthropic, Google, Mistral, and more.
Cost Tracking & Observability
Track token savings, costs, latency, and errors in real-time. Export data for analysis and budgeting.
Intelligent Routing
Policy-based routing, automatic failover, and cost-aware model selection for optimal spend.
Privacy Controls
Configurable logging and retention, plus provider-side ZDR where available.
Edge Tools
Run shared tools at the edge so LLM calls have real capabilities without hard-coding tool glue everywhere.
What to expect (right now)
- A product in progress: features and APIs may evolve as we learn from real production use-cases.
- Clear defaults, configurable controls: the goal is to reduce “LLM glue code” while keeping you in charge.
- Docs expanding quickly: each feature page will get deeper guides, examples, and best practices as we ship.