What is Edgee?
What is Edgee?
Edgee is a unified AI Gateway that sits between your application and LLM providers like OpenAI, Anthropic, Google, and Mistral. It provides a single API to access 200+ models, with built-in intelligent routing, cost tracking, automatic failovers, and full observability.
How is Edgee different from using LLM APIs directly?
How is Edgee different from using LLM APIs directly?
When you use LLM APIs directly, you’re locked into a single provider’s API format, have no visibility into costs until your bill arrives, no automatic failovers when providers go down, and scattered logs across multiple dashboards.Edgee gives you:
- One API for all providers — switch models with a single line change
- Real-time cost tracking — know exactly what each request costs
- Automatic failovers — when OpenAI is down, Claude takes over seamlessly
- Unified observability — all your AI logs in one place
- Intelligent routing — optimize for cost or performance automatically
Which LLM providers does Edgee support?
Which LLM providers does Edgee support?
Edgee supports all major LLM providers:
- OpenAI (GPT-4, GPT-4o, GPT-3.5, o1, etc.)
- Anthropic (Claude 3.5, Claude 3 Opus/Sonnet/Haiku)
- Google (Gemini Pro, Gemini Ultra)
- Mistral (Mistral Large, Medium, Small)
- Meta (Llama 3.1, Llama 3)
- Cohere (Command R+, Command R)
- AWS Bedrock (all supported models)
- Azure OpenAI (all GPT models)
- And 200+ more models
How much latency does Edgee add?
How much latency does Edgee add?
Edgee adds less than 10ms of latency at the p99 level. Our edge network processes requests at the point of presence closest to your application, minimizing round-trip time.For most AI applications, where LLM inference takes 500ms-5s, this overhead is negligible — typically less than 1-2% of total request time.
How does intelligent routing work?
How does intelligent routing work?
Edgee’s routing engine analyzes each request and selects the optimal model based on your configuration:
- Cost strategy: Routes to the cheapest model capable of handling the request
- Performance strategy: Always uses the fastest, most capable model
- Balanced strategy: Finds the optimal trade-off within your latency and cost budgets
What happens when a provider goes down?
What happens when a provider goes down?
Edgee automatically handles provider failures:
- Detection: We detect issues within seconds through health checks and error monitoring
- Retry: For transient errors, we retry with exponential backoff
- Failover: For persistent issues, we route to your configured backup models
How does cost tracking work?
How does cost tracking work?
Every response from Edgee includes a
cost field showing exactly how much that request cost in USD. You can also:- View aggregated costs by model, project, or time period in the dashboard
- Set budget alerts at 80%, 90%, 100% of your limit
- Receive webhook notifications when thresholds are crossed
- Export usage data for your own analysis
Can I use my own API keys for LLM providers?
Can I use my own API keys for LLM providers?
Yes! Edgee supports two modes:
- Edgee-managed keys: We handle provider accounts and billing. Simple, but you pay our prices (with volume discounts available).
- Bring Your Own Key (BYOK): Use your existing provider API keys. You get your negotiated rates, we just route and observe.
Is Edgee compliant with GDPR, SOC 2?
Is Edgee compliant with GDPR, SOC 2?
Yes. Edgee is designed for compliance-sensitive workloads:
- SOC 2 Type II certified
- GDPR compliant with DPA available
- Regional routing to keep data in specific jurisdictions
How can I contact support?
How can I contact support?
We’re here to help:
- Email: [email protected]
- Discord: Join our community
- GitHub: Open an issue