AI Gateway feature

Edge Models

In progress

Run lightweight models at the edge to add “reflexes” before calling an LLM: classification, redaction, enrichment, and routing decisions.

We’re actively building this. Talk to us if you want to help shape the first use cases.

Capabilities

  • Edge-side classification and intent routing
  • PII detection and redaction as a pre-inference step
  • Lightweight enrichment (language detection, summarization, tagging)
  • Policy hooks that can influence provider/model selection

Common use cases

  • Detect “sensitive” prompts and route to providers with stricter data policies
  • Classify requests and pick a cheaper model for low-stakes tasks
  • Auto-redact PII before logging/export and before provider calls
  • Gate tool usage by detecting intent and risk

Lower cost without losing quality

Use small models to compress, filter, or pre-process so large models see only what matters.

Faster p95 latency

Make quick decisions close to the user (and the provider) before the expensive call.

Safer inputs by default

Apply privacy layers (e.g., PII detection/redaction) before prompts are forwarded.

FAQ

Answers reflect current direction and may evolve as the platform ships.

Ship faster

Start with one key. Scale with policies.

Use Edgee’s unified access to get moving quickly, then add routing, budgets, and privacy controls as your AI usage grows.

Contact
Edge Models — Edgee AI Gateway