Radicalbit AI Gateway
The Radicalbit AI Gateway is a centralized access point to the generative AI models used across your organization.
It sits between your AI applications and the underlying models — analyzing traffic, filtering information based on configurable conditions, and recording metrics and events to give you full visibility into every operation. The result is a single, controlled layer that makes your AI applications observable, governed, and cost-efficient.
The Gateway is application-agnostic and supports any model that adheres to the OpenAI standard, making it compatible with the vast majority of AI applications and frameworks.
Three Pillars
The Radicalbit AI Gateway is built on three core pillars that define its long-term vision:
1. Governance & Security
The Gateway lets you govern and secure your AI applications through authenticated access, configurable guardrails, and mechanisms to limit calls to the underlying models:
- Inbound and Outbound Guardrails: text control (contains, starts with, ends with, regex), PII detection and masking, and LLM-as-a-Judge for custom logic evaluation.
- Rate and Token Limits: control the volume and cost of requests at the route level.
- Model Fallbacks and Routing: automatic failover and intelligent traffic distribution across models.
2. Cost Control
The Gateway provides a dedicated UI with a detailed cost dashboard, as well as a set of features designed to actively reduce application costs:
- Cost Dashboard: monitor spending by group and API key.
- Token and Rate Limiting: cap the resources consumed by each application or team.
- Semantic and Exact Caching: avoid redundant model calls by reusing previous responses.
3. Observability
The Gateway gives you all the tools needed to make your AI applications — and the Gateway itself — fully inspectable:
- Detailed Metrics: investigate performance and usage across routes and models.
- Gateway Tracing: end-to-end traces of every request processed by the Gateway.
- Event Notifications: receive alerts based on configured conditions.
Who Is It For
The Radicalbit AI Gateway serves different roles within an organization:
| Role | Responsibilities |
|---|---|
| DevOps | Configure the Gateway, manage infrastructure, set up observability |
| AI Engineers | Build AI applications and integrate them with the Gateway |
| System Administrators | Create users and API keys, manage groups, maintain an administrative view of AI usage |
| Project Managers | Monitor application costs, detect unwanted events, and request technical investigations |
How It Works
The Gateway is designed to be decoupled from your application development. The two are independent processes:
- Build your application — a chat assistant, a RAG pipeline, or any GenAI app. Focus on your core logic without worrying about guardrails, caching, or metrics.
- Create a project and configure the Gateway — every configuration lives inside a project. Write a
config.yamlthat defines your models and routes, then load, approve, and serve it. - Integrate — point your LLM client to the Gateway using:
- the Gateway base URL
- the Gateway API Key (generated from the UI)
project-name/route-nameas the model parameter
All traffic flows through the Gateway from that point on.
Get Started
For Developers
- Quick Start — Up and running in 5 minutes
- Basic Configuration — Essential configuration
- Advanced Configuration — Practical configurations
For Operations
- Monitoring — Observability and exposed metrics
- Telemetry — Gateway traces
- Troubleshooting — Common issues and solutions
Ready to get started? Begin with the Quick Start guide.
Supported Providers
The gateway has native support for OpenAI, Anthropic, Google Gemini, DeepSeek, Mistral, and Azure OpenAI. Any provider exposing an OpenAI-compatible endpoint — Ollama, vLLM, OpenRouter, or an on-premises deployment — also works out of the box. See Model Configuration for the full list and configuration examples. Additional integrations can be developed on request — contact us.