One line of code.
Every AI call governed.
Change your base URL. Identity, guardrails, rate limits, token saving, and routing — all enforced automatically. Works with every provider.
/openai_compatible/OpenAI Compatible/anthropic_messages/Anthropic Messages/vertex_ai/Vertex AI/sdk/v1/checkSDK ModeThree Ways to Integrate
OpenAI-compatible proxy, native Anthropic support, or SDK mode for any provider.
/openai_compatible/OpenAI Compatible
Drop-in replacement for any OpenAI SDK call. Chat completions, embeddings, assistants — all governed.
guardrails.quilr.aiAutomatic routingus.guardrails.quilr.aiUS East (Virginia)in.guardrails.quilr.aiAsia South (Mumbai)One Gateway. Every Connection Governed.
AI systems on the left. Tools and providers on the right. QuilrAI sits in the middle, every LLM call and MCP tool invocation passes through the Decision Engine.
Models
Agents
LLM Gateway + MCP Gateway
Decision Engine
── AI Systems ──
LLM Gateway + MCP Gateway
── LLM Providers ──
── Tools & MCP Servers ──
+ 140 more servers
Any model or agent connects via one base_url change. OpenAI-compatible, Anthropic, Vertex AI, or MCP.
Every call passes through Identity, Guardrails, Guardian Agent, and the Decision Engine. ~40ms overhead.
Approved calls route to 5+ LLM providers or 150+ MCP servers. Automatic failover. Token optimization.
Pipeline Architecture
Every request passes through a multi-stage pipeline. Toggle between LLM and MCP views.
MCP Gateway
150+ managed MCP servers via one URL. Dynamic Tool Calling. Auto-detected agents.
MCP Multiplexing
One single URL for all MCPs. Agents connect to one endpoint — the gateway routes to the right server based on the tool call.
mcp.quilr.ai/mcp/<slug>/Dynamic Tool Calling
Reduces tool selection context from 10-20K tokens to ~200 tokens. Higher accuracy — LLMs pick the right tool without noise.
Web Search MCP
Built-in web search with enterprise security gateway integration. URL filtering enforced through your existing security stack.
MCP Library — 150+ Servers
Tool Risk Categorization
get_filelist_repossearch_issuesread_channelcreate_issuesend_messageupdate_recordpush_commitdelete_repodrop_tableremove_userpurge_cacheAuto-Detected Agents
User-Agent: cursor/*User-Agent: claude-code/*User-Agent: openai-agents/*User-Agent: gemini-cli/*Agents are automatically identified via User-Agent header matching. Per-agent policies, rate limits, and tool access controls apply instantly. Add custom agents with your own keywords.
MCP/AI Portal
Self-service portal for end users to browse available MCPs, connect their accounts via OAuth, and start using tools. Not admin-only — engineers can self-serve.
Auth Mediation
Gateway brokers OAuth tokens — agents never see raw credentials. Modes: OAuth→Token, Token→Token, No Auth→OAuth. Bearer token + mcpuser header, OAuth DCR, OAuth Manual.
Built for Production
Cost control, intelligent routing, prompt management, and custom classifiers — all built in.
Routing Groups
Weighted distribution across providers with automatic fallback
gpt-4oclaude-sonnet-4gpt-4oToken Saving — JSON → TOON
Lossless compression cuts token count by 43%
Prompt Store
Versioned, centralized prompts with variable injection. No code deploys.
{{role}}{{company}}{{format}}{{focus_areas}}{{max_words}}Custom Intents
Train classifiers with examples. Block, monitor, or redact matches.
competitor_mentionblock