Managing model usage across providers used to mean juggling separate billing accounts and API keys. Now you can purchase credits directly in Orq and access 250+ models across 20+ providers with a single API key.
What’s Available:

- Purchase credits starting at $5 — buy credits directly in the platform and immediately unlock access to OpenAI, Anthropic, Google, Mistral, Meta, and 15+ other providers without configuring individual API keys.
- Unified usage tracking — monitor credit consumption in real-time with detailed transaction history across all API keys, so you know exactly where spend is going.
- Automatic top-up — configure balance thresholds and credits replenish automatically when you hit them, preventing service interruption mid-request.
- VAT-compliant invoicing — get proper invoices with VAT included for international purchases, not just payment receipts.
Learn how to purchase and manage credits in the AI Router Credits Documentation.
When agent behavior changes in production, you need to know what happened, when, and which version was running. Activity tracking gives you a complete audit trail, version control lets you invoke specific versions programmatically, and environments let you manage agent lifecycle from development to production.
What’s Tracked:

- Track all activities across agents and related entities with full audit trail support showing what changed, when, and by whom.
- Version-level invocation tracking — invoke and compare specific agent versions to measure performance across iterations and roll back when needed.
- Custom environments — create and tag environments like staging, production, or testing to manage and promote agents across your development lifecycle.
Learn more in the Agent Studio Documentation.
Enterprise teams shouldn’t have to manage separate credentials for every tool. Single sign-on lets your team authenticate with the accounts they already use.
What’s New:

- Sign in with Microsoft — authenticate using Azure AD or Entra ID.
- Sign in with Okta — authenticate with Okta as your identity provider.
- Centralized access management — grant and revoke workspace access directly through your identity provider instead of managing users manually in Orq.ai.
Learn how to configure SSO in the Enterprise SSO Documentation.
Running multiple deployment variants used to mean choosing between shared configuration (risky for testing) or duplicating entire deployments (messy to manage). Now you can configure behavior independently for each variant.What’s New:
- Variant-level configuration — set evaluators, guardrails, caching, TTL settings, and security masking controls independently per variant for safe A/B testing and progressive rollouts without affecting other versions.
Learn more in the Deployment Variants Documentation.
Connecting to authenticated MCP servers used to require workarounds. Now you can configure custom headers directly on MCP tools.What’s New:
- Custom header configuration per tool — add authentication tokens or API keys as custom headers when configuring MCP tools to connect securely to private or enterprise MCP servers.
Learn more in the MCP Tools Documentation.
Large knowledge bases return too much irrelevant context. Metadata filtering lets you narrow results before semantic search runs.
What’s New:

- Add, edit, and view metadata in UI — manage key-value metadata on knowledge base chunks via API or UI (previously API-only) for better organization by topic, source, client, or any custom attribute.
- Filter by metadata before query execution — narrow results by, for example, department, project, version, or customer segment before semantic search runs to retrieve only relevant context.
Learn more in the Knowledge Base Documentation.
Running better experiments requires both structured review workflows and execution transparency. Evaluatorq-sourced experiments now include a dedicated UI with side-by-side comparison, while agent experiments expose complete behavior details showing every tool call, parameter, and outcome.
What’s New:

- Full Evaluatorq support — run Evaluatorq-sourced experiments from code and get dedicated review tabs and export functionality in the UI for structured human feedback workflows.
- Complete agent messages — view every message and tool call from agents run via the Orq.ai experiments UI, including parameter configuration, tool responses, and the complete conversation flow.
Learn more in the Experiments Documentation.
Workspace management and navigation improvements.What’s New:
- Updated members page — redesigned layout aligned with the modern design system.
- Improved environment configuration — better interface for setting up and managing environments.
- Keyboard shortcuts — navigate between applications using keyboard shortcuts in the top-left app switcher.
- Model Garden filtering — filter models by enabled API keys to see only what you can actually use.
Learn more in the Workspace Settings Documentation.
Platform-level upgrades for reliability and structured outputs.What’s New:
- Improved load balancing — better request distribution for higher reliability under traffic spikes.
- Sonar structured outputs — Perplexity’s Sonar models now support structured output mode for reliable JSON generation.
Explore models in the Model Garden or via the AI Router.
Anthropic’s high-intelligence model with advanced reasoning, vision, and PDF processing capabilities.Model Specs:
- 200K context window with 128K maximum output tokens
- Tool calling, vision, PDF processing — full multimodal support with function calling and streaming
- Advanced reasoning — designed for complex, multi-step tasks requiring deep analysis
- $5.00 per 1M input tokens
- $25.00 per 1M output tokens
Explore Claude Opus 4.6 in the Model Garden or via the AI Router.
MiniMax expands with new M2.5 variants hosted in Singapore, offering strong reasoning, coding, and tool-calling capabilities.New Models:
- MiniMax-M2.5 — advanced foundation model with 204,800 context window and 204,800 max output tokens for reasoning and agent workflows. $0.15/$1.20 per 1M tokens.
- MiniMax-M2.5-lightning — speed-optimized version of M2.5 with the same 204,800 context window and output limits. $0.30/$2.40 per 1M tokens.
- minimax-m2, minimax-m2-1, minimax-m2-1-lightning, minimax-m2-5, minimax-m2-5-lightning, minimax-m2-her
Explore MiniMax models in the Model Garden or via the AI Router.
OpenAI’s next-generation coding and multimodal models with extended context and reasoning support.New Models:
- GPT-5.3 Codex — coding-optimized model with 400K context window and 128K max output tokens. Supports tool calling, JSON mode, vision, and streaming. $1.75/$14.00 per 1M tokens.
- GPT-5.3 Codex Spark — faster variant with same pricing, 128K context window and 32K max output tokens, optimized for lightweight coding workflows.
Explore GPT-5.3 models in the Model Garden or via the AI Router.
Z.ai’s 744B parameter Mixture-of-Experts model (40B active parameters) designed for agentic engineering.Model Specs:
- 200K context window with 128K maximum output tokens
- Enhanced coding and reasoning — optimized for agentic workflows with improved tool capabilities
- Tool calling support — full function calling and streaming support
- $1.00 per 1M input tokens
- $3.20 per 1M output tokens
Explore GLM-5 in the Model Garden or via the AI Router.