AI Router
Route your LLM calls through the AI Router with a single base URL change. Zero vendor lock-in: always run on the best model at the lowest cost for your use case.
Observability
Attach the native Orq callback handler to your LangGraph to capture traces for every LLM call, graph node, tool use, and retrieval.
AI Router
Overview
LangChain is a framework for building LLM-powered applications through composable chains, agents, and integrations with external data sources. By connecting LangChain to Orq.ai’s AI Router, you access 300+ models through a single base URL change.Key Benefits
Orq.ai’s AI Router enhances your LangChain applications with:Complete Observability
Track every chain step, tool use, and LLM call with detailed traces
Built-in Reliability
Automatic fallbacks, retries, and load balancing for production resilience
Cost Optimization
Real-time cost tracking and spend management across all your AI operations
Multi-Provider Access
Access 300+ LLMs and 20+ providers through a single, unified integration
Prerequisites
Before integrating LangChain with Orq.ai, ensure you have:- An Orq.ai account and API Key
- Python 3.8 or higher
To set up your API key, see API keys & Endpoints.
Installation
Configuration
Configure LangChain to use Orq.ai’s AI Router viaChatOpenAI with a custom base_url:
Python
base_url: https://api.orq.ai/v2/router
Basic Example
Python
Chains
Build composable chains using LangChain’s pipe operator:Python
Streaming
Python
Model Selection
With Orq.ai, you can use any supported model from 20+ providers:Python
Observability
OrqLangchainCallback is a native LangChain callback handler from the Orq.ai Python SDK. Attach it once to your compiled graph and it automatically captures the full execution hierarchy across every run: graph nodes, LLM calls, tool executions, and retrievals.
Zero configuration
No OpenTelemetry setup, no exporters, no environment variables. Two lines of code and tracing is live.
Full graph visibility
Traces preserve the parent-child structure of your LangGraph so you see exactly which node triggered each LLM call or tool use.
Token usage and costs
Input and output token counts are captured on every LLM call and synced to Orq.ai for cost tracking.
Retrieval tracking
Retrieval events include the query and all returned documents, making RAG pipelines fully inspectable.
Installation
orq-ai-sdk is the Orq.ai Python SDK. See the repository for the full reference and changelog.Basic Example
Python
.with_config({"callbacks": [orq_handler]}) call bakes the handler into the compiled graph so all subsequent invocations are traced automatically without passing callbacks manually each time.
What Gets Traced
| Event | Details captured |
|---|---|
| Graph nodes (chains) | Node name, inputs, outputs, duration |
| LLM calls | Messages, model, token usage, finish reason |
| Tool executions | Tool name, input, output, duration |
| Retrievals | Query, returned documents |
| Agent actions | Action taken, finish output |
Viewing Traces
Traces appear in the Orq.ai Studio under the Traces tab. Each run is captured as a tree reflecting your graph structure: top-level chain spans for each node, with LLM calls, tool executions, and retrievals nested underneath.