AI Router
Route your LLM calls through the AI Router with a single base URL change. Zero vendor lock-in: always run on the best model at the lowest cost for your use case.
Observability
Attach the native Orq callback handler to your LangGraph to capture traces for every LLM call, graph node, tool use, and retrieval.
AI Router
Overview
LangChain is a framework for building LLM-powered applications through composable chains, agents, and integrations with external data sources. By connecting LangChain to Orq.ai’s AI Router, you access 300+ models through a single base URL change.Key Benefits
Orq.ai’s AI Router enhances your LangChain applications with:Complete Observability
Track every chain step, tool use, and LLM call with detailed traces
Built-in Reliability
Automatic fallbacks, retries, and load balancing for production resilience
Cost Optimization
Real-time cost tracking and spend management across all your AI operations
Multi-Provider Access
Access 300+ LLMs and 20+ providers through a single, unified integration
Prerequisites
Before integrating LangChain with Orq.ai, ensure you have:- An Orq.ai account and API Key
- Python 3.8 or higher
To set up your API key, see API keys & Endpoints.
Installation
Configuration
Configure LangChain to use Orq.ai’s AI Router viaChatOpenAI with a custom base_url:
Python
base_url: https://api.orq.ai/v3/router
Basic Example
Python
Chains
Build composable chains using LangChain’s pipe operator:Python
Streaming
Python
Model Selection
With Orq.ai, you can use any supported model from 20+ providers:Python
Observability
orq_ai_sdk.langchain provides a global setup() function that automatically instruments all LangChain and LangGraph components. Call it once at the top of your application and every LLM call, graph node, tool execution, and retrieval is traced automatically, no callback wiring needed.
Zero configuration
One
setup() call and tracing is live, no callbacks, no OpenTelemetry exporters, no extra wiring.Full graph visibility
Traces preserve the parent-child structure of your LangGraph so you see exactly which node triggered each LLM call or tool use.
Token usage and costs
Input and output token counts are captured on every LLM call and synced to Orq.ai for cost tracking.
Retrieval tracking
Retrieval events include the query and all returned documents, making RAG pipelines fully inspectable.
Installation
orq-ai-sdk is the Orq.ai Python SDK. See the repository for the full reference and changelog.Environment Variables
Set your API keys before running your application:Python
Basic Example
setup() must be called before importing or using any LangChain components. It globally instruments LangChain so that all subsequent graphs and chains are traced automatically.Python
Async Example
Python
graph.ainvoke() instead of graph.invoke() for async execution. The setup() instrumentation works with both sync and async invocations.
OrqLangchainCallback is still available and fully backward compatible. Existing code using the callback handler will continue to work. However, the recommended approach is to use the global setup() function for simpler integration and automatic instrumentation.Viewing Traces
Traces appear in the Orq.ai Studio under the Traces tab. Each run is captured as a tree reflecting your graph structure: top-level chain spans for each node, with LLM calls, tool executions, and retrievals nested underneath.
What Gets Traced
| Event | Details captured |
|---|---|
| Graph nodes (chains) | Node name, inputs, outputs, duration |
| LLM calls | Messages, model, token usage, finish reason |
| Tool executions | Tool name, input, output, duration |
| Retrievals | Query, returned documents |
| Agent actions | Action taken, finish output |