AI Proxy
The Orq AI Proxy is a powerful tool that enables you to access models from OpenAI, Anthropic, Google, AWS, and third-party inference providers using an Unified API.
List of supported models
Chat models
curl https://api.orq.ai/v2/proxy/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $ORQ_API_KEY" \
-d '{
"model": "openai/gpt-4o",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "Hello!"
}
]
}'
from openai import OpenAI
import os
client = OpenAI(
base_url="https://api.orq.ai/v2/proxy",
api_key=os.getenv("ORQ_API_KEY"),
)
completion = client.chat.completions.create(
model="openai/gpt-4o",
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Hello!"}
]
)
print(completion.choices[0].message)
import OpenAI from "openai";
const openai = new OpenAI({
baseURL: 'https://api.orq.ai/v2/proxy',
apiKey: process.env.ORQ_API_KEY,
});
async function main() {
const completion = await openai.chat.completions.create({
messages: [{ role: "system", content: "You are a helpful assistant." }],
model: "openai/gpt-4o"
});
console.log(completion.choices[0]);
}
main();
Provider | Model |
---|---|
Anthropic | anthropic/claude-2.0 |
Anthropic | anthropic/claude-2.1 |
Anthropic | anthropic/claude-3-5-haiku-20241022 |
Anthropic | anthropic/claude-3-5-sonnet-20240620 |
Anthropic | anthropic/claude-3-5-sonnet-20241022 |
Anthropic | anthropic/claude-3-7-sonnet-20250219 |
Anthropic | anthropic/claude-3-7-sonnet-latest |
Anthropic | anthropic/claude-3-haiku-20240307 |
Anthropic | anthropic/claude-3-opus-20240229 |
Anthropic | anthropic/claude-3-sonnet-20240229 |
Anyscale | anyscale/meta-llama/Llama-2-13b-chat-hf |
Anyscale | anyscale/meta-llama/Llama-2-70b-chat-hf |
Anyscale | anyscale/meta-llama/Llama-2-7b-chat-hf |
Anyscale | anyscale/meta-llama/Llama-3-70b-chat-hf |
Anyscale | anyscale/meta-llama/Llama-3-8b-chat-hf |
AWS Bedrock | aws/anthropic.claude-3-5-sonnet-20241022-v2:0 |
AWS Bedrock | aws/anthropic.claude-3-haiku-20240307-v1:0 |
AWS Bedrock | aws/anthropic.claude-3-opus-20240229-v1:0 |
AWS Bedrock | aws/anthropic.claude-3-sonnet-20240229-v1:0 |
AWS Bedrock | aws/eu.anthropic.claude-3-5-sonnet-20240620-v1:0 |
AWS Bedrock | aws/eu.anthropic.claude-3-7-sonnet-20250219-v1:0 |
Azure | azure/gpt-35-turbo |
Azure | azure/gpt-35-turbo-16k |
Azure | azure/gpt-4 |
Azure | azure/gpt-4-32k |
Azure | azure/gpt-4.1 |
Azure | azure/gpt-4.1-mini |
Azure | azure/gpt-4.1-nano |
Azure | azure/gpt-4o |
Azure | azure/gpt-4o-mini |
Azure | azure/llama-2-13b-chat |
Azure | azure/llama-2-70b-chat |
Azure | azure/llama-2-7b-chat |
Azure | azure/llama-3-70B-instruct |
Azure | azure/llama-3-8B-instruct |
Azure | azure/llama-3.1-405B-instruct |
Azure | azure/llama-3.1-70B-instruct |
Azure | azure/llama-3.1-8B |
Azure | azure/o1 |
Azure | azure/o1-mini |
Azure | azure/o3-mini |
Cohere | cohere/command |
Cohere | cohere/command-r |
Cohere | cohere/command-r-plus |
Vertex AI | google/chat-bison-32k@002 |
Vertex AI | google/chat-bison@001 |
Vertex AI | google/chat-bison@002 |
Vertex AI | google/claude-3-5-haiku@20241022 |
Vertex AI | google/claude-3-5-sonnet-v2@20241022 |
Vertex AI | google/claude-3-5-sonnet@20240620 |
Vertex AI | google/claude-3-7-sonnet@20250219 |
Vertex AI | google/claude-3-haiku@20240307 |
Vertex AI | google/claude-3-opus@20240229 |
Vertex AI | google/gemini-1.0-pro-001 |
Vertex AI | google/gemini-1.0-pro-vision-001 |
Vertex AI | google/gemini-1.5-flash-001 |
Vertex AI | google/gemini-1.5-pro-001 |
Vertex AI | google/gemini-2.0-flash-001 |
Vertex AI | google/gemini-2.0-flash-exp |
Vertex AI | google/gemini-2.5-flash-preview-04-17 |
Vertex AI | google/gemini-2.5-pro-preview-03-25 |
Vertex AI | google/meta/llama-3.3-70b-instruct-maas |
Vertex AI | google/mistral-large-2411 |
Google AI | google-ai/gemini-1.0-pro |
Google AI | google-ai/gemini-1.5-flash |
Google AI | google-ai/gemini-1.5-flash-8b-exp-0827 |
Google AI | google-ai/gemini-1.5-flash-exp-0827 |
Google AI | google-ai/gemini-1.5-pro |
Google AI | google-ai/gemini-2.0-flash |
Google AI | google-ai/gemini-2.0-flash-001 |
Google AI | google-ai/gemini-2.0-flash-exp |
Google AI | google-ai/gemini-2.0-flash-lite-001 |
Google AI | google-ai/gemini-2.0-flash-thinking-exp-01-21 |
Google AI | google-ai/gemini-2.0-pro-exp-02-05 |
Google AI | google-ai/gemini-2.5-flash-preview-04-17 |
Google AI | google-ai/gemini-2.5-pro-preview-05-06 |
Groq | groq/deepseek-r1-distill-llama-70b |
Groq | groq/deepseek-r1-distill-qwen-32b |
Groq | groq/gemma-7b-it |
Groq | groq/gemma2-9b-it |
Groq | groq/llama-3.1-8b-instant |
Groq | groq/llama-3.2-11b-vision-preview |
Groq | groq/llama-3.2-1b-preview |
Groq | groq/llama-3.2-3b-preview |
Groq | groq/llama-3.2-90b-vision-preview |
Groq | groq/llama-3.3-70b-specdec |
Groq | groq/llama-3.3-70b-versatile |
Groq | groq/llama-guard-3-8b |
Groq | groq/llama2-70b-4096 |
Groq | groq/llama3-70b-8192 |
Groq | groq/llama3-8b-8192 |
Groq | groq/meta-llama/llama-4-scout-17b-16e-instruct |
Groq | groq/mistral-saba-24b |
Groq | groq/mixtral-8x7b-32768 |
Groq | groq/qwen-2.5-32b |
Groq | groq/qwen-2.5-coder-32b |
Groq | groq/qwen-qwq-32b |
Nvidia | nvidia/meta/llama-3.1-8b-instruct |
OpenAI | openai/chatgpt-4o-latest |
OpenAI | openai/ft:gpt-4o-2024-08-06:contentoo:dundle-nl-v3:BJm7PrnC |
OpenAI | openai/gpt-3.5-turbo |
OpenAI | openai/gpt-3.5-turbo-0125 |
OpenAI | openai/gpt-3.5-turbo-0301 |
OpenAI | openai/gpt-3.5-turbo-0613 |
OpenAI | openai/gpt-3.5-turbo-1106 |
OpenAI | openai/gpt-3.5-turbo-16k |
OpenAI | openai/gpt-3.5-turbo-16k-0613 |
OpenAI | openai/gpt-4 |
OpenAI | openai/gpt-4-0125-preview |
OpenAI | openai/gpt-4-0314 |
OpenAI | openai/gpt-4-0613 |
OpenAI | openai/gpt-4-1106-preview |
OpenAI | openai/gpt-4-32k |
OpenAI | openai/gpt-4-32k-0613 |
OpenAI | openai/gpt-4-turbo |
OpenAI | openai/gpt-4-turbo-2024-04-09 |
OpenAI | openai/gpt-4-turbo-preview |
OpenAI | openai/gpt-4-vision-preview |
OpenAI | openai/gpt-4.1 |
OpenAI | openai/gpt-4.1-2025-04-14 |
OpenAI | openai/gpt-4.1-mini |
OpenAI | openai/gpt-4.1-mini-2025-04-14 |
OpenAI | openai/gpt-4.1-nano |
OpenAI | openai/gpt-4.1-nano-2025-04-14 |
OpenAI | openai/gpt-4.5-preview |
OpenAI | openai/gpt-4.5-preview-2025-02-27 |
OpenAI | openai/gpt-4o |
OpenAI | openai/gpt-4o-2024-05-13 |
OpenAI | openai/gpt-4o-2024-08-06 |
OpenAI | openai/gpt-4o-mini |
OpenAI | openai/gpt-4o-mini-2024-07-18 |
OpenAI | openai/o1 |
OpenAI | openai/o1-2024-12-17 |
OpenAI | openai/o1-mini |
OpenAI | openai/o1-mini-2024-09-12 |
OpenAI | openai/o1-preview |
OpenAI | openai/o1-preview-2024-09-12 |
OpenAI | openai/o3 |
OpenAI | openai/o3-2025-04-16 |
OpenAI | openai/o3-mini |
OpenAI | openai/o3-mini-2025-01-31 |
OpenAI | openai/o4-mini |
OpenAI | openai/o4-mini-2025-04-16 |
Perplexity | perplexity/r1-1776 |
Perplexity | perplexity/sonar |
Perplexity | perplexity/sonar-deep-research |
Perplexity | perplexity/sonar-pro |
Perplexity | perplexity/sonar-reasoning |
Perplexity | perplexity/sonar-reasoning-pro |
Together AI | togetherai/deepseek-ai/deepseek-llm-67b-chat |
Together AI | togetherai/deepseek-ai/DeepSeek-R1 |
Together AI | togetherai/deepseek-ai/DeepSeek-V3 |
Together AI | togetherai/meta-llama/Llama-3.3-70B-Instruct-Turbo |
Together AI | togetherai/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8 |
Together AI | togetherai/meta-llama/Llama-4-Scout-17B-16E-Instruct |
Together AI | togetherai/meta-llama/Meta-Llama-Guard-3-8B |
Completion models
curl https://api.orq.ai/v2/proxy/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $ORQ_API_KEY" \
-d '{
"model": "openai/gpt-3.5-turbo-instruct",
"prompt": "Once upon a time",
"max_tokens": 100
}'
from openai import OpenAI
import os
client = OpenAI(
base_url="https://api.orq.ai/v2/proxy",
api_key=os.getenv("ORQ_API_KEY"),
)
completion = client.completions.create(
model="openai/gpt-3.5-turbo-instruct",
prompt="Once upon a time",
max_tokens=100
)
print(completion.choices[0].text)
import OpenAI from "openai";
const openai = new OpenAI({
baseURL: 'https://api.orq.ai/v2/proxy',
apiKey: process.env.ORQ_API_KEY,
});
async function main() {
const completion = await openai.completions.create({
model: "openai/gpt-3.5-turbo-instruct",
prompt: "Once upon a time",
max_tokens: 100
});
console.log(completion.choices[0].text);
}
main();
Provider | Model |
---|---|
Anthropic | anthropic/claude-instant-1.2 |
AWS Bedrock | aws/ai21.j2-mid-v1 |
AWS Bedrock | aws/ai21.j2-ultra-v1 |
AWS Bedrock | aws/amazon.titan-text-express-v1 |
AWS Bedrock | aws/anthropic.claude-instant-v1 |
AWS Bedrock | aws/anthropic.claude-v2 |
AWS Bedrock | aws/anthropic.claude-v2:1 |
AWS Bedrock | aws/cohere.command-light-text-v14 |
AWS Bedrock | aws/cohere.command-text-v14 |
Azure | azure/gpt-35-turbo-instruct |
Azure | azure/llama-2-13b |
Azure | azure/llama-2-70b |
Azure | azure/llama-2-7b |
Cohere | cohere/command |
Cohere | cohere/command-light |
Vertex AI | google/text-bison-32k@002 |
Vertex AI | google/text-bison@001 |
Vertex AI | google/text-bison@002 |
Vertex AI | google/text-unicorn@001 |
OpenAI | openai/gpt-3.5-turbo-instruct |
Embedding models
curl https://api.orq.ai/v2/proxy/embeddings \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $ORQ_API_KEY" \
-d '{
"model": "openai/text-embedding-3-small",
"input": "Hello world"
}'
from openai import OpenAI
import os
client = OpenAI(
base_url="https://api.orq.ai/v2/proxy",
api_key=os.getenv("ORQ_API_KEY"),
)
embedding = client.embeddings.create(
model="openai/text-embedding-3-small",
input="Hello world"
)
print(embedding.data[0].embedding)
import OpenAI from "openai";
const openai = new OpenAI({
baseURL: 'https://api.orq.ai/v2/proxy',
apiKey: process.env.ORQ_API_KEY,
});
async function main() {
const embedding = await openai.embeddings.create({
model: "openai/text-embedding-3-small",
input: "Hello world"
});
console.log(embedding.data[0].embedding);
}
main();
Provider | Model |
---|---|
Azure | azure/text-embedding-3-small |
Azure | azure/text-embedding-ada-002 |
Cohere | cohere/embed-english-light-v3.0 |
Cohere | cohere/embed-english-v3.0 |
Cohere | cohere/embed-multilingual-light-v3.0 |
Cohere | cohere/embed-multilingual-v3.0 |
Jina AI | jina/jina-clip-v1 |
Jina AI | jina/jina-clip-v2 |
Jina AI | jina/jina-embeddings-v2-base-code |
Jina AI | jina/jina-embeddings-v2-base-de |
Jina AI | jina/jina-embeddings-v2-base-en |
Jina AI | jina/jina-embeddings-v2-base-es |
Jina AI | jina/jina-embeddings-v2-base-zh |
Jina AI | jina/jina-embeddings-v3 |
OpenAI | openai/text-embedding-3-large |
OpenAI | openai/text-embedding-3-small |
OpenAI | openai/text-embedding-ada-002 |
Image models
curl https://api.orq.ai/v2/proxy/images/generations \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $ORQ_API_KEY" \
-d '{
"model": "openai/dall-e-3",
"prompt": "A beautiful sunset over mountains",
"n": 1,
"size": "1024x1024"
}'
from openai import OpenAI
import os
client = OpenAI(
base_url="https://api.orq.ai/v2/proxy",
api_key=os.getenv("ORQ_API_KEY"),
)
response = client.images.generate(
model="openai/dall-e-3",
prompt="A beautiful sunset over mountains",
n=1,
size="1024x1024"
)
print(response.data[0].url)
import OpenAI from "openai";
const openai = new OpenAI({
baseURL: 'https://api.orq.ai/v2/proxy',
apiKey: process.env.ORQ_API_KEY,
});
async function main() {
const response = await openai.images.generate({
model: "openai/dall-e-3",
prompt: "A beautiful sunset over mountains",
n: 1,
size: "1024x1024"
});
console.log(response.data[0].url);
}
main();
Provider | Model |
---|---|
Azure | azure/dall-e-3 |
FAL | fal/flux-pro/new |
FAL | fal/flux/dev |
FAL | fal/flux/schnell |
Vertex AI | google/imagegeneration@006 |
Vertex AI | google/imagen-3.0-fast-generate-001 |
Vertex AI | google/imagen-3.0-generate-001 |
Leonardo AI | leonardoai/leonard-diffusion-xl |
Leonardo AI | leonardoai/leonard-kino-xl |
Leonardo AI | leonardoai/leonard-lightning-xl |
Leonardo AI | leonardoai/leonard-vision-xl |
OpenAI | openai/dall-e-2 |
OpenAI | openai/dall-e-3 |
OpenAI | openai/gpt-image-1 |
Moderations models
curl https://api.orq.ai/v2/proxy/moderations \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $ORQ_API_KEY" \
-d '{
"model": "openai/text-moderation-latest",
"input": "I want to check if this text is appropriate."
}'
from openai import OpenAI
import os
client = OpenAI(
base_url="https://api.orq.ai/v2/proxy",
api_key=os.getenv("ORQ_API_KEY"),
)
response = client.moderations.create(
model="openai/text-moderation-latest",
input="I want to check if this text is appropriate."
)
print(response.results[0])
import OpenAI from "openai";
const openai = new OpenAI({
baseURL: 'https://api.orq.ai/v2/proxy',
apiKey: process.env.ORQ_API_KEY,
});
async function main() {
const response = await openai.moderations.create({
model: "openai/text-moderation-latest",
input: "I want to check if this text is appropriate.",
});
console.log(response.results[0]);
}
main();
Provider | Model |
---|---|
OpenAI | openai/omni-moderation-2024-09-26 |
OpenAI | openai/omni-moderation-latest |
OpenAI | openai/text-moderation-007 |
OpenAI | openai/text-moderation-latest |
OpenAI | openai/text-moderation-stable |
Reranking models
curl https://api.orq.ai/v2/proxy/rerank \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $ORQ_API_KEY" \
-d '{
"model": "cohere/rerank-english-v3.0",
"query": "What is machine learning?",
"documents": [
"Machine learning is a branch of AI",
"Machine learning uses data to improve",
"AI is changing the world"
]
}'
Provider | Model |
---|---|
Cohere | cohere/rerank-english-v3.0 |
Cohere | cohere/rerank-multilingual-v3.0 |
Cohere | cohere/rerank-v3.5 |
Jina AI | jina/jina-colbert-v2 |
Jina AI | jina/jina-reranker-v1-base-en |
Jina AI | jina/jina-reranker-v1-tiny-en |
Jina AI | jina/jina-reranker-v1-turbo-en |
Jina AI | jina/jina-reranker-v2-base-multilingual |
Speech-to-Text models
curl https://api.orq.ai/v2/proxy/audio/transcriptions \
-H "Authorization: Bearer $ORQ_API_KEY" \
-F file="@/path/to/audio.mp3" \
-F model="openai/whisper-1"
from openai import OpenAI
import os
client = OpenAI(
base_url="https://api.orq.ai/v2/proxy",
api_key=os.getenv("ORQ_API_KEY"),
)
audio_file = open("speech.mp3", "rb")
transcript = client.audio.transcriptions.create(
model="openai/whisper-1",
file=audio_file
)
print(transcript.text)
import OpenAI from "openai";
import fs from "fs";
const openai = new OpenAI({
baseURL: 'https://api.orq.ai/v2/proxy',
apiKey: process.env.ORQ_API_KEY,
});
async function main() {
const transcript = await openai.audio.transcriptions.create({
file: fs.createReadStream("speech.mp3"),
model: "openai/whisper-1",
});
console.log(transcript.text);
}
main();
Provider | Model |
---|---|
Azure | azure/whisper |
OpenAI | openai/whisper-1 |
Text-to-Speech models
curl https://api.orq.ai/v2/proxy/audio/speech \
-H "Authorization: Bearer $ORQ_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "openai/tts-1",
"input": "Hello world!",
"voice": "alloy"
}' --output speech.mp3
from openai import OpenAI
import os
client = OpenAI(
base_url="https://api.orq.ai/v2/proxy",
api_key=os.getenv("ORQ_API_KEY"),
)
response = client.audio.speech.create(
model="openai/tts-1",
voice="alloy",
input="Hello world!"
)
response.stream_to_file("speech.mp3")
import OpenAI from "openai";
import fs from "fs";
const openai = new OpenAI({
baseURL: 'https://api.orq.ai/v2/proxy',
apiKey: process.env.ORQ_API_KEY,
});
async function main() {
const mp3 = await openai.audio.speech.create({
model: "openai/tts-1",
voice: "alloy",
input: "Hello world!",
});
const buffer = Buffer.from(await mp3.arrayBuffer());
await fs.promises.writeFile("speech.mp3", buffer);
}
main();
Provider | Model |
---|---|
Eleven Labs | elevenlabs/eleven_flash_v2 |
Eleven Labs | elevenlabs/eleven_flash_v2_5 |
Eleven Labs | elevenlabs/eleven_multilingual_v2 |
Eleven Labs | elevenlabs/eleven_turbo_v2_5 |
OpenAI | openai/tts-1 |
OpenAI | openai/tts-1-hd |
Text-to-Speech Voices
The following voices are available for Text-to-Speech models:
OpenAI
alloy
: Neutral, versatile voiceecho
: Neutral, soft-spoken voicefable
: Expressive, narrative-focused voiceonyx
: Deep, authoritative voicenova
: Warm, natural voiceshimmer
: Clear, optimistic voice
ElevenLabs
aria
: Neutral, versatile voiceroger
: Deep, authoritative voicesarah
: Warm, friendly voicelaura
: Soft, gentle voicecharlie
: Casual, conversational voicegeorge
: Professional, articulate voicecallum
: Youthful, energetic voiceriver
: Calm, soothing voiceliam
: Clear, confident voicecharlotte
: Elegant, refined voicealice
: Bright, cheerful voicematilda
: Thoughtful, measured voicewill
: Reliable, trustworthy voicejessica
: Engaging, expressive voiceeric
: Authoritative, commanding voicechris
: Friendly, approachable voicebrian
: Mature, distinguished voicedaniel
: Versatile, balanced voicelily
: Sweet, melodious voicebill
: Grounded, authentic voice
Updated 8 days ago