Skip to main content

Configuring Private OpenAI-Compatible Models

Private OpenAI-compatible models allow you to connect and manage your own OpenAI-like endpoints directly from the Model Garden.
This enables full flexibility to integrate models from providers such as Groq, Together AI, Mistral, or any custom OpenAI-compatible deployment, all with workspace-level isolation and ownership.
Once configured, these models appear in your Model Garden alongside public models and can be used seamlessly within Deployments, Experiments, and the AI Gateway.

Supported Capabilities

Private OpenAI-compatible models fully support the same capabilities as standard OpenAI models, including:
  • Chat completion with function calling and structured outputs.
  • Embedding generation for vector search and semantic retrieval.
  • Image generation endpoints (DALL-E compatible).
  • Completion (legacy models and instruction tuning).
  • Vision capabilities when supported by the provider.

Adding a Private OpenAI-Compatible Model

To create a private model, head to the Model Garden in the orq.ai Studio and select Add Model → OpenAI-like.
You’ll see a configuration form where you can define all connection details for your model.
OpenAI Compatible Configuration ModalOpenAI Compatible Models configuration screen

Connection Settings

FieldDescription
TypeSelect the type of model to connect: Chat Completion, Completion, Embedding, or Image.
Base URLEnter the base API URL of your OpenAI-compatible endpoint. For example: https://api.groq.com/openai/v1 or https://api.openai.com/v1.
API KeyYour authentication key for the service. This will be stored securely and used for all subsequent requests.

General Settings

FieldDescription
Model NameThe name that will appear in your Model Garden and across orq.ai. Example: Custom Groq Llama 3.3.
Model IDThe model identifier as defined by your provider. Example: llama-3.3-70b-versatile or gpt-4o-mini.
RegionSelect the deployment region for your model, such as United States or Europe.
Description(Optional) Add a short note about what this model is used for.

Advanced Configuration

FieldDescription
Max TokensMaximum token limit for model outputs.
TemperatureControls randomness in the model output.
Input Price (per 1M tokens)Define the cost per million input tokens for billing and analytics.
Output Price (per 1M tokens)Define the cost per million output tokens.

Saving and Validating Your Configuration

After filling in the configuration form, click Add Model.
orq.ai automatically validates your setup by:
  • Checking endpoint connectivity.
  • Verifying your API key and authentication.
  • Testing supported model capabilities.
Once validated, your model will appear in the Model Garden under the “Private Models” section and can be used immediately in: