curl --request POST \
--url https://api.orq.ai/v2/prompts \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: application/json' \
--data '
{
"display_name": "<string>",
"path": "Default",
"description": "<string>",
"metadata": {
"language": "English"
},
"prompt_config": {
"messages": [
{
"role": "system",
"content": "<string>",
"tool_calls": [
{
"type": "function",
"function": {
"name": "<string>",
"arguments": "<string>"
},
"id": "<string>",
"index": 123
}
],
"tool_call_id": "<string>"
}
],
"model": "openai/gpt-4o",
"model_parameters": {
"temperature": 0.7,
"maxTokens": 1000
}
},
"prompt": {
"model": "openai/gpt-4o",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant"
},
{
"role": "user",
"content": "What is the weather today?"
}
],
"temperature": 0.7,
"max_tokens": 1000
}
}
'{
"_id": "<string>",
"type": "prompt",
"owner": "<string>",
"domain_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"created": "<string>",
"updated": "<string>",
"display_name": "<string>",
"prompt_config": {
"messages": [
{
"role": "system",
"content": "<string>",
"tool_calls": [
{
"type": "function",
"function": {
"name": "<string>",
"arguments": "<string>"
},
"id": "<string>",
"index": 123
}
],
"tool_call_id": "<string>"
}
],
"stream": true,
"model": "<string>",
"model_db_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"model_type": "chat",
"model_parameters": {
"temperature": 123,
"maxTokens": 123,
"topK": 123,
"topP": 123,
"frequencyPenalty": 123,
"presencePenalty": 123,
"numImages": 123,
"seed": 123,
"format": "url",
"dimensions": "<string>",
"quality": "<string>",
"style": "<string>",
"responseFormat": {
"type": "json_schema",
"json_schema": {
"name": "<string>",
"schema": {},
"description": "<string>",
"strict": true
},
"display_name": "<string>"
},
"photoRealVersion": "v1",
"encoding_format": "float",
"reasoningEffort": "none",
"budgetTokens": 123,
"verbosity": "low",
"thinkingLevel": "low"
},
"provider": "openai",
"integration_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"version": "<string>"
},
"created_by_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"updated_by_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"description": "<string>",
"metadata": {
"language": "English"
}
}curl --request POST \
--url https://api.orq.ai/v2/prompts \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: application/json' \
--data '
{
"display_name": "<string>",
"path": "Default",
"description": "<string>",
"metadata": {
"language": "English"
},
"prompt_config": {
"messages": [
{
"role": "system",
"content": "<string>",
"tool_calls": [
{
"type": "function",
"function": {
"name": "<string>",
"arguments": "<string>"
},
"id": "<string>",
"index": 123
}
],
"tool_call_id": "<string>"
}
],
"model": "openai/gpt-4o",
"model_parameters": {
"temperature": 0.7,
"maxTokens": 1000
}
},
"prompt": {
"model": "openai/gpt-4o",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant"
},
{
"role": "user",
"content": "What is the weather today?"
}
],
"temperature": 0.7,
"max_tokens": 1000
}
}
'{
"_id": "<string>",
"type": "prompt",
"owner": "<string>",
"domain_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"created": "<string>",
"updated": "<string>",
"display_name": "<string>",
"prompt_config": {
"messages": [
{
"role": "system",
"content": "<string>",
"tool_calls": [
{
"type": "function",
"function": {
"name": "<string>",
"arguments": "<string>"
},
"id": "<string>",
"index": 123
}
],
"tool_call_id": "<string>"
}
],
"stream": true,
"model": "<string>",
"model_db_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"model_type": "chat",
"model_parameters": {
"temperature": 123,
"maxTokens": 123,
"topK": 123,
"topP": 123,
"frequencyPenalty": 123,
"presencePenalty": 123,
"numImages": 123,
"seed": 123,
"format": "url",
"dimensions": "<string>",
"quality": "<string>",
"style": "<string>",
"responseFormat": {
"type": "json_schema",
"json_schema": {
"name": "<string>",
"schema": {},
"description": "<string>",
"strict": true
},
"display_name": "<string>"
},
"photoRealVersion": "v1",
"encoding_format": "float",
"reasoningEffort": "none",
"budgetTokens": 123,
"verbosity": "low",
"thinkingLevel": "low"
},
"provider": "openai",
"integration_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"version": "<string>"
},
"created_by_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"updated_by_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
"description": "<string>",
"metadata": {
"language": "English"
}
}Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
The prompt’s name, meant to be displayable in the UI.
128Entity storage path in the format: project/folder/subfolder/...
The first element identifies the project, followed by nested folders (auto-created as needed).
With project-based API keys, the first element is treated as a folder name, as the project is predetermined by the API key.
"Default"
The prompt’s description, meant to be displayable in the UI. Use this field to optionally store a long form explanation of the prompt for your own purpose
Show child attributes
A list of use cases that the prompt is meant to be used for. Use this field to categorize the prompt for your own purpose
Agents simulations, Agents, API interaction, Autonomous Agents, Chatbots, Classification, Code understanding, Code writing, Conversation, Documents QA, Evaluation, Extraction, Multi-modal, Self-checking, Sentiment analysis, SQL, Summarization, Tagging, Translation (document), Translation (sentences) The language that the prompt is written in. Use this field to categorize the prompt for your own purpose
Chinese, Dutch, English, French, German, Russian, Spanish [DEPRECATED]. Please use the prompt property instead. The current prompt_config will keep working but it will be deprecated in future versions. Configuration for the prompt including model and messages.
Show child attributes
Array of messages that make up the conversation.
Show child attributes
The role of the prompt message
system, assistant, user, exception, tool, prompt, correction, expected_output The contents of the user message. Either the text content of the message or an array of content parts with a defined type, each can be of type text or image_url when passing in images. You can pass multiple images by adding multiple image_url content parts. Can be null for tool messages in certain scenarios.
Show child attributes
function Model ID used to generate the response, like openai/gpt-4o or google/gemini-2.5-pro. The full list of models can be found at https://docs.orq.ai/docs/ai-gateway-supported-models. Only chat models are supported.
"openai/gpt-4o"
Optional model parameters like temperature and maxTokens.
Show child attributes
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on image models.
Best effort deterministic seed for the model. Currently only OpenAI models support these
Only supported on image models.
url, b64_json, text, json_object Only supported on image models.
Only supported on image models.
Only supported on image models.
An object specifying the format that the model must output.
Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema
Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.
Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.
Show child attributes
json_schema Show child attributes
The version of photoReal to use. Must be v1 or v2. Only available for leonardoai provider
v1, v2 The format to return the embeddings
float, base64 Constrains effort on reasoning for reasoning models. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
none, disable, minimal, low, medium, high Gives the model enhanced reasoning capabilities for complex tasks. A value of 0 disables thinking. The minimum budget tokens for thinking are 1024. The Budget Tokens should never exceed the Max Tokens parameter. Only supported by Anthropic
Controls the verbosity of the model output.
low, medium, high The level of thinking to use for the model. Only supported by Google AI
low, high { "temperature": 0.7, "maxTokens": 1000 }
Prompt configuration with model and messages. Either this field or prompt_config must be provided.
Show child attributes
Array of messages that make up the conversation. Each message has a role (system, user, assistant, or tool) and content.
Developer-provided instructions that the model should follow, regardless of messages sent by the user.
Show child attributes
The role of the messages author, in this case system.
system The contents of the system message.
An optional name for the participant. Provides the model information to differentiate between participants of the same role.
[
{
"role": "system",
"content": "You are a helpful assistant"
},
{
"role": "user",
"content": "What is the weather today?"
}
]
Model ID used to generate the response, like openai/gpt-4o or anthropic/claude-3-5-sonnet-20241022. The full list of models can be found at https://docs.orq.ai/docs/ai-gateway-supported-models. Only chat models are supported.
"openai/gpt-4o"
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.
0 <= x <= 2[Deprecated]. The maximum number of tokens that can be generated in the chat completion. This value can be used to control costs for text generated via API.
This value is now deprecated in favor of max_completion_tokens, and is not compatible with o1 series models.
{
"model": "openai/gpt-4o",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant"
},
{
"role": "user",
"content": "What is the weather today?"
}
],
"temperature": 0.7,
"max_tokens": 1000
}
Prompt created.
A prompt entity with configuration, metadata, and versioning.
prompt The prompt’s name, meant to be displayable in the UI.
128A list of messages compatible with the openAI schema
Show child attributes
Show child attributes
The role of the prompt message
system, assistant, user, exception, tool, prompt, correction, expected_output The contents of the user message. Either the text content of the message or an array of content parts with a defined type, each can be of type text or image_url when passing in images. You can pass multiple images by adding multiple image_url content parts. Can be null for tool messages in certain scenarios.
Show child attributes
function The id of the resource
The modality of the model
chat, completion, embedding, image, tts, stt, rerank, moderation, vision Model Parameters: Not all parameters apply to every model
Show child attributes
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on chat and completion models.
Only supported on image models.
Best effort deterministic seed for the model. Currently only OpenAI models support these
Only supported on image models.
url, b64_json, text, json_object Only supported on image models.
Only supported on image models.
Only supported on image models.
An object specifying the format that the model must output.
Setting to { "type": "json_schema", "json_schema": {...} } enables Structured Outputs which ensures the model will match your supplied JSON schema
Setting to { "type": "json_object" } enables JSON mode, which ensures the message the model generates is valid JSON.
Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length.
Show child attributes
json_schema Show child attributes
The version of photoReal to use. Must be v1 or v2. Only available for leonardoai provider
v1, v2 The format to return the embeddings
float, base64 Constrains effort on reasoning for reasoning models. Reducing reasoning effort can result in faster responses and fewer tokens used on reasoning in a response.
none, disable, minimal, low, medium, high Gives the model enhanced reasoning capabilities for complex tasks. A value of 0 disables thinking. The minimum budget tokens for thinking are 1024. The Budget Tokens should never exceed the Max Tokens parameter. Only supported by Anthropic
Controls the verbosity of the model output.
low, medium, high The level of thinking to use for the model. Only supported by Google AI
low, high openai, groq, cohere, azure, aws, google, google-ai, huggingface, togetherai, perplexity, anthropic, leonardoai, fal, nvidia, jina, elevenlabs, litellm, cerebras, openailike, bytedance, mistral, deepseek, contextualai, moonshotai The ID of the integration to use
The prompt’s description, meant to be displayable in the UI. Use this field to optionally store a long form explanation of the prompt for your own purpose
Show child attributes
A list of use cases that the prompt is meant to be used for. Use this field to categorize the prompt for your own purpose
Agents simulations, Agents, API interaction, Autonomous Agents, Chatbots, Classification, Code understanding, Code writing, Conversation, Documents QA, Evaluation, Extraction, Multi-modal, Self-checking, Sentiment analysis, SQL, Summarization, Tagging, Translation (document), Translation (sentences) The language that the prompt is written in. Use this field to categorize the prompt for your own purpose
Chinese, Dutch, English, French, German, Russian, Spanish Was this page helpful?