OpenAI's fast, lightweight reasoning model optimized for multi-step problem solving at lower cost.
curl https://api.aigateway.sh/v1/chat/completions \
-H "Authorization: Bearer $AIGATEWAY_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "openai/o4-mini",
"messages": [{"role":"user","content":"hello"}],
"stream": true
}'message.reasoning_content (non-streaming) and delta.reasoning_content (streaming). Safe to display or ignore — it's separate from content.max_completion_tokens instead of max_tokens. Our gateway accepts either and translates.{
"model": "openai/o4-mini",
"messages": [
{ "role": "system", "content": "You are a helpful assistant." },
{ "role": "user", "content": "Hello!" }
],
"temperature": 0.7,
"top_p": 0.95,
"max_completion_tokens": 1024,
"stream": false
}{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1776947082,
"model": "openai/o4-mini",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?",
"reasoning_content": "The user asked..." // o4-mini chain-of-thought
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 24,
"completion_tokens": 12,
"total_tokens": 36
}
}"stream": true// 1. Role announcement (first chunk):
data: {"choices":[{"index":0,"delta":{"role":"assistant"},"finish_reason":null}]}
// 2. Reasoning chunks (o4-mini thinks first):
data: {"choices":[{"index":0,"delta":{"reasoning_content":"The user "},"finish_reason":null}]}
data: {"choices":[{"index":0,"delta":{"reasoning_content":"wants..."},"finish_reason":null}]}
// 3. Content chunks (final answer):
data: {"choices":[{"index":0,"delta":{"content":"Hello"},"finish_reason":null}]}
data: {"choices":[{"index":0,"delta":{"content":"!"},"finish_reason":null}]}
// Finish chunk:
data: {"choices":[{"index":0,"delta":{},"finish_reason":"stop"}]}
// Terminator:
data: [DONE]# pip install aigateway-py openai
# aigateway-py adds sub-accounts, evals, replays, jobs, webhook verify.
# openai SDK covers chat — drop-in per our SDK's own guidance.
from openai import OpenAI
client = OpenAI(
base_url="https://api.aigateway.sh/v1",
api_key="sk-aig-...",
)
stream = client.chat.completions.create(
model="openai/o4-mini",
messages=[{"role": "user", "content": "Hello!"}],
stream=True,
)
for chunk in stream:
print(chunk.choices[0].delta.content or "", end="", flush=True)
# o4-mini returns chain-of-thought in message.reasoning_content —
# display it in a collapsed "show thinking" UI or just ignore it.