OpenAI provides developer APIs for GPT models, and Codex is also available as a ChatGPT-plan coding agent through OpenAI’s Codex clients. OpenClaw keeps those surfaces separate so config stays predictable. OpenClaw usesDocumentation Index
Fetch the complete documentation index at: https://docs.openclaw.ai/llms.txt
Use this file to discover all available pages before exploring further.
openai/* as the canonical OpenAI model route. Embedded agent
turns on OpenAI models run through the native Codex app-server runtime by
default; direct OpenAI API-key auth remains available for non-agent OpenAI
surfaces such as images, embeddings, speech, and realtime.
- Agent models -
openai/*models through the Codex runtime; sign in with Codex auth for ChatGPT/Codex subscription use, or configure a Codex-compatible OpenAI API-key backup when you intentionally want API-key auth. - Non-agent OpenAI APIs - direct OpenAI Platform access with usage-based
billing through
OPENAI_API_KEYor OpenAI API-key onboarding. - Legacy config -
openai-codex/*model refs are repaired byopenclaw doctor --fixtoopenai/*plus the Codex runtime.
Quick choice
| Goal | Use | Notes |
|---|---|---|
| ChatGPT/Codex subscription with native Codex runtime | openai/gpt-5.5 | Default OpenAI agent setup. Sign in with Codex auth. |
| Direct API-key billing for agent models | openai/gpt-5.5 plus a Codex-compatible API-key profile | Use auth.order.openai to place the backup after subscription auth. |
| Direct API-key billing through explicit PI | openai/gpt-5.5 plus provider/model runtime pi | Select a normal openai API-key profile. |
| Latest ChatGPT Instant API alias | openai/chat-latest | Direct API-key only. Moving alias for experiments, not the default. |
| ChatGPT/Codex subscription auth through explicit PI | openai/gpt-5.5 plus provider/model runtime pi | Select an openai-codex auth profile for the compatibility route. |
| Image generation or editing | openai/gpt-image-2 | Works with either OPENAI_API_KEY or OpenAI Codex OAuth. |
| Transparent-background images | openai/gpt-image-1.5 | Use outputFormat=png or webp and openai.background=transparent. |
Naming map
The names are similar but not interchangeable:| Name you see | Layer | Meaning |
|---|---|---|
openai | Provider prefix | Canonical OpenAI model route; agent turns use the Codex runtime. |
openai-codex | Legacy auth/profile prefix | Older OpenAI Codex OAuth/subscription profile namespace. Existing profiles and auth.order.openai-codex still work. |
codex plugin | Plugin | Bundled OpenClaw plugin that provides native Codex app-server runtime and /codex chat controls. |
provider/model agentRuntime.id: codex | Agent runtime | Force the native Codex app-server harness for matching embedded turns. |
/codex ... | Chat command set | Bind/control Codex app-server threads from a conversation. |
runtime: "acp", agentId: "codex" | ACP session route | Explicit fallback path that runs Codex through ACP/acpx. |
openai/* model refs while auth
profiles still point at Codex-compatible credentials. Prefer auth.order.openai
for new config; existing openai-codex:* profiles and auth.order.openai-codex
remain supported. openclaw doctor --fix rewrites legacy openai-codex/* model
refs to the canonical OpenAI model route.
GPT-5.5 is available through both direct OpenAI Platform API-key access and
subscription/OAuth routes. For ChatGPT/Codex subscription plus native Codex
execution, use
openai/gpt-5.5; unset runtime config now selects the Codex
harness for OpenAI agent turns. Use OpenAI API-key profiles only when you want
direct API-key auth for an OpenAI agent model.OpenAI agent model turns require the bundled Codex app-server plugin. Explicit
PI runtime config remains available as an opt-in compatibility route. When PI is
explicitly selected with an
openai-codex auth profile, OpenClaw keeps the
public model ref as openai/* and routes PI internally through the legacy
Codex-auth transport. Run openclaw doctor --fix to repair stale
openai-codex/* model refs or old PI session pins that do not come from
explicit runtime config.OpenClaw feature coverage
| OpenAI capability | OpenClaw surface | Status |
|---|---|---|
| Chat / Responses | openai/<model> model provider | Yes |
| Codex subscription models | openai/<model> with openai-codex OAuth | Yes |
| Legacy Codex model refs | openai-codex/<model> | Repaired by doctor to openai/<model> |
| Codex app-server harness | openai/<model> with omitted runtime or provider/model agentRuntime.id: codex | Yes |
| Server-side web search | Native OpenAI Responses tool | Yes, when web search is enabled and no provider pinned |
| Images | image_generate | Yes |
| Videos | video_generate | Yes |
| Text-to-speech | messages.tts.provider: "openai" / tts | Yes |
| Batch speech-to-text | tools.media.audio / media understanding | Yes |
| Streaming speech-to-text | Voice Call streaming.provider: "openai" | Yes |
| Realtime voice | Voice Call realtime.provider: "openai" / Control UI Talk | Yes |
| Embeddings | memory embedding provider | Yes |
Memory embeddings
OpenClaw can use OpenAI, or an OpenAI-compatible embedding endpoint, formemory_search indexing and query embeddings:
queryInputType and documentInputType under memorySearch. OpenClaw forwards
those as provider-specific input_type request fields: query embeddings use
queryInputType; indexed memory chunks and batch indexing use
documentInputType. See the Memory configuration reference for the full example.
Getting started
Choose your preferred auth method and follow the setup steps.- API key (OpenAI Platform)
- Codex subscription
Best for: direct API access and usage-based billing.
To try ChatGPT’s current Instant model from the OpenAI API, set the model
to
Get your API key
Create or copy an API key from the OpenAI Platform dashboard.
Route summary
| Model ref | Runtime config | Route | Auth |
|---|---|---|---|
openai/gpt-5.5 | omitted / provider/model agentRuntime.id: "codex" | Codex app-server harness | Codex-compatible OpenAI profile |
openai/gpt-5.4-mini | omitted / provider/model agentRuntime.id: "codex" | Codex app-server harness | Codex-compatible OpenAI profile |
openai/gpt-5.5 | provider/model agentRuntime.id: "pi" | PI embedded runtime | openai profile or selected openai-codex profile |
openai/* agent models use the Codex app-server harness. To use API-key
auth for an agent model, create a Codex-compatible API-key profile and order
it with auth.order.openai; OPENAI_API_KEY remains the direct fallback for
non-agent OpenAI API surfaces. Older auth.order.openai-codex entries still
work.Config example
openai/chat-latest:chat-latest is a moving alias. OpenAI documents it as the latest Instant
model used in ChatGPT and recommends gpt-5.5 for production API usage, so
keep openai/gpt-5.5 as the stable default unless you explicitly want that
alias behavior. The alias currently accepts only medium text verbosity, so
OpenClaw normalizes incompatible OpenAI text-verbosity overrides for this
model.Native Codex app-server auth
The native Codex app-server harness usesopenai/* model refs plus omitted
runtime config or provider/model agentRuntime.id: "codex", but its auth is
still account-based. OpenClaw selects auth in this order:
- Ordered OpenAI auth profiles for the agent, preferably under
auth.order.openai. Existingopenai-codex:*profiles andauth.order.openai-codexremain valid for older installs. - The app-server’s existing account, such as a local Codex CLI ChatGPT sign-in.
- For local stdio app-server launches only,
CODEX_API_KEY, thenOPENAI_API_KEY, when the app-server reports no account and still requires OpenAI auth.
OPENAI_API_KEY for direct OpenAI models
or embeddings. Env API-key fallback is only the local stdio no-account path; it
is not sent to WebSocket app-server connections. When a subscription-style Codex
profile is selected, OpenClaw also keeps CODEX_API_KEY and OPENAI_API_KEY
out of the spawned stdio app-server child and sends the selected credentials
through the app-server login RPC. When that subscription profile is blocked by a
Codex usage limit, OpenClaw can rotate to the next ordered openai:* API-key
profile without changing the selected model or dropping out of the Codex
harness. Once the subscription reset time passes, the subscription profile is
eligible again.
Image generation
The bundledopenai plugin registers image generation through the image_generate tool.
It supports both OpenAI API-key image generation and Codex OAuth image
generation through the same openai/gpt-image-2 model ref.
| Capability | OpenAI API key | Codex OAuth |
|---|---|---|
| Model ref | openai/gpt-image-2 | openai/gpt-image-2 |
| Auth | OPENAI_API_KEY | OpenAI Codex OAuth sign-in |
| Transport | OpenAI Images API | Codex Responses backend |
| Max images per request | 4 | 4 |
| Edit mode | Enabled (up to 5 reference images) | Enabled (up to 5 reference images) |
| Size overrides | Supported, including 2K/4K sizes | Supported, including 2K/4K sizes |
| Aspect ratio / resolution | Not forwarded to OpenAI Images API | Mapped to a supported size when safe |
See Image Generation for shared tool parameters, provider selection, and failover behavior.
gpt-image-2 is the default for both OpenAI text-to-image generation and image
editing. gpt-image-1.5, gpt-image-1, and gpt-image-1-mini remain usable as
explicit model overrides. Use openai/gpt-image-1.5 for transparent-background
PNG/WebP output; the current gpt-image-2 API rejects
background: "transparent".
For a transparent-background request, agents should call image_generate with
model: "openai/gpt-image-1.5", outputFormat: "png" or "webp", and
background: "transparent"; the older openai.background provider option is
still accepted. OpenClaw also protects the public OpenAI and
OpenAI Codex OAuth routes by rewriting default openai/gpt-image-2 transparent
requests to gpt-image-1.5; Azure and custom OpenAI-compatible endpoints keep
their configured deployment/model names.
The same setting is exposed for headless CLI runs:
--output-format and --background flags with
openclaw infer image edit when starting from an input file.
--openai-background remains available as an OpenAI-specific alias.
For Codex OAuth installs, keep the same openai/gpt-image-2 ref. When an
openai-codex OAuth profile is configured, OpenClaw resolves that stored OAuth
access token and sends image requests through the Codex Responses backend. It
does not first try OPENAI_API_KEY or silently fall back to an API key for that
request. Configure models.providers.openai explicitly with an API key,
custom base URL, or Azure endpoint when you want the direct OpenAI Images API
route instead.
If that custom image endpoint is on a trusted LAN/private address, also set
browser.ssrfPolicy.dangerouslyAllowPrivateNetwork: true; OpenClaw keeps
private/internal OpenAI-compatible image endpoints blocked unless this opt-in is
present.
Generate:
Video generation
The bundledopenai plugin registers video generation through the video_generate tool.
| Capability | Value |
|---|---|
| Default model | openai/sora-2 |
| Modes | Text-to-video, image-to-video, single-video edit |
| Reference inputs | 1 image or 1 video |
| Size overrides | Supported |
| Other overrides | aspectRatio, resolution, audio, watermark are ignored with a tool warning |
See Video Generation for shared tool parameters, provider selection, and failover behavior.
GPT-5 prompt contribution
OpenClaw adds a shared GPT-5 prompt contribution for GPT-5-family runs across providers. It applies by model id, soopenai/gpt-5.5, legacy pre-repair refs such as openai-codex/gpt-5.5, openrouter/openai/gpt-5.5, opencode/gpt-5.5, and other compatible GPT-5 refs receive the same overlay. Older GPT-4.x models do not.
The bundled native Codex harness uses the same GPT-5 behavior and heartbeat overlay through Codex app-server developer instructions, so openai/gpt-5.x sessions routed through Codex keep the same follow-through and proactive heartbeat guidance even though Codex owns the rest of the harness prompt.
The GPT-5 contribution adds a tagged behavior contract for persona persistence, execution safety, tool discipline, output shape, completion checks, and verification. Channel-specific reply and silent-message behavior stays in the shared OpenClaw system prompt and outbound delivery policy. The GPT-5 guidance is always enabled for matching models. The friendly interaction-style layer is separate and configurable.
| Value | Effect |
|---|---|
"friendly" (default) | Enable the friendly interaction-style layer |
"on" | Alias for "friendly" |
"off" | Disable only the friendly style layer |
- Config
- CLI
Legacy
plugins.entries.openai.config.personality is still read as a compatibility fallback when the shared agents.defaults.promptOverlays.gpt5.personality setting is not set.Voice and speech
Speech synthesis (TTS)
Speech synthesis (TTS)
The bundled
Available models:
openai plugin registers speech synthesis for the messages.tts surface.| Setting | Config path | Default |
|---|---|---|
| Model | messages.tts.providers.openai.model | gpt-4o-mini-tts |
| Voice | messages.tts.providers.openai.voice | coral |
| Speed | messages.tts.providers.openai.speed | (unset) |
| Instructions | messages.tts.providers.openai.instructions | (unset, gpt-4o-mini-tts only) |
| Format | messages.tts.providers.openai.responseFormat | opus for voice notes, mp3 for files |
| API key | messages.tts.providers.openai.apiKey | Falls back to OPENAI_API_KEY |
| Base URL | messages.tts.providers.openai.baseUrl | https://api.openai.com/v1 |
| Extra body | messages.tts.providers.openai.extraBody / extra_body | (unset) |
gpt-4o-mini-tts, tts-1, tts-1-hd. Available voices: alloy, ash, ballad, cedar, coral, echo, fable, juniper, marin, onyx, nova, sage, shimmer, verse.extraBody is merged into /audio/speech request JSON after OpenClaw’s generated fields, so use it for OpenAI-compatible endpoints that require additional keys such as lang. Prototype keys are ignored.Set
OPENAI_TTS_BASE_URL to override the TTS base URL without affecting the chat API endpoint. OpenAI TTS is still configured through an API key; for OAuth-only live talk-back, use the Realtime voice path instead of agent-mode STT -> TTS speech.Speech-to-text
Speech-to-text
The bundled Language and prompt hints are forwarded to OpenAI when supplied by the
shared audio media config or per-call transcription request.
openai plugin registers batch speech-to-text through
OpenClaw’s media-understanding transcription surface.- Default model:
gpt-4o-transcribe - Endpoint: OpenAI REST
/v1/audio/transcriptions - Input path: multipart audio file upload
- Supported by OpenClaw wherever inbound audio transcription uses
tools.media.audio, including Discord voice-channel segments and channel audio attachments
Realtime transcription
Realtime transcription
The bundled
openai plugin registers realtime transcription for the Voice Call plugin.| Setting | Config path | Default |
|---|---|---|
| Model | plugins.entries.voice-call.config.streaming.providers.openai.model | gpt-4o-transcribe |
| Language | ...openai.language | (unset) |
| Prompt | ...openai.prompt | (unset) |
| Silence duration | ...openai.silenceDurationMs | 800 |
| VAD threshold | ...openai.vadThreshold | 0.5 |
| Auth | ...openai.apiKey, OPENAI_API_KEY, or openai-codex OAuth | API keys connect directly; OAuth mints a Realtime transcription client secret |
Uses a WebSocket connection to
wss://api.openai.com/v1/realtime with G.711 u-law (g711_ulaw / audio/pcmu) audio. When only openai-codex OAuth is configured, the Gateway mints an ephemeral Realtime transcription client secret before opening the WebSocket. This streaming provider is for Voice Call’s realtime transcription path; Discord voice currently records short segments and uses the batch tools.media.audio transcription path instead.Realtime voice
Realtime voice
The bundled
Available built-in Realtime voices for
openai plugin registers realtime voice for the Voice Call plugin.| Setting | Config path | Default |
|---|---|---|
| Model | plugins.entries.voice-call.config.realtime.providers.openai.model | gpt-realtime-2 |
| Voice | ...openai.voice | alloy |
| Temperature (Azure deployment bridge) | ...openai.temperature | 0.8 |
| VAD threshold | ...openai.vadThreshold | 0.5 |
| Silence duration | ...openai.silenceDurationMs | 500 |
| Prefix padding | ...openai.prefixPaddingMs | 300 |
| Reasoning effort | ...openai.reasoningEffort | (unset) |
| Auth | ...openai.apiKey, OPENAI_API_KEY, or openai-codex OAuth | Browser Talk and non-Azure backend bridges can use Codex OAuth |
gpt-realtime-2: alloy, ash,
ballad, coral, echo, sage, shimmer, verse, marin, cedar.
OpenAI recommends marin and cedar for the best Realtime quality. This
is a separate set from the Text-to-speech voices above; do not assume a TTS
voice such as fable, nova, or onyx is valid for Realtime sessions.Backend OpenAI realtime bridges use the GA Realtime WebSocket session shape, which does not accept
session.temperature. Azure OpenAI deployments remain available via azureEndpoint and azureDeployment and keep the deployment-compatible session shape. Supports bidirectional tool calling and G.711 u-law audio.Realtime voice is selected when the session is created. OpenAI allows most
session fields to change later, but the voice cannot be changed after the
model has emitted audio in that session. OpenClaw currently exposes the
built-in Realtime voice ids as strings.
Control UI Talk uses OpenAI browser realtime sessions with a Gateway-minted
ephemeral client secret and a direct browser WebRTC SDP exchange against the
OpenAI Realtime API. When no direct OpenAI API key is configured, the
Gateway can mint that client secret with the selected
openai-codex OAuth
profile. Gateway relay and Voice Call backend realtime WebSocket bridges use
the same OAuth fallback for native OpenAI endpoints. Maintainer live
verification is available with
OPENAI_API_KEY=... GEMINI_API_KEY=... node --import tsx scripts/dev/realtime-talk-live-smoke.ts;
the OpenAI legs verify both the backend WebSocket bridge and the browser
WebRTC SDP exchange without logging secrets.Azure OpenAI endpoints
The bundledopenai provider can target an Azure OpenAI resource for image
generation by overriding the base URL. On the image-generation path, OpenClaw
detects Azure hostnames on models.providers.openai.baseUrl and switches to
Azure’s request shape automatically.
Realtime voice uses a separate configuration path
(
plugins.entries.voice-call.config.realtime.providers.openai.azureEndpoint)
and is not affected by models.providers.openai.baseUrl. See the Realtime
voice accordion under Voice and speech for its Azure
settings.- You already have an Azure OpenAI subscription, quota, or enterprise agreement
- You need regional data residency or compliance controls Azure provides
- You want to keep traffic inside an existing Azure tenancy
Configuration
For Azure image generation through the bundledopenai provider, point
models.providers.openai.baseUrl at your Azure resource and set apiKey to
the Azure OpenAI key (not an OpenAI Platform key):
*.openai.azure.com*.services.ai.azure.com*.cognitiveservices.azure.com
- Sends the
api-keyheader instead ofAuthorization: Bearer - Uses deployment-scoped paths (
/openai/deployments/{deployment}/...) - Appends
?api-version=...to each request - Uses a 600s default request timeout for Azure image-generation calls.
Per-call
timeoutMsvalues still override this default.
Azure routing for the
openai provider’s image-generation path requires
OpenClaw 2026.4.22 or later. Earlier versions treat any custom
openai.baseUrl like the public OpenAI endpoint and will fail against Azure
image deployments.API version
SetAZURE_OPENAI_API_VERSION to pin a specific Azure preview or GA version
for the Azure image-generation path:
2024-12-01-preview when the variable is unset.
Model names are deployment names
Azure OpenAI binds models to deployments. For Azure image-generation requests routed through the bundledopenai provider, the model field in OpenClaw
must be the Azure deployment name you configured in the Azure portal, not
the public OpenAI model id.
If you create a deployment called gpt-image-2-prod that serves gpt-image-2:
openai provider.
Regional availability
Azure image generation is currently available only in a subset of regions (for exampleeastus2, swedencentral, polandcentral, westus3,
uaenorth). Check Microsoft’s current region list before creating a
deployment, and confirm the specific model is offered in your region.
Parameter differences
Azure OpenAI and public OpenAI do not always accept the same image parameters. Azure may reject options that public OpenAI allows (for example certainbackground values on gpt-image-2) or expose them only on specific model
versions. These differences come from Azure and the underlying model, not
OpenClaw. If an Azure request fails with a validation error, check the
parameter set supported by your specific deployment and API version in the
Azure portal.
Azure OpenAI uses native transport and compat behavior but does not receive
OpenClaw’s hidden attribution headers — see the Native vs OpenAI-compatible
routes accordion under Advanced configuration.For chat or Responses traffic on Azure (beyond image generation), use the
onboarding flow or a dedicated Azure provider config —
openai.baseUrl alone
does not pick up the Azure API/auth shape. A separate
azure-openai-responses/* provider exists; see
the Server-side compaction accordion below.Advanced configuration
Transport (WebSocket vs SSE)
Transport (WebSocket vs SSE)
OpenClaw uses WebSocket-first with SSE fallback (
Related OpenAI docs:
"auto") for openai/*.In "auto" mode, OpenClaw:- Retries one early WebSocket failure before falling back to SSE
- After a failure, marks WebSocket as degraded for ~60 seconds and uses SSE during cool-down
- Attaches stable session and turn identity headers for retries and reconnects
- Normalizes usage counters (
input_tokens/prompt_tokens) across transport variants
| Value | Behavior |
|---|---|
"auto" (default) | WebSocket first, SSE fallback |
"sse" | Force SSE only |
"websocket" | Force WebSocket only |
Fast mode
Fast mode
OpenClaw exposes a shared fast-mode toggle for
openai/*:- Chat/UI:
/fast status|on|off - Config:
agents.defaults.models["<provider>/<model>"].params.fastMode
service_tier = "priority"). Existing service_tier values are preserved, and fast mode does not rewrite reasoning or text.verbosity.Session overrides win over config. Clearing the session override in the Sessions UI returns the session to the configured default.
Priority processing (service_tier)
Priority processing (service_tier)
OpenAI’s API exposes priority processing via Supported values:
service_tier. Set it per model in OpenClaw:auto, default, flex, priority.Server-side compaction (Responses API)
Server-side compaction (Responses API)
For direct OpenAI Responses models (
openai/* on api.openai.com), the OpenAI plugin’s Pi-harness stream wrapper auto-enables server-side compaction:- Forces
store: true(unless model compat setssupportsStore: false) - Injects
context_management: [{ type: "compaction", compact_threshold: ... }] - Default
compact_threshold: 70% ofcontextWindow(or80000when unavailable)
- Enable explicitly
- Custom threshold
- Disable
Useful for compatible endpoints like Azure OpenAI Responses:
responsesServerCompaction only controls context_management injection. Direct OpenAI Responses models still force store: true unless compat sets supportsStore: false.Strict-agentic GPT mode
Strict-agentic GPT mode
For GPT-5-family runs on With
openai/*, OpenClaw can use a stricter embedded execution contract:strict-agentic, OpenClaw:- No longer treats a plan-only turn as successful progress when a tool action is available
- Retries the turn with an act-now steer
- Auto-enables
update_planfor substantial work - Surfaces an explicit blocked state if the model keeps planning without acting
Scoped to OpenAI and Codex GPT-5-family runs only. Other providers and older model families keep default behavior.
Native vs OpenAI-compatible routes
Native vs OpenAI-compatible routes
OpenClaw treats direct OpenAI, Codex, and Azure OpenAI endpoints differently from generic OpenAI-compatible
/v1 proxies:Native routes (openai/*, Azure OpenAI):- Keep
reasoning: { effort: "none" }only for models that support the OpenAInoneeffort - Omit disabled reasoning for models or proxies that reject
reasoning.effort: "none" - Default tool schemas to strict mode
- Attach hidden attribution headers on verified native hosts only
- Keep OpenAI-only request shaping (
service_tier,store, reasoning-compat, prompt-cache hints)
- Use looser compat behavior
- Strip Completions
storefrom non-nativeopenai-completionspayloads - Accept advanced
params.extra_body/params.extraBodypass-through JSON for OpenAI-compatible Completions proxies - Accept
params.chat_template_kwargsfor OpenAI-compatible Completions proxies such as vLLM - Do not force strict tool schemas or native-only headers
Related
Model selection
Choosing providers, model refs, and failover behavior.
Image generation
Shared image tool parameters and provider selection.
Video generation
Shared video tool parameters and provider selection.
OAuth and auth
Auth details and credential reuse rules.