NVIDIA provides an OpenAI-compatible API atDocumentation Index
Fetch the complete documentation index at: https://docs.openclaw.ai/llms.txt
Use this file to discover all available pages before exploring further.
https://integrate.api.nvidia.com/v1 for
open models for free. Authenticate with an API key from
build.nvidia.com.
Getting started
Get your API key
Create an API key at build.nvidia.com.
Config example
Built-in catalog
| Model ref | Name | Context | Max output |
|---|---|---|---|
nvidia/nvidia/nemotron-3-super-120b-a12b | NVIDIA Nemotron 3 Super 120B | 262,144 | 8,192 |
nvidia/moonshotai/kimi-k2.5 | Kimi K2.5 | 262,144 | 8,192 |
nvidia/minimaxai/minimax-m2.5 | Minimax M2.5 | 196,608 | 8,192 |
nvidia/z-ai/glm5 | GLM 5 | 202,752 | 8,192 |
Advanced configuration
Auto-enable behavior
Auto-enable behavior
The provider auto-enables when the
NVIDIA_API_KEY environment variable is set.
No explicit provider config is required beyond the key.Catalog and pricing
Catalog and pricing
The bundled catalog is static. Costs default to
0 in source since NVIDIA
currently offers free API access for the listed models.OpenAI-compatible endpoint
OpenAI-compatible endpoint
NVIDIA uses the standard
/v1 completions endpoint. Any OpenAI-compatible
tooling should work out of the box with the NVIDIA base URL.Slow custom provider responses
Slow custom provider responses
Some NVIDIA-hosted custom models can take longer than the default model idle
watchdog before they emit a first response chunk. For custom NVIDIA provider
entries, raise the provider timeout instead of raising the whole agent
runtime timeout:
Related
Model selection
Choosing providers, model refs, and failover behavior.
Configuration reference
Full config reference for agents, models, and providers.