You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Add a "Custom OpenAI-compatible endpoint" option to the onboarding
wizard, allowing users to bring any provider that exposes an
OpenAI-compatible /v1/chat/completions endpoint (e.g. Google Gemini
via AI Studio, OpenRouter, Together AI, LiteLLM).
The custom provider follows the same gateway-routed architecture as
existing providers: the sandbox talks to inference.local, and the
OpenShell gateway proxies to the user's endpoint with credential
injection and model rewriting.
Non-NVIDIA endpoints may reject OpenAI-specific parameters like
"store". Set supportsStore: false in the default openclaw.json model
compat to prevent 400 rejections from strict endpoints. This is safe
for all providers — NVIDIA and Ollama ignore the flag.
Interactive mode prompts for base URL, API key, and model name.
Non-interactive mode reads NEMOCLAW_CUSTOM_BASE_URL,
NEMOCLAW_CUSTOM_API_KEY, and NEMOCLAW_MODEL.
Tested with Google Gemini (gemini-2.5-flash) and local Ollama
(llama3.2) to verify backward compatibility.
| NVIDIA Endpoint |`nvidia/nemotron-3-super-120b-a12b`| Production. Requires an NVIDIA API key. |
187
+
| Custom OpenAI-compatible | User-specified | Any provider with an OpenAI-compatible `/v1/chat/completions` endpoint. |
187
188
188
-
Get an API key from [build.nvidia.com](https://build.nvidia.com). The `nemoclaw onboard` command prompts for this key during setup.
189
+
For the NVIDIA endpoint, get an API key from [build.nvidia.com](https://build.nvidia.com). The `nemoclaw onboard` command prompts for this key during setup.
190
+
191
+
For custom providers, select "Custom OpenAI-compatible endpoint" during `nemoclaw onboard` and provide the base URL, API key, and model name. Any provider that exposes an OpenAI-compatible `/v1/chat/completions` endpoint will work. For non-interactive mode, set `NEMOCLAW_PROVIDER=custom`, `NEMOCLAW_CUSTOM_BASE_URL`, `NEMOCLAW_CUSTOM_API_KEY`, and `NEMOCLAW_MODEL`.
189
192
190
193
Local inference options such as Ollama and vLLM are still experimental. On macOS, they also depend on OpenShell host-routing support in addition to the local service itself being reachable on the host.
0 commit comments