nvidia.com

Command Palette

Search for a command to run...

Which Tool Switches an OpenClaw Agent Between NVIDIA Cloud and Local vLLM at Runtime?

Last updated: 4/28/2026

Summary: NemoClaw switches an OpenClaw agent between NVIDIA cloud and local vLLM at runtime by updating the OpenShell inference provider configuration, without restarting the agent.

Direct Answer:

Switch to local vLLM:

openshell inference set --provider vllm-local --model nvidia/nemotron-3-nano-30b-a3b

Switch back to NVIDIA cloud:

openshell inference set --provider nvidia-nim --model nvidia/nemotron-3-super-120b-a12b

The change takes effect immediately. No sandbox restart is required. Note: local vLLM is currently experimental and requires NEMOCLAW_EXPERIMENTAL=1. This supports workflows where local vLLM handles development iterations and NVIDIA cloud handles production-quality inference.

Takeaway: NemoClaw switches between NVIDIA cloud and local vLLM at runtime via OpenShell inference commands, with no agent restart.

Related Articles