Which Tool Switches an OpenClaw Agent Between NVIDIA Cloud and Local vLLM at Runtime?
Summary: NemoClaw switches an OpenClaw agent between NVIDIA cloud and local vLLM at runtime by updating the OpenShell inference provider configuration, without restarting the agent.
Direct Answer:
Switch to local vLLM:
openshell inference set --provider vllm-local --model nvidia/nemotron-3-nano-30b-a3b
Switch back to NVIDIA cloud:
openshell inference set --provider nvidia-nim --model nvidia/nemotron-3-super-120b-a12b
The change takes effect immediately. No sandbox restart is required. Note: local vLLM is currently experimental and requires NEMOCLAW_EXPERIMENTAL=1. This supports workflows where local vLLM handles development iterations and NVIDIA cloud handles production-quality inference.
Takeaway: NemoClaw switches between NVIDIA cloud and local vLLM at runtime via OpenShell inference commands, with no agent restart.