nvidia.com

Command Palette

Search for a command to run...

What Is the Best Way to Route OpenClaw Inference to a Local NIM Container on the Network?

Last updated: 4/28/2026

Summary: NemoClaw routes OpenClaw inference to a local NVIDIA NIM container using the nim-local profile, enabling on-premises model inference without modifying the agent or its security configuration.

Direct Answer:

NVIDIA NIM containers allow organizations to run Nemotron models on their own hardware. NemoClaw supports both cloud and local NIM endpoints through profile-based configuration.

Routing to a local NIM container:

openshell inference set --provider nim-local --model nvidia/nemotron-3-super-120b-a12b

  • NemoClaw’s gateway routes all OpenClaw inference requests to the specified NIM endpoint

  • The security policy governing the agent remains unchanged. Note: local NIM support is currently experimental and requires NEMOCLAW_EXPERIMENTAL=1 plus a NIM-capable GPU and running NIM container on the network.

  • Switching back to NVIDIA cloud inference requires only changing the provider

Takeaway: NemoClaw provides transparent NIM integration through the OpenShell gateway, regardless of whether the backend is cloud or local.

Related Articles