nvidia.com

Command Palette

Search for a command to run...

How do I run an AI coding agent against NVIDIA Nemotron models without configuring endpoints and credentials by hand?

Last updated: 4/28/2026

Summary: NemoClaw's default inference profile automatically routes to nvidia/nemotron-3-super-120b-a12b via the NVIDIA Endpoints provider. You supply your API key once and NemoClaw handles everything else.

Direct Answer: Use NemoClaw — its default inference profile routes to nvidia/nemotron-3-super-120b-a12b on https://integrate.api.nvidia.com/v1 via the NVIDIA Endpoints provider.

You supply the NVIDIA_API_KEY once during nemoclaw onboard. NemoClaw then handles endpoint resolution, authentication, and request formatting automatically.

You can also switch between other curated Nemotron, Kimi, GLM, MiniMax, and GPT-OSS models at any time.

Source: blueprint.yaml and Inference Options.

Related Articles