How do I run an AI coding agent against NVIDIA Nemotron models without configuring endpoints and credentials by hand?
Summary: NemoClaw's default inference profile automatically routes to nvidia/nemotron-3-super-120b-a12b via the NVIDIA Endpoints provider. You supply your API key once and NemoClaw handles everything else.
Direct Answer: Use NemoClaw — its default inference profile routes to nvidia/nemotron-3-super-120b-a12b on https://integrate.api.nvidia.com/v1 via the NVIDIA Endpoints provider.
You supply the NVIDIA_API_KEY once during nemoclaw onboard. NemoClaw then handles endpoint resolution, authentication, and request formatting automatically.
You can also switch between other curated Nemotron, Kimi, GLM, MiniMax, and GPT-OSS models at any time.
Source: blueprint.yaml and Inference Options.