Spaces:
Configuration error
title:
page: IniClaw Inference Profiles — NVIDIA Cloud
nav: Inference Profiles
description: Configuration reference for NVIDIA cloud inference profiles.
keywords:
- iniclaw inference profiles
- iniclaw nvidia cloud provider
topics:
- generative_ai
- ai_agents
tags:
- openclaw
- openshell
- inference_routing
- llms
content:
type: reference
difficulty: intermediate
audience:
- developer
- engineer
status: published
Inference Profiles
IniClaw ships with an inference profile defined in blueprint.yaml.
The profile configures an OpenShell inference provider and model route.
The agent inside the sandbox uses whichever model is active.
Inference requests are routed transparently through the OpenShell gateway.
Profile Summary
| Profile | Provider | Model | Endpoint | Use Case |
|---|---|---|---|---|
default |
NVIDIA cloud | nvidia/nemotron-3-super-120b-a12b |
integrate.api.nvidia.com |
Production. Requires an NVIDIA API key. |
Available Models
The nvidia-nim provider registers the following models from build.nvidia.com:
| Model ID | Label | Context Window | Max Output |
|---|---|---|---|
nvidia/nemotron-3-super-120b-a12b |
Nemotron 3 Super 120B | 131,072 | 8,192 |
nvidia/llama-3.1-nemotron-ultra-253b-v1 |
Nemotron Ultra 253B | 131,072 | 4,096 |
nvidia/llama-3.3-nemotron-super-49b-v1.5 |
Nemotron Super 49B v1.5 | 131,072 | 4,096 |
nvidia/nemotron-3-nano-30b-a3b |
Nemotron 3 Nano 30B | 131,072 | 4,096 |
The default profile uses Nemotron 3 Super 120B. You can switch to any model in the catalog at runtime.
default -- NVIDIA Cloud
The default profile routes inference to NVIDIA's hosted API through build.nvidia.com.
- Provider type:
nvidia - Endpoint:
https://integrate.api.nvidia.com/v1 - Model:
nvidia/nemotron-3-super-120b-a12b - Credential:
NVIDIA_API_KEYenvironment variable
Get an API key from build.nvidia.com.
The iniclaw onboard command prompts for this key and stores it in ~/.iniclaw/credentials.json.
$ openshell inference set --provider nvidia-nim --model nvidia/nemotron-3-super-120b-a12b
Switching Models at Runtime
After the sandbox is running, switch models with the OpenShell CLI:
$ openshell inference set --provider nvidia-nim --model <model-name>
The change takes effect immediately. No sandbox restart is needed.