Skip to content

Instantly share code, notes, and snippets.

@hmseeb
Last active March 6, 2026 17:21
Show Gist options
  • Select an option

  • Save hmseeb/2e53af5f8582614a3fe742e7216923f7 to your computer and use it in GitHub Desktop.

Select an option

Save hmseeb/2e53af5f8582614a3fe742e7216923f7 to your computer and use it in GitHub Desktop.
OpenClaw NVIDIA provider config for Railway deployment
{
"meta": {
"lastTouchedVersion": "2026.2.9",
"lastTouchedAt": "2026-03-06T11:17:33.023Z"
},
"wizard": {
"lastRunAt": "2026-03-06T11:17:33.016Z",
"lastRunVersion": "2026.2.9",
"lastRunCommand": "doctor",
"lastRunMode": "local"
},
"models": {
"mode": "merge",
"providers": {
"nvidia": {
"baseUrl": "https://integrate.api.nvidia.com/v1",
"apiKey": "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",
"api": "openai-completions",
"models": [
{
"id": "meta/llama-3.3-70b-instruct",
"name": "Llama 3.3 70B",
"contextWindow": 131072,
"maxTokens": 8192
},
{
"id": "mistralai/mistral-large-3-675b-instruct-2512",
"name": "Mistral Large 3 675B",
"contextWindow": 131072,
"maxTokens": 8192
}
]
}
}
},
"agents": {
"defaults": {
"model": {
"primary": "nvidia/mistralai/mistral-large-3-675b-instruct-2512",
"fallbacks": ["nvidia/meta/llama-3.3-70b-instruct"]
},
"models": {
"nvidia/mistralai/mistral-large-3-675b-instruct-2512": {
"alias": "mistral"
},
"nvidia/meta/llama-3.3-70b-instruct": {
"alias": "llama"
}
},
"workspace": "/data/workspace",
"compaction": {
"mode": "safeguard"
},
"maxConcurrent": 4,
"subagents": {
"maxConcurrent": 8
}
}
},
"messages": {
"ackReactionScope": "group-mentions"
},
"commands": {
"native": "auto",
"nativeSkills": "auto",
"restart": true
},
"channels": {
"telegram": {
"enabled": true,
"dmPolicy": "pairing",
"botToken": "8595214672:AAESnwfloBgI9SOm3_yQV76pZ363fC0PAzk",
"groupPolicy": "allowlist",
"streamMode": "partial"
}
},
"gateway": {
"port": 18789,
"mode": "local",
"bind": "loopback",
"auth": {
"mode": "token",
"token": "9a348afdae1ec7f92facf4a98efcbf42961067ba067c49ac5400495a12885684"
},
"trustedProxies": [
"127.0.0.1"
],
"tailscale": {
"mode": "off",
"resetOnExit": false
},
"remote": {
"token": "9a348afdae1ec7f92facf4a98efcbf42961067ba067c49ac5400495a12885684"
}
},
"skills": {
"install": {
"nodeManager": "npm"
}
},
"plugins": {
"entries": {
"telegram": {
"enabled": true
}
}
}
}
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment