Quick Start Safety Check
Use a clean environment only. Run this demo on a fresh device or VM with no personal data, confidential information, or sensitive credentials. Think of it like a sandbox—keep it isolated.
By installing this demo, you're taking responsibility for all third-party components, including reviewing their licenses, terms, and security posture. Read and accept before you install or use.
This experience is provided "AS IS" for demonstration purposes only—no warranties, no guarantees. This is a demo, not a production-ready solution. You'll need to implement appropriate security controls for your specific environment and use case.
The demo showcases experimental AI agent capabilities. Even with cutting-edge open-source tools like OpenShell in your toolkit, you need to layer in proper security measures for your specific threat model.
Be mindful of these risks with AI agents:
No system is perfect, but these practices help keep your information and systems safe:
By participating in this demo, you acknowledge and agree that you are solely responsible for your configuration and for any data, accounts, and tools you connect.
To the maximum extent permitted by law, NVIDIA is not responsible for any loss of data, device damage, security incidents, or other harm arising from your configuration or use of demo materials, including OpenClaw or any connected tools or services.
This guide introduces the installation of OpenClaw, powered by Nemotron 3 Super via Ollama, on a DGX Spark (powered by GB10).
Verify that the OS, GPU, Docker, and Python are available before proceeding with this demo.
head -n 2 /etc/os-release
nvidia-smi
python3 --versionExpected output should show Ubuntu 24.04 (DGX OS), an NVIDIA DGX Spark GB10 GPU, and Python 3.12+.
curl -fsSL https://openclaw.ai/install.sh | bash -s -- --no-onboardInstall Ollama (if not already present). Ollama is a local-first platform for inferencing with models.
curl -fsSL https://ollama.ai/install.sh | sh
ollama --versionVerify Ollama is running (it auto-starts as a service after installation). Run this command as follows:
curl http://localhost:11434Expected output should be: Ollama is running. If not, start it manually:
ollama serve &Go to Step 2b if you have a flash drive with pre-downloaded models.
Pull and run the Nemotron 3 Super model from Ollama. This will take a while—the model is about 86GB.
To test the model via Ollama, run:
ollama pull nemotron-3-superTo load and run the model, please run:
ollama run nemotron-3-superUse this path when the target hardware has no Internet or you were given a flash drive.
Skip this if you already have a flash drive with the ./models folder. On a machine with internet access, save the script below as download.sh, then run it. Optionally move the resulting ./models folder to an external disk or flash drive.
#!/bin/bash
models=("qwen3.5:9b" "nemotron-3-super:120b")
for model in "${models[@]}"; do
dir_name="${model//:/-}"
ollama pull "$model" && echo "Successfully pulled $model"
(
mkdir -p "models/$dir_name" && cd "models/$dir_name" || exit 1
ollama show "$model" --modelfile > Modelfile
from_path=$(awk '/^FROM /{print $2; exit}' Modelfile)
cp "$from_path" model.gguf
awk '/^FROM / { $0="FROM ./model.gguf" } { print }' Modelfile > Modelfile.tmp && mv Modelfile.tmp Modelfile
echo "Successfully setup Modelfile for $model"
) || exit 1
done
models_line="models=("
for m in "${models[@]}"; do models_line+=" \"$m\""; done
models_line+=" )"
cat > models/setup.sh << EOF
#!/bin/bash
$models_line
for model in "\${models[@]}"; do
dir_name="\${model//:/-}"
(cd "\$(dirname "\$0")/\$dir_name" && ollama create "\${dir_name}_local") && echo "Successfully created Ollama model \${dir_name}_local"
done
EOF
chmod +x models/setup.sh
echo "Successfully created setup script for Ollama models"bash download.sh./models to target hardware (e.g. DGX Spark)Copy the ./models folder onto your flash drive, then on the target machine (e.g. DGX Spark) mount the drive and copy the data.
1. Find the drive and partition (e.g. /dev/sdb1 or /dev/nvme1n1p1):
lsblk2. Create a mount point and mount the partition:
sudo mkdir -p /mnt/ssd
sudo mount /dev/sdb1 /mnt/ssd # replace sdb1 with your actual partition3. Copy the models folder to your desired location:
cp -R /mnt/ssd/path/to/models ~/where/you/want/it4. Unmount the drive when done:
sudo umount /mnt/ssdOn the target machine, go to the path where you copied the data. The GGUF files and a setup.sh script are in the ./models folder. Run:
cd ./models
bash setup.shThe local model is registered as <model_name>_local. Use that name in the --model flag for all later commands (e.g. nemotron-3-super-120b_local or qwen3.5-9b_local).
To launch OpenClaw, run (use your _local model name if you followed the offline path in Step 2b):
ollama launch openclaw --model nemotron-3-superThe expected output of this should resemble the following screenshot: 
Some of the key expected outputs you should see in the logs include:
OpenClaw is running
Open the Web UI:
http://localhost:18789/#token=ollama
session agent:main:mainOnce the OpenClaw agent is ready, you can either enter a prompt on the terminal to interact with the agent, or you can click on the URL (e.g., http://localhost:18789/ …) to open the control UI. From there, you can start messaging the Nemotron 3 Super models, and see all the other features.
This is what the OpenClaw Control UI should look like:

openclaw models list # should show nemotron-3-super (or your _local model) as default
openclaw dashboard # opens a simple dashboardOnce OpenClaw is running, you can connect external messaging channels such as WhatsApp, Telegram, and Discord.
To open the channel configuration workflow, run:
openclaw configure --section channelsYou will be prompted to choose which channel you want to set up.
Once pairing succeeds, OpenClaw will be connected to your WhatsApp account.
To verify that WhatsApp is working correctly:
You can repeat the same openclaw configure --section channels flow later to connect additional platforms such as Telegram or Discord.
This guide introduces the installation of OpenClaw, powered by Qwen 3.5 9B via Ollama, on GeForce RTX devices (native Linux or WSL).
If you are on Windows and WSL is not installed, install it first by running in PowerShell:
wsl --installThen launch a WSL window:
wslInstall Ollama (if not already present). Ollama is a local-first platform for inferencing with models.
curl -fsSL https://ollama.ai/install.sh | sh
ollama --versionExpected output should be as follows:

Verify Ollama is running (it auto-starts as a service after installation):
curl http://localhost:11434Expected output should be: Ollama is running. If not, start it manually:
ollama serve &Next, pull down and run the Qwen model from Ollama.
To test the model via Ollama, run:
ollama pull qwen3.5:9bTo load and run the model, please run:
ollama run qwen3.5:9bThe expected output should be as follows:

ollama launch openclaw --model qwen3.5:9bThe expected output should resemble the following screenshot:

Some of the key expected outputs you should see in the logs include:
OpenClaw is running
Open the Web UI:
http://localhost:18789/#token=ollama
session agent:main:mainOnce the OpenClaw agent is ready, you can enter a prompt in the terminal or click the Web UI link to open the control UI and start messaging the model.
This is what the OpenClaw Control UI should look like:

Click the Web UI link displayed after a successful launch to access the dashboard.
Once OpenClaw is running, you can connect external messaging channels such as WhatsApp, Telegram, and Discord.
To open the channel configuration workflow, run:
openclaw configure --section channelsYou will be prompted to choose which channel you want to set up.
Once pairing succeeds, OpenClaw will be connected to your WhatsApp account.
To verify that WhatsApp is working correctly:
You can repeat the same openclaw configure --section channels flow later to connect additional platforms such as Telegram or Discord.
NVIDIA Brev provides easy access to cloud compute instances across major cloud partners. Developers can select from a broad variety of NVIDIA GPUs and CPUs within a dedicated environment, separate from corporate systems, networks, and data. Use pre-configured Launchables to start building immediately.
For more information, see the NVIDIA Brev documentation.
Deploy OpenClaw on a Brev cloud instance using a one-click Launchable. The Launchable bootstraps OpenClaw on a fresh Ubuntu-based Brev environment.
No local GPU or manual setup required.
On first launch, code-server will automatically open a terminal running configure.sh from the cloned launch-openclaw repo.
Once the gateway is running, you will see output like:
OpenClaw Gateway Started
========================
code-server:
https://code-server0-<id>.brevlab.com
URL:
https://openclaw0-<id>.brevlab.com/chat?session=mainOpen the OpenClaw URL in your browser and start chatting with the agent.
Once OpenClaw is running, you can connect external messaging channels such as WhatsApp, Telegram, and Discord.
To open the channel configuration workflow, run:
openclaw configure --section channelsYou will be prompted to choose which channel you want to set up.
Once pairing succeeds, OpenClaw will be connected to your WhatsApp account.
To verify that WhatsApp is working correctly:
You can repeat the same openclaw configure --section channels flow later to connect additional platforms such as Telegram or Discord.