Get Started with Cosmos Reason1 on Brev: Inference and Post-Training
Author: Saurav Nanda Organization: NVIDIA
This guide walks you through setting up NVIDIA Cosmos Reason1 on a Brev H100 GPU instance for both inference and post-training workflows. Brev provides on-demand cloud GPUs with pre-configured environments, making it easy to get started with Cosmos models.
Overview
Brev.dev is a cloud GPU platform that provides instant access to high-performance GPUs like the H100. This guide will help you:
- Set up a Brev instance with H100 GPU
- Configure the environment for Cosmos Reason1
- Run inference on the Reason1 model
- Perform post-training (SFT) on custom datasets
Prerequisites
- A Brev account (sign up here)
- Install the CLI as shown in https://docs.nvidia.com/brev/latest/brev-cli.html
- See the quickstart to get a feel for the platform: https://docs.nvidia.com/brev/latest/quick-start.html. The handy Brev docs are linked from the Brev page too.
- A Hugging Face account with access to Cosmos-Reason1-7B
The cheat code: Launchables
Launchables are an easy way to bundle a hardware and software environment into an easily shareable link. Once you've dialed in your Cosmos setup, a Launchable is the most convenient way to save time and share your configuration with others.
Note: Cosmos and Brev are evolving. You may encounter minor UI and other differences in the steps below over time.
Step 1: Create a Brev Instance
- Log in to your Brev account
- Find the Launchable section of the Brev website.

- Click the Create Launchable button.

-
Enter the Cosmos Reason GitHub URL

-
Choose H100 GPU instance with 80GB VRAM.

-
If you don't need Jupyter, remove it. Tell Brev which ports to open if you plan to setup some other custom server(if any).

-
Add a setup script for Cosmos Reason. See sample setup script for an example.

-
Name your Launchable and configure access.(usually takes 2-3 minutes).

Step 2: Connect to Your Instance
Once your instance is ready, Brev will provide SSH connection details. 
Option 1: Open Jupyter Notebook

Option 2: Copy the SSH command from your Brev dashboard
Open a terminal locally
OR Open in Code Editor
Step 3: Authenticate Hugging Face CLI
The Hugging Face Token is required to download the Cosmos Reason1 model:
When prompted, enter your Hugging Face token. You can create a token at https://huggingface.co/settings/tokens.
Important: Make sure you have access to the Cosmos-Reason1-7B model. Request access if needed.
Step 4: Run Inference and Post-Training
Now you're ready to run inference with Cosmos Reason1!
Follow the steps provided in the Cosmos Reason GitHub repo to run the inference and post-training examples.
Troubleshooting
Model Download Issues
If the model fails to download:
- Verify your Hugging Face authentication:
~/.local/bin/hf whoami - Ensure you have access to the Cosmos-Reason1-7B model
- Check your internet connection
- Try downloading manually:
huggingface-cli download nvidia/Cosmos-Reason1-7B
SSH Connection Issues
If you lose SSH connection:
- Brev instances may pause after inactivity
- Check your Brev dashboard for instance status
- Restart the instance if needed
- Reconnect using the SSH command
Resource Management
Stopping Your Instance
To avoid unnecessary charges:
- Go to your Brev dashboard
- Select your instance
- Click "Stop" or "Delete" when done
Saving Your Work
Before stopping your instance:
# Save model checkpoints to cloud storage (e.g., S3, GCS)
# Or download them to your local machine
scp -r ubuntu@<your-instance-ip>:~/cosmos-reason1/examples/post_training_hf/outputs ./local-outputs
Additional Resources
- Cosmos Reason1 GitHub Repository
- Cosmos Reason1 Model on Hugging Face
- Cosmos Reason1 Paper
- Brev Documentation
- Cosmos Cookbook
Support
For issues related to:
- Cosmos Reason1: Open an issue on the GitHub repository
- Brev Platform: Contact Brev support