[Experimental] TensorRT-LLM
Cosmos-RL supports TensorRT-LLM as the backend of rollout generation.
Note
To use TensorRT-LLM as the rollout backend, you have to build docker image with the file docker/Dockerfile.trtllm
in the root directory of cosmos-rl
project.
Enable TensorRT-LLM
To enable TensorRT-LLM, you need to set the fields of rollout
section in the config file:
[rollout]
backend = "trtllm"
For now, tested models are:
Qwen3-moe
Qwen2-5
Qwen2-5 VL
Note
We just support rollout replica within a single node now.