This repository provides an optimized environment for running TensorRT-LLM on Ubuntu 24.04 with CUDA 13.0.
This Docker image is designed for performance, reproducibility, and ease of use:
- Base Image:
nvidia/cuda:13.0.0-devel-ubuntu24.04 - Modern Tooling: Powered by uv for lightning-fast package management.
- Pre-installed Dependencies:
- PyTorch: 2.9.1+cu130
- TensorRT-LLM: 1.3.0rc1
- System Deps:
libopenmpi-dev,python3.12-dev,nano,git,curl
- Automatic Environment: The virtual environment is located at
/workspace/trrt-llm/.venvand is automatically activated when you open a shell in the container.
docker build -t trt-llm-opt .Ensure you have the NVIDIA Container Toolkit installed.
docker run --gpus all -it --rm trt-llm-optOnce inside the container, you can verify the installation immediately:
python -c "import torch; import tensorrt_llm; print(f'Torch: {torch.__version__}'); print('TensorRT-LLM installed successfully!')"The environment is managed via pyproject.toml and uv.lock.
- To add new packages:
uv add <package> - To sync the environment:
uv sync
Note: The working directory is set to /workspace/trrt-llm.