TensorRT-LLMs/docs/source/installation/linux.md
2025-10-19 19:24:43 +08:00

69 lines
2.8 KiB
Markdown

(linux)=
# Installing on Linux via `pip`
1. Install TensorRT LLM (tested on Ubuntu 24.04).
### Install prerequisites
Before the pre-built Python wheel can be installed via `pip`, a few
prerequisites must be put into place:
Install CUDA Toolkit following the [CUDA Installation Guide for Linux](https://docs.nvidia.com/cuda/cuda-installation-guide-linux/) and
make sure `CUDA_HOME` environment variable is properly set.
```{tip}
:name: installation-linux-tip-cuda-version
TensorRT LLM 1.1 supports both CUDA 12.9 and 13.0. The wheel package release only supports CUDA 12.9, while CUDA 13.0 is only supported through NGC container release.
```
```bash
# Optional step: Only required for NVIDIA Blackwell GPUs and SBSA platform
pip3 install torch==2.7.1 torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128
sudo apt-get -y install libopenmpi-dev
# Optional step: Only required for disagg-serving
sudo apt-get -y install libzmq3-dev
```
PyTorch CUDA 12.8 package is required for supporting NVIDIA Blackwell GPUs and SBSA platform. On prior GPUs or Linux x86_64 platform, this extra installation is not required.
```{tip}
Instead of manually installing the preqrequisites as described
above, it is also possible to use the pre-built [TensorRT LLM Develop container
image hosted on NGC](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/tensorrt-llm/containers/devel)
(see [here](containers) for information on container tags).
```
### Install pre-built TensorRT LLM wheel
Once all prerequisites are in place, TensorRT LLM can be installed as follows:
```bash
pip3 install --upgrade pip setuptools && pip3 install tensorrt_llm
```
**This project will download and install additional third-party open source software projects. Review the license terms of these open source projects before use.**
2. Sanity check the installation by running the following in Python (tested on Python 3.12):
```{literalinclude} ../../../examples/llm-api/quickstart_example.py
:language: python
:linenos:
```
**Known limitations**
There are some known limitations when you pip install pre-built TensorRT LLM wheel package.
1. MPI in the Slurm environment
If you encounter an error while running TensorRT LLM in a Slurm-managed cluster, you need to reconfigure the MPI installation to work with Slurm.
The setup methods depends on your slurm configuration, pls check with your admin. This is not a TensorRT LLM specific, rather a general mpi+slurm issue.
```
The application appears to have been direct launched using "srun",
but OMPI was not built with SLURM support. This usually happens
when OMPI was not configured --with-slurm and we weren't able
to discover a SLURM installation in the usual places.
```