mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
* doc: [TRTLLM-325]Integrate the NGC image in Makefile automation and documentation Signed-off-by: Martin Marciniszyn Mehringer <11665257+MartinMarciniszyn@users.noreply.github.com> * WAR against https://github.com/advisories/GHSA-vqfr-h8mv-ghfj Signed-off-by: Martin Marciniszyn Mehringer <11665257+MartinMarciniszyn@users.noreply.github.com> * Fix default assignment for CUDA architectures in SBSA build Signed-off-by: Martin Marciniszyn Mehringer <11665257+MartinMarciniszyn@users.noreply.github.com> * Push new docker images Signed-off-by: Martin Marciniszyn Mehringer <11665257+MartinMarciniszyn@users.noreply.github.com> * Handle constraints.txt in setup.py Signed-off-by: Martin Marciniszyn Mehringer <11665257+MartinMarciniszyn@users.noreply.github.com> --------- Signed-off-by: Martin Marciniszyn Mehringer <11665257+MartinMarciniszyn@users.noreply.github.com>
116 lines
4.3 KiB
Markdown
116 lines
4.3 KiB
Markdown
# The Docker Build System
|
|
|
|
## Multi-stage Builds with Docker
|
|
|
|
TensorRT-LLM can be compiled in Docker using a multi-stage build implemented in [`Dockerfile.multi`](Dockerfile.multi).
|
|
The following build stages are defined:
|
|
|
|
* `devel`: this image provides all dependencies for building TensorRT-LLM.
|
|
* `wheel`: this image contains the source code and the compiled binary distribution.
|
|
* `release`: this image has the binaries installed and contains TensorRT-LLM examples in `/app/tensorrt_llm`.
|
|
|
|
## Building Docker Images with GNU `make`
|
|
|
|
The GNU [`Makefile`](Makefile) in the `docker` directory provides targets for building, pushing, and running each stage
|
|
of the Docker build. The corresponding target names are composed of two components, namely, `<stage>` and `<action>`
|
|
separated by `_`. The following actions are available:
|
|
|
|
* `<stage>_build`: builds the docker image for the stage.
|
|
* `<stage>_push`: pushes the docker image for the stage to a docker registry (implies `<stage>_build`).
|
|
* `<stage>_run`: runs the docker image for the stage in a new container.
|
|
|
|
For example, the `release` stage is built and pushed from the top-level directory of TensorRT-LLM as follows:
|
|
|
|
```bash
|
|
make -C docker release_push
|
|
```
|
|
|
|
Note that pushing the image to a docker registry is optional. After building an image, run it in a new container with
|
|
```bash
|
|
make -C docker release_run
|
|
```
|
|
|
|
### Building and Running Options
|
|
|
|
The full image name and tag can be controlled by supplying `IMAGE_WITH_TAG` to `make`:
|
|
|
|
```bash
|
|
make -C docker devel_push IMAGE_WITH_TAG="urm.nvidia.com/sw-tensorrt-docker/tensorrt-llm:dev"
|
|
```
|
|
|
|
Containers can be started with the local user instead of `root` by appending `LOCAL_USER=1` to the run target:
|
|
|
|
```bash
|
|
make -C docker devel_run LOCAL_USER=1
|
|
```
|
|
|
|
Specific CUDA architectures supported by the `wheel` can be specified WITH `CUDA_ARCHS`:
|
|
|
|
```bash
|
|
make -C docker release_build CUDA_ARCHS="80-real;90-real"
|
|
```
|
|
|
|
For more build options, see the variables defined in [`Makefile`](Makefile).
|
|
|
|
### NGC Integration
|
|
|
|
When building from source, one can conveniently download a docker image for development from
|
|
the [NVIDIA NGC Catalog](https://catalog.ngc.nvidia.com/) and start it like so:
|
|
|
|
```bash
|
|
make -C docker ngc-devel_run LOCAL_USER=1 DOCKER_PULL=1
|
|
```
|
|
|
|
As before, specifying `LOCAL_USER=1` will run the container with the local user's identity. Specifying `DOCKER_PULL=1`
|
|
is optional, but it will pull the latest image from the NGC Catalog. This will map the source code into the container
|
|
in the directory `/code/tensorrt_llm`.
|
|
|
|
We also provide an image with pre-installed binaries for release. This can be used like so:
|
|
|
|
```bash
|
|
make -C docker ngc-release_run LOCAL_USER=1 DOCKER_PULL=1
|
|
```
|
|
|
|
If you want to deploy a specific version of TensorRT-LLM, you can specify the version with
|
|
`TRT_LLM_VERSION=<version_tag>`. The application examples and benchmarks are installed in `/app/tensorrt_llm`.
|
|
|
|
### Jenkins Integration
|
|
|
|
[`Makefile`](Makefile) has special targets for building, pushing and running the Docker build image used on Jenkins.
|
|
The full image name and tag is defined in [`L0_MergeRequest.groovy`](../jenkins/L0_MergeRequest.groovy). The `make`
|
|
system will parse this name as the value of `LLM_DOCKER_IMAGE`. To build and push a new Docker image for Jenkins,
|
|
define a new image name and tag in [`L0_MergeRequest.groovy`](../jenkins/L0_MergeRequest.groovy) and run
|
|
|
|
```bash
|
|
make -C docker jenkins_push
|
|
```
|
|
|
|
Start a new container using the same image as Jenkins using your local user account with
|
|
|
|
```bash
|
|
make -C docker jenkins_run LOCAL_USER=1
|
|
```
|
|
|
|
One may also build a release image based on the Jenkins development image:
|
|
|
|
```bash
|
|
make -C docker trtllm_build CUDA_ARCHS="80-real;90-real"
|
|
```
|
|
|
|
These images can be pushed to
|
|
the [internal artifact repository](https://urm.nvidia.com/artifactory/sw-tensorrt-docker/tensorrt-llm-staging/release/):
|
|
|
|
```bash
|
|
make -C docker trtllm_push
|
|
```
|
|
|
|
Generally, only images built for all CUDA architectures should be pushed to the artifact repository. These images can
|
|
be deployed in docker in the usual way:
|
|
|
|
```bash
|
|
make -C docker trtllm_run LOCAL_USER=1 DOCKER_PULL=1
|
|
```
|
|
|
|
The argument `DOCKER_PULL=1` instructs `make` to pull the latest version of the image before deploying it in the container.
|
|
By default, images are tagged by their `git` branch name and may be frequently updated.
|