TensorRT-LLMs/docs
dhansen-nvidia 80235e53cf [None][feat] Add documentation on configuring CPU affinity in TRT-LLM (#10678)
Signed-off-by: Dan Hansen <1+dhansen-nvidia@users.noreply.github.com>
Signed-off-by: dhansen-nvidia <218031328+dhansen-nvidia@users.noreply.github.com>
Co-authored-by: Dan Hansen <1+dhansen-nvidia@users.noreply.github.com>
Co-authored-by: Kaiyu Xie <26294424+kaiyux@users.noreply.github.com>
Signed-off-by: Wangshanshan <30051912+dominicshanshan@users.noreply.github.com>
2026-02-15 19:57:03 +08:00
..
source [None][feat] Add documentation on configuring CPU affinity in TRT-LLM (#10678) 2026-02-15 19:57:03 +08:00
Doxygen Update TensorRT-LLM (#1315) 2024-03-19 17:36:42 +08:00
make.bat Kaiyu/update main (#5) 2023-10-18 22:38:53 +08:00
Makefile Kaiyu/update main (#5) 2023-10-18 22:38:53 +08:00
README.md Kaiyu/update main (#5) 2023-10-18 22:38:53 +08:00
requirements.txt [#10966][feat] AutoDeploy: kv cache manager integration [2/2] (#11149) 2026-02-04 09:44:27 -05:00

Docs

This directory contains the stuff for building static html documentations based on sphinx.

Build the docs

Firstly, install the sphinx:

apt-get install python3-sphinx doxygen python3-pip graphviz

Secondly, install the packages:

python3 -m pip install -r ./requirements.txt

And then, make the docs:

doxygen Doxygen # build C++ docs

make html

And the finally the generated html pages will locate in the build/html directory.

Preview the docs locally

The basic way to preview the docs is using the http.serve:

cd build/html

python3 -m http.server 8081

And you can visit the page with your web browser with url http://localhost:8081.