[None][doc] Hardware support update (#10719)

Signed-off-by: Patrice Castonguay <55748270+pcastonguay@users.noreply.github.com>
Signed-off-by: Wangshanshan <30051912+dominicshanshan@users.noreply.github.com>
This commit is contained in:
Patrice Castonguay 2026-01-29 13:44:37 -05:00 committed by Yanchao Lu
parent ed404f9298
commit d9f787a8d2
3 changed files with 8 additions and 7 deletions

View File

@ -14,6 +14,7 @@ Welcome to TensorRT LLM's Documentation!
overview.md
quick-start-guide.md
installation/index.rst
supported-hardware.md
.. toctree::

View File

@ -51,13 +51,6 @@ TensorRT LLM strives to support the most popular models on **Day 0**.
- **[Guided Decoding](./features/guided-decoding.md)**: Advanced sampling with stop words, bad words, and custom constraints
- **[Disaggregated Serving (Beta)](./features/disagg-serving.md)**: Separate context and generation phases across different GPUs for optimal resource utilization
### 🔧 **Latest GPU Architecture Support**
TensorRT LLM supports the full spectrum of NVIDIA GPU architectures:
- **NVIDIA Blackwell**: B200, GB200, B300, GB300, and RTX Pro 6000 SE with FP4 optimization
- **NVIDIA Ada Lovelace**: L40/L40S, RTX 40 series with FP8 acceleration
- **NVIDIA Ampere**: A100, RTX 30 series for production workloads
## What Can You Do With TensorRT LLM?
Whether you're building the next generation of AI applications, optimizing existing LLM deployments, or exploring the frontiers of large language model technology, TensorRT LLM provides the tools, performance, and flexibility you need to succeed in the era of generative AI.To get started, refer to the {ref}`quick-start-guide`.

View File

@ -0,0 +1,7 @@
# Supported Hardware
TensorRT LLM supports the full spectrum of NVIDIA GPU architectures:
- **NVIDIA Blackwell**: B200, GB200, B300, GB300, DGX Spark
- **NVIDIA Hopper**: H100, H200, GH200
- **NVIDIA Ada Lovelace**: L20, L40/L40S
- **NVIDIA Ampere**: A100