diff --git a/docs/source/index.rst b/docs/source/index.rst index f4246d40a5..a4ca7804e0 100644 --- a/docs/source/index.rst +++ b/docs/source/index.rst @@ -14,6 +14,7 @@ Welcome to TensorRT LLM's Documentation! overview.md quick-start-guide.md installation/index.rst + supported-hardware.md .. toctree:: diff --git a/docs/source/overview.md b/docs/source/overview.md index e7aeed43a2..0eee6b6f9f 100644 --- a/docs/source/overview.md +++ b/docs/source/overview.md @@ -51,13 +51,6 @@ TensorRT LLM strives to support the most popular models on **Day 0**. - **[Guided Decoding](./features/guided-decoding.md)**: Advanced sampling with stop words, bad words, and custom constraints - **[Disaggregated Serving (Beta)](./features/disagg-serving.md)**: Separate context and generation phases across different GPUs for optimal resource utilization -### 🔧 **Latest GPU Architecture Support** - -TensorRT LLM supports the full spectrum of NVIDIA GPU architectures: -- **NVIDIA Blackwell**: B200, GB200, B300, GB300, and RTX Pro 6000 SE with FP4 optimization -- **NVIDIA Ada Lovelace**: L40/L40S, RTX 40 series with FP8 acceleration -- **NVIDIA Ampere**: A100, RTX 30 series for production workloads - ## What Can You Do With TensorRT LLM? Whether you're building the next generation of AI applications, optimizing existing LLM deployments, or exploring the frontiers of large language model technology, TensorRT LLM provides the tools, performance, and flexibility you need to succeed in the era of generative AI.To get started, refer to the {ref}`quick-start-guide`. diff --git a/docs/source/supported-hardware.md b/docs/source/supported-hardware.md new file mode 100644 index 0000000000..a8c96e99ee --- /dev/null +++ b/docs/source/supported-hardware.md @@ -0,0 +1,7 @@ +# Supported Hardware + +TensorRT LLM supports the full spectrum of NVIDIA GPU architectures: +- **NVIDIA Blackwell**: B200, GB200, B300, GB300, DGX Spark +- **NVIDIA Hopper**: H100, H200, GH200 +- **NVIDIA Ada Lovelace**: L20, L40/L40S +- **NVIDIA Ampere**: A100