TensorRT-LLMs/examples/configs
Anish Shanbhag 6a6317727b
[TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173)
Signed-off-by: Anish Shanbhag <ashanbhag@nvidia.com>
2025-11-03 17:42:41 -08:00
..
deepseek-r1-deepgemm.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
deepseek-r1-latency.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
deepseek-r1-throughput.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
gpt-oss-120b-latency.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
gpt-oss-120b-throughput.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
llama-3.3-70b.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
llama-4-scout.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
qwen3-disagg-prefill.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
qwen3-next.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
qwen3.yaml [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00
README.md [TRTLLM-8680][doc] Add table with one-line deployment commands to docs (#8173) 2025-11-03 17:42:41 -08:00

Recommended LLM API Configuration Settings

This directory contains recommended LLM API performance settings for popular models. They can be used out-of-the-box with trtllm-serve via the --extra_llm_api_options CLI flag, or you can adjust them to your specific use case.

For model-specific deployment guides, please refer to the official documentation.