mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-13 22:18:36 +08:00
Signed-off-by: Jatin Gangani <jgangani@dc2-container-xterm-014.prd.it.nvidia.com> Co-authored-by: Jatin Gangani <jgangani@dc2-container-xterm-014.prd.it.nvidia.com> |
||
|---|---|---|
| .. | ||
| curated | ||
| database | ||
| __init__.py | ||
| README.md | ||
Recommended LLM API Configuration Settings
This directory contains recommended LLM API performance settings for popular models. They can be used out-of-the-box with trtllm-serve via the --config CLI flag, or you can adjust them to your specific use case.
For model-specific deployment guides, please refer to the official documentation.