mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
Signed-off-by: Frank Di Natale <3429989+FrankD412@users.noreply.github.com> Signed-off-by: Venky Ganesh <23023424+venkywonka@users.noreply.github.com> Co-authored-by: Frank Di Natale <3429989+FrankD412@users.noreply.github.com> |
||
|---|---|---|
| .. | ||
| curated | ||
| database | ||
| README.md | ||
Recommended LLM API Configuration Settings
This directory contains recommended LLM API performance settings for popular models. They can be used out-of-the-box with trtllm-serve via the --extra_llm_api_options CLI flag, or you can adjust them to your specific use case.
For model-specific deployment guides, please refer to the official documentation.