mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
* tests: skip writing prepare_dataset output to logs Signed-off-by: Ruodi <200874449+ruodil@users.noreply.github.com> * test: add llama_v3.1_8b_fp8 model, llama_v3.1_405b model and llama_nemotron_49b model in perf test, and modify original llama models dtype from float16 to bfloat16 according to README.md Signed-off-by: Ruodi <200874449+ruodil@users.noreply.github.com> --------- Signed-off-by: Ruodi <200874449+ruodil@users.noreply.github.com> Signed-off-by: Larry <197874197+LarryXFly@users.noreply.github.com> Co-authored-by: Larry <197874197+LarryXFly@users.noreply.github.com> |
||
|---|---|---|
| .. | ||
| .gitignore | ||
| examples_test_list.txt | ||
| llm_multinodes_function_test.txt | ||
| llm_release_perf_multinode_test.txt | ||
| llm_sanity_test.txt | ||
| trt_llm_integration_perf_sanity_test.yml | ||
| trt_llm_integration_perf_test.yml | ||
| trt_llm_release_perf_cluster_test.yml | ||
| trt_llm_release_perf_sanity_test.yml | ||
| trt_llm_release_perf_test.yml | ||