mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
[None][ci] parallelize unit tests of auto deploy in B200 (#7291)
Signed-off-by: junq <22017000+QiJune@users.noreply.github.com>
This commit is contained in:
parent
8dc62ffac4
commit
d09add5ede
@ -103,5 +103,7 @@ unittest/trt/model/test_mamba.py,NVIDIA H100,10,
|
||||
"unittest/trt/attention/test_gpt_attention.py -k ""xqa_generic""",NVIDIA L40,3,
|
||||
unittest/_torch/speculative,NVIDIA Graphics Device,4,B200 Bring Up Board
|
||||
unittest/_torch/thop,NVIDIA Graphics Device,32,B200 Bring Up Board
|
||||
"unittest/_torch/auto_deploy/unit/singlegpu -k ""not test_trtllm_bench_backend_comparison""",NVIDIA Graphics Device,4,B200 Bring Up Board
|
||||
unittest/_torch/speculative,NVIDIA B200,4,
|
||||
unittest/_torch/thop,NVIDIA B200,32,
|
||||
"unittest/_torch/auto_deploy/unit/singlegpu -k ""not test_trtllm_bench_backend_comparison""",NVIDIA B200,4,
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user