mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
[Test] - Correct waive the Slurm test stage (#4680)
Signed-off-by: Yanchao Lu <yanchaol@nvidia.com>
This commit is contained in:
parent
8eafe83c37
commit
cbb6a264be
@ -1454,9 +1454,8 @@ def launchTestJobs(pipeline, testFilter, dockerNode=null)
|
||||
|
||||
fullSet = parallelJobs.keySet()
|
||||
|
||||
turtleSlurmConfigs = [
|
||||
"RTXPro6000-PyTorch-[Post-Merge]-1": ["rtx-pro-6000", "l0_rtx_pro_6000", 1, 1],
|
||||
]
|
||||
turtleSlurmConfigs = [:]
|
||||
// "RTXPro6000-PyTorch-[Post-Merge]-1": ["rtx-pro-6000", "l0_rtx_pro_6000", 1, 1],
|
||||
|
||||
// TODO: use cpu pod to launch slurm job
|
||||
parallelSlurmJobs = turtleSlurmConfigs.collectEntries{key, values -> [key, [createKubernetesPodConfig(LLM_DOCKER_IMAGE, "a10", "amd64", values[4] ?: 1, key.contains("Perf")), {
|
||||
|
||||
Loading…
Reference in New Issue
Block a user