mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
[https://nvbugspro.nvidia.com/bug/5270564][test] skip per-hopper for llama4 (#4211)
skip per-hopper for llama4 Signed-off-by: Ivy Zhang <25222398+crazydemo@users.noreply.github.com>
This commit is contained in:
parent
ba13b51a58
commit
ee92edf2b4
@ -202,6 +202,7 @@ class TestLlama4MaverickInstruct(LlmapiAccuracyTestHarness):
|
||||
MODEL_NAME = "meta-llama/Llama-4-Maverick-17B-128E-Instruct"
|
||||
MODEL_PATH = f"{llm_models_root()}/llama4-models/Llama-4-Maverick-17B-128E-Instruct"
|
||||
|
||||
@skip_pre_hopper
|
||||
@pytest.mark.skip_less_device(8)
|
||||
@parametrize_with_ids("cuda_graph", [False, True])
|
||||
@pytest.mark.parametrize("tp_size,pp_size,ep_size", [(8, 1, 1), (8, 1, 4),
|
||||
@ -223,6 +224,7 @@ class TestLlama4ScoutInstruct(LlmapiAccuracyTestHarness):
|
||||
MODEL_NAME = "meta-llama/Llama-4-Scout-17B-16E-Instruct"
|
||||
MODEL_PATH = f"{llm_models_root()}/llama4-models/Llama-4-Scout-17B-16E-Instruct"
|
||||
|
||||
@skip_pre_hopper
|
||||
@pytest.mark.skip_less_device(8)
|
||||
@parametrize_with_ids("cuda_graph", [False, True])
|
||||
@pytest.mark.parametrize("tp_size,pp_size,ep_size", [(8, 1, 1), (8, 1, 4),
|
||||
|
||||
Loading…
Reference in New Issue
Block a user