mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
[https://nvbugspro.nvidia.com/bug/5260676]test: skip fp8 quantization case for pre-ada (#4095)
skip pre ada Signed-off-by: Ivy Zhang <25222398+crazydemo@users.noreply.github.com>
This commit is contained in:
parent
c9cac432dc
commit
c2d4c2adb6
@ -299,6 +299,7 @@ def test_mistral_eagle_1gpu(llm_mistral_model_root,
|
||||
llm_rouge_root=llm_rouge_root)
|
||||
|
||||
|
||||
@skip_pre_ada
|
||||
@pytest.mark.parametrize("use_dynamic_tree", [False, True],
|
||||
ids=['eagle1', 'eagle2'])
|
||||
@pytest.mark.parametrize("mistral_nemo_model_root", ['Mistral-Nemo-12b-Base'],
|
||||
|
||||
Loading…
Reference in New Issue
Block a user