mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
[CI] waive two multi-gpu test cases (#4206)
waive two multi-gpu test cases Signed-off-by: junq <22017000+QiJune@users.noreply.github.com>
This commit is contained in:
parent
7db368c72c
commit
f021afa241
@ -481,5 +481,6 @@ accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_fp8_block_scales_4gpu
|
||||
accuracy/test_disaggregated_serving.py::TestLlama3_1_8B::test_auto_dtype[False] SKIP (https://nvbugs/5266257)
|
||||
accuracy/test_disaggregated_serving.py::TestLlama3_1_8B::test_auto_dtype[True] SKIP (https://nvbugs/5266257)
|
||||
disaggregated/test_disaggregated.py::test_disaggregated_cuda_graph[TinyLlama-1.1B-Chat-v1.0] SKIP (https://nvbugs/5247271)
|
||||
disaggregated/test_disaggregated.py::test_disaggregated_deepseek_v3_lite_fp8_tp1_attention_dp_overlap_one_mtp[DeepSeek-V3-Lite-fp8] SKIP (https://nvbugspro.nvidia.com/bug/5273945)
|
||||
accuracy/test_llm_api_pytorch.py::TestLlama4MaverickInstruct::test_auto_dtype[tp8-] SKIP (https://nvbugs/5270564)
|
||||
accuracy/test_llm_api_pytorch.py::TestLlama4ScoutInstruct::test_auto_dtype[tp8-] SKIP (https://nvbugs/5270564)
|
||||
|
||||
@ -77,6 +77,7 @@ def engine_from_checkpoint() -> tempfile.TemporaryDirectory:
|
||||
@pytest.mark.part0
|
||||
def test_llm_loading_from_ckpt_for_tp2(
|
||||
engine_from_checkpoint: tempfile.TemporaryDirectory):
|
||||
pytest.skip(reason="https://nvbugspro.nvidia.com/bug/5273941")
|
||||
tokenizer = TransformersTokenizer.from_pretrained(llama_model_path)
|
||||
llm_test_harness(engine_from_checkpoint.name,
|
||||
prompts, ["D E F G H I J K"],
|
||||
|
||||
Loading…
Reference in New Issue
Block a user