mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-02-04 02:02:01 +08:00
[https://nvbugs/5800521][ci] Move test_openai_chat_guided_decoding to H100 stage (to avoid potential OOM) (#10703)
Signed-off-by: Enwei Zhu <21126786+syuoni@users.noreply.github.com>
This commit is contained in:
parent
ce561b6a8e
commit
9f741fb254
@ -53,7 +53,6 @@ l0_a10:
|
||||
- disaggregated/test_disaggregated_single_gpu.py::test_disaggregated_simple_llama[False-True-TinyLlama-1.1B-Chat-v1.0]
|
||||
- disaggregated/test_disaggregated_single_gpu.py::test_disaggregated_simple_llama[True-False-TinyLlama-1.1B-Chat-v1.0]
|
||||
- disaggregated/test_disaggregated_single_gpu.py::test_disaggregated_simple_llama[True-True-TinyLlama-1.1B-Chat-v1.0]
|
||||
- test_e2e.py::test_openai_chat_guided_decoding
|
||||
- test_e2e.py::test_get_ci_container_port
|
||||
- test_e2e.py::test_openai_chat_multimodal_example ISOLATION
|
||||
- test_e2e.py::test_openai_mmencoder_example
|
||||
|
||||
@ -110,6 +110,7 @@ l0_h100:
|
||||
- test_e2e.py::test_trtllm_bench_help_sanity[meta-llama/Llama-3.1-8B]
|
||||
- test_e2e.py::test_openai_chat_harmony
|
||||
- test_e2e.py::test_openai_responses
|
||||
- test_e2e.py::test_openai_chat_guided_decoding
|
||||
- test_e2e.py::test_trtllm_benchmark_serving[llama-3.1-model/Meta-Llama-3.1-8B]
|
||||
- condition:
|
||||
ranges:
|
||||
|
||||
Loading…
Reference in New Issue
Block a user