TensorRT-LLMs/tests/unittest/llmapi
Yan Chunwei 5999fab146 [https://nvbugs/5427043][fix] cherrypick: request length exceeds max_num_tokens (#7718)
Signed-off-by: Superjomn <328693+Superjomn@users.noreply.github.com>
Signed-off-by: Wangshanshan <30051912+dominicshanshan@users.noreply.github.com>
2025-09-25 21:02:35 +08:00
..
apps [TRTLLM-5235][feat] Enable regex and EBNF grammar in trtllm-serve (#7925) 2025-09-24 18:30:23 +08:00
__init__.py test: reorganize tests folder hierarchy (#2996) 2025-03-27 12:07:53 +08:00
_run_mpi_comm_task.py fix[nvbug5298640]: trtllm-llmapi-launch multiple LLM instances (#4727) 2025-06-19 06:13:53 +08:00
_test_remote_mpi_session.sh fix [nvbug/5351244]: address remote mpi session submit (#5664) 2025-07-22 12:48:00 +08:00
fake.sh doc: fix path after examples migration (#3814) 2025-04-24 02:36:45 +08:00
lora_test_utils.py Add basic Nemo Ckpt Lora Loading in pytorch flow (#6019) 2025-07-22 19:42:45 -07:00
run_llm_exit.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
run_llm_with_postproc.py [nvbug 5327706][fix] fix mgmn postprocess error (#5835) 2025-07-14 17:17:30 +08:00
run_llm.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_build_cache.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
test_executor.py [https://nvbugs/5477359][fix] Removing test waivers (#7877) 2025-09-22 08:59:13 -07:00
test_gc_utils.py [nvbug 5273941] fix: broken cyclic reference detect (#5417) 2025-07-01 20:12:55 +08:00
test_llm_args.py [None][chore] Remove executor config in create_py_executor (#7599) 2025-09-18 14:24:58 +08:00
test_llm_download.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_llm_kv_cache_events.py [https://nvbugs/5444937][chore] Fixing KV events tests (#7004) 2025-08-19 11:18:04 -04:00
test_llm_models.py [https://nvbugs/5371480][fix] Enable test_phi3_small_8k (#6938) 2025-08-19 09:42:35 +08:00
test_llm_multi_gpu_pytorch.py [None][fix] Refactoring to avoid circular import when importing torch models (#6720) 2025-08-11 18:00:42 -04:00
test_llm_multi_gpu.py [None][infra] Skip failed test for nvbugs 5532023 (#7905) 2025-09-22 03:49:44 -07:00
test_llm_pytorch.py [https://nvbugs/5427043][fix] cherrypick: request length exceeds max_num_tokens (#7718) 2025-09-25 21:02:35 +08:00
test_llm_quant.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_llm_utils.py [nvbug/5387226] chore: add propogation for trust_remote_code to AutoConfig (#6001) 2025-07-16 16:05:38 +08:00
test_llm.py [#7692][fix] recognize RequestError as per-request error in background handler (#7726) 2025-09-24 11:11:17 +08:00
test_mpi_session.py [https://nvbugs/5351244][fix] test_mpi_session (#7501) 2025-09-22 14:28:38 +08:00
test_reasoning_parser.py feat: add deepseek-r1 reasoning parser to trtllm-serve (#3354) 2025-05-06 08:13:04 +08:00
test_serialization.py [TRTLLM-4971]: Use safe deserialization in ParallelConfig (#4630) 2025-06-27 09:58:41 +08:00
test_utils.py [None][chore] Mass integration of release/1.0 - 3rd (#7519) 2025-09-08 14:03:04 +08:00