TensorRT-LLMs/tests/unittest/llmapi
JunyiXu-nv 743486b2ea
[TRTLLM-6842][feat] Support Response API for general purpose (#9392)
Signed-off-by: Junyi Xu <219237550+JunyiXu-nv@users.noreply.github.com>
2025-12-03 16:49:26 +08:00
..
apps [TRTLLM-6842][feat] Support Response API for general purpose (#9392) 2025-12-03 16:49:26 +08:00
__init__.py test: reorganize tests folder hierarchy (#2996) 2025-03-27 12:07:53 +08:00
_run_mpi_comm_task.py fix[nvbug5298640]: trtllm-llmapi-launch multiple LLM instances (#4727) 2025-06-19 06:13:53 +08:00
_run_multi_llm_tasks.py [https://nvbugs/5437384][test] CHERRY-PICK: fix trtllm-llmapi-launch multi tests (#8567) 2025-11-01 06:49:33 -07:00
_run_multi_mpi_comm_tasks.py [https://nvbugs/5437384][test] CHERRY-PICK: fix trtllm-llmapi-launch multi tests (#8567) 2025-11-01 06:49:33 -07:00
_test_remote_mpi_session.sh fix [nvbug/5351244]: address remote mpi session submit (#5664) 2025-07-22 12:48:00 +08:00
fake.sh doc: fix path after examples migration (#3814) 2025-04-24 02:36:45 +08:00
lora_test_utils.py [https://nvbugs/5510879][fix] Fix pytorch & TRT-python flows fused LoRA adapter modules weight split with TP>1 (#8063) 2025-10-12 12:29:52 -07:00
run_llm_exit.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
run_llm_with_postproc.py [nvbug 5327706][fix] fix mgmn postprocess error (#5835) 2025-07-14 17:17:30 +08:00
run_llm.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_additional_model_outputs.py [TRTLLM-4517] [feat] Additional model outputs (#7206) 2025-10-13 15:33:18 +02:00
test_build_cache.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
test_executor.py [https://nvbugs/5477359][fix] Removing test waivers (#7877) 2025-09-22 08:59:13 -07:00
test_gc_utils.py [nvbug 5273941] fix: broken cyclic reference detect (#5417) 2025-07-01 20:12:55 +08:00
test_llm_args.py [TRTLLM-9302][chore] Move build config from BaseLlmArgs to TrtLlmArgs (#9249) 2025-11-24 10:54:41 +08:00
test_llm_download.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_llm_kv_cache_events.py [https://nvbugs/5444937][chore] Fixing KV events tests (#7004) 2025-08-19 11:18:04 -04:00
test_llm_models.py [https://nvbugs/5371480][fix] Enable test_phi3_small_8k (#6938) 2025-08-19 09:42:35 +08:00
test_llm_multi_gpu_pytorch.py [TRTLLM-9144][fix] enhance RPC robustness (#8711) 2025-12-02 21:37:59 +08:00
test_llm_multi_gpu.py [TRTLLM-8682][chore] Remove auto_parallel module (#8329) 2025-10-22 20:53:08 -04:00
test_llm_pytorch.py [TRTLLM-9144][fix] enhance RPC robustness (#8711) 2025-12-02 21:37:59 +08:00
test_llm_quant.py [https://nvbugs/5558117][fix] Allow per-layer quant config from hf_quant_config.json (#8617) 2025-10-31 04:41:44 -07:00
test_llm_utils.py [TRTLLM-8189][chore] enhance GenerationExecutor with RPC (part1) (#5543) 2025-10-05 17:28:20 +08:00
test_llm.py [TRTLLM-909][feat] Overlap context chunks in pipeline parallel mode (#9308) 2025-11-25 22:11:51 +01:00
test_memory_profiling.py [TRTLLM-6780][fix] Add multimodal data to dummy requests during memory profiling (#7539) 2025-10-16 17:49:22 +02:00
test_mpi_session.py [https://nvbugs/5437384][test] CHERRY-PICK: fix trtllm-llmapi-launch multi tests (#8567) 2025-11-01 06:49:33 -07:00
test_reasoning_parser.py [None][feat] Support Qwen3 reasoning parser (#8000) 2025-10-21 14:08:39 +08:00
test_serialization.py [TRTLLM-8682][chore] Remove auto_parallel module (#8329) 2025-10-22 20:53:08 -04:00
test_utils.py [None][chore] Mass integration of release/1.0 - 3rd (#7519) 2025-09-08 14:03:04 +08:00