TensorRT-LLMs/tests/unittest/llmapi
Kaiyu Xie bb5b16fcb9
feat: Return context response immediately when stream_interval > 1 (#5836)
Signed-off-by: Kaiyu Xie <26294424+kaiyux@users.noreply.github.com>
2025-07-09 00:19:57 +09:00
..
apps [feat] Detokenize option in /v1/completions request (#5382) 2025-07-08 19:36:04 +08:00
__init__.py test: reorganize tests folder hierarchy (#2996) 2025-03-27 12:07:53 +08:00
_run_mpi_comm_task.py fix[nvbug5298640]: trtllm-llmapi-launch multiple LLM instances (#4727) 2025-06-19 06:13:53 +08:00
_test_remote_mpi_session.sh fix [nvbug5351244]: test_mpi_session submit sync/async (#5608) 2025-07-04 13:14:13 +08:00
fake.sh doc: fix path after examples migration (#3814) 2025-04-24 02:36:45 +08:00
run_llm_exit.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
run_llm_with_postproc.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
run_llm.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_build_cache.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
test_executor.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_gc_utils.py [nvbug 5273941] fix: broken cyclic reference detect (#5417) 2025-07-01 20:12:55 +08:00
test_llm_args.py Revert "chore: [Breaking Change] Rename cuda_graph_config padding_enabled fie…" (#5818) 2025-07-08 13:15:30 +09:00
test_llm_download.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_llm_kv_cache_events.py [Infra] - Add import pytest (#5565) 2025-06-29 11:06:14 +08:00
test_llm_models.py [Infra] - Waive failed tests for main 0702 (#5671) 2025-07-02 22:05:07 -04:00
test_llm_multi_gpu_pytorch.py [ci] small multigpu speedups (#5643) 2025-07-03 08:06:10 -04:00
test_llm_multi_gpu.py [NvBug 5362426] fix: Fix prompt adapter TP2 case (#5782) 2025-07-08 16:01:36 +09:00
test_llm_pytorch.py [ci] small multigpu speedups (#5643) 2025-07-03 08:06:10 -04:00
test_llm_quant.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_llm_utils.py [TRTLLM-5208][BREAKING CHANGE] chore: make pytorch LLM the default (#5312) 2025-06-20 03:01:10 +08:00
test_llm.py feat: Return context response immediately when stream_interval > 1 (#5836) 2025-07-09 00:19:57 +09:00
test_mpi_session.py [Infra] - Waive failed cases on release/0.21 (#5674) 2025-07-04 13:14:13 +08:00
test_reasoning_parser.py feat: add deepseek-r1 reasoning parser to trtllm-serve (#3354) 2025-05-06 08:13:04 +08:00
test_serialization.py [TRTLLM-4971]: Use safe deserialization in ParallelConfig (#4630) 2025-06-27 09:58:41 +08:00