TensorRT-LLMs/tests/unittest/_torch
Wanli Jiang 2a30f11d63
[None][chore] Upgrade transformers to 4.56.0 (#7523)
Signed-off-by: Wanli Jiang <35160485+Wanli-Jiang@users.noreply.github.com>
Signed-off-by: Enwei Zhu <21126786+syuoni@users.noreply.github.com>
Co-authored-by: Enwei Zhu <21126786+syuoni@users.noreply.github.com>
2025-09-22 22:20:16 +08:00
..
attention [https://nvbugs/5453806][unwaive] Unwaive fp8 kvcache attention test (#7243) 2025-09-05 12:13:57 -04:00
auto_deploy [#7308] [feat] AutoDeploy: graph-less transformers mode for HF (#7635) 2025-09-18 10:44:24 +08:00
compilation [TRTLLM-3105][feat] Add Piecewise CUDA Graph Support (#3804) 2025-05-09 11:04:01 +08:00
debugger Fix: fix nvbug 5356427 (#5464) 2025-06-25 22:24:26 +08:00
executor [https://nvbugs/5503440][fix] Fix potential hang due to wrong type of ZMQ socket and protocol for worker_init_status_queue (#7646) 2025-09-19 18:13:33 +08:00
misc [TRTLLM-4629] [feat] Add support of CUDA13 and sm103 devices (#7568) 2025-09-16 09:56:18 +08:00
modeling [None][chore] Upgrade transformers to 4.56.0 (#7523) 2025-09-22 22:20:16 +08:00
models/checkpoints/hf [None][feat] Skip prefetching consolidated safetensors when appropriate (#7013) 2025-08-25 23:56:21 -04:00
modules [TRTLLM-7385][feat] Optimize Qwen2/2.5-VL performance (#7250) 2025-09-22 03:40:02 -07:00
multi_gpu [TRTLLM-4629] [feat] Add support of CUDA13 and sm103 devices (#7568) 2025-09-16 09:56:18 +08:00
multi_gpu_modeling [None][ci] Test waives for the release/1.0 branch 09/15 (#7700) 2025-09-22 14:28:38 +08:00
multimodal [TRTLLM-7385][feat] Optimize Qwen2/2.5-VL performance (#7250) 2025-09-22 03:40:02 -07:00
sampler [TRTLLM-7155][feat] Unify sampler handle logits implementation. (#6867) 2025-08-22 08:09:30 +02:00
speculative [https://nvbugs/5517404][fix] Use the correct cuda graph for dynamic spec dec (#7728) 2025-09-21 08:20:48 +08:00
thop [None][ci] set TORCHINDUCTOR_COMPILE_THREADS correctly (#7800) 2025-09-19 07:19:50 +08:00
helpers.py [https://nvbugs/5517404][fix] Use the correct cuda graph for dynamic spec dec (#7728) 2025-09-21 08:20:48 +08:00
pattern_watcher.py [TRTLLM-3105][feat] Add Piecewise CUDA Graph Support (#3804) 2025-05-09 11:04:01 +08:00
test_connector.py [None][feat] KV Cache Connector API (#7228) 2025-08-28 23:09:27 -04:00