TensorRT LLM
|
ba1cb6831d
|
[None][infra] Check in most recent lock file from nightly pipeline
Signed-off-by: TensorRT LLM <90828364+tensorrt-cicd@users.noreply.github.com>
|
2026-01-13 03:08:08 +00:00 |
|
fredricz-20070104
|
bbe535fddf
|
[None][chore] Fix disagg assert (#10596)
Signed-off-by: FredricZ-2007 <226039983+fredricz-20070104@users.noreply.github.com>
|
2026-01-12 21:39:57 -05:00 |
|
xxi
|
ba1037ca4a
|
[https://nvbugs/5762336][fix] support to parse the keyword modules_to_not_convert of the HF model config" (#10527)
Signed-off-by: xxi <xxi@nvidia.com>
|
2026-01-12 20:21:01 -05:00 |
|
Iman Tabrizian
|
48b09e5a25
|
[https://nvbugs/5689235][fix] Fix cancellation+chunked prefill+disagg (#10111)
Signed-off-by: Iman Tabrizian <10105175+tabrizian@users.noreply.github.com>
|
2026-01-12 18:23:26 -05:00 |
|
Gal Hubara-Agam
|
18a33764b5
|
[None][chore] Print correct backend name in benchmark report (#10597)
Signed-off-by: Gal Hubara Agam <96368689+galagam@users.noreply.github.com>
|
2026-01-12 14:46:00 -05:00 |
|
Anish Shanbhag
|
dacc881993
|
[https://nvbugs/5761391][fix] Use correct model names for config database regression tests (#10192)
Signed-off-by: Anish Shanbhag <ashanbhag@nvidia.com>
|
2026-01-12 10:55:07 -08:00 |
|
Suyog Gupta
|
a1385243e1
|
[#10580][fix] re-enable NemotronH MOE MMLU test (#10594)
Signed-off-by: Suyog Gupta <41447211+suyoggupta@users.noreply.github.com>
|
2026-01-12 09:26:07 -08:00 |
|
Emma Qiao
|
9f044b9dd9
|
[None][infra] Waive failed tests for main 01/12 (#10604)
Signed-off-by: qqiao <qqiao@nvidia.com>
|
2026-01-12 10:24:54 -05:00 |
|
mpikulski
|
bf7998f1b8
|
[TRTLLM-9522][test] cover LLM API multi_modal_embeddings (#9963)
Signed-off-by: ixlmar <206748156+ixlmar@users.noreply.github.com>
|
2026-01-12 11:38:22 +01:00 |
|
Wanli Jiang
|
11da7e3605
|
[None][fix] Solve pillow version conflict (#10537)
Signed-off-by: Wanli Jiang <35160485+Wanli-Jiang@users.noreply.github.com>
|
2026-01-12 04:05:54 -05:00 |
|
Zhenhuan Chen
|
3bd319dc8e
|
[https://nvbugs/5794796][chore] waive test blocking premerge (#10593)
Signed-off-by: Zhenhuan Chen <zhenhuanc@nvidia.com>
|
2026-01-12 15:39:07 +08:00 |
|
yufeiwu-nv
|
8e806abac3
|
[None][test] Remove most TRT-backend test cases in llm_perf_nim.yml (#10572)
Signed-off-by: Fanrong Li <23290157+lfr-0531@users.noreply.github.com>
Signed-off-by: yufeiwu-nv <230315618+yufeiwu-nv@users.noreply.github.com>
Co-authored-by: Fanrong Li <23290157+lfr-0531@users.noreply.github.com>
|
2026-01-12 15:34:55 +08:00 |
|
yingguo-trt
|
c5914f9085
|
[None][chore] update deepseekv3.2 test parameter (#10595)
Signed-off-by: yingguo-trt <244492186+yingguo-trt@users.noreply.github.com>
|
2026-01-12 01:43:22 -05:00 |
|
chenfeiz0326
|
54459377d2
|
[TRTLLM-10248][feat] Support Bot to Send Perf Regression Msg to Slack Channel (#10489)
Signed-off-by: Chenfei Zhang <chenfeiz@nvidia.com>
|
2026-01-12 14:23:23 +08:00 |
|
Xianjie Qiao
|
3a9a00b544
|
[None][feat] Add ExpertStatistic and DUMMY_ALLREDUCE for configurable_moe (#10401)
Signed-off-by: Xianjie <5410381+qiaoxj07@users.noreply.github.com>
|
2026-01-12 14:10:31 +08:00 |
|
Jie Li
|
5e0dbba0c9
|
[None][chore]: update waive list (#10577)
Signed-off-by: Jie Li <lijie@nvidia.com>
|
2026-01-11 22:18:04 -05:00 |
|
TensorRT LLM
|
2de22f1a70
|
[None][infra] Check in most recent lock file from nightly pipeline
Signed-off-by: TensorRT LLM <90828364+tensorrt-cicd@users.noreply.github.com>
|
2026-01-12 03:09:53 +00:00 |
|
Pengbo Wang
|
c0e25e5418
|
[TRTLLM-10022][feat] Add hopper xqa decode support for skip softmax attention (#10264)
Signed-off-by: Pengbo Wang <221450789+pengbowang-nv@users.noreply.github.com>
|
2026-01-11 19:26:10 -05:00 |
|
Eran Geva
|
c5d5af9e7f
|
[#8391][chore] removed llama and added deepseek to AutoDeploy's L0 perf test (#10585)
Signed-off-by: Eran Geva <19514940+MrGeva@users.noreply.github.com>
|
2026-01-11 16:31:24 -05:00 |
|
Ivy Zhang
|
7f018c89e9
|
[None][test] update core test list (#10538)
Signed-off-by: Ivy Zhang <25222398+crazydemo@users.noreply.github.com>
|
2026-01-11 14:08:20 -05:00 |
|
Yechan Kim
|
8e0d20d901
|
[TRTLLM-10195][feat] K-EXAONE support (#10355)
Signed-off-by: Jaedeok Kim <jaedeokk@nvidia.com>
Signed-off-by: yechank <161688079+yechank-nvidia@users.noreply.github.com>
Co-authored-by: Jaedeok Kim <jaedeokk@nvidia.com>
|
2026-01-12 00:29:51 +09:00 |
|
Yanchao Lu
|
80649a8b78
|
[None][ci] Workaround OCI-NRT slowdown issue (#10587)
Signed-off-by: Yanchao Lu <yanchaol@nvidia.com>
|
2026-01-11 22:08:19 +08:00 |
|
Guoming Zhang
|
0371cbfd88
|
[None][doc] Update Qwen3-Next doc by adding known issues section (#10582)
Signed-off-by: nv-guomingz <137257613+nv-guomingz@users.noreply.github.com>
|
2026-01-11 14:47:47 +08:00 |
|
TensorRT LLM
|
b2e2538fcd
|
[None][infra] Check in most recent lock file from nightly pipeline
Signed-off-by: TensorRT LLM <90828364+tensorrt-cicd@users.noreply.github.com>
|
2026-01-11 03:07:48 +00:00 |
|
HuiGao-NV
|
3c65ec3c55
|
[None][chore] waive test case (#10581)
Signed-off-by: Hui Gao <huig@nvidia.com>
|
2026-01-10 18:53:36 -05:00 |
|
fredricz-20070104
|
f6045fac09
|
[None][chore] Fix Gitlab CI termination issues (#10576)
Signed-off-by: FredricZ-2007 <226039983+fredricz-20070104@users.noreply.github.com>
Signed-off-by: yufeiwu-nv <230315618+yufeiwu-nv@users.noreply.github.com>
Co-authored-by: yufeiwu-nv <230315618+yufeiwu-nv@users.noreply.github.com>
|
2026-01-10 07:51:18 -05:00 |
|
tcherckez-nvidia
|
f6c4dd885f
|
[None][chore] Update AutoDeploy model list (#10505)
Signed-off-by: Tal Cherckez <127761168+tcherckez-nvidia@users.noreply.github.com>
|
2026-01-10 08:47:37 +02:00 |
|
TensorRT LLM
|
6ab996d635
|
[None][infra] Check in most recent lock file from nightly pipeline
Signed-off-by: TensorRT LLM <90828364+tensorrt-cicd@users.noreply.github.com>
|
2026-01-10 03:09:09 +00:00 |
|
William Zhang
|
ff7eb93f31
|
[https://nvbugs/5669097][tests] Add MMMU test for mistral small (#10530)
Signed-off-by: William Zhang <133824995+2ez4bz@users.noreply.github.com>
|
2026-01-09 16:09:28 -08:00 |
|
Chenghao Zhang
|
38f249b479
|
[https://nvbugs/5548861][fix] AutoDeploy: Fix the test (#10521)
Signed-off-by: Chenghao Zhang <211069071+nvchenghaoz@users.noreply.github.com>
|
2026-01-09 13:30:24 -08:00 |
|
Linda
|
82dfef2e56
|
[https://nvbugs/5628848][fix] Fix nanobind stub generation (#10516)
Signed-off-by: Linda-Stadter <57756729+Linda-Stadter@users.noreply.github.com>
|
2026-01-09 11:32:21 -08:00 |
|
Faraz
|
fdbdbba540
|
[https://nvbugs/5752687][fix] Choose register model config over root config for VLM (#10553)
Signed-off-by: Faraz Khoubsirat <58580514+farazkh80@users.noreply.github.com>
|
2026-01-09 12:10:52 -05:00 |
|
yingguo-trt
|
d80f01d205
|
[None][feat] Add support for DeepSeek v3.2 tests (#10561)
Signed-off-by: yingguo-trt <244492186+yingguo-trt@users.noreply.github.com>
|
2026-01-09 10:20:29 -05:00 |
|
Yechan Kim
|
7295af68ba
|
[None][fix] Enable AttentionDP on Qwen3-VL and fix test (#10435)
Signed-off-by: yechank <161688079+yechank-nvidia@users.noreply.github.com>
|
2026-01-10 00:13:26 +09:00 |
|
Kaiyu Xie
|
1c69aad850
|
[TRTLLM-10309] [feat] Optimize qk rope/nope concat for DSA (#10571)
Signed-off-by: Kaiyu Xie <26294424+kaiyux@users.noreply.github.com>
|
2026-01-09 09:50:57 -05:00 |
|
Iman Tabrizian
|
ced88424ef
|
[https://nvbugs/5756008][fix] unwaive test (#10523)
Signed-off-by: Iman Tabrizian <10105175+tabrizian@users.noreply.github.com>
|
2026-01-09 09:40:07 -05:00 |
|
Jie Li
|
627d306df9
|
[None][chore] remove some model support; add device constraint (#10563)
Signed-off-by: Jie Li <lijie@nvidia.com>
|
2026-01-09 09:36:23 -05:00 |
|
ruodil
|
2b72d33fdc
|
[TRTLLM-9932][test] add kimi_k2 single node perf test (#10436)
Signed-off-by: Ruodi Lu <ruodil@users.noreply.github.com>
Co-authored-by: Ruodi Lu <ruodil@users.noreply.github.com>
|
2026-01-09 05:36:50 -05:00 |
|
Fanrong Li
|
4632a8642d
|
[None][doc] blog: Optimizing DeepSeek-V3.2 on NVIDIA Blackwell GPUs (#10565)
Signed-off-by: Fanrong Li <23290157+lfr-0531@users.noreply.github.com>
|
2026-01-09 05:16:00 -05:00 |
|
Yuxian Qiu
|
80f261ea36
|
[https://nvbugs/5622938][feat] Run sample_async on extra stream. (#10215)
Signed-off-by: Yuxian Qiu <142763828+yuxianq@users.noreply.github.com>
|
2026-01-09 18:15:18 +08:00 |
|
Chang Liu
|
78bb245554
|
[https://nvbugs/5787453][fix] Better align MLA chunking with indexer chunking when chunked prefill enabled for DSV32 (#10552)
|
2026-01-09 00:49:39 -08:00 |
|
bhsueh_NV
|
4a09acd012
|
[https://nvbugs/5785206][infra] unwaive the accuracy/test_llm_api_pytorch.py::TestQwen3_30B_A3B (#10560)
Signed-off-by: bhsueh <11360707+byshiue@users.noreply.github.com>
|
2026-01-09 03:13:29 -05:00 |
|
JadoTu
|
4c498bfe58
|
[TRTLLM-9676][fix] Fix mamba_cache_manager when enabling cuda_graph_padding and let test cover this case (#9873)
Signed-off-by: JadoTu <107457950+JadoTu@users.noreply.github.com>
|
2026-01-09 14:50:16 +08:00 |
|
Yukun He
|
c5331e6dbb
|
[None][fix] Setup dist for AutoTuner in Layerwise benchmarking. (#10534)
Signed-off-by: Yukun He <23156053+hyukn@users.noreply.github.com>
|
2026-01-09 14:16:39 +08:00 |
|
Jie Li
|
6fcd4e7099
|
[None][chore] Add failed cases into waives.txt (#10541)
Signed-off-by: Jie Li <lijie@nvidia.com>
|
2026-01-09 01:03:47 -05:00 |
|
TensorRT LLM
|
5df03b2ea7
|
[None][infra] Check in most recent lock file from nightly pipeline
Signed-off-by: TensorRT LLM <90828364+tensorrt-cicd@users.noreply.github.com>
|
2026-01-09 03:43:08 +00:00 |
|
ruodil
|
d707286ca8
|
[None][test] restrict max_num_tokens in disagg mtp config (#10442)
Signed-off-by: Ruodi Lu <ruodil@users.noreply.github.com>
Co-authored-by: Ruodi Lu <ruodil@users.noreply.github.com>
|
2026-01-08 21:53:24 -05:00 |
|
Yuxian Qiu
|
afa55c12b6
|
[None][fix] revert https://github.com/NVIDIA/TensorRT-LLM/pull/10445. (#10547)
Signed-off-by: Yuxian Qiu <142763828+yuxianq@users.noreply.github.com>
|
2026-01-08 21:50:04 -05:00 |
|
Balaram Buddharaju
|
56e779d09f
|
[None][chore] Waive tests blocking premerge 01/08 (#10555)
Signed-off-by: Balaram Buddharaju <169953907+brb-nv@users.noreply.github.com>
|
2026-01-08 20:22:28 -05:00 |
|
Mike Iovine
|
4092a87b6f
|
[https://nvbugs/5740075][fix] Fix sm120 speculation (#10049)
Signed-off-by: Mike Iovine <miovine@nvidia.com>
|
2026-01-08 19:55:43 -05:00 |
|