TensorRT-LLMs/tensorrt_llm/_torch/auto_deploy
Chenghao Zhang e033929221
[None][feat] AutoDeploy: Flashinfer kernels bringup (#10867)
Signed-off-by: nvchenghaoz <211069071+nvchenghaoz@users.noreply.github.com>
2026-01-29 14:59:29 -08:00
..
compile [None][feat] AutoDeploy: prepare_metadata revisited (#9764) 2025-12-12 20:14:14 +08:00
config [#10013][feat] AutoDeploy: native cache manager integration (#10635) 2026-01-27 11:23:22 -05:00
custom_ops [None][feat] AutoDeploy: Flashinfer kernels bringup (#10867) 2026-01-29 14:59:29 -08:00
distributed [None][refactor] Unify the usage of MPIDist and TorchDist. (#10380) 2026-01-14 14:05:47 +08:00
export [None][fix] fix Qwen2/3 export for AutoDeploy (#11007) 2026-01-28 16:53:21 -08:00
models [#10245][feat] AutoDeploy: Add Minimax M2 support (#10525) 2026-01-28 17:22:32 -05:00
shim [#10013][feat] AutoDeploy: native cache manager integration (#10635) 2026-01-27 11:23:22 -05:00
transform [https://nvbugs/5761391][fix] Include triton-kernels as a packaged dependency (#10471) 2026-01-28 19:56:32 -08:00
utils [#10245][feat] AutoDeploy: Add Minimax M2 support (#10525) 2026-01-28 17:22:32 -05:00
__init__.py [AutoDeploy] merge feat/ad-2025-07-07 (#6196) 2025-07-23 05:11:04 +08:00
llm_args.py [#10013][feat] AutoDeploy: native cache manager integration (#10635) 2026-01-27 11:23:22 -05:00
llm.py [TRTLLM-9065][chore] remove PyTorchConfig completely (#8856) 2025-11-06 22:37:03 -08:00