TensorRT-LLMs/cpp
Yukun He ab2f663101
fix: Reduce memory usage in fused moe op associated with AutoTuning and fix moe fallback issue. (#3793)
* Reduce memory usage in fused moe op associated with AutoTuning.
* Replace pre-defined bucket size strategy with a generating function based on the tune_max_num_tokens.
* Add free_memory logic of workspace in min_latency_mode fused moe path.

Signed-off-by: Yukun He <23156053+hyukn@users.noreply.github.com>

* Fix fused_moe fallback issue. (#3652)

min_latency_mode is only set to False during warmup phase. Thus when it becomes true during inference, all tactics fall back to the default one and thus cause perf regression.

Signed-off-by: Yukun He <23156053+hyukn@users.noreply.github.com>

---------

Signed-off-by: Yukun He <23156053+hyukn@users.noreply.github.com>
2025-04-24 10:14:26 +08:00
..
cmake fix #3109: early exit cmake if find_library() does not find any lib (#3113) 2025-03-29 19:59:03 +08:00
include/tensorrt_llm chore: Clean up cpp runtime (#3537) 2025-04-15 16:06:14 +08:00
micro_benchmarks feat: Add FP8 support for SM 120 (#3248) 2025-04-14 16:05:41 -07:00
tensorrt_llm fix: Reduce memory usage in fused moe op associated with AutoTuning and fix moe fallback issue. (#3793) 2025-04-24 10:14:26 +08:00
tests chore: Clean up cpp runtime (#3537) 2025-04-15 16:06:14 +08:00
CMakeLists.txt Revert "infra: move nvrtc_wrapper to conan (#3282)" (#3573) 2025-04-15 22:45:13 +08:00