TensorRT-LLMs/cpp
Nikita Korobov fe569f0594
[None][feat] bias for FP4 TRT-LLM Gen MoE (#9220)
Signed-off-by: Nikita Korobov <14355239+nekorobov@users.noreply.github.com>
2025-11-18 09:59:47 -08:00
..
cmake [None][doc] Rename TensorRT-LLM to TensorRT LLM for homepage and the … (#7850) 2025-09-25 21:02:35 +08:00
include/tensorrt_llm [None][refactor] decoding inputs, part 2 (#5799) 2025-11-18 14:38:51 +01:00
kernels [https://nvbugs/5498478][fix] Fix eagle3 fp8 kv target model + bf16 draft model + chunked prefill (#8910) 2025-11-06 07:41:21 -08:00
micro_benchmarks [None][fix] Fix the performance issue of FP8 blockwise grouped GEMM when using attention DP (#8501) 2025-10-27 10:18:19 +08:00
tensorrt_llm [None][feat] bias for FP4 TRT-LLM Gen MoE (#9220) 2025-11-18 09:59:47 -08:00
tests [None][refactor] decoding inputs, part 2 (#5799) 2025-11-18 14:38:51 +01:00
CMakeLists.txt [TRTLLM-8535][feat] Support DeepSeek V3.2 with FP8 + BF16 KV cache/NVFP4 + BF16 KV cache (#8405) 2025-10-24 13:40:41 -04:00
conandata.yml
conanfile.py feat: large-scale EP(part 6: Online EP load balancer integration for GB200 nvfp4) (#4818) 2025-06-08 10:25:18 +08:00
libnuma_conan.py fix cuda driver link issue with driver version less than 12.3 (#5025) 2025-06-10 15:27:39 +08:00