TensorRT-LLMs/cpp/tensorrt_llm
Erin 8fe7bdeacf
feat: LogitsProcessor in PyTorch backend (#3145)
* support lp in pytorch backend

Signed-off-by: Erin Ho <14718778+hchings@users.noreply.github.com>

* fix tp

Signed-off-by: Erin Ho <14718778+hchings@users.noreply.github.com>

---------

Signed-off-by: Erin Ho <14718778+hchings@users.noreply.github.com>
2025-05-01 14:15:30 -07:00
..
batch_manager cacheTransceiver buffer manager (#3798) 2025-04-27 11:48:15 +08:00
common fix: [https://nvbugspro.nvidia.com/bug/5243482] If FlashMLA is used, the existence of FMHA based MLA kernels should not be checked. (#3862) 2025-04-30 14:27:38 +08:00
cutlass_extensions/include/cutlass_extensions TRTLLM-4624 feat: Add nvfp4 gemm and moe support for SM120 (#3770) 2025-04-29 11:19:11 -04:00
executor feat: Add multimodal embedding field in LlmRequest (#3855) 2025-05-01 12:23:30 +08:00
executor_worker Update TensorRT-LLM (#2792) 2025-02-18 21:27:39 +08:00
kernels feat: add relaxed acceptance for DS (#3865) 2025-05-01 21:50:36 +08:00
layers fix: Eagle decoding (#3456) 2025-04-11 22:06:38 +08:00
plugins TRTLLM-4624 feat: Add nvfp4 gemm and moe support for SM120 (#3770) 2025-04-29 11:19:11 -04:00
pybind feat: LogitsProcessor in PyTorch backend (#3145) 2025-05-01 14:15:30 -07:00
runtime fix: 5197419 and removed unused runtime kernels (#3631) 2025-04-23 18:04:50 +02:00
thop feat: add relaxed acceptance for DS (#3865) 2025-05-01 21:50:36 +08:00
CMakeLists.txt infra: open source XQA kernels (#3762) 2025-04-30 18:05:15 +08:00