TensorRT-LLMs/triton_backend
Vivian Chen 34212e2e36
[TRTLLM-6104] feat: add request_perf_metrics to triton LLMAPI backend (#5554)
Signed-off-by: Vivian Chen <140748220+xuanzic@users.noreply.github.com>
2025-06-30 21:34:42 -07:00
..
all_models [TRTLLM-6104] feat: add request_perf_metrics to triton LLMAPI backend (#5554) 2025-06-30 21:34:42 -07:00
ci Move Triton backend to TRT-LLM main (#3549) 2025-05-16 07:15:23 +08:00
inflight_batcher_llm [nvbugs/5309940] Add support for input output token counts (#5445) 2025-06-28 04:39:39 +08:00
scripts Add testing for trtllm-llmapi-launch with tritonserver (#5528) 2025-06-27 11:19:52 +08:00
tools [nvbug 5283506] fix: Fix spec decode triton test (#4845) 2025-06-09 08:40:17 -04:00
requirements.txt Move Triton backend to TRT-LLM main (#3549) 2025-05-16 07:15:23 +08:00