| .. |
|
__init__.py
|
Update TensorRT-LLM (#2755)
|
2025-02-11 03:01:00 +00:00 |
|
_util.py
|
[TRTLLM-5826][feat] Support pytorch LoRA adapter eviction (#5616)
|
2025-07-20 08:00:14 +03:00 |
|
config_utils.py
|
feat: support kv cache reuse for MLA (#3571)
|
2025-05-15 15:22:21 +08:00 |
|
config.py
|
[TRTLLM-4279] feat: Multistream initial support for torch compile flow (#5847)
|
2025-07-21 19:10:22 +08:00 |
|
cuda_graph_runner.py
|
[refactor] Simplification of Speculative decoding configs (#5639)
|
2025-07-10 11:37:30 -04:00 |
|
executor_request_queue.py
|
feat: Refactor the fetching request logic (#5786)
|
2025-07-22 09:16:28 +08:00 |
|
finish_reason.py
|
[TRTLLM-5974][feat] Support disaggregated serving in TRTLLM Sampler (#5328)
|
2025-06-25 17:41:36 +02:00 |
|
grammar_matcher.py
|
feat: TRTLLM-5941 Upgrade xgrammar to 0.1.18 (#5364)
|
2025-07-01 20:12:55 +08:00 |
|
guided_decoder.py
|
[TRTLLM-6406] feat: Enable guided decoding with overlap scheduler (#6000)
|
2025-07-17 17:46:10 +08:00 |
|
handle_logits.py
|
feat: Optimize TRTLLM Sampler perf single beam single step (#5550)
|
2025-07-07 15:44:47 +02:00 |
|
kv_cache_transceiver.py
|
chore:[BREAKING CHANGE] use cacheTransceiverConfig as knobs for disagg service (#5234)
|
2025-07-17 17:42:07 +08:00 |
|
layerwise_nvtx_marker.py
|
Update TensorRT-LLM (#2849)
|
2025-03-04 18:44:00 +08:00 |
|
llm_request.py
|
feat: add MultimodalParams & putting all multimodal params into it and refactor HyperCLOVAX & Qwen2/2.5-VL (#5522)
|
2025-07-07 18:03:12 -07:00 |
|
make_decoding_batch_input_output.py
|
feat: Optimize TRTLLM Sampler perf single beam single step (#5550)
|
2025-07-07 15:44:47 +02:00 |
|
model_engine.py
|
fix: fix index out of bounds error in spec decoding (#5954)
|
2025-07-22 12:48:00 +08:00 |
|
py_executor_creator.py
|
[TRTLLM-6452][feat]: Two-model engine KV cache reuse support (#6133)
|
2025-07-19 13:17:15 +08:00 |
|
py_executor.py
|
feat: Refactor the fetching request logic (#5786)
|
2025-07-22 09:16:28 +08:00 |
|
resource_manager.py
|
[TRTLLM-5826][feat] Support pytorch LoRA adapter eviction (#5616)
|
2025-07-20 08:00:14 +03:00 |
|
sampler.py
|
[nvbug/5393888][nvbug/5393042] Always use py_seq_slot (#6147)
|
2025-07-18 22:45:16 +03:00 |
|
scheduler.py
|
[TRTLLM-5826][feat] Support pytorch LoRA adapter eviction (#5616)
|
2025-07-20 08:00:14 +03:00 |
|
seq_slot_manager.py
|
feat: Optimize TRTLLM Sampler perf single beam single step (#5550)
|
2025-07-07 15:44:47 +02:00 |