| .. |
|
__init__.py
|
Update TensorRT-LLM (#2755)
|
2025-02-11 03:01:00 +00:00 |
|
_util.py
|
[TRTLLM-10321][feat] Support different KV cache layout for one-model spec dec (#10502)
|
2026-02-10 05:16:02 +08:00 |
|
config_utils.py
|
[None][feat] Eagle: MLA Based Eagle (#9677)
|
2026-01-02 13:45:07 -05:00 |
|
cuda_graph_runner.py
|
[TRTLLM-10321][feat] Support different KV cache layout for one-model spec dec (#10502)
|
2026-02-10 05:16:02 +08:00 |
|
executor_request_queue.py
|
[TRTLLM-10666][chore] Refactor request fetching logic for better separation of concerns (#10988)
|
2026-02-02 10:36:08 +08:00 |
|
finish_reason.py
|
[TRTLLM-5974][feat] Support disaggregated serving in TRTLLM Sampler (#5328)
|
2025-06-25 17:41:36 +02:00 |
|
grammar_matcher.py
|
[TRTLLM-8763][chore] Deprecate pybind based GuidedDecodingConfig usage in torch backend (#8717)
|
2025-10-29 20:37:14 +08:00 |
|
guided_decoder.py
|
[None][fix] Always reset drafting states for GuidedDecoder (#10899)
|
2026-02-02 16:26:46 +08:00 |
|
handle_additional_outputs.py
|
[TRTLLM-8831][feat] Enable early exit with overlap scheduler (#8587)
|
2025-11-17 18:07:13 +01:00 |
|
handle_logits.py
|
[TRTLLM-8831][feat] Enable early exit with overlap scheduler (#8587)
|
2025-11-17 18:07:13 +01:00 |
|
hang_detector.py
|
[None][feat] Hang detection for executor loop and worker. (#10480)
|
2026-01-13 02:34:32 -05:00 |
|
kv_cache_connector.py
|
[None][feat] Add priority-based KV cache offload filtering support (#10751)
|
2026-02-05 05:22:56 -05:00 |
|
kv_cache_transceiver.py
|
[TRTLLM-8921][feat] implement gen-first disagg_service (#11020)
|
2026-02-03 15:46:11 -05:00 |
|
layerwise_nvtx_marker.py
|
Update TensorRT-LLM (#2849)
|
2025-03-04 18:44:00 +08:00 |
|
llm_request.py
|
[TRTLLM-10030][chore] promote SampleState to TypeVar + typing fixes (#11281)
|
2026-02-05 16:33:22 +01:00 |
|
make_decoding_batch_input_output.py
|
[None][refactor] decoding inputs, part 2 (#5799)
|
2025-11-18 14:38:51 +01:00 |
|
mamba_cache_manager.py
|
[#10013][feat] AutoDeploy: native cache manager integration (#10635)
|
2026-01-27 11:23:22 -05:00 |
|
model_engine.py
|
[TRTLLM-10321][feat] Support different KV cache layout for one-model spec dec (#10502)
|
2026-02-10 05:16:02 +08:00 |
|
model_loader.py
|
[TRTLLM-9457][feat] Add cute dsl fp8 gemm for Blackwell (#10130)
|
2026-02-06 09:49:30 +08:00 |
|
py_executor_creator.py
|
[TRTLLM-10321][feat] Support different KV cache layout for one-model spec dec (#10502)
|
2026-02-10 05:16:02 +08:00 |
|
py_executor.py
|
[https://nvbugs/5829097][fix] Disaggregated serving: Only send finished context requests to the KV cache transceiver (#11354)
|
2026-02-09 17:11:45 +08:00 |
|
request_utils.py
|
[TRTLLM-10666][chore] Refactor request fetching logic for better separation of concerns (#10988)
|
2026-02-02 10:36:08 +08:00 |
|
resource_manager.py
|
[TRTLLM-10321][feat] Support different KV cache layout for one-model spec dec (#10502)
|
2026-02-10 05:16:02 +08:00 |
|
sampler.py
|
[https://nvbugs/5769815][fix] Fix offset calculation in _are_stop_words when using speculative decoding (#10854)
|
2026-02-09 23:53:40 +08:00 |
|
sampling_utils_flashinfer.py
|
[TRTLLM-10030][perf] beam search (remove GPU sync + fix batching + refactor) (#11276)
|
2026-02-05 15:33:51 +01:00 |
|
sampling_utils.py
|
[TRTLLM-10030][perf] beam search (remove GPU sync + fix batching + refactor) (#11276)
|
2026-02-05 15:33:51 +01:00 |
|
scheduler.py
|
[TRTLLM-9766][feat] Integration of the KVCacheManager V2 to TRTLLM Runtime (#10659)
|
2026-02-02 14:29:02 +08:00 |
|
seq_slot_manager.py
|
[https://nvbugs/5394392][fix] Enlarge scheduler capacity under disagg bs == 1 (#6537)
|
2025-08-15 09:52:06 -07:00 |