TensorRT-LLMs/cpp/include/tensorrt_llm/runtime
Daniel Cámpora 1299f27c74
fix: Fix C++ decoder synchronization in PyTorch (#3106)
* Use updateDecoderBuffers in python decoder.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Fix synchronize in trtllm decoder.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Enable by default.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Use guided_decoder to setup seqslots and free them.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Use always decode_async and update_requests.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Update decoder buffers.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Fix speculative decoding tests.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Send new_tensors_host instead of assuming dict.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Make default False in enable_trtllm_decoder.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Partially fix mtp, partially fix py_executor.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Update request states before sending disagg ctx cache.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Fix disagg test for torch decoder.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Make isend_tensor_list and recv_tensor_list for sending the tensors_host.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Formatting.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Fix rebase.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Add disagg serving case to guided decoder.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Get overlap scheduling to work.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Update cutlass to main.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Update after rebasing.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Formatting.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Update to use decode async and update requests.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Properly pass information to update_requests

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Formatting.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Make disaggregated serving a step closer to working.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Fix rebase.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Fix rebase and format.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Copy new device tokens more pythonic.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Restore MTP add dummy reqs.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Add ordereddict import to py_executor.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Formatting.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Added seq slot manager. Add test.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Use transmission for single tensor except when list of tensors is received.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Add TRTLLMDecoder allocation to estimate max kv cache tokens.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Add stream synchronization

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Formatting.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Make memory calculation of decoder adapt to the chosen decoder. Recognize decoder option passed in executorconfig. Make overlap scheduler test run on TinyLlama.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Format

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Add decoder creation to estimate max kv.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Formatting.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

* Update submodule UCXX inline with main.

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>

---------

Signed-off-by: Daniel Campora <961215+dcampora@users.noreply.github.com>
2025-04-23 23:55:27 +08:00
..
utils feat: Add numNodes to ParallelConfig (#3346) 2025-04-13 13:55:04 +02:00
bufferManager.h Update TensorRT-LLM (#2755) 2025-02-11 03:01:00 +00:00
common.h Update TensorRT-LLM (#2502) 2024-11-26 16:51:34 +08:00
cudaEvent.h Update TensorRT-LLM (#787) 2024-01-02 17:54:32 +08:00
cudaStream.h fix: Fix C++ decoder synchronization in PyTorch (#3106) 2025-04-23 23:55:27 +08:00
decoderState.h refactor: decoder buffers (#3307) 2025-04-12 11:41:24 +02:00
decodingInput.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
decodingOutput.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
eagleBuffers.h Update TensorRT-LLM (#2873) 2025-03-11 21:13:42 +08:00
eagleModule.h Update TensorRT-LLM (#2792) 2025-02-18 21:27:39 +08:00
explicitDraftTokensBuffers.h Update TensorRT-LLM (#2873) 2025-03-11 21:13:42 +08:00
generationInput.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
generationOutput.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
gptDecoder.h Update TensorRT-LLM (#2582) 2024-12-16 21:50:47 -08:00
gptDecoderBatched.h Reapply "refactor: Replace DecoderFinishedEvent with CudaEvent in decoder clas…" (#3183) (#3195) 2025-04-04 15:56:28 +02:00
gptJsonConfig.h Update TensorRT-LLM (#2532) 2024-12-04 21:16:56 +08:00
gptSession.h feat: Integrate GPUDirect Storage (GDS) into Executor API (#3582) 2025-04-18 15:59:21 +08:00
iBuffer.h Update TensorRT-LLM (#2755) 2025-02-11 03:01:00 +00:00
iGptDecoderBatched.h chore: Clean up cpp runtime (#3537) 2025-04-15 16:06:14 +08:00
ipcNvlsMemory.h chore: Stabilize ABI boundary for internal kernel library (#3117) 2025-04-11 15:07:50 +08:00
ipcUtils.h Update TensorRT-LLM (#2873) 2025-03-11 21:13:42 +08:00
iStatefulGptDecoder.h refactor: Remove speculative decoding parameters from stateful decoders (#3024) 2025-03-26 20:16:26 +08:00
iTensor.h Update TensorRT-LLM (#2532) 2024-12-04 21:16:56 +08:00
lookaheadBuffers.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
lookaheadModule.h Update TensorRT-LLM (#2582) 2024-12-16 21:50:47 -08:00
loraCache.h Update TensorRT-LLM (#2755) 2025-02-11 03:01:00 +00:00
loraCachePageManagerConfig.h Update TensorRT-LLM (#1598) 2024-05-14 16:43:41 +08:00
loraModule.h Update TensorRT-LLM (#2755) 2025-02-11 03:01:00 +00:00
medusaModule.h Update TensorRT-LLM (#2436) 2024-11-12 15:27:49 +08:00
memoryCounters.h Update TensorRT-LLM (#2110) 2024-08-13 22:34:33 +08:00
modelConfig.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
promptTuningParams.h Update TensorRT-LLM (#1598) 2024-05-14 16:43:41 +08:00
rawEngine.h Update TensorRT-LLM (#2215) 2024-09-10 18:21:22 +08:00
request.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
runtimeDefaults.h Update TensorRT-LLM (#2436) 2024-11-12 15:27:49 +08:00
samplingConfig.h Feat: Variable-Beam-Width-Search (VBWS) part3 (#3338) 2025-04-08 23:51:27 +08:00
speculativeDecodingMode.h Update TensorRT-LLM (#2436) 2024-11-12 15:27:49 +08:00
speculativeDecodingModule.h Update TensorRT-LLM (#1763) 2024-06-11 16:59:02 +08:00
statefulGptDecoderBatched.h chore: Clean up cpp runtime (#3537) 2025-04-15 16:06:14 +08:00
tllmLogger.h Update TensorRT-LLM (#787) 2024-01-02 17:54:32 +08:00
worldConfig.h bug: Fix hang bug when context server doesn't have enough capacity for KV Cache (#3095) 2025-04-21 15:16:55 +08:00