TensorRT-LLMs/tensorrt_llm/executor
Lucas Liebenwein 74105a45d9
[#6120][feat] AutoDeploy: flexible args for sequence interface + AD multi-modal input processor + llama4 VLM example (#7221)
Signed-off-by: Lucas Liebenwein <11156568+lucaslie@users.noreply.github.com>
2025-09-05 22:10:48 -04:00
..
__init__.py chore: rename ExecutorBindingsWorker/Proxy (#4716) 2025-05-29 10:32:35 +08:00
executor.py [None][chore] Use llm args in create_py_executor (#7239) 2025-09-01 16:27:55 -07:00
ipc.py [fix]: Fall back to HMAC to Avoid IPC Serialization Churn (#5074) 2025-06-13 11:37:50 +08:00
postproc_worker.py [None][feat] Core Metrics Implementation (#5785) 2025-08-09 02:48:53 -04:00
proxy.py [None][feat] KV Cache Connector API (#7228) 2025-08-28 23:09:27 -04:00
request.py [None][feat] Add support of scheduling attention dp request (#6246) 2025-08-01 20:38:01 -04:00
result.py [TRTLLM-7326][feat] Add standalone multimodal encoder (#6743) 2025-08-19 21:42:50 -07:00
utils.py fix[nvbug5298640]: trtllm-llmapi-launch multiple LLM instances (#4727) 2025-06-19 06:13:53 +08:00
worker.py [#6120][feat] AutoDeploy: flexible args for sequence interface + AD multi-modal input processor + llama4 VLM example (#7221) 2025-09-05 22:10:48 -04:00