TensorRT-LLMs/tensorrt_llm/inputs/__init__.py
Chang Liu 23500b55c3
[TRTLLM-7398][feat] Support KV cache salting for secure KV cache reuse (#7106)
Signed-off-by: Chang Liu (Enterprise Products) <9713593+chang-l@users.noreply.github.com>
Signed-off-by: Chang Liu <9713593+chang-l@users.noreply.github.com>
2025-09-06 17:58:32 -04:00

50 lines
1.9 KiB
Python

from .data import PromptInputs, TextPrompt, TokensPrompt, prompt_inputs
from .multimodal import MultimodalInput
from .registry import (BaseMultimodalInputProcessor, ExtraProcessedInputs,
InputProcessor, MultimodalPlaceholderMetadata,
MultimodalPlaceholderPlacement, create_input_processor,
create_input_processor_with_hash,
register_input_processor)
from .utils import (ALL_SUPPORTED_AUDIO_MODELS, ALL_SUPPORTED_IMAGE_MODELS,
ALL_SUPPORTED_MULTIMODAL_MODELS, ALL_SUPPORTED_VIDEO_MODELS,
ConversationMessage, MultimodalData, MultimodalDataTracker,
add_multimodal_placeholders, apply_chat_template,
async_load_audio, async_load_image, async_load_video,
convert_image_mode, default_multimodal_input_loader,
encode_base64_content_from_url, get_cache_salt_id,
load_image, load_video)
__all__ = [
"ALL_SUPPORTED_MULTIMODAL_MODELS",
"ALL_SUPPORTED_IMAGE_MODELS",
"ALL_SUPPORTED_VIDEO_MODELS",
"ALL_SUPPORTED_AUDIO_MODELS",
"PromptInputs",
"prompt_inputs",
"TextPrompt",
"TokensPrompt",
"InputProcessor",
"create_input_processor",
"create_input_processor_with_hash",
"register_input_processor",
"ExtraProcessedInputs",
"BaseMultimodalInputProcessor",
"MultimodalPlaceholderMetadata",
"MultimodalPlaceholderPlacement",
"ConversationMessage",
"MultimodalDataTracker",
"MultimodalData",
"MultimodalInput",
"async_load_audio",
"async_load_image",
"async_load_video",
"add_multimodal_placeholders",
"apply_chat_template",
"convert_image_mode",
"default_multimodal_input_loader",
"encode_base64_content_from_url",
"load_image",
"load_video",
"get_cache_salt_id",
]