TensorRT-LLMs/tensorrt_llm/inputs/__init__.py
Yechan Kim 0893afae3d
[TRTLLM-6771][feat] Support MMMU for multimodal models (#6828)
Signed-off-by: yechank <161688079+yechank-nvidia@users.noreply.github.com>
2025-08-21 08:54:12 +08:00

47 lines
1.7 KiB
Python

from .data import PromptInputs, TextPrompt, TokensPrompt, prompt_inputs
from .multimodal import MultimodalInput
from .registry import (ExtraProcessedInputs, InputProcessor,
MultimodalPlaceholderMetadata,
MultimodalPlaceholderPlacement, create_input_processor,
create_input_processor_with_hash,
register_input_processor)
from .utils import (ALL_SUPPORTED_AUDIO_MODELS, ALL_SUPPORTED_IMAGE_MODELS,
ALL_SUPPORTED_MULTIMODAL_MODELS, ALL_SUPPORTED_VIDEO_MODELS,
ConversationMessage, MultimodalData, MultimodalDataTracker,
add_multimodal_placeholders, apply_chat_template,
async_load_audio, async_load_image, async_load_video,
convert_image_mode, default_multimodal_input_loader,
encode_base64_content_from_url, load_image, load_video)
__all__ = [
"ALL_SUPPORTED_MULTIMODAL_MODELS",
"ALL_SUPPORTED_IMAGE_MODELS",
"ALL_SUPPORTED_VIDEO_MODELS",
"ALL_SUPPORTED_AUDIO_MODELS",
"PromptInputs",
"prompt_inputs",
"TextPrompt",
"TokensPrompt",
"InputProcessor",
"create_input_processor",
"create_input_processor_with_hash",
"register_input_processor",
"ExtraProcessedInputs",
"MultimodalPlaceholderMetadata",
"MultimodalPlaceholderPlacement",
"ConversationMessage",
"MultimodalDataTracker",
"MultimodalData",
"MultimodalInput",
"async_load_audio",
"async_load_image",
"async_load_video",
"add_multimodal_placeholders",
"apply_chat_template",
"convert_image_mode",
"default_multimodal_input_loader",
"encode_base64_content_from_url",
"load_image",
"load_video",
]