| .. |
|
checkpoints
|
[None][feat] Support Qwen3 next (#7892)
|
2025-09-29 21:16:07 +08:00 |
|
__init__.py
|
[None][feat] Support Qwen3 next (#7892)
|
2025-09-29 21:16:07 +08:00 |
|
modeling_auto.py
|
[TRTLLM-6746][feat] Enable two-model spec dec for MTP Eagle (#7001)
|
2025-09-18 12:05:36 -04:00 |
|
modeling_bert.py
|
feat: Remove not used padding_idx in models (#5385)
|
2025-06-25 17:19:59 +08:00 |
|
modeling_clip.py
|
[feat] Enable TP and batching for PixtralVisionModel / Mistral3VLM (#6152)
|
2025-07-22 11:06:41 -07:00 |
|
modeling_deepseekv3.py
|
[None][chore] Waive failing MNNVL alltoall multi-gpu test (#8106)
|
2025-09-30 20:05:42 -04:00 |
|
modeling_exaone4.py
|
[#6186][feat] Introduce QKNormRoPEAttention module (#6830)
|
2025-09-05 14:04:41 +02:00 |
|
modeling_gemma3.py
|
[#6186][feat] Introduce QKNormRoPEAttention module (#6830)
|
2025-09-05 14:04:41 +02:00 |
|
modeling_gemma3vl.py
|
[TRTLLM-7440][fix] Split fused_input_embed to separate out host sync (#7280)
|
2025-09-06 23:11:39 -04:00 |
|
modeling_gpt_oss.py
|
[TRTLLM-7775][feat] Integrate tinygemm2 for gpt-oss (#7916)
|
2025-10-02 10:47:04 -07:00 |
|
modeling_hunyuan_dense.py
|
[None][feat] Add Tencent HunYuanDenseV1 model support (#7081)
|
2025-09-23 09:27:29 +08:00 |
|
modeling_hunyuan_moe.py
|
[TRTLLM-7385][feat] Optimize Qwen2/2.5-VL performance (#7250)
|
2025-09-22 03:40:02 -07:00 |
|
modeling_hyperclovax.py
|
[https://nvbugs/5549081][fix] Fix device id assignment for some vision models (#8070)
|
2025-10-01 23:28:05 -04:00 |
|
modeling_llama_min_latency.py
|
[https://nvbugs/5516710][fix] fix Llama 3.3 TP PP case (#7717)
|
2025-09-25 21:02:35 +08:00 |
|
modeling_llama.py
|
[https://nvbugs/5549081][fix] Fix device id assignment for some vision models (#8070)
|
2025-10-01 23:28:05 -04:00 |
|
modeling_llava_next.py
|
[https://nvbugs/5549081][fix] Fix device id assignment for some vision models (#8070)
|
2025-10-01 23:28:05 -04:00 |
|
modeling_mistral.py
|
[TRTLLM-7410][feat] Enable KV cache reuse and chunked prefill for mistral3.1 (#7628)
|
2025-09-17 08:11:16 -07:00 |
|
modeling_mixtral.py
|
[TRTLLM-7408][feat] Wrap MOE with custom op. (#7277)
|
2025-09-09 12:18:56 -04:00 |
|
modeling_mllama.py
|
feat : support duplicate_kv_weight for qwen3 blockwise scale (#5459)
|
2025-06-30 11:49:22 +08:00 |
|
modeling_multimodal_encoder.py
|
Update TensorRT-LLM (#2873)
|
2025-03-11 21:13:42 +08:00 |
|
modeling_multimodal_utils.py
|
[TRTLLM-7328][feat] E-PD Disagg Support via llmapi (3/N) (#7577)
|
2025-09-22 19:07:18 -07:00 |
|
modeling_nanov2vlm.py
|
[TRTLLM-6577][feat] Support nano_v2_vlm in pytorch backend (#7207)
|
2025-09-18 16:26:20 +08:00 |
|
modeling_nemotron_h.py
|
[None][fix] enable NvFP4/FP8 quantization for Nemotron-H architecture (#7589)
|
2025-09-09 11:42:22 +03:00 |
|
modeling_nemotron_nas.py
|
[Perf]: Add residual, norm for nemotron_nas models (#6455)
|
2025-07-30 09:10:38 -07:00 |
|
modeling_nemotron.py
|
feat: Remove not used padding_idx in models (#5385)
|
2025-06-25 17:19:59 +08:00 |
|
modeling_phi3.py
|
[None][fix] Accommodate Phi3/4 to work with ModelOpt's FP8 ckpts in Torch (#6761)
|
2025-08-19 09:22:46 -07:00 |
|
modeling_phi4mm.py
|
[TRTLLM-7758][feat] Phi4-mm image modality inference optimization (#7918)
|
2025-09-25 15:58:29 +08:00 |
|
modeling_pixtral.py
|
[TRTLLM-7442][model] Remove unnecessary D2H copies (#7273)
|
2025-09-03 23:14:20 -04:00 |
|
modeling_qwen2vl.py
|
[TRTLLM-7385][feat] Optimize Qwen2/2.5-VL performance (#7250)
|
2025-09-22 03:40:02 -07:00 |
|
modeling_qwen3_moe.py
|
[None] [feat] Enable run_post_quant_allgather for MoE TRTLLM backend (#6794)
|
2025-09-23 08:24:21 +08:00 |
|
modeling_qwen3_next.py
|
[None][chore] Refine qwen3-next implementation. (#8064)
|
2025-09-30 15:05:13 -04:00 |
|
modeling_qwen3.py
|
[None][feat] Support Qwen3 next (#7892)
|
2025-09-29 21:16:07 +08:00 |
|
modeling_qwen_moe.py
|
[TRTLLM-7408][feat] Wrap MOE with custom op. (#7277)
|
2025-09-09 12:18:56 -04:00 |
|
modeling_qwen.py
|
[TRTLLM-7385][feat] Optimize Qwen2/2.5-VL performance (#7250)
|
2025-09-22 03:40:02 -07:00 |
|
modeling_radio.py
|
[TRTLLM-6577][feat] Support nano_v2_vlm in pytorch backend (#7207)
|
2025-09-18 16:26:20 +08:00 |
|
modeling_seedoss.py
|
[None][feat] Support Seed-OSS model in pytorch backend (#7496)
|
2025-09-24 03:57:12 -07:00 |
|
modeling_siglip.py
|
feat: Update Gemma3 Vision Encoder (#5973)
|
2025-07-14 22:38:10 +08:00 |
|
modeling_speculative.py
|
[TRTLLM-6746][feat] Enable two-model spec dec for MTP Eagle (#7001)
|
2025-09-18 12:05:36 -04:00 |
|
modeling_utils.py
|
[None][feat] Support Qwen3 next (#7892)
|
2025-09-29 21:16:07 +08:00 |
|
modeling_vila.py
|
[TRTLLM-7440][fix] Split fused_input_embed to separate out host sync (#7280)
|
2025-09-06 23:11:39 -04:00 |