| .. |
|
test_modeling_bert.py
|
feat: no-cache attention in PyTorch workflow (#3085)
|
2025-04-05 01:54:32 +08:00 |
|
test_modeling_clip.py
|
feat: add Pytorch support of Vision Encoder for multimodal models (#3791)
|
2025-05-03 05:13:47 +08:00 |
|
test_modeling_deepseek.py
|
[TRTLLM-5530][BREAKING CHANGE]: enhance the llm args pytorch config part 1(cuda_graph_config) (#5014)
|
2025-06-30 11:05:40 +08:00 |
|
test_modeling_gemma3.py
|
fix: Skip rope scaling for local layers in Gemma3 VLM (#5857)
|
2025-07-09 10:10:33 +08:00 |
|
test_modeling_llama_min_latency.py
|
[fix] speedup modeling unittests (#5579)
|
2025-06-30 06:30:45 +03:00 |
|
test_modeling_llama.py
|
[fix] speedup modeling unittests (#5579)
|
2025-06-30 06:30:45 +03:00 |
|
test_modeling_mixtral.py
|
[fix] speedup modeling unittests (#5579)
|
2025-06-30 06:30:45 +03:00 |
|
test_modeling_mllama.py
|
Update transformers to 4.53.0 (#5747)
|
2025-07-09 09:32:24 -07:00 |
|
test_modeling_nemotron_h.py
|
[TRTLLM-4923][feat] Enable CUDA graphs for Nemotron-H (#5646)
|
2025-07-03 11:07:51 +03:00 |
|
test_modeling_nemotron_nas.py
|
[fix][test] Speedup Nemotron NAS unittests (#5202)
|
2025-06-15 11:26:03 +03:00 |
|
test_modeling_nemotron.py
|
[fix] speedup modeling unittests (#5579)
|
2025-06-30 06:30:45 +03:00 |
|
test_modeling_out_of_tree.py
|
chores: merge examples for v1.0 doc (#5736)
|
2025-07-08 21:00:42 -07:00 |
|
test_modeling_qwen_moe.py
|
feat: add qwen2 moe to torch flow; fix wrong imported KvCacheConfig in gpqa… (#3369)
|
2025-04-10 22:45:57 +08:00 |
|
test_modeling_qwen.py
|
feat: no-cache attention in PyTorch workflow (#3085)
|
2025-04-05 01:54:32 +08:00 |
|
test_modeling_siglip.py
|
feat: add Pytorch support of Vision Encoder for multimodal models (#3791)
|
2025-05-03 05:13:47 +08:00 |
|
test_modeling_vila.py
|
feat: llama4 input processor (#3383)
|
2025-04-25 16:47:14 -07:00 |