TensorRT-LLMs/tests/unittest/_torch/attention
Chang Liu e47c787dd7
[TRTLLM-8535][feat] Support DeepSeek V3.2 with FP8 + BF16 KV cache/NVFP4 + BF16 KV cache (#8405)
Signed-off-by: Fanrong Li <23290157+lfr-0531@users.noreply.github.com>
Signed-off-by: Chang Liu <9713593+chang-l@users.noreply.github.com>
Signed-off-by: Tracin <10434017+Tracin@users.noreply.github.com>
2025-10-24 13:40:41 -04:00
..
sparse [TRTLLM-8535][feat] Support DeepSeek V3.2 with FP8 + BF16 KV cache/NVFP4 + BF16 KV cache (#8405) 2025-10-24 13:40:41 -04:00
test_attention_mla.py [https://nvbugs/5453806][unwaive] Unwaive fp8 kvcache attention test (#7243) 2025-09-05 12:13:57 -04:00
test_attention_no_cache.py [None][ci] move unittests to sub-directories (#6635) 2025-08-20 05:42:22 -04:00
test_attention.py [None][ci] move unittests to sub-directories (#6635) 2025-08-20 05:42:22 -04:00
test_flashinfer_attention.py [None][ci] move unittests to sub-directories (#6635) 2025-08-20 05:42:22 -04:00
test_flashinfer_star_attn.py [None][ci] move unittests to sub-directories (#6635) 2025-08-20 05:42:22 -04:00
test_vanilla_attention.py [None][ci] move unittests to sub-directories (#6635) 2025-08-20 05:42:22 -04:00