TensorRT-LLMs/tests/unittest
Fridah-nv ec723fa993
feat:[AutoDeploy] Enhance RoPE support (#3115)
* add test to map flashinfer rope op with triton custom rope ops and pytorch rope in fused_mha

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* add rope matcher and unit tests

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* capture cos and sin from graph

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* revert fuse_mha op change

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* minor update to address comment and remove redundant unit test

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* move view and transpose into graph nodes and update unit test to test custom op directly

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* move view into custom op, update bfs with bound, update custom op return type to be half precision

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* custom op update to support 3D input

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* handle bnsd and bsnd format, update tests, handle 3D cos/sin input to the custom op

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* add llama4 rope test, update custom op with is_neox flag

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* add llama4 style rope to matcher and update unit test

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* separate into two transformations

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* fix when num_head != num_kv_head; add support for cached position_ids and cos_sin_cache in graph; update unit tests

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* minor update, cache locally and propagate meta info of qk nodes

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* minor: fix cos_sin_cache not float

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

* minor: move cache into matcher

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>

---------

Signed-off-by: Frida Hou <201670829+Fridah-nv@users.noreply.github.com>
2025-04-11 23:51:24 +08:00
..
_torch feat:[AutoDeploy] Enhance RoPE support (#3115) 2025-04-11 23:51:24 +08:00
api_stability chore: refactor the LlmArgs with Pydantic and migrate remaining pybinding configs to python (#3025) 2025-04-05 13:31:48 +08:00
bindings feat: support abort disconnected requests (#3214) 2025-04-07 16:14:58 +08:00
llmapi test: fix memory leak of tests (#3392) 2025-04-10 14:31:40 +08:00
others test: reorganize tests folder hierarchy (#2996) 2025-03-27 12:07:53 +08:00
scaffolding feat: Enhance the integrated robustness of scaffolding with __init__.py #3305 (#3312) 2025-04-09 21:13:47 +08:00
tools Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
trt Waive unittest/trt/model/test_mamba.py::TestMamba::test_loaders_mamba_130m_hf_from_checkpoint. Will fix it later. (#3356) 2025-04-07 22:36:35 -07:00
utils test: reorganize tests folder hierarchy (#2996) 2025-03-27 12:07:53 +08:00
conftest.py Add thread leak check and fix thread/memory leak issues. (#3270) 2025-04-08 19:03:18 +08:00
dump_checkpoint_stats.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
profile_utils.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
pytest.ini chore: Refine attention backend interface. (#3271) 2025-04-09 02:34:53 +08:00
test_model_runner_cpp.py Update TensorRT-LLM (#2936) 2025-03-18 21:25:19 +08:00
test_pip_install.py relax the limitation of setuptools (#2992) 2025-03-24 13:36:10 +08:00