mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-25 21:22:57 +08:00
This merge request attempts to support more SWA KV cache functionality inside the KV cache manager. Before this merge request, the KV cache for sliding window attention (SWA) only holds "window size" number of blocks and reuse them in a cyclic manner. We will not be able to utilize more GPU memory with this design, leading to a limited max batch size throughput. Additionally, we will not be able to support KV cache reuse with this design. In this MR, we change such behavior to let the manager write blocks in a linear manner. With a linear block writing behavior, as the attention window moves on, the out-of-window (OOW) blocks will be detached. Right now for the sake of a correct feature first, we directly offload the OOW block from the primary block pool (GPU memory) to the secondary block pool (host memory). We will improve this in the future by delegating the block movement to the eviction policy. KV cache reuse for SWA is not developed in this merge request and will be amended in a follow-up merge request. Writing the blocks linearly, the maximum number of blocks allocated for a sequence(`GenerationRequest`) is the "max sequence length" specified. The `GenerationRequest` that stores the cache block bookkeeping structure will now keep "max sequence length" tokens of blocks. Given the above, main changes are (more context in the MR): - Remove "cyclic" concept under the kv cache manager, such concept originally guards the block reuse under kv cache manager. - Add detach mechanism and have it under `KVCacheManager::addToken`. Please note that detach is still guarded off for SWA when reuse is enabled. A follow-up merge request will proceed to improve this. - Enforce "max sequence length" to be a non-optional parameter to the `KVCacheManager`/`BlockManager` - Let all window size resource pool get identical proportion of memory - Fix free memory calculation under `resource_manager.py` Signed-off-by: eopXD <yuehtingc@nvidia.com> Co-authored-by: Tomer Asida <tasida@nvidia.com> |
||
|---|---|---|
| .. | ||
| l0_a10.yml | ||
| l0_a30.yml | ||
| l0_a100.yml | ||
| l0_b200.yml | ||
| l0_dgx_b200.yml | ||
| l0_dgx_b300.yml | ||
| l0_dgx_h100.yml | ||
| l0_dgx_h200.yml | ||
| l0_gb200_multi_gpus.yml | ||
| l0_gb200_multi_nodes.yml | ||
| l0_gb202.yml | ||
| l0_gb203.yml | ||
| l0_gb300_multi_gpus.yml | ||
| l0_gh200.yml | ||
| l0_h100.yml | ||
| l0_l40s.yml | ||
| l0_perf.yml | ||
| l0_rtx_pro_6000.yml | ||
| l0_sanity_check.yml | ||
| README.md | ||
Description
This folder contains test definition which is consumed by trt-test-db tool based on system specifications.
Installation
Install trt-test-db using the following command:
pip3 install --extra-index-url https://urm.nvidia.com/artifactory/api/pypi/sw-tensorrt-pypi/simple --ignore-installed trt-test-db==1.8.5+bc6df7
Test Definition
Test definitions are stored in YAML files located in ${TRT_LLM_ROOT}/tests/integration/test_lists/test-db/. These files define test conditions and the tests to be executed.
Example YAML Structure
version: 0.0.1
l0_e2e:
- condition:
terms:
supports_fp8: true
ranges:
system_gpu_count:
gte: 4
lte: 4
wildcards:
gpu:
- '*h100*'
linux_distribution_name: ubuntu*
tests:
- examples/test_llama.py::test_llm_llama_v3_1_1node_multi_gpus[llama-3.1-8b-enable_fp8]
- examples/test_llama.py::test_llm_llama_v3_1_1node_multi_gpus[llama-3.1-70b-enable_fp8]
Generating Test Lists
Use trt-test-db to generate a test list based on the system configuration:
trt-test-db -d /TensorRT-LLM/src/tests/integration/test_lists/test-db \
--context l0_e2e \
--test-names \
--output /TensorRT-LLM/src/l0_e2e.txt \
--match-exact '{"chip":"ga102gl-a","compute_capability":"8.6","cpu":"x86_64","gpu":"A10","gpu_memory":"23028.0","host_mem_available_mib":"989937","host_mem_total_mib":"1031949","is_aarch64":false,"is_linux":true,"linux_distribution_name":"ubuntu","linux_version":"22.04","supports_fp8":false,"supports_int8":true,"supports_tf32":true,"sysname":"Linux","system_gpu_count":"1",...}'
This command generates a test list file (l0_e2e.txt) based on the specified context and system configuration.
Running Tests
Execute the tests using pytest with the generated test list:
pytest -v --test-list=/TensorRT-LLM/src/l0_e2e.txt --output-dir=/tmp/logs
This command runs the tests specified in the test list and outputs the results to the specified directory.
Additional Information
- The
--contextparameter in thetrt-test-dbcommand specifies which context to search in the YAML files. - The
--match-exactparameter provides system information used to filter tests based on the conditions defined in the YAML files. - Modify the YAML files to add or update test conditions and test cases as needed.
For more detailed information on
trt-test-dbandpytestusage, refer to their respective documentation.