diff --git a/latest/.buildinfo b/latest/.buildinfo index 1f24c7367c..e1872b4a6c 100644 --- a/latest/.buildinfo +++ b/latest/.buildinfo @@ -1,4 +1,4 @@ # Sphinx build info version 1 # This file hashes the configuration used when building these files. When it is not found, a full rebuild will be done. -config: a9c5c8c57021602368f541d74d22523d +config: 5b10b2153627779ea5be4dbb07d82396 tags: 645f666f9bcd5a90fca523b33c5a78b7 diff --git a/latest/_cpp_gen/executor.html b/latest/_cpp_gen/executor.html index 182be9c38a..824e5f6543 100644 --- a/latest/_cpp_gen/executor.html +++ b/latest/_cpp_gen/executor.html @@ -32,7 +32,7 @@ - + @@ -44,6 +44,8 @@ + + @@ -59,20 +61,24 @@ + + + - + +
@@ -80,6 +86,8 @@ + + @@ -506,8 +514,8 @@Public Functions
+Constructs a DisaggExecutorOrchestrator object.
+ctxEnginePaths – A vector of file paths to context engine files.
genEnginePaths – A vector of file paths to generation engine files.
ctxExecutorConfigs – A vector of ExecutorConfig for context executors.
genExecutorConfigs – A vector of ExecutorConfig for generation executors.
hasContextAwaitThreads – Whether or not there are threads that receive response for each generation executor.
hasGenAwaitThreads – Whether or not there are threads that receive response for each generation executor.
Enqueue context-only requests to context executors.
+requests – A vector of context-only requests.
selectContextId – The index of the context executor to use. If std::nullopt, the executor that has the smallest number of inflight requests will be used.
batch – If true,enqueue requests in same context executor.If false, will try to use a different executor for each request.
A vector of global request ids, corresponding to the order of the requests in requests, the id returned may be different from the request id in each executor.
Enqueue generation-only requests to generation executors.
+requests – A vector of generation-only requests.
globalRequestIds – A vector of global request ids, corresponding to the order of the requests,and must be the ids returned by the enqueueContext function.
selectGenIdx – The index of the generation executor to use. If std::nullopt, the executor that has the smallest number of inflight requests will be used.
batch – If true,enqueue requests in same generation executor.If false, will try to use a different executor for each request.
Await for context responses.
+timeout – The maximum time to wait for new responses
contextIdx – The index of the context executor to use. If std::nullopt, return ready responses in all context executors,if hasContextAwaitThreads is true, then this parameter must be std::nullopt.
A vector of responses with corresponding global request ids
+Await for generation responses.
+timeout – The maximum time to wait for new responses.
genIdx – The index of the generation executor to use. If std::nullopt, return ready responses in all generation executors,if hasGenAwaitThreads is true, then this parameter must be std::nullopt.
A vector of responses with corresponding global request ids.
+Indicates if the current process is allowed to enqueueRequests.
+Get context executors.
+Get generation executors.
+Private Members
+Public Functions
+Public Types
+Public Types
+Public Functions
+Returns a pointer to underlying array.
+Returns a pointer to underlying array.
+Returns the memory type of the buffer.
+Returns the number of elements in the tensor.
+Returns the size of the tensor in bytes.
+Set the entire memory to zero.
+stream – Must be a valid CUDA stream if the memory type is GPU.
+Copy the data and shape from another tensor.
+other – A tensor to copy from.
stream – Must be a valid CUDA stream if the memory type is GPU.
Public Static Functions
+Allocate a cpu tensor with the given shape and data type.
+shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a cpu tensor in pinned memory with the given shape and data type.
+shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a cpu tensor in pooled pinned memory with the given shape and data type.
+shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a tensor in managed memory (UVM) with the given shape and data type.
+shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a gpu tensor with the given shape and data type on a particular cuda stream.
+shape – The shape of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
dataType – The data type of the tensor.
Wrap a data pointer into a tensor without taking ownership.
+shape – The shape of the tensor.
dataType – The data type of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
Wrap a data pointer into a tensor without taking ownership.
+shape – The shape of the tensor.
dataType – The data type of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
Wrap any container into a tensor without taking ownership.
+shape – The shape of the tensor.
dataType – The data type of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
Private Functions
+ + +Typedefs
+Public Static Functions
+Public Functions
-Public Functions
-Public Functions
-Constructs a DisaggExecutorOrchestrator object.
-ctxEnginePaths – A vector of file paths to context engine files.
genEnginePaths – A vector of file paths to generation engine files.
ctxExecutorConfigs – A vector of ExecutorConfig for context executors.
genExecutorConfigs – A vector of ExecutorConfig for generation executors.
hasContextAwaitThreads – Whether or not there are threads that receive response for each generation executor.
hasGenAwaitThreads – Whether or not there are threads that receive response for each generation executor.
Enqueue context-only requests to context executors.
-requests – A vector of context-only requests.
selectContextId – The index of the context executor to use. If std::nullopt, the executor that has the smallest number of inflight requests will be used.
batch – If true,enqueue requests in same context executor.If false, will try to use a different executor for each request.
A vector of global request ids, corresponding to the order of the requests in requests, the id returned may be different from the request id in each executor.
Enqueue generation-only requests to generation executors.
-requests – A vector of generation-only requests.
globalRequestIds – A vector of global request ids, corresponding to the order of the requests,and must be the ids returned by the enqueueContext function.
selectGenIdx – The index of the generation executor to use. If std::nullopt, the executor that has the smallest number of inflight requests will be used.
batch – If true,enqueue requests in same generation executor.If false, will try to use a different executor for each request.
Await for context responses.
-timeout – The maximum time to wait for new responses
contextIdx – The index of the context executor to use. If std::nullopt, return ready responses in all context executors,if hasContextAwaitThreads is true, then this parameter must be std::nullopt.
A vector of responses with corresponding global request ids
-Await for generation responses.
-timeout – The maximum time to wait for new responses.
genIdx – The index of the generation executor to use. If std::nullopt, return ready responses in all generation executors,if hasGenAwaitThreads is true, then this parameter must be std::nullopt.
A vector of responses with corresponding global request ids.
-Indicates if the current process is allowed to enqueueRequests.
-Get context executors.
-Get generation executors.
-Private Members
-Public Functions
-Public Types
-Public Types
-Public Functions
-Returns a pointer to underlying array.
-Returns a pointer to underlying array.
-Returns the memory type of the buffer.
-Returns the number of elements in the tensor.
-Returns the size of the tensor in bytes.
-Set the entire memory to zero.
-stream – Must be a valid CUDA stream if the memory type is GPU.
-Copy the data and shape from another tensor.
-other – A tensor to copy from.
stream – Must be a valid CUDA stream if the memory type is GPU.
Public Static Functions
-Allocate a cpu tensor with the given shape and data type.
-shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a cpu tensor in pinned memory with the given shape and data type.
-shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a cpu tensor in pooled pinned memory with the given shape and data type.
-shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a tensor in managed memory (UVM) with the given shape and data type.
-shape – The shape of the tensor.
dataType – The data type of the tensor.
Allocate a gpu tensor with the given shape and data type on a particular cuda stream.
-shape – The shape of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
dataType – The data type of the tensor.
Wrap a data pointer into a tensor without taking ownership.
-shape – The shape of the tensor.
dataType – The data type of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
Wrap a data pointer into a tensor without taking ownership.
-shape – The shape of the tensor.
dataType – The data type of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
Wrap any container into a tensor without taking ownership.
-shape – The shape of the tensor.
dataType – The data type of the tensor.
stream – Specifies the CUDA stream on which to allocate the tensor for GPU memory.
Private Functions
- - -Public Functions
Public Functions
Friends
@@ -10263,1736 +11937,131 @@ -Public Static Functions
+Public Functions
Public Functions
+In order to prepare a dataset, you can use the provided script. +
In order to prepare a dataset, you can use the provided script. To generate a synthetic dataset, run the following command:
python benchmarks/cpp/prepare_dataset.py --tokenizer=$model_name --stdout token-norm-dist --num-requests=$num_requests --input-mean=$isl --output-mean=$osl --input-stdev=0 --output-stdev=0 > $dataset_file
To run the benchmark with the generated data set, simply use the trtllm-bench throughput subcommand. The benchmarker will
run an offline maximum throughput scenario such that all requests are queued in rapid succession. You simply need to provide
-a model name (HuggingFace reference or path to a local model), a generated dataset, and a file containing any desired extra options to the LLM APIs (details in tensorrt_llm/llmapi/llm_args.py:LlmArgs).
For dense / non-MoE models:
trtllm-bench --tp $tp_size --pp $pp_size --model $model_name throughput --dataset $dataset_file --backend pytorch --extra_llm_api_options $llm_options
--kv_cach
--kv_cach
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -1688,9 +1693,9 @@ using the --kv_cach
diff --git a/latest/examples/curl_chat_client.html b/latest/examples/curl_chat_client.html
index 7efbc5c8f6..ef1f3ea393 100644
--- a/latest/examples/curl_chat_client.html
+++ b/latest/examples/curl_chat_client.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -508,7 +516,7 @@
Curl Chat Client#
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#! /usr/bin/env bash
2
3curl http://localhost:8000/v1/chat/completions \
@@ -574,10 +582,12 @@
+
+
@@ -599,12 +609,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -652,9 +657,9 @@
diff --git a/latest/examples/curl_chat_client_for_multimodal.html b/latest/examples/curl_chat_client_for_multimodal.html
index e9bfbd2497..76a86c268c 100644
--- a/latest/examples/curl_chat_client_for_multimodal.html
+++ b/latest/examples/curl_chat_client_for_multimodal.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -508,7 +516,7 @@
Curl Chat Client For Multimodal#
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#! /usr/bin/env bash
2
3# SINGLE IMAGE INFERENCE
@@ -651,10 +659,12 @@
+
+
@@ -676,12 +686,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -729,9 +734,9 @@
diff --git a/latest/examples/curl_completion_client.html b/latest/examples/curl_completion_client.html
index 0e1e24e5d1..3f6a5c50a9 100644
--- a/latest/examples/curl_completion_client.html
+++ b/latest/examples/curl_completion_client.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -508,7 +516,7 @@
Curl Completion Client#
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#! /usr/bin/env bash
2
3curl http://localhost:8000/v1/completions \
@@ -573,10 +581,12 @@
+
+
@@ -598,12 +608,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -651,9 +656,9 @@
diff --git a/latest/examples/customization.html b/latest/examples/customization.html
index 3e57b1265f..162d0ed41e 100644
--- a/latest/examples/customization.html
+++ b/latest/examples/customization.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -690,10 +698,12 @@
+
+
@@ -715,12 +725,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -768,9 +773,9 @@
diff --git a/latest/examples/deepseek_r1_reasoning_parser.html b/latest/examples/deepseek_r1_reasoning_parser.html
index 2ac1161c9f..166c34a10f 100644
--- a/latest/examples/deepseek_r1_reasoning_parser.html
+++ b/latest/examples/deepseek_r1_reasoning_parser.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -508,7 +516,7 @@
Deepseek R1 Reasoning Parser#
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#! /usr/bin/env bash
2
3trtllm-serve \
@@ -573,10 +581,12 @@
+
+
@@ -598,12 +608,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -651,9 +656,9 @@
diff --git a/latest/examples/dynamo_k8s_example.html b/latest/examples/dynamo_k8s_example.html
index ebe12d8954..e6eb6a5862 100644
--- a/latest/examples/dynamo_k8s_example.html
+++ b/latest/examples/dynamo_k8s_example.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -570,10 +578,12 @@ for more details.
+
+
@@ -595,12 +605,7 @@ for more details.
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -648,9 +653,9 @@ for more details.
diff --git a/latest/examples/genai_perf_client.html b/latest/examples/genai_perf_client.html
index 0e4e390f0b..8b32681ce0 100644
--- a/latest/examples/genai_perf_client.html
+++ b/latest/examples/genai_perf_client.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -508,7 +516,7 @@
Genai Perf Client#
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#! /usr/bin/env bash
2
3genai-perf profile \
@@ -579,10 +587,12 @@
+
+
@@ -604,12 +614,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -657,9 +662,9 @@
diff --git a/latest/examples/genai_perf_client_for_multimodal.html b/latest/examples/genai_perf_client_for_multimodal.html
index a86994e98f..727eaf5636 100644
--- a/latest/examples/genai_perf_client_for_multimodal.html
+++ b/latest/examples/genai_perf_client_for_multimodal.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -508,7 +516,7 @@
Genai Perf Client For Multimodal#
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#! /usr/bin/env bash
2
3genai-perf profile \
@@ -582,10 +590,12 @@
+
+
@@ -607,12 +617,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -660,9 +665,9 @@
diff --git a/latest/examples/index.html b/latest/examples/index.html
index 4404e1a0ce..17b2c3ed1e 100644
--- a/latest/examples/index.html
+++ b/latest/examples/index.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -580,10 +588,12 @@
+
+
@@ -605,12 +615,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -658,9 +663,9 @@
diff --git a/latest/examples/kvcacheconfig.html b/latest/examples/kvcacheconfig.html
index 985b238cc5..284bdc450b 100644
--- a/latest/examples/kvcacheconfig.html
+++ b/latest/examples/kvcacheconfig.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -575,10 +583,12 @@
+
+
@@ -600,12 +610,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -653,9 +658,9 @@
diff --git a/latest/examples/kvcacheretentionconfig.html b/latest/examples/kvcacheretentionconfig.html
index 36d0785c47..006422be18 100644
--- a/latest/examples/kvcacheretentionconfig.html
+++ b/latest/examples/kvcacheretentionconfig.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -598,10 +606,12 @@
+
+
@@ -623,12 +633,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -676,9 +681,9 @@
diff --git a/latest/examples/llm_api_examples.html b/latest/examples/llm_api_examples.html
index a63f76e16c..7660b640ea 100644
--- a/latest/examples/llm_api_examples.html
+++ b/latest/examples/llm_api_examples.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -615,10 +623,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1from tensorrt_llm import LLM, SamplingParams
2from tensorrt_llm.llmapi import GuidedDecodingParams
3
@@ -606,10 +614,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1from tensorrt_llm import LLM, SamplingParams
2
3
@@ -594,10 +602,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1import asyncio
2
3from tensorrt_llm import LLM, SamplingParams
@@ -602,10 +610,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1import asyncio
2
3from tensorrt_llm import LLM, SamplingParams
@@ -623,10 +631,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1from tensorrt_llm import LLM, SamplingParams
2
3
@@ -603,10 +611,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1
2import os
3import sys
@@ -806,10 +814,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1'''
2This script demonstrates the effectiveness of KV cache host offloading in TensorRT-LLM.
3
@@ -693,10 +701,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1from typing import List, Optional
2
3import torch
@@ -687,10 +695,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#!/bin/bash
2#SBATCH -A <account> # parameter
3#SBATCH -p <partition> # parameter
@@ -614,10 +622,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#!/bin/bash
2#SBATCH -A <account>
3#SBATCH -p <partition>
@@ -654,10 +662,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1#!/bin/bash
2#SBATCH -A <account>
3#SBATCH -p <partition>
@@ -615,10 +623,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1
2import argparse
3from typing import Optional
@@ -648,10 +656,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1
2import argparse
3
@@ -655,10 +663,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1"""
2This example demonstrates various sampling techniques available in TensorRT-LLM.
3It showcases different sampling parameters and their effects on text generation.
@@ -788,10 +796,12 @@
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1"""
2This example demonstrates how to use sparse attention with TensorRT-LLM.
3
4Supported sparse attention algorithms:
5- RocketKV
- 6
- 7Usage:
- 8```bash
- 9python llm_sparse_attention.py --algo RocketKV --attention_backend TRTLLM --window_size 32 --kernel_size 63 --prompt_budget 2048
- 10```
- 11"""
- 12import argparse
- 13import json
- 14
- 15from tensorrt_llm import LLM, SamplingParams
- 16from tensorrt_llm.llmapi import (CudaGraphConfig, DeepSeekSparseAttentionConfig,
- 17 KvCacheConfig, MoeConfig,
- 18 RocketSparseAttentionConfig)
- 19
+ 6- DSA
+ 7
+ 8Usage:
+ 9```bash
+ 10python llm_sparse_attention.py --algo ROCKETKV --attention_backend TRTLLM --window_size 32 --kernel_size 63 --prompt_budget 2048
+ 11```
+ 12"""
+ 13import argparse
+ 14import json
+ 15
+ 16from tensorrt_llm import LLM, SamplingParams
+ 17from tensorrt_llm.llmapi import (CudaGraphConfig, DeepSeekSparseAttentionConfig,
+ 18 KvCacheConfig, MoeConfig,
+ 19 RocketSparseAttentionConfig)
20
- 21def read_input(input_file):
- 22 results = []
- 23 with open(input_file, 'r') as f:
- 24 for line in f:
- 25 ret = json.loads(line)
- 26 results.append(ret)
- 27 return results
- 28
+ 21
+ 22def read_input(input_file):
+ 23 results = []
+ 24 with open(input_file, 'r') as f:
+ 25 for line in f:
+ 26 ret = json.loads(line)
+ 27 results.append(ret)
+ 28 return results
29
- 30def parse_arguments():
- 31 parser = argparse.ArgumentParser()
- 32 parser.add_argument(
- 33 '--model_path',
- 34 type=str,
- 35 default=
- 36 "/home/scratch.trt_llm_data/llm-models/llama-3.1-model/Llama-3.1-8B-Instruct"
- 37 )
- 38 parser.add_argument(
- 39 '--input_file',
- 40 type=str,
- 41 default="tests/unittest/_torch/multi_gpu/test_star_attention_input.jsonl"
- 42 )
- 43 # Build config
- 44 parser.add_argument('--algo',
- 45 type=str,
- 46 default='ROCKETKV',
- 47 choices=['ROCKETKV', 'DSA'])
- 48 parser.add_argument('--attention_backend',
- 49 type=str,
- 50 default='TRTLLM',
- 51 choices=['VANILLA', 'TRTLLM'])
- 52 parser.add_argument('--window_size',
- 53 type=int,
- 54 default=32,
- 55 help="The window size for RocketKV.")
- 56 parser.add_argument('--kernel_size',
- 57 type=int,
- 58 default=63,
- 59 help="The kernel size for RocketKV.")
- 60 parser.add_argument('--prompt_budget',
- 61 type=int,
- 62 default=2048,
- 63 help="The prompt budget for RocketKV.")
- 64 parser.add_argument('--index_max_chunk_size',
- 65 type=int,
- 66 default=32768,
- 67 help="The maximum chunk size for the indexer.")
- 68 parser.add_argument("--max_seq_len",
- 69 type=int,
- 70 default=8192,
- 71 help="The maximum sequence length.")
- 72 parser.add_argument("--max_batch_size",
- 73 type=int,
- 74 default=256,
- 75 help="The maximum batch size.")
- 76 parser.add_argument("--max_new_tokens",
- 77 type=int,
- 78 default=128,
- 79 help="The maximum new tokens.")
- 80 parser.add_argument(
- 81 "--max_num_tokens",
- 82 type=int,
- 83 default=8192,
- 84 help=
- 85 "The maximum total tokens (context + generation) across all sequences in a batch."
- 86 )
- 87
- 88 # Parallelism
- 89 parser.add_argument('--moe_backend',
- 90 type=str,
- 91 default='CUTLASS',
- 92 choices=[
- 93 'CUTLASS', 'TRTLLM', 'VANILLA', 'WIDEEP',
- 94 'DEEPGEMM', 'CUTEDSL', 'TRITON'
- 95 ])
- 96 parser.add_argument('--tp_size', type=int, default=1)
- 97 parser.add_argument('--moe_ep_size', type=int, default=-1)
- 98 parser.add_argument('--enable_attention_dp',
- 99 default=False,
-100 action='store_true')
-101
-102 # KV cache
-103 parser.add_argument('--kv_cache_dtype', type=str, default='auto')
-104 parser.add_argument("--kv_cache_fraction", type=float, default=None)
-105 parser.add_argument('--num_samples', type=int, default=10)
-106
-107 # Runtime
-108 parser.add_argument('--print_iter_log',
-109 default=False,
-110 action='store_true',
-111 help='Print iteration logs during execution')
-112 parser.add_argument('--use_cuda_graph', default=False, action='store_true')
-113 parser.add_argument('--cuda_graph_padding_enabled',
-114 default=False,
-115 action='store_true')
-116 parser.add_argument('--cuda_graph_batch_sizes',
-117 nargs='+',
-118 type=int,
-119 default=None)
-120 args = parser.parse_args()
-121 return args
-122
+ 30
+ 31def parse_arguments():
+ 32 parser = argparse.ArgumentParser()
+ 33 parser.add_argument(
+ 34 '--model_path',
+ 35 type=str,
+ 36 default=
+ 37 "/home/scratch.trt_llm_data/llm-models/llama-3.1-model/Llama-3.1-8B-Instruct"
+ 38 )
+ 39 parser.add_argument(
+ 40 '--input_file',
+ 41 type=str,
+ 42 default="tests/unittest/_torch/multi_gpu/test_star_attention_input.jsonl"
+ 43 )
+ 44 # Build config
+ 45 parser.add_argument('--algo',
+ 46 type=str,
+ 47 default='ROCKETKV',
+ 48 choices=['ROCKETKV', 'DSA'])
+ 49 parser.add_argument('--attention_backend',
+ 50 type=str,
+ 51 default='TRTLLM',
+ 52 choices=['VANILLA', 'TRTLLM'])
+ 53 parser.add_argument('--window_size',
+ 54 type=int,
+ 55 default=32,
+ 56 help="The window size for RocketKV.")
+ 57 parser.add_argument('--kernel_size',
+ 58 type=int,
+ 59 default=63,
+ 60 help="The kernel size for RocketKV.")
+ 61 parser.add_argument('--prompt_budget',
+ 62 type=int,
+ 63 default=2048,
+ 64 help="The prompt budget for RocketKV.")
+ 65 parser.add_argument('--index_max_chunk_size',
+ 66 type=int,
+ 67 default=32768,
+ 68 help="The maximum chunk size for the indexer.")
+ 69 parser.add_argument("--max_seq_len",
+ 70 type=int,
+ 71 default=10240,
+ 72 help="The maximum sequence length.")
+ 73 parser.add_argument("--max_batch_size",
+ 74 type=int,
+ 75 default=256,
+ 76 help="The maximum batch size.")
+ 77 parser.add_argument("--max_new_tokens",
+ 78 type=int,
+ 79 default=128,
+ 80 help="The maximum new tokens.")
+ 81 parser.add_argument(
+ 82 "--max_num_tokens",
+ 83 type=int,
+ 84 default=81920,
+ 85 help=
+ 86 "The maximum total tokens (context + generation) across all sequences in a batch."
+ 87 )
+ 88
+ 89 # Parallelism
+ 90 parser.add_argument('--moe_backend',
+ 91 type=str,
+ 92 default='CUTLASS',
+ 93 choices=[
+ 94 'CUTLASS', 'TRTLLM', 'VANILLA', 'WIDEEP',
+ 95 'DEEPGEMM', 'CUTEDSL', 'TRITON'
+ 96 ])
+ 97 parser.add_argument('--tp_size', type=int, default=1)
+ 98 parser.add_argument('--moe_ep_size', type=int, default=-1)
+ 99 parser.add_argument('--enable_attention_dp',
+100 default=False,
+101 action='store_true')
+102
+103 # KV cache
+104 parser.add_argument('--kv_cache_dtype', type=str, default='auto')
+105 parser.add_argument("--kv_cache_fraction", type=float, default=0.7)
+106 parser.add_argument('--num_samples', type=int, default=10)
+107
+108 # Runtime
+109 parser.add_argument('--print_iter_log',
+110 default=False,
+111 action='store_true',
+112 help='Print iteration logs during execution')
+113 parser.add_argument('--use_cuda_graph', default=False, action='store_true')
+114 parser.add_argument('--cuda_graph_padding_enabled',
+115 default=False,
+116 action='store_true')
+117 parser.add_argument('--cuda_graph_batch_sizes',
+118 nargs='+',
+119 type=int,
+120 default=None)
+121 args = parser.parse_args()
+122 return args
123
-124def run_llm(args, sparse_attention_config):
-125 data = read_input(args.input_file)
-126 num_samples = args.num_samples if args.num_samples is not None else len(
-127 data)
-128 data = data[:num_samples]
-129
-130 kv_cache_config = KvCacheConfig(
-131 enable_block_reuse=
-132 False, # sparse attention does not support kv cache reuse now
-133 free_gpu_memory_fraction=args.kv_cache_fraction,
-134 dtype=args.kv_cache_dtype,
-135 )
-136
-137 cuda_graph_config = CudaGraphConfig(
-138 batch_sizes=args.cuda_graph_batch_sizes,
-139 enable_padding=args.cuda_graph_padding_enabled,
-140 ) if args.use_cuda_graph else None
-141
-142 llm = LLM(
-143 model=args.model_path,
-144 backend='pytorch',
-145 kv_cache_config=kv_cache_config,
-146 attn_backend=args.attention_backend,
-147 sparse_attention_config=sparse_attention_config,
-148 max_batch_size=args.max_batch_size,
-149 max_seq_len=args.max_seq_len,
-150 max_num_tokens=args.max_num_tokens,
-151 tensor_parallel_size=args.tp_size,
-152 moe_expert_parallel_size=args.moe_ep_size,
-153 enable_attention_dp=args.enable_attention_dp,
-154 cuda_graph_config=cuda_graph_config,
-155 print_iter_log=args.print_iter_log,
-156 enable_iter_perf_stats=args.print_iter_log,
-157 moe_config=MoeConfig(backend=args.moe_backend),
-158 )
-159
-160 prompts = []
-161 reference = []
-162 for sample in data:
-163 prompts.append(
-164 {'prompt': sample['input_context'] + sample['input_query']})
-165 reference.append(sample['outputs'])
-166
-167 sampling_params = SamplingParams(add_special_tokens=False,
-168 max_tokens=args.max_new_tokens,
-169 temperature=0.8,
-170 top_p=0.95)
-171
-172 outputs = llm.generate(prompts, sampling_params)
-173 for idx, output in enumerate(outputs):
-174 print(
-175 f'Generated text: {output.outputs[0].text!r}, ref: {reference[idx]}'
-176 )
-177
+124
+125def run_llm(args, sparse_attention_config):
+126 data = read_input(args.input_file)
+127 num_samples = args.num_samples if args.num_samples is not None else len(
+128 data)
+129 data = data[:num_samples]
+130
+131 kv_cache_config = KvCacheConfig(
+132 enable_block_reuse=
+133 False, # sparse attention does not support kv cache reuse now
+134 free_gpu_memory_fraction=args.kv_cache_fraction,
+135 dtype=args.kv_cache_dtype,
+136 )
+137
+138 cuda_graph_config = CudaGraphConfig(
+139 batch_sizes=args.cuda_graph_batch_sizes,
+140 enable_padding=args.cuda_graph_padding_enabled,
+141 ) if args.use_cuda_graph else None
+142
+143 llm = LLM(
+144 model=args.model_path,
+145 backend='pytorch',
+146 kv_cache_config=kv_cache_config,
+147 attn_backend=args.attention_backend,
+148 sparse_attention_config=sparse_attention_config,
+149 max_batch_size=args.max_batch_size,
+150 max_seq_len=args.max_seq_len,
+151 max_num_tokens=args.max_num_tokens,
+152 tensor_parallel_size=args.tp_size,
+153 moe_expert_parallel_size=args.moe_ep_size,
+154 enable_attention_dp=args.enable_attention_dp,
+155 cuda_graph_config=cuda_graph_config,
+156 print_iter_log=args.print_iter_log,
+157 enable_iter_perf_stats=args.print_iter_log,
+158 moe_config=MoeConfig(backend=args.moe_backend),
+159 )
+160
+161 prompts = []
+162 reference = []
+163 for sample in data:
+164 prompts.append(
+165 {'prompt': sample['input_context'] + sample['input_query']})
+166 reference.append(sample['outputs'])
+167
+168 sampling_params = SamplingParams(add_special_tokens=False,
+169 max_tokens=args.max_new_tokens,
+170 temperature=0.8,
+171 top_p=0.95)
+172
+173 outputs = llm.generate(prompts, sampling_params)
+174 for idx, output in enumerate(outputs):
+175 print(
+176 f'Generated text: {output.outputs[0].text!r}, ref: {reference[idx]}'
+177 )
178
-179def run_RocketKV(args):
-180 sparse_attention_config = RocketSparseAttentionConfig(
-181 window_size=args.window_size,
-182 kernel_size=args.kernel_size,
-183 prompt_budget=args.prompt_budget,
-184 )
-185 run_llm(args, sparse_attention_config)
-186
+179
+180def run_RocketKV(args):
+181 sparse_attention_config = RocketSparseAttentionConfig(
+182 window_size=args.window_size,
+183 kernel_size=args.kernel_size,
+184 prompt_budget=args.prompt_budget,
+185 )
+186 run_llm(args, sparse_attention_config)
187
-188def run_DSA(args):
-189 sparse_attention_config = DeepSeekSparseAttentionConfig(
-190 indexer_max_chunk_size=args.index_max_chunk_size, )
-191 run_llm(args, sparse_attention_config)
-192
+188
+189def run_DSA(args):
+190 sparse_attention_config = DeepSeekSparseAttentionConfig(
+191 indexer_max_chunk_size=args.index_max_chunk_size, )
+192 run_llm(args, sparse_attention_config)
193
-194def main():
-195 args = parse_arguments()
-196 if args.algo == 'ROCKETKV':
-197 run_RocketKV(args)
-198 elif args.algo == 'DSA':
-199 run_DSA(args)
-200 else:
-201 raise ValueError(f"Invalid algorithm: {args.algo}")
-202
+194
+195def main():
+196 args = parse_arguments()
+197 if args.algo == 'ROCKETKV':
+198 run_RocketKV(args)
+199 elif args.algo == 'DSA':
+200 run_DSA(args)
+201 else:
+202 raise ValueError(f"Invalid algorithm: {args.algo}")
203
-204if __name__ == "__main__":
-205 main()
+204
+205if __name__ == "__main__":
+206 main()
Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1from typing import Optional
2
3import click
@@ -654,10 +662,12 @@
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1
2from openai import OpenAI
3
@@ -583,10 +591,12 @@
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1
2import os
3from pathlib import Path
@@ -691,10 +699,12 @@
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1
2from openai import OpenAI
3
@@ -577,10 +585,12 @@
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1### OpenAI Completion Client
2
3import os
@@ -593,10 +601,12 @@
Refer to the trtllm-serve documentation for starting a server.
-Source NVIDIA/TensorRT-LLM.
+Source NVIDIA/TensorRT-LLM.
1
2# This example requires to specify `guided_decoding_backend` as
3# `xgrammar` or `llguidance` in the extra_llm_api_options.yaml file.
@@ -614,10 +622,12 @@
sequence.additi
sequence.additi
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -719,9 +724,9 @@ and sequence.additi
diff --git a/latest/features/attention.html b/latest/features/attention.html
index aa7325c0f2..5a50ed43c7 100644
--- a/latest/features/attention.html
+++ b/latest/features/attention.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,22 +61,26 @@
+
+
+
-
+
+
@@ -82,6 +88,8 @@
+
+
@@ -844,7 +852,7 @@ reach that point).
the different requests by a cache manager during processing. That cache manager
keeps track of the sequences, allocates new blocks from a pool and recycles those
blocks when required. See the implementation of
-KVCacheManager.
+KVCacheManager.
INT8/FP8 KV Caches#
@@ -1060,10 +1068,12 @@ is computed as:
+
@@ -681,12 +691,7 @@ We welcome community contributions, see Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -734,9 +739,9 @@ We welcome community contributions, see
diff --git a/latest/features/auto_deploy/support_matrix.html b/latest/features/auto_deploy/support_matrix.html
index 624198c15f..6156140473 100644
--- a/latest/features/auto_deploy/support_matrix.html
+++ b/latest/features/auto_deploy/support_matrix.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -722,10 +730,12 @@ In addition, the following models have been officially validated using the defau
The following examples demonstrate how to use TensorRT LLM’s multimodal support in various scenarios, including quick run examples, serving endpoints, and performance benchmarking.
Quickly try out TensorRT LLM’s multimodal support using our LLM-API and a ready-to-run example:
python3 quickstart_multimodal.py --model_dir Efficient-Large-Model/NVILA-8B --modality image
+Quickly try out TensorRT LLM’s multimodal support using our LLM-API and a ready-to-run example:
+python3 quickstart_multimodal.py --model_dir Efficient-Large-Model/NVILA-8B --modality image --disable_kv_cache_reuse
@@ -549,7 +557,7 @@
trtllm-serve Qwen/Qwen2-VL-7B-Instruct --backend pytorch
-You can then send OpenAI-compatible requests, such as via curl or API clients, to the server endpoint. See curl chat client for multimodal script as an example.
+You can then send OpenAI-compatible requests, such as via curl or API clients, to the server endpoint. See curl chat client for multimodal script as an example.
Run with trtllm-bench#
@@ -632,10 +640,12 @@
tensorrt_llm.runtime.KVCacheManager.
+tensorrt_llm.runtime.KVCacheManager.
A more efficient C++ implementation is included in the
-Batch Manager.
+Batch Manager.
To use the feature:
Enable the enable_trtllm_sampler option in the LLM class
Pass a SamplingParams object with the desired options to the generate() function
Pass a SamplingParams object with the desired options to the generate() function
The following example prepares two identical prompts which will give different results due to the sampling parameters chosen:
from tensorrt_llm import LLM, SamplingParams
@@ -579,7 +587,7 @@
To enable guided decoding, you must:
Set the guided_decoding_backend parameter to 'xgrammar' or 'llguidance' in the LLM class
-Create a GuidedDecodingParams object with the desired format specification
+Create a GuidedDecodingParams object with the desired format specification
Note: Depending on the type of format, a different parameter needs to be chosen to construct the object (json, regex, grammar, structural_tag).
@@ -600,14 +608,14 @@
llm.generate("Generate a JSON response", sampling_params)
You can find a more detailed example on guided decoding here.
+You can find a more detailed example on guided decoding here.
Logits processors allow you to modify the logits produced by the network before sampling, enabling custom generation behavior and constraints.
To use a custom logits processor:
Create a custom class that inherits from LogitsProcessor and implements the __call__ method
Create a custom class that inherits from LogitsProcessor and implements the __call__ method
Pass an instance of this class to the logits_processor parameter of SamplingParams
The following example demonstrates logits processing:
@@ -635,7 +643,7 @@ llm.generate(["Hello, my name is"], sampling_params)You can find a more detailed example on logits processors here.
+You can find a more detailed example on logits processors here.
To use the C++ benchmark scripts under benchmark/cpp, for example gptManagerBenchmark.cpp, add the --benchmarks option:
To use the C++ benchmark scripts under benchmark/cpp, for example gptManagerBenchmark.cpp, add the --benchmarks option:
python3 ./scripts/build_wheel.py --benchmarks
Container image tags
In the example shell commands, x.y.z corresponds to the TensorRT-LLM container
version to use. If omitted, IMAGE_TAG will default to tensorrt_llm.__version__
-(e.g., this documentation was generated from the 1.2.0rc2 source tree).
+(e.g., this documentation was generated from the 1.2.0rc3 source tree).
If this does not work, e.g., because a container for the version you are
currently working with has not been released yet, you can try using a
container published for a previous
@@ -575,10 +583,12 @@ for all related options.
Prevent pip from replacing existing PyTorch installation
On certain systems, particularly Ubuntu 22.04, users installing TensorRT LLM would find that their existing, CUDA 13.0 compatible PyTorch installation (e.g., torch==2.9.0+cu130) was being uninstalled by pip. It was then replaced by a CUDA 12.8 version (torch==2.9.0), causing the TensorRT LLM installation to be unusable and leading to runtime errors.
The solution is to create a pip constraints file, locking torch to the currently installed version. Here is an example of how this can be done manually:
CURRENT_TORCH_VERSION=$(python3 -c "import torch; print(torch.__version__)")
+echo "torch==$CURRENT_TORCH_VERSION" > /tmp/torch-constraint.txt
+pip3 install --upgrade pip setuptools && pip3 install tensorrt_llm -c /tmp/torch-constraint.txt
+TensorRT-LLM includes a high-level C++ API called the Executor API which allows you to execute requests asynchronously, with in-flight batching, and without the need to define callbacks.
A software component (referred to as “the client” in the text that follows) can interact
-with the executor using the API defined in the executor.h file.
+with the executor using the API defined in the executor.h file.
For details about the API, refer to the _cpp_gen/executor.rst.
The following sections provide an overview of the main classes defined in the Executor API.
Refer to tensorrt_llm/llmapi/tokenizer.py for more details. You may dump these materials to disk, and reload them to C++ runtime for use.
Refer to tensorrt_llm/llmapi/tokenizer.py for more details. You may dump these materials to disk, and reload them to C++ runtime for use.
Each request can be optionally specified with a GuidedDecodingParams, which defines the desired structured format. Currently, it supports four types:
GuidedDecodingParams::GuideType::kJSON: The generated text is amenable to JSON format;
Two C++ examples are provided that shows how to use the Executor API and can be found in the examples/cpp/executor folder.
Two C++ examples are provided that shows how to use the Executor API and can be found in the examples/cpp/executor folder.
Python bindings for the Executor API are also available to use the Executor API from Python. The Python bindings are defined in bindings.cpp and once built, are available in package tensorrt_llm.bindings.executor. Running 'help('tensorrt_llm.bindings.executor') in a Python interpreter will provide an overview of the classes available.
In addition, three Python examples are provided to demonstrate how to use the Python bindings to the Executor API for single and multi-GPU models. They can be found in examples/bindings.
Python bindings for the Executor API are also available to use the Executor API from Python. The Python bindings are defined in bindings.cpp and once built, are available in package tensorrt_llm.bindings.executor. Running 'help('tensorrt_llm.bindings.executor') in a Python interpreter will provide an overview of the classes available.
In addition, three Python examples are provided to demonstrate how to use the Python bindings to the Executor API for single and multi-GPU models. They can be found in examples/bindings.
tensorrt_llm.runtime.KVCacheManager.
+tensorrt_llm.runtime.KVCacheManager.
A more efficient C++ implementation is included in the
-Batch Manager.
+Batch Manager.
GptDecoder
GptDecoderManage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -1017,9 +1022,9 @@ The GptDecoder
LoraWe
LoraWe
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -827,9 +832,9 @@ The shape of LoraWe
diff --git a/latest/legacy/advanced/lowprecision-pcie-allreduce.html b/latest/legacy/advanced/lowprecision-pcie-allreduce.html
index 2eb881e30d..903e68306c 100644
--- a/latest/legacy/advanced/lowprecision-pcie-allreduce.html
+++ b/latest/legacy/advanced/lowprecision-pcie-allreduce.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -618,10 +626,12 @@ This feature is optimized for PCIe-based GPU topologies and may affect model acc
In TensorRT-LLM, the tensorrt_llm.Builder class
+
In TensorRT-LLM, the tensorrt_llm.Builder class
contains a
tensorrt.Builder
object. That instance is used in the tensorrt_llm.Builder.create_network
@@ -521,7 +529,7 @@ method to create an instance of the
tensorrt.INetworkDefinition
class. The INetworkDefinition object can then be populated using the free
functions defined in the
-tensorrt_llm.functional.
tensorrt_llm.functional.
A simple example of such a free function is tensorrt_llm.activation that inserts a
tensorrt.IActivationLayer
node in the graph of the model:
The plugins are nodes inserted in the network graph definition that map to user-defined
GPU kernels. TensorRT-LLM uses a number of such plugins. They can be found in
-the cpp/tensorrt_llm/plugins directory.
cpp/tensorrt_llm/plugins directory.
Plugins are written in C++ and follow a well-defined interface described in the
Extending TensorRT with Custom Layers
section of the TensorRT
Developer Guide.
When executed within a TensorRT engine, plugins trigger the execution of
their encapsulated GPU kernels. A fairly simple example of plugins is the
-QuantizeTensorPlugin that
+QuantizeTensorPlugin that
triggers a CUDA kernel in the QuantizeTensorPlugin::enqueue member function:
// In cpp/tensorrt_llm/plugins/quantizeTensorPlugin/quantizeTensorPlugin.cpp:
@@ -707,7 +715,7 @@ using TensorRT plugins that wrap communication primitives from the
plugin that optimize the All-Reduce primitive in the presence of All-to-all
connections between GPUs (through NVSwitch in DGX systems).
The communication plugins can be found in
-cpp/tensorrt_llm/plugins/ncclPlugin
+cpp/tensorrt_llm/plugins/ncclPlugin
and the multi-GPU functions are exposed in the TensorRT-LLM Model Definition API
as:
# In tensorrt_llm/functional.py:
@@ -919,10 +927,12 @@ srun \
+
+
@@ -944,12 +954,7 @@ srun \
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -997,9 +1002,9 @@ srun \
diff --git a/latest/legacy/architecture/model-weights-loader.html b/latest/legacy/architecture/model-weights-loader.html
index b042b00390..5dd1c45207 100644
--- a/latest/legacy/architecture/model-weights-loader.html
+++ b/latest/legacy/architecture/model-weights-loader.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -857,10 +865,12 @@ The support for Qwen-1 is in
+
+
@@ -882,12 +892,7 @@ The support for Qwen-1 is in Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -935,9 +940,9 @@ The support for Qwen-1 is in
diff --git a/latest/legacy/architecture/workflow.html b/latest/legacy/architecture/workflow.html
index 896a1bfce8..d6bc4432ac 100644
--- a/latest/legacy/architecture/workflow.html
+++ b/latest/legacy/architecture/workflow.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -745,10 +753,12 @@ The usage of this API looks like this:
Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -1579,9 +1584,9 @@ The choices are specified with a YAML file like the following example (
KVCacheConfig
KVCacheConfig
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -835,9 +840,9 @@ via KVCacheConfig
diff --git a/latest/legacy/python-api/tensorrt_llm.functional.html b/latest/legacy/python-api/tensorrt_llm.functional.html
index cb88b651ef..0145aa492d 100644
--- a/latest/legacy/python-api/tensorrt_llm.functional.html
+++ b/latest/legacy/python-api/tensorrt_llm.functional.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -6955,10 +6963,12 @@ The index for rebuilding the sequence
TensorRT-LLM C++ runtime is using stream-ordered memory allocator to allocate and free buffers, see BufferManager::initMemoryPool, which uses the default memory pool managed by the CUDA driver. When a TrtGptModel object is destroyed, memory is returned to the memory pool and can be reused by the next instance of a TrtGptModel object. Memory will be released from the pool if it is required for other memory allocations.
However, nvidia-smi may still show high memory occupation after memory is returned to the CUDA driver’s memory pool. This should not be a concern and is intended behavior. The amount of reserved and free memory in the pool can be inspected by BufferManager::memoryPoolReserved()) and BufferManager::memoryPoolFree()), respectively.
TensorRT-LLM C++ runtime is using stream-ordered memory allocator to allocate and free buffers, see BufferManager::initMemoryPool, which uses the default memory pool managed by the CUDA driver. When a TrtGptModel object is destroyed, memory is returned to the memory pool and can be reused by the next instance of a TrtGptModel object. Memory will be released from the pool if it is required for other memory allocations.
However, nvidia-smi may still show high memory occupation after memory is returned to the CUDA driver’s memory pool. This should not be a concern and is intended behavior. The amount of reserved and free memory in the pool can be inspected by BufferManager::memoryPoolReserved()) and BufferManager::memoryPoolFree()), respectively.
Examples of how to enable SmoothQuant for GPT, GPT-J and LLaMA can be found in -the examples/quantization folder of that release.
+the examples/quantization folder of that release.To use INT4/INT8 Weight-Only methods, the user must determine the scaling factors to use to quantize and dequantize the weights of the model.
- +weight_only_groupwise_quant_matmul
+weight_only_groupwise_quant_matmul
Python function, for details.
-This release includes examples of applying GPTQ to GPT-NeoX -and LLaMA-v2, as well as an example of using AWQ with -GPT-J.
+This release includes examples of applying GPTQ to GPT-NeoX +and LLaMA-v2, as well as an example of using AWQ with +GPT-J.
This release of TensorRT-LLM contains implementations of FP8 for GPT-NeMo, GPT-J and LLaMA. Those examples can be found in -examples/quantization.
+examples/quantization.QuantMode Flags#The quantization method is controlled by the
-QuantMode flags. The different fields
+QuantMode flags. The different fields
are:
INT4_WEIGHTS, the weights are quantized to 4 bits (W4A*),
moe_expert_parallel_size (Optional[int]) – stable The expert parallel size for MoE models’s expert weights. Defaults to None.
enable_attention_dp (bool) – beta Enable attention data parallel. Defaults to False.
enable_lm_head_tp_in_adp (bool) – prototype Enable LM head TP in attention dp. Defaults to False.
pp_partition (Optional[List[int]]) – prototype Pipeline parallel partition, a list of each rank’s layer number. Defaults to None.
cp_config (Optional[dict]) – prototype Context parallel config. Defaults to None.
load_format (Union[str, tensorrt_llm.llmapi.llm_args.LoadFormat]) – stable How to load the model weights. By default, detect the weight type from the model checkpoint. Defaults to 0.
fail_fast_on_attention_window_too_large (bool) – prototype Fail fast when attention window is too large to fit even a single sequence in the KV cache. Defaults to False.
Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+Bases: DecodingBaseConfig
Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+The communication backend type to use for the cache transceiver.
Timeout in milliseconds to wait for the sender future to be ready when scheduled batch size is 0. This allows the request to be eventually cancelled by the user or because of kv_transfer_timeout_ms
+gt = 0
Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+prototype The path to the tokenizer directory for postprocessing.
prototype Pipeline parallel partition, a list of each rank’s layer number.
The path to the tokenizer directory for postprocessing.
Pipeline parallel partition, a list of each rank’s layer number.
+Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+Do any additional error checking here.
Validate and sort draft_len_schedule by batch size thresholds.
+LookaheadDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduleload_formatmax_concurrencymax_draft_lensupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortvalidate_positive_valuesdecoding_typemodel_computed_fieldsMedusaDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduleload_formatmax_concurrencymax_draft_lensupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typemodel_computed_fieldsmodel_configEagleDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduledynamic_tree_max_topKeagle3_layers_to_captureeagle3_one_modelsupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typeis_linear_treemodel_computed_fieldsMTPDecodingConfigBEGIN_THINKING_PHASE_TOKENEND_THINKING_PHASE_TOKENacceptance_length_thresholdacceptance_windowbegin_thinking_phase_tokendraft_len_scheduleend_thinking_phase_tokenload_formatmax_concurrencymax_draft_lensupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typemodel_computed_fieldsmodel_configCacheTransceiverConfigbackendkv_transfer_sender_future_timeout_mskv_transfer_timeout_msmax_tokens_in_bufferConfigNGramDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduleis_keep_allis_public_poolis_use_oldestsupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typemodel_computed_fieldsmodel_configUserProvidedDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduledrafterload_formatmax_concurrencysupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typemodel_computed_fieldsmodel_configDraftTargetDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduleload_formatmax_concurrencymax_draft_lensupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typemodel_computed_fieldsmodel_configperf_metrics_max_requestspipeline_parallel_sizepostprocess_tokenizer_dirpp_partitionprint_iter_logray_worker_extension_clsreasoning_parserconvert_load_formatfrom_kwargs()get_executor_config()get_pytorch_backend_config()get_runtime_sizes()init_backendinit_build_configpeft_cache_configpipeline_parallel_sizepostprocess_tokenizer_dirpp_partitionquant_configreasoning_parserrequest_stats_max_iterationsAutoDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduleload_formatmax_concurrencymax_draft_lensupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typemodel_computed_fieldsmodel_configSaveHiddenStatesDecodingConfigacceptance_length_thresholdacceptance_windowdraft_len_scheduleeagle3_layers_to_captureeagle_choicesfile_prefixsupports_backend()update_forward_refs()validate()validate_draft_len_schedule_and_sortdecoding_typemodel_computed_fieldsmodel_configdict()from_dict()from_orm()get_indices_block_size()json()model_construct()model_copy()dict()from_dict()from_orm()get_indices_block_size()json()model_construct()model_copy()TensorRT LLM provides a high-level Python LLM API that supports a wide range of inference setups - from single-GPU to multi-GPU or multi-node deployments. It includes built-in support for various parallelism strategies and advanced features. The LLM API integrates seamlessly with the broader inference ecosystem, including NVIDIA Dynamo and the Triton Inference Server.
-TensorRT LLM is designed to be modular and easy to modify. Its PyTorch-native architecture allows developers to experiment with the runtime or extend functionality. Several popular models are also pre-defined and can be customized using native PyTorch code, making it easy to adapt the system to specific needs.
+TensorRT LLM is designed to be modular and easy to modify. Its PyTorch-native architecture allows developers to experiment with the runtime or extend functionality. Several popular models are also pre-defined and can be customized using native PyTorch code, making it easy to adapt the system to specific needs.
The TensorRT LLM container maintained by NVIDIA contains all of the required dependencies pre-installed. You can start the container on a machine with NVIDIA GPUs via:
-docker run --rm -it --ipc host --gpus all --ulimit memlock=-1 --ulimit stack=67108864 -p 8000:8000 nvcr.io/nvidia/tensorrt-llm/release:1.2.0rc2
+docker run --rm -it --ipc host --gpus all --ulimit memlock=-1 --ulimit stack=67108864 -p 8000:8000 nvcr.io/nvidia/tensorrt-llm/release:1.2.0rc3
@@ -716,10 +724,12 @@ To learn more about the LLM API, check out the
trtllm-serve", "trtllm-serve", "Deployment Guide for DeepSeek R1 on TensorRT LLM - Blackwell & Hopper Hardware", "Deployment Guide for GPT-OSS on TensorRT-LLM - Blackwell Hardware", "Deployment Guide for Llama3.3 70B on TensorRT LLM - Blackwell & Hopper Hardware", "Deployment Guide for Llama4 Scout 17B on TensorRT LLM - Blackwell & Hopper Hardware", "Deployment Guide for Qwen3 Next on TensorRT LLM - Blackwell & Hopper Hardware", "Model Recipes", "LLM API Change Guide", "Continuous Integration Overview", "Using Dev Containers", "Introduction to KV Cache Transmission", "Architecture Overview", "Performance Analysis", "TensorRT LLM Benchmarking", "Overview", "Curl Chat Client", "Curl Chat Client For Multimodal", "Curl Completion Client", "LLM Common Customizations", "Deepseek R1 Reasoning Parser", "Dynamo K8s Example", "Genai Perf Client", "Genai Perf Client For Multimodal", "LLM Examples Introduction", "How to Change KV Cache Behavior", "How to Change Block Priorities", "LLM Examples", "Generate text with guided decoding", "Generate text", "Generate text asynchronously", "Generate text in streaming", "Distributed LLM Generation", "KV Cache Connector", "KV Cache Offloading", "Control generated text using logits processor", "Run LLM-API with pytorch backend on Slurm", "Run trtllm-bench with pytorch backend on Slurm", "Run trtllm-serve with pytorch backend on Slurm", "Generate text with multiple LoRA adapters", "Runtime Configuration Examples", "Sampling Techniques Showcase", "Sparse Attention", "Speculative Decoding", "OpenAI Chat Client", "OpenAI Chat Client for Multimodal", "OpenAI Completion Client", "Openai Completion Client For Lora", "OpenAI Completion Client with JSON Schema", "Online Serving Examples", "Additional Outputs", "Multi-Head, Multi-Query, and Group-Query Attention", "Benchmarking with trtllm-bench", "Example Run Script", "Expert Configuration of LLM API", "Logging Level", "Incorporating auto_deploy into your own workflow", "AutoDeploy (Prototype)", "Support Matrix", "Checkpoint Loading", "Disaggregated Serving", "Feature Combination Matrix", "KV Cache System", "Long Sequences", "LoRA (Low-Rank Adaptation)", "Multimodal Support in TensorRT LLM", "Overlap Scheduler", "Paged Attention, IFB, and Request Scheduling", "Parallelism in TensorRT LLM", "Quantization", "Ray Orchestrator (Prototype)", "Sampling", "Speculative Decoding", "Torch Compile & Piecewise CUDA Graph", "Welcome to TensorRT LLM\u2019s Documentation!", "Building from Source Code on Linux", "Pre-built release container images on NGC", "Installation", "Installing on Linux via pip", "Disaggregated-Service (Prototype)", "Executor API", "Expert Parallelism in TensorRT-LLM", "Multi-Head, Multi-Query, and Group-Query Attention", "C++ GPT Runtime", "Graph Rewriting Module", "KV Cache Management: Pools, Blocks, and Events", "KV cache reuse", "Run gpt-2b + LoRA using Executor / cpp runtime", "Low-Precision-AllReduce", "<no title>", "Speculative Sampling", "Running With Weight Streaming to Reduce GPU Memory Consumption", "Adding a Model", "TensorRT LLM Checkpoint", "Model Definition", "TensorRT-LLM Model Weights Loader", "TensorRT-LLM Build Workflow", "Build the TensorRT LLM Docker Image", "Develop TensorRT LLM on Runpod", "Key Features", "Performance Analysis", "TensorRT-LLM Benchmarking", "Benchmarking Default Performance", "Deciding Model Sharding Strategy", "FP8 Quantization", "Performance Tuning Guide", "Prerequisite Knowledge", "Tuning Max Batch Size and Max Num Tokens", "Useful Build-Time Flags", "Useful Runtime Options", "Functionals", "Layers", "Models", "Plugin", "Quantization", "Runtime", "Memory Usage of TensorRT-LLM", "Multimodal Feature Support Matrix (PyTorch Backend)", "Numerical Precision", "Support Matrix", "Troubleshooting", "LLM API with TensorRT Engine", "PyTorch Backend", "LLM API Introduction", "API Reference", "Adding a New Model", "Supported Models", "Overview", "Quick Start Guide", "Release Notes", "Adding a New Model in PyTorch Backend", "Architecture Ovewiew", "Attention", "Benchmarking with trtllm-bench", "Example Run Script", "Expert Configuration of LLM API", "Logging Level", "Serving with trtllm-serve", "Incorporating auto_deploy into your own workflow", "AutoDeploy", "Support Matrix", "Checkpoint Loading", "LoRA (Low-Rank Adaptation)", "Overlap Scheduler", "Quantization", "Sampling", "KV Cache Manager", "Scheduler"], "titleterms": {"": [4, 7, 77, 100, 108], "0": [149, 155], "000": [4, 5], "0528": 2, "0rc0": 9, "1": [2, 8, 9, 16, 18, 19, 20, 34, 41, 85, 94, 101, 118, 120, 142, 149, 155, 167], "10": [4, 155], "100m": 4, "1024": 17, "11": 155, "12": [5, 155], "1200": 17, "120b": [9, 21], "13": 155, "13b": 5, "14": 155, "15": 155, "16": 155, "17": 155, "17b": 31, "18": 155, "180b": 3, "19": 155, "2": [2, 6, 8, 16, 18, 19, 20, 34, 85, 94, 101, 118, 142, 149, 155, 167], "20": 155, "21": 155, "235b": 41, "256": 17, "2b": 113, "3": [2, 12, 16, 17, 18, 20, 30, 34, 41, 85, 94, 98, 118, 120, 127, 142, 145, 167], "30b": 41, "4": [2, 4, 18, 20, 41, 85, 118, 167], "405b": [41, 120], "4096": 17, "4400": 17, "4x": 6, "5": [2, 18], "6": [2, 3, 18], "6x": 4, "7": [18, 155], "70b": [3, 6, 30, 41, 120, 127], "7x": 3, "8": 155, "8192": 17, "8b": 41, "9": 155, "A": [11, 15], "As": 106, "For": [43, 49, 73], "In": [77, 93, 106, 108], "It": [92, 169], "Not": 142, "One": [13, 101], "The": [11, 16, 77, 88, 93, 106, 144], "To": 128, "With": [117, 153], "_prepare_draft_request": 98, "_prepare_draft_token": 98, "_torchllm": 34, "a100": [3, 4], "a22b": 41, "a3b": 41, "about": [24, 26, 27, 116, 129, 153], "absorb": 15, "accept": [13, 14], "access": [30, 31, 123], "account": 124, "accuraci": [7, 14, 28, 29, 30, 31, 114], "achiev": [4, 5, 14], "acknowledg": [8, 10, 11, 12, 13, 14, 15, 16, 17, 20], "across": 88, "activ": [137, 142], "ad": [34, 118, 151, 156], "adapt": [40, 65, 90, 127, 168], "add": [12, 34], "addit": [76, 106], "additional_model_output": 76, "address": 20, "adp": [8, 15], "advanc": [78, 83, 90, 101, 153, 159, 165, 168], "after": 149, "agent": 37, "algorithm": [19, 114], "alibi": [77, 108], "alloc": 88, "allreduc": 114, "alltoal": 12, "also": 163, "altern": 18, "an": [11, 34, 111], "analysi": [8, 10, 39, 126], "announc": 155, "api": [27, 28, 29, 30, 31, 32, 34, 62, 80, 100, 106, 110, 117, 122, 128, 147, 149, 150, 154, 155, 157, 161], "appli": 10, "approach": 20, "arbitrari": 106, "architect": 153, "architectur": [11, 13, 34, 38, 94, 96, 98, 153, 157], "argument": [23, 34, 80, 161], "art": 153, "artifact": 18, "asynchron": 56, "asyncio": 45, "aten": 99, "attempt": 20, "attent": [13, 14, 15, 68, 77, 84, 88, 89, 93, 94, 108, 119, 133, 134, 135, 137, 158, 166], "attention_backend": [26, 28, 30, 31], "attentionbackend": [77, 158], "attentionmetadata": [77, 158], "auto": [19, 99], "auto_deploi": [82, 164], "autodeploi": [83, 165], "autodeploy_config": [78, 159], "autoregress": 13, "auxiliari": 20, "avoid": [20, 35, 128], "awq": [3, 119, 144], "b200": [2, 9, 13], "backend": [13, 17, 21, 28, 29, 62, 63, 64, 77, 84, 86, 99, 143, 145, 148, 152, 156, 158, 166], "background": [8, 10, 11, 13, 14, 19, 90, 91, 99, 168], "balanc": [8, 13, 16, 20, 94], "base": [9, 11, 14, 45], "basecheckpointload": [85, 167], "baseconfigload": [85, 167], "baselin": [8, 130], "baseweightload": [85, 167], "baseweightmapp": [85, 167], "basic": [14, 28, 29, 30, 31, 32, 53, 78, 85, 88, 90, 96, 159, 167, 168], "batch": [77, 93, 106, 108, 133], "beam": [77, 97, 106, 108], "befor": [40, 127, 128], "begin": 128, "behavior": [51, 88, 127], "bench": [22, 39, 63, 78, 90, 91, 98, 126, 128, 159, 168], "benchmark": [2, 7, 26, 28, 29, 30, 31, 32, 40, 41, 78, 127, 128, 159], "best": [2, 7, 34, 35, 94], "beta": 34, "between": 10, "bf16": 144, "bia": 108, "bind": [20, 101, 106, 120], "blackwel": [15, 28, 29, 30, 31, 32, 144], "block": [52, 111], "blockmanag": 111, "blog": 100, "boost": [40, 127], "boundari": [12, 13], "break": 99, "breakdown": 26, "budget": 6, "buffer": [77, 108, 130, 142], "buffermanag": 1, "build": [2, 18, 22, 23, 45, 101, 119, 122, 123, 124, 127, 128, 134], "build_and_run_ad": [80, 161], "built": [80, 85, 102, 161, 167], "c": [16, 101, 106, 109, 142], "cach": [2, 17, 37, 51, 59, 60, 77, 86, 88, 90, 93, 95, 98, 108, 111, 112, 119, 130, 135, 142, 168, 172], "cachecommun": 0, "callback": 10, "can": [112, 149, 153], "cannot": 149, "capabl": 153, "capac": 135, "captur": 10, "case": [19, 21, 133], "cast": 137, "caveat": [40, 127], "challeng": 10, "chang": [34, 51, 52, 117, 133, 155], "chat": [9, 27, 42, 43, 70, 71], "check": 9, "checkpoint": [85, 119, 167], "choos": 7, "chunk": [2, 77, 89, 93, 108, 133, 135], "ci": 35, "class": [34, 106], "classic": 110, "cli": [80, 100, 122, 128, 161], "client": [42, 43, 44, 48, 49, 70, 71, 72, 73, 74, 90, 168], "clock": [2, 40, 127], "clone": 18, "close": [3, 6], "cluster": 86, "cnn_dailymail": 24, "code": [28, 29, 30, 31, 32, 101], "collect": [16, 39, 126], "combin": [2, 10, 87], "come": 7, "command": 41, "commit": 34, "common": [1, 34, 45, 99], "commun": [13, 16, 20, 37, 94, 129, 149], "compat": [10, 91], "compil": [2, 12, 84, 99, 101, 120, 166], "complet": [9, 27, 44, 72, 73, 74], "complex": 19, "compon": [37, 85, 109, 148, 167], "compos": 36, "comprehens": 153, "comput": [10, 11], "concat": 12, "conclus": [8, 130, 133, 134], "config": [23, 85, 119, 167], "configur": [8, 9, 13, 16, 21, 28, 29, 30, 31, 32, 34, 36, 45, 66, 78, 80, 90, 106, 109, 113, 124, 130, 134, 151, 156, 159, 161, 163, 168], "connect": [37, 124], "connector": 59, "consider": 114, "constructor": 34, "consumpt": 117, "contain": [2, 9, 21, 26, 28, 29, 30, 31, 32, 36, 101, 102, 123, 154], "content": [2, 8, 10, 11, 12, 13, 14, 15, 16, 19, 20, 35, 37, 85, 90, 99, 131, 132, 151, 156, 167, 168], "context": [2, 8, 12, 77, 89, 93, 94, 106, 108, 133, 134, 135], "contigu": [77, 93, 108], "continu": 35, "control": [11, 61, 88, 106], "conv": 137, "convers": [19, 118, 122], "cooper": 10, "coordin": [8, 39, 126], "copi": [12, 20], "core": [11, 16, 78, 85, 151, 156, 159, 167], "cot": 11, "coverag": 27, "cp": 94, "cpp": 113, "cpu": [10, 20], "creat": [9, 21, 85, 124, 167], "cross": [77, 88, 108], "cuda": [10, 13, 38, 78, 99, 159], "cuda_graph_config": [26, 28, 29, 30, 31, 32], "cudaev": 1, "cudamemcpyasync": 20, "cudastream": 1, "curl": [42, 43, 44], "current": 99, "custom": [37, 45, 53, 85, 94, 99, 121, 167, 172, 173], "cutlass": 13, "cyclic": [77, 108], "data": [10, 15, 94], "dataset": [2, 8, 16, 17, 19, 22, 40, 41, 127, 128], "datatransceiverst": 0, "datatyp": 88, "deadlock": [10, 20], "debug": [39, 86, 105, 126, 146], "decid": 129, "decod": [9, 10, 14, 19, 23, 54, 69, 88, 97, 98, 106, 116, 142, 157], "decoderst": 1, "decodinginput": 1, "decodingoutput": 1, "decor": 110, "deep": [15, 80, 161], "deepseek": [2, 13, 14, 15, 17, 28, 41, 46], "default": [2, 13, 80, 127, 128, 161], "definit": [35, 120, 151, 156], "dens": [13, 94], "depend": [12, 13], "deploi": 154, "deploy": [28, 29, 30, 31, 32, 33, 83, 100, 165], "deprec": [34, 88], "dequant": 144, "descript": [39, 126], "design": 16, "detail": [34, 113, 144], "dev": 36, "develop": [15, 98, 99, 100, 124, 148], "diagram": 13, "differ": 106, "disabl": [35, 45, 88], "disaggreg": [17, 27, 86, 105, 116], "disaggregated_mpi_work": 27, "disaggserverutil": 0, "distribut": [8, 58], "dive": 15, "do": 153, "docker": [18, 21, 28, 29, 30, 31, 32, 36, 101, 123, 124, 149, 154], "dockerhub": [123, 124], "document": [100, 155], "dora": 113, "dot": [80, 161], "download": [2, 9, 18], "dp": 94, "dq": 144, "draft": [10, 98, 116], "dynamo": [17, 47, 86], "dynasor": 11, "e2": [16, 26, 28, 29, 30, 31, 146], "eagl": [14, 98, 116], "eagle3": [9, 14, 18], "eaglebuff": 1, "eaglemodul": 1, "effect": [16, 19], "embed": [77, 108, 137], "enabl": [2, 19, 39, 88, 94, 107, 112, 123, 126, 130, 134], "encapsul": 37, "end": [12, 20, 26, 28, 29, 30, 31], "endpoint": [27, 28, 29, 30, 31, 32], "engin": [119, 120, 127, 128, 147, 157], "enhanc": 155, "environ": [34, 86, 105], "ep": [15, 16, 94], "eplb": [16, 20, 94], "error": 146, "errorcod": 149, "etp": 13, "eval": 24, "evalu": [14, 16, 28, 29, 30, 31, 119], "event": 111, "everyth": 13, "evolut": 37, "exampl": [11, 24, 39, 47, 50, 53, 66, 75, 78, 79, 91, 106, 113, 119, 120, 121, 126, 127, 149, 159, 160], "except": 142, "exchang": [17, 86], "execut": 146, "executor": [0, 106, 113], "exist": 34, "exp": 12, "expand": 16, "expect": [2, 112], "experi": 8, "experiment": 19, "expert": [12, 13, 15, 16, 20, 28, 80, 94, 107, 161], "explicitdrafttokensbuff": 1, "explor": 2, "extens": 16, "extra": 21, "extra_llm_api_opt": 26, "face": 149, "factor": [77, 108, 119], "fail": 35, "failur": 99, "falcon": 3, "faq": [86, 105, 142], "fast": 35, "faster": 3, "featur": [2, 11, 39, 77, 83, 87, 94, 96, 100, 125, 126, 143, 148, 152, 153, 155, 165], "ffn": 94, "field": 34, "file": [80, 101, 161], "find": 35, "first": [4, 19, 26, 28, 29, 30, 31], "fix": 155, "flag": [134, 144], "flayerinfo": 110, "flight": [77, 93, 106, 108], "flow": [40, 127], "fmha": [12, 77, 108], "format": [2, 85, 90, 113, 167, 168], "formatt": 37, "fp16": [2, 144], "fp32": 144, "fp4": [12, 41, 153], "fp8": [2, 4, 12, 41, 77, 95, 108, 119, 130, 144, 153], "fraction": 135, "framework": 11, "free": 135, "free_gpu_memory_fract": 32, "from": [34, 83, 101, 149, 165], "full": [8, 101], "fulli": 121, "function": [110, 121, 136], "further": 20, "fuse": 12, "fuse_a_gemm": 13, "fusion": [12, 13, 99, 120, 130, 134], "futur": [11, 13, 14, 15, 17, 20, 45], "garbag": [39, 126], "gate": 130, "gb200": [9, 16], "gc": [39, 126], "gemm": [12, 13, 130, 134], "genai": [48, 49], "gener": [16, 45, 54, 55, 56, 57, 58, 61, 65, 77, 86, 97, 99, 105, 108, 149], "get": [2, 9, 78, 83, 100, 159, 165], "gil": [10, 39, 126], "got": 149, "gpqa_diamond": 24, "gpqa_extend": 24, "gpqa_main": 24, "gpt": [9, 21, 29, 109, 113], "gptdecod": 1, "gptdecoderbatch": 1, "gptjsonconfig": 1, "gptq": 144, "gpu": [2, 3, 10, 13, 15, 16, 40, 117, 120, 127, 135, 142, 153], "gqa": 88, "gram": 19, "grammar": 10, "graph": [10, 13, 38, 78, 99, 110, 159], "group": [13, 77, 108], "gsm8k": [16, 24], "guarante": 34, "guid": [10, 28, 29, 30, 31, 32, 33, 34, 54, 94, 97, 98, 99, 100, 106, 131, 148, 151, 154, 156], "h": [0, 1], "h100": [4, 5, 21], "h200": [2, 3, 5, 6, 21], "ha": 4, "hang": 149, "hardwar": [8, 28, 29, 30, 31, 32, 41, 95, 145], "hbm": 5, "head": [12, 77, 108], "header": 101, "health": 9, "heurist": 19, "hierarch": 34, "hierarchi": 111, "high": [16, 21, 110], "highlight": [19, 20], "hopper": [28, 30, 31, 32, 144], "host": [10, 16, 20, 88, 112, 149], "how": [2, 11, 13, 14, 15, 18, 51, 52, 88, 92, 94, 107, 112, 127, 129, 133, 169], "hub": 149, "hug": 149, "huggingfac": [85, 167], "i": [4, 88, 129, 142], "ibuff": 1, "id": 113, "ifb": 93, "igptdecoderbatch": 1, "imag": [18, 36, 101, 102, 123, 124], "implement": [8, 11, 13, 14, 16, 20, 34, 77, 118, 158], "import": 108, "improv": 116, "incorpor": [82, 164], "increas": 6, "indic": 100, "infer": [7, 11, 14, 16, 17, 21, 27, 106, 142, 154], "inform": [39, 110, 126], "infrastructur": 155, "initi": 20, "inplac": 99, "input": [77, 108, 149], "instal": [2, 103, 104, 146], "int4": [3, 144], "int8": [77, 108, 144], "integr": [10, 35, 84, 166], "inter": [20, 26, 28, 29, 30, 31], "interfac": [16, 172], "intern": 109, "interv": 20, "introduct": [11, 15, 16, 26, 28, 29, 30, 31, 32, 37, 50, 149, 151, 156, 172, 173], "invok": 149, "ipcnvlsmemori": 1, "ipcutil": 1, "ir": 99, "isl": [2, 17], "issu": [2, 20, 99, 142, 148, 149, 155], "itensor": 1, "iter": [39, 126], "itl": [26, 28, 29, 30, 31], "jenkin": 35, "json": 74, "json_mode_ev": 24, "k": [12, 93], "k8": 47, "kei": [13, 26, 28, 29, 30, 31, 37, 83, 94, 121, 124, 125, 129, 148, 152, 153, 155, 165], "kernel": [6, 10, 12, 13, 16, 20, 21, 94], "knob": 34, "knowledg": [99, 131, 132], "known": [99, 101, 142, 148, 155], "kv": [2, 17, 37, 51, 59, 60, 77, 86, 88, 93, 95, 98, 108, 111, 112, 119, 130, 135, 142, 172], "kv_cache_config": [26, 28, 30, 31, 32], "kv_cache_free_gpu_memory_fract": [28, 29, 30, 31], "kvcacheeventmanag": 111, "kvcachemanag": 157, "larg": 16, "latenc": [2, 6, 8, 13, 21, 22, 26, 28, 29, 30, 31, 127, 128, 130], "latest": [5, 153], "launch": [9, 12, 13, 18, 21, 26, 28, 29, 30, 31, 32, 39, 86, 126, 154], "layer": [13, 15, 137], "layernorm": 119, "layout": [17, 86, 121], "level": [13, 16, 37, 81, 94, 110, 157, 162], "librari": 37, "licens": [30, 31], "light": 8, "limit": [40, 88, 101, 116, 127, 155, 163], "linear": 137, "link": [95, 100, 101], "linux": [101, 104], "list": 11, "llama": [3, 6, 41, 120, 127, 130, 134], "llama2": 5, "llama3": 30, "llama4": [18, 31], "llm": [2, 4, 5, 7, 9, 10, 11, 12, 14, 16, 17, 18, 19, 20, 21, 28, 29, 30, 31, 32, 34, 40, 45, 50, 53, 58, 62, 80, 83, 91, 94, 95, 99, 100, 101, 107, 116, 119, 121, 122, 123, 124, 127, 128, 133, 142, 145, 147, 149, 153, 154, 155, 161, 165], "llmarg": 34, "lm": 12, "load": [8, 16, 20, 85, 94, 121, 151, 156, 167], "loader": [85, 121, 167], "local": [12, 149], "log": [18, 81, 162], "logic": [16, 37], "logit": [23, 61, 97, 106], "logprob": 76, "long": 89, "longbench_v2": 24, "lookahead": 116, "lookaheadbuff": 1, "lookaheadmodul": 1, "lora": [23, 40, 65, 73, 90, 113, 127, 168], "loracach": [1, 113], "loracachepagemanagerconfig": 1, "loramodul": 1, "low": [12, 21, 37, 90, 114, 127, 130, 168], "lower": 12, "machin": [16, 17], "major": 11, "make": [10, 119], "manag": [20, 34, 40, 90, 110, 111, 127, 168, 172], "map": [113, 127], "mapper": [85, 167], "mark": 106, "marker": [39, 126], "mask": 10, "match": 120, "mathemat": 8, "matrix": [28, 29, 83, 84, 87, 91, 95, 143, 144, 145, 152, 165, 166], "maverick": [18, 41], "max": [2, 21, 127, 133, 135], "max_batch_s": [28, 29, 30, 31, 32, 93], "max_num_token": [28, 29, 30, 31, 32, 93], "max_seq_len": [28, 29, 30, 31, 32, 93], "maximum": 135, "measur": [17, 41], "mechan": 8, "medusa": [116, 127], "medusamodul": 1, "memori": [2, 5, 20, 88, 112, 117, 135, 142], "memorycount": 1, "merg": [35, 80, 161], "method": [7, 34, 110], "methodologi": [17, 26], "metric": [26, 27, 28, 29, 30, 31], "migrat": 20, "min": 2, "miscellan": 16, "mix": 13, "mixtur": [94, 107], "mla": [2, 15], "mlp": [119, 130, 137], "mlperf": 4, "mm_embedding_serv": 27, "mmlu": 24, "mmmu": 24, "modal": [27, 40, 127, 145], "mode": [40, 127], "model": [2, 8, 9, 10, 13, 14, 28, 29, 30, 31, 32, 33, 40, 41, 83, 84, 91, 94, 95, 98, 100, 109, 116, 118, 120, 121, 127, 129, 130, 134, 138, 145, 146, 149, 151, 152, 153, 155, 156, 157, 165, 166], "modelconfig": 1, "modelopt": 95, "modif": 37, "modifi": 34, "modul": [14, 15, 94, 110, 113], "moe": [15, 20, 21, 28, 29, 94, 107], "moe_backend": 13, "moe_config": [26, 28, 29, 30, 31, 32], "moe_expert_parallel_s": [28, 29, 30, 31, 32], "monitor": 18, "more": [2, 6, 12, 39, 126], "motiv": [8, 10, 11, 16, 17, 19, 86, 94, 96], "mount": 36, "mpi_abort": 149, "mpi_comm_world": 149, "mqa": [15, 88], "mtp": [12, 13, 14, 20, 98], "much": 88, "multi": [13, 17, 19, 20, 27, 40, 77, 86, 90, 99, 108, 120, 127, 145, 149, 168], "multimod": [26, 27, 43, 49, 71, 91, 143, 152], "multipl": [65, 134], "multithread": 20, "mutex": 10, "n": 19, "name": [23, 34, 35, 121], "nativ": [12, 121], "nearli": 5, "nemo": [90, 168], "net": 149, "network": [12, 40, 127], "new": [6, 34, 77, 118, 151, 156, 158], "next": [7, 32, 154], "ngc": [18, 26, 102], "ngram": [98, 116], "node": [27, 120, 149], "non": [34, 127], "norm": [130, 134], "normal": 137, "notat": [80, 161], "note": [106, 108, 155], "nsight": [39, 126], "num": 133, "numa": 20, "numer": 144, "nvfp4": 144, "nvidia": [13, 15, 39, 126], "nvtx": [39, 126], "o": 142, "observ": 16, "obtain": 106, "off": 8, "offlin": [16, 95, 154], "offload": [60, 88, 112], "one": 16, "onli": [13, 21, 39, 99, 101, 126, 144, 149], "onlin": [16, 20, 75, 154], "op": 99, "openai": [21, 70, 71, 72, 73, 74, 91], "oper": 99, "optim": [8, 12, 13, 14, 15, 17, 20, 38, 77, 78, 86, 91, 99, 108, 134, 153, 159], "option": [2, 18, 21, 28, 29, 30, 31, 32, 76, 78, 101, 130, 134, 135, 159], "orchestr": 96, "osl": [2, 17], "oss": [9, 21, 29], "other": 127, "out": [2, 151, 156], "outlook": 37, "output": [26, 28, 29, 30, 31, 76, 106, 127], "over": [3, 16, 34], "overhead": 20, "overlap": [12, 17, 38, 86, 92, 169], "overload": 37, "overrid": 36, "overview": [12, 34, 35, 38, 41, 85, 94, 109, 119, 121, 122, 153, 167], "ovewiew": 157, "own": [82, 164, 173], "p": 112, "pack": [77, 108], "pad": [99, 108], "page": [20, 77, 93, 108, 111, 133, 134, 135], "parallel": [12, 13, 15, 16, 20, 28, 94, 107, 113, 127, 129, 134], "paramet": 109, "pareto": 8, "parser": 46, "part": [12, 16, 20, 118], "partial": 88, "pattern": [110, 120], "pdl": 12, "per": [26, 28, 29, 30, 31], "perf": [48, 49], "perform": [2, 4, 7, 8, 10, 12, 13, 16, 17, 18, 19, 20, 21, 28, 29, 30, 31, 32, 39, 78, 99, 112, 114, 116, 126, 128, 130, 131, 134, 153, 159], "persist": [40, 127], "phase": [12, 77, 108], "piecewis": 99, "pip": 104, "pipelin": [35, 94, 129, 134], "pitfal": 128, "plugin": [23, 120, 130, 134, 139], "pod": 124, "polici": [88, 135], "pool": [111, 137, 142], "popular": 33, "posit": [77, 108], "post": [35, 106], "postprocess": 121, "power": [40, 127], "pp": 94, "practic": [7, 34, 35, 94], "pre": [95, 102], "preced": [80, 161], "precis": [12, 13, 15, 84, 114, 144, 166], "predict": 20, "prefer": 34, "prefil": 93, "prepar": [2, 18, 40, 41, 119, 124, 127, 128], "prepare_dataset": 22, "prerequisit": [2, 9, 18, 21, 28, 29, 30, 31, 32, 101, 131, 132, 151, 156], "prevent": [20, 112], "principl": 34, "prioriti": 52, "probe": 11, "process": 20, "processor": [61, 97, 106], "product": 153, "profil": [13, 39, 126, 134], "programmat": [12, 13], "promot": 34, "prompt_logprob": 76, "prompttuningparam": 1, "properti": 88, "propos": 10, "prototyp": [83, 96, 105, 148], "provid": [6, 98], "push": [12, 13, 18], "py": [22, 35, 80, 161], "pyexecutor": 157, "python": [10, 16, 101, 106, 142], "pytorch": [12, 39, 40, 62, 63, 64, 83, 126, 127, 143, 145, 148, 152, 153, 156, 165], "q": [12, 144], "qkv": [77, 108], "quantiz": [7, 12, 40, 45, 90, 95, 119, 122, 127, 130, 140, 144, 168, 170], "quantmod": 144, "queri": [77, 108], "quick": [9, 33, 91, 95, 98, 100, 148, 149, 154, 163], "quickstart": 127, "quit": 149, "qwen": 17, "qwen3": [32, 41], "r1": [2, 13, 14, 15, 17, 28, 41, 46], "rab": 108, "race": 10, "rai": 96, "rank": [90, 119, 149, 168], "ratio": 8, "rawengin": 1, "re": [13, 99], "reason": 46, "receiv": 37, "recip": 33, "recommend": [28, 29, 30, 31, 32, 130, 134, 142], "recompil": 99, "record_signatur": 110, "redraft": 116, "reduc": [20, 117, 130, 134], "reduct": 12, "refer": [11, 78, 92, 94, 100, 118, 150, 159, 169], "regist": 118, "registr": [151, 156], "registri": 18, "reject": 10, "rel": 108, "relat": 110, "relax": [13, 14], "releas": [18, 102, 155], "replic": 94, "repositori": 18, "reproduc": [2, 13, 15, 16, 17, 41], "request": [9, 18, 21, 26, 88, 93, 106], "requir": [110, 114], "resourcemanag": 157, "respons": 106, "result": [2, 8, 39, 41, 106, 126, 128], "retent": 88, "rethink": 12, "retriev": 110, "return_context_logit": 76, "return_generation_logit": 76, "reus": [88, 112], "revisit": [93, 133], "rewind": 98, "rewrit": 110, "right": 7, "roadmap": [83, 96, 165], "robin": 8, "roll": [77, 108], "rope": [77, 108], "rotari": [77, 108], "round": 8, "router": 13, "routergemm": 13, "run": [2, 9, 14, 16, 21, 26, 28, 29, 30, 31, 32, 34, 39, 40, 41, 62, 63, 64, 79, 90, 91, 95, 113, 117, 126, 127, 128, 149, 154, 160, 168], "runpod": 124, "runtim": [1, 10, 15, 37, 38, 45, 66, 84, 101, 109, 113, 120, 135, 141, 142, 166], "runtimedefault": 1, "runtimeerror": 149, "salt": 88, "same": 6, "sampl": [9, 18, 21, 45, 67, 97, 98, 109, 116, 171], "samplingconfig": 1, "save": 128, "scaffold": 11, "scaffoldingllm": 11, "scale": [12, 16, 20, 77, 108, 119], "scatter": 134, "schedul": [8, 38, 92, 93, 133, 135, 157, 169, 173], "schema": [34, 74], "scout": [31, 41], "script": [79, 160], "seamless": [83, 165], "seamlessli": 10, "search": [77, 97, 108], "sec": 5, "second": [26, 28, 29, 30, 31], "secur": 88, "see": 163, "select": [21, 36], "send": 106, "sender": 37, "sequenc": 89, "serial": 0, "serv": [17, 21, 25, 26, 27, 39, 64, 75, 86, 90, 91, 98, 116, 126, 154, 163, 168], "server": [9, 17, 18, 21, 27, 28, 29, 30, 31, 32, 86, 90, 91, 106, 168], "servic": [26, 105], "session": 149, "set": [8, 26, 28, 29, 30, 31, 32, 40, 78, 127, 129, 159], "setup": 19, "sever": 12, "shard": 129, "share": 12, "shoot": 121, "show": [28, 29, 30, 31, 32], "showcas": 67, "singl": [3, 90, 149, 168], "situat": 112, "size": [88, 133, 135, 142], "slide": [77, 89, 108], "slurm": [27, 53, 62, 63, 64, 86, 149], "smart": 13, "smoothquant": 144, "softwar": 145, "sol": 8, "sota": 7, "sourc": 101, "spars": [12, 13, 68], "specif": [33, 39, 126], "specul": [9, 10, 14, 19, 23, 69, 88, 98, 116], "speculativedecodingmod": 1, "speculativedecodingmodul": 1, "speed": [7, 8, 11, 19], "speedup": 14, "ssh": [123, 124], "stabil": 34, "stage": 35, "start": [9, 18, 26, 27, 33, 78, 83, 90, 91, 98, 100, 148, 149, 154, 159, 163, 165, 168], "state": [10, 153], "statist": 16, "statu": [20, 99], "step": [2, 16, 17, 18, 28, 29, 30, 31, 32, 85, 101, 118, 151, 154, 156, 167], "stop": 18, "strategi": [8, 13, 15, 94, 129], "stream": [13, 20, 57, 99, 117], "streamingllm": [77, 108], "structur": [12, 106], "studi": [14, 16, 17, 19, 133], "style": 45, "subcommand": 127, "summari": [8, 127, 130, 134], "support": [12, 14, 17, 20, 28, 29, 40, 83, 84, 86, 90, 91, 95, 101, 120, 121, 127, 143, 144, 145, 152, 153, 165, 166, 168], "swiglu": 130, "syntax": [24, 27], "synthet": 17, "system": [13, 39, 88, 126], "tabl": [2, 8, 10, 11, 12, 13, 14, 15, 16, 19, 20, 35, 37, 85, 90, 99, 100, 131, 132, 151, 156, 167, 168], "tag": [18, 102], "target": [10, 98, 116], "technic": 144, "techniqu": [7, 11, 67], "templat": 124, "tensor": [0, 12, 77, 94, 106, 107, 108, 110, 113, 129, 142], "tensor_parallel_s": [28, 29, 30, 31, 32], "tensorrt": [2, 4, 5, 7, 9, 10, 11, 12, 14, 16, 17, 18, 19, 20, 21, 28, 29, 30, 31, 32, 40, 83, 91, 94, 95, 99, 100, 101, 107, 116, 119, 120, 121, 122, 123, 124, 127, 128, 133, 142, 145, 147, 153, 155, 165], "test": [18, 21, 28, 29, 30, 31, 32, 34, 35, 146], "test_to_stage_map": 35, "text": [54, 55, 56, 57, 61, 65], "theoret": 8, "think": 129, "thought": 16, "thrash": 20, "throughput": [2, 6, 8, 15, 21, 22, 26, 28, 29, 30, 31, 40, 41, 127, 128], "time": [11, 26, 28, 29, 30, 31, 134, 142], "tip": [18, 21, 28, 29, 30, 31, 32, 78, 99, 128, 146, 149, 159, 163], "tlb": 20, "tllmlogger": 1, "tok": 4, "token": [4, 5, 8, 10, 20, 26, 28, 29, 30, 31, 45, 133, 135], "token_norm_dist": 22, "token_unif_dist": 22, "tool": 122, "top": 157, "topic": 101, "topologi": 114, "torch": [12, 99], "torchllmarg": 34, "total": [26, 28, 29, 30, 31], "tp": [8, 26, 28, 29, 30, 31, 94], "tpot": [26, 28, 29, 30, 31], "trace": [10, 99], "trade": 8, "tradeoff": [92, 169], "transceiv": 37, "transfer": 37, "transferag": 0, "transform": [17, 86], "translat": [16, 17, 19, 121], "transmiss": 37, "tree": [14, 116, 151, 156], "trigger": [35, 111], "triton": [17, 21, 106], "troubl": 121, "troubleshoot": [10, 18, 21, 28, 29, 30, 31, 32, 86, 105, 128, 146, 149], "trt": 7, "trtllm": [13, 17, 22, 23, 24, 25, 26, 27, 39, 63, 64, 78, 86, 90, 91, 98, 126, 128, 154, 159, 163, 168], "trtllmattent": 77, "trust_remote_cod": [28, 29, 30, 31, 32], "ttft": [26, 28, 29, 30, 31], "tune": [2, 18, 99, 112, 131, 133], "turn": 19, "two": [10, 98], "type": [0, 34, 111], "understand": [133, 142], "unit": [35, 146], "unnecessari": 35, "up": [3, 6, 7, 11, 19, 26], "updat": [20, 34, 155], "upload": [123, 124], "upper": 37, "us": [12, 18, 21, 35, 36, 61, 85, 110, 113, 116, 134, 135, 142, 149, 167], "usag": [24, 35, 78, 83, 85, 86, 90, 92, 95, 96, 97, 98, 99, 114, 142, 159, 165, 167, 168, 169], "user": [98, 130], "v": [5, 8, 12, 107], "valid": [34, 40, 127], "vanilla": 14, "variabl": [34, 41, 86, 105], "verif": [13, 98], "verifi": [28, 29, 30, 31, 118], "via": [91, 104, 128, 163], "virtualmemori": 1, "visual": [39, 126], "volum": 36, "vote": 11, "w4a16": 144, "w8a16": 144, "w8a8": 144, "wa": 149, "wai": 18, "wait": 8, "waiv": 35, "weekli": 18, "weight": [15, 20, 85, 117, 118, 119, 120, 121, 142, 144, 151, 156, 167], "welcom": 100, "what": [4, 7, 111, 153], "when": [13, 85, 110, 149, 167], "wide": [28, 94], "width": 106, "window": [77, 88, 89, 108, 135], "windowblockmanag": 111, "wip": 2, "within": 6, "without": 101, "wo": 12, "work": [11, 13, 14, 15, 17, 20, 92, 127, 169], "worker": 11, "workflow": [34, 37, 39, 40, 82, 110, 121, 122, 126, 127, 164], "workload": 13, "world": 109, "worldconfig": 1, "write": 118, "xqa": [6, 77, 108], "yaml": [28, 29, 30, 31, 32, 78, 80, 90, 159, 161, 163, 168], "you": [128, 153], "your": [18, 82, 164, 173]}})
\ No newline at end of file
+Search.setIndex({"alltitles": {"(H200 Only) Using OpenAI Triton Kernels for MoE": [[21, "h200-only-using-openai-triton-kernels-for-moe"]], "(H200/H100 Only) Using OpenAI Triton Kernels for MoE": [[21, "h200-h100-only-using-openai-triton-kernels-for-moe"]], "1. Add the field to TorchLlmArgs": [[34, "add-the-field-to-torchllmargs"]], "1. Balance Ratio": [[8, "balance-ratio"]], "1. Committed APIs": [[34, "committed-apis"]], "1. Download TensorRT LLM": [[2, "download-tensorrt-llm"]], "1. Expert Replication and Load Balancing": [[94, "expert-replication-and-load-balancing"]], "1. Implement the method in _TorchLLM": [[34, "implement-the-method-in-torchllm"]], "1. Initial Approach for Weight Updating - cudaMemcpyAsync": [[20, "initial-approach-for-weight-updating-cudamemcpyasync"]], "1. Knob Naming": [[34, "knob-naming"]], "1. Using a Model from the Hugging Face Hub": [[149, "using-a-model-from-the-hugging-face-hub"]], "1. Weights size": [[142, "weights-size"]], "2. Activation size": [[142, "activation-size"]], "2. Avoiding Deadlock - Multithreaded CPU Copy with Managed Memory": [[20, "avoiding-deadlock-multithreaded-cpu-copy-with-managed-memory"]], "2. Custom EP Communication Kernels": [[94, "custom-ep-communication-kernels"]], "2. Download the DeepSeek R1 models": [[2, "download-the-deepseek-r1-models"]], "2. Hierarchical Configuration": [[34, "hierarchical-configuration"]], "2. Non-committed APIs": [[34, "non-committed-apis"]], "2. Speed-of-Light Throughput (SOL TPS)": [[8, "speed-of-light-throughput-sol-tps"]], "2. Update the API schema": [[34, "update-the-api-schema"], [34, "id1"]], "2. Using a Local Hugging Face Model": [[149, "using-a-local-hugging-face-model"]], "3. Build and run TensorRT LLM container": [[2, "build-and-run-tensorrt-llm-container"]], "3. Expert Parallelism Load Balancer (EPLB)": [[94, "expert-parallelism-load-balancer-eplb"]], "3. I/O tensors": [[142, "i-o-tensors"]], "3. NUMA Memory to Prevent Page Migration": [[20, "numa-memory-to-prevent-page-migration"]], "3. Prefer LlmArgs Over Environment Variables": [[34, "prefer-llmargs-over-environment-variables"]], "3. Run validation tests": [[34, "run-validation-tests"]], "3.1 Runtime and decoder buffers except KV cache tensor": [[142, "runtime-and-decoder-buffers-except-kv-cache-tensor"]], "3.2 KV cache tensor": [[142, "kv-cache-tensor"]], "4. Addressing the TLB Thrashing Issue": [[20, "addressing-the-tlb-thrashing-issue"]], "4. Compile and Install TensorRT LLM": [[2, "compile-and-install-tensorrt-llm"]], "5. Optional: Tune GPU clocks": [[2, "optional-tune-gpu-clocks"]], "6. Dataset preparation": [[2, "dataset-preparation"]], "@record_signature to Decorate Functionals Requiring FLayerInfo": [[110, "record-signature-to-decorate-functionals-requiring-flayerinfo"]], "ADP Balance Strategy": [[8, null]], "ADP Balance Strategy: Coordinated Waiting Mechanism": [[8, "adp-balance-strategy-coordinated-waiting-mechanism"]], "ADP Balance with Context Wait Implementation": [[8, "adp-balance-with-context-wait-implementation"]], "ADP Balance with Full Strategy Implementation": [[8, "adp-balance-with-full-strategy-implementation"]], "ALiBi": [[77, "alibi"], [108, "alibi"]], "API": [[106, "api"]], "API Change Principles": [[34, "api-change-principles"]], "API Changes": [[117, "api-changes"], [155, "api-changes"], [155, "id4"], [155, "id9"], [155, "id13"], [155, "id24"], [155, "id29"], [155, "id34"], [155, "id39"], [155, "id46"], [155, "id51"], [155, "id57"], [155, "id63"], [155, "id69"]], "API Reference": [[100, null], [150, null]], "API Schema Management": [[34, "api-schema-management"]], "API Types and Stability Guarantees": [[34, "api-types-and-stability-guarantees"]], "ATen IR Optimization": [[99, "aten-ir-optimization"]], "AWQ Quantization Scaling Factors": [[119, "awq-quantization-scaling-factors"]], "About": [[24, "about"], [27, "about"]], "About Speculative Sampling": [[116, "about-speculative-sampling"]], "About TensorRT LLM": [[153, "about-tensorrt-llm"]], "About extra_llm_api_options": [[26, "about-extra-llm-api-options"]], "Access & Licensing": [[30, "access-licensing"], [31, "access-licensing"]], "Accuracy": [[7, "accuracy"]], "Accuracy studies for Relaxed Acceptance": [[14, "accuracy-studies-for-relaxed-acceptance"]], "Achieving speedup with MTP speculative decoding": [[14, "achieving-speedup-with-mtp-speculative-decoding"]], "Acknowledgement": [[8, "acknowledgement"], [16, "acknowledgement"], [17, "acknowledgement"]], "Acknowledgements": [[10, "acknowledgements"], [11, "acknowledgements"], [12, "acknowledgements"], [20, "acknowledgements"]], "Acknowledgment": [[13, "acknowledgment"], [14, "acknowledgment"], [15, "acknowledgment"]], "Activation": [[137, "module-tensorrt_llm.layers.activation"]], "Adding a Model": [[118, null]], "Adding a New Argument": [[34, "adding-a-new-argument"]], "Adding a New Method": [[34, "adding-a-new-method"]], "Adding a New Model": [[151, null]], "Adding a New Model in PyTorch Backend": [[156, null]], "Additional Outputs": [[76, null]], "Advanced Configuration": [[78, "advanced-configuration"], [159, "advanced-configuration"]], "Advanced Usage": [[83, "advanced-usage"], [90, "advanced-usage"], [165, "advanced-usage"], [168, "advanced-usage"]], "Advanced topics": [[101, "advanced-topics"]], "Algorithm": [[114, "algorithm"]], "Algorithm & Complexity": [[19, "algorithm-complexity"]], "An Example: Implement Dynasor-CoT on Scaffolding": [[11, "an-example-implement-dynasor-cot-on-scaffolding"]], "Announcements": [[155, "announcements"], [155, "id67"]], "Architecture": [[11, "architecture"], [34, "architecture"], [96, "architecture"]], "Architecture Overview": [[38, null], [94, "architecture-overview"]], "Architecture Ovewiew": [[157, null]], "Asyncio-Based Generation": [[45, "asyncio-based-generation"]], "Attempts at Online EPLB Implementation": [[20, "attempts-at-online-eplb-implementation"]], "Attention": [[137, "module-tensorrt_llm.layers.attention"], [158, null]], "Attention Backends": [[77, "attention-backends"], [158, "attention-backends"]], "Attention Kernel": [[13, "attention-kernel"]], "Attention Module": [[94, "attention-module"]], "Attention Weights": [[119, "attention-weights"]], "Attention Window Size": [[88, "attention-window-size"]], "Attention backends": [[84, "attention-backends"], [166, "attention-backends"]], "Attention for MTP": [[14, "attention-for-mtp"]], "Auto Multi-stream": [[99, "auto-multi-stream"]], "AutoDeploy": [[165, null]], "AutoDeploy (Prototype)": [[83, null]], "Autoregressive MTP Layers": [[13, "autoregressive-mtp-layers"]], "Auto\u2011Enablement with Heuristic": [[19, "autoenablement-with-heuristic"]], "Avoiding unnecessary --disable-fail-fast usage": [[35, "avoiding-unnecessary-disable-fail-fast-usage"]], "B200 max-throughput for R1 with FP16 KV cache": [[2, "b200-max-throughput-for-r1-with-fp16-kv-cache"]], "B200 max-throughput for R1-0528 with FP8 KV cache": [[2, "b200-max-throughput-for-r1-0528-with-fp8-kv-cache"]], "B200 min-latency": [[2, "b200-min-latency"]], "Background": [[13, "background"], [14, "background"], [90, "background"], [91, "background"], [168, "background"]], "Background & Motivation": [[19, "background-motivation"]], "Background Knowledge": [[99, "background-knowledge"]], "Background and Challenges": [[10, "background-and-challenges"]], "Background and Motivation": [[11, "background-and-motivation"]], "BaseCheckpointLoader": [[85, "basecheckpointloader"], [167, "basecheckpointloader"]], "BaseConfigLoader": [[85, "baseconfigloader"], [167, "baseconfigloader"]], "BaseWeightLoader": [[85, "baseweightloader"], [167, "baseweightloader"]], "BaseWeightMapper": [[85, "baseweightmapper"], [167, "baseweightmapper"]], "Baseline Performance": [[8, "baseline-performance"]], "Baseline: Round-Robin Token Distribution": [[8, "baseline-round-robin-token-distribution"]], "Basic Implementation": [[14, "basic-implementation"]], "Basic Performance Configuration (autodeploy_config.yaml)": [[78, "basic-performance-configuration-autodeploy-config-yaml"], [159, "basic-performance-configuration-autodeploy-config-yaml"]], "Basic Test": [[28, "basic-test"], [29, "basic-test"], [30, "basic-test"], [31, "basic-test"], [32, "basic-test"]], "Basic Usage": [[78, "basic-usage"], [85, "basic-usage"], [90, "basic-usage"], [96, "basic-usage"], [159, "basic-usage"], [167, "basic-usage"], [168, "basic-usage"]], "Basics": [[53, "basics"]], "Beam search": [[97, "beam-search"]], "Beam-Search": [[77, "beam-search"], [108, "beam-search"]], "Before Benchmarking": [[40, "before-benchmarking"], [127, "before-benchmarking"]], "Before You Begin: TensorRT-LLM LLM-API": [[128, "before-you-begin-tensorrt-llm-llm-api"]], "Benchmark": [[2, "benchmark"], [2, "id1"], [7, "benchmark"]], "Benchmarking Default Performance": [[128, null]], "Benchmarking Performance": [[28, "benchmarking-performance"], [29, "benchmarking-performance"], [30, "benchmarking-performance"], [31, "benchmarking-performance"], [32, "benchmarking-performance"]], "Benchmarking a non-Medusa Low Latency Engine": [[127, "benchmarking-a-non-medusa-low-latency-engine"]], "Benchmarking with LoRA Adapters in PyTorch workflow": [[40, "benchmarking-with-lora-adapters-in-pytorch-workflow"], [127, "benchmarking-with-lora-adapters-in-pytorch-workflow"]], "Benchmarking with trtllm-bench": [[78, null], [128, "benchmarking-with-trtllm-bench"], [159, null]], "Best Practices": [[34, "best-practices"], [94, "best-practices"]], "Best practices to choose the right quantization methods": [[7, "best-practices-to-choose-the-right-quantization-methods"]], "Block": [[111, "block"]], "Blogs": [[100, null]], "Boost settings": [[40, "boost-settings"], [127, "boost-settings"]], "Build APIs": [[122, "build-apis"]], "Build Checkpoint into TensorRT Engine": [[119, "build-checkpoint-into-tensorrt-engine"]], "Build Configuration": [[45, "build-configuration"]], "Build TensorRT LLM": [[101, "build-tensorrt-llm"]], "Build the TensorRT LLM Docker Image": [[123, null]], "Build the TensorRT LLM Docker Image and Upload to DockerHub": [[123, "build-the-tensorrt-llm-docker-image-and-upload-to-dockerhub"], [124, "build-the-tensorrt-llm-docker-image-and-upload-to-dockerhub"]], "Building a Benchmark Engine": [[127, "building-a-benchmark-engine"]], "Building a Medusa Low-Latency Engine": [[127, "building-a-medusa-low-latency-engine"]], "Building a TensorRT LLM Docker Image": [[101, "building-a-tensorrt-llm-docker-image"]], "Building and Saving Engines via CLI": [[128, "building-and-saving-engines-via-cli"]], "Building and Saving the Engine": [[128, "building-and-saving-the-engine"]], "Building from Source Code on Linux": [[101, null]], "Building the Python Bindings for the C++ Runtime": [[101, "building-the-python-bindings-for-the-c-runtime"]], "Built-in Checkpoint Formats": [[85, "built-in-checkpoint-formats"], [167, "built-in-checkpoint-formats"]], "Built-in Default Configuration": [[80, "built-in-default-configuration"], [161, "built-in-default-configuration"]], "C++ Executor API Example": [[106, "c-executor-api-example"]], "C++ GPT Runtime": [[109, null]], "C++ extension": [[16, "c-extension"]], "C++ runtime": [[142, "c-runtime"], [142, "id1"]], "CI pipelines": [[35, "ci-pipelines"]], "CLI Arguments with Dot Notation": [[80, "cli-arguments-with-dot-notation"], [161, "cli-arguments-with-dot-notation"]], "CLI Reference": [[100, null]], "CLI Tools": [[122, "cli-tools"]], "CUDA Callback": [[10, "cuda-callback"]], "CUDA Graph": [[38, "cuda-graph"]], "CUDA Graph & Programmatic Dependent Launch": [[13, "cuda-graph-programmatic-dependent-launch"]], "CUDA Graph Compatibility: Grammar Computation": [[10, "cuda-graph-compatibility-grammar-computation"]], "CUDA Graph Compatibility: Mask Applying Kernel": [[10, "cuda-graph-compatibility-mask-applying-kernel"]], "CUDA Graph Optimization": [[78, "cuda-graph-optimization"], [159, "cuda-graph-optimization"]], "CUTLASS Backend (default backend)": [[13, "cutlass-backend-default-backend"]], "Cache Layout Transformation": [[17, "cache-layout-transformation"], [86, "cache-layout-transformation"]], "Cache Management": [[90, "cache-management"], [168, "cache-management"]], "Cannot quit after generation": [[149, "cannot-quit-after-generation"]], "Capacity Scheduler Policy": [[135, "capacity-scheduler-policy"]], "Case 1 with Conversation Dataset": [[19, "case-1-with-conversation-dataset"]], "Case 2 with Translation Dataset": [[19, "case-2-with-translation-dataset"]], "Cast": [[137, "module-tensorrt_llm.layers.cast"]], "Chat API": [[27, "chat-api"]], "Checkpoint Loading": [[85, null], [167, null]], "Chunked Context": [[77, "chunked-context"], [89, "chunked-context"], [108, "chunked-context"]], "Chunked Context (a.k.a Chunked Prefill)": [[93, "chunked-context-a-k-a-chunked-prefill"]], "Chunked attention": [[89, "chunked-attention"]], "Classical Workflow": [[110, "classical-workflow"]], "Client Usage": [[90, "client-usage"], [168, "client-usage"]], "Closing": [[3, "closing"], [6, "closing"]], "Collect PyTorch profiler results": [[39, "collect-pytorch-profiler-results"], [126, "collect-pytorch-profiler-results"]], "Combining Guided Decoding and Speculative Decoding: Making CPU and GPU Cooperate Seamlessly": [[10, null]], "Command Overview": [[41, "command-overview"]], "Common Trace Failure": [[99, "common-trace-failure"]], "Common Workflows": [[34, "common-workflows"]], "Communication Kernel": [[13, "communication-kernel"]], "Communication Kernels": [[20, "communication-kernels"]], "Compilation": [[120, "compilation"]], "Compile Backends": [[84, "compile-backends"], [166, "compile-backends"]], "Completions API": [[27, "completions-api"]], "Conclusion": [[8, "conclusion"], [130, "conclusion"], [133, "conclusion"], [134, "conclusion"]], "Config": [[119, "config"]], "Configuration Examples": [[78, "configuration-examples"], [159, "configuration-examples"]], "Configuration Options Reference": [[78, "configuration-options-reference"], [159, "configuration-options-reference"]], "Configuration Precedence and Deep Merging": [[80, "configuration-precedence-and-deep-merging"], [161, "configuration-precedence-and-deep-merging"]], "Configuration via YAML": [[163, "configuration-via-yaml"]], "Configure SSH Key": [[124, "configure-ssh-key"]], "Configure The Executor": [[106, "configure-the-executor"]], "Connect to the Pod": [[124, "connect-to-the-pod"]], "Connection": [[37, "connection"]], "Container image selection": [[36, "container-image-selection"]], "Container image tags": [[102, null]], "Context Chunking Policy": [[135, "context-chunking-policy"]], "Context Parallelism (CP)": [[94, "context-parallelism-cp"]], "Context Phase": [[77, "context-phase"], [108, "context-phase"]], "Context and Generation Phases": [[77, "context-and-generation-phases"], [108, "context-and-generation-phases"]], "Context phase Q/K/V concat optimization": [[12, "context-phase-q-k-v-concat-optimization"]], "Contiguous KV Cache": [[77, "contiguous-kv-cache"], [93, "contiguous-kv-cache"], [108, "contiguous-kv-cache"]], "Continuous Integration Overview": [[35, null]], "Control generated text using logits processor": [[61, null]], "Controller": [[11, "controller"]], "Controlling KV Cache Behavior": [[88, "controlling-kv-cache-behavior"]], "Controlling output with Logits Post-Processor": [[106, "controlling-output-with-logits-post-processor"]], "Conv": [[137, "module-tensorrt_llm.layers.conv"]], "Conversion APIs": [[122, "conversion-apis"]], "Coordinating with NVIDIA Nsight Systems Launch": [[39, "coordinating-with-nvidia-nsight-systems-launch"], [126, "coordinating-with-nvidia-nsight-systems-launch"]], "Coordinating with PyTorch profiler (PyTorch workflow only)": [[39, "coordinating-with-pytorch-profiler-pytorch-workflow-only"], [126, "coordinating-with-pytorch-profiler-pytorch-workflow-only"]], "Core Components": [[85, "core-components"], [167, "core-components"]], "Core Features": [[11, "core-features"]], "Core Models": [[151, "core-models"], [156, "core-models"]], "Core Performance Settings": [[78, "core-performance-settings"], [159, "core-performance-settings"]], "Core implementations of the GPU logic": [[16, "core-implementations-of-the-gpu-logic"]], "Core implementations of the host logic": [[16, "core-implementations-of-the-host-logic"]], "Create a Pod Template": [[124, "create-a-pod-template"]], "Create a Runpod account": [[124, "create-a-runpod-account"]], "Create the Eagle3 Configuration": [[9, "create-the-eagle3-configuration"]], "Creating Custom Checkpoint Loaders": [[85, "creating-custom-checkpoint-loaders"], [167, "creating-custom-checkpoint-loaders"]], "Creating the Extra Options Configuration": [[21, "creating-the-extra-options-configuration"], [21, "id1"]], "Cross Attention": [[77, "cross-attention"], [108, "cross-attention"]], "Curl Chat Client": [[42, null]], "Curl Chat Client For Multimodal": [[43, null]], "Curl Completion Client": [[44, null]], "Current Status": [[99, "current-status"]], "Custom Op": [[99, "custom-op"]], "Customization": [[37, "customization"], [53, "customization"]], "Customize KV Cache Manager": [[172, "customize-kv-cache-manager"]], "Customize Your Own Scheduler": [[173, "customize-your-own-scheduler"]], "Data Parallel for Attention module (ADP)": [[15, "data-parallel-for-attention-module-adp"]], "Data Parallelism (DP)": [[94, "data-parallelism-dp"], [94, "id2"]], "Dataset Configuration": [[8, "dataset-configuration"]], "Datatype": [[88, "datatype"]], "Debug Execution Errors": [[146, "debug-execution-errors"]], "Debug on E2E Models": [[146, "debug-on-e2e-models"]], "Debug on Unit Tests": [[146, "debug-on-unit-tests"]], "Debugging FAQs": [[86, "debugging-faqs"], [105, "debugging-faqs"]], "Deciding Model Sharding Strategy": [[129, null]], "Decoder": [[157, "decoder"]], "DeepSeek R1": [[17, "deepseek-r1"], [41, "deepseek-r1"]], "DeepSeek R1 MTP Implementation and Optimization": [[14, null]], "Deepseek R1 Reasoning Parser": [[46, null]], "Default Build Behavior": [[127, "default-build-behavior"]], "Dense GEMM optimization": [[13, "dense-gemm-optimization"]], "Dense Models": [[94, "dense-models"]], "Deploy Online Serving with trtllm-serve": [[154, "deploy-online-serving-with-trtllm-serve"]], "Deployment Guide": [[100, null]], "Deployment Guide for DeepSeek R1 on TensorRT LLM - Blackwell & Hopper Hardware": [[28, null]], "Deployment Guide for GPT-OSS on TensorRT-LLM - Blackwell Hardware": [[29, null]], "Deployment Guide for Llama3.3 70B on TensorRT LLM - Blackwell & Hopper Hardware": [[30, null]], "Deployment Guide for Llama4 Scout 17B on TensorRT LLM - Blackwell & Hopper Hardware": [[31, null]], "Deployment Guide for Qwen3 Next on TensorRT LLM - Blackwell & Hopper Hardware": [[32, null]], "Deployment Steps": [[28, "deployment-steps"], [29, "deployment-steps"], [30, "deployment-steps"], [31, "deployment-steps"], [32, "deployment-steps"]], "Deprecated Properties": [[88, "deprecated-properties"]], "Deprecating an API": [[34, "deprecating-an-api"]], "Develop TensorRT LLM on Runpod": [[124, null]], "Developer Guide": [[98, "developer-guide"], [100, null], [148, "developer-guide"]], "Development Guide": [[99, "development-guide"]], "Disable Tokenizer": [[45, "disable-tokenizer"]], "Disaggregated Serving": [[86, null], [116, "disaggregated-serving"]], "Disaggregated Serving in TensorRT LLM": [[17, null], [17, "id1"]], "Disaggregated-Service (Prototype)": [[105, null]], "Distributed LLM Generation": [[58, null]], "DoRA": [[113, "dora"]], "Documentation": [[155, "documentation"], [155, "id43"]], "Download Artifacts": [[18, "download-artifacts"]], "Download the models (Base + Eagle3)": [[9, "download-the-models-base-eagle3"]], "Draft Model": [[10, "draft-model"]], "Draft-Target-Model": [[116, "draft-target-model"]], "Draft/Target": [[98, "draft-target"]], "Dynamo": [[17, "dynamo"], [86, "dynamo"]], "Dynamo K8s Example": [[47, null]], "E2E evaluation": [[16, "e2e-evaluation"]], "EAGLE": [[116, "eagle"]], "EAGLE 3": [[98, "eagle-3"]], "EP Load Balancer": [[16, "ep-load-balancer"]], "EP communication kernels": [[16, "ep-communication-kernels"]], "EP communication kernels implementation": [[16, "ep-communication-kernels-implementation"]], "Eagle3 support": [[14, "eagle3-support"]], "Effect of Multi-turn conversation": [[19, "effect-of-multi-turn-conversation"]], "Embedding": [[137, "module-tensorrt_llm.layers.embedding"]], "Enable GIL information in NVTX markers": [[39, "enable-gil-information-in-nvtx-markers"], [126, "enable-gil-information-in-nvtx-markers"]], "Enable Offloading to Host Memory": [[88, "enable-offloading-to-host-memory"]], "Enable garbage collection (GC) NVTX markers": [[39, "enable-garbage-collection-gc-nvtx-markers"], [126, "enable-garbage-collection-gc-nvtx-markers"]], "Enable kv cache reuse for p-tuning": [[112, "enable-kv-cache-reuse-for-p-tuning"]], "Enable more NVTX markers for debugging": [[39, "enable-more-nvtx-markers-for-debugging"], [126, "enable-more-nvtx-markers-for-debugging"]], "Enable ssh access to the container": [[123, "enable-ssh-access-to-the-container"]], "Enable/Disable Cross Request Reuse": [[88, "enable-disable-cross-request-reuse"]], "Enabling GEMM + SwiGLU Fusion": [[130, "enabling-gemm-swiglu-fusion"]], "Enabling GEMM Plugin": [[134, "enabling-gemm-plugin"]], "Enabling Low Latency GEMM plugin": [[130, "enabling-low-latency-gemm-plugin"]], "Enabling Paged Context Attention": [[134, "enabling-paged-context-attention"]], "Enabling Quantization": [[130, "enabling-quantization"]], "Enabling Quantized KV Cache": [[130, "enabling-quantized-kv-cache"]], "Enabling Reduce Norm Fusion Plugin": [[134, "enabling-reduce-norm-fusion-plugin"]], "Enabling Reduce Norm Fusion with User Buffers": [[130, "enabling-reduce-norm-fusion-with-user-buffers"]], "Enabling building with multiple profiles": [[134, "enabling-building-with-multiple-profiles"]], "Encapsulation and Overloading of Low-Level Communication Libraries": [[37, "encapsulation-and-overloading-of-low-level-communication-libraries"]], "End-to-End (E2E) Latency": [[26, "end-to-end-e2e-latency"], [28, "end-to-end-e2e-latency"], [29, "end-to-end-e2e-latency"], [30, "end-to-end-e2e-latency"], [31, "end-to-end-e2e-latency"]], "End-to-End Performance": [[12, "end-to-end-performance"], [20, "end-to-end-performance"]], "Environment Variables": [[86, "environment-variables"], [105, "environment-variables"]], "Evaluation": [[14, "evaluation"]], "Events in KVCacheEventManager": [[111, "events-in-kvcacheeventmanager"]], "Everything in One Diagram": [[13, "everything-in-one-diagram"]], "Evolution Outlook": [[37, "evolution-outlook"]], "Example": [[119, "example"]], "Example LoRA tensors": [[113, "example-lora-tensors"]], "Example Run Script": [[79, null], [160, null]], "Example of Build Subcommand Output:": [[127, "example-of-build-subcommand-output"]], "Examples": [[39, "examples"], [91, "examples"], [120, "examples"], [121, "examples"], [126, "examples"]], "Executor": [[0, null]], "Executor API": [[106, null]], "Expanded thoughts": [[16, "expanded-thoughts"]], "Expected Result Format": [[2, "expected-result-format"], [2, "id2"], [2, "id3"], [2, "id4"]], "Expected Results": [[2, "expected-results"]], "Experimental Setup": [[19, "experimental-setup"]], "Experiments": [[8, "experiments"]], "Expert Configuration of LLM API": [[80, null], [161, null]], "Expert Configuration of build_and_run_ad.py": [[80, "expert-configuration-of-build-and-run-ad-py"], [161, "expert-configuration-of-build-and-run-ad-py"]], "Expert Parallelism (EP)": [[94, "expert-parallelism-ep"]], "Expert Parallelism Load Balancer (EPLB)": [[20, "expert-parallelism-load-balancer-eplb"]], "Expert Parallelism in TensorRT-LLM": [[107, null]], "Expert parallel for MoE (EP)": [[15, "expert-parallel-for-moe-ep"]], "Exploring more ISL/OSL combinations": [[2, "exploring-more-isl-osl-combinations"]], "FAQ": [[142, "faq"]], "FFN Module": [[94, "ffn-module"]], "FLayerInfo for Retrieving High-Level Information for a Functional": [[110, "flayerinfo-for-retrieving-high-level-information-for-a-functional"]], "FP32, FP16 and BF16": [[144, "fp32-fp16-and-bf16"]], "FP4 Models": [[41, "fp4-models"]], "FP4 Support": [[153, "fp4-support"]], "FP8 (Hopper)": [[144, "fp8-hopper"]], "FP8 Context FMHA": [[77, "fp8-context-fmha"], [108, "fp8-context-fmha"]], "FP8 KV Cache": [[95, "fp8-kv-cache"]], "FP8 Models": [[41, "fp8-models"]], "FP8 Quantization": [[130, null]], "FP8 Quantization Scaling Factors": [[119, "fp8-quantization-scaling-factors"]], "FP8 Support": [[153, "fp8-support"]], "FP8 context FMHA support": [[12, "fp8-context-fmha-support"]], "FP8 \u201cBaseline\u201d Performance": [[130, "fp8-baseline-performance"]], "Falcon-180B on a single H200 GPU with INT4 AWQ, and 6.7x faster Llama-70B over A100": [[3, null]], "Falcon-180B on a single H200 with INT4 AWQ": [[3, "falcon-180b-on-a-single-h200-with-int4-awq"]], "Feature Combination Matrix": [[87, null]], "Feature Descriptions": [[39, "feature-descriptions"], [126, "feature-descriptions"]], "Feature List on Scaffolding": [[11, "feature-list-on-scaffolding"]], "Features": [[96, "features"], [100, null], [148, "features"]], "Finding the stage for a test": [[35, "finding-the-stage-for-a-test"]], "Fixed Issues": [[155, "fixed-issues"], [155, "id5"], [155, "id10"], [155, "id14"], [155, "id26"], [155, "id30"], [155, "id36"], [155, "id41"], [155, "id48"], [155, "id53"], [155, "id59"], [155, "id65"], [155, "id71"], [155, "id76"]], "Formatter": [[37, "formatter"]], "Fully customized": [[121, "fully-customized"]], "Functionals": [[136, null]], "Further Performance Optimization": [[20, "further-performance-optimization"]], "Fuse add (sparse exp and shared exp) into local reduction": [[12, "fuse-add-sparse-exp-and-shared-exp-into-local-reduction"]], "Fuse several AlltoAll kernels": [[12, "fuse-several-alltoall-kernels"]], "Fuse_A_GEMM": [[13, "fuse-a-gemm"]], "Future Work": [[11, "future-work"], [17, "future-work"], [20, "future-work"]], "Future Works": [[13, "future-works"], [14, "future-works"], [15, "future-works"]], "Future-Style Generation": [[45, "future-style-generation"]], "GEMM + SwiGLU Fusion in Gated-MLP": [[130, "gemm-swiglu-fusion-in-gated-mlp"]], "GEMM Plugin": [[134, "gemm-plugin"]], "GPTQ and AWQ (W4A16)": [[144, "gptq-and-awq-w4a16"]], "GPU Clock Management": [[40, "gpu-clock-management"], [127, "gpu-clock-management"]], "Genai Perf Client": [[48, null]], "Genai Perf Client For Multimodal": [[49, null]], "General FAQs": [[86, "general-faqs"], [105, "general-faqs"]], "General usage": [[97, "general-usage"]], "Generate text": [[55, null]], "Generate text asynchronously": [[56, null]], "Generate text in streaming": [[57, null]], "Generate text with guided decoding": [[54, null]], "Generate text with multiple LoRA adapters": [[65, null]], "Generation": [[45, "generation"]], "Generation Phase": [[77, "generation-phase"], [108, "generation-phase"]], "Get Started": [[83, "get-started"], [165, "get-started"]], "Get the TensorRT LLM Container (1.1.0rc0)": [[9, "get-the-tensorrt-llm-container-1-1-0rc0"]], "Getting Started": [[78, "getting-started"], [100, null], [159, "getting-started"]], "Graph Break": [[99, "graph-break"]], "Graph Rewriting APIs": [[110, "graph-rewriting-apis"]], "Graph Rewriting Module": [[110, null]], "Grouped GEMM": [[13, "grouped-gemm"]], "Guided Decoding": [[10, "guided-decoding"]], "Guided decoding": [[97, "guided-decoding"]], "H100 has 4.6x A100 Performance in TensorRT LLM, achieving 10,000 tok/s at 100ms to first token": [[4, null]], "H200 achieves nearly 12,000 tokens/sec on Llama2-13B with TensorRT LLM": [[5, null]], "H200 max-throughput": [[2, "h200-max-throughput"]], "H200 min-latency": [[2, "h200-min-latency"]], "H200 vs H100": [[5, "h200-vs-h100"]], "Hang issue on Slurm Node": [[149, "hang-issue-on-slurm-node"]], "Hardware": [[41, "hardware"], [145, "hardware"]], "Hardware Support Matrix": [[95, "hardware-support-matrix"]], "Hardware and Model Configuration": [[8, "hardware-and-model-configuration"]], "Hierarchy: Pool, Block, and Page": [[111, "hierarchy-pool-block-and-page"]], "High-level design introduction": [[16, "high-level-design-introduction"]], "Highlights": [[19, "highlights"]], "Host Overhead Optimization": [[20, "host-overhead-optimization"]], "How It Works": [[92, "how-it-works"], [169, "how-it-works"]], "How Much Memory is Allocated to KV Cache": [[88, "how-much-memory-is-allocated-to-kv-cache"]], "How it speeds up inference": [[11, "how-it-speeds-up-inference"]], "How the Benchmarker Works": [[127, "how-the-benchmarker-works"]], "How to Change Block Priorities": [[52, null]], "How to Change KV Cache Behavior": [[51, null]], "How to Enable": [[107, "how-to-enable"]], "How to Enable Attention Parallelism": [[94, "how-to-enable-attention-parallelism"]], "How to Enable MoE Parallelism": [[94, "how-to-enable-moe-parallelism"]], "How to Think about Model Sharding: Communication is Key": [[129, "how-to-think-about-model-sharding-communication-is-key"]], "How to change Max Batch Size": [[133, "how-to-change-max-batch-size"]], "How to change Max Num Tokens": [[133, "how-to-change-max-num-tokens"]], "How to enable kv cache reuse": [[112, "how-to-enable-kv-cache-reuse"]], "How to get best performance on DeepSeek-R1 in TensorRT LLM": [[2, null]], "How to launch Llama4 Maverick + Eagle3 TensorRT LLM server": [[18, null]], "How to reproduce": [[13, "how-to-reproduce"], [15, "how-to-reproduce"]], "How to run DeepSeek models with MTP": [[14, "how-to-run-deepseek-models-with-mtp"]], "How to run the DeepSeek-R1 model with Relaxed Acceptance": [[14, "how-to-run-the-deepseek-r1-model-with-relaxed-acceptance"]], "How to set Tensor Parallelism and Pipeline Parallelism": [[129, "how-to-set-tensor-parallelism-and-pipeline-parallelism"]], "HuggingFace Format": [[85, "huggingface-format"], [167, "huggingface-format"]], "INT4 and INT8 Weight-Only (W4A16 and W8A16)": [[144, "int4-and-int8-weight-only-w4a16-and-w8a16"]], "INT8 SmoothQuant (W8A8)": [[144, "int8-smoothquant-w8a8"]], "INT8/FP8 KV Caches": [[77, "int8-fp8-kv-caches"], [108, "int8-fp8-kv-caches"]], "ISL 4096 - OSL 1024 (Machine Translation Dataset)": [[17, "isl-4096-osl-1024-machine-translation-dataset"]], "ISL 4400 - OSL 1200 (Machine Translation Dataset)": [[17, "isl-4400-osl-1200-machine-translation-dataset"]], "ISL 8192 - OSL 1024 (Machine Translation Dataset)": [[17, "isl-8192-osl-1024-machine-translation-dataset"]], "ISL 8192 - OSL 256 (Synthetic Dataset)": [[17, "isl-8192-osl-256-synthetic-dataset"]], "Implement AttentionBackend": [[77, "implement-attentionbackend"], [158, "implement-attentionbackend"]], "Implement AttentionMetadata": [[77, "implement-attentionmetadata"], [158, "implement-attentionmetadata"]], "Implement Dynasor-CoT based Majority Voting in Scaffolding": [[11, "implement-dynasor-cot-based-majority-voting-in-scaffolding"]], "Implement Dynasor-CoT in Scaffolding": [[11, "implement-dynasor-cot-in-scaffolding"]], "Implement a New Attention Backend": [[77, "implement-a-new-attention-backend"], [158, "implement-a-new-attention-backend"]], "Implementation Configuration": [[13, "implementation-configuration"]], "Implementation Details": [[34, "implementation-details"]], "Important Note": [[108, "important-note"]], "In-flight Batching": [[77, "in-flight-batching"], [93, "in-flight-batching"], [108, "in-flight-batching"]], "In-flight Batching with the Triton Inference Server": [[106, "in-flight-batching-with-the-triton-inference-server"]], "Incorporating auto_deploy into your own workflow": [[82, null], [164, null]], "Indices and tables": [[100, "indices-and-tables"]], "Inference Endpoints": [[27, "inference-endpoints"]], "Inference Time Compute Implementation in TensorRT LLM": [[11, null]], "Infrastructure Changes": [[155, "infrastructure-changes"], [155, "id3"], [155, "id8"], [155, "id15"], [155, "id19"], [155, "id22"], [155, "id27"], [155, "id31"], [155, "id37"], [155, "id42"], [155, "id49"], [155, "id54"], [155, "id60"]], "Infrastructure changes": [[155, "id66"]], "Input QKV tensor": [[77, "input-qkv-tensor"], [108, "input-qkv-tensor"]], "Installation": [[103, null]], "Installation Errors": [[146, "installation-errors"]], "Installing on Linux via pip": [[104, null]], "Integration to TensorRT LLM Python Runtime": [[10, "integration-to-tensorrt-llm-python-runtime"]], "Interfaces": [[172, "interfaces"]], "Internal Components": [[109, "internal-components"]], "Introduction": [[15, "introduction"], [28, "introduction"], [29, "introduction"], [30, "introduction"], [31, "introduction"], [32, "introduction"], [151, "introduction"], [156, "introduction"]], "Introduction for Dynasor-CoT": [[11, "introduction-for-dynasor-cot"]], "Introduction for Scaffolding: A Framework for inference-time compute": [[11, "introduction-for-scaffolding-a-framework-for-inference-time-compute"]], "Introduction to KV Cache Transmission": [[37, null]], "Jenkins stage names": [[35, "jenkins-stage-names"]], "KV Cache": [[77, "kv-cache"], [93, "kv-cache"], [108, "kv-cache"]], "KV Cache Connector": [[59, null]], "KV Cache Exchange": [[17, "kv-cache-exchange"], [86, "kv-cache-exchange"]], "KV Cache Management: Pools, Blocks, and Events": [[111, null]], "KV Cache Manager": [[172, null]], "KV Cache Manager Introduction": [[172, "kv-cache-manager-introduction"]], "KV Cache Offloading": [[60, null]], "KV Cache Pool Management": [[111, "kv-cache-pool-management"]], "KV Cache Quantization Scaling Factors": [[119, "kv-cache-quantization-scaling-factors"]], "KV Cache Rewind": [[98, "kv-cache-rewind"]], "KV Cache Salting for Secure Reuse": [[88, "kv-cache-salting-for-secure-reuse"]], "KV Cache System": [[88, null]], "KV cache reuse": [[112, null]], "KVCacheManager": [[157, "kvcachemanager"]], "Kernel Level optimizations": [[13, "kernel-level-optimizations"]], "Kernel Optimizations": [[20, "kernel-optimizations"]], "Kernel fusion": [[13, "kernel-fusion"]], "Key Capabilities": [[153, "key-capabilities"]], "Key Components": [[37, "key-components"], [148, "key-components"]], "Key Feature:": [[165, "key-feature"]], "Key Features": [[83, "key-features"], [125, null]], "Key Features and Enhancements": [[155, "key-features-and-enhancements"], [155, "id2"], [155, "id7"], [155, "id12"], [155, "id17"], [155, "id18"], [155, "id20"], [155, "id23"], [155, "id28"], [155, "id33"], [155, "id38"], [155, "id45"], [155, "id50"], [155, "id56"], [155, "id62"], [155, "id68"], [155, "id72"], [155, "id74"]], "Key Features of Wide-EP": [[94, "key-features-of-wide-ep"]], "Key Metrics": [[26, "key-metrics"], [28, "key-metrics"], [29, "key-metrics"], [30, "key-metrics"], [31, "key-metrics"]], "Key Optimizations": [[13, "key-optimizations"]], "Known Issue": [[99, "known-issue"]], "Known Issues": [[142, "known-issues"], [148, "known-issues"], [155, "known-issues"], [155, "id6"], [155, "id11"], [155, "id16"], [155, "id21"], [155, "id25"], [155, "id32"], [155, "id44"], [155, "id55"], [155, "id61"], [155, "id77"]], "Known Limitations": [[101, "known-limitations"]], "LLM API Change Guide": [[34, null]], "LLM API Introduction": [[149, null]], "LLM API Options (YAML Configuration)": [[28, "llm-api-options-yaml-configuration"], [29, "llm-api-options-yaml-configuration"], [30, "llm-api-options-yaml-configuration"], [31, "llm-api-options-yaml-configuration"], [32, "llm-api-options-yaml-configuration"]], "LLM API with TensorRT Engine": [[147, null]], "LLM Common Customizations": [[45, null]], "LLM Examples": [[53, null]], "LLM Examples Introduction": [[50, null]], "LLM Models": [[145, "llm-models"]], "Latest HBM Memory": [[5, "latest-hbm-memory"]], "Launch Docker Container": [[154, "launch-docker-container"]], "Launch the NGC container": [[26, "launch-the-ngc-container"]], "Launch the Server (Eagle3 Speculative Decoding)": [[9, "launch-the-server-eagle3-speculative-decoding"]], "Launch the TensorRT LLM Server": [[28, "launch-the-tensorrt-llm-server"], [29, "launch-the-tensorrt-llm-server"], [30, "launch-the-tensorrt-llm-server"], [31, "launch-the-tensorrt-llm-server"], [32, "launch-the-tensorrt-llm-server"]], "Launch the TensorRT-LLM Server": [[21, "launch-the-tensorrt-llm-server"]], "Launching TensorRT LLM Serve": [[21, "launching-tensorrt-llm-serve"], [21, "id2"]], "Launching disaggregated servers on SLURM clusters": [[86, "launching-disaggregated-servers-on-slurm-clusters"]], "Launching the TensorRT LLM docker container": [[21, "launching-the-tensorrt-llm-docker-container"]], "Launching the server": [[18, "launching-the-server"]], "LayerNorm Weights": [[119, "layernorm-weights"]], "Layers": [[137, null]], "Limitations": [[116, "limitations"], [155, "limitations"]], "Limitations and Caveats": [[40, "limitations-and-caveats"], [127, "limitations-and-caveats"]], "Limitations and tips": [[163, "limitations-and-tips"]], "Limited Attention Window Size": [[88, "limited-attention-window-size"]], "Linear": [[137, "module-tensorrt_llm.layers.linear"]], "Linking with the TensorRT LLM C++ Runtime": [[101, "linking-with-the-tensorrt-llm-c-runtime"]], "Llama 3.1 405B": [[41, "llama-3-1-405b"], [120, "llama-3-1-405b"]], "Llama 3.1 70B": [[120, "llama-3-1-70b"]], "Llama 3.1 8B": [[41, "llama-3-1-8b"]], "Llama 3.3 70B": [[41, "llama-3-3-70b"]], "Llama 4 Maverick": [[41, "llama-4-maverick"]], "Llama 4 Scout": [[41, "llama-4-scout"]], "Llama-70B on H200 up to 2.4x increased throughput with XQA within same latency budget": [[6, "llama-70b-on-h200-up-to-2-4x-increased-throughput-with-xqa-within-same-latency-budget"]], "Llama-70B on H200 up to 6.7x A100": [[3, "llama-70b-on-h200-up-to-6-7x-a100"]], "LoRA (Low-Rank Adaptation)": [[90, null], [168, null]], "LoRA Module id mapping": [[113, "lora-module-id-mapping"]], "LoRA arguments": [[23, "tensorrt_llm.commands.build-parse_arguments-lora-arguments"]], "LoRA tensor format details": [[113, "lora-tensor-format-details"]], "LoRA with Quantization": [[90, "lora-with-quantization"], [168, "lora-with-quantization"]], "LoRA with tensor parallel": [[113, "lora-with-tensor-parallel"]], "Loading function": [[121, "loading-function"]], "Logging Level": [[81, null], [162, null]], "Logits arguments": [[23, "tensorrt_llm.commands.build-parse_arguments-logits-arguments"]], "Logits processor": [[97, "logits-processor"]], "Long Sequences": [[89, null]], "Lookahead Decoding": [[116, "lookahead-decoding"]], "LoraCache configuration": [[113, "loracache-configuration"]], "Low Latency Benchmark": [[127, "low-latency-benchmark"]], "Low Latency GEMM Plugin": [[130, "low-latency-gemm-plugin"]], "Low Latency TensorRT-LLM Engine for Llama-3 70B": [[127, "low-latency-tensorrt-llm-engine-for-llama-3-70b"]], "Low precision AlltoAll": [[12, "low-precision-alltoall"]], "Low-Precision-AllReduce": [[114, null]], "Low-latency Use-Case": [[21, "low-latency-use-case"]], "Lower precision": [[12, "lower-precision"]], "MLA Layers Optimizations": [[15, "mla-layers-optimizations"]], "MLA chunked context": [[2, "mla-chunked-context"]], "MLP": [[137, "module-tensorrt_llm.layers.mlp"]], "MLP Weights": [[119, "mlp-weights"]], "MLPerf on H100 with FP8": [[4, "mlperf-on-h100-with-fp8"]], "MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD with errorcode 1.": [[149, "mpi-abort-was-invoked-on-rank-1-in-communicator-mpi-comm-world-with-errorcode-1"]], "MQA / GQA": [[88, "mqa-gqa"]], "MTP": [[13, "mtp"], [98, "mtp"]], "MTP Eagle": [[14, "mtp-eagle"]], "MTP LM head tensor parallelism": [[12, "mtp-lm-head-tensor-parallelism"]], "MTP Modules": [[14, "mtp-modules"]], "MTP Vanilla": [[14, "mtp-vanilla"]], "MTP for inference": [[14, "mtp-for-inference"]], "MTP implementation in TensorRT LLM": [[14, "mtp-implementation-in-tensorrt-llm"]], "MTP optimization - Relaxed Acceptance": [[14, "mtp-optimization-relaxed-acceptance"]], "Make Evaluation": [[119, "make-evaluation"]], "Make Grammar Computation Capturable by CUDA Graph": [[10, "make-grammar-computation-capturable-by-cuda-graph"]], "Mark Tensors As Output": [[106, "mark-tensors-as-output"]], "Mathematical Modeling": [[8, "mathematical-modeling"]], "Max Throughput Benchmark": [[127, "max-throughput-benchmark"]], "Max Tokens in Paged KV Cache and KV Cache Free GPU Memory Fraction": [[135, "max-tokens-in-paged-kv-cache-and-kv-cache-free-gpu-memory-fraction"]], "Max-Throughput Use Case": [[21, "max-throughput-use-case"]], "Maximum Attention Window Size": [[135, "maximum-attention-window-size"]], "Measurement Methodology": [[17, "measurement-methodology"]], "Medusa": [[116, "medusa"]], "Medusa Tree": [[116, "medusa-tree"]], "Memory Usage of TensorRT-LLM": [[142, null]], "Memory pool": [[142, "memory-pool"]], "Methodology Introduction": [[26, "methodology-introduction"]], "Metrics Endpoint": [[27, "metrics-endpoint"]], "Miscellaneous": [[16, "miscellaneous"]], "Mixed ETP": [[13, "mixed-etp"]], "Mixture of Experts (MoE)": [[94, "mixture-of-experts-moe"], [107, "mixture-of-experts-moe"]], "MoE Auxiliary Kernels": [[20, "moe-auxiliary-kernels"]], "MoE Backend Support Matrix": [[28, "moe-backend-support-matrix"], [29, "moe-backend-support-matrix"]], "MoE Layers Optimizations": [[15, "moe-layers-optimizations"]], "Model Architecture": [[13, "model-architecture"]], "Model Configuration": [[109, "model-configuration"], [151, "model-configuration"], [156, "model-configuration"]], "Model Definition": [[120, null], [151, "model-definition"], [156, "model-definition"]], "Model Engine": [[120, "model-engine"], [157, "model-engine"]], "Model Input": [[149, "model-input"]], "Model Recipes": [[33, null]], "Model Registration": [[151, "model-registration"], [156, "model-registration"]], "Model Support Matrix": [[91, "model-support-matrix"]], "Model Supported Matrix": [[95, "model-supported-matrix"]], "Model Updates": [[155, "model-updates"], [155, "id35"], [155, "id40"], [155, "id47"], [155, "id52"], [155, "id58"], [155, "id64"], [155, "id70"], [155, "id73"], [155, "id75"]], "Model-Feature Support Matrix(Key Models)": [[152, "model-feature-support-matrix-key-models"]], "Model-Specific Deployment Guides": [[33, "model-specific-deployment-guides"]], "Models": [[28, "models"], [29, "models"], [30, "models"], [31, "models"], [32, "models"], [100, null], [138, null]], "Models (PyTorch Backend)": [[145, "models-pytorch-backend"]], "Models (TensorRT Backend)": [[145, "models-tensorrt-backend"]], "Models with customized key names": [[121, "models-with-customized-key-names"]], "Models with customized weight layout": [[121, "models-with-customized-weight-layout"]], "Modifications to Upper-Level Runtime Logic": [[37, "modifications-to-upper-level-runtime-logic"]], "Modifying Existing Methods": [[34, "modifying-existing-methods"]], "Modifying LLM Class Methods": [[34, "modifying-llm-class-methods"]], "Modifying LLM Constructor Arguments": [[34, "modifying-llm-constructor-arguments"]], "Module-level Parallelism Guide": [[94, "module-level-parallelism-guide"]], "More kernel overlap, fusion and optimization": [[12, "more-kernel-overlap-fusion-and-optimization"]], "Motivation": [[10, "motivation"], [17, "motivation"], [86, "motivation"], [96, "motivation"]], "Motivation and Background": [[8, "motivation-and-background"]], "Motivation for Wide-EP": [[94, "motivation-for-wide-ep"]], "Motivation for large-scale EP": [[16, "motivation-for-large-scale-ep"]], "Motivation of Dynasor-CoT": [[11, "motivation-of-dynasor-cot"]], "Motivation of EP communication kernels for GB200": [[16, "motivation-of-ep-communication-kernels-for-gb200"]], "Multi-GPU and Multi-Node Support": [[120, "multi-gpu-and-multi-node-support"]], "Multi-Head, Multi-Query, and Group-Query Attention": [[77, null], [108, null]], "Multi-LoRA Support": [[90, "multi-lora-support"], [168, "multi-lora-support"]], "Multi-Modal Models 3": [[145, "multi-modal-models"]], "Multi-Token Prediction (MTP)": [[20, "multi-token-prediction-mtp"]], "Multi-backend Support": [[17, "multi-backend-support"], [86, "multi-backend-support"]], "Multi-node Serving with Slurm": [[27, "multi-node-serving-with-slurm"]], "Multi-streams": [[13, "multi-streams"]], "Multimodal Benchmarking": [[26, "multimodal-benchmarking"]], "Multimodal Chat API": [[27, "multimodal-chat-api"]], "Multimodal Feature Support Matrix (PyTorch Backend)": [[143, null], [152, "multimodal-feature-support-matrix-pytorch-backend"]], "Multimodal Modality Coverage": [[27, "multimodal-modality-coverage"]], "Multimodal Serving": [[27, "multimodal-serving"]], "Multimodal Serving and Benchmarking": [[26, "multimodal-serving-and-benchmarking"]], "Multimodal Support in TensorRT LLM": [[91, null]], "Multiple Profiles": [[134, "multiple-profiles"]], "N-Gram\u202fSpeculative\u202fDecoding\u202fin TensorRT LLM": [[19, null]], "NGram": [[98, "ngram"], [116, "ngram"]], "NVFP4 (Blackwell)": [[144, "nvfp4-blackwell"]], "Named Arguments": [[23, "tensorrt_llm.commands.build-parse_arguments-named-arguments"]], "Natively supported models": [[121, "natively-supported-models"]], "NeMo LoRA Format": [[90, "nemo-lora-format"], [168, "nemo-lora-format"]], "New XQA-kernel provides 2.4x more Llama-70B throughput within the same latency budget": [[6, null]], "Next Steps": [[154, "next-steps"]], "Normalization": [[137, "module-tensorrt_llm.layers.normalization"]], "Note on context outputs": [[106, "note-on-context-outputs"]], "Numerical Precision": [[144, null]], "Observation over GSM8K dataset": [[16, "observation-over-gsm8k-dataset"]], "Observations over one machine translation dataset": [[16, "observations-over-one-machine-translation-dataset"]], "Obtaining Arbitrary Output Tensors": [[106, "obtaining-arbitrary-output-tensors"]], "Offline EP Load Balancer": [[16, "offline-ep-load-balancer"], [16, "id1"]], "Offline Quantization with ModelOpt": [[95, "offline-quantization-with-modelopt"]], "Offloading to host memory": [[112, "offloading-to-host-memory"]], "Online EP Load Balancer": [[16, "online-ep-load-balancer"], [16, "id2"]], "Online Serving Examples": [[75, null]], "Only collect specific iterations": [[39, "only-collect-specific-iterations"], [126, "only-collect-specific-iterations"]], "OpenAI Chat Client": [[70, null]], "OpenAI Chat Client for Multimodal": [[71, null]], "OpenAI Completion Client": [[72, null]], "OpenAI Completion Client with JSON Schema": [[74, null]], "OpenAI-Compatible Server via trtllm-serve": [[91, "openai-compatible-server-via-trtllm-serve"]], "Openai Completion Client For Lora": [[73, null]], "Operation Fusion": [[99, "operation-fusion"]], "Optimization Highlights": [[20, "optimization-highlights"]], "Optimizations": [[91, "optimizations"]], "Optimize PyTorch native copy and concat using torch.compile": [[12, "optimize-pytorch-native-copy-and-concat-using-torch-compile"]], "Optimizing DeepSeek R1 Throughput on NVIDIA Blackwell GPUs: A Deep Dive for Developers": [[15, null]], "Option 1. Use weekly release NGC docker image": [[18, "option-1-use-weekly-release-ngc-docker-image"]], "Option 1: Build TensorRT LLM in One Step": [[101, "option-1-build-tensorrt-llm-in-one-step"]], "Option 1: Full Build with C++ Compilation": [[101, "option-1-full-build-with-c-compilation"]], "Option 2. Build TensorRT LLM Docker image (Alternative way)": [[18, "option-2-build-tensorrt-llm-docker-image-alternative-way"]], "Option 2: Container for building TensorRT LLM Step-by-Step": [[101, "option-2-container-for-building-tensorrt-llm-step-by-step"]], "Option 2: Python-Only Build without C++ Compilation": [[101, "option-2-python-only-build-without-c-compilation"]], "Options": [[76, "options"]], "Other Build Modes": [[127, "other-build-modes"]], "Out of memory issues": [[2, "out-of-memory-issues"]], "Out-of-Tree Models": [[151, "out-of-tree-models"], [156, "out-of-tree-models"]], "Overlap Optimization": [[17, "overlap-optimization"], [86, "overlap-optimization"]], "Overlap Scheduler": [[38, "overlap-scheduler"], [92, null], [169, null]], "Overlap kernels using programmatic dependent launch (PDL)": [[12, "overlap-kernels-using-programmatic-dependent-launch-pdl"]], "Overriding Docker Compose configuration": [[36, "overriding-docker-compose-configuration"]], "Overview": [[12, "overview"], [34, "overview"], [41, null], [85, "overview"], [109, "overview"], [119, "overview"], [121, "overview"], [122, "overview"], [153, null], [167, "overview"]], "Overview of Parallelism Strategies": [[94, "overview-of-parallelism-strategies"]], "Packed Tensors": [[77, "packed-tensors"]], "Padded and Packed Tensors": [[108, "padded-and-packed-tensors"]], "Page": [[111, "page"]], "Paged Attention, IFB, and Request Scheduling": [[93, null]], "Paged Context Attention": [[134, "paged-context-attention"]], "Paged KV Cache": [[77, "paged-kv-cache"], [93, "paged-kv-cache"], [108, "paged-kv-cache"]], "Parallel strategy": [[15, "parallel-strategy"]], "Parallelism Mapping Support": [[127, "parallelism-mapping-support"]], "Parallelism Strategy": [[13, "parallelism-strategy"]], "Parallelism in TensorRT LLM": [[94, null]], "Pareto Analysis: Throughput-Latency Trade-off Optimization": [[8, "pareto-analysis-throughput-latency-trade-off-optimization"]], "Partial Reuse": [[88, "partial-reuse"]], "Pattern and Pattern Manager": [[110, "pattern-and-pattern-manager"]], "Pattern-Matching and Fusion": [[120, "pattern-matching-and-fusion"]], "Performance": [[7, "performance"], [134, "performance"]], "Performance Analysis": [[39, null], [126, null]], "Performance Analysis: Baseline vs. ADP Balance": [[8, "performance-analysis-baseline-vs-adp-balance"]], "Performance Improvements": [[116, "performance-improvements"]], "Performance Optimization Tips": [[78, "performance-optimization-tips"], [159, "performance-optimization-tips"]], "Performance Results": [[8, "performance-results"]], "Performance Studies": [[17, "performance-studies"]], "Performance Study": [[19, "performance-study"]], "Performance Summary": [[8, "performance-summary"]], "Performance Tuning": [[18, "performance-tuning"], [99, "performance-tuning"]], "Performance Tuning Guide": [[131, null]], "Performance and Accuracy Considerations": [[114, "performance-and-accuracy-considerations"]], "Performance and Analysis": [[10, "performance-and-analysis"]], "Performance expectations": [[112, "performance-expectations"]], "Performance study": [[16, "performance-study"]], "Performance with GEMM + SwiGLU Fusion": [[130, "performance-with-gemm-swiglu-fusion"]], "Performance with GEMM Plugin": [[134, "performance-with-gemm-plugin"]], "Performance with Low Latency GEMM plugin": [[130, "performance-with-low-latency-gemm-plugin"]], "Performance with Quantized KV Cache": [[130, "performance-with-quantized-kv-cache"]], "Performance with Reduce Norm Fusion": [[134, "performance-with-reduce-norm-fusion"]], "Performance with Reduce Norm Fusion + User Buffers:": [[130, "performance-with-reduce-norm-fusion-user-buffers"]], "Performance with multiple profiles": [[134, "performance-with-multiple-profiles"]], "Persistence mode": [[40, "persistence-mode"], [127, "persistence-mode"]], "Piecewise CUDA Graph": [[99, "piecewise-cuda-graph"]], "Piecewise CUDA Graph & Generation Only CUDA Graph": [[99, "piecewise-cuda-graph-generation-only-cuda-graph"]], "Piecewise CUDA Graph Padding": [[99, "piecewise-cuda-graph-padding"]], "Pipeline Parallel Reduce Scatter Optimization": [[134, "pipeline-parallel-reduce-scatter-optimization"]], "Pipeline Parallelism (PP)": [[94, "pipeline-parallelism-pp"]], "Plugin": [[139, null]], "Plugin config arguments": [[23, "tensorrt_llm.commands.build-parse_arguments-plugin-config-arguments"]], "Plugins": [[120, "plugins"]], "Pool": [[111, "pool"]], "Pooling": [[137, "module-tensorrt_llm.layers.pooling"]], "Postprocessing functions": [[121, "postprocessing-functions"]], "Pre-built release container images on NGC": [[102, null]], "Precision Strategy": [[13, "precision-strategy"]], "Precision Support": [[84, "precision-support"], [166, "precision-support"]], "Precision strategy": [[15, "precision-strategy"]], "Prepare": [[124, "prepare"]], "Prepare Dataset": [[128, "prepare-dataset"]], "Prepare the TensorRT LLM Checkpoint": [[119, "prepare-the-tensorrt-llm-checkpoint"]], "Preparing a Dataset": [[40, "preparing-a-dataset"], [41, "preparing-a-dataset"], [127, "preparing-a-dataset"]], "Prerequisite Knowledge": [[131, "prerequisite-knowledge"], [132, null]], "Prerequisites": [[9, "prerequisites"], [18, "prerequisites"], [21, "prerequisites"], [28, "prerequisites"], [29, "prerequisites"], [30, "prerequisites"], [31, "prerequisites"], [32, "prerequisites"], [101, "prerequisites"], [151, "prerequisites"], [156, "prerequisites"]], "Prerequisites: Install TensorRT LLM and download models": [[2, "prerequisites-install-tensorrt-llm-and-download-models"]], "Profiling specific iterations on a trtllm-bench/trtllm-serve run": [[39, "profiling-specific-iterations-on-a-trtllm-bench-trtllm-serve-run"], [126, "profiling-specific-iterations-on-a-trtllm-bench-trtllm-serve-run"]], "Promoting an API from Beta to Committed": [[34, "promoting-an-api-from-beta-to-committed"]], "Prototype Features": [[148, "prototype-features"]], "Pushing Latency Boundaries: Optimizing DeepSeek-R1 Performance on NVIDIA B200 GPUs": [[13, null]], "PyExecutor": [[157, "pyexecutor"]], "PyTorch Backend": [[148, null]], "Python Bindings for the Executor API": [[106, "python-bindings-for-the-executor-api"]], "Python Interface": [[16, "python-interface"]], "Python runtime (Not recommended to be used)": [[142, "python-runtime-not-recommended-to-be-used"]], "Quantization": [[45, "quantization"], [95, null], [140, null], [170, null]], "Quantization APIs": [[122, "quantization-apis"]], "Quantization and Dequantization (Q/DQ)": [[144, "quantization-and-dequantization-q-dq"]], "Quantization in TensorRT LLM": [[95, "quantization-in-tensorrt-llm"]], "Quantization in TensorRT-LLM": [[7, "quantization-in-tensorrt-llm"]], "Quantization in the PyTorch Flow": [[40, "quantization-in-the-pytorch-flow"], [127, "quantization-in-the-pytorch-flow"]], "Quantized KV-Cache": [[130, "quantized-kv-cache"]], "Quick Health Check": [[9, "quick-health-check"]], "Quick Links": [[95, "quick-links"], [100, null]], "Quick Start": [[98, "quick-start"], [148, "quick-start"]], "Quick Start Example": [[149, "quick-start-example"]], "Quick Start Guide": [[154, null]], "Quick Start for Popular Models": [[33, "quick-start-for-popular-models"]], "Quick start": [[91, "quick-start"], [163, "quick-start"]], "Quickstart": [[127, "quickstart"]], "Qwen 3": [[17, "qwen-3"]], "Qwen3-235B-A22B": [[41, "qwen3-235b-a22b"]], "Qwen3-30B-A3B": [[41, "qwen3-30b-a3b"]], "Rank Weights": [[119, "rank-weights"]], "Ray Orchestrator (Prototype)": [[96, null]], "Re-balanced the sparse experts": [[13, "re-balanced-the-sparse-experts"]], "Re-inplace Optimization": [[99, "re-inplace-optimization"]], "ReDrafter": [[116, "redrafter"]], "Recommended Performance Settings": [[28, "recommended-performance-settings"], [29, "recommended-performance-settings"], [30, "recommended-performance-settings"], [31, "recommended-performance-settings"], [32, "recommended-performance-settings"]], "Recompilation": [[99, "recompilation"]], "Reduce Binding and Inter-Process Communication Overhead": [[20, "reduce-binding-and-inter-process-communication-overhead"]], "Reduce Norm Fusion Plugin for Llama models:": [[134, "reduce-norm-fusion-plugin-for-llama-models"]], "Reduce Norm Fusion with User Buffers for Llama Models": [[130, "reduce-norm-fusion-with-user-buffers-for-llama-models"]], "Reference": [[11, "reference"], [118, "reference"]], "References": [[92, "references"], [94, "references"], [169, "references"]], "Relative Attention Bias (RAB)": [[108, "relative-attention-bias-rab"]], "Relax Acceptance Verification": [[13, "relax-acceptance-verification"]], "Relaxed Acceptance": [[14, "relaxed-acceptance"]], "Release Notes": [[155, null]], "Reproducing Benchmarked Results": [[41, "reproducing-benchmarked-results"]], "Reproducing Steps": [[17, "reproducing-steps"]], "Reproducing steps": [[2, "reproducing-steps"], [16, "reproducing-steps"]], "Request Additional Output": [[106, "request-additional-output"]], "Request Time Breakdown": [[26, "request-time-breakdown"]], "ResourceManager": [[157, "resourcemanager"]], "Results": [[128, "results"]], "Retention Policy": [[88, "retention-policy"]], "Rethink network structure": [[12, "rethink-network-structure"]], "Reuse Across Requests": [[88, "reuse-across-requests"]], "Revisiting Paged Context Attention and Context Chunking": [[93, "revisiting-paged-context-attention-and-context-chunking"], [133, "revisiting-paged-context-attention-and-context-chunking"]], "Roadmap": [[83, "roadmap"], [96, "roadmap"], [165, "roadmap"]], "Rotary Positional Embedding (RoPE)": [[77, "rotary-positional-embedding-rope"], [108, "rotary-positional-embedding-rope"]], "RouterGEMM": [[13, "routergemm"]], "Run Docker Container": [[28, "run-docker-container"], [29, "run-docker-container"], [30, "run-docker-container"], [31, "run-docker-container"], [32, "run-docker-container"]], "Run LLM-API with pytorch backend on Slurm": [[62, null]], "Run Offline Inference with LLM API": [[154, "run-offline-inference-with-llm-api"]], "Run benchmarking with trtllm-serve": [[26, null]], "Run gpt-2b + LoRA using Executor / cpp runtime": [[113, null]], "Run the benchmark": [[26, "run-the-benchmark"]], "Run trtllm-bench": [[90, "run-trtllm-bench"], [168, "run-trtllm-bench"]], "Run trtllm-bench with pytorch backend on Slurm": [[63, null]], "Run trtllm-serve with pytorch backend on Slurm": [[64, null]], "Run with trtllm-bench": [[91, "run-with-trtllm-bench"]], "Running Evaluations to Verify Accuracy (Optional)": [[28, "running-evaluations-to-verify-accuracy-optional"], [29, "running-evaluations-to-verify-accuracy-optional"], [30, "running-evaluations-to-verify-accuracy-optional"], [31, "running-evaluations-to-verify-accuracy-optional"]], "Running GPT-OSS-120B with Eagle3 Speculative Decoding on GB200/B200 (TensorRT LLM)": [[9, null]], "Running Pre-quantized Models": [[95, "running-pre-quantized-models"]], "Running Tests": [[34, "running-tests"]], "Running Throughput and Latency Benchmarks": [[128, "running-throughput-and-latency-benchmarks"]], "Running With Weight Streaming to Reduce GPU Memory Consumption": [[117, null]], "Running a High Performance GPT-OSS-120B Inference Server with TensorRT LLM": [[21, null]], "Running multi-modal models in the PyTorch Workflow": [[40, "running-multi-modal-models-in-the-pytorch-workflow"], [127, "running-multi-modal-models-in-the-pytorch-workflow"]], "Running the Benchmark": [[41, "running-the-benchmark"]], "Running the TensorRT LLM Server": [[21, "running-the-tensorrt-llm-server"]], "Running with the PyTorch Workflow": [[40, "running-with-the-pytorch-workflow"], [127, "running-with-the-pytorch-workflow"]], "Runtime": [[1, null], [120, "runtime"], [141, null]], "Runtime Configuration Examples": [[66, null]], "Runtime Customization": [[45, "runtime-customization"]], "Runtime Integrations": [[84, "runtime-integrations"], [166, "runtime-integrations"]], "Runtime Optimizations": [[15, "runtime-optimizations"], [38, "runtime-optimizations"]], "RuntimeError: only rank 0 can start multi-node session, got 1": [[149, "runtimeerror-only-rank-0-can-start-multi-node-session-got-1"]], "Sample Chat Completions Request": [[9, "sample-chat-completions-request"]], "Sampling": [[45, "sampling"], [97, null], [171, null]], "Sampling Parameters": [[109, "sampling-parameters"]], "Sampling Techniques Showcase": [[67, null]], "ScaffoldingLlm": [[11, "scaffoldingllm"]], "Scaling Expert Parallelism in TensorRT LLM (Part 1: Design and Implementation of Large-scale EP)": [[16, null]], "Scaling Expert Parallelism in TensorRT LLM (Part 2: Performance Status and Optimization)": [[20, null]], "Scaling Expert Parallelism in TensorRT LLM (Part 3: Pushing the Performance Boundary)": [[12, null]], "Scaling factor(s)": [[77, "scaling-factor-s"], [108, "scaling-factor-s"]], "Scheduler": [[157, "scheduler"], [173, null]], "Scheduler Introduction": [[173, "scheduler-introduction"]], "Scheduling Strategies for Load Balancing": [[8, "scheduling-strategies-for-load-balancing"]], "Seamless Model Deployment from PyTorch to TensorRT LLM": [[83, "seamless-model-deployment-from-pytorch-to-tensorrt-llm"]], "Seamless Model Deployment from PyTorch to TensorRT-LLM": [[165, "seamless-model-deployment-from-pytorch-to-tensorrt-llm"]], "See also": [[163, "see-also"]], "Selecting Triton as the MoE backend": [[21, "selecting-triton-as-the-moe-backend"], [21, "id3"]], "Sender and Receiver": [[37, "sender-and-receiver"]], "Sending Requests with Different Beam Widths": [[106, "sending-requests-with-different-beam-widths"]], "Serving with trtllm-serve": [[163, null]], "Set power limits": [[40, "set-power-limits"], [127, "set-power-limits"]], "Setting": [[8, "setting"]], "Setting up Multimodal Serving": [[26, "setting-up-multimodal-serving"]], "Show code": [[28, null], [28, null], [29, null], [29, null], [30, null], [31, null], [32, null]], "Single LoRA Adapter": [[90, "single-lora-adapter"], [168, "single-lora-adapter"]], "Single node hanging when using docker run --net=host": [[149, "single-node-hanging-when-using-docker-run-net-host"]], "Situations that can prevent kv cache reuse": [[112, "situations-that-can-prevent-kv-cache-reuse"]], "Sliding Window Attention": [[89, "sliding-window-attention"]], "Sliding Window Attention, Cyclic (Rolling Buffer) KV Cache": [[77, "sliding-window-attention-cyclic-rolling-buffer-kv-cache"], [108, "sliding-window-attention-cyclic-rolling-buffer-kv-cache"]], "Slurm": [[53, "slurm"]], "Smart Router": [[13, "smart-router"]], "Software": [[145, "software"]], "Sparse Attention": [[68, null]], "Sparse Experts as GEMMs (only works when moe_backend=CUTLASS)": [[13, "sparse-experts-as-gemms-only-works-when-moe-backend-cutlass"]], "Speculative Decoding": [[10, "speculative-decoding"], [69, null], [88, "speculative-decoding"], [98, null]], "Speculative Sampling": [[116, null]], "Speculative decoding arguments": [[23, "tensorrt_llm.commands.build-parse_arguments-speculative-decoding-arguments"]], "Speed up inference with SOTA quantization techniques in TRT-LLM": [[7, null]], "Speed-up for the First Turn": [[19, "speed-up-for-the-first-turn"]], "Start the TensorRT LLM Container": [[9, "start-the-tensorrt-llm-container"]], "Start the trtllm-serve service": [[26, "start-the-trtllm-serve-service"]], "Starting a Server": [[27, "starting-a-server"]], "Starting the Server": [[90, "starting-the-server"], [168, "starting-the-server"]], "Step 1. Write Modeling Part": [[118, "step-1-write-modeling-part"]], "Step 1: Clone the repository": [[18, "step-1-clone-the-repository"]], "Step 1: Create the Checkpoint Loader": [[85, "step-1-create-the-checkpoint-loader"], [167, "step-1-create-the-checkpoint-loader"]], "Step 1: Run inference and collect statistics": [[16, "step-1-run-inference-and-collect-statistics"]], "Step 2. Implement Weight Conversion": [[118, "step-2-implement-weight-conversion"]], "Step 2: Create the Checkpoint Weight Loader": [[85, "step-2-create-the-checkpoint-weight-loader"], [167, "step-2-create-the-checkpoint-weight-loader"]], "Step 2: Generate the EPLB configuration": [[16, "step-2-generate-the-eplb-configuration"]], "Step 2: Prepare the TensorRT LLM release Docker image": [[18, "step-2-prepare-the-tensorrt-llm-release-docker-image"]], "Step 3. Register New Model": [[118, "step-3-register-new-model"]], "Step 3: (Optional) Tag and push the Docker image to your registry": [[18, "step-3-optional-tag-and-push-the-docker-image-to-your-registry"]], "Step 3: Create the Checkpoint Config Loader": [[85, "step-3-create-the-checkpoint-config-loader"], [167, "step-3-create-the-checkpoint-config-loader"]], "Step 3: Run inference with the EPLB configuration": [[16, "step-3-run-inference-with-the-eplb-configuration"]], "Step 4. Verify New Model": [[118, "step-4-verify-new-model"]], "Step 4: Create the Checkpoint Weight Mapper": [[85, "step-4-create-the-checkpoint-weight-mapper"], [167, "step-4-create-the-checkpoint-weight-mapper"]], "Step 4: Start the TensorRT LLM server": [[18, "step-4-start-the-tensorrt-llm-server"]], "Step 5: Test the server with a sample request": [[18, "step-5-test-the-server-with-a-sample-request"]], "Step 6: (Optional) Monitor server logs": [[18, "step-6-optional-monitor-server-logs"]], "Step 7: (Optional) Stop the server": [[18, "step-7-optional-stop-the-server"]], "Step-by-Step Guide": [[151, "step-by-step-guide"], [156, "step-by-step-guide"]], "StreamingLLM": [[77, "streamingllm"], [108, "streamingllm"]], "Structured output with guided decoding": [[106, "structured-output-with-guided-decoding"]], "Summary": [[127, "summary"]], "Summary of Configuration Option Recommendations:": [[130, "summary-of-configuration-option-recommendations"], [134, "summary-of-configuration-option-recommendations"]], "Support Matrix": [[83, "support-matrix"], [84, null], [145, null], [165, "support-matrix"], [166, null]], "Support Models": [[84, "support-models"], [166, "support-models"]], "Support Stream Interval": [[20, "support-stream-interval"]], "Support matrix": [[144, "support-matrix"]], "Supported C++ Header Files": [[101, "supported-c-header-files"]], "Supported Models": [[152, null]], "Supported Quantization Modes": [[40, "supported-quantization-modes"], [127, "supported-quantization-modes"]], "Syntax": [[24, "syntax"], [27, "syntax"]], "System Level optimizations": [[13, "system-level-optimizations"]], "TRTLLM Backend": [[13, "trtllm-backend"]], "TRTLLM bench with LORA": [[90, "trtllm-bench-with-lora"], [168, "trtllm-bench-with-lora"]], "TRTLLM serve with LoRA": [[90, "trtllm-serve-with-lora"], [168, "trtllm-serve-with-lora"]], "Table of Contents": [[2, "table-of-contents"], [8, "table-of-contents"], [10, "table-of-contents"], [11, "table-of-contents"], [12, "table-of-contents"], [13, "table-of-contents"], [14, "table-of-contents"], [15, "table-of-contents"], [16, "table-of-contents"], [19, "table-of-contents"], [20, "table-of-contents"], [35, "table-of-contents"], [37, "table-of-contents"], [85, "table-of-contents"], [90, "table-of-contents"], [99, "table-of-contents"], [131, "table-of-contents"], [132, "table-of-contents"], [151, "table-of-contents"], [156, "table-of-contents"], [167, "table-of-contents"], [168, "table-of-contents"]], "Target Model": [[10, "target-model"]], "Technical Detail: The QuantMode Flags": [[144, "technical-detail-the-quantmode-flags"]], "Tensor Parallel vs Expert Parallel": [[107, "tensor-parallel-vs-expert-parallel"]], "Tensor Parallelism (TP)": [[94, "tensor-parallelism-tp"], [94, "id1"]], "Tensor-Related Methods": [[110, "tensor-related-methods"]], "TensorRT Compiler": [[120, "tensorrt-compiler"]], "TensorRT LLM Benchmarking": [[40, null]], "TensorRT LLM Checkpoint": [[119, null]], "TensorRT LLM Custom Backend": [[99, "tensorrt-llm-custom-backend"]], "TensorRT-LLM Benchmarking": [[127, null]], "TensorRT-LLM Build Workflow": [[122, null]], "TensorRT-LLM Model Weights Loader": [[121, null]], "TensorRT-LLM Release 0.10.0": [[155, "tensorrt-llm-release-0-10-0"]], "TensorRT-LLM Release 0.11.0": [[155, "tensorrt-llm-release-0-11-0"]], "TensorRT-LLM Release 0.12.0": [[155, "tensorrt-llm-release-0-12-0"]], "TensorRT-LLM Release 0.13.0": [[155, "tensorrt-llm-release-0-13-0"]], "TensorRT-LLM Release 0.14.0": [[155, "tensorrt-llm-release-0-14-0"]], "TensorRT-LLM Release 0.15.0": [[155, "tensorrt-llm-release-0-15-0"]], "TensorRT-LLM Release 0.16.0": [[155, "tensorrt-llm-release-0-16-0"]], "TensorRT-LLM Release 0.17.0": [[155, "tensorrt-llm-release-0-17-0"]], "TensorRT-LLM Release 0.18.0": [[155, "tensorrt-llm-release-0-18-0"]], "TensorRT-LLM Release 0.18.1": [[155, "tensorrt-llm-release-0-18-1"]], "TensorRT-LLM Release 0.18.2": [[155, "tensorrt-llm-release-0-18-2"]], "TensorRT-LLM Release 0.19.0": [[155, "tensorrt-llm-release-0-19-0"]], "TensorRT-LLM Release 0.20.0": [[155, "tensorrt-llm-release-0-20-0"]], "TensorRT-LLM Release 0.21.0": [[155, "tensorrt-llm-release-0-21-0"]], "TensorRT-LLM Release 0.7.1": [[155, "tensorrt-llm-release-0-7-1"]], "TensorRT-LLM Release 0.8.0": [[155, "tensorrt-llm-release-0-8-0"]], "TensorRT-LLM Release 0.9.0": [[155, "tensorrt-llm-release-0-9-0"]], "TensorRT-LLM Release 1.0": [[155, "tensorrt-llm-release-1-0"]], "Test definitions": [[35, "test-definitions"]], "Test the Server with a Sample Request": [[21, "test-the-server-with-a-sample-request"]], "Testing API Endpoint": [[28, "testing-api-endpoint"], [29, "testing-api-endpoint"], [30, "testing-api-endpoint"], [31, "testing-api-endpoint"], [32, "testing-api-endpoint"]], "The Basics": [[88, "the-basics"]], "The Executor Class": [[106, "the-executor-class"]], "The Features of the TrtllmAttention Backend": [[77, "the-features-of-the-trtllmattention-backend"]], "The Request Class": [[106, "the-request-class"]], "The Response Class": [[106, "the-response-class"]], "The Result Class": [[106, "the-result-class"]], "The effect of EP Load Balancer": [[16, "the-effect-of-ep-load-balancer"], [16, "id3"]], "The schedulers": [[93, "the-schedulers"]], "The \u201cProbe\u201d technique": [[11, "the-probe-technique"]], "Theoretical Analysis and Modeling": [[8, "theoretical-analysis-and-modeling"]], "Throughput Benchmarking": [[40, "throughput-benchmarking"], [127, "throughput-benchmarking"]], "Throughput Measurements": [[41, "throughput-measurements"]], "Time Per Output Token (TPOT) and Inter-Token Latency (ITL)": [[26, "time-per-output-token-tpot-and-inter-token-latency-itl"], [28, "time-per-output-token-tpot-and-inter-token-latency-itl"], [29, "time-per-output-token-tpot-and-inter-token-latency-itl"], [30, "time-per-output-token-tpot-and-inter-token-latency-itl"], [31, "time-per-output-token-tpot-and-inter-token-latency-itl"]], "Time to First Token (TTFT)": [[26, "time-to-first-token-ttft"], [28, "time-to-first-token-ttft"], [29, "time-to-first-token-ttft"], [30, "time-to-first-token-ttft"], [31, "time-to-first-token-ttft"]], "Tips": [[146, "tips"]], "Tips and Troubleshooting": [[149, "tips-and-troubleshooting"]], "Tips for Piecewise CUDA Graph": [[99, "tips-for-piecewise-cuda-graph"]], "Tokenizer Customization": [[45, "tokenizer-customization"]], "Tokens Per Second (TPS) or Output Token Throughput": [[26, "tokens-per-second-tps-or-output-token-throughput"], [28, "tokens-per-second-tps-or-output-token-throughput"], [29, "tokens-per-second-tps-or-output-token-throughput"], [30, "tokens-per-second-tps-or-output-token-throughput"], [31, "tokens-per-second-tps-or-output-token-throughput"]], "Top Level API": [[157, "top-level-api"]], "Topology Requirements": [[114, "topology-requirements"]], "Torch Compile & Piecewise CUDA Graph": [[99, null]], "Torch IR Optimization": [[99, "torch-ir-optimization"]], "Total Token Throughput": [[26, "total-token-throughput"], [28, "total-token-throughput"], [29, "total-token-throughput"], [30, "total-token-throughput"], [31, "total-token-throughput"]], "Trace Grammar State for Draft Token Proposal and Rejection": [[10, "trace-grammar-state-for-draft-token-proposal-and-rejection"]], "Tradeoff": [[92, "tradeoff"], [169, "tradeoff"]], "Transceiver": [[37, "transceiver"]], "Transfer Agent": [[37, "transfer-agent"]], "Translator": [[121, "translator"]], "Tree-based speculative decoding support": [[14, "tree-based-speculative-decoding-support"]], "Triggering CI Best Practices": [[35, "triggering-ci-best-practices"]], "Triggering Post-merge tests": [[35, "triggering-post-merge-tests"]], "Triton Inference Server": [[17, "triton-inference-server"]], "Trouble shooting": [[121, "trouble-shooting"]], "Troubleshooting": [[146, null]], "Troubleshooting Tips": [[18, "troubleshooting-tips"], [21, "troubleshooting-tips"], [28, "troubleshooting-tips"], [29, "troubleshooting-tips"], [30, "troubleshooting-tips"], [31, "troubleshooting-tips"], [32, "troubleshooting-tips"]], "Troubleshooting Tips and Pitfalls To Avoid": [[128, "troubleshooting-tips-and-pitfalls-to-avoid"]], "Troubleshooting and FAQ": [[86, "troubleshooting-and-faq"], [105, "troubleshooting-and-faq"]], "Troubleshooting: Data Race between Host and CUDA Callback": [[10, "troubleshooting-data-race-between-host-and-cuda-callback"]], "Troubleshooting: Deadlock by GIL and CUDA Mutex": [[10, "troubleshooting-deadlock-by-gil-and-cuda-mutex"]], "Tuning Case Study": [[133, "tuning-case-study"], [133, "id2"]], "Tuning Max Batch Size": [[133, "tuning-max-batch-size"]], "Tuning Max Batch Size and Max Num Tokens": [[133, null]], "Tuning Max Num Tokens": [[133, "tuning-max-num-tokens"]], "Two Challenges": [[10, "two-challenges"]], "Two Model Speculative Decoding Architecture": [[98, "two-model-speculative-decoding-architecture"]], "Types of Events": [[111, "types-of-events"]], "Understand inference time GPU memory usage": [[142, "understand-inference-time-gpu-memory-usage"]], "Understanding the TensorRT-LLM scheduler": [[133, "understanding-the-tensorrt-llm-scheduler"]], "Unit tests": [[35, "unit-tests"]], "Upload the Docker Image to DockerHub": [[123, "upload-the-docker-image-to-dockerhub"]], "Usage": [[86, "usage"], [92, "usage"], [95, "usage"], [99, "usage"], [114, "usage"], [169, "usage"]], "Usage and Examples": [[24, "usage-and-examples"]], "Usage with trtllm-bench and trtllm-serve": [[98, "usage-with-trtllm-bench-and-trtllm-serve"]], "Useful Build-Time Flags": [[134, null]], "Useful Runtime Options": [[135, null]], "User-provided drafting": [[98, "user-provided-drafting"]], "Using Checkpoint Loaders": [[85, "using-checkpoint-loaders"], [167, "using-checkpoint-loaders"]], "Using Dev Containers": [[36, null]], "Using Medusa with TensorRT-LLM": [[116, "using-medusa-with-tensorrt-llm"]], "Using test_to_stage_mapping.py": [[35, "using-test-to-stage-mapping-py"]], "Validated Networks for Benchmarking": [[40, "validated-networks-for-benchmarking"], [127, "validated-networks-for-benchmarking"]], "Variables": [[41, "variables"]], "Verification and Sampling": [[98, "verification-and-sampling"]], "Visualize the PyTorch profiler results": [[39, "visualize-the-pytorch-profiler-results"], [126, "visualize-the-pytorch-profiler-results"]], "Volume Mounts": [[36, "volume-mounts"]], "WIP: Enable more features by default": [[2, "wip-enable-more-features-by-default"]], "Waiving tests": [[35, "waiving-tests"]], "Weight Bindings": [[120, "weight-bindings"]], "Weight Loading": [[151, "weight-loading"], [156, "weight-loading"]], "Weights absorb and MQA": [[15, "weights-absorb-and-mqa"]], "Welcome to TensorRT LLM\u2019s Documentation!": [[100, null]], "What Can You Do With TensorRT LLM?": [[153, "what-can-you-do-with-tensorrt-llm"]], "What Triggers an Event?": [[111, "what-triggers-an-event"]], "What is H100 FP8?": [[4, "what-is-h100-fp8"]], "What\u2019s coming next": [[7, "whats-coming-next"]], "When to Create Custom Components": [[85, "when-to-create-custom-components"], [167, "when-to-create-custom-components"]], "When to Use Graph Rewriting?": [[110, "when-to-use-graph-rewriting"]], "Wide Expert Parallelism": [[28, "wide-expert-parallelism"]], "Wide Expert Parallelism (Wide-EP)": [[94, "wide-expert-parallelism-wide-ep"], [94, "id3"]], "WindowBlockManager/BlockManager": [[111, "windowblockmanager-blockmanager"]], "Worker": [[11, "worker"]], "Workflow": [[37, "workflow"], [121, "workflow"], [127, "workflow"]], "Workload Profile": [[13, "workload-profile"]], "World Configuration": [[109, "world-configuration"]], "XQA Optimization": [[77, "xqa-optimization"], [108, "xqa-optimization"]], "YAML Configuration": [[90, "yaml-configuration"], [90, "id1"], [168, "yaml-configuration"], [168, "id1"]], "YAML Configuration Files": [[80, "yaml-configuration-files"], [161, "yaml-configuration-files"]], "_prepare_draft_requests": [[98, "prepare-draft-requests"]], "_prepare_draft_tokens": [[98, "prepare-draft-tokens"]], "additional_model_outputs": [[76, "additional-model-outputs"]], "attention_backend": [[26, "attention-backend"], [28, "attention-backend"], [30, "attention-backend"], [31, "attention-backend"]], "bufferManager.h": [[1, "buffermanager-h"]], "build": [[22, "trtllm-bench-build"]], "cacheCommunicator.h": [[0, "cachecommunicator-h"]], "cnn_dailymail": [[24, "trtllm-eval-cnn-dailymail"]], "common.h": [[1, "common-h"]], "cudaEvent.h": [[1, "cudaevent-h"]], "cudaStream.h": [[1, "cudastream-h"]], "cuda_graph_config": [[26, "cuda-graph-config"], [28, "cuda-graph-config"], [29, "cuda-graph-config"], [30, "cuda-graph-config"], [31, "cuda-graph-config"], [32, "cuda-graph-config"]], "dataTransceiverState.h": [[0, "datatransceiverstate-h"]], "dataset": [[22, "dataset"]], "decoderState.h": [[1, "decoderstate-h"]], "decodingInput.h": [[1, "decodinginput-h"]], "decodingOutput.h": [[1, "decodingoutput-h"]], "disaggServerUtil.h": [[0, "disaggserverutil-h"]], "disaggregated": [[27, "trtllm-serve-disaggregated"]], "disaggregated_mpi_worker": [[27, "trtllm-serve-disaggregated-mpi-worker"]], "eagleBuffers.h": [[1, "eaglebuffers-h"]], "eagleModule.h": [[1, "eaglemodule-h"]], "executor.h": [[0, "executor-h"]], "explicitDraftTokensBuffers.h": [[1, "explicitdrafttokensbuffers-h"]], "gpqa_diamond": [[24, "trtllm-eval-gpqa-diamond"]], "gpqa_extended": [[24, "trtllm-eval-gpqa-extended"]], "gpqa_main": [[24, "trtllm-eval-gpqa-main"]], "gptDecoder.h": [[1, "gptdecoder-h"]], "gptDecoderBatched.h": [[1, "gptdecoderbatched-h"]], "gptJsonConfig.h": [[1, "gptjsonconfig-h"]], "gsm8k": [[24, "trtllm-eval-gsm8k"]], "iBuffer.h": [[1, "ibuffer-h"]], "iGptDecoderBatched.h": [[1, "igptdecoderbatched-h"]], "iTensor.h": [[1, "itensor-h"]], "ipcNvlsMemory.h": [[1, "ipcnvlsmemory-h"]], "ipcUtils.h": [[1, "ipcutils-h"]], "json_mode_eval": [[24, "trtllm-eval-json-mode-eval"]], "kv_cache_config": [[26, "kv-cache-config"], [28, "kv-cache-config"], [30, "kv-cache-config"], [31, "kv-cache-config"]], "kv_cache_config.free_gpu_memory_fraction": [[32, "kv-cache-config-free-gpu-memory-fraction"]], "kv_cache_free_gpu_memory_fraction": [[28, "kv-cache-free-gpu-memory-fraction"], [29, "kv-cache-free-gpu-memory-fraction"], [30, "kv-cache-free-gpu-memory-fraction"], [31, "kv-cache-free-gpu-memory-fraction"]], "latency": [[22, "trtllm-bench-latency"]], "logprobs": [[76, "logprobs"]], "longbench_v2": [[24, "trtllm-eval-longbench-v2"]], "lookaheadBuffers.h": [[1, "lookaheadbuffers-h"]], "lookaheadModule.h": [[1, "lookaheadmodule-h"]], "loraCache.h": [[1, "loracache-h"]], "loraCachePageManagerConfig.h": [[1, "loracachepagemanagerconfig-h"]], "loraModule.h": [[1, "loramodule-h"]], "max_batch_size": [[28, "max-batch-size"], [29, "max-batch-size"], [30, "max-batch-size"], [31, "max-batch-size"], [32, "max-batch-size"], [93, "max-batch-size"]], "max_batch_size, max_seq_len and max_num_tokens": [[93, "max-batch-size-max-seq-len-and-max-num-tokens"]], "max_num_tokens": [[28, "max-num-tokens"], [29, "max-num-tokens"], [30, "max-num-tokens"], [31, "max-num-tokens"], [32, "max-num-tokens"], [93, "max-num-tokens"]], "max_seq_len": [[28, "max-seq-len"], [29, "max-seq-len"], [30, "max-seq-len"], [31, "max-seq-len"], [32, "max-seq-len"], [93, "max-seq-len"]], "medusaModule.h": [[1, "medusamodule-h"]], "memoryCounters.h": [[1, "memorycounters-h"]], "mm_embedding_serve": [[27, "trtllm-serve-mm-embedding-serve"]], "mmlu": [[24, "trtllm-eval-mmlu"]], "mmmu": [[24, "trtllm-eval-mmmu"]], "modelConfig.h": [[1, "modelconfig-h"]], "moe_config": [[26, "moe-config"], [28, "moe-config"], [29, "moe-config"], [30, "moe-config"], [31, "moe-config"], [32, "moe-config"]], "moe_expert_parallel_size": [[28, "moe-expert-parallel-size"], [29, "moe-expert-parallel-size"], [30, "moe-expert-parallel-size"], [31, "moe-expert-parallel-size"], [32, "moe-expert-parallel-size"]], "prepare_dataset": [[22, "prepare-dataset"]], "prepare_dataset.py": [[22, "prepare-dataset-py"]], "promptTuningParams.h": [[1, "prompttuningparams-h"]], "prompt_logprobs": [[76, "prompt-logprobs"]], "rawEngine.h": [[1, "rawengine-h"]], "return_context_logits": [[76, "return-context-logits"]], "return_generation_logits": [[76, "return-generation-logits"]], "runtimeDefaults.h": [[1, "runtimedefaults-h"]], "samplingConfig.h": [[1, "samplingconfig-h"]], "serialization.h": [[0, "serialization-h"]], "serve": [[27, "trtllm-serve-serve"]], "speculativeDecodingMode.h": [[1, "speculativedecodingmode-h"]], "speculativeDecodingModule.h": [[1, "speculativedecodingmodule-h"]], "tensor.h": [[0, "tensor-h"]], "tensor_parallel_size": [[28, "tensor-parallel-size"], [29, "tensor-parallel-size"], [30, "tensor-parallel-size"], [31, "tensor-parallel-size"], [32, "tensor-parallel-size"]], "throughput": [[22, "trtllm-bench-throughput"]], "tllmLogger.h": [[1, "tllmlogger-h"]], "token_norm_dist": [[22, "token-norm-dist"]], "token_unif_dist": [[22, "token-unif-dist"]], "transferAgent.h": [[0, "transferagent-h"]], "trtllm-bench": [[22, null], [22, "trtllm-bench"]], "trtllm-build": [[23, null]], "trtllm-eval": [[24, null], [24, "trtllm-eval"]], "trtllm-serve": [[17, "trtllm-serve"], [25, null], [27, null], [27, "trtllm-serve"], [86, "trtllm-serve"]], "trust_remote_code": [[28, "trust-remote-code"], [29, "trust-remote-code"], [30, "trust-remote-code"], [31, "trust-remote-code"], [32, "trust-remote-code"]], "types.h": [[0, "types-h"]], "virtualMemory.h": [[1, "virtualmemory-h"]], "wo GEMM FP4 quantization": [[12, "wo-gemm-fp4-quantization"]], "worldConfig.h": [[1, "worldconfig-h"]], "\u26a1 State-of-the-Art Performance": [[153, "state-of-the-art-performance"]], "\ud83c\udfaf Comprehensive Model Support": [[153, "comprehensive-model-support"]], "\ud83d\udd25 Architected on Pytorch": [[153, "architected-on-pytorch"]], "\ud83d\udd27 Latest GPU Architecture Support": [[153, "latest-gpu-architecture-support"]], "\ud83d\ude80 Advanced Optimization & Production Features": [[153, "advanced-optimization-production-features"]]}, "docnames": ["_cpp_gen/executor", "_cpp_gen/runtime", "blogs/Best_perf_practice_on_DeepSeek-R1_in_TensorRT-LLM", "blogs/Falcon180B-H200", "blogs/H100vsA100", "blogs/H200launch", "blogs/XQA-kernel", "blogs/quantization-in-TRT-LLM", "blogs/tech_blog/blog10_ADP_Balance_Strategy", "blogs/tech_blog/blog11_GPT_OSS_Eagle3", "blogs/tech_blog/blog12_Combining_Guided_Decoding_and_Speculative_Decoding", "blogs/tech_blog/blog13_Inference_Time_Compute_Implementation_in_TensorRT-LLM", "blogs/tech_blog/blog14_Scaling_Expert_Parallelism_in_TensorRT-LLM_part3", "blogs/tech_blog/blog1_Pushing_Latency_Boundaries_Optimizing_DeepSeek-R1_Performance_on_NVIDIA_B200_GPUs", "blogs/tech_blog/blog2_DeepSeek_R1_MTP_Implementation_and_Optimization", "blogs/tech_blog/blog3_Optimizing_DeepSeek_R1_Throughput_on_NVIDIA_Blackwell_GPUs", "blogs/tech_blog/blog4_Scaling_Expert_Parallelism_in_TensorRT-LLM", "blogs/tech_blog/blog5_Disaggregated_Serving_in_TensorRT-LLM", "blogs/tech_blog/blog6_Llama4_maverick_eagle_guide", "blogs/tech_blog/blog7_NGram_performance_Analysis_And_Auto_Enablement", "blogs/tech_blog/blog8_Scaling_Expert_Parallelism_in_TensorRT-LLM_part2", "blogs/tech_blog/blog9_Deploying_GPT_OSS_on_TRTLLM", "commands/trtllm-bench", "commands/trtllm-build", "commands/trtllm-eval", "commands/trtllm-serve/index", "commands/trtllm-serve/run-benchmark-with-trtllm-serve", "commands/trtllm-serve/trtllm-serve", "deployment-guide/deployment-guide-for-deepseek-r1-on-trtllm", "deployment-guide/deployment-guide-for-gpt-oss-on-trtllm", "deployment-guide/deployment-guide-for-llama3.3-70b-on-trtllm", "deployment-guide/deployment-guide-for-llama4-scout-on-trtllm", "deployment-guide/deployment-guide-for-qwen3-next-on-trtllm", "deployment-guide/index", "developer-guide/api-change", "developer-guide/ci-overview", "developer-guide/dev-containers", "developer-guide/kv-transfer", "developer-guide/overview", "developer-guide/perf-analysis", "developer-guide/perf-benchmarking", "developer-guide/perf-overview", "examples/curl_chat_client", "examples/curl_chat_client_for_multimodal", "examples/curl_completion_client", "examples/customization", "examples/deepseek_r1_reasoning_parser", "examples/dynamo_k8s_example", "examples/genai_perf_client", "examples/genai_perf_client_for_multimodal", "examples/index", "examples/kvcacheconfig", "examples/kvcacheretentionconfig", "examples/llm_api_examples", "examples/llm_guided_decoding", "examples/llm_inference", "examples/llm_inference_async", "examples/llm_inference_async_streaming", "examples/llm_inference_distributed", "examples/llm_kv_cache_connector", "examples/llm_kv_cache_offloading", "examples/llm_logits_processor", "examples/llm_mgmn_llm_distributed", "examples/llm_mgmn_trtllm_bench", "examples/llm_mgmn_trtllm_serve", "examples/llm_multilora", "examples/llm_runtime", "examples/llm_sampling", "examples/llm_sparse_attention", "examples/llm_speculative_decoding", "examples/openai_chat_client", "examples/openai_chat_client_for_multimodal", "examples/openai_completion_client", "examples/openai_completion_client_for_lora", "examples/openai_completion_client_json_schema", "examples/trtllm_serve_examples", "features/additional-outputs", "features/attention", "features/auto_deploy/advanced/benchmarking_with_trtllm_bench", "features/auto_deploy/advanced/example_run", "features/auto_deploy/advanced/expert_configurations", "features/auto_deploy/advanced/logging", "features/auto_deploy/advanced/workflow", "features/auto_deploy/auto-deploy", "features/auto_deploy/support_matrix", "features/checkpoint-loading", "features/disagg-serving", "features/feature-combination-matrix", "features/kvcache", "features/long-sequence", "features/lora", "features/multi-modality", "features/overlap-scheduler", "features/paged-attention-ifb-scheduler", "features/parallel-strategy", "features/quantization", "features/ray-orchestrator", "features/sampling", "features/speculative-decoding", "features/torch_compile_and_piecewise_cuda_graph", "index", "installation/build-from-source-linux", "installation/containers", "installation/index", "installation/linux", "legacy/advanced/disaggregated-service", "legacy/advanced/executor", "legacy/advanced/expert-parallelism", "legacy/advanced/gpt-attention", "legacy/advanced/gpt-runtime", "legacy/advanced/graph-rewriting", "legacy/advanced/kv-cache-management", "legacy/advanced/kv-cache-reuse", "legacy/advanced/lora", "legacy/advanced/lowprecision-pcie-allreduce", "legacy/advanced/open-sourced-cutlass-kernels", "legacy/advanced/speculative-decoding", "legacy/advanced/weight-streaming", "legacy/architecture/add-model", "legacy/architecture/checkpoint", "legacy/architecture/core-concepts", "legacy/architecture/model-weights-loader", "legacy/architecture/workflow", "legacy/dev-on-cloud/build-image-to-dockerhub", "legacy/dev-on-cloud/dev-on-runpod", "legacy/key-features", "legacy/performance/perf-analysis", "legacy/performance/perf-benchmarking", "legacy/performance/performance-tuning-guide/benchmarking-default-performance", "legacy/performance/performance-tuning-guide/deciding-model-sharding-strategy", "legacy/performance/performance-tuning-guide/fp8-quantization", "legacy/performance/performance-tuning-guide/index", "legacy/performance/performance-tuning-guide/introduction", "legacy/performance/performance-tuning-guide/tuning-max-batch-size-and-max-num-tokens", "legacy/performance/performance-tuning-guide/useful-build-time-flags", "legacy/performance/performance-tuning-guide/useful-runtime-flags", "legacy/python-api/tensorrt_llm.functional", "legacy/python-api/tensorrt_llm.layers", "legacy/python-api/tensorrt_llm.models", "legacy/python-api/tensorrt_llm.plugin", "legacy/python-api/tensorrt_llm.quantization", "legacy/python-api/tensorrt_llm.runtime", "legacy/reference/memory", "legacy/reference/multimodal-feature-support-matrix", "legacy/reference/precision", "legacy/reference/support-matrix", "legacy/reference/troubleshooting", "legacy/tensorrt_quickstart", "legacy/torch", "llm-api/index", "llm-api/reference", "models/adding-new-model", "models/supported-models", "overview", "quick-start-guide", "release-notes", "torch/adding_new_model", "torch/arch_overview", "torch/attention", "torch/auto_deploy/advanced/benchmarking_with_trtllm_bench", "torch/auto_deploy/advanced/example_run", "torch/auto_deploy/advanced/expert_configurations", "torch/auto_deploy/advanced/logging", "torch/auto_deploy/advanced/serving_with_trtllm_serve", "torch/auto_deploy/advanced/workflow", "torch/auto_deploy/auto-deploy", "torch/auto_deploy/support_matrix", "torch/features/checkpoint_loading", "torch/features/lora", "torch/features/overlap_scheduler", "torch/features/quantization", "torch/features/sampling", "torch/kv_cache_manager", "torch/scheduler"], "envversion": {"sphinx": 62, "sphinx.domains.c": 3, "sphinx.domains.changeset": 1, "sphinx.domains.citation": 1, "sphinx.domains.cpp": 9, "sphinx.domains.index": 1, "sphinx.domains.javascript": 3, "sphinx.domains.math": 2, "sphinx.domains.python": 4, "sphinx.domains.rst": 2, "sphinx.domains.std": 2, "sphinx.ext.todo": 2, "sphinx.ext.viewcode": 1}, "filenames": ["_cpp_gen/executor.rst", "_cpp_gen/runtime.rst", "blogs/Best_perf_practice_on_DeepSeek-R1_in_TensorRT-LLM.md", "blogs/Falcon180B-H200.md", "blogs/H100vsA100.md", "blogs/H200launch.md", "blogs/XQA-kernel.md", "blogs/quantization-in-TRT-LLM.md", "blogs/tech_blog/blog10_ADP_Balance_Strategy.md", "blogs/tech_blog/blog11_GPT_OSS_Eagle3.md", "blogs/tech_blog/blog12_Combining_Guided_Decoding_and_Speculative_Decoding.md", "blogs/tech_blog/blog13_Inference_Time_Compute_Implementation_in_TensorRT-LLM.md", "blogs/tech_blog/blog14_Scaling_Expert_Parallelism_in_TensorRT-LLM_part3.md", "blogs/tech_blog/blog1_Pushing_Latency_Boundaries_Optimizing_DeepSeek-R1_Performance_on_NVIDIA_B200_GPUs.md", "blogs/tech_blog/blog2_DeepSeek_R1_MTP_Implementation_and_Optimization.md", "blogs/tech_blog/blog3_Optimizing_DeepSeek_R1_Throughput_on_NVIDIA_Blackwell_GPUs.md", "blogs/tech_blog/blog4_Scaling_Expert_Parallelism_in_TensorRT-LLM.md", "blogs/tech_blog/blog5_Disaggregated_Serving_in_TensorRT-LLM.md", "blogs/tech_blog/blog6_Llama4_maverick_eagle_guide.md", "blogs/tech_blog/blog7_NGram_performance_Analysis_And_Auto_Enablement.md", "blogs/tech_blog/blog8_Scaling_Expert_Parallelism_in_TensorRT-LLM_part2.md", "blogs/tech_blog/blog9_Deploying_GPT_OSS_on_TRTLLM.md", "commands/trtllm-bench.rst", "commands/trtllm-build.rst", "commands/trtllm-eval.rst", "commands/trtllm-serve/index.rst", "commands/trtllm-serve/run-benchmark-with-trtllm-serve.md", "commands/trtllm-serve/trtllm-serve.rst", "deployment-guide/deployment-guide-for-deepseek-r1-on-trtllm.md", "deployment-guide/deployment-guide-for-gpt-oss-on-trtllm.md", "deployment-guide/deployment-guide-for-llama3.3-70b-on-trtllm.md", "deployment-guide/deployment-guide-for-llama4-scout-on-trtllm.md", "deployment-guide/deployment-guide-for-qwen3-next-on-trtllm.md", "deployment-guide/index.rst", "developer-guide/api-change.md", "developer-guide/ci-overview.md", "developer-guide/dev-containers.md", "developer-guide/kv-transfer.md", "developer-guide/overview.md", "developer-guide/perf-analysis.md", "developer-guide/perf-benchmarking.md", "developer-guide/perf-overview.md", "examples/curl_chat_client.rst", "examples/curl_chat_client_for_multimodal.rst", "examples/curl_completion_client.rst", "examples/customization.md", "examples/deepseek_r1_reasoning_parser.rst", "examples/dynamo_k8s_example.rst", "examples/genai_perf_client.rst", "examples/genai_perf_client_for_multimodal.rst", "examples/index.rst", "examples/kvcacheconfig.md", "examples/kvcacheretentionconfig.md", "examples/llm_api_examples.rst", "examples/llm_guided_decoding.rst", "examples/llm_inference.rst", "examples/llm_inference_async.rst", "examples/llm_inference_async_streaming.rst", "examples/llm_inference_distributed.rst", "examples/llm_kv_cache_connector.rst", "examples/llm_kv_cache_offloading.rst", "examples/llm_logits_processor.rst", "examples/llm_mgmn_llm_distributed.rst", "examples/llm_mgmn_trtllm_bench.rst", "examples/llm_mgmn_trtllm_serve.rst", "examples/llm_multilora.rst", "examples/llm_runtime.rst", "examples/llm_sampling.rst", "examples/llm_sparse_attention.rst", "examples/llm_speculative_decoding.rst", "examples/openai_chat_client.rst", "examples/openai_chat_client_for_multimodal.rst", "examples/openai_completion_client.rst", "examples/openai_completion_client_for_lora.rst", "examples/openai_completion_client_json_schema.rst", "examples/trtllm_serve_examples.rst", "features/additional-outputs.md", "features/attention.md", "features/auto_deploy/advanced/benchmarking_with_trtllm_bench.md", "features/auto_deploy/advanced/example_run.md", "features/auto_deploy/advanced/expert_configurations.md", "features/auto_deploy/advanced/logging.md", "features/auto_deploy/advanced/workflow.md", "features/auto_deploy/auto-deploy.md", "features/auto_deploy/support_matrix.md", "features/checkpoint-loading.md", "features/disagg-serving.md", "features/feature-combination-matrix.md", "features/kvcache.md", "features/long-sequence.md", "features/lora.md", "features/multi-modality.md", "features/overlap-scheduler.md", "features/paged-attention-ifb-scheduler.md", "features/parallel-strategy.md", "features/quantization.md", "features/ray-orchestrator.md", "features/sampling.md", "features/speculative-decoding.md", "features/torch_compile_and_piecewise_cuda_graph.md", "index.rst", "installation/build-from-source-linux.md", "installation/containers.md", "installation/index.rst", "installation/linux.md", "legacy/advanced/disaggregated-service.md", "legacy/advanced/executor.md", "legacy/advanced/expert-parallelism.md", "legacy/advanced/gpt-attention.md", "legacy/advanced/gpt-runtime.md", "legacy/advanced/graph-rewriting.md", "legacy/advanced/kv-cache-management.md", "legacy/advanced/kv-cache-reuse.md", "legacy/advanced/lora.md", "legacy/advanced/lowprecision-pcie-allreduce.md", "legacy/advanced/open-sourced-cutlass-kernels.md", "legacy/advanced/speculative-decoding.md", "legacy/advanced/weight-streaming.md", "legacy/architecture/add-model.md", "legacy/architecture/checkpoint.md", "legacy/architecture/core-concepts.md", "legacy/architecture/model-weights-loader.md", "legacy/architecture/workflow.md", "legacy/dev-on-cloud/build-image-to-dockerhub.md", "legacy/dev-on-cloud/dev-on-runpod.md", "legacy/key-features.md", "legacy/performance/perf-analysis.md", "legacy/performance/perf-benchmarking.md", "legacy/performance/performance-tuning-guide/benchmarking-default-performance.md", "legacy/performance/performance-tuning-guide/deciding-model-sharding-strategy.md", "legacy/performance/performance-tuning-guide/fp8-quantization.md", "legacy/performance/performance-tuning-guide/index.rst", "legacy/performance/performance-tuning-guide/introduction.md", "legacy/performance/performance-tuning-guide/tuning-max-batch-size-and-max-num-tokens.md", "legacy/performance/performance-tuning-guide/useful-build-time-flags.md", "legacy/performance/performance-tuning-guide/useful-runtime-flags.md", "legacy/python-api/tensorrt_llm.functional.rst", "legacy/python-api/tensorrt_llm.layers.rst", "legacy/python-api/tensorrt_llm.models.rst", "legacy/python-api/tensorrt_llm.plugin.rst", "legacy/python-api/tensorrt_llm.quantization.rst", "legacy/python-api/tensorrt_llm.runtime.rst", "legacy/reference/memory.md", "legacy/reference/multimodal-feature-support-matrix.md", "legacy/reference/precision.md", "legacy/reference/support-matrix.md", "legacy/reference/troubleshooting.md", "legacy/tensorrt_quickstart.md", "legacy/torch.md", "llm-api/index.md", "llm-api/reference.rst", "models/adding-new-model.md", "models/supported-models.md", "overview.md", "quick-start-guide.md", "release-notes.md", "torch/adding_new_model.md", "torch/arch_overview.md", "torch/attention.md", "torch/auto_deploy/advanced/benchmarking_with_trtllm_bench.md", "torch/auto_deploy/advanced/example_run.md", "torch/auto_deploy/advanced/expert_configurations.md", "torch/auto_deploy/advanced/logging.md", "torch/auto_deploy/advanced/serving_with_trtllm_serve.md", "torch/auto_deploy/advanced/workflow.md", "torch/auto_deploy/auto-deploy.md", "torch/auto_deploy/support_matrix.md", "torch/features/checkpoint_loading.md", "torch/features/lora.md", "torch/features/overlap_scheduler.md", "torch/features/quantization.md", "torch/features/sampling.md", "torch/kv_cache_manager.md", "torch/scheduler.md"], "indexentries": {"--accuracy_threshold": [[24, "cmdoption-trtllm-eval-mmlu-accuracy_threshold", false]], "--apply_chat_template": [[24, "cmdoption-trtllm-eval-cnn_dailymail-apply_chat_template", false], [24, "cmdoption-trtllm-eval-gpqa_diamond-apply_chat_template", false], [24, "cmdoption-trtllm-eval-gpqa_extended-apply_chat_template", false], [24, "cmdoption-trtllm-eval-gpqa_main-apply_chat_template", false], [24, "cmdoption-trtllm-eval-gsm8k-apply_chat_template", false], [24, "cmdoption-trtllm-eval-longbench_v2-apply_chat_template", false], [24, "cmdoption-trtllm-eval-mmlu-apply_chat_template", false]], "--backend": [[22, "cmdoption-trtllm-bench-latency-backend", false], [22, "cmdoption-trtllm-bench-throughput-backend", false], [24, "cmdoption-trtllm-eval-backend", false], [27, "cmdoption-trtllm-serve-serve-backend", false]], "--beam_width": [[22, "cmdoption-trtllm-bench-latency-beam_width", false], [22, "cmdoption-trtllm-bench-throughput-beam_width", false]], "--chat_template_kwargs": [[24, "cmdoption-trtllm-eval-gpqa_diamond-chat_template_kwargs", false], [24, "cmdoption-trtllm-eval-gpqa_extended-chat_template_kwargs", false], [24, "cmdoption-trtllm-eval-gpqa_main-chat_template_kwargs", false], [24, "cmdoption-trtllm-eval-gsm8k-chat_template_kwargs", false], [24, "cmdoption-trtllm-eval-mmlu-chat_template_kwargs", false], [24, "cmdoption-trtllm-eval-mmmu-chat_template_kwargs", false]], "--check_accuracy": [[24, "cmdoption-trtllm-eval-mmlu-check_accuracy", false]], "--cluster_size": [[22, "cmdoption-trtllm-bench-throughput-cluster_size", false], [27, "cmdoption-trtllm-serve-serve-cluster_size", false]], "--concurrency": [[22, "cmdoption-trtllm-bench-latency-concurrency", false], [22, "cmdoption-trtllm-bench-throughput-concurrency", false]], "--config_file": [[27, "cmdoption-trtllm-serve-disaggregated-c", false], [27, "cmdoption-trtllm-serve-disaggregated_mpi_worker-c", false]], "--cot": [[24, "cmdoption-trtllm-eval-longbench_v2-cot", false]], "--custom_module_dirs": [[22, "cmdoption-trtllm-bench-throughput-custom_module_dirs", false]], "--data_device": [[22, "cmdoption-trtllm-bench-throughput-data_device", false]], "--dataset": [[22, "cmdoption-trtllm-bench-build-dataset", false], [22, "cmdoption-trtllm-bench-latency-dataset", false], [22, "cmdoption-trtllm-bench-throughput-dataset", false]], "--dataset_path": [[24, "cmdoption-trtllm-eval-cnn_dailymail-dataset_path", false], [24, "cmdoption-trtllm-eval-gpqa_diamond-dataset_path", false], [24, "cmdoption-trtllm-eval-gpqa_extended-dataset_path", false], [24, "cmdoption-trtllm-eval-gpqa_main-dataset_path", false], [24, "cmdoption-trtllm-eval-gsm8k-dataset_path", false], [24, "cmdoption-trtllm-eval-json_mode_eval-dataset_path", false], [24, "cmdoption-trtllm-eval-longbench_v2-dataset_path", false], [24, "cmdoption-trtllm-eval-mmlu-dataset_path", false], [24, "cmdoption-trtllm-eval-mmmu-dataset_path", false]], "--difficulty": [[24, "cmdoption-trtllm-eval-longbench_v2-difficulty", false]], "--disable_chunked_context": [[22, "cmdoption-trtllm-bench-throughput-enable_chunked_context", false]], "--disable_kv_cache_reuse": [[24, "cmdoption-trtllm-eval-disable_kv_cache_reuse", false]], "--disagg_cluster_uri": [[27, "cmdoption-trtllm-serve-serve-disagg_cluster_uri", false]], "--domain": [[24, "cmdoption-trtllm-eval-longbench_v2-domain", false]], "--enable_chunked_context": [[22, "cmdoption-trtllm-bench-throughput-enable_chunked_context", false]], "--enable_chunked_prefill": [[27, "cmdoption-trtllm-serve-serve-enable_chunked_prefill", false]], "--engine_dir": [[22, "cmdoption-trtllm-bench-latency-engine_dir", false], [22, "cmdoption-trtllm-bench-throughput-engine_dir", false]], "--eos_id": [[22, "cmdoption-trtllm-bench-throughput-eos_id", false]], "--ep": [[22, "cmdoption-trtllm-bench-latency-ep", false], [22, "cmdoption-trtllm-bench-throughput-ep", false]], "--ep_size": [[24, "cmdoption-trtllm-eval-ep_size", false], [27, "cmdoption-trtllm-serve-serve-ep_size", false]], "--extra_encoder_options": [[27, "cmdoption-trtllm-serve-mm_embedding_serve-extra_encoder_options", false]], "--extra_llm_api_options": [[22, "cmdoption-trtllm-bench-latency-extra_llm_api_options", false], [22, "cmdoption-trtllm-bench-throughput-extra_llm_api_options", false], [24, "cmdoption-trtllm-eval-extra_llm_api_options", false], [27, "cmdoption-trtllm-serve-serve-extra_llm_api_options", false]], "--fail_fast_on_attention_window_too_large": [[27, "cmdoption-trtllm-serve-serve-fail_fast_on_attention_window_too_large", false]], "--fewshot_as_multiturn": [[24, "cmdoption-trtllm-eval-gsm8k-fewshot_as_multiturn", false]], "--gpus_per_node": [[24, "cmdoption-trtllm-eval-gpus_per_node", false], [27, "cmdoption-trtllm-serve-mm_embedding_serve-gpus_per_node", false], [27, "cmdoption-trtllm-serve-serve-gpus_per_node", false]], "--host": [[27, "cmdoption-trtllm-serve-mm_embedding_serve-host", false], [27, "cmdoption-trtllm-serve-serve-host", false]], "--image_data_format": [[22, "cmdoption-trtllm-bench-throughput-image_data_format", false]], "--iteration_log": [[22, "cmdoption-trtllm-bench-latency-iteration_log", false], [22, "cmdoption-trtllm-bench-throughput-iteration_log", false]], "--kv_cache_free_gpu_mem_fraction": [[22, "cmdoption-trtllm-bench-latency-kv_cache_free_gpu_mem_fraction", false], [22, "cmdoption-trtllm-bench-throughput-kv_cache_free_gpu_mem_fraction", false]], "--kv_cache_free_gpu_memory_fraction": [[24, "cmdoption-trtllm-eval-kv_cache_free_gpu_memory_fraction", false], [27, "cmdoption-trtllm-serve-serve-kv_cache_free_gpu_memory_fraction", false]], "--length": [[24, "cmdoption-trtllm-eval-longbench_v2-length", false]], "--log_level": [[22, "cmdoption-trtllm-bench-log_level", false], [24, "cmdoption-trtllm-eval-log_level", false], [27, "cmdoption-trtllm-serve-disaggregated-l", false], [27, "cmdoption-trtllm-serve-disaggregated_mpi_worker-log_level", false], [27, "cmdoption-trtllm-serve-mm_embedding_serve-log_level", false], [27, "cmdoption-trtllm-serve-serve-log_level", false]], "--max_batch_size": [[22, "cmdoption-trtllm-bench-build-max_batch_size", false], [22, "cmdoption-trtllm-bench-throughput-max_batch_size", false], [24, "cmdoption-trtllm-eval-max_batch_size", false], [27, "cmdoption-trtllm-serve-mm_embedding_serve-max_batch_size", false], [27, "cmdoption-trtllm-serve-serve-max_batch_size", false]], "--max_beam_width": [[24, "cmdoption-trtllm-eval-max_beam_width", false], [27, "cmdoption-trtllm-serve-serve-max_beam_width", false]], "--max_input_len": [[22, "cmdoption-trtllm-bench-latency-max_input_len", false], [22, "cmdoption-trtllm-bench-throughput-max_input_len", false]], "--max_input_length": [[24, "cmdoption-trtllm-eval-cnn_dailymail-max_input_length", false], [24, "cmdoption-trtllm-eval-gpqa_diamond-max_input_length", false], [24, "cmdoption-trtllm-eval-gpqa_extended-max_input_length", false], [24, "cmdoption-trtllm-eval-gpqa_main-max_input_length", false], [24, "cmdoption-trtllm-eval-gsm8k-max_input_length", false], [24, "cmdoption-trtllm-eval-json_mode_eval-max_input_length", false], [24, "cmdoption-trtllm-eval-longbench_v2-max_input_length", false], [24, "cmdoption-trtllm-eval-mmlu-max_input_length", false], [24, "cmdoption-trtllm-eval-mmmu-max_input_length", false]], "--max_len": [[24, "cmdoption-trtllm-eval-longbench_v2-max_len", false]], "--max_num_tokens": [[22, "cmdoption-trtllm-bench-build-max_num_tokens", false], [22, "cmdoption-trtllm-bench-throughput-max_num_tokens", false], [24, "cmdoption-trtllm-eval-max_num_tokens", false], [27, "cmdoption-trtllm-serve-mm_embedding_serve-max_num_tokens", false], [27, "cmdoption-trtllm-serve-serve-max_num_tokens", false]], "--max_output_length": [[24, "cmdoption-trtllm-eval-cnn_dailymail-max_output_length", false], [24, "cmdoption-trtllm-eval-gpqa_diamond-max_output_length", false], [24, "cmdoption-trtllm-eval-gpqa_extended-max_output_length", false], [24, "cmdoption-trtllm-eval-gpqa_main-max_output_length", false], [24, "cmdoption-trtllm-eval-gsm8k-max_output_length", false], [24, "cmdoption-trtllm-eval-json_mode_eval-max_output_length", false], [24, "cmdoption-trtllm-eval-longbench_v2-max_output_length", false], [24, "cmdoption-trtllm-eval-mmlu-max_output_length", false], [24, "cmdoption-trtllm-eval-mmmu-max_output_length", false]], "--max_seq_len": [[22, "cmdoption-trtllm-bench-build-max_seq_len", false], [22, "cmdoption-trtllm-bench-latency-max_seq_len", false], [22, "cmdoption-trtllm-bench-throughput-max_seq_len", false], [24, "cmdoption-trtllm-eval-max_seq_len", false], [27, "cmdoption-trtllm-serve-serve-max_seq_len", false]], "--media_io_kwargs": [[27, "cmdoption-trtllm-serve-serve-media_io_kwargs", false]], "--medusa_choices": [[22, "cmdoption-trtllm-bench-latency-medusa_choices", false]], "--metadata_server_config_file": [[27, "cmdoption-trtllm-serve-disaggregated-m", false], [27, "cmdoption-trtllm-serve-mm_embedding_serve-metadata_server_config_file", false], [27, "cmdoption-trtllm-serve-serve-metadata_server_config_file", false]], "--metrics-log-interval": [[27, "cmdoption-trtllm-serve-disaggregated-metrics-log-interval", false]], "--modality": [[22, "cmdoption-trtllm-bench-latency-modality", false], [22, "cmdoption-trtllm-bench-throughput-modality", false]], "--model": [[22, "cmdoption-trtllm-bench-m", false], [24, "cmdoption-trtllm-eval-model", false]], "--model_path": [[22, "cmdoption-trtllm-bench-model_path", false]], "--no_context": [[24, "cmdoption-trtllm-eval-longbench_v2-no_context", false]], "--no_skip_tokenizer_init": [[22, "cmdoption-trtllm-bench-throughput-no_skip_tokenizer_init", false]], "--no_weights_loading": [[22, "cmdoption-trtllm-bench-build-no_weights_loading", false]], "--num_fewshot": [[24, "cmdoption-trtllm-eval-mmlu-num_fewshot", false]], "--num_postprocess_workers": [[27, "cmdoption-trtllm-serve-serve-num_postprocess_workers", false]], "--num_requests": [[22, "cmdoption-trtllm-bench-latency-num_requests", false], [22, "cmdoption-trtllm-bench-throughput-num_requests", false]], "--num_samples": [[24, "cmdoption-trtllm-eval-cnn_dailymail-num_samples", false], [24, "cmdoption-trtllm-eval-gpqa_diamond-num_samples", false], [24, "cmdoption-trtllm-eval-gpqa_extended-num_samples", false], [24, "cmdoption-trtllm-eval-gpqa_main-num_samples", false], [24, "cmdoption-trtllm-eval-gsm8k-num_samples", false], [24, "cmdoption-trtllm-eval-json_mode_eval-num_samples", false], [24, "cmdoption-trtllm-eval-longbench_v2-num_samples", false], [24, "cmdoption-trtllm-eval-mmlu-num_samples", false], [24, "cmdoption-trtllm-eval-mmmu-num_samples", false]], "--otlp_traces_endpoint": [[27, "cmdoption-trtllm-serve-serve-otlp_traces_endpoint", false]], "--output_dir": [[24, "cmdoption-trtllm-eval-longbench_v2-output_dir", false]], "--output_json": [[22, "cmdoption-trtllm-bench-throughput-output_json", false]], "--port": [[27, "cmdoption-trtllm-serve-mm_embedding_serve-port", false], [27, "cmdoption-trtllm-serve-serve-port", false]], "--pp": [[22, "cmdoption-trtllm-bench-latency-pp", false], [22, "cmdoption-trtllm-bench-throughput-pp", false]], "--pp_size": [[22, "cmdoption-trtllm-bench-build-pp", false], [24, "cmdoption-trtllm-eval-pp_size", false], [27, "cmdoption-trtllm-serve-serve-pp_size", false]], "--prompts_dir": [[24, "cmdoption-trtllm-eval-longbench_v2-prompts_dir", false]], "--quantization": [[22, "cmdoption-trtllm-bench-build-q", false]], "--rag": [[24, "cmdoption-trtllm-eval-longbench_v2-rag", false]], "--random_seed": [[24, "cmdoption-trtllm-eval-cnn_dailymail-random_seed", false], [24, "cmdoption-trtllm-eval-gpqa_diamond-random_seed", false], [24, "cmdoption-trtllm-eval-gpqa_extended-random_seed", false], [24, "cmdoption-trtllm-eval-gpqa_main-random_seed", false], [24, "cmdoption-trtllm-eval-gsm8k-random_seed", false], [24, "cmdoption-trtllm-eval-json_mode_eval-random_seed", false], [24, "cmdoption-trtllm-eval-longbench_v2-random_seed", false], [24, "cmdoption-trtllm-eval-mmlu-random_seed", false], [24, "cmdoption-trtllm-eval-mmmu-random_seed", false]], "--reasoning_parser": [[27, "cmdoption-trtllm-serve-serve-reasoning_parser", false]], "--report_json": [[22, "cmdoption-trtllm-bench-latency-report_json", false], [22, "cmdoption-trtllm-bench-throughput-report_json", false]], "--request_json": [[22, "cmdoption-trtllm-bench-throughput-request_json", false]], "--request_timeout": [[27, "cmdoption-trtllm-serve-disaggregated-r", false]], "--rouge_path": [[24, "cmdoption-trtllm-eval-cnn_dailymail-rouge_path", false]], "--sampler_options": [[22, "cmdoption-trtllm-bench-latency-sampler_options", false], [22, "cmdoption-trtllm-bench-throughput-sampler_options", false]], "--scheduler_policy": [[22, "cmdoption-trtllm-bench-throughput-scheduler_policy", false]], "--server_role": [[27, "cmdoption-trtllm-serve-serve-server_role", false]], "--server_start_timeout": [[27, "cmdoption-trtllm-serve-disaggregated-t", false]], "--start_idx": [[24, "cmdoption-trtllm-eval-longbench_v2-start_idx", false]], "--streaming": [[22, "cmdoption-trtllm-bench-throughput-streaming", false]], "--system_prompt": [[24, "cmdoption-trtllm-eval-cnn_dailymail-system_prompt", false], [24, "cmdoption-trtllm-eval-gpqa_diamond-system_prompt", false], [24, "cmdoption-trtllm-eval-gpqa_extended-system_prompt", false], [24, "cmdoption-trtllm-eval-gpqa_main-system_prompt", false], [24, "cmdoption-trtllm-eval-gsm8k-system_prompt", false], [24, "cmdoption-trtllm-eval-json_mode_eval-system_prompt", false], [24, "cmdoption-trtllm-eval-longbench_v2-system_prompt", false], [24, "cmdoption-trtllm-eval-mmlu-system_prompt", false], [24, "cmdoption-trtllm-eval-mmmu-system_prompt", false]], "--target_input_len": [[22, "cmdoption-trtllm-bench-build-target_input_len", false], [22, "cmdoption-trtllm-bench-throughput-target_input_len", false]], "--target_output_len": [[22, "cmdoption-trtllm-bench-build-target_output_len", false], [22, "cmdoption-trtllm-bench-throughput-target_output_len", false]], "--tokenizer": [[24, "cmdoption-trtllm-eval-tokenizer", false], [27, "cmdoption-trtllm-serve-serve-tokenizer", false]], "--tool_parser": [[27, "cmdoption-trtllm-serve-serve-tool_parser", false]], "--tp": [[22, "cmdoption-trtllm-bench-latency-tp", false], [22, "cmdoption-trtllm-bench-throughput-tp", false]], "--tp_size": [[22, "cmdoption-trtllm-bench-build-tp", false], [24, "cmdoption-trtllm-eval-tp_size", false], [27, "cmdoption-trtllm-serve-serve-tp_size", false]], "--trust_remote_code": [[22, "cmdoption-trtllm-bench-build-trust_remote_code", false], [24, "cmdoption-trtllm-eval-trust_remote_code", false], [27, "cmdoption-trtllm-serve-mm_embedding_serve-trust_remote_code", false], [27, "cmdoption-trtllm-serve-serve-trust_remote_code", false]], "--warmup": [[22, "cmdoption-trtllm-bench-latency-warmup", false], [22, "cmdoption-trtllm-bench-throughput-warmup", false]], "--workspace": [[22, "cmdoption-trtllm-bench-w", false]], "-c": [[27, "cmdoption-trtllm-serve-disaggregated-c", false], [27, "cmdoption-trtllm-serve-disaggregated_mpi_worker-c", false]], "-l": [[27, "cmdoption-trtllm-serve-disaggregated-l", false]], "-m": [[22, "cmdoption-trtllm-bench-m", false], [27, "cmdoption-trtllm-serve-disaggregated-m", false]], "-pp": [[22, "cmdoption-trtllm-bench-build-pp", false]], "-q": [[22, "cmdoption-trtllm-bench-build-q", false]], "-r": [[27, "cmdoption-trtllm-serve-disaggregated-r", false]], "-t": [[27, "cmdoption-trtllm-serve-disaggregated-t", false]], "-tp": [[22, "cmdoption-trtllm-bench-build-tp", false]], "-w": [[22, "cmdoption-trtllm-bench-w", false]], "__init__ (tensorrt_llm.llmapi.kvcacheretentionconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.__init__", false]], "__init__ (tensorrt_llm.llmapi.kvcacheretentionconfig.tokenrangeretentionconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.TokenRangeRetentionConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.__init__", false]], "__init__() (tensorrt_llm.llmapi.buildcacheconfig method)": [[150, "tensorrt_llm.llmapi.BuildCacheConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.__init__", false]], "__init__() (tensorrt_llm.llmapi.completionoutput method)": [[150, "tensorrt_llm.llmapi.CompletionOutput.__init__", false]], "__init__() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.__init__", false]], "__init__() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.disaggregatedparams method)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.__init__", false]], "__init__() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.guideddecodingparams method)": [[150, "tensorrt_llm.llmapi.GuidedDecodingParams.__init__", false]], "__init__() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.__init__", false]], "__init__() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.lorarequest method)": [[150, "tensorrt_llm.llmapi.LoRARequest.__init__", false]], "__init__() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.__init__", false]], "__init__() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.__init__", false]], "__init__() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.__init__", false]], "__init__() (tensorrt_llm.llmapi.quantconfig method)": [[150, "tensorrt_llm.llmapi.QuantConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.requesterror method)": [[150, "tensorrt_llm.llmapi.RequestError.__init__", false]], "__init__() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.__init__", false]], "__init__() (tensorrt_llm.llmapi.requestoutput.postprocworker method)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.__init__", false]], "__init__() (tensorrt_llm.llmapi.requestoutput.postprocworker.input method)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Input.__init__", false]], "__init__() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.samplingparams method)": [[150, "tensorrt_llm.llmapi.SamplingParams.__init__", false]], "__init__() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.__init__", false]], "__init__() (tensorrt_llm.llmapi.torchllmargs method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.__init__", false]], "__init__() (tensorrt_llm.llmapi.trtllmargs method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.__init__", false]], "__init__() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.__init__", false]], "abort() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.abort", false]], "abort() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.abort", false]], "aborted() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.aborted", false]], "abs() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.abs", false]], "abs() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.abs", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.acceptance_length_threshold", false]], "acceptance_length_threshold (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.acceptance_length_threshold", false]], "acceptance_window (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.acceptance_window", false]], "acceptance_window (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.acceptance_window", false]], "activation() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.activation", false]], "adalayernorm (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNorm", false]], "adalayernormcontinuous (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNormContinuous", false]], "adalayernormzero (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNormZero", false]], "adalayernormzerosingle (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNormZeroSingle", false]], "adapter_id (tensorrt_llm.llmapi.lorarequest property)": [[150, "tensorrt_llm.llmapi.LoRARequest.adapter_id", false]], "add() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.add", false]], "add_input() (tensorrt_llm.functional.conditional method)": [[136, "tensorrt_llm.functional.Conditional.add_input", false]], "add_note() (tensorrt_llm.llmapi.requesterror method)": [[150, "tensorrt_llm.llmapi.RequestError.add_note", false]], "add_output() (tensorrt_llm.functional.conditional method)": [[136, "tensorrt_llm.functional.Conditional.add_output", false]], "add_sequence() (tensorrt_llm.runtime.kvcachemanager method)": [[141, "tensorrt_llm.runtime.KVCacheManager.add_sequence", false]], "add_special_tokens (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.add_special_tokens", false]], "additional_context_outputs (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.additional_context_outputs", false]], "additional_generation_outputs (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.additional_generation_outputs", false]], "additional_model_outputs (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.additional_model_outputs", false]], "algorithm (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.algorithm", false]], "algorithm (tensorrt_llm.llmapi.rocketsparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.algorithm", false]], "alibi (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.alibi", false]], "alibi_with_scale (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.alibi_with_scale", false]], "allgather() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.allgather", false]], "allreduce() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.allreduce", false]], "allreduce_strategy (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.allreduce_strategy", false]], "allreducefusionop (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.AllReduceFusionOp", false]], "allreduceparams (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.AllReduceParams", false]], "allreducestrategy (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.AllReduceStrategy", false]], "apply_batched_logits_processor (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.apply_batched_logits_processor", false]], "apply_llama3_scaling() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.apply_llama3_scaling", false]], "apply_rotary_pos_emb() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.apply_rotary_pos_emb", false]], "apply_rotary_pos_emb_chatglm() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.apply_rotary_pos_emb_chatglm", false]], "apply_rotary_pos_emb_cogvlm() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.apply_rotary_pos_emb_cogvlm", false]], "arange() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.arange", false]], "aresult() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.aresult", false]], "argmax() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.argmax", false]], "args (tensorrt_llm.llmapi.requesterror attribute)": [[150, "tensorrt_llm.llmapi.RequestError.args", false]], "assert_valid_quant_algo() (tensorrt_llm.models.gemmaforcausallm class method)": [[138, "tensorrt_llm.models.GemmaForCausalLM.assert_valid_quant_algo", false]], "assertion() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.assertion", false]], "attention (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.Attention", false]], "attention_dp_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.attention_dp_config", false]], "attention_dp_events_gather_period_ms (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.attention_dp_events_gather_period_ms", false]], "attentiondpconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig", false]], "attentiondpconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.Config", false]], "attentionmaskparams (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.AttentionMaskParams", false]], "attentionmasktype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.AttentionMaskType", false]], "attentionparams (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.AttentionParams", false]], "attn_backend (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.attn_backend", false]], "attn_processors (tensorrt_llm.models.sd3transformer2dmodel property)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.attn_processors", false]], "audio_engine_dir (tensorrt_llm.runtime.multimodalmodelrunner property)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.audio_engine_dir", false]], "auto (tensorrt_llm.functional.allreducestrategy attribute)": [[136, "tensorrt_llm.functional.AllReduceStrategy.AUTO", false]], "auto (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.AUTO", false]], "autodecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig", false]], "autodecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.Config", false]], "avg_pool2d() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.avg_pool2d", false]], "avgpool2d (class in tensorrt_llm.layers.pooling)": [[137, "tensorrt_llm.layers.pooling.AvgPool2d", false]], "axes (tensorrt_llm.functional.sliceinputtype attribute)": [[136, "tensorrt_llm.functional.SliceInputType.axes", false]], "backend (tensorrt_llm.llmapi.cachetransceiverconfig attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.backend", false]], "backend (tensorrt_llm.llmapi.moeconfig attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.backend", false]], "backend (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.backend", false]], "backend (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.backend", false]], "bad (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.bad", false]], "bad_token_ids (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.bad_token_ids", false]], "bad_words_list (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.bad_words_list", false]], "baichuanforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.BaichuanForCausalLM", false]], "batch_size (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.batch_size", false]], "batch_sizes (tensorrt_llm.llmapi.cudagraphconfig attribute)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.batch_sizes", false]], "batch_wait_max_tokens_ratio (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.batch_wait_max_tokens_ratio", false]], "batch_wait_timeout_iters (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.batch_wait_timeout_iters", false]], "batch_wait_timeout_ms (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.batch_wait_timeout_ms", false]], "batched_logits_processor (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.batched_logits_processor", false]], "batched_logits_processor (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.batched_logits_processor", false]], "batching_type (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.batching_type", false]], "batching_wait_iters (tensorrt_llm.llmapi.attentiondpconfig attribute)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.batching_wait_iters", false]], "batchingtype (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.BatchingType", false]], "beam_search_diversity_rate (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.beam_search_diversity_rate", false]], "beam_search_diversity_rate (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.beam_search_diversity_rate", false]], "beam_width_array (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.beam_width_array", false]], "begin_thinking_phase_token (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.begin_thinking_phase_token", false]], "bert_attention() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.bert_attention", false]], "bert_attention_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.bert_attention_plugin", false]], "bert_context_fmha_fp32_acc (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.bert_context_fmha_fp32_acc", false]], "bertattention (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.BertAttention", false]], "bertforquestionanswering (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.BertForQuestionAnswering", false]], "bertforsequenceclassification (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.BertForSequenceClassification", false]], "bertmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.BertModel", false]], "best_of (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.best_of", false]], "bidirectional (tensorrt_llm.functional.attentionmasktype attribute)": [[136, "tensorrt_llm.functional.AttentionMaskType.bidirectional", false]], "bidirectionalglm (tensorrt_llm.functional.attentionmasktype attribute)": [[136, "tensorrt_llm.functional.AttentionMaskType.bidirectionalglm", false]], "blocksparse (tensorrt_llm.functional.attentionmasktype attribute)": [[136, "tensorrt_llm.functional.AttentionMaskType.blocksparse", false]], "blocksparseattnparams (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.BlockSparseAttnParams", false]], "bloomforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.BloomForCausalLM", false]], "bloommodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.BloomModel", false]], "broadcast_helper() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.broadcast_helper", false]], "buffer_allocated (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.buffer_allocated", false]], "build_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.build_config", false]], "build_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.build_config", false]], "buildcacheconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.BuildCacheConfig", false]], "buildconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.BuildConfig", false]], "cache_root (tensorrt_llm.llmapi.buildcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildCacheConfig.cache_root", false]], "cache_root (tensorrt_llm.llmapi.buildcacheconfig property)": [[150, "id13", false]], "cache_transceiver_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.cache_transceiver_config", false]], "cache_transceiver_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.cache_transceiver_config", false]], "cachetransceiverconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig", false]], "cachetransceiverconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.Config", false]], "calculate_speculative_resource() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.calculate_speculative_resource", false]], "calib_batch_size (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.calib_batch_size", false]], "calib_batches (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.calib_batches", false]], "calib_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.calib_config", false]], "calib_dataset (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.calib_dataset", false]], "calib_max_seq_length (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.calib_max_seq_length", false]], "calibconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CalibConfig", false]], "calibconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CalibConfig.Config", false]], "capacity_scheduler_policy (tensorrt_llm.llmapi.schedulerconfig attribute)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.capacity_scheduler_policy", false]], "capacityschedulerpolicy (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy", false]], "capitalize() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.capitalize", false]], "capitalize() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.capitalize", false]], "capitalize() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.capitalize", false]], "capitalize() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.capitalize", false]], "capture_num_tokens (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.capture_num_tokens", false]], "casefold() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.casefold", false]], "casefold() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.casefold", false]], "casefold() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.casefold", false]], "casefold() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.casefold", false]], "cast (class in tensorrt_llm.layers.cast)": [[137, "tensorrt_llm.layers.cast.Cast", false]], "cast() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.cast", false]], "cast() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.cast", false]], "categorical_sample() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.categorical_sample", false]], "causal (tensorrt_llm.functional.attentionmasktype attribute)": [[136, "tensorrt_llm.functional.AttentionMaskType.causal", false]], "center() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.center", false]], "center() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.center", false]], "center() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.center", false]], "center() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.center", false]], "chatglm (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.chatglm", false]], "chatglmconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.ChatGLMConfig", false]], "chatglmforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.ChatGLMForCausalLM", false]], "chatglmgenerationsession (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.ChatGLMGenerationSession", false]], "chatglmmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.ChatGLMModel", false]], "check_config() (tensorrt_llm.models.decodermodel method)": [[138, "tensorrt_llm.models.DecoderModel.check_config", false]], "check_config() (tensorrt_llm.models.dit method)": [[138, "tensorrt_llm.models.DiT.check_config", false]], "check_config() (tensorrt_llm.models.encodermodel method)": [[138, "tensorrt_llm.models.EncoderModel.check_config", false]], "check_config() (tensorrt_llm.models.falconforcausallm method)": [[138, "tensorrt_llm.models.FalconForCausalLM.check_config", false]], "check_config() (tensorrt_llm.models.mptforcausallm method)": [[138, "tensorrt_llm.models.MPTForCausalLM.check_config", false]], "check_config() (tensorrt_llm.models.optforcausallm method)": [[138, "tensorrt_llm.models.OPTForCausalLM.check_config", false]], "check_config() (tensorrt_llm.models.phiforcausallm method)": [[138, "tensorrt_llm.models.PhiForCausalLM.check_config", false]], "check_config() (tensorrt_llm.models.pretrainedmodel method)": [[138, "tensorrt_llm.models.PretrainedModel.check_config", false]], "check_eagle_choices() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.check_eagle_choices", false]], "checkpoint_format (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.checkpoint_format", false]], "checkpoint_loader (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.checkpoint_loader", false]], "choices() (tensorrt_llm.functional.positionembeddingtype static method)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.choices", false]], "chunk() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.chunk", false]], "ckpt_source (tensorrt_llm.llmapi.lorarequest property)": [[150, "tensorrt_llm.llmapi.LoRARequest.ckpt_source", false]], "clamp_val (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.clamp_val", false]], "clear_logprob_params() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.clear_logprob_params", false]], "client_id (tensorrt_llm.llmapi.requestoutput.postprocworker.output attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.client_id", false]], "clip() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.clip", false]], "clipvisiontransformer (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.CLIPVisionTransformer", false]], "cogvlmattention (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.CogVLMAttention", false]], "cogvlmconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.CogVLMConfig", false]], "cogvlmforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.CogVLMForCausalLM", false]], "cohereforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.CohereForCausalLM", false]], "collect_and_bias() (tensorrt_llm.layers.linear.linear method)": [[137, "tensorrt_llm.layers.linear.Linear.collect_and_bias", false]], "collect_and_bias() (tensorrt_llm.layers.linear.linearbase method)": [[137, "tensorrt_llm.layers.linear.LinearBase.collect_and_bias", false]], "collect_and_bias() (tensorrt_llm.layers.linear.rowlinear method)": [[137, "tensorrt_llm.layers.linear.RowLinear.collect_and_bias", false]], "columnlinear (in module tensorrt_llm.layers.linear)": [[137, "tensorrt_llm.layers.linear.ColumnLinear", false]], "combinedtimesteplabelembeddings (class in tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.CombinedTimestepLabelEmbeddings", false]], "combinedtimesteptextprojembeddings (class in tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.CombinedTimestepTextProjEmbeddings", false]], "completionoutput (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CompletionOutput", false]], "compute_relative_bias() (in module tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.compute_relative_bias", false]], "concat() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.concat", false]], "conditional (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.Conditional", false]], "config_class (tensorrt_llm.models.baichuanforcausallm attribute)": [[138, "tensorrt_llm.models.BaichuanForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.chatglmforcausallm attribute)": [[138, "tensorrt_llm.models.ChatGLMForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.cogvlmforcausallm attribute)": [[138, "tensorrt_llm.models.CogVLMForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.cohereforcausallm attribute)": [[138, "tensorrt_llm.models.CohereForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.dbrxforcausallm attribute)": [[138, "tensorrt_llm.models.DbrxForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.deepseekforcausallm attribute)": [[138, "tensorrt_llm.models.DeepseekForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.deepseekv2forcausallm attribute)": [[138, "tensorrt_llm.models.DeepseekV2ForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.eagleforcausallm attribute)": [[138, "tensorrt_llm.models.EagleForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.falconforcausallm attribute)": [[138, "tensorrt_llm.models.FalconForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.gemmaforcausallm attribute)": [[138, "tensorrt_llm.models.GemmaForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.gptforcausallm attribute)": [[138, "tensorrt_llm.models.GPTForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.gptjforcausallm attribute)": [[138, "tensorrt_llm.models.GPTJForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.llamaforcausallm attribute)": [[138, "tensorrt_llm.models.LLaMAForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.mambaforcausallm attribute)": [[138, "tensorrt_llm.models.MambaForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.medusaforcausallm attribute)": [[138, "tensorrt_llm.models.MedusaForCausalLm.config_class", false]], "config_class (tensorrt_llm.models.mllamaforcausallm attribute)": [[138, "tensorrt_llm.models.MLLaMAForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.phi3forcausallm attribute)": [[138, "tensorrt_llm.models.Phi3ForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.phiforcausallm attribute)": [[138, "tensorrt_llm.models.PhiForCausalLM.config_class", false]], "config_class (tensorrt_llm.models.sd3transformer2dmodel attribute)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.config_class", false]], "constant() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.constant", false]], "constant_to_tensor_() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.constant_to_tensor_", false]], "constants_to_tensors_() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.constants_to_tensors_", false]], "construct() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.construct", false]], "construct() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.construct", false]], "construct() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.construct", false]], "construct() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.construct", false]], "construct() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.construct", false]], "construct() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.construct", false]], "construct() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.construct", false]], "construct() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.construct", false]], "construct() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.construct", false]], "construct() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.construct", false]], "construct() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.construct", false]], "construct() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.construct", false]], "construct() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.construct", false]], "construct() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.construct", false]], "construct() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.construct", false]], "context (tensorrt_llm.runtime.session property)": [[141, "tensorrt_llm.runtime.Session.context", false]], "context_chunking_policy (tensorrt_llm.llmapi.schedulerconfig attribute)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.context_chunking_policy", false]], "context_fmha (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.context_fmha", false]], "context_fmha_type (tensorrt_llm.plugin.pluginconfig property)": [[139, "tensorrt_llm.plugin.PluginConfig.context_fmha_type", false]], "context_logits (tensorrt_llm.llmapi.requestoutput attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.context_logits", false]], "context_logits (tensorrt_llm.llmapi.requestoutput property)": [[150, "id6", false]], "context_mem_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.context_mem_size", false]], "context_mem_size (tensorrt_llm.runtime.session property)": [[141, "tensorrt_llm.runtime.Session.context_mem_size", false]], "context_parallel_size (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.context_parallel_size", false]], "context_parallel_size (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.context_parallel_size", false]], "contextchunkingpolicy (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy", false]], "conv1d (class in tensorrt_llm.layers.conv)": [[137, "tensorrt_llm.layers.conv.Conv1d", false]], "conv1d() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.conv1d", false]], "conv2d (class in tensorrt_llm.layers.conv)": [[137, "tensorrt_llm.layers.conv.Conv2d", false]], "conv2d() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.conv2d", false]], "conv3d (class in tensorrt_llm.layers.conv)": [[137, "tensorrt_llm.layers.conv.Conv3d", false]], "conv3d() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.conv3d", false]], "conv_kernel (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.conv_kernel", false]], "conv_kernel (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.conv_kernel", false]], "conv_transpose2d() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.conv_transpose2d", false]], "convert_enable_disable() (tensorrt_llm.plugin.pluginconfig class method)": [[139, "tensorrt_llm.plugin.PluginConfig.convert_enable_disable", false]], "convert_load_format() (tensorrt_llm.llmapi.torchllmargs class method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.convert_load_format", false]], "convtranspose2d (class in tensorrt_llm.layers.conv)": [[137, "tensorrt_llm.layers.conv.ConvTranspose2d", false]], "copy() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.copy", false]], "copy() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.copy", false]], "copy() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.copy", false]], "copy() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.copy", false]], "copy() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.copy", false]], "copy() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.copy", false]], "copy() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.copy", false]], "copy() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.copy", false]], "copy() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.copy", false]], "copy() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.copy", false]], "copy() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.copy", false]], "copy() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.copy", false]], "copy() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.copy", false]], "copy() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.copy", false]], "copy() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.copy", false]], "copy_on_partial_reuse (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.copy_on_partial_reuse", false]], "cos() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.cos", false]], "count() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.count", false]], "count() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.count", false]], "count() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.count", false]], "count() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.count", false]], "count() (tensorrt_llm.llmapi.requestoutput.postprocworker.output method)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.count", false]], "cp_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.cp_config", false]], "cp_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.cp_config", false]], "cp_split_plugin() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.cp_split_plugin", false]], "cpp_e2e (tensorrt_llm.runtime.multimodalmodelrunner property)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.cpp_e2e", false]], "cpp_llm_only (tensorrt_llm.runtime.multimodalmodelrunner property)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.cpp_llm_only", false]], "create_allreduce_plugin() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.create_allreduce_plugin", false]], "create_attention_const_params() (tensorrt_llm.layers.attention.attention static method)": [[137, "tensorrt_llm.layers.attention.Attention.create_attention_const_params", false]], "create_fake_weight() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.create_fake_weight", false]], "create_runtime_defaults() (tensorrt_llm.models.pretrainedconfig static method)": [[138, "tensorrt_llm.models.PretrainedConfig.create_runtime_defaults", false]], "create_sinusoidal_positions() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.create_sinusoidal_positions", false]], "create_sinusoidal_positions_for_attention_plugin() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.create_sinusoidal_positions_for_attention_plugin", false]], "create_sinusoidal_positions_for_cogvlm_attention_plugin() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.create_sinusoidal_positions_for_cogvlm_attention_plugin", false]], "create_sinusoidal_positions_long_rope() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.create_sinusoidal_positions_long_rope", false]], "create_sinusoidal_positions_long_rope_for_attention_plugin() (tensorrt_llm.functional.ropeembeddingutils method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.create_sinusoidal_positions_long_rope_for_attention_plugin", false]], "create_sinusoidal_positions_yarn() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.create_sinusoidal_positions_yarn", false]], "cropped_pos_embed() (tensorrt_llm.layers.embedding.sd3patchembed method)": [[137, "tensorrt_llm.layers.embedding.SD3PatchEmbed.cropped_pos_embed", false]], "cross_attention (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.cross_attention", false]], "cross_attention (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.cross_attention", false]], "cross_kv_cache_fraction (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.cross_kv_cache_fraction", false]], "ctx_request_id (tensorrt_llm.llmapi.disaggregatedparams attribute)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.ctx_request_id", false]], "cuda_graph_cache_size (tensorrt_llm.llmapi.extendedruntimeperfknobconfig attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.cuda_graph_cache_size", false]], "cuda_graph_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.cuda_graph_config", false]], "cuda_graph_mode (tensorrt_llm.llmapi.extendedruntimeperfknobconfig attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.cuda_graph_mode", false]], "cuda_graph_mode (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.cuda_graph_mode", false]], "cuda_stream_guard() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.cuda_stream_guard", false]], "cuda_stream_sync() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.cuda_stream_sync", false]], "cudagraphconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig", false]], "cudagraphconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.Config", false]], "cumsum() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.cumsum", false]], "cumulative_logprob (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.cumulative_logprob", false]], "custom_mask (tensorrt_llm.functional.attentionmasktype attribute)": [[136, "tensorrt_llm.functional.AttentionMaskType.custom_mask", false]], "data (tensorrt_llm.functional.sliceinputtype attribute)": [[136, "tensorrt_llm.functional.SliceInputType.data", false]], "dbrxconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.DbrxConfig", false]], "dbrxforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.DbrxForCausalLM", false]], "debug_mode (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.debug_mode", false]], "debug_tensors_to_save (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.debug_tensors_to_save", false]], "decode() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.decode", false]], "decode_batch() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.decode_batch", false]], "decode_duration_ms (tensorrt_llm.llmapi.kvcacheretentionconfig property)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.decode_duration_ms", false]], "decode_regular() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.decode_regular", false]], "decode_retention_priority (tensorrt_llm.llmapi.kvcacheretentionconfig property)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.decode_retention_priority", false]], "decode_stream() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.decode_stream", false]], "decode_words_list() (in module tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.decode_words_list", false]], "decodermodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.DecoderModel", false]], "decoding_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.decoding_config", false]], "decoding_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.decoding_config", false]], "decoding_type (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.decoding_type", false]], "decoding_type (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.decoding_type", false]], "deepseekforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.DeepseekForCausalLM", false]], "deepseeksparseattentionconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig", false]], "deepseeksparseattentionconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.Config", false]], "deepseekv2attention (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.DeepseekV2Attention", false]], "deepseekv2forcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.DeepseekV2ForCausalLM", false]], "default_plugin_config() (tensorrt_llm.models.cogvlmforcausallm method)": [[138, "tensorrt_llm.models.CogVLMForCausalLM.default_plugin_config", false]], "default_plugin_config() (tensorrt_llm.models.llamaforcausallm method)": [[138, "tensorrt_llm.models.LLaMAForCausalLM.default_plugin_config", false]], "default_record_creator() (tensorrt_llm.llmapi.requestoutput.postprocworker static method)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.default_record_creator", false]], "deferred (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.deferred", false]], "detokenize (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.detokenize", false]], "device (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.device", false]], "device (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.device", false]], "dict() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.dict", false]], "dict() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.dict", false]], "dict() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.dict", false]], "dict() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.dict", false]], "dict() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.dict", false]], "dict() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.dict", false]], "dict() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.dict", false]], "dict() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.dict", false]], "dict() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.dict", false]], "dict() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.dict", false]], "dict() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.dict", false]], "dict() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.dict", false]], "dict() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.dict", false]], "dict() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.dict", false]], "dict() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.dict", false]], "diffusersattention (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.DiffusersAttention", false]], "dimrange (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.DimRange", false]], "directory (tensorrt_llm.llmapi.kvcacheretentionconfig property)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.directory", false]], "disable (tensorrt_llm.functional.sidestreamidtype attribute)": [[136, "tensorrt_llm.functional.SideStreamIDType.disable", false]], "disable_finalize_fusion (tensorrt_llm.llmapi.moeconfig attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.disable_finalize_fusion", false]], "disable_forward_chunking() (tensorrt_llm.models.sd3transformer2dmodel method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.disable_forward_chunking", false]], "disable_overlap_scheduler (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.disable_overlap_scheduler", false]], "disaggregated_params (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.disaggregated_params", false]], "disaggregated_params (tensorrt_llm.llmapi.requestoutput.postprocworker.output attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.disaggregated_params", false]], "disaggregatedparams (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams", false]], "dit (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.DiT", false]], "div() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.div", false]], "do_tracing() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.do_tracing", false]], "dora_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.dora_plugin", false]], "dora_plugin() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.dora_plugin", false]], "draft_len_schedule (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.draft_len_schedule", false]], "draft_len_schedule (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.draft_len_schedule", false]], "draft_tokens (tensorrt_llm.llmapi.disaggregatedparams attribute)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.draft_tokens", false]], "draft_tokens_external (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.DRAFT_TOKENS_EXTERNAL", false]], "drafter (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.drafter", false]], "drafttargetdecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig", false]], "drafttargetdecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.Config", false]], "dry_run (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.dry_run", false]], "dtype (tensorrt_llm.functional.tensor property)": [[136, "tensorrt_llm.functional.Tensor.dtype", false]], "dtype (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.dtype", false]], "dtype (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.dtype", false]], "dtype (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.dtype", false]], "dtype (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.dtype", false]], "dtype (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.dtype", false]], "dtype (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.dtype", false]], "dtype (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.dtype", false]], "dtype (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.dtype", false]], "dtype (tensorrt_llm.runtime.tensorinfo attribute)": [[141, "tensorrt_llm.runtime.TensorInfo.dtype", false]], "dump_debug_buffers() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.dump_debug_buffers", false]], "duration_ms (tensorrt_llm.llmapi.kvcacheretentionconfig.tokenrangeretentionconfig property)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.TokenRangeRetentionConfig.duration_ms", false]], "dynamic (tensorrt_llm.functional.rotaryscalingtype attribute)": [[136, "tensorrt_llm.functional.RotaryScalingType.dynamic", false]], "dynamic_batch_config (tensorrt_llm.llmapi.schedulerconfig attribute)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.dynamic_batch_config", false]], "dynamic_batch_moving_average_window (tensorrt_llm.llmapi.dynamicbatchconfig attribute)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.dynamic_batch_moving_average_window", false]], "dynamic_tree_max_topk (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.dynamic_tree_max_topK", false]], "dynamicbatchconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig", false]], "dynamicbatchconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.Config", false]], "eagle (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.EAGLE", false]], "eagle3_layers_to_capture (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.eagle3_layers_to_capture", false]], "eagle3_layers_to_capture (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.eagle3_layers_to_capture", false]], "eagle3_one_model (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.eagle3_one_model", false]], "eagle_choices (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.eagle_choices", false]], "eagle_choices (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.eagle_choices", false]], "eagledecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig", false]], "eagledecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.Config", false]], "eagleforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.EagleForCausalLM", false]], "early_stop_criteria() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.early_stop_criteria", false]], "early_stopping (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.early_stopping", false]], "early_stopping (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.early_stopping", false]], "einsum() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.einsum", false]], "elementwise_binary() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.elementwise_binary", false]], "embedding (class in tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.Embedding", false]], "embedding() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.embedding", false]], "embedding_bias (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.embedding_bias", false]], "embedding_parallel_mode (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.embedding_parallel_mode", false]], "enable_attention_dp (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_attention_dp", false]], "enable_attention_dp (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.enable_attention_dp", false]], "enable_autotuner (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_autotuner", false]], "enable_balance (tensorrt_llm.llmapi.attentiondpconfig attribute)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.enable_balance", false]], "enable_batch_size_tuning (tensorrt_llm.llmapi.dynamicbatchconfig attribute)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.enable_batch_size_tuning", false]], "enable_block_reuse (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.enable_block_reuse", false]], "enable_build_cache (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.enable_build_cache", false]], "enable_chunked_prefill (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_chunked_prefill", false]], "enable_chunked_prefill (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.enable_chunked_prefill", false]], "enable_context_fmha_fp32_acc (tensorrt_llm.llmapi.extendedruntimeperfknobconfig attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.enable_context_fmha_fp32_acc", false]], "enable_debug_output (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.enable_debug_output", false]], "enable_forward_chunking() (tensorrt_llm.models.sd3transformer2dmodel method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.enable_forward_chunking", false]], "enable_fullgraph (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.enable_fullgraph", false]], "enable_inductor (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.enable_inductor", false]], "enable_iter_perf_stats (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_iter_perf_stats", false]], "enable_iter_req_stats (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_iter_req_stats", false]], "enable_layerwise_nvtx_marker (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_layerwise_nvtx_marker", false]], "enable_lm_head_tp_in_adp (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_lm_head_tp_in_adp", false]], "enable_lm_head_tp_in_adp (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.enable_lm_head_tp_in_adp", false]], "enable_lora (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_lora", false]], "enable_lora (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.enable_lora", false]], "enable_max_num_tokens_tuning (tensorrt_llm.llmapi.dynamicbatchconfig attribute)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.enable_max_num_tokens_tuning", false]], "enable_min_latency (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_min_latency", false]], "enable_padding (tensorrt_llm.llmapi.cudagraphconfig attribute)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.enable_padding", false]], "enable_paged_kv_cache() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.enable_paged_kv_cache", false]], "enable_partial_reuse (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.enable_partial_reuse", false]], "enable_piecewise_cuda_graph (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.enable_piecewise_cuda_graph", false]], "enable_prompt_adapter (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.enable_prompt_adapter", false]], "enable_sleep (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.enable_sleep", false]], "enable_tqdm (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.enable_tqdm", false]], "enable_userbuffers (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.enable_userbuffers", false]], "encdecmodelrunner (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.EncDecModelRunner", false]], "encode() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.encode", false]], "encode() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.encode", false]], "encode() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.encode", false]], "encode() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.encode", false]], "encoder_run() (tensorrt_llm.runtime.encdecmodelrunner method)": [[141, "tensorrt_llm.runtime.EncDecModelRunner.encoder_run", false]], "encodermodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.EncoderModel", false]], "end_id (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.end_id", false]], "end_id (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.end_id", false]], "end_thinking_phase_token (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.end_thinking_phase_token", false]], "endswith() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.endswith", false]], "endswith() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.endswith", false]], "endswith() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.endswith", false]], "endswith() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.endswith", false]], "engine (tensorrt_llm.runtime.session property)": [[141, "tensorrt_llm.runtime.Session.engine", false]], "engine_inspector (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.engine_inspector", false]], "eq() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.eq", false]], "equal_progress (tensorrt_llm.llmapi.contextchunkingpolicy attribute)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.EQUAL_PROGRESS", false]], "error (tensorrt_llm.llmapi.requestoutput.postprocworker.output attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.error", false]], "event_buffer_max_size (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.event_buffer_max_size", false]], "exclude_input_from_output (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.exclude_input_from_output", false]], "exclude_modules (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.exclude_modules", false]], "exp() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.exp", false]], "expand() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.expand", false]], "expand_dims() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.expand_dims", false]], "expand_dims_like() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.expand_dims_like", false]], "expand_mask() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.expand_mask", false]], "expandtabs() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.expandtabs", false]], "expandtabs() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.expandtabs", false]], "expandtabs() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.expandtabs", false]], "expandtabs() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.expandtabs", false]], "explicit_draft_tokens (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.EXPLICIT_DRAFT_TOKENS", false]], "extended_runtime_perf_knob_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.extended_runtime_perf_knob_config", false]], "extendedruntimeperfknobconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig", false]], "extendedruntimeperfknobconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.Config", false]], "extra (tensorrt_llm.llmapi.attentiondpconfig.config attribute)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.autodecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.cachetransceiverconfig.config attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.calibconfig.config attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.cudagraphconfig.config attribute)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.deepseeksparseattentionconfig.config attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.drafttargetdecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.dynamicbatchconfig.config attribute)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.eagledecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.extendedruntimeperfknobconfig.config attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.kvcacheconfig.config attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.lookaheaddecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.medusadecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.moeconfig.config attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.mtpdecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.ngramdecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.rocketsparseattentionconfig.config attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.schedulerconfig.config attribute)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.torchcompileconfig.config attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.Config.extra", false]], "extra (tensorrt_llm.llmapi.torchllmargs.config attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.Config.extra", false]], "extra (tensorrt_llm.llmapi.trtllmargs.config attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.Config.extra", false]], "extra (tensorrt_llm.llmapi.userprovideddecodingconfig.config attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.Config.extra", false]], "extra_resource_managers (tensorrt_llm.llmapi.torchllmargs property)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.extra_resource_managers", false]], "fail_fast_on_attention_window_too_large (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.fail_fast_on_attention_window_too_large", false]], "fail_fast_on_attention_window_too_large (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.fail_fast_on_attention_window_too_large", false]], "falconconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.FalconConfig", false]], "falconforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.FalconForCausalLM", false]], "falconmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.FalconModel", false]], "fast_build (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.fast_build", false]], "fc_gate() (tensorrt_llm.layers.mlp.fusedgatedmlp method)": [[137, "tensorrt_llm.layers.mlp.FusedGatedMLP.fc_gate", false]], "fc_gate_dora() (in module tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.fc_gate_dora", false]], "fc_gate_lora() (in module tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.fc_gate_lora", false]], "fc_gate_plugin() (tensorrt_llm.layers.mlp.fusedgatedmlp method)": [[137, "tensorrt_llm.layers.mlp.FusedGatedMLP.fc_gate_plugin", false]], "field_name (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.field_name", false]], "field_name (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.field_name", false]], "file_prefix (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.file_prefix", false]], "fill_attention_const_params_for_long_rope() (tensorrt_llm.layers.attention.attentionparams method)": [[137, "tensorrt_llm.layers.attention.AttentionParams.fill_attention_const_params_for_long_rope", false]], "fill_attention_const_params_for_rope() (tensorrt_llm.layers.attention.attentionparams method)": [[137, "tensorrt_llm.layers.attention.AttentionParams.fill_attention_const_params_for_rope", false]], "fill_attention_params() (tensorrt_llm.layers.attention.attention static method)": [[137, "tensorrt_llm.layers.attention.Attention.fill_attention_params", false]], "fill_none_tensor_list() (tensorrt_llm.layers.attention.keyvaluecacheparams method)": [[137, "tensorrt_llm.layers.attention.KeyValueCacheParams.fill_none_tensor_list", false]], "fill_value (tensorrt_llm.functional.sliceinputtype attribute)": [[136, "tensorrt_llm.functional.SliceInputType.fill_value", false]], "filter_medusa_logits() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.filter_medusa_logits", false]], "finalize_decoder() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.finalize_decoder", false]], "find() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.find", false]], "find() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.find", false]], "find() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.find", false]], "find() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.find", false]], "find_best_medusa_path() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.find_best_medusa_path", false]], "finish_reason (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.finish_reason", false]], "finished (tensorrt_llm.llmapi.requestoutput attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.finished", false]], "finished (tensorrt_llm.llmapi.requestoutput property)": [[150, "id7", false]], "first_come_first_served (tensorrt_llm.llmapi.contextchunkingpolicy attribute)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.FIRST_COME_FIRST_SERVED", false]], "first_gen_tokens (tensorrt_llm.llmapi.disaggregatedparams attribute)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.first_gen_tokens", false]], "first_layer (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.first_layer", false]], "flatten() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.flatten", false]], "flatten() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.flatten", false]], "flip() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.flip", false]], "floordiv() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.floordiv", false]], "fmt_dim (c macro)": [[1, "c.FMT_DIM", false]], "for_each_rank() (tensorrt_llm.models.pretrainedconfig method)": [[138, "tensorrt_llm.models.PretrainedConfig.for_each_rank", false]], "force_dynamic_quantization (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.force_dynamic_quantization", false]], "force_num_profiles (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.force_num_profiles", false]], "format() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.format", false]], "format() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.format", false]], "format() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.format", false]], "format() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.format", false]], "format_map() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.format_map", false]], "format_map() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.format_map", false]], "format_map() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.format_map", false]], "format_map() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.format_map", false]], "forward() (tensorrt_llm.layers.activation.mish method)": [[137, "tensorrt_llm.layers.activation.Mish.forward", false]], "forward() (tensorrt_llm.layers.attention.attention method)": [[137, "tensorrt_llm.layers.attention.Attention.forward", false]], "forward() (tensorrt_llm.layers.attention.bertattention method)": [[137, "tensorrt_llm.layers.attention.BertAttention.forward", false]], "forward() (tensorrt_llm.layers.attention.cogvlmattention method)": [[137, "tensorrt_llm.layers.attention.CogVLMAttention.forward", false]], "forward() (tensorrt_llm.layers.attention.deepseekv2attention method)": [[137, "tensorrt_llm.layers.attention.DeepseekV2Attention.forward", false]], "forward() (tensorrt_llm.layers.attention.diffusersattention method)": [[137, "tensorrt_llm.layers.attention.DiffusersAttention.forward", false]], "forward() (tensorrt_llm.layers.cast.cast method)": [[137, "tensorrt_llm.layers.cast.Cast.forward", false]], "forward() (tensorrt_llm.layers.conv.conv1d method)": [[137, "tensorrt_llm.layers.conv.Conv1d.forward", false]], "forward() (tensorrt_llm.layers.conv.conv2d method)": [[137, "tensorrt_llm.layers.conv.Conv2d.forward", false]], "forward() (tensorrt_llm.layers.conv.conv3d method)": [[137, "tensorrt_llm.layers.conv.Conv3d.forward", false]], "forward() (tensorrt_llm.layers.conv.convtranspose2d method)": [[137, "tensorrt_llm.layers.conv.ConvTranspose2d.forward", false]], "forward() (tensorrt_llm.layers.embedding.combinedtimesteplabelembeddings method)": [[137, "tensorrt_llm.layers.embedding.CombinedTimestepLabelEmbeddings.forward", false]], "forward() (tensorrt_llm.layers.embedding.combinedtimesteptextprojembeddings method)": [[137, "tensorrt_llm.layers.embedding.CombinedTimestepTextProjEmbeddings.forward", false]], "forward() (tensorrt_llm.layers.embedding.embedding method)": [[137, "tensorrt_llm.layers.embedding.Embedding.forward", false]], "forward() (tensorrt_llm.layers.embedding.labelembedding method)": [[137, "tensorrt_llm.layers.embedding.LabelEmbedding.forward", false]], "forward() (tensorrt_llm.layers.embedding.pixartalphatextprojection method)": [[137, "tensorrt_llm.layers.embedding.PixArtAlphaTextProjection.forward", false]], "forward() (tensorrt_llm.layers.embedding.prompttuningembedding method)": [[137, "tensorrt_llm.layers.embedding.PromptTuningEmbedding.forward", false]], "forward() (tensorrt_llm.layers.embedding.sd3patchembed method)": [[137, "tensorrt_llm.layers.embedding.SD3PatchEmbed.forward", false]], "forward() (tensorrt_llm.layers.embedding.timestepembedding method)": [[137, "tensorrt_llm.layers.embedding.TimestepEmbedding.forward", false]], "forward() (tensorrt_llm.layers.embedding.timesteps method)": [[137, "tensorrt_llm.layers.embedding.Timesteps.forward", false]], "forward() (tensorrt_llm.layers.linear.linearbase method)": [[137, "tensorrt_llm.layers.linear.LinearBase.forward", false]], "forward() (tensorrt_llm.layers.mlp.fusedgatedmlp method)": [[137, "tensorrt_llm.layers.mlp.FusedGatedMLP.forward", false]], "forward() (tensorrt_llm.layers.mlp.gatedmlp method)": [[137, "tensorrt_llm.layers.mlp.GatedMLP.forward", false]], "forward() (tensorrt_llm.layers.mlp.linearactivation method)": [[137, "tensorrt_llm.layers.mlp.LinearActivation.forward", false]], "forward() (tensorrt_llm.layers.mlp.linearapproximategelu method)": [[137, "tensorrt_llm.layers.mlp.LinearApproximateGELU.forward", false]], "forward() (tensorrt_llm.layers.mlp.lineargeglu method)": [[137, "tensorrt_llm.layers.mlp.LinearGEGLU.forward", false]], "forward() (tensorrt_llm.layers.mlp.lineargelu method)": [[137, "tensorrt_llm.layers.mlp.LinearGELU.forward", false]], "forward() (tensorrt_llm.layers.mlp.linearswiglu method)": [[137, "tensorrt_llm.layers.mlp.LinearSwiGLU.forward", false]], "forward() (tensorrt_llm.layers.mlp.mlp method)": [[137, "tensorrt_llm.layers.mlp.MLP.forward", false]], "forward() (tensorrt_llm.layers.normalization.adalayernorm method)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNorm.forward", false]], "forward() (tensorrt_llm.layers.normalization.adalayernormcontinuous method)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNormContinuous.forward", false]], "forward() (tensorrt_llm.layers.normalization.adalayernormzero method)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNormZero.forward", false]], "forward() (tensorrt_llm.layers.normalization.adalayernormzerosingle method)": [[137, "tensorrt_llm.layers.normalization.AdaLayerNormZeroSingle.forward", false]], "forward() (tensorrt_llm.layers.normalization.groupnorm method)": [[137, "tensorrt_llm.layers.normalization.GroupNorm.forward", false]], "forward() (tensorrt_llm.layers.normalization.layernorm method)": [[137, "tensorrt_llm.layers.normalization.LayerNorm.forward", false]], "forward() (tensorrt_llm.layers.normalization.rmsnorm method)": [[137, "tensorrt_llm.layers.normalization.RmsNorm.forward", false]], "forward() (tensorrt_llm.layers.normalization.sd35adalayernormzerox method)": [[137, "tensorrt_llm.layers.normalization.SD35AdaLayerNormZeroX.forward", false]], "forward() (tensorrt_llm.layers.pooling.avgpool2d method)": [[137, "tensorrt_llm.layers.pooling.AvgPool2d.forward", false]], "forward() (tensorrt_llm.models.bertforquestionanswering method)": [[138, "tensorrt_llm.models.BertForQuestionAnswering.forward", false]], "forward() (tensorrt_llm.models.bertforsequenceclassification method)": [[138, "tensorrt_llm.models.BertForSequenceClassification.forward", false]], "forward() (tensorrt_llm.models.bertmodel method)": [[138, "tensorrt_llm.models.BertModel.forward", false]], "forward() (tensorrt_llm.models.bloommodel method)": [[138, "tensorrt_llm.models.BloomModel.forward", false]], "forward() (tensorrt_llm.models.chatglmmodel method)": [[138, "tensorrt_llm.models.ChatGLMModel.forward", false]], "forward() (tensorrt_llm.models.clipvisiontransformer method)": [[138, "tensorrt_llm.models.CLIPVisionTransformer.forward", false]], "forward() (tensorrt_llm.models.decodermodel method)": [[138, "tensorrt_llm.models.DecoderModel.forward", false]], "forward() (tensorrt_llm.models.dit method)": [[138, "tensorrt_llm.models.DiT.forward", false]], "forward() (tensorrt_llm.models.eagleforcausallm method)": [[138, "tensorrt_llm.models.EagleForCausalLM.forward", false]], "forward() (tensorrt_llm.models.encodermodel method)": [[138, "tensorrt_llm.models.EncoderModel.forward", false]], "forward() (tensorrt_llm.models.falconmodel method)": [[138, "tensorrt_llm.models.FalconModel.forward", false]], "forward() (tensorrt_llm.models.gptjmodel method)": [[138, "tensorrt_llm.models.GPTJModel.forward", false]], "forward() (tensorrt_llm.models.gptmodel method)": [[138, "tensorrt_llm.models.GPTModel.forward", false]], "forward() (tensorrt_llm.models.gptneoxmodel method)": [[138, "tensorrt_llm.models.GPTNeoXModel.forward", false]], "forward() (tensorrt_llm.models.llamamodel method)": [[138, "tensorrt_llm.models.LLaMAModel.forward", false]], "forward() (tensorrt_llm.models.llavanextvisionwrapper method)": [[138, "tensorrt_llm.models.LlavaNextVisionWrapper.forward", false]], "forward() (tensorrt_llm.models.mambaforcausallm method)": [[138, "tensorrt_llm.models.MambaForCausalLM.forward", false]], "forward() (tensorrt_llm.models.mllamaforcausallm method)": [[138, "tensorrt_llm.models.MLLaMAForCausalLM.forward", false]], "forward() (tensorrt_llm.models.mptmodel method)": [[138, "tensorrt_llm.models.MPTModel.forward", false]], "forward() (tensorrt_llm.models.optmodel method)": [[138, "tensorrt_llm.models.OPTModel.forward", false]], "forward() (tensorrt_llm.models.phi3model method)": [[138, "tensorrt_llm.models.Phi3Model.forward", false]], "forward() (tensorrt_llm.models.phimodel method)": [[138, "tensorrt_llm.models.PhiModel.forward", false]], "forward() (tensorrt_llm.models.recurrentgemmaforcausallm method)": [[138, "tensorrt_llm.models.RecurrentGemmaForCausalLM.forward", false]], "forward() (tensorrt_llm.models.sd3transformer2dmodel method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.forward", false]], "forward() (tensorrt_llm.models.whisperencoder method)": [[138, "tensorrt_llm.models.WhisperEncoder.forward", false]], "forward_with_cfg() (tensorrt_llm.models.dit method)": [[138, "tensorrt_llm.models.DiT.forward_with_cfg", false]], "forward_without_cfg() (tensorrt_llm.models.dit method)": [[138, "tensorrt_llm.models.DiT.forward_without_cfg", false]], "fp8 (tensorrt_llm.llmapi.quantalgo attribute)": [[150, "tensorrt_llm.llmapi.QuantAlgo.FP8", false]], "fp8_block_scales (tensorrt_llm.llmapi.quantalgo attribute)": [[150, "tensorrt_llm.llmapi.QuantAlgo.FP8_BLOCK_SCALES", false]], "fp8_per_channel_per_token (tensorrt_llm.llmapi.quantalgo attribute)": [[150, "tensorrt_llm.llmapi.QuantAlgo.FP8_PER_CHANNEL_PER_TOKEN", false]], "fp8_rowwise_gemm_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.fp8_rowwise_gemm_plugin", false]], "free_gpu_memory_fraction (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.free_gpu_memory_fraction", false]], "frequency_penalty (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.frequency_penalty", false]], "frequency_penalty (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.frequency_penalty", false]], "from_arguments() (tensorrt_llm.models.speculativedecodingmode static method)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.from_arguments", false]], "from_arguments() (tensorrt_llm.plugin.pluginconfig class method)": [[139, "tensorrt_llm.plugin.PluginConfig.from_arguments", false]], "from_checkpoint() (tensorrt_llm.models.pretrainedconfig class method)": [[138, "tensorrt_llm.models.PretrainedConfig.from_checkpoint", false]], "from_checkpoint() (tensorrt_llm.models.pretrainedmodel class method)": [[138, "tensorrt_llm.models.PretrainedModel.from_checkpoint", false]], "from_config() (tensorrt_llm.models.pretrainedmodel class method)": [[138, "tensorrt_llm.models.PretrainedModel.from_config", false]], "from_dict() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.quantconfig class method)": [[150, "tensorrt_llm.llmapi.QuantConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.from_dict", false]], "from_dict() (tensorrt_llm.models.pretrainedconfig class method)": [[138, "tensorrt_llm.models.PretrainedConfig.from_dict", false]], "from_dir() (tensorrt_llm.runtime.modelrunner class method)": [[141, "tensorrt_llm.runtime.ModelRunner.from_dir", false]], "from_dir() (tensorrt_llm.runtime.modelrunnercpp class method)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.from_dir", false]], "from_engine() (tensorrt_llm.runtime.encdecmodelrunner class method)": [[141, "tensorrt_llm.runtime.EncDecModelRunner.from_engine", false]], "from_engine() (tensorrt_llm.runtime.modelrunner class method)": [[141, "tensorrt_llm.runtime.ModelRunner.from_engine", false]], "from_engine() (tensorrt_llm.runtime.session static method)": [[141, "tensorrt_llm.runtime.Session.from_engine", false]], "from_hugging_face() (tensorrt_llm.models.baichuanforcausallm class method)": [[138, "tensorrt_llm.models.BaichuanForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.chatglmconfig class method)": [[138, "tensorrt_llm.models.ChatGLMConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.chatglmforcausallm class method)": [[138, "tensorrt_llm.models.ChatGLMForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.cogvlmforcausallm class method)": [[138, "tensorrt_llm.models.CogVLMForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.cohereforcausallm class method)": [[138, "tensorrt_llm.models.CohereForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.deepseekforcausallm class method)": [[138, "tensorrt_llm.models.DeepseekForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.deepseekv2forcausallm class method)": [[138, "tensorrt_llm.models.DeepseekV2ForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.eagleforcausallm class method)": [[138, "tensorrt_llm.models.EagleForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.falconconfig class method)": [[138, "tensorrt_llm.models.FalconConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.falconforcausallm class method)": [[138, "tensorrt_llm.models.FalconForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.gemmaconfig class method)": [[138, "tensorrt_llm.models.GemmaConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.gemmaforcausallm class method)": [[138, "tensorrt_llm.models.GemmaForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.gptconfig class method)": [[138, "tensorrt_llm.models.GPTConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.gptforcausallm class method)": [[138, "tensorrt_llm.models.GPTForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.gptjconfig class method)": [[138, "tensorrt_llm.models.GPTJConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.gptjforcausallm class method)": [[138, "tensorrt_llm.models.GPTJForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.llamaconfig class method)": [[138, "tensorrt_llm.models.LLaMAConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.llamaforcausallm class method)": [[138, "tensorrt_llm.models.LLaMAForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.llavanextvisionconfig class method)": [[138, "tensorrt_llm.models.LlavaNextVisionConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.llavanextvisionwrapper class method)": [[138, "tensorrt_llm.models.LlavaNextVisionWrapper.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.mambaforcausallm class method)": [[138, "tensorrt_llm.models.MambaForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.medusaconfig class method)": [[138, "tensorrt_llm.models.MedusaConfig.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.medusaforcausallm class method)": [[138, "tensorrt_llm.models.MedusaForCausalLm.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.mllamaforcausallm class method)": [[138, "tensorrt_llm.models.MLLaMAForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.phi3forcausallm class method)": [[138, "tensorrt_llm.models.Phi3ForCausalLM.from_hugging_face", false]], "from_hugging_face() (tensorrt_llm.models.phiforcausallm class method)": [[138, "tensorrt_llm.models.PhiForCausalLM.from_hugging_face", false]], "from_json_file() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.from_json_file", false]], "from_json_file() (tensorrt_llm.models.pretrainedconfig class method)": [[138, "tensorrt_llm.models.PretrainedConfig.from_json_file", false]], "from_kwargs() (tensorrt_llm.llmapi.torchllmargs class method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.from_kwargs", false]], "from_kwargs() (tensorrt_llm.llmapi.trtllmargs class method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.from_kwargs", false]], "from_meta_ckpt() (tensorrt_llm.models.llamaconfig class method)": [[138, "tensorrt_llm.models.LLaMAConfig.from_meta_ckpt", false]], "from_meta_ckpt() (tensorrt_llm.models.llamaforcausallm class method)": [[138, "tensorrt_llm.models.LLaMAForCausalLM.from_meta_ckpt", false]], "from_model_config_cpp() (tensorrt_llm.runtime.modelconfig class method)": [[141, "tensorrt_llm.runtime.ModelConfig.from_model_config_cpp", false]], "from_nemo() (tensorrt_llm.models.gptconfig class method)": [[138, "tensorrt_llm.models.GPTConfig.from_nemo", false]], "from_nemo() (tensorrt_llm.models.gptforcausallm class method)": [[138, "tensorrt_llm.models.GPTForCausalLM.from_nemo", false]], "from_orm() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.from_orm", false]], "from_orm() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.from_orm", false]], "from_pretrained() (tensorrt_llm.models.sd3transformer2dmodel class method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.from_pretrained", false]], "from_pybind() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.from_pybind", false]], "from_pybind() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.from_pybind", false]], "from_pybind() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.from_pybind", false]], "from_pybind() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.from_pybind", false]], "from_pybind() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.from_pybind", false]], "from_pybind() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.from_pybind", false]], "from_serialized_engine() (tensorrt_llm.runtime.session static method)": [[141, "tensorrt_llm.runtime.Session.from_serialized_engine", false]], "from_string() (tensorrt_llm.functional.positionembeddingtype static method)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.from_string", false]], "from_string() (tensorrt_llm.functional.rotaryscalingtype static method)": [[136, "tensorrt_llm.functional.RotaryScalingType.from_string", false]], "fuse_fp4_quant (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.fuse_fp4_quant", false]], "fuse_qkv_projections() (tensorrt_llm.models.sd3transformer2dmodel method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.fuse_qkv_projections", false]], "fusedgatedmlp (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.FusedGatedMLP", false]], "fusedgatedmlp (tensorrt_llm.functional.mlptype attribute)": [[136, "tensorrt_llm.functional.MLPType.FusedGatedMLP", false]], "garbage_collection_gen0_threshold (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.garbage_collection_gen0_threshold", false]], "gatedmlp (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.GatedMLP", false]], "gatedmlp (tensorrt_llm.functional.mlptype attribute)": [[136, "tensorrt_llm.functional.MLPType.GatedMLP", false]], "gather() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gather", false]], "gather_context_logits (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.gather_context_logits", false]], "gather_context_logits (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.gather_context_logits", false]], "gather_context_logits (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.gather_context_logits", false]], "gather_context_logits (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.gather_context_logits", false]], "gather_context_logits (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.gather_context_logits", false]], "gather_generation_logits (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.gather_generation_logits", false]], "gather_generation_logits (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.gather_generation_logits", false]], "gather_generation_logits (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.gather_generation_logits", false]], "gather_generation_logits (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.gather_generation_logits", false]], "gather_generation_logits (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.gather_generation_logits", false]], "gather_generation_logits (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.gather_generation_logits", false]], "gather_generation_logits (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.gather_generation_logits", false]], "gather_last_token_logits() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gather_last_token_logits", false]], "gather_nd() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gather_nd", false]], "gegelu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gegelu", false]], "geglu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.geglu", false]], "gelu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gelu", false]], "gemm_allreduce() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gemm_allreduce", false]], "gemm_allreduce_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.gemm_allreduce_plugin", false]], "gemm_allreduce_plugin (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.gemm_allreduce_plugin", false]], "gemm_allreduce_plugin (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.gemm_allreduce_plugin", false]], "gemm_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.gemm_plugin", false]], "gemm_swiglu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gemm_swiglu", false]], "gemm_swiglu_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.gemm_swiglu_plugin", false]], "gemma2_added_fields (tensorrt_llm.models.gemmaconfig attribute)": [[138, "tensorrt_llm.models.GemmaConfig.GEMMA2_ADDED_FIELDS", false]], "gemma2_config() (tensorrt_llm.models.gemmaconfig method)": [[138, "tensorrt_llm.models.GemmaConfig.gemma2_config", false]], "gemma3_added_fields (tensorrt_llm.models.gemmaconfig attribute)": [[138, "tensorrt_llm.models.GemmaConfig.GEMMA3_ADDED_FIELDS", false]], "gemma3_config() (tensorrt_llm.models.gemmaconfig method)": [[138, "tensorrt_llm.models.GemmaConfig.gemma3_config", false]], "gemma_added_fields (tensorrt_llm.models.gemmaconfig attribute)": [[138, "tensorrt_llm.models.GemmaConfig.GEMMA_ADDED_FIELDS", false]], "gemmaconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GemmaConfig", false]], "gemmaforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GemmaForCausalLM", false]], "generate() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.generate", false]], "generate() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.generate", false]], "generate() (tensorrt_llm.runtime.encdecmodelrunner method)": [[141, "tensorrt_llm.runtime.EncDecModelRunner.generate", false]], "generate() (tensorrt_llm.runtime.modelrunner method)": [[141, "tensorrt_llm.runtime.ModelRunner.generate", false]], "generate() (tensorrt_llm.runtime.modelrunnercpp method)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.generate", false]], "generate() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.generate", false]], "generate() (tensorrt_llm.runtime.qwenforcausallmgenerationsession method)": [[141, "tensorrt_llm.runtime.QWenForCausalLMGenerationSession.generate", false]], "generate_alibi_biases() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.generate_alibi_biases", false]], "generate_alibi_slopes() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.generate_alibi_slopes", false]], "generate_async() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.generate_async", false]], "generate_async() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.generate_async", false]], "generate_logn_scaling() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.generate_logn_scaling", false]], "generation_logits (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.generation_logits", false]], "generationsequence (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.GenerationSequence", false]], "generationsession (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.GenerationSession", false]], "get_1d_sincos_pos_embed_from_grid() (in module tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.get_1d_sincos_pos_embed_from_grid", false]], "get_2d_sincos_pos_embed() (in module tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.get_2d_sincos_pos_embed", false]], "get_2d_sincos_pos_embed_from_grid() (in module tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.get_2d_sincos_pos_embed_from_grid", false]], "get_audio_features() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.get_audio_features", false]], "get_batch_idx() (tensorrt_llm.runtime.generationsequence method)": [[141, "tensorrt_llm.runtime.GenerationSequence.get_batch_idx", false]], "get_block_offsets() (tensorrt_llm.runtime.kvcachemanager method)": [[141, "tensorrt_llm.runtime.KVCacheManager.get_block_offsets", false]], "get_comm() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.get_comm", false]], "get_config_group() (tensorrt_llm.models.pretrainedconfig method)": [[138, "tensorrt_llm.models.PretrainedConfig.get_config_group", false]], "get_context_phase_params() (tensorrt_llm.llmapi.disaggregatedparams method)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.get_context_phase_params", false]], "get_executor_config() (tensorrt_llm.llmapi.torchllmargs method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.get_executor_config", false]], "get_first_past_key_value() (tensorrt_llm.layers.attention.keyvaluecacheparams method)": [[137, "tensorrt_llm.layers.attention.KeyValueCacheParams.get_first_past_key_value", false]], "get_hf_config() (tensorrt_llm.models.gemmaconfig static method)": [[138, "tensorrt_llm.models.GemmaConfig.get_hf_config", false]], "get_indices_block_size() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.get_indices_block_size", false]], "get_indices_block_size() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.get_indices_block_size", false]], "get_kv_cache_events() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.get_kv_cache_events", false]], "get_kv_cache_events() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.get_kv_cache_events", false]], "get_kv_cache_events_async() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.get_kv_cache_events_async", false]], "get_kv_cache_events_async() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.get_kv_cache_events_async", false]], "get_next_medusa_tokens() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.get_next_medusa_tokens", false]], "get_num_heads_kv() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.get_num_heads_kv", false]], "get_parent() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.get_parent", false]], "get_pybind_enum_fields() (tensorrt_llm.llmapi.cachetransceiverconfig static method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.get_pybind_enum_fields", false]], "get_pybind_enum_fields() (tensorrt_llm.llmapi.dynamicbatchconfig static method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.get_pybind_enum_fields", false]], "get_pybind_enum_fields() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig static method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.get_pybind_enum_fields", false]], "get_pybind_enum_fields() (tensorrt_llm.llmapi.kvcacheconfig static method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.get_pybind_enum_fields", false]], "get_pybind_enum_fields() (tensorrt_llm.llmapi.lookaheaddecodingconfig static method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.get_pybind_enum_fields", false]], "get_pybind_enum_fields() (tensorrt_llm.llmapi.schedulerconfig static method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.get_pybind_enum_fields", false]], "get_pybind_variable_fields() (tensorrt_llm.llmapi.cachetransceiverconfig static method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.get_pybind_variable_fields", false]], "get_pybind_variable_fields() (tensorrt_llm.llmapi.dynamicbatchconfig static method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.get_pybind_variable_fields", false]], "get_pybind_variable_fields() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig static method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.get_pybind_variable_fields", false]], "get_pybind_variable_fields() (tensorrt_llm.llmapi.kvcacheconfig static method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.get_pybind_variable_fields", false]], "get_pybind_variable_fields() (tensorrt_llm.llmapi.lookaheaddecodingconfig static method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.get_pybind_variable_fields", false]], "get_pybind_variable_fields() (tensorrt_llm.llmapi.schedulerconfig static method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.get_pybind_variable_fields", false]], "get_request_type() (tensorrt_llm.llmapi.disaggregatedparams method)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.get_request_type", false]], "get_rope_index() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.get_rope_index", false]], "get_runtime_sizes() (tensorrt_llm.llmapi.torchllmargs method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.get_runtime_sizes", false]], "get_runtime_sizes() (tensorrt_llm.llmapi.trtllmargs method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.get_runtime_sizes", false]], "get_seq_idx() (tensorrt_llm.runtime.generationsequence method)": [[141, "tensorrt_llm.runtime.GenerationSequence.get_seq_idx", false]], "get_stats() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.get_stats", false]], "get_stats() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.get_stats", false]], "get_stats_async() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.get_stats_async", false]], "get_stats_async() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.get_stats_async", false]], "get_timestep_embedding() (in module tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.get_timestep_embedding", false]], "get_users() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.get_users", false]], "get_visual_features() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.get_visual_features", false]], "get_weight() (tensorrt_llm.layers.linear.linearbase method)": [[137, "tensorrt_llm.layers.linear.LinearBase.get_weight", false]], "gpt_attention() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gpt_attention", false]], "gpt_attention_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.gpt_attention_plugin", false]], "gpt_attention_plugin (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.gpt_attention_plugin", false]], "gptconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTConfig", false]], "gptforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTForCausalLM", false]], "gptjconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTJConfig", false]], "gptjforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTJForCausalLM", false]], "gptjmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTJModel", false]], "gptmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTModel", false]], "gptneoxforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTNeoXForCausalLM", false]], "gptneoxmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.GPTNeoXModel", false]], "gpu_weights_percent (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.gpu_weights_percent", false]], "gpus_per_node (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.gpus_per_node", false]], "gpus_per_node (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.gpus_per_node", false]], "grammar (tensorrt_llm.llmapi.guideddecodingparams attribute)": [[150, "tensorrt_llm.llmapi.GuidedDecodingParams.grammar", false]], "greedy_sampling (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.greedy_sampling", false]], "group_norm() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.group_norm", false]], "group_size (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.group_size", false]], "groupnorm (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.GroupNorm", false]], "groupnorm (tensorrt_llm.functional.layernormtype attribute)": [[136, "tensorrt_llm.functional.LayerNormType.GroupNorm", false]], "gt() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.gt", false]], "guaranteed_no_evict (tensorrt_llm.llmapi.capacityschedulerpolicy attribute)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.GUARANTEED_NO_EVICT", false]], "guided_decoding (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.guided_decoding", false]], "guided_decoding_backend (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.guided_decoding_backend", false]], "guided_decoding_backend (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.guided_decoding_backend", false]], "guideddecodingparams (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.GuidedDecodingParams", false]], "handle_per_step() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.handle_per_step", false]], "has_affine() (tensorrt_llm.functional.allreduceparams method)": [[136, "tensorrt_llm.functional.AllReduceParams.has_affine", false]], "has_bias() (tensorrt_llm.functional.allreduceparams method)": [[136, "tensorrt_llm.functional.AllReduceParams.has_bias", false]], "has_config_group() (tensorrt_llm.models.pretrainedconfig method)": [[138, "tensorrt_llm.models.PretrainedConfig.has_config_group", false]], "has_position_embedding (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.has_position_embedding", false]], "has_position_embedding (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.has_position_embedding", false]], "has_scale() (tensorrt_llm.functional.allreduceparams method)": [[136, "tensorrt_llm.functional.AllReduceParams.has_scale", false]], "has_token_type_embedding (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.has_token_type_embedding", false]], "has_token_type_embedding (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.has_token_type_embedding", false]], "has_zero_point (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.has_zero_point", false]], "head_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.head_size", false]], "head_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.head_size", false]], "hidden_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.hidden_size", false]], "hidden_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.hidden_size", false]], "hidden_size (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.hidden_size", false]], "hidden_size (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.hidden_size", false]], "host_cache_size (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.host_cache_size", false]], "identity() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.identity", false]], "identity_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.identity_plugin", false]], "ignore_eos (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.ignore_eos", false]], "include_stop_str_in_output (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.include_stop_str_in_output", false]], "index (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.index", false]], "index() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.index", false]], "index() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.index", false]], "index() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.index", false]], "index() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.index", false]], "index() (tensorrt_llm.llmapi.requestoutput.postprocworker.output method)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.index", false]], "index_head_dim (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.index_head_dim", false]], "index_n_heads (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.index_n_heads", false]], "index_select() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.index_select", false]], "index_topk (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.index_topk", false]], "indexer_max_chunk_size (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.indexer_max_chunk_size", false]], "infer_shapes() (tensorrt_llm.runtime.session method)": [[141, "tensorrt_llm.runtime.Session.infer_shapes", false]], "inflight (tensorrt_llm.llmapi.batchingtype attribute)": [[150, "tensorrt_llm.llmapi.BatchingType.INFLIGHT", false]], "init_audio_encoder() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.init_audio_encoder", false]], "init_backend() (tensorrt_llm.llmapi.torchllmargs class method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.init_backend", false]], "init_build_config() (tensorrt_llm.llmapi.torchllmargs method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.init_build_config", false]], "init_build_config() (tensorrt_llm.llmapi.trtllmargs method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.init_build_config", false]], "init_calib_config() (tensorrt_llm.llmapi.trtllmargs class method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.init_calib_config", false]], "init_image_encoder() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.init_image_encoder", false]], "init_llm() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.init_llm", false]], "init_processor() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.init_processor", false]], "init_tokenizer() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.init_tokenizer", false]], "input_timing_cache (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.input_timing_cache", false]], "int8 (tensorrt_llm.llmapi.quantalgo attribute)": [[150, "tensorrt_llm.llmapi.QuantAlgo.INT8", false]], "int_clip() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.int_clip", false]], "interpolate() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.interpolate", false]], "is_alibi() (tensorrt_llm.functional.positionembeddingtype method)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.is_alibi", false]], "is_comm_session() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.is_comm_session", false]], "is_context_fmha_enabled() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.is_context_fmha_enabled", false]], "is_deferred() (tensorrt_llm.functional.positionembeddingtype method)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.is_deferred", false]], "is_dynamic() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.is_dynamic", false]], "is_final (tensorrt_llm.llmapi.requestoutput.postprocworker.output attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.is_final", false]], "is_gated_activation() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.is_gated_activation", false]], "is_gemma_2 (tensorrt_llm.models.gemmaconfig property)": [[138, "tensorrt_llm.models.GemmaConfig.is_gemma_2", false]], "is_gemma_3 (tensorrt_llm.models.gemmaconfig property)": [[138, "tensorrt_llm.models.GemmaConfig.is_gemma_3", false]], "is_keep_all (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.is_keep_all", false]], "is_linear_tree (tensorrt_llm.llmapi.eagledecodingconfig property)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.is_linear_tree", false]], "is_medusa_mode (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.is_medusa_mode", false]], "is_module_excluded_from_quantization() (tensorrt_llm.llmapi.quantconfig method)": [[150, "tensorrt_llm.llmapi.QuantConfig.is_module_excluded_from_quantization", false]], "is_mrope() (tensorrt_llm.functional.positionembeddingtype method)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.is_mrope", false]], "is_public_pool (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.is_public_pool", false]], "is_redrafter_mode (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.is_redrafter_mode", false]], "is_rope() (tensorrt_llm.functional.positionembeddingtype method)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.is_rope", false]], "is_trt_wrapper() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.is_trt_wrapper", false]], "is_use_oldest (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.is_use_oldest", false]], "is_valid() (tensorrt_llm.functional.moeallreduceparams method)": [[136, "tensorrt_llm.functional.MoEAllReduceParams.is_valid", false]], "is_valid() (tensorrt_llm.layers.attention.attentionparams method)": [[137, "tensorrt_llm.layers.attention.AttentionParams.is_valid", false]], "is_valid() (tensorrt_llm.layers.attention.keyvaluecacheparams method)": [[137, "tensorrt_llm.layers.attention.KeyValueCacheParams.is_valid", false]], "is_valid_cross_attn() (tensorrt_llm.layers.attention.attentionparams method)": [[137, "tensorrt_llm.layers.attention.AttentionParams.is_valid_cross_attn", false]], "isalnum() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isalnum", false]], "isalnum() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isalnum", false]], "isalnum() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isalnum", false]], "isalnum() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isalnum", false]], "isalpha() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isalpha", false]], "isalpha() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isalpha", false]], "isalpha() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isalpha", false]], "isalpha() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isalpha", false]], "isascii() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isascii", false]], "isascii() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isascii", false]], "isascii() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isascii", false]], "isascii() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isascii", false]], "isdecimal() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isdecimal", false]], "isdecimal() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isdecimal", false]], "isdecimal() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isdecimal", false]], "isdecimal() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isdecimal", false]], "isdigit() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isdigit", false]], "isdigit() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isdigit", false]], "isdigit() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isdigit", false]], "isdigit() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isdigit", false]], "isidentifier() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isidentifier", false]], "isidentifier() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isidentifier", false]], "isidentifier() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isidentifier", false]], "isidentifier() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isidentifier", false]], "islower() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.islower", false]], "islower() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.islower", false]], "islower() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.islower", false]], "islower() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.islower", false]], "isnumeric() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isnumeric", false]], "isnumeric() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isnumeric", false]], "isnumeric() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isnumeric", false]], "isnumeric() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isnumeric", false]], "isprintable() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isprintable", false]], "isprintable() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isprintable", false]], "isprintable() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isprintable", false]], "isprintable() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isprintable", false]], "isspace() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isspace", false]], "isspace() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isspace", false]], "isspace() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isspace", false]], "isspace() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isspace", false]], "istitle() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.istitle", false]], "istitle() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.istitle", false]], "istitle() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.istitle", false]], "istitle() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.istitle", false]], "isupper() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.isupper", false]], "isupper() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.isupper", false]], "isupper() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.isupper", false]], "isupper() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.isupper", false]], "iter_stats_max_iterations (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.iter_stats_max_iterations", false]], "iter_stats_max_iterations (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.iter_stats_max_iterations", false]], "join() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.join", false]], "join() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.join", false]], "join() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.join", false]], "join() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.join", false]], "joint_attn_forward() (tensorrt_llm.layers.attention.diffusersattention method)": [[137, "tensorrt_llm.layers.attention.DiffusersAttention.joint_attn_forward", false]], "json (tensorrt_llm.llmapi.guideddecodingparams attribute)": [[150, "tensorrt_llm.llmapi.GuidedDecodingParams.json", false]], "json() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.json", false]], "json() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.json", false]], "json() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.json", false]], "json() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.json", false]], "json() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.json", false]], "json() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.json", false]], "json() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.json", false]], "json() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.json", false]], "json() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.json", false]], "json() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.json", false]], "json() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.json", false]], "json() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.json", false]], "json() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.json", false]], "json() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.json", false]], "json() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.json", false]], "json_object (tensorrt_llm.llmapi.guideddecodingparams attribute)": [[150, "tensorrt_llm.llmapi.GuidedDecodingParams.json_object", false]], "kernel_size (tensorrt_llm.llmapi.rocketsparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.kernel_size", false]], "keyvaluecacheparams (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.KeyValueCacheParams", false]], "kv_cache_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.kv_cache_config", false]], "kv_cache_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.kv_cache_config", false]], "kv_cache_quant_algo (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.kv_cache_quant_algo", false]], "kv_cache_type (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.kv_cache_type", false]], "kv_cache_type (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.kv_cache_type", false]], "kv_cache_type (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.kv_cache_type", false]], "kv_connector_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.kv_connector_config", false]], "kv_dtype (tensorrt_llm.models.pretrainedconfig property)": [[138, "tensorrt_llm.models.PretrainedConfig.kv_dtype", false]], "kv_transfer_sender_future_timeout_ms (tensorrt_llm.llmapi.cachetransceiverconfig attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.kv_transfer_sender_future_timeout_ms", false]], "kv_transfer_timeout_ms (tensorrt_llm.llmapi.cachetransceiverconfig attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.kv_transfer_timeout_ms", false]], "kvcacheconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.KvCacheConfig", false]], "kvcacheconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.Config", false]], "kvcachemanager (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.KVCacheManager", false]], "kvcacheretentionconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig", false]], "kvcacheretentionconfig.tokenrangeretentionconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.TokenRangeRetentionConfig", false]], "labelembedding (class in tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.LabelEmbedding", false]], "language_adapter_config (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.language_adapter_config", false]], "last_layer (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.last_layer", false]], "last_process_for_ub (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.LAST_PROCESS_FOR_UB", false]], "layer_norm() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.layer_norm", false]], "layer_quant_mode (tensorrt_llm.llmapi.quantconfig property)": [[150, "tensorrt_llm.llmapi.QuantConfig.layer_quant_mode", false]], "layer_types (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.layer_types", false]], "layernorm (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.LayerNorm", false]], "layernorm (tensorrt_llm.functional.layernormtype attribute)": [[136, "tensorrt_llm.functional.LayerNormType.LayerNorm", false]], "layernorm_quantization_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.layernorm_quantization_plugin", false]], "layernormpositiontype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.LayerNormPositionType", false]], "layernormtype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.LayerNormType", false]], "learned_absolute (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.learned_absolute", false]], "length (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.length", false]], "length (tensorrt_llm.llmapi.completionoutput property)": [[150, "id2", false]], "length_penalty (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.length_penalty", false]], "length_penalty (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.length_penalty", false]], "linear (class in tensorrt_llm.layers.linear)": [[137, "tensorrt_llm.layers.linear.Linear", false]], "linear (tensorrt_llm.functional.rotaryscalingtype attribute)": [[136, "tensorrt_llm.functional.RotaryScalingType.linear", false]], "linearactivation (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.LinearActivation", false]], "linearapproximategelu (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.LinearApproximateGELU", false]], "linearbase (class in tensorrt_llm.layers.linear)": [[137, "tensorrt_llm.layers.linear.LinearBase", false]], "lineargeglu (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.LinearGEGLU", false]], "lineargelu (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.LinearGELU", false]], "linearswiglu (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.LinearSwiGLU", false]], "ljust() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.ljust", false]], "ljust() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.ljust", false]], "ljust() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.ljust", false]], "ljust() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.ljust", false]], "llama3 (tensorrt_llm.functional.rotaryscalingtype attribute)": [[136, "tensorrt_llm.functional.RotaryScalingType.llama3", false]], "llamaconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.LLaMAConfig", false]], "llamaforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.LLaMAForCausalLM", false]], "llamamodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.LLaMAModel", false]], "llavanextvisionconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.LlavaNextVisionConfig", false]], "llavanextvisionwrapper (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.LlavaNextVisionWrapper", false]], "llm (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.LLM", false]], "llm_engine_dir (tensorrt_llm.runtime.multimodalmodelrunner property)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.llm_engine_dir", false]], "llm_id (tensorrt_llm.llmapi.llm attribute)": [[150, "tensorrt_llm.llmapi.LLM.llm_id", false]], "llm_id (tensorrt_llm.llmapi.llm property)": [[150, "id0", false]], "llm_id (tensorrt_llm.llmapi.multimodalencoder property)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.llm_id", false]], "llmargs (in module tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.LlmArgs", false]], "load() (tensorrt_llm.models.pretrainedmodel method)": [[138, "tensorrt_llm.models.PretrainedModel.load", false]], "load() (tensorrt_llm.models.sd3transformer2dmodel method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.load", false]], "load_balancer (tensorrt_llm.llmapi.moeconfig attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.load_balancer", false]], "load_format (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.load_format", false]], "load_format (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.load_format", false]], "load_format (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.load_format", false]], "load_format (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.load_format", false]], "load_test_audio() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.load_test_audio", false]], "load_test_data() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.load_test_data", false]], "locate_accepted_draft_tokens() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.locate_accepted_draft_tokens", false]], "location (tensorrt_llm.functional.tensor property)": [[136, "tensorrt_llm.functional.Tensor.location", false]], "log() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.log", false]], "log() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.log", false]], "log_field_changes() (tensorrt_llm.plugin.pluginconfig class method)": [[139, "tensorrt_llm.plugin.PluginConfig.log_field_changes", false]], "log_softmax() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.log_softmax", false]], "logits_processor (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.logits_processor", false]], "logitsprocessor (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.LogitsProcessor", false]], "logitsprocessorlist (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.LogitsProcessorList", false]], "logprobs (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.logprobs", false]], "logprobs (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.logprobs", false]], "logprobs_diff (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.logprobs_diff", false]], "logprobs_diff (tensorrt_llm.llmapi.completionoutput property)": [[150, "id3", false]], "long_rope (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.long_rope", false]], "longrope (tensorrt_llm.functional.rotaryscalingtype attribute)": [[136, "tensorrt_llm.functional.RotaryScalingType.longrope", false]], "lookahead_config (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.lookahead_config", false]], "lookahead_decoding (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.LOOKAHEAD_DECODING", false]], "lookaheaddecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig", false]], "lookaheaddecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.Config", false]], "lora_ckpt_source (tensorrt_llm.llmapi.lorarequest attribute)": [[150, "tensorrt_llm.llmapi.LoRARequest.lora_ckpt_source", false]], "lora_config (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.lora_config", false]], "lora_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.lora_config", false]], "lora_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.lora_config", false]], "lora_int_id (tensorrt_llm.llmapi.lorarequest attribute)": [[150, "tensorrt_llm.llmapi.LoRARequest.lora_int_id", false]], "lora_name (tensorrt_llm.llmapi.lorarequest attribute)": [[150, "tensorrt_llm.llmapi.LoRARequest.lora_name", false]], "lora_path (tensorrt_llm.llmapi.lorarequest attribute)": [[150, "tensorrt_llm.llmapi.LoRARequest.lora_path", false]], "lora_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.lora_plugin", false]], "lora_plugin (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.lora_plugin", false]], "lora_plugin() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.lora_plugin", false]], "lora_target_modules (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.lora_target_modules", false]], "lorarequest (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.LoRARequest", false]], "low_latency_gemm() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.low_latency_gemm", false]], "low_latency_gemm_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.low_latency_gemm_plugin", false]], "low_latency_gemm_swiglu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.low_latency_gemm_swiglu", false]], "low_latency_gemm_swiglu_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.low_latency_gemm_swiglu_plugin", false]], "lower() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.lower", false]], "lower() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.lower", false]], "lower() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.lower", false]], "lower() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.lower", false]], "lowprecision (tensorrt_llm.functional.allreducestrategy attribute)": [[136, "tensorrt_llm.functional.AllReduceStrategy.LOWPRECISION", false]], "lstrip() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.lstrip", false]], "lstrip() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.lstrip", false]], "lstrip() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.lstrip", false]], "lstrip() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.lstrip", false]], "lt() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.lt", false]], "make_causal_mask() (in module tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.make_causal_mask", false]], "maketrans() (tensorrt_llm.llmapi.batchingtype static method)": [[150, "tensorrt_llm.llmapi.BatchingType.maketrans", false]], "maketrans() (tensorrt_llm.llmapi.capacityschedulerpolicy static method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.maketrans", false]], "maketrans() (tensorrt_llm.llmapi.contextchunkingpolicy static method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.maketrans", false]], "maketrans() (tensorrt_llm.llmapi.quantalgo static method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.maketrans", false]], "mamba_conv1d() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.mamba_conv1d", false]], "mamba_conv1d_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.mamba_conv1d_plugin", false]], "mamba_conv1d_plugin (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.mamba_conv1d_plugin", false]], "mamba_ssm_cache_dtype (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.mamba_ssm_cache_dtype", false]], "mamba_ssm_cache_dtype (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.mamba_ssm_cache_dtype", false]], "mambaforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.MambaForCausalLM", false]], "manage_weights (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.manage_weights", false]], "mapping (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.mapping", false]], "mapping (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.mapping", false]], "mark_output() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.mark_output", false]], "masked_scatter() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.masked_scatter", false]], "masked_select() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.masked_select", false]], "matmul() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.matmul", false]], "max() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.max", false]], "max() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.max", false]], "max_attention_window (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.max_attention_window", false]], "max_attention_window_size (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.max_attention_window_size", false]], "max_batch_size (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_batch_size", false]], "max_batch_size (tensorrt_llm.llmapi.cudagraphconfig attribute)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.max_batch_size", false]], "max_batch_size (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.max_batch_size", false]], "max_batch_size (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.max_batch_size", false]], "max_batch_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.max_batch_size", false]], "max_beam_width (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_beam_width", false]], "max_beam_width (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.max_beam_width", false]], "max_beam_width (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.max_beam_width", false]], "max_beam_width (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.max_beam_width", false]], "max_cache_storage_gb (tensorrt_llm.llmapi.buildcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildCacheConfig.max_cache_storage_gb", false]], "max_cache_storage_gb (tensorrt_llm.llmapi.buildcacheconfig property)": [[150, "id14", false]], "max_concurrency (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.max_concurrency", false]], "max_concurrency (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.max_concurrency", false]], "max_draft_len (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.max_draft_len", false]], "max_draft_len (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.max_draft_len", false]], "max_draft_tokens (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.max_draft_tokens", false]], "max_encoder_input_len (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_encoder_input_len", false]], "max_gpu_total_bytes (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.max_gpu_total_bytes", false]], "max_input_len (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_input_len", false]], "max_input_len (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.max_input_len", false]], "max_input_len (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.max_input_len", false]], "max_matching_ngram_size (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.max_matching_ngram_size", false]], "max_medusa_tokens (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.max_medusa_tokens", false]], "max_new_tokens (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.max_new_tokens", false]], "max_ngram_size (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.max_ngram_size", false]], "max_non_leaves_per_layer (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.max_non_leaves_per_layer", false]], "max_num_streams (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.max_num_streams", false]], "max_num_tokens (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_num_tokens", false]], "max_num_tokens (tensorrt_llm.llmapi.moeconfig attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.max_num_tokens", false]], "max_num_tokens (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.max_num_tokens", false]], "max_num_tokens (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.max_num_tokens", false]], "max_prompt_adapter_token (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.max_prompt_adapter_token", false]], "max_prompt_embedding_table_size (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_prompt_embedding_table_size", false]], "max_prompt_embedding_table_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.max_prompt_embedding_table_size", false]], "max_prompt_embedding_table_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.max_prompt_embedding_table_size", false]], "max_prompt_embedding_table_size (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.max_prompt_embedding_table_size", false]], "max_prompt_embedding_table_size (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.max_prompt_embedding_table_size", false]], "max_records (tensorrt_llm.llmapi.buildcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildCacheConfig.max_records", false]], "max_records (tensorrt_llm.llmapi.buildcacheconfig property)": [[150, "id15", false]], "max_seq_len (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.max_seq_len", false]], "max_seq_len (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.max_seq_len", false]], "max_seq_len (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.max_seq_len", false]], "max_sequence_length (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.max_sequence_length", false]], "max_sequence_length (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.max_sequence_length", false]], "max_tokens (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.max_tokens", false]], "max_tokens (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.max_tokens", false]], "max_tokens_in_buffer (tensorrt_llm.llmapi.cachetransceiverconfig attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.max_tokens_in_buffer", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.max_total_draft_tokens", false]], "max_total_draft_tokens (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.max_total_draft_tokens", false]], "max_utilization (tensorrt_llm.llmapi.capacityschedulerpolicy attribute)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.MAX_UTILIZATION", false]], "max_verification_set_size (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.max_verification_set_size", false]], "max_window_size (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.max_window_size", false]], "maximum() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.maximum", false]], "maybe_to_pybind() (tensorrt_llm.llmapi.cachetransceiverconfig static method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.maybe_to_pybind", false]], "maybe_to_pybind() (tensorrt_llm.llmapi.dynamicbatchconfig static method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.maybe_to_pybind", false]], "maybe_to_pybind() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig static method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.maybe_to_pybind", false]], "maybe_to_pybind() (tensorrt_llm.llmapi.kvcacheconfig static method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.maybe_to_pybind", false]], "maybe_to_pybind() (tensorrt_llm.llmapi.lookaheaddecodingconfig static method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.maybe_to_pybind", false]], "maybe_to_pybind() (tensorrt_llm.llmapi.schedulerconfig static method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.maybe_to_pybind", false]], "mean() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.mean", false]], "mean() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.mean", false]], "medusa (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.MEDUSA", false]], "medusa_choices (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.medusa_choices", false]], "medusa_decode_and_verify() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.medusa_decode_and_verify", false]], "medusa_paths (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.medusa_paths", false]], "medusa_position_offsets (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.medusa_position_offsets", false]], "medusa_temperature (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.medusa_temperature", false]], "medusa_topks (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.medusa_topks", false]], "medusa_tree_ids (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.medusa_tree_ids", false]], "medusaconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.MedusaConfig", false]], "medusadecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig", false]], "medusadecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.Config", false]], "medusaforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.MedusaForCausalLm", false]], "meshgrid2d() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.meshgrid2d", false]], "metrics (tensorrt_llm.llmapi.requestoutput.postprocworker.output attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.metrics", false]], "min() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.min", false]], "min_latency (tensorrt_llm.functional.allreducestrategy attribute)": [[136, "tensorrt_llm.functional.AllReduceStrategy.MIN_LATENCY", false]], "min_length (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.min_length", false]], "min_p (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.min_p", false]], "min_p (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.min_p", false]], "min_tokens (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.min_tokens", false]], "minimum() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.minimum", false]], "mirror_pybind_enum() (tensorrt_llm.llmapi.cachetransceiverconfig static method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.mirror_pybind_enum", false]], "mirror_pybind_enum() (tensorrt_llm.llmapi.dynamicbatchconfig static method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.mirror_pybind_enum", false]], "mirror_pybind_enum() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig static method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.mirror_pybind_enum", false]], "mirror_pybind_enum() (tensorrt_llm.llmapi.kvcacheconfig static method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.mirror_pybind_enum", false]], "mirror_pybind_enum() (tensorrt_llm.llmapi.lookaheaddecodingconfig static method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.mirror_pybind_enum", false]], "mirror_pybind_enum() (tensorrt_llm.llmapi.schedulerconfig static method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.mirror_pybind_enum", false]], "mirror_pybind_fields() (tensorrt_llm.llmapi.cachetransceiverconfig static method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.mirror_pybind_fields", false]], "mirror_pybind_fields() (tensorrt_llm.llmapi.dynamicbatchconfig static method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.mirror_pybind_fields", false]], "mirror_pybind_fields() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig static method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.mirror_pybind_fields", false]], "mirror_pybind_fields() (tensorrt_llm.llmapi.kvcacheconfig static method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.mirror_pybind_fields", false]], "mirror_pybind_fields() (tensorrt_llm.llmapi.lookaheaddecodingconfig static method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.mirror_pybind_fields", false]], "mirror_pybind_fields() (tensorrt_llm.llmapi.schedulerconfig static method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.mirror_pybind_fields", false]], "mish (class in tensorrt_llm.layers.activation)": [[137, "tensorrt_llm.layers.activation.Mish", false]], "mixed_precision (tensorrt_llm.llmapi.quantalgo attribute)": [[150, "tensorrt_llm.llmapi.QuantAlgo.MIXED_PRECISION", false]], "mllamaforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.MLLaMAForCausalLM", false]], "mlp (class in tensorrt_llm.layers.mlp)": [[137, "tensorrt_llm.layers.mlp.MLP", false]], "mlp (tensorrt_llm.functional.mlptype attribute)": [[136, "tensorrt_llm.functional.MLPType.MLP", false]], "mlptype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.MLPType", false]], "mm_embedding_handle (tensorrt_llm.llmapi.requestoutput attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.mm_embedding_handle", false]], "mm_embedding_handle (tensorrt_llm.llmapi.requestoutput property)": [[150, "id8", false]], "mm_encoder_only (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.mm_encoder_only", false]], "mnnvl (tensorrt_llm.functional.allreducestrategy attribute)": [[136, "tensorrt_llm.functional.AllReduceStrategy.MNNVL", false]], "model": [[27, "cmdoption-trtllm-serve-mm_embedding_serve-arg-MODEL", false], [27, "cmdoption-trtllm-serve-serve-arg-MODEL", false]], "model (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.model", false]], "model (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.model", false]], "model_computed_fields (tensorrt_llm.llmapi.attentiondpconfig attribute)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.cachetransceiverconfig attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.cudagraphconfig attribute)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.dynamicbatchconfig attribute)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.extendedruntimeperfknobconfig attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.moeconfig attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.rocketsparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.schedulerconfig attribute)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_computed_fields", false]], "model_computed_fields (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_computed_fields", false]], "model_config (tensorrt_llm.llmapi.attentiondpconfig attribute)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.cachetransceiverconfig attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.cudagraphconfig attribute)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.dynamicbatchconfig attribute)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.extendedruntimeperfknobconfig attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.moeconfig attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.rocketsparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.schedulerconfig attribute)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_config", false]], "model_config (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_config", false]], "model_construct() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_construct", false]], "model_construct() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_construct", false]], "model_copy() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_copy", false]], "model_copy() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_copy", false]], "model_dump() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_dump", false]], "model_dump() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_dump", false]], "model_dump_json() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_dump_json", false]], "model_dump_json() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_dump_json", false]], "model_extra (tensorrt_llm.llmapi.attentiondpconfig property)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.autodecodingconfig property)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.buildconfig property)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.cachetransceiverconfig property)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.calibconfig property)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.cudagraphconfig property)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.deepseeksparseattentionconfig property)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.drafttargetdecodingconfig property)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.dynamicbatchconfig property)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.eagledecodingconfig property)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.extendedruntimeperfknobconfig property)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.kvcacheconfig property)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.lookaheaddecodingconfig property)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.medusadecodingconfig property)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.moeconfig property)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.mtpdecodingconfig property)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.ngramdecodingconfig property)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.rocketsparseattentionconfig property)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig property)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.schedulerconfig property)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.torchcompileconfig property)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_extra", false]], "model_extra (tensorrt_llm.llmapi.userprovideddecodingconfig property)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_extra", false]], "model_fields (tensorrt_llm.llmapi.attentiondpconfig attribute)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.cachetransceiverconfig attribute)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.cudagraphconfig attribute)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.deepseeksparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.dynamicbatchconfig attribute)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.extendedruntimeperfknobconfig attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.moeconfig attribute)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.rocketsparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.schedulerconfig attribute)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.torchcompileconfig attribute)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_fields", false]], "model_fields (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_fields", false]], "model_fields_set (tensorrt_llm.llmapi.attentiondpconfig property)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.autodecodingconfig property)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.buildconfig property)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.cachetransceiverconfig property)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.calibconfig property)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.cudagraphconfig property)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.deepseeksparseattentionconfig property)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.drafttargetdecodingconfig property)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.dynamicbatchconfig property)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.eagledecodingconfig property)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.extendedruntimeperfknobconfig property)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.kvcacheconfig property)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.lookaheaddecodingconfig property)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.medusadecodingconfig property)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.moeconfig property)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.mtpdecodingconfig property)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.ngramdecodingconfig property)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.rocketsparseattentionconfig property)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig property)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.schedulerconfig property)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.torchcompileconfig property)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_fields_set", false]], "model_fields_set (tensorrt_llm.llmapi.userprovideddecodingconfig property)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_fields_set", false]], "model_format (tensorrt_llm.llmapi.torchllmargs property)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.model_format", false]], "model_format (tensorrt_llm.llmapi.trtllmargs property)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.model_format", false]], "model_json_schema() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_json_schema", false]], "model_json_schema() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_json_schema", false]], "model_name (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.model_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_parametrized_name", false]], "model_parametrized_name() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_parametrized_name", false]], "model_post_init() (tensorrt_llm.llmapi.attentiondpconfig method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.buildconfig method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.cachetransceiverconfig method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.calibconfig method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.cudagraphconfig method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.dynamicbatchconfig method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.kvcacheconfig method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.moeconfig method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.schedulerconfig method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.torchcompileconfig method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_post_init", false]], "model_post_init() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.model_post_init", false]], "model_rebuild() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_rebuild", false]], "model_rebuild() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_rebuild", false]], "model_validate() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_validate", false]], "model_validate() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_validate", false]], "model_validate_json() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_validate_json", false]], "model_validate_json() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_validate_json", false]], "model_validate_strings() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.model_validate_strings", false]], "model_validate_strings() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.model_validate_strings", false]], "modelconfig (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.ModelConfig", false]], "modelrunner (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.ModelRunner", false]], "modelrunnercpp (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp", false]], "module": [[136, "module-tensorrt_llm", false], [136, "module-tensorrt_llm.functional", false], [137, "module-tensorrt_llm", false], [137, "module-tensorrt_llm.layers.activation", false], [137, "module-tensorrt_llm.layers.attention", false], [137, "module-tensorrt_llm.layers.cast", false], [137, "module-tensorrt_llm.layers.conv", false], [137, "module-tensorrt_llm.layers.embedding", false], [137, "module-tensorrt_llm.layers.linear", false], [137, "module-tensorrt_llm.layers.mlp", false], [137, "module-tensorrt_llm.layers.normalization", false], [137, "module-tensorrt_llm.layers.pooling", false], [138, "module-tensorrt_llm", false], [138, "module-tensorrt_llm.models", false], [139, "module-tensorrt_llm", false], [139, "module-tensorrt_llm.plugin", false], [140, "module-tensorrt_llm", false], [140, "module-tensorrt_llm.quantization", false], [141, "module-tensorrt_llm", false], [141, "module-tensorrt_llm.runtime", false]], "modulo() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.modulo", false]], "moe (tensorrt_llm.functional.sidestreamidtype attribute)": [[136, "tensorrt_llm.functional.SideStreamIDType.moe", false]], "moe_cluster_parallel_size (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.moe_cluster_parallel_size", false]], "moe_cluster_parallel_size (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.moe_cluster_parallel_size", false]], "moe_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.moe_config", false]], "moe_expert_parallel_size (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.moe_expert_parallel_size", false]], "moe_expert_parallel_size (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.moe_expert_parallel_size", false]], "moe_finalize_allreduce_residual_rms_norm (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.MOE_FINALIZE_ALLREDUCE_RESIDUAL_RMS_NORM", false]], "moe_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.moe_plugin", false]], "moe_tensor_parallel_size (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.moe_tensor_parallel_size", false]], "moe_tensor_parallel_size (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.moe_tensor_parallel_size", false]], "moeallreduceparams (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.MoEAllReduceParams", false]], "moeconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MoeConfig", false]], "moeconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MoeConfig.Config", false]], "monitor_memory (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.monitor_memory", false]], "mpi_session (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.mpi_session", false]], "mpi_session (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.mpi_session", false]], "mpicommsession (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MpiCommSession", false]], "mptforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.MPTForCausalLM", false]], "mptmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.MPTModel", false]], "mrope (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.mrope", false]], "mrope (tensorrt_llm.functional.rotaryscalingtype attribute)": [[136, "tensorrt_llm.functional.RotaryScalingType.mrope", false]], "mropeparams (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.MropeParams", false]], "msg (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.msg", false]], "msg (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.msg", false]], "mtp_eagle_one_model (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.mtp_eagle_one_model", false]], "mtpdecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig", false]], "mtpdecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.Config", false]], "mul() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.mul", false]], "multi_block_mode (tensorrt_llm.llmapi.extendedruntimeperfknobconfig attribute)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.multi_block_mode", false]], "multimodal_embedding_handles (tensorrt_llm.llmapi.disaggregatedparams attribute)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.multimodal_embedding_handles", false]], "multimodal_hashes (tensorrt_llm.llmapi.disaggregatedparams attribute)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.multimodal_hashes", false]], "multimodalencoder (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder", false]], "multimodalmodelrunner (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner", false]], "multiple_profiles (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.multiple_profiles", false]], "multiply_and_lora() (tensorrt_llm.layers.linear.linearbase method)": [[137, "tensorrt_llm.layers.linear.LinearBase.multiply_and_lora", false]], "multiply_collect() (tensorrt_llm.layers.linear.linearbase method)": [[137, "tensorrt_llm.layers.linear.LinearBase.multiply_collect", false]], "multiply_collect() (tensorrt_llm.layers.linear.rowlinear method)": [[137, "tensorrt_llm.layers.linear.RowLinear.multiply_collect", false]], "n (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.n", false]], "name (tensorrt_llm.functional.tensor property)": [[136, "tensorrt_llm.functional.Tensor.name", false]], "name (tensorrt_llm.llmapi.lorarequest property)": [[150, "tensorrt_llm.llmapi.LoRARequest.name", false]], "name (tensorrt_llm.runtime.tensorinfo attribute)": [[141, "tensorrt_llm.runtime.TensorInfo.name", false]], "native_quant_flow (tensorrt_llm.models.gemmaforcausallm attribute)": [[138, "tensorrt_llm.models.GemmaForCausalLM.NATIVE_QUANT_FLOW", false]], "nccl (tensorrt_llm.functional.allreducestrategy attribute)": [[136, "tensorrt_llm.functional.AllReduceStrategy.NCCL", false]], "nccl_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.nccl_plugin", false]], "nccl_symmetric (tensorrt_llm.functional.allreducestrategy attribute)": [[136, "tensorrt_llm.functional.AllReduceStrategy.NCCL_SYMMETRIC", false]], "ndim() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.ndim", false]], "network (tensorrt_llm.functional.tensor property)": [[136, "tensorrt_llm.functional.Tensor.network", false]], "next_medusa_input_ids() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.next_medusa_input_ids", false]], "ngram (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.NGRAM", false]], "ngramdecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig", false]], "ngramdecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.Config", false]], "no_quant (tensorrt_llm.llmapi.quantalgo attribute)": [[150, "tensorrt_llm.llmapi.QuantAlgo.NO_QUANT", false]], "no_repeat_ngram_size (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.no_repeat_ngram_size", false]], "no_repeat_ngram_size (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.no_repeat_ngram_size", false]], "non_gated_version() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.non_gated_version", false]], "none (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.NONE", false]], "none (tensorrt_llm.functional.rotaryscalingtype attribute)": [[136, "tensorrt_llm.functional.RotaryScalingType.none", false]], "none (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.NONE", false]], "nonzero() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.nonzero", false]], "norm_quant_fusion (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.norm_quant_fusion", false]], "normalize_log_probs (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.normalize_log_probs", false]], "not_op() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.not_op", false]], "num_beams (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.num_beams", false]], "num_capture_layers (tensorrt_llm.llmapi.eagledecodingconfig property)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.num_capture_layers", false]], "num_capture_layers (tensorrt_llm.llmapi.mtpdecodingconfig property)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.num_capture_layers", false]], "num_capture_layers (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig property)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.num_capture_layers", false]], "num_draft_tokens (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.num_draft_tokens", false]], "num_eagle_layers (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.num_eagle_layers", false]], "num_heads (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.num_heads", false]], "num_heads (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.num_heads", false]], "num_heads (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.num_heads", false]], "num_heads (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.num_heads", false]], "num_kv_heads (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.num_kv_heads", false]], "num_kv_heads_per_cross_attn_layer (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.num_kv_heads_per_cross_attn_layer", false]], "num_kv_heads_per_layer (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.num_kv_heads_per_layer", false]], "num_layers (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.num_layers", false]], "num_layers (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.num_layers", false]], "num_layers (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.num_layers", false]], "num_layers (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.num_layers", false]], "num_medusa_heads (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.num_medusa_heads", false]], "num_medusa_heads (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.num_medusa_heads", false]], "num_medusa_heads (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.num_medusa_heads", false]], "num_nextn_predict_layers (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.num_nextn_predict_layers", false]], "num_nextn_predict_layers_from_model_config (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.num_nextn_predict_layers_from_model_config", false]], "num_postprocess_workers (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.num_postprocess_workers", false]], "num_postprocess_workers (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.num_postprocess_workers", false]], "num_return_sequences (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.num_return_sequences", false]], "numel() (tensorrt_llm.runtime.tensorinfo method)": [[141, "tensorrt_llm.runtime.TensorInfo.numel", false]], "nvfp4 (tensorrt_llm.llmapi.quantalgo attribute)": [[150, "tensorrt_llm.llmapi.QuantAlgo.NVFP4", false]], "nvinfer1 (c++ type)": [[1, "_CPPv48nvinfer1", false]], "onboard_blocks (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.onboard_blocks", false]], "oneshot (tensorrt_llm.functional.allreducestrategy attribute)": [[136, "tensorrt_llm.functional.AllReduceStrategy.ONESHOT", false]], "op_and() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.op_and", false]], "op_or() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.op_or", false]], "op_xor() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.op_xor", false]], "opaque_state (tensorrt_llm.llmapi.disaggregatedparams attribute)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.opaque_state", false]], "opt_batch_size (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.opt_batch_size", false]], "opt_num_tokens (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.opt_num_tokens", false]], "optforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.OPTForCausalLM", false]], "optmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.OPTModel", false]], "orchestrator_type (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.orchestrator_type", false]], "orchestrator_type (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.orchestrator_type", false]], "otlp_traces_endpoint (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.otlp_traces_endpoint", false]], "otlp_traces_endpoint (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.otlp_traces_endpoint", false]], "outer() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.outer", false]], "output_cum_log_probs (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.output_cum_log_probs", false]], "output_directory (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.output_directory", false]], "output_log_probs (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.output_log_probs", false]], "output_sequence_lengths (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.output_sequence_lengths", false]], "output_timing_cache (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.output_timing_cache", false]], "outputs (tensorrt_llm.llmapi.requestoutput attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.outputs", false]], "outputs (tensorrt_llm.llmapi.requestoutput property)": [[150, "id9", false]], "pad() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.pad", false]], "pad_id (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.pad_id", false]], "pad_id (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.pad_id", false]], "padding (tensorrt_llm.functional.attentionmasktype attribute)": [[136, "tensorrt_llm.functional.AttentionMaskType.padding", false]], "page_size (tensorrt_llm.llmapi.rocketsparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.page_size", false]], "paged_kv_cache (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.paged_kv_cache", false]], "paged_kv_cache (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.paged_kv_cache", false]], "paged_state (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.paged_state", false]], "paged_state (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.paged_state", false]], "paged_state (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.paged_state", false]], "parallel_config (tensorrt_llm.llmapi.torchllmargs property)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.parallel_config", false]], "parallel_config (tensorrt_llm.llmapi.trtllmargs property)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.parallel_config", false]], "params_imply_greedy_decoding() (tensorrt_llm.llmapi.samplingparams static method)": [[150, "tensorrt_llm.llmapi.SamplingParams.params_imply_greedy_decoding", false]], "parse_file() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.parse_file", false]], "parse_file() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.parse_file", false]], "parse_obj() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.parse_obj", false]], "parse_obj() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.parse_obj", false]], "parse_raw() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.parse_raw", false]], "parse_raw() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.parse_raw", false]], "partition() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.partition", false]], "partition() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.partition", false]], "partition() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.partition", false]], "partition() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.partition", false]], "path (tensorrt_llm.llmapi.lorarequest property)": [[150, "tensorrt_llm.llmapi.LoRARequest.path", false]], "peft_cache_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.peft_cache_config", false]], "peft_cache_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.peft_cache_config", false]], "perf_metrics_max_requests (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.perf_metrics_max_requests", false]], "permute() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.permute", false]], "permute() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.permute", false]], "phi3forcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.Phi3ForCausalLM", false]], "phi3model (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.Phi3Model", false]], "phiforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.PhiForCausalLM", false]], "phimodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.PhiModel", false]], "pipeline_parallel_size (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.pipeline_parallel_size", false]], "pipeline_parallel_size (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.pipeline_parallel_size", false]], "pixartalphatextprojection (class in tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.PixArtAlphaTextProjection", false]], "plugin_config (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.plugin_config", false]], "positionembeddingtype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.PositionEmbeddingType", false]], "post_layernorm (tensorrt_llm.functional.layernormpositiontype attribute)": [[136, "tensorrt_llm.functional.LayerNormPositionType.post_layernorm", false]], "posterior_threshold (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.posterior_threshold", false]], "postproc_params (tensorrt_llm.llmapi.requestoutput.postprocworker.input attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Input.postproc_params", false]], "postprocess() (tensorrt_llm.layers.attention.attention method)": [[137, "tensorrt_llm.layers.attention.Attention.postprocess", false]], "postprocess() (tensorrt_llm.layers.attention.deepseekv2attention method)": [[137, "tensorrt_llm.layers.attention.DeepseekV2Attention.postprocess", false]], "postprocess() (tensorrt_llm.layers.embedding.embedding method)": [[137, "tensorrt_llm.layers.embedding.Embedding.postprocess", false]], "postprocess() (tensorrt_llm.layers.linear.linear method)": [[137, "tensorrt_llm.layers.linear.Linear.postprocess", false]], "postprocess_tokenizer_dir (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.postprocess_tokenizer_dir", false]], "postprocess_tokenizer_dir (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.postprocess_tokenizer_dir", false]], "pow() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.pow", false]], "pp_communicate_final_output_ids() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.pp_communicate_final_output_ids", false]], "pp_communicate_new_tokens() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.pp_communicate_new_tokens", false]], "pp_partition (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.pp_partition", false]], "pp_partition (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.pp_partition", false]], "pp_reduce_scatter (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.pp_reduce_scatter", false]], "pre_layernorm (tensorrt_llm.functional.layernormpositiontype attribute)": [[136, "tensorrt_llm.functional.LayerNormPositionType.pre_layernorm", false]], "pre_quant_scale (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.pre_quant_scale", false]], "precompute_relative_attention_bias() (tensorrt_llm.models.decodermodel method)": [[138, "tensorrt_llm.models.DecoderModel.precompute_relative_attention_bias", false]], "precompute_relative_attention_bias() (tensorrt_llm.models.encodermodel method)": [[138, "tensorrt_llm.models.EncoderModel.precompute_relative_attention_bias", false]], "precompute_relative_attention_bias() (tensorrt_llm.models.whisperencoder method)": [[138, "tensorrt_llm.models.WhisperEncoder.precompute_relative_attention_bias", false]], "prepare_inputs() (tensorrt_llm.models.chatglmforcausallm method)": [[138, "tensorrt_llm.models.ChatGLMForCausalLM.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.decodermodel method)": [[138, "tensorrt_llm.models.DecoderModel.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.dit method)": [[138, "tensorrt_llm.models.DiT.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.eagleforcausallm method)": [[138, "tensorrt_llm.models.EagleForCausalLM.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.encodermodel method)": [[138, "tensorrt_llm.models.EncoderModel.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.llavanextvisionwrapper method)": [[138, "tensorrt_llm.models.LlavaNextVisionWrapper.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.mambaforcausallm method)": [[138, "tensorrt_llm.models.MambaForCausalLM.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.mllamaforcausallm method)": [[138, "tensorrt_llm.models.MLLaMAForCausalLM.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.pretrainedmodel method)": [[138, "tensorrt_llm.models.PretrainedModel.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.recurrentgemmaforcausallm method)": [[138, "tensorrt_llm.models.RecurrentGemmaForCausalLM.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.sd3transformer2dmodel method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.prepare_inputs", false]], "prepare_inputs() (tensorrt_llm.models.whisperencoder method)": [[138, "tensorrt_llm.models.WhisperEncoder.prepare_inputs", false]], "prepare_position_ids_for_cogvlm() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.prepare_position_ids_for_cogvlm", false]], "prepare_recurrent_inputs() (tensorrt_llm.models.recurrentgemmaforcausallm method)": [[138, "tensorrt_llm.models.RecurrentGemmaForCausalLM.prepare_recurrent_inputs", false]], "preprocess() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.preprocess", false]], "presence_penalty (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.presence_penalty", false]], "presence_penalty (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.presence_penalty", false]], "pretrainedconfig (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.PretrainedConfig", false]], "pretrainedmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.PretrainedModel", false]], "print_iter_log (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.print_iter_log", false]], "priority (tensorrt_llm.llmapi.kvcacheretentionconfig.tokenrangeretentionconfig property)": [[150, "tensorrt_llm.llmapi.KvCacheRetentionConfig.TokenRangeRetentionConfig.priority", false]], "process_input() (tensorrt_llm.runtime.encdecmodelrunner method)": [[141, "tensorrt_llm.runtime.EncDecModelRunner.process_input", false]], "process_logits_including_draft() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.process_logits_including_draft", false]], "prod() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.prod", false]], "profiler (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.profiler", false]], "profiling_verbosity (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.profiling_verbosity", false]], "prompt (tensorrt_llm.llmapi.requestoutput attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.prompt", false]], "prompt (tensorrt_llm.llmapi.requestoutput property)": [[150, "id10", false]], "prompt_budget (tensorrt_llm.llmapi.rocketsparseattentionconfig attribute)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.prompt_budget", false]], "prompt_ignore_length (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.prompt_ignore_length", false]], "prompt_ignore_length (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.prompt_ignore_length", false]], "prompt_logprobs (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.prompt_logprobs", false]], "prompt_logprobs (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.prompt_logprobs", false]], "prompt_token_ids (tensorrt_llm.llmapi.requestoutput attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.prompt_token_ids", false]], "prompt_token_ids (tensorrt_llm.llmapi.requestoutput property)": [[150, "id11", false]], "prompttuningembedding (class in tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.PromptTuningEmbedding", false]], "ptuning_setup() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.ptuning_setup", false]], "ptuning_setup_fuyu() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.ptuning_setup_fuyu", false]], "ptuning_setup_llava_next() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.ptuning_setup_llava_next", false]], "ptuning_setup_phi3() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.ptuning_setup_phi3", false]], "ptuning_setup_pixtral() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.ptuning_setup_pixtral", false]], "pybind_equals() (tensorrt_llm.llmapi.cachetransceiverconfig static method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.pybind_equals", false]], "pybind_equals() (tensorrt_llm.llmapi.dynamicbatchconfig static method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.pybind_equals", false]], "pybind_equals() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig static method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.pybind_equals", false]], "pybind_equals() (tensorrt_llm.llmapi.kvcacheconfig static method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.pybind_equals", false]], "pybind_equals() (tensorrt_llm.llmapi.lookaheaddecodingconfig static method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.pybind_equals", false]], "pybind_equals() (tensorrt_llm.llmapi.schedulerconfig static method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.pybind_equals", false]], "python_e2e (tensorrt_llm.runtime.multimodalmodelrunner property)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.python_e2e", false]], "qserve_gemm_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.qserve_gemm_plugin", false]], "quant_algo (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.quant_algo", false]], "quant_algo (tensorrt_llm.models.pretrainedconfig property)": [[138, "tensorrt_llm.models.PretrainedConfig.quant_algo", false]], "quant_config (tensorrt_llm.llmapi.torchllmargs property)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.quant_config", false]], "quant_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.quant_config", false]], "quant_mode (tensorrt_llm.llmapi.quantconfig property)": [[150, "tensorrt_llm.llmapi.QuantConfig.quant_mode", false]], "quant_mode (tensorrt_llm.models.pretrainedconfig property)": [[138, "tensorrt_llm.models.PretrainedConfig.quant_mode", false]], "quant_mode (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.quant_mode", false]], "quant_mode (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.quant_mode", false]], "quantalgo (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.QuantAlgo", false]], "quantalgo (class in tensorrt_llm.quantization)": [[140, "tensorrt_llm.quantization.QuantAlgo", false]], "quantconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.QuantConfig", false]], "quantize() (tensorrt_llm.models.baichuanforcausallm class method)": [[138, "tensorrt_llm.models.BaichuanForCausalLM.quantize", false]], "quantize() (tensorrt_llm.models.chatglmforcausallm class method)": [[138, "tensorrt_llm.models.ChatGLMForCausalLM.quantize", false]], "quantize() (tensorrt_llm.models.cogvlmforcausallm class method)": [[138, "tensorrt_llm.models.CogVLMForCausalLM.quantize", false]], "quantize() (tensorrt_llm.models.gemmaforcausallm class method)": [[138, "tensorrt_llm.models.GemmaForCausalLM.quantize", false]], "quantize() (tensorrt_llm.models.gptforcausallm class method)": [[138, "tensorrt_llm.models.GPTForCausalLM.quantize", false]], "quantize() (tensorrt_llm.models.llamaforcausallm class method)": [[138, "tensorrt_llm.models.LLaMAForCausalLM.quantize", false]], "quantize() (tensorrt_llm.models.pretrainedmodel class method)": [[138, "tensorrt_llm.models.PretrainedModel.quantize", false]], "quantize_and_export() (in module tensorrt_llm.quantization)": [[140, "tensorrt_llm.quantization.quantize_and_export", false]], "quantize_per_token_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.quantize_per_token_plugin", false]], "quantize_tensor_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.quantize_tensor_plugin", false]], "quantmode (class in tensorrt_llm.quantization)": [[140, "tensorrt_llm.quantization.QuantMode", false]], "quick_gelu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.quick_gelu", false]], "qwenforcausallmgenerationsession (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.QWenForCausalLMGenerationSession", false]], "rand() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.rand", false]], "random_seed (tensorrt_llm.llmapi.calibconfig attribute)": [[150, "tensorrt_llm.llmapi.CalibConfig.random_seed", false]], "random_seed (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.random_seed", false]], "rank() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.rank", false]], "ray_worker_extension_cls (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.ray_worker_extension_cls", false]], "rearrange() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.rearrange", false]], "reasoning_parser (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.reasoning_parser", false]], "reasoning_parser (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.reasoning_parser", false]], "record_stats() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.record_stats", false]], "recurrentgemmaforcausallm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.RecurrentGemmaForCausalLM", false]], "recv() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.recv", false]], "redrafter_draft_len_per_beam (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.redrafter_draft_len_per_beam", false]], "redrafter_num_beams (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.redrafter_num_beams", false]], "redrafterforllamalm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.ReDrafterForLLaMALM", false]], "redrafterforqwenlm (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.ReDrafterForQWenLM", false]], "reduce() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.reduce", false]], "reduce_fusion (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.reduce_fusion", false]], "reduce_scatter() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.reduce_scatter", false]], "regex (tensorrt_llm.llmapi.guideddecodingparams attribute)": [[150, "tensorrt_llm.llmapi.GuidedDecodingParams.regex", false]], "relative (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.relative", false]], "relaxed_delta (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.relaxed_delta", false]], "relaxed_topk (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.relaxed_topk", false]], "release() (tensorrt_llm.models.pretrainedmodel method)": [[138, "tensorrt_llm.models.PretrainedModel.release", false]], "relu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.relu", false]], "remove_input_padding (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.remove_input_padding", false]], "remove_input_padding (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.remove_input_padding", false]], "remove_input_padding (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.remove_input_padding", false]], "remove_input_padding (tensorrt_llm.runtime.modelrunner property)": [[141, "tensorrt_llm.runtime.ModelRunner.remove_input_padding", false]], "remove_input_padding (tensorrt_llm.runtime.modelrunnercpp property)": [[141, "tensorrt_llm.runtime.ModelRunnerCpp.remove_input_padding", false]], "removeprefix() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.removeprefix", false]], "removeprefix() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.removeprefix", false]], "removeprefix() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.removeprefix", false]], "removeprefix() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.removeprefix", false]], "removesuffix() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.removesuffix", false]], "removesuffix() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.removesuffix", false]], "removesuffix() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.removesuffix", false]], "removesuffix() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.removesuffix", false]], "reorder_kv_cache_for_beam_search() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.reorder_kv_cache_for_beam_search", false]], "repeat() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.repeat", false]], "repeat() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.repeat", false]], "repeat_interleave() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.repeat_interleave", false]], "repetition_penalty (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.repetition_penalty", false]], "repetition_penalty (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.repetition_penalty", false]], "replace() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.replace", false]], "replace() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.replace", false]], "replace() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.replace", false]], "replace() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.replace", false]], "replace_all_uses_with() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.replace_all_uses_with", false]], "request_id (tensorrt_llm.llmapi.requestoutput attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.request_id", false]], "request_id (tensorrt_llm.llmapi.requestoutput property)": [[150, "id12", false]], "request_perf_metrics (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.request_perf_metrics", false]], "request_perf_metrics (tensorrt_llm.llmapi.requestoutput.postprocworker.output attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.request_perf_metrics", false]], "request_stats_max_iterations (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.request_stats_max_iterations", false]], "request_stats_max_iterations (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.request_stats_max_iterations", false]], "request_type (tensorrt_llm.llmapi.disaggregatedparams attribute)": [[150, "tensorrt_llm.llmapi.DisaggregatedParams.request_type", false]], "requesterror (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.RequestError", false]], "requestoutput (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.RequestOutput", false]], "requestoutput.postprocworker (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker", false]], "requestoutput.postprocworker.input (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Input", false]], "requestoutput.postprocworker.output (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output", false]], "res (tensorrt_llm.llmapi.requestoutput.postprocworker.output attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Output.res", false]], "residual_rms_norm (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.RESIDUAL_RMS_NORM", false]], "residual_rms_norm_out_quant_fp8 (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.RESIDUAL_RMS_NORM_OUT_QUANT_FP8", false]], "residual_rms_norm_out_quant_nvfp4 (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.RESIDUAL_RMS_NORM_OUT_QUANT_NVFP4", false]], "residual_rms_norm_quant_fp8 (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.RESIDUAL_RMS_NORM_QUANT_FP8", false]], "residual_rms_norm_quant_nvfp4 (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.RESIDUAL_RMS_NORM_QUANT_NVFP4", false]], "residual_rms_prepost_norm (tensorrt_llm.functional.allreducefusionop attribute)": [[136, "tensorrt_llm.functional.AllReduceFusionOp.RESIDUAL_RMS_PREPOST_NORM", false]], "resource_manager (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.resource_manager", false]], "result() (tensorrt_llm.llmapi.requestoutput method)": [[150, "tensorrt_llm.llmapi.RequestOutput.result", false]], "return_context_logits (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.return_context_logits", false]], "return_dict (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.return_dict", false]], "return_encoder_output (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.return_encoder_output", false]], "return_generation_logits (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.return_generation_logits", false]], "return_perf_metrics (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.return_perf_metrics", false]], "return_perf_metrics (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.return_perf_metrics", false]], "return_perf_metrics (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.return_perf_metrics", false]], "revision (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.revision", false]], "revision (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.revision", false]], "rfind() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.rfind", false]], "rfind() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.rfind", false]], "rfind() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.rfind", false]], "rfind() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.rfind", false]], "rg_lru() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.rg_lru", false]], "rindex() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.rindex", false]], "rindex() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.rindex", false]], "rindex() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.rindex", false]], "rindex() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.rindex", false]], "rjust() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.rjust", false]], "rjust() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.rjust", false]], "rjust() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.rjust", false]], "rjust() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.rjust", false]], "rms_norm() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.rms_norm", false]], "rmsnorm (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.RmsNorm", false]], "rmsnorm (tensorrt_llm.functional.layernormtype attribute)": [[136, "tensorrt_llm.functional.LayerNormType.RmsNorm", false]], "rmsnorm_quantization_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.rmsnorm_quantization_plugin", false]], "rnn_conv_dim_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.rnn_conv_dim_size", false]], "rnn_conv_dim_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.rnn_conv_dim_size", false]], "rnn_head_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.rnn_head_size", false]], "rnn_head_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.rnn_head_size", false]], "rnn_hidden_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.rnn_hidden_size", false]], "rnn_hidden_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.rnn_hidden_size", false]], "robertaforquestionanswering (in module tensorrt_llm.models)": [[138, "tensorrt_llm.models.RobertaForQuestionAnswering", false]], "robertaforsequenceclassification (in module tensorrt_llm.models)": [[138, "tensorrt_llm.models.RobertaForSequenceClassification", false]], "robertamodel (in module tensorrt_llm.models)": [[138, "tensorrt_llm.models.RobertaModel", false]], "rocketsparseattentionconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig", false]], "rocketsparseattentionconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.Config", false]], "rope_gpt_neox (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.rope_gpt_neox", false]], "rope_gptj (tensorrt_llm.functional.positionembeddingtype attribute)": [[136, "tensorrt_llm.functional.PositionEmbeddingType.rope_gptj", false]], "ropeembeddingutils (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils", false]], "rotaryscalingtype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.RotaryScalingType", false]], "rotate_every_two() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.rotate_every_two", false]], "rotate_half() (tensorrt_llm.functional.ropeembeddingutils static method)": [[136, "tensorrt_llm.functional.RopeEmbeddingUtils.rotate_half", false]], "round() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.round", false]], "rowlinear (class in tensorrt_llm.layers.linear)": [[137, "tensorrt_llm.layers.linear.RowLinear", false]], "rpartition() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.rpartition", false]], "rpartition() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.rpartition", false]], "rpartition() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.rpartition", false]], "rpartition() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.rpartition", false]], "rsp (tensorrt_llm.llmapi.requestoutput.postprocworker.input attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Input.rsp", false]], "rsplit() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.rsplit", false]], "rsplit() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.rsplit", false]], "rsplit() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.rsplit", false]], "rsplit() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.rsplit", false]], "rstrip() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.rstrip", false]], "rstrip() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.rstrip", false]], "rstrip() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.rstrip", false]], "rstrip() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.rstrip", false]], "run() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.run", false]], "run() (tensorrt_llm.runtime.session method)": [[141, "tensorrt_llm.runtime.Session.run", false]], "runtime (tensorrt_llm.runtime.generationsession attribute)": [[141, "tensorrt_llm.runtime.GenerationSession.runtime", false]], "runtime (tensorrt_llm.runtime.session property)": [[141, "tensorrt_llm.runtime.Session.runtime", false]], "sampler_type (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.sampler_type", false]], "sampling_params (tensorrt_llm.llmapi.requestoutput.postprocworker.input attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Input.sampling_params", false]], "samplingconfig (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.SamplingConfig", false]], "samplingparams (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.SamplingParams", false]], "save_checkpoint() (tensorrt_llm.models.llavanextvisionwrapper method)": [[138, "tensorrt_llm.models.LlavaNextVisionWrapper.save_checkpoint", false]], "save_checkpoint() (tensorrt_llm.models.pretrainedmodel method)": [[138, "tensorrt_llm.models.PretrainedModel.save_checkpoint", false]], "save_hidden_states (tensorrt_llm.models.speculativedecodingmode attribute)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode.SAVE_HIDDEN_STATES", false]], "savehiddenstatesdecodingconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig", false]], "savehiddenstatesdecodingconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.Config", false]], "scatter() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.scatter", false]], "scatter_nd() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.scatter_nd", false]], "scheduler_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.scheduler_config", false]], "scheduler_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.scheduler_config", false]], "schedulerconfig (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.SchedulerConfig", false]], "schedulerconfig.config (class in tensorrt_llm.llmapi)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.Config", false]], "schema() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.schema", false]], "schema() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.schema", false]], "schema() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.schema", false]], "schema() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.schema", false]], "schema() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.schema", false]], "schema() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.schema", false]], "schema() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.schema", false]], "schema() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.schema", false]], "schema() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.schema", false]], "schema() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.schema", false]], "schema() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.schema", false]], "schema() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.schema", false]], "schema() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.schema", false]], "schema() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.schema", false]], "schema() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.schema", false]], "schema_json() (tensorrt_llm.llmapi.attentiondpconfig class method)": [[150, "tensorrt_llm.llmapi.AttentionDpConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.autodecodingconfig class method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.buildconfig class method)": [[150, "tensorrt_llm.llmapi.BuildConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.cachetransceiverconfig class method)": [[150, "tensorrt_llm.llmapi.CacheTransceiverConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.calibconfig class method)": [[150, "tensorrt_llm.llmapi.CalibConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.cudagraphconfig class method)": [[150, "tensorrt_llm.llmapi.CudaGraphConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.deepseeksparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.drafttargetdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.dynamicbatchconfig class method)": [[150, "tensorrt_llm.llmapi.DynamicBatchConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.eagledecodingconfig class method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.extendedruntimeperfknobconfig class method)": [[150, "tensorrt_llm.llmapi.ExtendedRuntimePerfKnobConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.kvcacheconfig class method)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.lookaheaddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.medusadecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.moeconfig class method)": [[150, "tensorrt_llm.llmapi.MoeConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.mtpdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.ngramdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.rocketsparseattentionconfig class method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig class method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.schedulerconfig class method)": [[150, "tensorrt_llm.llmapi.SchedulerConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.torchcompileconfig class method)": [[150, "tensorrt_llm.llmapi.TorchCompileConfig.schema_json", false]], "schema_json() (tensorrt_llm.llmapi.userprovideddecodingconfig class method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.schema_json", false]], "sd35adalayernormzerox (class in tensorrt_llm.layers.normalization)": [[137, "tensorrt_llm.layers.normalization.SD35AdaLayerNormZeroX", false]], "sd3patchembed (class in tensorrt_llm.layers.embedding)": [[137, "tensorrt_llm.layers.embedding.SD3PatchEmbed", false]], "sd3transformer2dmodel (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.SD3Transformer2DModel", false]], "secondary_offload_min_priority (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.secondary_offload_min_priority", false]], "seed (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.seed", false]], "select() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.select", false]], "select() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.select", false]], "selective_scan() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.selective_scan", false]], "send() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.send", false]], "serialize_engine() (tensorrt_llm.runtime.modelrunner method)": [[141, "tensorrt_llm.runtime.ModelRunner.serialize_engine", false]], "session (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.Session", false]], "set_attn_processor() (tensorrt_llm.models.sd3transformer2dmodel method)": [[138, "tensorrt_llm.models.SD3Transformer2DModel.set_attn_processor", false]], "set_context_fmha() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.set_context_fmha", false]], "set_default_max_input_len() (tensorrt_llm.llmapi.torchllmargs method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.set_default_max_input_len", false]], "set_default_max_input_len() (tensorrt_llm.llmapi.trtllmargs method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.set_default_max_input_len", false]], "set_dora_plugin() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.set_dora_plugin", false]], "set_fp8_rowwise_quant_plugins() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.set_fp8_rowwise_quant_plugins", false]], "set_from_optional (c macro)": [[1, "c.SET_FROM_OPTIONAL", false]], "set_if_not_exist() (tensorrt_llm.models.pretrainedconfig method)": [[138, "tensorrt_llm.models.PretrainedConfig.set_if_not_exist", false]], "set_lora_plugin() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.set_lora_plugin", false]], "set_nccl_plugin() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.set_nccl_plugin", false]], "set_qserve_plugins() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.set_qserve_plugins", false]], "set_rank() (tensorrt_llm.models.pretrainedconfig method)": [[138, "tensorrt_llm.models.PretrainedConfig.set_rank", false]], "set_rel_attn_table() (tensorrt_llm.layers.attention.attention method)": [[137, "tensorrt_llm.layers.attention.Attention.set_rel_attn_table", false]], "set_runtime_knobs_from_build_config() (tensorrt_llm.llmapi.torchllmargs method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.set_runtime_knobs_from_build_config", false]], "set_runtime_knobs_from_build_config() (tensorrt_llm.llmapi.trtllmargs method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.set_runtime_knobs_from_build_config", false]], "set_shapes() (tensorrt_llm.runtime.session method)": [[141, "tensorrt_llm.runtime.Session.set_shapes", false]], "set_smooth_quant_plugins() (tensorrt_llm.plugin.pluginconfig method)": [[139, "tensorrt_llm.plugin.PluginConfig.set_smooth_quant_plugins", false]], "setup() (tensorrt_llm.runtime.generationsession method)": [[141, "tensorrt_llm.runtime.GenerationSession.setup", false]], "setup_embedding_parallel_mode() (tensorrt_llm.llmapi.trtllmargs method)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.setup_embedding_parallel_mode", false]], "setup_fake_prompts() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.setup_fake_prompts", false]], "setup_fake_prompts_qwen2vl() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.setup_fake_prompts_qwen2vl", false]], "setup_fake_prompts_vila() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.setup_fake_prompts_vila", false]], "setup_inputs() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.setup_inputs", false]], "shape (tensorrt_llm.functional.tensor property)": [[136, "tensorrt_llm.functional.Tensor.shape", false]], "shape (tensorrt_llm.runtime.tensorinfo attribute)": [[141, "tensorrt_llm.runtime.TensorInfo.shape", false]], "shape() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.shape", false]], "shutdown() (tensorrt_llm.llmapi.llm method)": [[150, "tensorrt_llm.llmapi.LLM.shutdown", false]], "shutdown() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.shutdown", false]], "shutdown() (tensorrt_llm.llmapi.multimodalencoder method)": [[150, "tensorrt_llm.llmapi.MultimodalEncoder.shutdown", false]], "shutdown_abort() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.shutdown_abort", false]], "sidestreamidtype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.SideStreamIDType", false]], "sigmoid() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.sigmoid", false]], "silu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.silu", false]], "sin() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.sin", false]], "sink_token_length (tensorrt_llm.llmapi.kvcacheconfig attribute)": [[150, "tensorrt_llm.llmapi.KvCacheConfig.sink_token_length", false]], "sink_token_length (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.sink_token_length", false]], "size (tensorrt_llm.functional.sliceinputtype attribute)": [[136, "tensorrt_llm.functional.SliceInputType.size", false]], "size() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.size", false]], "skip_cross_attn_blocks (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.skip_cross_attn_blocks", false]], "skip_cross_kv (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.skip_cross_kv", false]], "skip_special_tokens (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.skip_special_tokens", false]], "skip_tokenizer_init (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.skip_tokenizer_init", false]], "skip_tokenizer_init (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.skip_tokenizer_init", false]], "slice() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.slice", false]], "sliceinputtype (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.SliceInputType", false]], "sliding_window_causal (tensorrt_llm.functional.attentionmasktype attribute)": [[136, "tensorrt_llm.functional.AttentionMaskType.sliding_window_causal", false]], "smooth_quant_gemm_plugin (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.smooth_quant_gemm_plugin", false]], "smooth_quant_plugins (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.smooth_quant_plugins", false]], "smoothquant_val (tensorrt_llm.llmapi.quantconfig attribute)": [[150, "tensorrt_llm.llmapi.QuantConfig.smoothquant_val", false]], "softmax() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.softmax", false]], "softplus() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.softplus", false]], "spaces_between_special_tokens (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.spaces_between_special_tokens", false]], "sparse_attention_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.sparse_attention_config", false]], "sparse_attention_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.sparse_attention_config", false]], "spec_dec_mode (tensorrt_llm.llmapi.autodecodingconfig property)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.drafttargetdecodingconfig property)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.eagledecodingconfig property)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.lookaheaddecodingconfig property)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.medusadecodingconfig property)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.mtpdecodingconfig property)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.ngramdecodingconfig property)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig property)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.spec_dec_mode", false]], "spec_dec_mode (tensorrt_llm.llmapi.userprovideddecodingconfig property)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.spec_dec_mode", false]], "specdecodingparams (class in tensorrt_llm.layers.attention)": [[137, "tensorrt_llm.layers.attention.SpecDecodingParams", false]], "speculative_config (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.speculative_config", false]], "speculative_config (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.speculative_config", false]], "speculative_decoding_mode (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.speculative_decoding_mode", false]], "speculative_model_dir (tensorrt_llm.llmapi.autodecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.drafttargetdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.eagledecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.lookaheaddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.medusadecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.mtpdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.ngramdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.torchllmargs property)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.trtllmargs property)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.speculative_model_dir", false]], "speculative_model_dir (tensorrt_llm.llmapi.userprovideddecodingconfig attribute)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.speculative_model_dir", false]], "speculative_model_format (tensorrt_llm.llmapi.torchllmargs property)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.speculative_model_format", false]], "speculative_model_format (tensorrt_llm.llmapi.trtllmargs property)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.speculative_model_format", false]], "speculativedecodingmode (class in tensorrt_llm.models)": [[138, "tensorrt_llm.models.SpeculativeDecodingMode", false]], "split() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.split", false]], "split() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.split", false]], "split() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.split", false]], "split() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.split", false]], "split() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.split", false]], "split() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.split", false]], "split_prompt_by_images() (tensorrt_llm.runtime.multimodalmodelrunner method)": [[141, "tensorrt_llm.runtime.MultimodalModelRunner.split_prompt_by_images", false]], "splitlines() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.splitlines", false]], "splitlines() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.splitlines", false]], "splitlines() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.splitlines", false]], "splitlines() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.splitlines", false]], "sqrt() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.sqrt", false]], "sqrt() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.sqrt", false]], "squared_relu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.squared_relu", false]], "squeeze() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.squeeze", false]], "squeeze() (tensorrt_llm.functional.tensor method)": [[136, "tensorrt_llm.functional.Tensor.squeeze", false]], "squeeze() (tensorrt_llm.runtime.tensorinfo method)": [[141, "tensorrt_llm.runtime.TensorInfo.squeeze", false]], "stack() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.stack", false]], "start (tensorrt_llm.functional.sliceinputtype attribute)": [[136, "tensorrt_llm.functional.SliceInputType.start", false]], "start() (tensorrt_llm.llmapi.requestoutput.postprocworker method)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.start", false]], "startswith() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.startswith", false]], "startswith() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.startswith", false]], "startswith() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.startswith", false]], "startswith() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.startswith", false]], "state_dtype (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.state_dtype", false]], "state_dtype (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.state_dtype", false]], "state_size (tensorrt_llm.runtime.generationsession property)": [[141, "tensorrt_llm.runtime.GenerationSession.state_size", false]], "state_size (tensorrt_llm.runtime.modelconfig attribute)": [[141, "tensorrt_llm.runtime.ModelConfig.state_size", false]], "static (tensorrt_llm.llmapi.batchingtype attribute)": [[150, "tensorrt_llm.llmapi.BatchingType.STATIC", false]], "static_batch (tensorrt_llm.llmapi.capacityschedulerpolicy attribute)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.STATIC_BATCH", false]], "step() (tensorrt_llm.runtime.kvcachemanager method)": [[141, "tensorrt_llm.runtime.KVCacheManager.step", false]], "stop (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.stop", false]], "stop_reason (tensorrt_llm.llmapi.completionoutput attribute)": [[150, "tensorrt_llm.llmapi.CompletionOutput.stop_reason", false]], "stop_token_ids (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.stop_token_ids", false]], "stop_words_list (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.stop_words_list", false]], "stoppingcriteria (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.StoppingCriteria", false]], "stoppingcriterialist (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.StoppingCriteriaList", false]], "stream_interval (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.stream_interval", false]], "streaming (tensorrt_llm.llmapi.requestoutput.postprocworker.input attribute)": [[150, "tensorrt_llm.llmapi.RequestOutput.PostprocWorker.Input.streaming", false]], "streamingllm (tensorrt_llm.plugin.pluginconfig attribute)": [[139, "tensorrt_llm.plugin.PluginConfig.streamingllm", false]], "stride (tensorrt_llm.functional.sliceinputtype attribute)": [[136, "tensorrt_llm.functional.SliceInputType.stride", false]], "strip() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.strip", false]], "strip() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.strip", false]], "strip() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.strip", false]], "strip() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.strip", false]], "strongly_typed (tensorrt_llm.llmapi.buildconfig attribute)": [[150, "tensorrt_llm.llmapi.BuildConfig.strongly_typed", false]], "structural_tag (tensorrt_llm.llmapi.guideddecodingparams attribute)": [[150, "tensorrt_llm.llmapi.GuidedDecodingParams.structural_tag", false]], "sub() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.sub", false]], "submit() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.submit", false]], "submit_sync() (tensorrt_llm.llmapi.mpicommsession method)": [[150, "tensorrt_llm.llmapi.MpiCommSession.submit_sync", false]], "sum() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.sum", false]], "supports_backend() (tensorrt_llm.llmapi.autodecodingconfig method)": [[150, "tensorrt_llm.llmapi.AutoDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.deepseeksparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.DeepSeekSparseAttentionConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.drafttargetdecodingconfig method)": [[150, "tensorrt_llm.llmapi.DraftTargetDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.eagledecodingconfig method)": [[150, "tensorrt_llm.llmapi.EagleDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.lookaheaddecodingconfig method)": [[150, "tensorrt_llm.llmapi.LookaheadDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.medusadecodingconfig method)": [[150, "tensorrt_llm.llmapi.MedusaDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.mtpdecodingconfig method)": [[150, "tensorrt_llm.llmapi.MTPDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.ngramdecodingconfig method)": [[150, "tensorrt_llm.llmapi.NGramDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.rocketsparseattentionconfig method)": [[150, "tensorrt_llm.llmapi.RocketSparseAttentionConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.savehiddenstatesdecodingconfig method)": [[150, "tensorrt_llm.llmapi.SaveHiddenStatesDecodingConfig.supports_backend", false]], "supports_backend() (tensorrt_llm.llmapi.userprovideddecodingconfig method)": [[150, "tensorrt_llm.llmapi.UserProvidedDecodingConfig.supports_backend", false]], "swapcase() (tensorrt_llm.llmapi.batchingtype method)": [[150, "tensorrt_llm.llmapi.BatchingType.swapcase", false]], "swapcase() (tensorrt_llm.llmapi.capacityschedulerpolicy method)": [[150, "tensorrt_llm.llmapi.CapacitySchedulerPolicy.swapcase", false]], "swapcase() (tensorrt_llm.llmapi.contextchunkingpolicy method)": [[150, "tensorrt_llm.llmapi.ContextChunkingPolicy.swapcase", false]], "swapcase() (tensorrt_llm.llmapi.quantalgo method)": [[150, "tensorrt_llm.llmapi.QuantAlgo.swapcase", false]], "swiglu() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.swiglu", false]], "sync_quant_config_with_kv_cache_config_dtype() (tensorrt_llm.llmapi.torchllmargs method)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.sync_quant_config_with_kv_cache_config_dtype", false]], "tanh() (in module tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.tanh", false]], "temperature (tensorrt_llm.llmapi.samplingparams attribute)": [[150, "tensorrt_llm.llmapi.SamplingParams.temperature", false]], "temperature (tensorrt_llm.runtime.samplingconfig attribute)": [[141, "tensorrt_llm.runtime.SamplingConfig.temperature", false]], "tensor (class in tensorrt_llm.functional)": [[136, "tensorrt_llm.functional.Tensor", false]], "tensor_parallel_size (tensorrt_llm.llmapi.torchllmargs attribute)": [[150, "tensorrt_llm.llmapi.TorchLlmArgs.tensor_parallel_size", false]], "tensor_parallel_size (tensorrt_llm.llmapi.trtllmargs attribute)": [[150, "tensorrt_llm.llmapi.TrtLlmArgs.tensor_parallel_size", false]], "tensorinfo (class in tensorrt_llm.runtime)": [[141, "tensorrt_llm.runtime.TensorInfo", false]], "tensorrt_llm": [[136, "module-tensorrt_llm", false], [137, "module-tensorrt_llm", false], [138, "module-tensorrt_llm", false], [139, "module-tensorrt_llm", false], [140, "module-tensorrt_llm", false], [141, "module-tensorrt_llm", false]], "tensorrt_llm (c++ type)": [[0, "_CPPv412tensorrt_llm", false], [1, "_CPPv412tensorrt_llm", false]], "tensorrt_llm.functional": [[136, "module-tensorrt_llm.functional", false]], "tensorrt_llm.layers.activation": [[137, "module-tensorrt_llm.layers.activation", false]], "tensorrt_llm.layers.attention": [[137, "module-tensorrt_llm.layers.attention", false]], "tensorrt_llm.layers.cast": [[137, "module-tensorrt_llm.layers.cast", false]], "tensorrt_llm.layers.conv": [[137, "module-tensorrt_llm.layers.conv", false]], "tensorrt_llm.layers.embedding": [[137, "module-tensorrt_llm.layers.embedding", false]], "tensorrt_llm.layers.linear": [[137, "module-tensorrt_llm.layers.linear", false]], "tensorrt_llm.layers.mlp": [[137, "module-tensorrt_llm.layers.mlp", false]], "tensorrt_llm.layers.normalization": [[137, "module-tensorrt_llm.layers.normalization", false]], "tensorrt_llm.layers.pooling": [[137, "module-tensorrt_llm.layers.pooling", false]], "tensorrt_llm.models": [[138, "module-tensorrt_llm.models", false]], "tensorrt_llm.plugin": [[139, "module-tensorrt_llm.plugin", false]], "tensorrt_llm.quantization": [[140, "module-tensorrt_llm.quantization", false]], "tensorrt_llm.runtime": [[141, "module-tensorrt_llm.runtime", false]], "tensorrt_llm::batch_manager (c++ type)": [[0, "_CPPv4N12tensorrt_llm13batch_managerE", false], [1, "_CPPv4N12tensorrt_llm13batch_managerE", false]], "tensorrt_llm::batch_manager::kv_cache_manager (c++ type)": [[0, "_CPPv4N12tensorrt_llm13batch_manager16kv_cache_managerE", false]], "tensorrt_llm::executor (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executorE", false]], "tensorrt_llm::executor::additionalmodeloutput (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor21AdditionalModelOutputE", false]], "tensorrt_llm::executor::additionalmodeloutput::additionalmodeloutput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor21AdditionalModelOutput21AdditionalModelOutputENSt6stringEb", false]], "tensorrt_llm::executor::additionalmodeloutput::gathercontext (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21AdditionalModelOutput13gatherContextE", false]], "tensorrt_llm::executor::additionalmodeloutput::name (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21AdditionalModelOutput4nameE", false]], "tensorrt_llm::executor::additionalmodeloutput::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor21AdditionalModelOutputeqERK21AdditionalModelOutput", false]], "tensorrt_llm::executor::additionaloutput (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutputE", false]], "tensorrt_llm::executor::additionaloutput::additionaloutput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutput16AdditionalOutputENSt6stringE6Tensor", false], [0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutput16AdditionalOutputERK16AdditionalOutput", false], [0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutput16AdditionalOutputERR16AdditionalOutput", false]], "tensorrt_llm::executor::additionaloutput::name (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutput4nameE", false]], "tensorrt_llm::executor::additionaloutput::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutputaSERK16AdditionalOutput", false], [0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutputaSERR16AdditionalOutput", false]], "tensorrt_llm::executor::additionaloutput::output (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutput6outputE", false]], "tensorrt_llm::executor::additionaloutput::~additionaloutput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor16AdditionalOutputD0Ev", false]], "tensorrt_llm::executor::batchingtype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor12BatchingTypeE", false]], "tensorrt_llm::executor::batchingtype::kinflight (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12BatchingType9kINFLIGHTE", false]], "tensorrt_llm::executor::batchingtype::kstatic (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12BatchingType7kSTATICE", false]], "tensorrt_llm::executor::beamtokens (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor10BeamTokensE", false]], "tensorrt_llm::executor::bufferview (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor10BufferViewE", false]], "tensorrt_llm::executor::cachesaltidtype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor15CacheSaltIDTypeE", false]], "tensorrt_llm::executor::cachetransceiverconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfigE", false]], "tensorrt_llm::executor::cachetransceiverconfig::backendtype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig11BackendTypeE", false]], "tensorrt_llm::executor::cachetransceiverconfig::backendtype::default (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig11BackendType7DEFAULTE", false]], "tensorrt_llm::executor::cachetransceiverconfig::backendtype::mpi (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig11BackendType3MPIE", false]], "tensorrt_llm::executor::cachetransceiverconfig::backendtype::nixl (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig11BackendType4NIXLE", false]], "tensorrt_llm::executor::cachetransceiverconfig::backendtype::ucx (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig11BackendType3UCXE", false]], "tensorrt_llm::executor::cachetransceiverconfig::cachetransceiverconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig22CacheTransceiverConfigENSt8optionalI11BackendTypeEENSt8optionalI6size_tEENSt8optionalIiEENSt8optionalIiEE", false]], "tensorrt_llm::executor::cachetransceiverconfig::getbackendtype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22CacheTransceiverConfig14getBackendTypeEv", false]], "tensorrt_llm::executor::cachetransceiverconfig::getkvtransfersenderfuturetimeoutms (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22CacheTransceiverConfig34getKvTransferSenderFutureTimeoutMsEv", false]], "tensorrt_llm::executor::cachetransceiverconfig::getkvtransfertimeoutms (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22CacheTransceiverConfig22getKvTransferTimeoutMsEv", false]], "tensorrt_llm::executor::cachetransceiverconfig::getmaxtokensinbuffer (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22CacheTransceiverConfig20getMaxTokensInBufferEv", false]], "tensorrt_llm::executor::cachetransceiverconfig::mbackendtype (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig12mBackendTypeE", false]], "tensorrt_llm::executor::cachetransceiverconfig::mkvtransfersenderfuturetimeoutms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig32mKvTransferSenderFutureTimeoutMsE", false]], "tensorrt_llm::executor::cachetransceiverconfig::mkvtransfertimeoutms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig20mKvTransferTimeoutMsE", false]], "tensorrt_llm::executor::cachetransceiverconfig::mmaxtokensinbuffer (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig18mMaxTokensInBufferE", false]], "tensorrt_llm::executor::cachetransceiverconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22CacheTransceiverConfigeqERK22CacheTransceiverConfig", false]], "tensorrt_llm::executor::cachetransceiverconfig::setbackendtype (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig14setBackendTypeENSt8optionalI11BackendTypeEE", false]], "tensorrt_llm::executor::cachetransceiverconfig::setkvtransfersenderfuturetimeoutms (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig34setKvTransferSenderFutureTimeoutMsENSt8optionalIiEE", false]], "tensorrt_llm::executor::cachetransceiverconfig::setkvtransfertimeoutms (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig22setKvTransferTimeoutMsENSt8optionalIiEE", false]], "tensorrt_llm::executor::cachetransceiverconfig::setmaxtokensinbuffer (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22CacheTransceiverConfig20setMaxTokensInBufferENSt8optionalI6size_tEE", false]], "tensorrt_llm::executor::capacityschedulerpolicy (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor23CapacitySchedulerPolicyE", false]], "tensorrt_llm::executor::capacityschedulerpolicy::kguaranteed_no_evict (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor23CapacitySchedulerPolicy20kGUARANTEED_NO_EVICTE", false]], "tensorrt_llm::executor::capacityschedulerpolicy::kmax_utilization (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor23CapacitySchedulerPolicy16kMAX_UTILIZATIONE", false]], "tensorrt_llm::executor::capacityschedulerpolicy::kstatic_batch (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor23CapacitySchedulerPolicy13kSTATIC_BATCHE", false]], "tensorrt_llm::executor::communicationmode (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor17CommunicationModeE", false]], "tensorrt_llm::executor::communicationmode::kleader (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor17CommunicationMode7kLEADERE", false]], "tensorrt_llm::executor::communicationmode::korchestrator (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor17CommunicationMode13kORCHESTRATORE", false]], "tensorrt_llm::executor::communicationtype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor17CommunicationTypeE", false]], "tensorrt_llm::executor::communicationtype::kmpi (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor17CommunicationType4kMPIE", false]], "tensorrt_llm::executor::contextchunkingpolicy (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor21ContextChunkingPolicyE", false]], "tensorrt_llm::executor::contextchunkingpolicy::kequal_progress (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor21ContextChunkingPolicy15kEQUAL_PROGRESSE", false]], "tensorrt_llm::executor::contextchunkingpolicy::kfirst_come_first_served (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor21ContextChunkingPolicy24kFIRST_COME_FIRST_SERVEDE", false]], "tensorrt_llm::executor::contextphaseparams (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParamsE", false]], "tensorrt_llm::executor::contextphaseparams::contextphaseparams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams18ContextPhaseParamsE9VecTokens13RequestIdTypeNSt8optionalI9VecTokensEE", false], [0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams18ContextPhaseParamsE9VecTokens13RequestIdTypePvNSt8optionalI9VecTokensEE", false], [0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams18ContextPhaseParamsE9VecTokens13RequestIdTypeRKNSt6vectorIcEENSt8optionalI9VecTokensEE", false], [0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams18ContextPhaseParamsERK18ContextPhaseParams", false], [0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams18ContextPhaseParamsERR18ContextPhaseParams", false]], "tensorrt_llm::executor::contextphaseparams::deleter (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams7deleterEPKv", false]], "tensorrt_llm::executor::contextphaseparams::getdrafttokens (c++ function)": [[0, "_CPPv4NKR12tensorrt_llm8executor18ContextPhaseParams14getDraftTokensEv", false]], "tensorrt_llm::executor::contextphaseparams::getfirstgentokens (c++ function)": [[0, "_CPPv4NKR12tensorrt_llm8executor18ContextPhaseParams17getFirstGenTokensEv", false]], "tensorrt_llm::executor::contextphaseparams::getreqid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18ContextPhaseParams8getReqIdEv", false]], "tensorrt_llm::executor::contextphaseparams::getserializedstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18ContextPhaseParams18getSerializedStateEv", false]], "tensorrt_llm::executor::contextphaseparams::getstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams8getStateEv", false], [0, "_CPPv4NK12tensorrt_llm8executor18ContextPhaseParams8getStateEv", false]], "tensorrt_llm::executor::contextphaseparams::mdrafttokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams12mDraftTokensE", false]], "tensorrt_llm::executor::contextphaseparams::mfirstgentokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams15mFirstGenTokensE", false]], "tensorrt_llm::executor::contextphaseparams::mreqid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams6mReqIdE", false]], "tensorrt_llm::executor::contextphaseparams::mstate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams6mStateE", false]], "tensorrt_llm::executor::contextphaseparams::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParamsaSERK18ContextPhaseParams", false], [0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParamsaSERR18ContextPhaseParams", false]], "tensorrt_llm::executor::contextphaseparams::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18ContextPhaseParamseqERK18ContextPhaseParams", false]], "tensorrt_llm::executor::contextphaseparams::popfirstgentokens (c++ function)": [[0, "_CPPv4NO12tensorrt_llm8executor18ContextPhaseParams17popFirstGenTokensEv", false]], "tensorrt_llm::executor::contextphaseparams::releasestate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams12releaseStateEv", false]], "tensorrt_llm::executor::contextphaseparams::requestidtype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams13RequestIdTypeE", false]], "tensorrt_llm::executor::contextphaseparams::stateptr (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParams8StatePtrE", false]], "tensorrt_llm::executor::contextphaseparams::~contextphaseparams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18ContextPhaseParamsD0Ev", false]], "tensorrt_llm::executor::datatransceiverstate (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor20DataTransceiverStateE", false]], "tensorrt_llm::executor::datatransceiverstate::datatransceiverstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20DataTransceiverState20DataTransceiverStateEN8kv_cache10CacheStateEN8kv_cache9CommStateE", false], [0, "_CPPv4N12tensorrt_llm8executor20DataTransceiverState20DataTransceiverStateEv", false]], "tensorrt_llm::executor::datatransceiverstate::getcachestate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20DataTransceiverState13getCacheStateEv", false]], "tensorrt_llm::executor::datatransceiverstate::getcommstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20DataTransceiverState12getCommStateEv", false]], "tensorrt_llm::executor::datatransceiverstate::mcachestate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20DataTransceiverState11mCacheStateE", false]], "tensorrt_llm::executor::datatransceiverstate::mcommstate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20DataTransceiverState10mCommStateE", false]], "tensorrt_llm::executor::datatransceiverstate::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20DataTransceiverStateeqERK20DataTransceiverState", false]], "tensorrt_llm::executor::datatransceiverstate::setcachestate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20DataTransceiverState13setCacheStateEN8kv_cache10CacheStateE", false]], "tensorrt_llm::executor::datatransceiverstate::setcommstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20DataTransceiverState12setCommStateEN8kv_cache9CommStateE", false]], "tensorrt_llm::executor::datatransceiverstate::tostring (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20DataTransceiverState8toStringEv", false]], "tensorrt_llm::executor::datatype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor8DataTypeE", false]], "tensorrt_llm::executor::datatype::kbf16 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType5kBF16E", false]], "tensorrt_llm::executor::datatype::kbool (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType5kBOOLE", false]], "tensorrt_llm::executor::datatype::kfp16 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType5kFP16E", false]], "tensorrt_llm::executor::datatype::kfp32 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType5kFP32E", false]], "tensorrt_llm::executor::datatype::kfp8 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType4kFP8E", false]], "tensorrt_llm::executor::datatype::kint32 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType6kINT32E", false]], "tensorrt_llm::executor::datatype::kint64 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType6kINT64E", false]], "tensorrt_llm::executor::datatype::kint8 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType5kINT8E", false]], "tensorrt_llm::executor::datatype::kuint8 (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType6kUINT8E", false]], "tensorrt_llm::executor::datatype::kunknown (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8DataType8kUNKNOWNE", false]], "tensorrt_llm::executor::debugconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfigE", false]], "tensorrt_llm::executor::debugconfig::debugconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig11DebugConfigEbb9StringVec10SizeType32", false]], "tensorrt_llm::executor::debugconfig::getdebuginputtensors (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11DebugConfig20getDebugInputTensorsEv", false]], "tensorrt_llm::executor::debugconfig::getdebugoutputtensors (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11DebugConfig21getDebugOutputTensorsEv", false]], "tensorrt_llm::executor::debugconfig::getdebugtensornames (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11DebugConfig19getDebugTensorNamesEv", false]], "tensorrt_llm::executor::debugconfig::getdebugtensorsmaxiterations (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11DebugConfig28getDebugTensorsMaxIterationsEv", false]], "tensorrt_llm::executor::debugconfig::mdebuginputtensors (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig18mDebugInputTensorsE", false]], "tensorrt_llm::executor::debugconfig::mdebugoutputtensors (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig19mDebugOutputTensorsE", false]], "tensorrt_llm::executor::debugconfig::mdebugtensornames (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig17mDebugTensorNamesE", false]], "tensorrt_llm::executor::debugconfig::mdebugtensorsmaxiterations (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig26mDebugTensorsMaxIterationsE", false]], "tensorrt_llm::executor::debugconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11DebugConfigeqERK11DebugConfig", false]], "tensorrt_llm::executor::debugconfig::setdebuginputtensors (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig20setDebugInputTensorsEb", false]], "tensorrt_llm::executor::debugconfig::setdebugoutputtensors (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig21setDebugOutputTensorsEb", false]], "tensorrt_llm::executor::debugconfig::setdebugtensornames (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig19setDebugTensorNamesERK9StringVec", false]], "tensorrt_llm::executor::debugconfig::setdebugtensorsmaxiterations (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig28setDebugTensorsMaxIterationsE10SizeType32", false]], "tensorrt_llm::executor::debugconfig::stringvec (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor11DebugConfig9StringVecE", false]], "tensorrt_llm::executor::debugtensorsperiteration (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor24DebugTensorsPerIterationE", false]], "tensorrt_llm::executor::debugtensorsperiteration::debugtensors (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor24DebugTensorsPerIteration12debugTensorsE", false]], "tensorrt_llm::executor::debugtensorsperiteration::iter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor24DebugTensorsPerIteration4iterE", false]], "tensorrt_llm::executor::decodingconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfigE", false]], "tensorrt_llm::executor::decodingconfig::decodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig14DecodingConfigENSt8optionalI12DecodingModeEENSt8optionalI23LookaheadDecodingConfigEENSt8optionalI13MedusaChoicesEENSt8optionalI11EagleConfigEE", false]], "tensorrt_llm::executor::decodingconfig::enableseamlesslookaheaddecoding (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig31enableSeamlessLookaheadDecodingEv", false]], "tensorrt_llm::executor::decodingconfig::getdecodingmode (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14DecodingConfig15getDecodingModeEv", false]], "tensorrt_llm::executor::decodingconfig::geteagleconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14DecodingConfig14getEagleConfigEv", false]], "tensorrt_llm::executor::decodingconfig::getlookaheaddecodingconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14DecodingConfig26getLookaheadDecodingConfigEv", false]], "tensorrt_llm::executor::decodingconfig::getlookaheaddecodingmaxnumrequest (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14DecodingConfig33getLookaheadDecodingMaxNumRequestEv", false]], "tensorrt_llm::executor::decodingconfig::getmedusachoices (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14DecodingConfig16getMedusaChoicesEv", false]], "tensorrt_llm::executor::decodingconfig::mdecodingmode (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig13mDecodingModeE", false]], "tensorrt_llm::executor::decodingconfig::meagleconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig12mEagleConfigE", false]], "tensorrt_llm::executor::decodingconfig::mlookaheaddecodingconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig24mLookaheadDecodingConfigE", false]], "tensorrt_llm::executor::decodingconfig::mlookaheaddecodingmaxnumrequest (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig31mLookaheadDecodingMaxNumRequestE", false]], "tensorrt_llm::executor::decodingconfig::mmedusachoices (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig14mMedusaChoicesE", false]], "tensorrt_llm::executor::decodingconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14DecodingConfigeqERK14DecodingConfig", false]], "tensorrt_llm::executor::decodingconfig::setdecodingmode (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig15setDecodingModeERK12DecodingMode", false]], "tensorrt_llm::executor::decodingconfig::seteagleconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig14setEagleConfigERK11EagleConfig", false]], "tensorrt_llm::executor::decodingconfig::setlookaheaddecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig26setLookaheadDecodingConfigERK23LookaheadDecodingConfig", false]], "tensorrt_llm::executor::decodingconfig::setmedusachoices (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14DecodingConfig16setMedusaChoicesERK13MedusaChoices", false]], "tensorrt_llm::executor::decodingmode (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingModeE", false]], "tensorrt_llm::executor::decodingmode::allbitset (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode9allBitSetE14UnderlyingType", false]], "tensorrt_llm::executor::decodingmode::anybitset (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode9anyBitSetE14UnderlyingType", false]], "tensorrt_llm::executor::decodingmode::auto (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode4AutoEv", false]], "tensorrt_llm::executor::decodingmode::beamsearch (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode10BeamSearchEv", false]], "tensorrt_llm::executor::decodingmode::decodingmode (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode12DecodingModeE14UnderlyingType", false]], "tensorrt_llm::executor::decodingmode::eagle (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode5EagleEv", false]], "tensorrt_llm::executor::decodingmode::explicitdrafttokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode19ExplicitDraftTokensEv", false]], "tensorrt_llm::executor::decodingmode::externaldrafttokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode19ExternalDraftTokensEv", false]], "tensorrt_llm::executor::decodingmode::getname (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode7getNameEv", false]], "tensorrt_llm::executor::decodingmode::getstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode8getStateEv", false]], "tensorrt_llm::executor::decodingmode::isauto (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode6isAutoEv", false]], "tensorrt_llm::executor::decodingmode::isbeamsearch (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode12isBeamSearchEv", false]], "tensorrt_llm::executor::decodingmode::iseagle (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode7isEagleEv", false]], "tensorrt_llm::executor::decodingmode::isexplicitdrafttokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode21isExplicitDraftTokensEv", false]], "tensorrt_llm::executor::decodingmode::isexternaldrafttokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode21isExternalDraftTokensEv", false]], "tensorrt_llm::executor::decodingmode::islookahead (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode11isLookaheadEv", false]], "tensorrt_llm::executor::decodingmode::ismedusa (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode8isMedusaEv", false]], "tensorrt_llm::executor::decodingmode::istopk (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode6isTopKEv", false]], "tensorrt_llm::executor::decodingmode::istopkandtopp (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode13isTopKandTopPEv", false]], "tensorrt_llm::executor::decodingmode::istopkortopp (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode12isTopKorTopPEv", false]], "tensorrt_llm::executor::decodingmode::istopp (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode6isTopPEv", false]], "tensorrt_llm::executor::decodingmode::isusebantokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode14isUseBanTokensEv", false]], "tensorrt_llm::executor::decodingmode::isusebanwords (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode13isUseBanWordsEv", false]], "tensorrt_llm::executor::decodingmode::isuseexpliciteosstop (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode20isUseExplicitEosStopEv", false]], "tensorrt_llm::executor::decodingmode::isusefrequencypenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode21isUseFrequencyPenaltyEv", false]], "tensorrt_llm::executor::decodingmode::isusemaxlengthstop (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode18isUseMaxLengthStopEv", false]], "tensorrt_llm::executor::decodingmode::isuseminlength (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode14isUseMinLengthEv", false]], "tensorrt_llm::executor::decodingmode::isuseminp (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode9isUseMinPEv", false]], "tensorrt_llm::executor::decodingmode::isusenorepeatngramsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode22isUseNoRepeatNgramSizeEv", false]], "tensorrt_llm::executor::decodingmode::isuseoccurrencepenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode22isUseOccurrencePenaltyEv", false]], "tensorrt_llm::executor::decodingmode::isusepenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode12isUsePenaltyEv", false]], "tensorrt_llm::executor::decodingmode::isusepresencepenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode20isUsePresencePenaltyEv", false]], "tensorrt_llm::executor::decodingmode::isuserepetitionpenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode22isUseRepetitionPenaltyEv", false]], "tensorrt_llm::executor::decodingmode::isusestopcriteria (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode17isUseStopCriteriaEv", false]], "tensorrt_llm::executor::decodingmode::isusestopwords (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode14isUseStopWordsEv", false]], "tensorrt_llm::executor::decodingmode::isusetemperature (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode16isUseTemperatureEv", false]], "tensorrt_llm::executor::decodingmode::isusevariablebeamwidthsearch (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingMode28isUseVariableBeamWidthSearchEv", false]], "tensorrt_llm::executor::decodingmode::kauto (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode5kAutoE", false]], "tensorrt_llm::executor::decodingmode::kbeamsearch (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode11kBeamSearchE", false]], "tensorrt_llm::executor::decodingmode::keagle (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode6kEagleE", false]], "tensorrt_llm::executor::decodingmode::kexplicitdrafttokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode20kExplicitDraftTokensE", false]], "tensorrt_llm::executor::decodingmode::kexternaldrafttokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode20kExternalDraftTokensE", false]], "tensorrt_llm::executor::decodingmode::klookahead (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode10kLookaheadE", false]], "tensorrt_llm::executor::decodingmode::kmedusa (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode7kMedusaE", false]], "tensorrt_llm::executor::decodingmode::knumflags (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode9kNumFlagsE", false]], "tensorrt_llm::executor::decodingmode::ktopk (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode5kTopKE", false]], "tensorrt_llm::executor::decodingmode::ktopktopp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode9kTopKTopPE", false]], "tensorrt_llm::executor::decodingmode::ktopp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode5kTopPE", false]], "tensorrt_llm::executor::decodingmode::kusebantokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode13kUseBanTokensE", false]], "tensorrt_llm::executor::decodingmode::kusebanwords (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode12kUseBanWordsE", false]], "tensorrt_llm::executor::decodingmode::kuseexpliciteosstop (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode19kUseExplicitEosStopE", false]], "tensorrt_llm::executor::decodingmode::kusefrequencypenalties (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode22kUseFrequencyPenaltiesE", false]], "tensorrt_llm::executor::decodingmode::kusemaxlengthstop (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode17kUseMaxLengthStopE", false]], "tensorrt_llm::executor::decodingmode::kuseminlength (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode13kUseMinLengthE", false]], "tensorrt_llm::executor::decodingmode::kuseminp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode8kUseMinPE", false]], "tensorrt_llm::executor::decodingmode::kusenorepeatngramsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode21kUseNoRepeatNgramSizeE", false]], "tensorrt_llm::executor::decodingmode::kuseoccurrencepenalties (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode23kUseOccurrencePenaltiesE", false]], "tensorrt_llm::executor::decodingmode::kusepenalties (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode13kUsePenaltiesE", false]], "tensorrt_llm::executor::decodingmode::kusepresencepenalties (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode21kUsePresencePenaltiesE", false]], "tensorrt_llm::executor::decodingmode::kuserepetitionpenalties (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode23kUseRepetitionPenaltiesE", false]], "tensorrt_llm::executor::decodingmode::kusestandardstopcriteria (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode24kUseStandardStopCriteriaE", false]], "tensorrt_llm::executor::decodingmode::kusestopwords (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode13kUseStopWordsE", false]], "tensorrt_llm::executor::decodingmode::kusetemperature (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode15kUseTemperatureE", false]], "tensorrt_llm::executor::decodingmode::kusevariablebeamwidthsearch (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode27kUseVariableBeamWidthSearchE", false]], "tensorrt_llm::executor::decodingmode::lookahead (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode9LookaheadEv", false]], "tensorrt_llm::executor::decodingmode::medusa (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode6MedusaEv", false]], "tensorrt_llm::executor::decodingmode::mstate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode6mStateE", false]], "tensorrt_llm::executor::decodingmode::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor12DecodingModeeqERK12DecodingMode", false]], "tensorrt_llm::executor::decodingmode::setbitto (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode8setBitToE14UnderlyingTypeb", false]], "tensorrt_llm::executor::decodingmode::topk (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode4TopKEv", false]], "tensorrt_llm::executor::decodingmode::topktopp (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode8TopKTopPEv", false]], "tensorrt_llm::executor::decodingmode::topp (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode4TopPEv", false]], "tensorrt_llm::executor::decodingmode::underlyingtype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode14UnderlyingTypeE", false]], "tensorrt_llm::executor::decodingmode::usebantokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode12useBanTokensEb", false]], "tensorrt_llm::executor::decodingmode::usebanwords (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode11useBanWordsEb", false]], "tensorrt_llm::executor::decodingmode::useexpliciteosstop (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode18useExplicitEosStopEb", false]], "tensorrt_llm::executor::decodingmode::usefrequencypenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode19useFrequencyPenaltyEb", false]], "tensorrt_llm::executor::decodingmode::usemaxlengthstop (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode16useMaxLengthStopEb", false]], "tensorrt_llm::executor::decodingmode::useminlength (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode12useMinLengthEb", false]], "tensorrt_llm::executor::decodingmode::useminp (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode7useMinPEb", false]], "tensorrt_llm::executor::decodingmode::usenorepeatngramsize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode20useNoRepeatNgramSizeEb", false]], "tensorrt_llm::executor::decodingmode::useoccurrencepenalties (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode22useOccurrencePenaltiesEb", false]], "tensorrt_llm::executor::decodingmode::usepresencepenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode18usePresencePenaltyEb", false]], "tensorrt_llm::executor::decodingmode::userepetitionpenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode20useRepetitionPenaltyEb", false]], "tensorrt_llm::executor::decodingmode::usestopwords (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode12useStopWordsEb", false]], "tensorrt_llm::executor::decodingmode::usetemperature (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode14useTemperatureEb", false]], "tensorrt_llm::executor::decodingmode::usevariablebeamwidthsearch (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12DecodingMode26useVariableBeamWidthSearchEb", false]], "tensorrt_llm::executor::detail (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor6detailE", false]], "tensorrt_llm::executor::detail::dimtype64 (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor6detail9DimType64E", false]], "tensorrt_llm::executor::detail::ofitensor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6detail9ofITensorENSt10shared_ptrIN7runtime7ITensorEEE", false]], "tensorrt_llm::executor::detail::toitensor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6detail9toITensorERK6Tensor", false]], "tensorrt_llm::executor::disagg_executor (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executorE", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestratorE", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::awaitcontextresponses (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator21awaitContextResponsesERKNSt8optionalINSt6chrono12millisecondsEEENSt8optionalIiEE", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::awaitgenerationresponses (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator24awaitGenerationResponsesERKNSt8optionalINSt6chrono12millisecondsEEENSt8optionalIiEE", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::canenqueue (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator10canEnqueueEv", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::disaggexecutororchestrator (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator26DisaggExecutorOrchestratorERKNSt6vectorINSt10filesystem4pathEEERKNSt6vectorINSt10filesystem4pathEEERKNSt6vectorIN8executor14ExecutorConfigEEERKNSt6vectorIN8executor14ExecutorConfigEEEbb", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::enqueuecontext (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator14enqueueContextERKNSt6vectorIN5texec7RequestEEENSt8optionalIiEEb", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::enqueuegeneration (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator17enqueueGenerationERKNSt6vectorIN5texec7RequestEEERKNSt6vectorI6IdTypeEENSt8optionalIiEEb", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::getcontextexecutors (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator19getContextExecutorsEv", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::getgenexecutors (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator15getGenExecutorsEv", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::mimpl (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestrator5mImplE", false]], "tensorrt_llm::executor::disagg_executor::disaggexecutororchestrator::~disaggexecutororchestrator (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor26DisaggExecutorOrchestratorD0Ev", false]], "tensorrt_llm::executor::disagg_executor::responsewithid (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithIdE", false]], "tensorrt_llm::executor::disagg_executor::responsewithid::gid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithId3gidE", false]], "tensorrt_llm::executor::disagg_executor::responsewithid::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithIdaSERK14ResponseWithId", false], [0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithIdaSERR14ResponseWithId", false]], "tensorrt_llm::executor::disagg_executor::responsewithid::response (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithId8responseE", false]], "tensorrt_llm::executor::disagg_executor::responsewithid::responsewithid (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithId14ResponseWithIdERK14ResponseWithId", false], [0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithId14ResponseWithIdERKN12tensorrt_llm8executor8ResponseE6IdType", false], [0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithId14ResponseWithIdERR14ResponseWithId", false], [0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithId14ResponseWithIdERRN12tensorrt_llm8executor8ResponseE6IdType", false]], "tensorrt_llm::executor::disagg_executor::responsewithid::~responsewithid (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15disagg_executor14ResponseWithIdD0Ev", false]], "tensorrt_llm::executor::disservingrequeststats (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor22DisServingRequestStatsE", false]], "tensorrt_llm::executor::disservingrequeststats::kvcachesize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22DisServingRequestStats11kvCacheSizeE", false]], "tensorrt_llm::executor::disservingrequeststats::kvcachetransferms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22DisServingRequestStats17kvCacheTransferMSE", false]], "tensorrt_llm::executor::dynamicbatchconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfigE", false]], "tensorrt_llm::executor::dynamicbatchconfig::dynamicbatchconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfig18DynamicBatchConfigEbb10SizeType32NSt6vectorINSt4pairI10SizeType3210SizeType32EEEE", false]], "tensorrt_llm::executor::dynamicbatchconfig::getbatchsizetable (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18DynamicBatchConfig17getBatchSizeTableEv", false]], "tensorrt_llm::executor::dynamicbatchconfig::getdynamicbatchmovingaveragewindow (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18DynamicBatchConfig34getDynamicBatchMovingAverageWindowEv", false]], "tensorrt_llm::executor::dynamicbatchconfig::getenablebatchsizetuning (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18DynamicBatchConfig24getEnableBatchSizeTuningEv", false]], "tensorrt_llm::executor::dynamicbatchconfig::getenablemaxnumtokenstuning (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18DynamicBatchConfig27getEnableMaxNumTokensTuningEv", false]], "tensorrt_llm::executor::dynamicbatchconfig::kdefaultbatchsizetable (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfig22kDefaultBatchSizeTableE", false]], "tensorrt_llm::executor::dynamicbatchconfig::kdefaultdynamicbatchmovingaveragewindow (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfig39kDefaultDynamicBatchMovingAverageWindowE", false]], "tensorrt_llm::executor::dynamicbatchconfig::mbatchsizetable (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfig15mBatchSizeTableE", false]], "tensorrt_llm::executor::dynamicbatchconfig::mdynamicbatchmovingaveragewindow (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfig32mDynamicBatchMovingAverageWindowE", false]], "tensorrt_llm::executor::dynamicbatchconfig::menablebatchsizetuning (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfig22mEnableBatchSizeTuningE", false]], "tensorrt_llm::executor::dynamicbatchconfig::menablemaxnumtokenstuning (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18DynamicBatchConfig25mEnableMaxNumTokensTuningE", false]], "tensorrt_llm::executor::eaglechoices (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor12EagleChoicesE", false]], "tensorrt_llm::executor::eagleconfig (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfigE", false]], "tensorrt_llm::executor::eagleconfig::checkposteriorvalue (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfig19checkPosteriorValueERKNSt8optionalIfEE", false]], "tensorrt_llm::executor::eagleconfig::eagleconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfig11EagleConfigENSt8optionalI12EagleChoicesEEbNSt8optionalIfEEbNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::eagleconfig::getdynamictreemaxtopk (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11EagleConfig21getDynamicTreeMaxTopKEv", false]], "tensorrt_llm::executor::eagleconfig::geteaglechoices (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11EagleConfig15getEagleChoicesEv", false]], "tensorrt_llm::executor::eagleconfig::getposteriorthreshold (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11EagleConfig21getPosteriorThresholdEv", false]], "tensorrt_llm::executor::eagleconfig::isgreedysampling (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11EagleConfig16isGreedySamplingEv", false]], "tensorrt_llm::executor::eagleconfig::mdynamictreemaxtopk (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfig19mDynamicTreeMaxTopKE", false]], "tensorrt_llm::executor::eagleconfig::meaglechoices (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfig13mEagleChoicesE", false]], "tensorrt_llm::executor::eagleconfig::mgreedysampling (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfig15mGreedySamplingE", false]], "tensorrt_llm::executor::eagleconfig::mposteriorthreshold (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfig19mPosteriorThresholdE", false]], "tensorrt_llm::executor::eagleconfig::musedynamictree (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11EagleConfig15mUseDynamicTreeE", false]], "tensorrt_llm::executor::eagleconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11EagleConfigeqERK11EagleConfig", false]], "tensorrt_llm::executor::eagleconfig::usedynamictree (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11EagleConfig14useDynamicTreeEv", false]], "tensorrt_llm::executor::executor (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8ExecutorE", false]], "tensorrt_llm::executor::executor::awaitresponses (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor14awaitResponsesERK6IdTypeRKNSt8optionalINSt6chrono12millisecondsEEE", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor14awaitResponsesERKNSt6vectorI6IdTypeEERKNSt8optionalINSt6chrono12millisecondsEEE", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor14awaitResponsesERKNSt8optionalINSt6chrono12millisecondsEEE", false]], "tensorrt_llm::executor::executor::cancelrequest (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor13cancelRequestE6IdType", false]], "tensorrt_llm::executor::executor::canenqueuerequests (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Executor18canEnqueueRequestsEv", false]], "tensorrt_llm::executor::executor::enqueuerequest (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor14enqueueRequestERK7Request", false]], "tensorrt_llm::executor::executor::enqueuerequests (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor15enqueueRequestsERKNSt6vectorI7RequestEE", false]], "tensorrt_llm::executor::executor::executor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorENSt10shared_ptrI5ModelEENSt10shared_ptrI5ModelEERK14ExecutorConfig", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorENSt10shared_ptrI5ModelEERK14ExecutorConfig", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorERK10BufferViewRKNSt6stringE9ModelTypeRK14ExecutorConfigRKNSt8optionalINSt3mapINSt6stringE6TensorEEEE", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorERK10BufferViewRKNSt6stringERK10BufferViewRKNSt6stringE9ModelTypeRK14ExecutorConfig", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorERK8Executor", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorERKNSt10filesystem4pathE9ModelTypeRK14ExecutorConfig", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorERKNSt10filesystem4pathERKNSt10filesystem4pathE9ModelTypeRK14ExecutorConfig", false], [0, "_CPPv4N12tensorrt_llm8executor8Executor8ExecutorERR8Executor", false]], "tensorrt_llm::executor::executor::getkvcacheeventmanager (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Executor22getKVCacheEventManagerEv", false]], "tensorrt_llm::executor::executor::getlatestdebugtensors (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor21getLatestDebugTensorsEv", false]], "tensorrt_llm::executor::executor::getlatestiterationstats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor23getLatestIterationStatsEv", false]], "tensorrt_llm::executor::executor::getlatestrequeststats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor21getLatestRequestStatsEv", false]], "tensorrt_llm::executor::executor::getnumresponsesready (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Executor20getNumResponsesReadyERKNSt8optionalI6IdTypeEE", false]], "tensorrt_llm::executor::executor::isparticipant (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Executor13isParticipantEv", false]], "tensorrt_llm::executor::executor::mimpl (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor5mImplE", false]], "tensorrt_llm::executor::executor::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8ExecutoraSERK8Executor", false], [0, "_CPPv4N12tensorrt_llm8executor8ExecutoraSERR8Executor", false]], "tensorrt_llm::executor::executor::shutdown (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Executor8shutdownEv", false]], "tensorrt_llm::executor::executor::~executor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8ExecutorD0Ev", false]], "tensorrt_llm::executor::executorconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfigE", false]], "tensorrt_llm::executor::executorconfig::executorconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig14ExecutorConfigE10SizeType3215SchedulerConfig13KvCacheConfigbb10SizeType3210SizeType3212BatchingTypeNSt8optionalI10SizeType32EENSt8optionalI10SizeType32EENSt8optionalI14ParallelConfigEERKNSt8optionalI15PeftCacheConfigEENSt8optionalI25LogitsPostProcessorConfigEENSt8optionalI14DecodingConfigEEbfNSt8optionalI10SizeType32EERK29ExtendedRuntimePerfKnobConfigNSt8optionalI11DebugConfigEE10SizeType328uint64_tNSt8optionalI25SpeculativeDecodingConfigEENSt8optionalI20GuidedDecodingConfigEENSt8optionalINSt6vectorI21AdditionalModelOutputEEEENSt8optionalI22CacheTransceiverConfigEEbbbb", false]], "tensorrt_llm::executor::executorconfig::getadditionalmodeloutputs (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig25getAdditionalModelOutputsEv", false]], "tensorrt_llm::executor::executorconfig::getbatchingtype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig15getBatchingTypeEv", false]], "tensorrt_llm::executor::executorconfig::getcachetransceiverconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig25getCacheTransceiverConfigEv", false]], "tensorrt_llm::executor::executorconfig::getdebugconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig14getDebugConfigEv", false]], "tensorrt_llm::executor::executorconfig::getdecodingconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig17getDecodingConfigEv", false]], "tensorrt_llm::executor::executorconfig::getenablechunkedcontext (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig23getEnableChunkedContextEv", false]], "tensorrt_llm::executor::executorconfig::getenabletrtoverlap (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig19getEnableTrtOverlapEv", false]], "tensorrt_llm::executor::executorconfig::getextendedruntimeperfknobconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig32getExtendedRuntimePerfKnobConfigEv", false]], "tensorrt_llm::executor::executorconfig::getfailfastonattentionwindowtoolarge (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig36getFailFastOnAttentionWindowTooLargeEv", false]], "tensorrt_llm::executor::executorconfig::getgathergenerationlogits (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig25getGatherGenerationLogitsEv", false]], "tensorrt_llm::executor::executorconfig::getgpuweightspercent (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig20getGpuWeightsPercentEv", false]], "tensorrt_llm::executor::executorconfig::getguideddecodingconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig23getGuidedDecodingConfigEv", false]], "tensorrt_llm::executor::executorconfig::getiterstatsmaxiterations (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig25getIterStatsMaxIterationsEv", false]], "tensorrt_llm::executor::executorconfig::getkvcacheconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig16getKvCacheConfigEv", false]], "tensorrt_llm::executor::executorconfig::getkvcacheconfigref (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig19getKvCacheConfigRefEv", false]], "tensorrt_llm::executor::executorconfig::getlogitspostprocessorconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig28getLogitsPostProcessorConfigEv", false]], "tensorrt_llm::executor::executorconfig::getmaxbatchsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig15getMaxBatchSizeEv", false]], "tensorrt_llm::executor::executorconfig::getmaxbeamwidth (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig15getMaxBeamWidthEv", false]], "tensorrt_llm::executor::executorconfig::getmaxnumtokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig15getMaxNumTokensEv", false]], "tensorrt_llm::executor::executorconfig::getmaxqueuesize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig15getMaxQueueSizeEv", false]], "tensorrt_llm::executor::executorconfig::getmaxseqidlemicroseconds (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig25getMaxSeqIdleMicrosecondsEv", false]], "tensorrt_llm::executor::executorconfig::getnormalizelogprobs (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig20getNormalizeLogProbsEv", false]], "tensorrt_llm::executor::executorconfig::getparallelconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig17getParallelConfigEv", false]], "tensorrt_llm::executor::executorconfig::getpeftcacheconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig18getPeftCacheConfigEv", false]], "tensorrt_llm::executor::executorconfig::getprompttableoffloading (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig24getPromptTableOffloadingEv", false]], "tensorrt_llm::executor::executorconfig::getrecvpollperiodms (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig19getRecvPollPeriodMsEv", false]], "tensorrt_llm::executor::executorconfig::getrequeststatsmaxiterations (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig28getRequestStatsMaxIterationsEv", false]], "tensorrt_llm::executor::executorconfig::getschedulerconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig18getSchedulerConfigEv", false]], "tensorrt_llm::executor::executorconfig::getschedulerconfigref (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig21getSchedulerConfigRefEv", false]], "tensorrt_llm::executor::executorconfig::getspecdecconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig16getSpecDecConfigEv", false]], "tensorrt_llm::executor::executorconfig::getusegpudirectstorage (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ExecutorConfig22getUseGpuDirectStorageEv", false]], "tensorrt_llm::executor::executorconfig::kdefaultiterstatsmaxiterations (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig30kDefaultIterStatsMaxIterationsE", false]], "tensorrt_llm::executor::executorconfig::kdefaultmaxseqidlemicroseconds (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig30kDefaultMaxSeqIdleMicrosecondsE", false]], "tensorrt_llm::executor::executorconfig::kdefaultrequeststatsmaxiterations (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig33kDefaultRequestStatsMaxIterationsE", false]], "tensorrt_llm::executor::executorconfig::madditionalmodeloutputs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig23mAdditionalModelOutputsE", false]], "tensorrt_llm::executor::executorconfig::mbatchingtype (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig13mBatchingTypeE", false]], "tensorrt_llm::executor::executorconfig::mcachetransceiverconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig23mCacheTransceiverConfigE", false]], "tensorrt_llm::executor::executorconfig::mdebugconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig12mDebugConfigE", false]], "tensorrt_llm::executor::executorconfig::mdecodingconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig15mDecodingConfigE", false]], "tensorrt_llm::executor::executorconfig::menablechunkedcontext (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig21mEnableChunkedContextE", false]], "tensorrt_llm::executor::executorconfig::menabletrtoverlap (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig17mEnableTrtOverlapE", false]], "tensorrt_llm::executor::executorconfig::mextendedruntimeperfknobconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig30mExtendedRuntimePerfKnobConfigE", false]], "tensorrt_llm::executor::executorconfig::mfailfastonattentionwindowtoolarge (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig34mFailFastOnAttentionWindowTooLargeE", false]], "tensorrt_llm::executor::executorconfig::mgathergenerationlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig23mGatherGenerationLogitsE", false]], "tensorrt_llm::executor::executorconfig::mgpuweightspercent (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig18mGpuWeightsPercentE", false]], "tensorrt_llm::executor::executorconfig::mguideddecodingconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig21mGuidedDecodingConfigE", false]], "tensorrt_llm::executor::executorconfig::miterstatsmaxiterations (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig23mIterStatsMaxIterationsE", false]], "tensorrt_llm::executor::executorconfig::mkvcacheconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig14mKvCacheConfigE", false]], "tensorrt_llm::executor::executorconfig::mlogitspostprocessorconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig26mLogitsPostProcessorConfigE", false]], "tensorrt_llm::executor::executorconfig::mmaxbatchsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig13mMaxBatchSizeE", false]], "tensorrt_llm::executor::executorconfig::mmaxbeamwidth (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig13mMaxBeamWidthE", false]], "tensorrt_llm::executor::executorconfig::mmaxnumtokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig13mMaxNumTokensE", false]], "tensorrt_llm::executor::executorconfig::mmaxqueuesize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig13mMaxQueueSizeE", false]], "tensorrt_llm::executor::executorconfig::mmaxseqidlemicroseconds (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig23mMaxSeqIdleMicrosecondsE", false]], "tensorrt_llm::executor::executorconfig::mnormalizelogprobs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig18mNormalizeLogProbsE", false]], "tensorrt_llm::executor::executorconfig::mparallelconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig15mParallelConfigE", false]], "tensorrt_llm::executor::executorconfig::mpeftcacheconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig16mPeftCacheConfigE", false]], "tensorrt_llm::executor::executorconfig::mprompttableoffloading (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig22mPromptTableOffloadingE", false]], "tensorrt_llm::executor::executorconfig::mrecvpollperiodms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig17mRecvPollPeriodMsE", false]], "tensorrt_llm::executor::executorconfig::mrequeststatsmaxiterations (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig26mRequestStatsMaxIterationsE", false]], "tensorrt_llm::executor::executorconfig::mschedulerconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig16mSchedulerConfigE", false]], "tensorrt_llm::executor::executorconfig::mspeculativedecodingconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig26mSpeculativeDecodingConfigE", false]], "tensorrt_llm::executor::executorconfig::musegpudirectstorage (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig20mUseGpuDirectStorageE", false]], "tensorrt_llm::executor::executorconfig::setadditionalmodeloutputs (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig25setAdditionalModelOutputsERKNSt6vectorI21AdditionalModelOutputEE", false]], "tensorrt_llm::executor::executorconfig::setbatchingtype (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig15setBatchingTypeE12BatchingType", false]], "tensorrt_llm::executor::executorconfig::setcachetransceiverconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig25setCacheTransceiverConfigERK22CacheTransceiverConfig", false]], "tensorrt_llm::executor::executorconfig::setdebugconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig14setDebugConfigERK11DebugConfig", false]], "tensorrt_llm::executor::executorconfig::setdecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig17setDecodingConfigERK14DecodingConfig", false]], "tensorrt_llm::executor::executorconfig::setenablechunkedcontext (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig23setEnableChunkedContextEb", false]], "tensorrt_llm::executor::executorconfig::setenabletrtoverlap (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig19setEnableTrtOverlapEb", false]], "tensorrt_llm::executor::executorconfig::setextendedruntimeperfknobconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig32setExtendedRuntimePerfKnobConfigERK29ExtendedRuntimePerfKnobConfig", false]], "tensorrt_llm::executor::executorconfig::setfailfastonattentionwindowtoolarge (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig36setFailFastOnAttentionWindowTooLargeEb", false]], "tensorrt_llm::executor::executorconfig::setgathergenerationlogits (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig25setGatherGenerationLogitsEb", false]], "tensorrt_llm::executor::executorconfig::setgpuweightspercent (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig20setGpuWeightsPercentERKf", false]], "tensorrt_llm::executor::executorconfig::setguideddecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig23setGuidedDecodingConfigERK20GuidedDecodingConfig", false]], "tensorrt_llm::executor::executorconfig::setiterstatsmaxiterations (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig25setIterStatsMaxIterationsE10SizeType32", false]], "tensorrt_llm::executor::executorconfig::setkvcacheconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig16setKvCacheConfigERK13KvCacheConfig", false]], "tensorrt_llm::executor::executorconfig::setlogitspostprocessorconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig28setLogitsPostProcessorConfigERK25LogitsPostProcessorConfig", false]], "tensorrt_llm::executor::executorconfig::setmaxbatchsize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig15setMaxBatchSizeE10SizeType32", false]], "tensorrt_llm::executor::executorconfig::setmaxbeamwidth (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig15setMaxBeamWidthE10SizeType32", false]], "tensorrt_llm::executor::executorconfig::setmaxnumtokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig15setMaxNumTokensE10SizeType32", false]], "tensorrt_llm::executor::executorconfig::setmaxqueuesize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig15setMaxQueueSizeERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::executorconfig::setmaxseqidlemicroseconds (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig25setMaxSeqIdleMicrosecondsE8uint64_t", false]], "tensorrt_llm::executor::executorconfig::setnormalizelogprobs (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig20setNormalizeLogProbsEb", false]], "tensorrt_llm::executor::executorconfig::setparallelconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig17setParallelConfigERK14ParallelConfig", false]], "tensorrt_llm::executor::executorconfig::setpeftcacheconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig18setPeftCacheConfigERK15PeftCacheConfig", false]], "tensorrt_llm::executor::executorconfig::setprompttableoffloading (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig24setPromptTableOffloadingEb", false]], "tensorrt_llm::executor::executorconfig::setrecvpollperiodms (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig19setRecvPollPeriodMsERK10SizeType32", false]], "tensorrt_llm::executor::executorconfig::setrequeststatsmaxiterations (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig28setRequestStatsMaxIterationsE10SizeType32", false]], "tensorrt_llm::executor::executorconfig::setschedulerconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig18setSchedulerConfigERK15SchedulerConfig", false]], "tensorrt_llm::executor::executorconfig::setspecdecconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig16setSpecDecConfigERK25SpeculativeDecodingConfig", false]], "tensorrt_llm::executor::executorconfig::setusegpudirectstorage (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ExecutorConfig22setUseGpuDirectStorageERKb", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfigE", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::extendedruntimeperfknobconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig29ExtendedRuntimePerfKnobConfigEbbb10SizeType32", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::getcudagraphcachesize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig21getCudaGraphCacheSizeEv", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::getcudagraphmode (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig16getCudaGraphModeEv", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::getenablecontextfmhafp32acc (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig27getEnableContextFMHAFP32AccEv", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::getmultiblockmode (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig17getMultiBlockModeEv", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::mcudagraphcachesize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig19mCudaGraphCacheSizeE", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::mcudagraphmode (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig14mCudaGraphModeE", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::menablecontextfmhafp32acc (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig25mEnableContextFMHAFP32AccE", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::mmultiblockmode (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig15mMultiBlockModeE", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfigeqERK29ExtendedRuntimePerfKnobConfig", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::setcudagraphcachesize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig21setCudaGraphCacheSizeE10SizeType32", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::setcudagraphmode (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig16setCudaGraphModeEb", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::setenablecontextfmhafp32acc (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig27setEnableContextFMHAFP32AccEb", false]], "tensorrt_llm::executor::extendedruntimeperfknobconfig::setmultiblockmode (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor29ExtendedRuntimePerfKnobConfig17setMultiBlockModeEb", false]], "tensorrt_llm::executor::externaldrafttokensconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor25ExternalDraftTokensConfigE", false]], "tensorrt_llm::executor::externaldrafttokensconfig::externaldrafttokensconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor25ExternalDraftTokensConfig25ExternalDraftTokensConfigE9VecTokensNSt8optionalI6TensorEERKNSt8optionalI9FloatTypeEERKNSt8optionalIbEE", false]], "tensorrt_llm::executor::externaldrafttokensconfig::getacceptancethreshold (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25ExternalDraftTokensConfig22getAcceptanceThresholdEv", false]], "tensorrt_llm::executor::externaldrafttokensconfig::getfastlogits (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25ExternalDraftTokensConfig13getFastLogitsEv", false]], "tensorrt_llm::executor::externaldrafttokensconfig::getlogits (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25ExternalDraftTokensConfig9getLogitsEv", false]], "tensorrt_llm::executor::externaldrafttokensconfig::gettokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25ExternalDraftTokensConfig9getTokensEv", false]], "tensorrt_llm::executor::externaldrafttokensconfig::macceptancethreshold (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25ExternalDraftTokensConfig20mAcceptanceThresholdE", false]], "tensorrt_llm::executor::externaldrafttokensconfig::mfastlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25ExternalDraftTokensConfig11mFastLogitsE", false]], "tensorrt_llm::executor::externaldrafttokensconfig::mlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25ExternalDraftTokensConfig7mLogitsE", false]], "tensorrt_llm::executor::externaldrafttokensconfig::mtokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25ExternalDraftTokensConfig7mTokensE", false]], "tensorrt_llm::executor::finishreason (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor12FinishReasonE", false]], "tensorrt_llm::executor::finishreason::kcancelled (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12FinishReason10kCANCELLEDE", false]], "tensorrt_llm::executor::finishreason::kend_id (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12FinishReason7kEND_IDE", false]], "tensorrt_llm::executor::finishreason::klength (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12FinishReason7kLENGTHE", false]], "tensorrt_llm::executor::finishreason::knot_finished (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12FinishReason13kNOT_FINISHEDE", false]], "tensorrt_llm::executor::finishreason::kstop_words (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12FinishReason11kSTOP_WORDSE", false]], "tensorrt_llm::executor::finishreason::ktimed_out (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12FinishReason10kTIMED_OUTE", false]], "tensorrt_llm::executor::floattype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor9FloatTypeE", false]], "tensorrt_llm::executor::guideddecodingconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfigE", false]], "tensorrt_llm::executor::guideddecodingconfig::getbackend (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingConfig10getBackendEv", false]], "tensorrt_llm::executor::guideddecodingconfig::getencodedvocab (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingConfig15getEncodedVocabEv", false]], "tensorrt_llm::executor::guideddecodingconfig::getstoptokenids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingConfig15getStopTokenIdsEv", false]], "tensorrt_llm::executor::guideddecodingconfig::gettokenizerstr (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingConfig15getTokenizerStrEv", false]], "tensorrt_llm::executor::guideddecodingconfig::guideddecodingbackend (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig21GuidedDecodingBackendE", false]], "tensorrt_llm::executor::guideddecodingconfig::guideddecodingbackend::kllguidance (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig21GuidedDecodingBackend11kLLGUIDANCEE", false]], "tensorrt_llm::executor::guideddecodingconfig::guideddecodingbackend::kxgrammar (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig21GuidedDecodingBackend9kXGRAMMARE", false]], "tensorrt_llm::executor::guideddecodingconfig::guideddecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig20GuidedDecodingConfigE21GuidedDecodingBackendNSt8optionalINSt6vectorINSt6stringEEEEENSt8optionalINSt6stringEEENSt8optionalINSt6vectorI11TokenIdTypeEEEE", false]], "tensorrt_llm::executor::guideddecodingconfig::mbackend (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig8mBackendE", false]], "tensorrt_llm::executor::guideddecodingconfig::mencodedvocab (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig13mEncodedVocabE", false]], "tensorrt_llm::executor::guideddecodingconfig::mstoptokenids (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig13mStopTokenIdsE", false]], "tensorrt_llm::executor::guideddecodingconfig::mtokenizerstr (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig13mTokenizerStrE", false]], "tensorrt_llm::executor::guideddecodingconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingConfigeqERK20GuidedDecodingConfig", false]], "tensorrt_llm::executor::guideddecodingconfig::setbackend (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig10setBackendERK21GuidedDecodingBackend", false]], "tensorrt_llm::executor::guideddecodingconfig::setencodedvocab (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig15setEncodedVocabERKNSt6vectorINSt6stringEEE", false]], "tensorrt_llm::executor::guideddecodingconfig::setstoptokenids (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig15setStopTokenIdsERKNSt6vectorI11TokenIdTypeEE", false]], "tensorrt_llm::executor::guideddecodingconfig::settokenizerstr (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingConfig15setTokenizerStrERKNSt6stringE", false]], "tensorrt_llm::executor::guideddecodingconfig::validate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingConfig8validateEv", false]], "tensorrt_llm::executor::guideddecodingparams (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParamsE", false]], "tensorrt_llm::executor::guideddecodingparams::getguide (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingParams8getGuideEv", false]], "tensorrt_llm::executor::guideddecodingparams::getguidetype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingParams12getGuideTypeEv", false]], "tensorrt_llm::executor::guideddecodingparams::guideddecodingparams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams20GuidedDecodingParamsE9GuideTypeNSt8optionalINSt6stringEEE", false]], "tensorrt_llm::executor::guideddecodingparams::guidetype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams9GuideTypeE", false]], "tensorrt_llm::executor::guideddecodingparams::guidetype::kebnf_grammar (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams9GuideType13kEBNF_GRAMMARE", false]], "tensorrt_llm::executor::guideddecodingparams::guidetype::kjson (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams9GuideType5kJSONE", false]], "tensorrt_llm::executor::guideddecodingparams::guidetype::kjson_schema (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams9GuideType12kJSON_SCHEMAE", false]], "tensorrt_llm::executor::guideddecodingparams::guidetype::kregex (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams9GuideType6kREGEXE", false]], "tensorrt_llm::executor::guideddecodingparams::guidetype::kstructural_tag (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams9GuideType15kSTRUCTURAL_TAGE", false]], "tensorrt_llm::executor::guideddecodingparams::mguide (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams6mGuideE", false]], "tensorrt_llm::executor::guideddecodingparams::mguidetype (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor20GuidedDecodingParams10mGuideTypeE", false]], "tensorrt_llm::executor::guideddecodingparams::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor20GuidedDecodingParamseqERK20GuidedDecodingParams", false]], "tensorrt_llm::executor::idtype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor6IdTypeE", false]], "tensorrt_llm::executor::inflightbatchingstats (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStatsE", false]], "tensorrt_llm::executor::inflightbatchingstats::avgnumdecodedtokensperiter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStats26avgNumDecodedTokensPerIterE", false]], "tensorrt_llm::executor::inflightbatchingstats::microbatchid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStats12microBatchIdE", false]], "tensorrt_llm::executor::inflightbatchingstats::numcontextrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStats18numContextRequestsE", false]], "tensorrt_llm::executor::inflightbatchingstats::numctxtokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStats12numCtxTokensE", false]], "tensorrt_llm::executor::inflightbatchingstats::numgenrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStats14numGenRequestsE", false]], "tensorrt_llm::executor::inflightbatchingstats::numpausedrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStats17numPausedRequestsE", false]], "tensorrt_llm::executor::inflightbatchingstats::numscheduledrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor21InflightBatchingStats20numScheduledRequestsE", false]], "tensorrt_llm::executor::iterationstats (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStatsE", false]], "tensorrt_llm::executor::iterationstats::cpumemusage (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats11cpuMemUsageE", false]], "tensorrt_llm::executor::iterationstats::crosskvcachestats (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats17crossKvCacheStatsE", false]], "tensorrt_llm::executor::iterationstats::gpumemusage (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats11gpuMemUsageE", false]], "tensorrt_llm::executor::iterationstats::inflightbatchingstats (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats21inflightBatchingStatsE", false]], "tensorrt_llm::executor::iterationstats::iter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats4iterE", false]], "tensorrt_llm::executor::iterationstats::iterlatencyms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats13iterLatencyMSE", false]], "tensorrt_llm::executor::iterationstats::kvcachestats (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats12kvCacheStatsE", false]], "tensorrt_llm::executor::iterationstats::maxbatchsizeruntime (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats19maxBatchSizeRuntimeE", false]], "tensorrt_llm::executor::iterationstats::maxbatchsizestatic (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats18maxBatchSizeStaticE", false]], "tensorrt_llm::executor::iterationstats::maxbatchsizetunerrecommended (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats28maxBatchSizeTunerRecommendedE", false]], "tensorrt_llm::executor::iterationstats::maxnumactiverequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats20maxNumActiveRequestsE", false]], "tensorrt_llm::executor::iterationstats::maxnumtokensruntime (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats19maxNumTokensRuntimeE", false]], "tensorrt_llm::executor::iterationstats::maxnumtokensstatic (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats18maxNumTokensStaticE", false]], "tensorrt_llm::executor::iterationstats::maxnumtokenstunerrecommended (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats28maxNumTokensTunerRecommendedE", false]], "tensorrt_llm::executor::iterationstats::newactiverequestsqueuelatencyms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats31newActiveRequestsQueueLatencyMSE", false]], "tensorrt_llm::executor::iterationstats::numactiverequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats17numActiveRequestsE", false]], "tensorrt_llm::executor::iterationstats::numcompletedrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats20numCompletedRequestsE", false]], "tensorrt_llm::executor::iterationstats::numnewactiverequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats20numNewActiveRequestsE", false]], "tensorrt_llm::executor::iterationstats::numqueuedrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats17numQueuedRequestsE", false]], "tensorrt_llm::executor::iterationstats::pinnedmemusage (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats14pinnedMemUsageE", false]], "tensorrt_llm::executor::iterationstats::specdecodingstats (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats17specDecodingStatsE", false]], "tensorrt_llm::executor::iterationstats::staticbatchingstats (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats19staticBatchingStatsE", false]], "tensorrt_llm::executor::iterationstats::timestamp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14IterationStats9timestampE", false]], "tensorrt_llm::executor::iterationtype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor13IterationTypeE", false]], "tensorrt_llm::executor::jsonserialization (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor17JsonSerializationE", false]], "tensorrt_llm::executor::jsonserialization::tojsonstr (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor17JsonSerialization9toJsonStrERK12RequestStats", false], [0, "_CPPv4N12tensorrt_llm8executor17JsonSerialization9toJsonStrERK14IterationStats", false], [0, "_CPPv4N12tensorrt_llm8executor17JsonSerialization9toJsonStrERK24RequestStatsPerIteration", false]], "tensorrt_llm::executor::kv_cache (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cacheE", false]], "tensorrt_llm::executor::kv_cache::agentdesc (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9AgentDescE", false]], "tensorrt_llm::executor::kv_cache::agentdesc::agentdesc (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9AgentDesc9AgentDescENSt6stringE", false]], "tensorrt_llm::executor::kv_cache::agentdesc::getbackendagentdesc (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9AgentDesc19getBackendAgentDescEv", false]], "tensorrt_llm::executor::kv_cache::agentdesc::mbackendagentdesc (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9AgentDesc17mBackendAgentDescE", false]], "tensorrt_llm::executor::kv_cache::agentstate (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10AgentStateE", false]], "tensorrt_llm::executor::kv_cache::agentstate::agentstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10AgentState10AgentStateENSt6stringENSt6stringE", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache10AgentState10AgentStateEv", false]], "tensorrt_llm::executor::kv_cache::agentstate::magentname (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10AgentState10mAgentNameE", false]], "tensorrt_llm::executor::kv_cache::agentstate::mconnectioninfo (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10AgentState15mConnectionInfoE", false]], "tensorrt_llm::executor::kv_cache::agentstate::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10AgentStateeqERK10AgentState", false]], "tensorrt_llm::executor::kv_cache::agentstate::tostring (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10AgentState8toStringEv", false]], "tensorrt_llm::executor::kv_cache::baseagentconfig (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15BaseAgentConfigE", false]], "tensorrt_llm::executor::kv_cache::baseagentconfig::mname (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15BaseAgentConfig5mNameE", false]], "tensorrt_llm::executor::kv_cache::baseagentconfig::multithread (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15BaseAgentConfig11multiThreadE", false]], "tensorrt_llm::executor::kv_cache::baseagentconfig::useprogthread (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15BaseAgentConfig13useProgThreadE", false]], "tensorrt_llm::executor::kv_cache::baseloopbackagent (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseLoopbackAgentE", false]], "tensorrt_llm::executor::kv_cache::baseloopbackagent::executeloopbackrequest (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseLoopbackAgent22executeLoopbackRequestERK11MemoryDescsRK9FileDescsb", false]], "tensorrt_llm::executor::kv_cache::baseloopbackagent::~baseloopbackagent (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseLoopbackAgentD0Ev", false]], "tensorrt_llm::executor::kv_cache::basetransferagent (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgentE", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::checkremotedescs (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent16checkRemoteDescsERKNSt6stringERK11MemoryDescs", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::deregistermemory (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent16deregisterMemoryERK13RegisterDescs", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::getlocalagentdesc (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent17getLocalAgentDescEv", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::getlocalconnectioninfo (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent22getLocalConnectionInfoEv", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::getnotifiedsyncmessages (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent23getNotifiedSyncMessagesEv", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::invalidateremoteagent (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent21invalidateRemoteAgentERKNSt6stringE", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::loadremoteagent (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent15loadRemoteAgentERKNSt6stringERK18ConnectionInfoType", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent15loadRemoteAgentERKNSt6stringERK9AgentDesc", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::notifysyncmessage (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent17notifySyncMessageERKNSt6stringERK11SyncMessage", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::registermemory (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent14registerMemoryERK13RegisterDescs", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::submittransferrequests (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgent22submitTransferRequestsERK15TransferRequest", false]], "tensorrt_llm::executor::kv_cache::basetransferagent::~basetransferagent (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17BaseTransferAgentD0Ev", false]], "tensorrt_llm::executor::kv_cache::cachestate (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheStateE", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentionconfig (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState15AttentionConfigE", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentionconfig::attentionconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState15AttentionConfig15AttentionConfigE13AttentionTypei", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentionconfig::mattentiontype (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState15AttentionConfig14mAttentionTypeE", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentionconfig::mkvfactor (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState15AttentionConfig9mKvFactorE", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentionconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState15AttentionConfigeqERK15AttentionConfig", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentiontype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState13AttentionTypeE", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentiontype::kdefault (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState13AttentionType8kDEFAULTE", false]], "tensorrt_llm::executor::kv_cache::cachestate::attentiontype::kmla (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState13AttentionType4kMLAE", false]], "tensorrt_llm::executor::kv_cache::cachestate::cachestate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState10CacheStateE10SizeType3210SizeType3210SizeType3210SizeType3210SizeType3210SizeType3210SizeType32RKNSt6vectorI10SizeType32EEN8nvinfer18DataTypeE13AttentionTypeibiibb10SizeType3210SizeType32", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState10CacheStateE11ModelConfigRKN7runtime11WorldConfigERKNSt6vectorI10SizeType32EEN8nvinfer18DataTypeE13AttentionTypeibb10SizeType3210SizeType32", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState10CacheStateENSt6vectorI10SizeType32EE10SizeType3210SizeType3210SizeType3210SizeType3210SizeType32RKNSt6vectorI10SizeType32EEN8nvinfer18DataTypeE13AttentionTypeibiibb10SizeType3210SizeType32", false]], "tensorrt_llm::executor::kv_cache::cachestate::getattentionconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState18getAttentionConfigEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::getdatatype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState11getDataTypeEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::getenableblockreuse (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState19getEnableBlockReuseEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::gethasindexerkcache (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState19getHasIndexerKCacheEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::getindexerdimperhead (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState20getIndexerDimPerHeadEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::getindexerkcachequantblocksize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState30getIndexerKCacheQuantBlockSizeEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::getmodelconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState14getModelConfigEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::getparallelconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState17getParallelConfigEv", false]], "tensorrt_llm::executor::kv_cache::cachestate::mattentionconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState16mAttentionConfigE", false]], "tensorrt_llm::executor::kv_cache::cachestate::mdatatype (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState9mDataTypeE", false]], "tensorrt_llm::executor::kv_cache::cachestate::menableblockreuse (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState17mEnableBlockReuseE", false]], "tensorrt_llm::executor::kv_cache::cachestate::mhasindexerkcache (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState17mHasIndexerKCacheE", false]], "tensorrt_llm::executor::kv_cache::cachestate::mindexerdimperhead (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState18mIndexerDimPerHeadE", false]], "tensorrt_llm::executor::kv_cache::cachestate::mindexerkcachequantblocksize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState28mIndexerKCacheQuantBlockSizeE", false]], "tensorrt_llm::executor::kv_cache::cachestate::mmodelconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState12mModelConfigE", false]], "tensorrt_llm::executor::kv_cache::cachestate::modelconfig (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState11ModelConfigE", false]], "tensorrt_llm::executor::kv_cache::cachestate::modelconfig::mnbkvheadsperlayer (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState11ModelConfig18mNbKvHeadsPerLayerE", false]], "tensorrt_llm::executor::kv_cache::cachestate::modelconfig::msizeperhead (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState11ModelConfig12mSizePerHeadE", false]], "tensorrt_llm::executor::kv_cache::cachestate::modelconfig::mtokensperblock (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState11ModelConfig15mTokensPerBlockE", false]], "tensorrt_llm::executor::kv_cache::cachestate::modelconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState11ModelConfigeqERK11ModelConfig", false]], "tensorrt_llm::executor::kv_cache::cachestate::mparallelconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState15mParallelConfigE", false]], "tensorrt_llm::executor::kv_cache::cachestate::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheStateeqERKN8kv_cache10CacheStateE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfigE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::mattentionlayernumperpp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfig23mAttentionLayerNumPerPPE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::mcontextparallelism (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfig19mContextParallelismE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::mdprank (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfig7mDPrankE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::mdpsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfig7mDPsizeE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::menableattentiondp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfig18mEnableAttentionDPE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::mpipelineparallelism (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfig20mPipelineParallelismE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::mtensorparallelism (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfig18mTensorParallelismE", false]], "tensorrt_llm::executor::kv_cache::cachestate::parallelconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState14ParallelConfigeqERK14ParallelConfig", false]], "tensorrt_llm::executor::kv_cache::cachestate::tostring (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10CacheState8toStringEv", false]], "tensorrt_llm::executor::kv_cache::commstate (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommStateE", false]], "tensorrt_llm::executor::kv_cache::commstate::commstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommState9CommStateENSt6vectorI10AgentStateEEi", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommState9CommStateENSt6vectorI10SizeType32EEi", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommState9CommStateENSt6vectorI11SocketStateEEi", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommState9CommStateENSt8uint16_tENSt6stringE", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommState9CommStateEv", false]], "tensorrt_llm::executor::kv_cache::commstate::getagentstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState13getAgentStateEv", false]], "tensorrt_llm::executor::kv_cache::commstate::getmpistate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState11getMpiStateEv", false]], "tensorrt_llm::executor::kv_cache::commstate::getselfidx (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState10getSelfIdxEv", false]], "tensorrt_llm::executor::kv_cache::commstate::getsocketstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState14getSocketStateEv", false]], "tensorrt_llm::executor::kv_cache::commstate::isagentstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState12isAgentStateEv", false]], "tensorrt_llm::executor::kv_cache::commstate::ismpistate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState10isMpiStateEv", false]], "tensorrt_llm::executor::kv_cache::commstate::issocketstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState13isSocketStateEv", false]], "tensorrt_llm::executor::kv_cache::commstate::mselfidx (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommState8mSelfIdxE", false]], "tensorrt_llm::executor::kv_cache::commstate::mstate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9CommState6mStateE", false]], "tensorrt_llm::executor::kv_cache::commstate::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommStateeqERK9CommState", false]], "tensorrt_llm::executor::kv_cache::commstate::tostring (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9CommState8toStringEv", false]], "tensorrt_llm::executor::kv_cache::connection (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10ConnectionE", false]], "tensorrt_llm::executor::kv_cache::connection::isthreadsafe (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10Connection12isThreadSafeEv", false]], "tensorrt_llm::executor::kv_cache::connection::recv (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10Connection4recvERK11DataContextPv6size_t", false]], "tensorrt_llm::executor::kv_cache::connection::send (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10Connection4sendERK11DataContextPKv6size_t", false]], "tensorrt_llm::executor::kv_cache::connection::~connection (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10ConnectionD0Ev", false]], "tensorrt_llm::executor::kv_cache::connectioninfotype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache18ConnectionInfoTypeE", false]], "tensorrt_llm::executor::kv_cache::connectionmanager (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17ConnectionManagerE", false]], "tensorrt_llm::executor::kv_cache::connectionmanager::getcommstate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache17ConnectionManager12getCommStateEv", false]], "tensorrt_llm::executor::kv_cache::connectionmanager::getconnections (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17ConnectionManager14getConnectionsERK9CommState", false]], "tensorrt_llm::executor::kv_cache::connectionmanager::recvconnect (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17ConnectionManager11recvConnectERK11DataContextPv6size_t", false]], "tensorrt_llm::executor::kv_cache::connectionmanager::~connectionmanager (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache17ConnectionManagerD0Ev", false]], "tensorrt_llm::executor::kv_cache::datacontext (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11DataContextE", false]], "tensorrt_llm::executor::kv_cache::datacontext::datacontext (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11DataContext11DataContextEi", false]], "tensorrt_llm::executor::kv_cache::datacontext::gettag (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache11DataContext6getTagEv", false]], "tensorrt_llm::executor::kv_cache::datacontext::mtag (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11DataContext4mTagE", false]], "tensorrt_llm::executor::kv_cache::dynlibloader (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoaderE", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::dlsym (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoader5dlSymEPvPKc", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::dynlibloader (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoader12DynLibLoaderERK12DynLibLoader", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoader12DynLibLoaderEv", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::getfunctionpointer (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor8kv_cache12DynLibLoader18getFunctionPointerE9FunctionTRKNSt6stringERKNSt6stringE", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::gethandle (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoader9getHandleERKNSt6stringE", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::getinstance (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoader11getInstanceEv", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::mdllmutex (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoader9mDllMutexE", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::mhandlers (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoader9mHandlersE", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoaderaSERK12DynLibLoader", false]], "tensorrt_llm::executor::kv_cache::dynlibloader::~dynlibloader (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache12DynLibLoaderD0Ev", false]], "tensorrt_llm::executor::kv_cache::filedesc (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDescE", false]], "tensorrt_llm::executor::kv_cache::filedesc::fd (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDesc2fdE", false]], "tensorrt_llm::executor::kv_cache::filedesc::filedesc (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDesc8FileDescERK8FileDesc", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDesc8FileDescERKNSt6stringEi6mode_t6size_t", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDesc8FileDescERR8FileDesc", false]], "tensorrt_llm::executor::kv_cache::filedesc::getfd (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache8FileDesc5getFdEv", false]], "tensorrt_llm::executor::kv_cache::filedesc::getlen (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache8FileDesc6getLenEv", false]], "tensorrt_llm::executor::kv_cache::filedesc::mlen (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDesc4mLenE", false]], "tensorrt_llm::executor::kv_cache::filedesc::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDescaSERK8FileDesc", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDescaSERR8FileDesc", false]], "tensorrt_llm::executor::kv_cache::filedesc::~filedesc (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8FileDescD0Ev", false]], "tensorrt_llm::executor::kv_cache::filedescs (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9FileDescsE", false]], "tensorrt_llm::executor::kv_cache::filedescs::filedescs (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9FileDescs9FileDescsERRNSt6vectorI8FileDescEE", false]], "tensorrt_llm::executor::kv_cache::filedescs::getdescs (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache9FileDescs8getDescsEv", false]], "tensorrt_llm::executor::kv_cache::filedescs::mdescs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache9FileDescs6mDescsE", false]], "tensorrt_llm::executor::kv_cache::makeloopbackagent (c++ function)": [[0, "_CPPv4IDpEN12tensorrt_llm8executor8kv_cache17makeLoopbackAgentENSt10shared_ptrI17BaseLoopbackAgentEERKNSt6stringEDpRR4Args", false]], "tensorrt_llm::executor::kv_cache::maketransferagent (c++ function)": [[0, "_CPPv4IDpEN12tensorrt_llm8executor8kv_cache17makeTransferAgentENSt10unique_ptrI17BaseTransferAgentEERKNSt6stringEDpRR4Args", false]], "tensorrt_llm::executor::kv_cache::memorydesc (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDescE", false]], "tensorrt_llm::executor::kv_cache::memorydesc::deserialize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc11deserializeERNSt7istreamE", false]], "tensorrt_llm::executor::kv_cache::memorydesc::getaddr (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10MemoryDesc7getAddrEv", false]], "tensorrt_llm::executor::kv_cache::memorydesc::getdeviceid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10MemoryDesc11getDeviceIdEv", false]], "tensorrt_llm::executor::kv_cache::memorydesc::getlen (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache10MemoryDesc6getLenEv", false]], "tensorrt_llm::executor::kv_cache::memorydesc::maddr (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc5mAddrE", false]], "tensorrt_llm::executor::kv_cache::memorydesc::mdeviceid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc9mDeviceIdE", false]], "tensorrt_llm::executor::kv_cache::memorydesc::memorydesc (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc10MemoryDescE9uintptr_t6size_t8uint32_t", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc10MemoryDescEPv6size_t8uint32_t", false], [0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc10MemoryDescERKNSt6vectorIcEE8uint32_t", false]], "tensorrt_llm::executor::kv_cache::memorydesc::mlen (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc4mLenE", false]], "tensorrt_llm::executor::kv_cache::memorydesc::serialize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc9serializeERK10MemoryDescRNSt7ostreamE", false]], "tensorrt_llm::executor::kv_cache::memorydesc::serializedsize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryDesc14serializedSizeERK10MemoryDesc", false]], "tensorrt_llm::executor::kv_cache::memorydescs (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11MemoryDescsE", false]], "tensorrt_llm::executor::kv_cache::memorydescs::getdescs (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache11MemoryDescs8getDescsEv", false]], "tensorrt_llm::executor::kv_cache::memorydescs::gettype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache11MemoryDescs7getTypeEv", false]], "tensorrt_llm::executor::kv_cache::memorydescs::mdescs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11MemoryDescs6mDescsE", false]], "tensorrt_llm::executor::kv_cache::memorydescs::memorydescs (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11MemoryDescs11MemoryDescsE10MemoryTypeNSt6vectorI10MemoryDescEE", false]], "tensorrt_llm::executor::kv_cache::memorydescs::mtype (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11MemoryDescs5mTypeE", false]], "tensorrt_llm::executor::kv_cache::memorytype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryTypeE", false]], "tensorrt_llm::executor::kv_cache::memorytype::kblk (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryType4kBLKE", false]], "tensorrt_llm::executor::kv_cache::memorytype::kdram (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryType5kDRAME", false]], "tensorrt_llm::executor::kv_cache::memorytype::kfile (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryType5kFILEE", false]], "tensorrt_llm::executor::kv_cache::memorytype::kobj (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryType4kOBJE", false]], "tensorrt_llm::executor::kv_cache::memorytype::kvram (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10MemoryType5kVRAME", false]], "tensorrt_llm::executor::kv_cache::mpistate (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8MpiStateE", false]], "tensorrt_llm::executor::kv_cache::mpistate::mranks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache8MpiState6mRanksE", false]], "tensorrt_llm::executor::kv_cache::mpistate::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache8MpiStateeqERK8MpiState", false]], "tensorrt_llm::executor::kv_cache::mpistate::tostring (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache8MpiState8toStringEv", false]], "tensorrt_llm::executor::kv_cache::registerdescs (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache13RegisterDescsE", false]], "tensorrt_llm::executor::kv_cache::socketstate (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11SocketStateE", false]], "tensorrt_llm::executor::kv_cache::socketstate::mip (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11SocketState3mIpE", false]], "tensorrt_llm::executor::kv_cache::socketstate::mport (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11SocketState5mPortE", false]], "tensorrt_llm::executor::kv_cache::socketstate::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache11SocketStateeqERK11SocketState", false]], "tensorrt_llm::executor::kv_cache::socketstate::tostring (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache11SocketState8toStringEv", false]], "tensorrt_llm::executor::kv_cache::syncmessage (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache11SyncMessageE", false]], "tensorrt_llm::executor::kv_cache::transferdescs (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache13TransferDescsE", false]], "tensorrt_llm::executor::kv_cache::transferop (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10TransferOpE", false]], "tensorrt_llm::executor::kv_cache::transferop::kread (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10TransferOp5kREADE", false]], "tensorrt_llm::executor::kv_cache::transferop::kwrite (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache10TransferOp6kWRITEE", false]], "tensorrt_llm::executor::kv_cache::transferrequest (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15TransferRequestE", false]], "tensorrt_llm::executor::kv_cache::transferrequest::getdstdescs (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache15TransferRequest11getDstDescsEv", false]], "tensorrt_llm::executor::kv_cache::transferrequest::getop (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache15TransferRequest5getOpEv", false]], "tensorrt_llm::executor::kv_cache::transferrequest::getremotename (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache15TransferRequest13getRemoteNameEv", false]], "tensorrt_llm::executor::kv_cache::transferrequest::getsrcdescs (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache15TransferRequest11getSrcDescsEv", false]], "tensorrt_llm::executor::kv_cache::transferrequest::getsyncmessage (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache15TransferRequest14getSyncMessageEv", false]], "tensorrt_llm::executor::kv_cache::transferrequest::mdstdescs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15TransferRequest9mDstDescsE", false]], "tensorrt_llm::executor::kv_cache::transferrequest::mop (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15TransferRequest3mOpE", false]], "tensorrt_llm::executor::kv_cache::transferrequest::mremotename (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15TransferRequest11mRemoteNameE", false]], "tensorrt_llm::executor::kv_cache::transferrequest::msrcdescs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15TransferRequest9mSrcDescsE", false]], "tensorrt_llm::executor::kv_cache::transferrequest::msyncmessage (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15TransferRequest12mSyncMessageE", false]], "tensorrt_llm::executor::kv_cache::transferrequest::transferrequest (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache15TransferRequest15TransferRequestE10TransferOp13TransferDescs13TransferDescsRKNSt6stringENSt8optionalI11SyncMessageEE", false]], "tensorrt_llm::executor::kv_cache::transferstatus (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache14TransferStatusE", false]], "tensorrt_llm::executor::kv_cache::transferstatus::iscompleted (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache14TransferStatus11isCompletedEv", false]], "tensorrt_llm::executor::kv_cache::transferstatus::wait (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8kv_cache14TransferStatus4waitEv", false]], "tensorrt_llm::executor::kv_cache::transferstatus::~transferstatus (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8kv_cache14TransferStatusD0Ev", false]], "tensorrt_llm::executor::kvcacheconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfigE", false]], "tensorrt_llm::executor::kvcacheconfig::fillemptyfieldsfromruntimedefaults (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig34fillEmptyFieldsFromRuntimeDefaultsERKN12tensorrt_llm7runtime15RuntimeDefaultsE", false]], "tensorrt_llm::executor::kvcacheconfig::getattentiondpeventsgatherperiodms (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig34getAttentionDpEventsGatherPeriodMsEv", false]], "tensorrt_llm::executor::kvcacheconfig::getcopyonpartialreuse (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig21getCopyOnPartialReuseEv", false]], "tensorrt_llm::executor::kvcacheconfig::getcrosskvcachefraction (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig23getCrossKvCacheFractionEv", false]], "tensorrt_llm::executor::kvcacheconfig::getenableblockreuse (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig19getEnableBlockReuseEv", false]], "tensorrt_llm::executor::kvcacheconfig::getenablepartialreuse (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig21getEnablePartialReuseEv", false]], "tensorrt_llm::executor::kvcacheconfig::geteventbuffermaxsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig21getEventBufferMaxSizeEv", false]], "tensorrt_llm::executor::kvcacheconfig::getfreegpumemoryfraction (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig24getFreeGpuMemoryFractionEv", false]], "tensorrt_llm::executor::kvcacheconfig::gethostcachesize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig16getHostCacheSizeEv", false]], "tensorrt_llm::executor::kvcacheconfig::getmaxattentionwindowvec (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig24getMaxAttentionWindowVecEv", false]], "tensorrt_llm::executor::kvcacheconfig::getmaxgputotalbytes (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig19getMaxGpuTotalBytesEv", false]], "tensorrt_llm::executor::kvcacheconfig::getmaxtokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig12getMaxTokensEv", false]], "tensorrt_llm::executor::kvcacheconfig::getonboardblocks (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig16getOnboardBlocksEv", false]], "tensorrt_llm::executor::kvcacheconfig::getsecondaryoffloadminpriority (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig30getSecondaryOffloadMinPriorityEv", false]], "tensorrt_llm::executor::kvcacheconfig::getsinktokenlength (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig18getSinkTokenLengthEv", false]], "tensorrt_llm::executor::kvcacheconfig::getuseuvm (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor13KvCacheConfig9getUseUvmEv", false]], "tensorrt_llm::executor::kvcacheconfig::kdefaultgpumemfraction (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig22kDefaultGpuMemFractionE", false]], "tensorrt_llm::executor::kvcacheconfig::kvcacheconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig13KvCacheConfigEbRKNSt8optionalI10SizeType32EERKNSt8optionalINSt6vectorI10SizeType32EEEERKNSt8optionalI10SizeType32EERKNSt8optionalI9FloatTypeEERKNSt8optionalI6size_tEEbRKNSt8optionalI9FloatTypeEENSt8optionalI17RetentionPriorityEE6size_tbbb10SizeType32RKNSt8optionalIN12tensorrt_llm7runtime15RuntimeDefaultsEEERK8uint64_t", false]], "tensorrt_llm::executor::kvcacheconfig::mattentiondpeventsgatherperiodms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig32mAttentionDpEventsGatherPeriodMsE", false]], "tensorrt_llm::executor::kvcacheconfig::mcopyonpartialreuse (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig19mCopyOnPartialReuseE", false]], "tensorrt_llm::executor::kvcacheconfig::mcrosskvcachefraction (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig21mCrossKvCacheFractionE", false]], "tensorrt_llm::executor::kvcacheconfig::menableblockreuse (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig17mEnableBlockReuseE", false]], "tensorrt_llm::executor::kvcacheconfig::menablepartialreuse (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig19mEnablePartialReuseE", false]], "tensorrt_llm::executor::kvcacheconfig::meventbuffermaxsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig19mEventBufferMaxSizeE", false]], "tensorrt_llm::executor::kvcacheconfig::mfreegpumemoryfraction (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig22mFreeGpuMemoryFractionE", false]], "tensorrt_llm::executor::kvcacheconfig::mhostcachesize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig14mHostCacheSizeE", false]], "tensorrt_llm::executor::kvcacheconfig::mmaxattentionwindowvec (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig22mMaxAttentionWindowVecE", false]], "tensorrt_llm::executor::kvcacheconfig::mmaxgputotalbytes (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig17mMaxGpuTotalBytesE", false]], "tensorrt_llm::executor::kvcacheconfig::mmaxtokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig10mMaxTokensE", false]], "tensorrt_llm::executor::kvcacheconfig::monboardblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig14mOnboardBlocksE", false]], "tensorrt_llm::executor::kvcacheconfig::msecondaryoffloadminpriority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig28mSecondaryOffloadMinPriorityE", false]], "tensorrt_llm::executor::kvcacheconfig::msinktokenlength (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig16mSinkTokenLengthE", false]], "tensorrt_llm::executor::kvcacheconfig::museuvm (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig7mUseUvmE", false]], "tensorrt_llm::executor::kvcacheconfig::setattentiondpeventsgatherperiodms (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig34setAttentionDpEventsGatherPeriodMsE10SizeType32", false]], "tensorrt_llm::executor::kvcacheconfig::setcopyonpartialreuse (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig21setCopyOnPartialReuseEb", false]], "tensorrt_llm::executor::kvcacheconfig::setcrosskvcachefraction (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig23setCrossKvCacheFractionE9FloatType", false]], "tensorrt_llm::executor::kvcacheconfig::setenableblockreuse (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig19setEnableBlockReuseEb", false]], "tensorrt_llm::executor::kvcacheconfig::setenablepartialreuse (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig21setEnablePartialReuseEb", false]], "tensorrt_llm::executor::kvcacheconfig::seteventbuffermaxsize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig21setEventBufferMaxSizeE6size_t", false]], "tensorrt_llm::executor::kvcacheconfig::setfreegpumemoryfraction (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig24setFreeGpuMemoryFractionE9FloatType", false]], "tensorrt_llm::executor::kvcacheconfig::sethostcachesize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig16setHostCacheSizeE6size_t", false]], "tensorrt_llm::executor::kvcacheconfig::setmaxattentionwindowvec (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig24setMaxAttentionWindowVecENSt6vectorI10SizeType32EE", false]], "tensorrt_llm::executor::kvcacheconfig::setmaxgputotalbytes (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig19setMaxGpuTotalBytesE8uint64_t", false]], "tensorrt_llm::executor::kvcacheconfig::setmaxtokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig12setMaxTokensENSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::kvcacheconfig::setonboardblocks (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig16setOnboardBlocksEb", false]], "tensorrt_llm::executor::kvcacheconfig::setsecondaryoffloadminpriority (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig30setSecondaryOffloadMinPriorityENSt8optionalI17RetentionPriorityEE", false]], "tensorrt_llm::executor::kvcacheconfig::setsinktokenlength (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig18setSinkTokenLengthE10SizeType32", false]], "tensorrt_llm::executor::kvcacheconfig::setuseuvm (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13KvCacheConfig9setUseUvmEb", false]], "tensorrt_llm::executor::kvcachecreateddata (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheCreatedDataE", false]], "tensorrt_llm::executor::kvcachecreateddata::numblockspercachelevel (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheCreatedData22numBlocksPerCacheLevelE", false]], "tensorrt_llm::executor::kvcacheevent (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor12KVCacheEventE", false]], "tensorrt_llm::executor::kvcacheevent::attentiondprank (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KVCacheEvent15attentionDpRankE", false]], "tensorrt_llm::executor::kvcacheevent::data (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KVCacheEvent4dataE", false]], "tensorrt_llm::executor::kvcacheevent::eventid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KVCacheEvent7eventIdE", false]], "tensorrt_llm::executor::kvcacheevent::kvcacheevent (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12KVCacheEvent12KVCacheEventE6IdType16KVCacheEventData10SizeType32NSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::kvcacheevent::windowsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KVCacheEvent10windowSizeE", false]], "tensorrt_llm::executor::kvcacheeventdata (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor16KVCacheEventDataE", false]], "tensorrt_llm::executor::kvcacheeventdiff (c++ struct)": [[0, "_CPPv4I0EN12tensorrt_llm8executor16KVCacheEventDiffE", false]], "tensorrt_llm::executor::kvcacheeventdiff::newvalue (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor16KVCacheEventDiff8newValueE", false]], "tensorrt_llm::executor::kvcacheeventdiff::oldvalue (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor16KVCacheEventDiff8oldValueE", false]], "tensorrt_llm::executor::kvcacheeventmanager (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor19KVCacheEventManagerE", false]], "tensorrt_llm::executor::kvcacheeventmanager::getlatestevents (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor19KVCacheEventManager15getLatestEventsENSt8optionalINSt6chrono12millisecondsEEE", false]], "tensorrt_llm::executor::kvcacheeventmanager::kvcacheeventmanager (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor19KVCacheEventManager19KVCacheEventManagerENSt10shared_ptrIN12tensorrt_llm13batch_manager16kv_cache_manager18BaseKVCacheManagerEEE", false]], "tensorrt_llm::executor::kvcacheeventmanager::kvcachemanager (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor19KVCacheEventManager14kvCacheManagerE", false]], "tensorrt_llm::executor::kvcacheremoveddata (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheRemovedDataE", false]], "tensorrt_llm::executor::kvcacheremoveddata::blockhashes (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheRemovedData11blockHashesE", false]], "tensorrt_llm::executor::kvcacheretentionconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfigE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::getdecodedurationms (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfig19getDecodeDurationMsEv", false]], "tensorrt_llm::executor::kvcacheretentionconfig::getdecoderetentionpriority (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfig26getDecodeRetentionPriorityEv", false]], "tensorrt_llm::executor::kvcacheretentionconfig::getdirectory (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfig12getDirectoryEv", false]], "tensorrt_llm::executor::kvcacheretentionconfig::getperblockretentionpriorityduration (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfig36getPerBlockRetentionPriorityDurationE10SizeType3210SizeType32", false]], "tensorrt_llm::executor::kvcacheretentionconfig::gettokenrangeretentionconfigs (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfig29getTokenRangeRetentionConfigsEv", false]], "tensorrt_llm::executor::kvcacheretentionconfig::gettransfermode (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfig15getTransferModeEv", false]], "tensorrt_llm::executor::kvcacheretentionconfig::kdefaultretentionpriority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig25kDefaultRetentionPriorityE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::kmaxretentionpriority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig21kMaxRetentionPriorityE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::kminretentionpriority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig21kMinRetentionPriorityE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::kvcacheretentionconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig22KvCacheRetentionConfigERKNSt6vectorI25TokenRangeRetentionConfigEE17RetentionPriorityNSt8optionalINSt6chrono12millisecondsEEE19KvCacheTransferModeRKNSt6stringE", false], [0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig22KvCacheRetentionConfigEv", false]], "tensorrt_llm::executor::kvcacheretentionconfig::mdecodedurationms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig17mDecodeDurationMsE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::mdecoderetentionpriority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig24mDecodeRetentionPriorityE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::mdirectory (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig10mDirectoryE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::mtokenrangeretentionconfigs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig27mTokenRangeRetentionConfigsE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::mtransfermode (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig13mTransferModeE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfigeqERK22KvCacheRetentionConfig", false]], "tensorrt_llm::executor::kvcacheretentionconfig::tokenrangeretentionconfig (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig25TokenRangeRetentionConfigE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::tokenrangeretentionconfig::durationms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig25TokenRangeRetentionConfig10durationMsE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::tokenrangeretentionconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor22KvCacheRetentionConfig25TokenRangeRetentionConfigeqERK25TokenRangeRetentionConfig", false]], "tensorrt_llm::executor::kvcacheretentionconfig::tokenrangeretentionconfig::priority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig25TokenRangeRetentionConfig8priorityE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::tokenrangeretentionconfig::tokenend (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig25TokenRangeRetentionConfig8tokenEndE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::tokenrangeretentionconfig::tokenrangeretentionconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig25TokenRangeRetentionConfig25TokenRangeRetentionConfigE10SizeType32NSt8optionalI10SizeType32EE17RetentionPriorityNSt8optionalINSt6chrono12millisecondsEEE", false]], "tensorrt_llm::executor::kvcacheretentionconfig::tokenrangeretentionconfig::tokenstart (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KvCacheRetentionConfig25TokenRangeRetentionConfig10tokenStartE", false]], "tensorrt_llm::executor::kvcachestats (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStatsE", false]], "tensorrt_llm::executor::kvcachestats::allocnewblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats14allocNewBlocksE", false]], "tensorrt_llm::executor::kvcachestats::alloctotalblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats16allocTotalBlocksE", false]], "tensorrt_llm::executor::kvcachestats::cachehitrate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats12cacheHitRateE", false]], "tensorrt_llm::executor::kvcachestats::freenumblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats13freeNumBlocksE", false]], "tensorrt_llm::executor::kvcachestats::maxnumblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats12maxNumBlocksE", false]], "tensorrt_llm::executor::kvcachestats::missedblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats12missedBlocksE", false]], "tensorrt_llm::executor::kvcachestats::reusedblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats12reusedBlocksE", false]], "tensorrt_llm::executor::kvcachestats::tokensperblock (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats14tokensPerBlockE", false]], "tensorrt_llm::executor::kvcachestats::usednumblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12KvCacheStats13usedNumBlocksE", false]], "tensorrt_llm::executor::kvcachestoredblockdata (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor22KVCacheStoredBlockDataE", false]], "tensorrt_llm::executor::kvcachestoredblockdata::blockhash (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KVCacheStoredBlockData9blockHashE", false]], "tensorrt_llm::executor::kvcachestoredblockdata::cachelevel (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KVCacheStoredBlockData10cacheLevelE", false]], "tensorrt_llm::executor::kvcachestoredblockdata::kvcachestoredblockdata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor22KVCacheStoredBlockData22KVCacheStoredBlockDataE6IdTypeN12tensorrt_llm7runtime15VecUniqueTokensENSt8optionalIN12tensorrt_llm7runtime14LoraTaskIdTypeEEE10SizeType3210SizeType32", false]], "tensorrt_llm::executor::kvcachestoredblockdata::loraid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KVCacheStoredBlockData6loraIdE", false]], "tensorrt_llm::executor::kvcachestoredblockdata::priority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KVCacheStoredBlockData8priorityE", false]], "tensorrt_llm::executor::kvcachestoredblockdata::tokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor22KVCacheStoredBlockData6tokensE", false]], "tensorrt_llm::executor::kvcachestoreddata (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor17KVCacheStoredDataE", false]], "tensorrt_llm::executor::kvcachestoreddata::blocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17KVCacheStoredData6blocksE", false]], "tensorrt_llm::executor::kvcachestoreddata::parenthash (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17KVCacheStoredData10parentHashE", false]], "tensorrt_llm::executor::kvcachetransfermode (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor19KvCacheTransferModeE", false]], "tensorrt_llm::executor::kvcachetransfermode::dram (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor19KvCacheTransferMode4DRAME", false]], "tensorrt_llm::executor::kvcachetransfermode::gds (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor19KvCacheTransferMode3GDSE", false]], "tensorrt_llm::executor::kvcachetransfermode::posix_debug_fallback (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor19KvCacheTransferMode20POSIX_DEBUG_FALLBACKE", false]], "tensorrt_llm::executor::kvcacheupdateddata (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedDataE", false]], "tensorrt_llm::executor::kvcacheupdateddata::blockhash (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedData9blockHashE", false]], "tensorrt_llm::executor::kvcacheupdateddata::cachelevel (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedData10cacheLevelE", false]], "tensorrt_llm::executor::kvcacheupdateddata::cachelevelupdated (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedData17cacheLevelUpdatedE10SizeType3210SizeType32", false]], "tensorrt_llm::executor::kvcacheupdateddata::kvcacheupdateddata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedData18KVCacheUpdatedDataE6IdType", false], [0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedData18KVCacheUpdatedDataE6IdTypeNSt8optionalI16KVCacheEventDiffI10SizeType32EEENSt8optionalI16KVCacheEventDiffI10SizeType32EEE", false]], "tensorrt_llm::executor::kvcacheupdateddata::priority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedData8priorityE", false]], "tensorrt_llm::executor::kvcacheupdateddata::priorityupdated (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18KVCacheUpdatedData15priorityUpdatedE10SizeType3210SizeType32", false]], "tensorrt_llm::executor::logitspostprocessor (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor19LogitsPostProcessorE", false]], "tensorrt_llm::executor::logitspostprocessorbatched (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor26LogitsPostProcessorBatchedE", false]], "tensorrt_llm::executor::logitspostprocessorconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfigE", false]], "tensorrt_llm::executor::logitspostprocessorconfig::getprocessorbatched (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25LogitsPostProcessorConfig19getProcessorBatchedEv", false]], "tensorrt_llm::executor::logitspostprocessorconfig::getprocessormap (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25LogitsPostProcessorConfig15getProcessorMapEv", false]], "tensorrt_llm::executor::logitspostprocessorconfig::getreplicate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25LogitsPostProcessorConfig12getReplicateEv", false]], "tensorrt_llm::executor::logitspostprocessorconfig::logitspostprocessorconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfig25LogitsPostProcessorConfigENSt8optionalI22LogitsPostProcessorMapEENSt8optionalI26LogitsPostProcessorBatchedEEb", false]], "tensorrt_llm::executor::logitspostprocessorconfig::mprocessorbatched (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfig17mProcessorBatchedE", false]], "tensorrt_llm::executor::logitspostprocessorconfig::mprocessormap (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfig13mProcessorMapE", false]], "tensorrt_llm::executor::logitspostprocessorconfig::mreplicate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfig10mReplicateE", false]], "tensorrt_llm::executor::logitspostprocessorconfig::setprocessorbatched (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfig19setProcessorBatchedERK26LogitsPostProcessorBatched", false]], "tensorrt_llm::executor::logitspostprocessorconfig::setprocessormap (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfig15setProcessorMapERK22LogitsPostProcessorMap", false]], "tensorrt_llm::executor::logitspostprocessorconfig::setreplicate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor25LogitsPostProcessorConfig12setReplicateEb", false]], "tensorrt_llm::executor::logitspostprocessormap (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor22LogitsPostProcessorMapE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfigE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::calculatespeculativeresource (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor23LookaheadDecodingConfig28calculateSpeculativeResourceEv", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::calculatespeculativeresourcetuple (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig33calculateSpeculativeResourceTupleE10SizeType3210SizeType3210SizeType32", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::get (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor23LookaheadDecodingConfig3getEv", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::getngramsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor23LookaheadDecodingConfig12getNgramSizeEv", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::getverificationsetsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor23LookaheadDecodingConfig22getVerificationSetSizeEv", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::getwindowsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor23LookaheadDecodingConfig13getWindowSizeEv", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::isle (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor23LookaheadDecodingConfig4isLEERK23LookaheadDecodingConfig", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::islegal (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig7isLegalE10SizeType3210SizeType3210SizeType32", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::kdefaultlookaheaddecodingngram (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig30kDefaultLookaheadDecodingNgramE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::kdefaultlookaheaddecodingverificationset (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig40kDefaultLookaheadDecodingVerificationSetE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::kdefaultlookaheaddecodingwindow (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig31kDefaultLookaheadDecodingWindowE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::lookaheaddecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig23LookaheadDecodingConfigE10SizeType3210SizeType3210SizeType32", false], [0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig23LookaheadDecodingConfigEv", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::mngramsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig10mNgramSizeE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::mverificationsetsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig20mVerificationSetSizeE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::mwindowsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor23LookaheadDecodingConfig11mWindowSizeE", false]], "tensorrt_llm::executor::lookaheaddecodingconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor23LookaheadDecodingConfigeqERK23LookaheadDecodingConfig", false]], "tensorrt_llm::executor::loraconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor10LoraConfigE", false]], "tensorrt_llm::executor::loraconfig::getconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor10LoraConfig9getConfigEv", false]], "tensorrt_llm::executor::loraconfig::gettaskid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor10LoraConfig9getTaskIdEv", false]], "tensorrt_llm::executor::loraconfig::getweights (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor10LoraConfig10getWeightsEv", false]], "tensorrt_llm::executor::loraconfig::loraconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor10LoraConfig10LoraConfigE6IdTypeNSt8optionalI6TensorEENSt8optionalI6TensorEE", false]], "tensorrt_llm::executor::loraconfig::mconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor10LoraConfig7mConfigE", false]], "tensorrt_llm::executor::loraconfig::mtaskid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor10LoraConfig7mTaskIdE", false]], "tensorrt_llm::executor::loraconfig::mweights (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor10LoraConfig8mWeightsE", false]], "tensorrt_llm::executor::medusachoices (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor13MedusaChoicesE", false]], "tensorrt_llm::executor::memorytype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor10MemoryTypeE", false]], "tensorrt_llm::executor::memorytype::kcpu (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor10MemoryType4kCPUE", false]], "tensorrt_llm::executor::memorytype::kcpu_pinned (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor10MemoryType11kCPU_PINNEDE", false]], "tensorrt_llm::executor::memorytype::kcpu_pinnedpool (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor10MemoryType15kCPU_PINNEDPOOLE", false]], "tensorrt_llm::executor::memorytype::kgpu (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor10MemoryType4kGPUE", false]], "tensorrt_llm::executor::memorytype::kunknown (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor10MemoryType8kUNKNOWNE", false]], "tensorrt_llm::executor::memorytype::kuvm (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor10MemoryType4kUVME", false]], "tensorrt_llm::executor::millisecondstype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor16MillisecondsTypeE", false]], "tensorrt_llm::executor::modeltype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor9ModelTypeE", false]], "tensorrt_llm::executor::modeltype::kdecoder_only (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor9ModelType13kDECODER_ONLYE", false]], "tensorrt_llm::executor::modeltype::kencoder_decoder (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor9ModelType16kENCODER_DECODERE", false]], "tensorrt_llm::executor::modeltype::kencoder_only (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor9ModelType13kENCODER_ONLYE", false]], "tensorrt_llm::executor::mropeconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor11MropeConfigE", false]], "tensorrt_llm::executor::mropeconfig::getmropepositiondeltas (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11MropeConfig22getMRopePositionDeltasEv", false]], "tensorrt_llm::executor::mropeconfig::getmroperotarycossin (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor11MropeConfig20getMRopeRotaryCosSinEv", false]], "tensorrt_llm::executor::mropeconfig::mmropepositiondeltas (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11MropeConfig20mMRopePositionDeltasE", false]], "tensorrt_llm::executor::mropeconfig::mmroperotarycossin (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor11MropeConfig18mMRopeRotaryCosSinE", false]], "tensorrt_llm::executor::mropeconfig::mropeconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor11MropeConfig11MropeConfigE6Tensor10SizeType32", false]], "tensorrt_llm::executor::multimodalinput (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor15MultimodalInputE", false]], "tensorrt_llm::executor::multimodalinput::getmultimodalhashes (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15MultimodalInput19getMultimodalHashesEv", false]], "tensorrt_llm::executor::multimodalinput::getmultimodallengths (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15MultimodalInput20getMultimodalLengthsEv", false]], "tensorrt_llm::executor::multimodalinput::getmultimodalpositions (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15MultimodalInput22getMultimodalPositionsEv", false]], "tensorrt_llm::executor::multimodalinput::mmultimodalhashes (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15MultimodalInput17mMultimodalHashesE", false]], "tensorrt_llm::executor::multimodalinput::mmultimodallengths (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15MultimodalInput18mMultimodalLengthsE", false]], "tensorrt_llm::executor::multimodalinput::mmultimodalpositions (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15MultimodalInput20mMultimodalPositionsE", false]], "tensorrt_llm::executor::multimodalinput::multimodalinput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15MultimodalInput15MultimodalInputENSt6vectorINSt6vectorI10SizeType32EEEENSt6vectorI10SizeType32EENSt6vectorI10SizeType32EE", false]], "tensorrt_llm::executor::operator<< (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executorlsERNSt7ostreamE21ContextChunkingPolicy", false], [0, "_CPPv4N12tensorrt_llm8executorlsERNSt7ostreamE23CapacitySchedulerPolicy", false]], "tensorrt_llm::executor::orchestratorconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfigE", false]], "tensorrt_llm::executor::orchestratorconfig::getisorchestrator (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18OrchestratorConfig17getIsOrchestratorEv", false]], "tensorrt_llm::executor::orchestratorconfig::getorchleadercomm (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18OrchestratorConfig17getOrchLeaderCommEv", false]], "tensorrt_llm::executor::orchestratorconfig::getspawnprocesses (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18OrchestratorConfig17getSpawnProcessesEv", false]], "tensorrt_llm::executor::orchestratorconfig::getworkerexecutablepath (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18OrchestratorConfig23getWorkerExecutablePathEv", false]], "tensorrt_llm::executor::orchestratorconfig::misorchestrator (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig15mIsOrchestratorE", false]], "tensorrt_llm::executor::orchestratorconfig::morchleadercomm (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig15mOrchLeaderCommE", false]], "tensorrt_llm::executor::orchestratorconfig::mspawnprocesses (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig15mSpawnProcessesE", false]], "tensorrt_llm::executor::orchestratorconfig::mworkerexecutablepath (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig21mWorkerExecutablePathE", false]], "tensorrt_llm::executor::orchestratorconfig::orchestratorconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig18OrchestratorConfigEbNSt6stringENSt10shared_ptrIN3mpi7MpiCommEEEb", false]], "tensorrt_llm::executor::orchestratorconfig::setisorchestrator (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig17setIsOrchestratorEb", false]], "tensorrt_llm::executor::orchestratorconfig::setorchleadercomm (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig17setOrchLeaderCommERKNSt10shared_ptrIN3mpi7MpiCommEEE", false]], "tensorrt_llm::executor::orchestratorconfig::setspawnprocesses (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig17setSpawnProcessesEb", false]], "tensorrt_llm::executor::orchestratorconfig::setworkerexecutablepath (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18OrchestratorConfig23setWorkerExecutablePathERKNSt6stringE", false]], "tensorrt_llm::executor::outputconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfigE", false]], "tensorrt_llm::executor::outputconfig::additionalmodeloutputs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig22additionalModelOutputsE", false]], "tensorrt_llm::executor::outputconfig::excludeinputfromoutput (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig22excludeInputFromOutputE", false]], "tensorrt_llm::executor::outputconfig::outputconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig12OutputConfigEbbbbbbNSt8optionalINSt6vectorI21AdditionalModelOutputEEEE", false]], "tensorrt_llm::executor::outputconfig::returncontextlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig19returnContextLogitsE", false]], "tensorrt_llm::executor::outputconfig::returnencoderoutput (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig19returnEncoderOutputE", false]], "tensorrt_llm::executor::outputconfig::returngenerationlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig22returnGenerationLogitsE", false]], "tensorrt_llm::executor::outputconfig::returnlogprobs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig14returnLogProbsE", false]], "tensorrt_llm::executor::outputconfig::returnperfmetrics (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12OutputConfig17returnPerfMetricsE", false]], "tensorrt_llm::executor::parallelconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfigE", false]], "tensorrt_llm::executor::parallelconfig::getcommunicationmode (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ParallelConfig20getCommunicationModeEv", false]], "tensorrt_llm::executor::parallelconfig::getcommunicationtype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ParallelConfig20getCommunicationTypeEv", false]], "tensorrt_llm::executor::parallelconfig::getdeviceids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ParallelConfig12getDeviceIdsEv", false]], "tensorrt_llm::executor::parallelconfig::getnumnodes (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ParallelConfig11getNumNodesEv", false]], "tensorrt_llm::executor::parallelconfig::getorchestratorconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ParallelConfig21getOrchestratorConfigEv", false]], "tensorrt_llm::executor::parallelconfig::getparticipantids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14ParallelConfig17getParticipantIdsEv", false]], "tensorrt_llm::executor::parallelconfig::mcommmode (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig9mCommModeE", false]], "tensorrt_llm::executor::parallelconfig::mcommtype (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig9mCommTypeE", false]], "tensorrt_llm::executor::parallelconfig::mdeviceids (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig10mDeviceIdsE", false]], "tensorrt_llm::executor::parallelconfig::mnumnodes (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig9mNumNodesE", false]], "tensorrt_llm::executor::parallelconfig::morchestratorconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig19mOrchestratorConfigE", false]], "tensorrt_llm::executor::parallelconfig::mparticipantids (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig15mParticipantIdsE", false]], "tensorrt_llm::executor::parallelconfig::parallelconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig14ParallelConfigE17CommunicationType17CommunicationModeNSt8optionalINSt6vectorI10SizeType32EEEENSt8optionalINSt6vectorI10SizeType32EEEERKNSt8optionalI18OrchestratorConfigEENSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::parallelconfig::setcommunicationmode (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig20setCommunicationModeE17CommunicationMode", false]], "tensorrt_llm::executor::parallelconfig::setcommunicationtype (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig20setCommunicationTypeE17CommunicationType", false]], "tensorrt_llm::executor::parallelconfig::setdeviceids (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig12setDeviceIdsERKNSt6vectorI10SizeType32EE", false]], "tensorrt_llm::executor::parallelconfig::setnumnodes (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig11setNumNodesE10SizeType32", false]], "tensorrt_llm::executor::parallelconfig::setorchestratorconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig21setOrchestratorConfigERK18OrchestratorConfig", false]], "tensorrt_llm::executor::parallelconfig::setparticipantids (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14ParallelConfig17setParticipantIdsERKNSt6vectorI10SizeType32EE", false]], "tensorrt_llm::executor::peftcacheconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfigE", false]], "tensorrt_llm::executor::peftcacheconfig::getdevicecachepercent (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig21getDeviceCachePercentEv", false]], "tensorrt_llm::executor::peftcacheconfig::gethostcachesize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig16getHostCacheSizeEv", false]], "tensorrt_llm::executor::peftcacheconfig::getloraprefetchdir (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig18getLoraPrefetchDirEv", false]], "tensorrt_llm::executor::peftcacheconfig::getmaxadaptersize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig17getMaxAdapterSizeEv", false]], "tensorrt_llm::executor::peftcacheconfig::getmaxpagesperblockdevice (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig25getMaxPagesPerBlockDeviceEv", false]], "tensorrt_llm::executor::peftcacheconfig::getmaxpagesperblockhost (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig23getMaxPagesPerBlockHostEv", false]], "tensorrt_llm::executor::peftcacheconfig::getnumcopystreams (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig17getNumCopyStreamsEv", false]], "tensorrt_llm::executor::peftcacheconfig::getnumdevicemodulelayer (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig23getNumDeviceModuleLayerEv", false]], "tensorrt_llm::executor::peftcacheconfig::getnumensureworkers (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig19getNumEnsureWorkersEv", false]], "tensorrt_llm::executor::peftcacheconfig::getnumhostmodulelayer (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig21getNumHostModuleLayerEv", false]], "tensorrt_llm::executor::peftcacheconfig::getnumputworkers (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig16getNumPutWorkersEv", false]], "tensorrt_llm::executor::peftcacheconfig::getoptimaladaptersize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfig21getOptimalAdapterSizeEv", false]], "tensorrt_llm::executor::peftcacheconfig::kdefaultmaxadaptersize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig22kDefaultMaxAdapterSizeE", false]], "tensorrt_llm::executor::peftcacheconfig::kdefaultmaxpagesperblockdevice (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig30kDefaultMaxPagesPerBlockDeviceE", false]], "tensorrt_llm::executor::peftcacheconfig::kdefaultmaxpagesperblockhost (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig28kDefaultMaxPagesPerBlockHostE", false]], "tensorrt_llm::executor::peftcacheconfig::kdefaultoptimaladaptersize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig26kDefaultOptimalAdapterSizeE", false]], "tensorrt_llm::executor::peftcacheconfig::mdevicecachepercent (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig19mDeviceCachePercentE", false]], "tensorrt_llm::executor::peftcacheconfig::mhostcachesize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig14mHostCacheSizeE", false]], "tensorrt_llm::executor::peftcacheconfig::mloraprefetchdir (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig16mLoraPrefetchDirE", false]], "tensorrt_llm::executor::peftcacheconfig::mmaxadaptersize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig15mMaxAdapterSizeE", false]], "tensorrt_llm::executor::peftcacheconfig::mmaxpagesperblockdevice (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig23mMaxPagesPerBlockDeviceE", false]], "tensorrt_llm::executor::peftcacheconfig::mmaxpagesperblockhost (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig21mMaxPagesPerBlockHostE", false]], "tensorrt_llm::executor::peftcacheconfig::mnumcopystreams (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig15mNumCopyStreamsE", false]], "tensorrt_llm::executor::peftcacheconfig::mnumdevicemodulelayer (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig21mNumDeviceModuleLayerE", false]], "tensorrt_llm::executor::peftcacheconfig::mnumensureworkers (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig17mNumEnsureWorkersE", false]], "tensorrt_llm::executor::peftcacheconfig::mnumhostmodulelayer (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig19mNumHostModuleLayerE", false]], "tensorrt_llm::executor::peftcacheconfig::mnumputworkers (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig14mNumPutWorkersE", false]], "tensorrt_llm::executor::peftcacheconfig::moptimaladaptersize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig19mOptimalAdapterSizeE", false]], "tensorrt_llm::executor::peftcacheconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15PeftCacheConfigeqERK15PeftCacheConfig", false]], "tensorrt_llm::executor::peftcacheconfig::peftcacheconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15PeftCacheConfig15PeftCacheConfigE10SizeType3210SizeType3210SizeType3210SizeType3210SizeType3210SizeType3210SizeType3210SizeType3210SizeType32RKNSt8optionalIfEERKNSt8optionalI6size_tEERKNSt8optionalINSt6stringEEE", false]], "tensorrt_llm::executor::prioritytype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor12PriorityTypeE", false]], "tensorrt_llm::executor::prompttuningconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor18PromptTuningConfigE", false]], "tensorrt_llm::executor::prompttuningconfig::getembeddingtable (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18PromptTuningConfig17getEmbeddingTableEv", false]], "tensorrt_llm::executor::prompttuningconfig::getinputtokenextraids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor18PromptTuningConfig21getInputTokenExtraIdsEv", false]], "tensorrt_llm::executor::prompttuningconfig::membeddingtable (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18PromptTuningConfig15mEmbeddingTableE", false]], "tensorrt_llm::executor::prompttuningconfig::minputtokenextraids (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18PromptTuningConfig19mInputTokenExtraIdsE", false]], "tensorrt_llm::executor::prompttuningconfig::prompttuningconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor18PromptTuningConfig18PromptTuningConfigE6TensorNSt8optionalI16VecTokenExtraIdsEE", false]], "tensorrt_llm::executor::randomseedtype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor14RandomSeedTypeE", false]], "tensorrt_llm::executor::request (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor7RequestE", false]], "tensorrt_llm::executor::request::getadditionaloutputnames (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request24getAdditionalOutputNamesEv", false]], "tensorrt_llm::executor::request::getallottedtimems (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request17getAllottedTimeMsEv", false]], "tensorrt_llm::executor::request::getbadwords (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request11getBadWordsEv", false]], "tensorrt_llm::executor::request::getcachesaltid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request14getCacheSaltIDEv", false]], "tensorrt_llm::executor::request::getclientid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request11getClientIdEv", false]], "tensorrt_llm::executor::request::getcontextphaseparams (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request21getContextPhaseParamsEv", false]], "tensorrt_llm::executor::request::getcrossattentionmask (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request21getCrossAttentionMaskEv", false]], "tensorrt_llm::executor::request::geteagleconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request14getEagleConfigEv", false]], "tensorrt_llm::executor::request::getembeddingbias (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request16getEmbeddingBiasEv", false]], "tensorrt_llm::executor::request::getencoderinputfeatures (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request23getEncoderInputFeaturesEv", false]], "tensorrt_llm::executor::request::getencoderinputtokenids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request23getEncoderInputTokenIdsEv", false]], "tensorrt_llm::executor::request::getencoderoutputlength (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request22getEncoderOutputLengthEv", false]], "tensorrt_llm::executor::request::getendid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request8getEndIdEv", false]], "tensorrt_llm::executor::request::getexternaldrafttokensconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request28getExternalDraftTokensConfigEv", false]], "tensorrt_llm::executor::request::getguideddecodingparams (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request23getGuidedDecodingParamsEv", false]], "tensorrt_llm::executor::request::getinputtokenids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request16getInputTokenIdsEv", false]], "tensorrt_llm::executor::request::getkvcacheretentionconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request25getKvCacheRetentionConfigEv", false]], "tensorrt_llm::executor::request::getlanguageadapteruid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request21getLanguageAdapterUidEv", false]], "tensorrt_llm::executor::request::getlogitspostprocessor (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request22getLogitsPostProcessorEv", false]], "tensorrt_llm::executor::request::getlogitspostprocessorname (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request26getLogitsPostProcessorNameEv", false]], "tensorrt_llm::executor::request::getlookaheadconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request18getLookaheadConfigEv", false]], "tensorrt_llm::executor::request::getloraconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request13getLoraConfigEv", false]], "tensorrt_llm::executor::request::getmaxtokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request12getMaxTokensEv", false]], "tensorrt_llm::executor::request::getmropeconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request14getMropeConfigEv", false]], "tensorrt_llm::executor::request::getmultimodalembedding (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request22getMultimodalEmbeddingEv", false]], "tensorrt_llm::executor::request::getmultimodalinput (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request18getMultimodalInputEv", false]], "tensorrt_llm::executor::request::getoutputconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request15getOutputConfigEv", false]], "tensorrt_llm::executor::request::getpadid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request8getPadIdEv", false]], "tensorrt_llm::executor::request::getpositionids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request14getPositionIdsEv", false]], "tensorrt_llm::executor::request::getpriority (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request11getPriorityEv", false]], "tensorrt_llm::executor::request::getprompttuningconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request21getPromptTuningConfigEv", false]], "tensorrt_llm::executor::request::getrequesttype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request14getRequestTypeEv", false]], "tensorrt_llm::executor::request::getreturnallgeneratedtokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request27getReturnAllGeneratedTokensEv", false]], "tensorrt_llm::executor::request::getsamplingconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request17getSamplingConfigEv", false]], "tensorrt_llm::executor::request::getskipcrossattnblocks (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request22getSkipCrossAttnBlocksEv", false]], "tensorrt_llm::executor::request::getstopwords (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request12getStopWordsEv", false]], "tensorrt_llm::executor::request::getstreaming (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor7Request12getStreamingEv", false]], "tensorrt_llm::executor::request::kbatchedpostprocessorname (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor7Request25kBatchedPostProcessorNameE", false]], "tensorrt_llm::executor::request::kdefaultpriority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor7Request16kDefaultPriorityE", false]], "tensorrt_llm::executor::request::kdynamicpostprocessornameprefix (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor7Request31kDynamicPostProcessorNamePrefixE", false]], "tensorrt_llm::executor::request::mimpl (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor7Request5mImplE", false]], "tensorrt_llm::executor::request::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7RequestaSERK7Request", false], [0, "_CPPv4N12tensorrt_llm8executor7RequestaSERR7Request", false]], "tensorrt_llm::executor::request::request (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request7RequestE9VecTokens10SizeType32bRK14SamplingConfigRK12OutputConfigRKNSt8optionalI10SizeType32EERKNSt8optionalI10SizeType32EENSt8optionalINSt6vectorI10SizeType32EEEENSt8optionalINSt4listI9VecTokensEEEENSt8optionalINSt4listI9VecTokensEEEENSt8optionalI6TensorEENSt8optionalI25ExternalDraftTokensConfigEENSt8optionalI18PromptTuningConfigEENSt8optionalI15MultimodalInputEENSt8optionalI6TensorEENSt8optionalI11MropeConfigEENSt8optionalI10LoraConfigEENSt8optionalI23LookaheadDecodingConfigEENSt8optionalI22KvCacheRetentionConfigEENSt8optionalINSt6stringEEENSt8optionalI19LogitsPostProcessorEENSt8optionalI9VecTokensEENSt8optionalI6IdTypeEEb12PriorityType11RequestTypeNSt8optionalI18ContextPhaseParamsEENSt8optionalI6TensorEENSt8optionalI10SizeType32EENSt8optionalI6TensorEE10SizeType32NSt8optionalI11EagleConfigEENSt8optionalI6TensorEENSt8optionalI20GuidedDecodingParamsEENSt8optionalI10SizeType32EENSt8optionalI16MillisecondsTypeEENSt8optionalI15CacheSaltIDTypeEE", false], [0, "_CPPv4N12tensorrt_llm8executor7Request7RequestERK7Request", false], [0, "_CPPv4N12tensorrt_llm8executor7Request7RequestERR7Request", false]], "tensorrt_llm::executor::request::setallottedtimems (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request17setAllottedTimeMsE16MillisecondsType", false]], "tensorrt_llm::executor::request::setbadwords (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request11setBadWordsERKNSt4listI9VecTokensEE", false]], "tensorrt_llm::executor::request::setcachesaltid (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request14setCacheSaltIDE15CacheSaltIDType", false]], "tensorrt_llm::executor::request::setclientid (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request11setClientIdE6IdType", false]], "tensorrt_llm::executor::request::setcontextphaseparams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request21setContextPhaseParamsE18ContextPhaseParams", false]], "tensorrt_llm::executor::request::setcrossattentionmask (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request21setCrossAttentionMaskE6Tensor", false]], "tensorrt_llm::executor::request::seteagleconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request14setEagleConfigERKNSt8optionalI11EagleConfigEE", false]], "tensorrt_llm::executor::request::setembeddingbias (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request16setEmbeddingBiasERK6Tensor", false]], "tensorrt_llm::executor::request::setencoderinputfeatures (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request23setEncoderInputFeaturesE6Tensor", false]], "tensorrt_llm::executor::request::setencoderinputtokenids (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request23setEncoderInputTokenIdsERK9VecTokens", false]], "tensorrt_llm::executor::request::setencoderoutputlength (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request22setEncoderOutputLengthE10SizeType32", false]], "tensorrt_llm::executor::request::setendid (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request8setEndIdE10SizeType32", false]], "tensorrt_llm::executor::request::setexternaldrafttokensconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request28setExternalDraftTokensConfigERK25ExternalDraftTokensConfig", false]], "tensorrt_llm::executor::request::setguideddecodingparams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request23setGuidedDecodingParamsERK20GuidedDecodingParams", false]], "tensorrt_llm::executor::request::setkvcacheretentionconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request25setKvCacheRetentionConfigERK22KvCacheRetentionConfig", false]], "tensorrt_llm::executor::request::setlanguageadapteruid (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request21setLanguageAdapterUidE10SizeType32", false]], "tensorrt_llm::executor::request::setlogitspostprocessor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request22setLogitsPostProcessorERKNSt8optionalI19LogitsPostProcessorEE", false]], "tensorrt_llm::executor::request::setlogitspostprocessorname (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request26setLogitsPostProcessorNameERKNSt6stringE", false]], "tensorrt_llm::executor::request::setlookaheadconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request18setLookaheadConfigERK23LookaheadDecodingConfig", false]], "tensorrt_llm::executor::request::setloraconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request13setLoraConfigERK10LoraConfig", false]], "tensorrt_llm::executor::request::setmropeconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request14setMropeConfigERK11MropeConfig", false]], "tensorrt_llm::executor::request::setmultimodalembedding (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request22setMultimodalEmbeddingERK6Tensor", false]], "tensorrt_llm::executor::request::setmultimodalinput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request18setMultimodalInputERK15MultimodalInput", false]], "tensorrt_llm::executor::request::setoutputconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request15setOutputConfigERK12OutputConfig", false]], "tensorrt_llm::executor::request::setpadid (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request8setPadIdE10SizeType32", false]], "tensorrt_llm::executor::request::setpositionids (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request14setPositionIdsERKNSt6vectorI10SizeType32EE", false]], "tensorrt_llm::executor::request::setpriority (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request11setPriorityE12PriorityType", false]], "tensorrt_llm::executor::request::setprompttuningconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request21setPromptTuningConfigERK18PromptTuningConfig", false]], "tensorrt_llm::executor::request::setrequesttype (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request14setRequestTypeERK11RequestType", false]], "tensorrt_llm::executor::request::setreturnallgeneratedtokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request27setReturnAllGeneratedTokensEb", false]], "tensorrt_llm::executor::request::setsamplingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request17setSamplingConfigERK14SamplingConfig", false]], "tensorrt_llm::executor::request::setskipcrossattnblocks (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request22setSkipCrossAttnBlocksE6Tensor", false]], "tensorrt_llm::executor::request::setstopwords (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request12setStopWordsERKNSt4listI9VecTokensEE", false]], "tensorrt_llm::executor::request::setstreaming (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7Request12setStreamingEb", false]], "tensorrt_llm::executor::request::~request (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor7RequestD0Ev", false]], "tensorrt_llm::executor::requestperfmetrics (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetricsE", false]], "tensorrt_llm::executor::requestperfmetrics::firstiter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics9firstIterE", false]], "tensorrt_llm::executor::requestperfmetrics::iter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics4iterE", false]], "tensorrt_llm::executor::requestperfmetrics::kvcachemetrics (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics14kvCacheMetricsE", false]], "tensorrt_llm::executor::requestperfmetrics::kvcachemetrics (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics14KvCacheMetricsE", false]], "tensorrt_llm::executor::requestperfmetrics::kvcachemetrics::kvcachehitrate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics14KvCacheMetrics14kvCacheHitRateE", false]], "tensorrt_llm::executor::requestperfmetrics::kvcachemetrics::nummissedblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics14KvCacheMetrics15numMissedBlocksE", false]], "tensorrt_llm::executor::requestperfmetrics::kvcachemetrics::numnewallocatedblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics14KvCacheMetrics21numNewAllocatedBlocksE", false]], "tensorrt_llm::executor::requestperfmetrics::kvcachemetrics::numreusedblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics14KvCacheMetrics15numReusedBlocksE", false]], "tensorrt_llm::executor::requestperfmetrics::kvcachemetrics::numtotalallocatedblocks (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics14KvCacheMetrics23numTotalAllocatedBlocksE", false]], "tensorrt_llm::executor::requestperfmetrics::lastiter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics8lastIterE", false]], "tensorrt_llm::executor::requestperfmetrics::speculativedecoding (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics19speculativeDecodingE", false]], "tensorrt_llm::executor::requestperfmetrics::speculativedecodingmetrics (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics26SpeculativeDecodingMetricsE", false]], "tensorrt_llm::executor::requestperfmetrics::speculativedecodingmetrics::acceptancerate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics26SpeculativeDecodingMetrics14acceptanceRateE", false]], "tensorrt_llm::executor::requestperfmetrics::speculativedecodingmetrics::totalaccepteddrafttokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics26SpeculativeDecodingMetrics24totalAcceptedDraftTokensE", false]], "tensorrt_llm::executor::requestperfmetrics::speculativedecodingmetrics::totaldrafttokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics26SpeculativeDecodingMetrics16totalDraftTokensE", false]], "tensorrt_llm::executor::requestperfmetrics::timepoint (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics9TimePointE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13timingMetricsE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetricsE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics::arrivaltime (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetrics11arrivalTimeE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics::firstscheduledtime (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetrics18firstScheduledTimeE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics::firsttokentime (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetrics14firstTokenTimeE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics::kvcachesize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetrics11kvCacheSizeE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics::kvcachetransferend (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetrics18kvCacheTransferEndE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics::kvcachetransferstart (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetrics20kvCacheTransferStartE", false]], "tensorrt_llm::executor::requestperfmetrics::timingmetrics::lasttokentime (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor18RequestPerfMetrics13TimingMetrics13lastTokenTimeE", false]], "tensorrt_llm::executor::requeststage (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStageE", false]], "tensorrt_llm::executor::requeststage::kcontext_in_progress (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStage20kCONTEXT_IN_PROGRESSE", false]], "tensorrt_llm::executor::requeststage::kencoder_in_progress (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStage20kENCODER_IN_PROGRESSE", false]], "tensorrt_llm::executor::requeststage::kgeneration_complete (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStage20kGENERATION_COMPLETEE", false]], "tensorrt_llm::executor::requeststage::kgeneration_in_progress (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStage23kGENERATION_IN_PROGRESSE", false]], "tensorrt_llm::executor::requeststage::kqueued (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStage7kQUEUEDE", false]], "tensorrt_llm::executor::requeststats (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStatsE", false]], "tensorrt_llm::executor::requeststats::allocnewblocksperrequest (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats24allocNewBlocksPerRequestE", false]], "tensorrt_llm::executor::requeststats::alloctotalblocksperrequest (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats26allocTotalBlocksPerRequestE", false]], "tensorrt_llm::executor::requeststats::avgnumdecodedtokensperiter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats26avgNumDecodedTokensPerIterE", false]], "tensorrt_llm::executor::requeststats::contextprefillposition (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats22contextPrefillPositionE", false]], "tensorrt_llm::executor::requeststats::disservingstats (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats15disServingStatsE", false]], "tensorrt_llm::executor::requeststats::id (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats2idE", false]], "tensorrt_llm::executor::requeststats::kvcachehitrateperrequest (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats24kvCacheHitRatePerRequestE", false]], "tensorrt_llm::executor::requeststats::missedblocksperrequest (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats22missedBlocksPerRequestE", false]], "tensorrt_llm::executor::requeststats::numgeneratedtokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats18numGeneratedTokensE", false]], "tensorrt_llm::executor::requeststats::paused (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats6pausedE", false]], "tensorrt_llm::executor::requeststats::reusedblocksperrequest (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats22reusedBlocksPerRequestE", false]], "tensorrt_llm::executor::requeststats::scheduled (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats9scheduledE", false]], "tensorrt_llm::executor::requeststats::stage (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor12RequestStats5stageE", false]], "tensorrt_llm::executor::requeststatsperiteration (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor24RequestStatsPerIterationE", false]], "tensorrt_llm::executor::requeststatsperiteration::iter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor24RequestStatsPerIteration4iterE", false]], "tensorrt_llm::executor::requeststatsperiteration::requeststats (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor24RequestStatsPerIteration12requestStatsE", false]], "tensorrt_llm::executor::requesttype (c++ enum)": [[0, "_CPPv4N12tensorrt_llm8executor11RequestTypeE", false]], "tensorrt_llm::executor::requesttype::request_type_context_and_generation (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor11RequestType35REQUEST_TYPE_CONTEXT_AND_GENERATIONE", false]], "tensorrt_llm::executor::requesttype::request_type_context_only (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor11RequestType25REQUEST_TYPE_CONTEXT_ONLYE", false]], "tensorrt_llm::executor::requesttype::request_type_generation_only (c++ enumerator)": [[0, "_CPPv4N12tensorrt_llm8executor11RequestType28REQUEST_TYPE_GENERATION_ONLYE", false]], "tensorrt_llm::executor::response (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor8ResponseE", false]], "tensorrt_llm::executor::response::getclientid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Response11getClientIdEv", false]], "tensorrt_llm::executor::response::geterrormsg (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Response11getErrorMsgEv", false]], "tensorrt_llm::executor::response::getrequestid (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Response12getRequestIdEv", false]], "tensorrt_llm::executor::response::getresult (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Response9getResultEv", false]], "tensorrt_llm::executor::response::haserror (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor8Response8hasErrorEv", false]], "tensorrt_llm::executor::response::mimpl (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor8Response5mImplE", false]], "tensorrt_llm::executor::response::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8ResponseaSERK8Response", false], [0, "_CPPv4N12tensorrt_llm8executor8ResponseaSERR8Response", false]], "tensorrt_llm::executor::response::response (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8Response8ResponseE6IdType6ResultNSt8optionalI6IdTypeEE", false], [0, "_CPPv4N12tensorrt_llm8executor8Response8ResponseE6IdTypeNSt6stringENSt8optionalI6IdTypeEE", false], [0, "_CPPv4N12tensorrt_llm8executor8Response8ResponseERK8Response", false], [0, "_CPPv4N12tensorrt_llm8executor8Response8ResponseERR8Response", false]], "tensorrt_llm::executor::response::~response (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor8ResponseD0Ev", false]], "tensorrt_llm::executor::result (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor6ResultE", false]], "tensorrt_llm::executor::result::additionaloutputs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result17additionalOutputsE", false]], "tensorrt_llm::executor::result::avgdecodedtokensperiter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result23avgDecodedTokensPerIterE", false]], "tensorrt_llm::executor::result::contextlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result13contextLogitsE", false]], "tensorrt_llm::executor::result::contextphaseparams (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result18contextPhaseParamsE", false]], "tensorrt_llm::executor::result::cumlogprobs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result11cumLogProbsE", false]], "tensorrt_llm::executor::result::decodingiter (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result12decodingIterE", false]], "tensorrt_llm::executor::result::encoderoutput (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result13encoderOutputE", false]], "tensorrt_llm::executor::result::finishreasons (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result13finishReasonsE", false]], "tensorrt_llm::executor::result::generationlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result16generationLogitsE", false]], "tensorrt_llm::executor::result::isfinal (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result7isFinalE", false]], "tensorrt_llm::executor::result::issequencefinal (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result15isSequenceFinalE", false]], "tensorrt_llm::executor::result::logprobs (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result8logProbsE", false]], "tensorrt_llm::executor::result::outputtokenids (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result14outputTokenIdsE", false]], "tensorrt_llm::executor::result::requestperfmetrics (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result18requestPerfMetricsE", false]], "tensorrt_llm::executor::result::sequenceindex (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result13sequenceIndexE", false]], "tensorrt_llm::executor::result::specdecfastlogitsinfo (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Result21specDecFastLogitsInfoE", false]], "tensorrt_llm::executor::retentionpriority (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor17RetentionPriorityE", false]], "tensorrt_llm::executor::retentionpriorityandduration (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor28RetentionPriorityAndDurationE", false]], "tensorrt_llm::executor::retentionpriorityandduration::durationms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor28RetentionPriorityAndDuration10durationMsE", false]], "tensorrt_llm::executor::retentionpriorityandduration::retentionpriority (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor28RetentionPriorityAndDuration17retentionPriorityE", false]], "tensorrt_llm::executor::retentionpriorityandduration::retentionpriorityandduration (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor28RetentionPriorityAndDuration28RetentionPriorityAndDurationERKNSt8optionalI17RetentionPriorityEERKNSt8optionalINSt6chrono12millisecondsEEE", false]], "tensorrt_llm::executor::samplingconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfigE", false]], "tensorrt_llm::executor::samplingconfig::checkbeamsearchdiversityrate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig28checkBeamSearchDiversityRateERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checkbeamwidth (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig14checkBeamWidthE10SizeType32", false]], "tensorrt_llm::executor::samplingconfig::checkbeamwidtharray (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig19checkBeamWidthArrayERKNSt8optionalINSt6vectorI10SizeType32EEEEK10SizeType32", false]], "tensorrt_llm::executor::samplingconfig::checkearlystopping (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig18checkEarlyStoppingERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::checklengthpenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig18checkLengthPenaltyERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checkminp (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig9checkMinPERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checkmintokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig14checkMinTokensERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::checknorepeatngramsize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig22checkNoRepeatNgramSizeERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::checknumreturnsequences (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig23checkNumReturnSequencesERKNSt8optionalI10SizeType32EE10SizeType32", false]], "tensorrt_llm::executor::samplingconfig::checkpromptignorelength (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig23checkPromptIgnoreLengthERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::checkrepetitionpenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig22checkRepetitionPenaltyERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checktemperature (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig16checkTemperatureERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checktopk (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig9checkTopKERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checktopp (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig9checkTopPERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checktoppdecay (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig14checkTopPDecayERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checktoppmin (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig12checkTopPMinERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::checktoppresetids (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig17checkTopPResetIdsERKNSt8optionalI11TokenIdTypeEE", false]], "tensorrt_llm::executor::samplingconfig::getbeamsearchdiversityrate (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig26getBeamSearchDiversityRateEv", false]], "tensorrt_llm::executor::samplingconfig::getbeamwidth (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig12getBeamWidthEv", false]], "tensorrt_llm::executor::samplingconfig::getbeamwidtharray (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig17getBeamWidthArrayEv", false]], "tensorrt_llm::executor::samplingconfig::getearlystopping (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig16getEarlyStoppingEv", false]], "tensorrt_llm::executor::samplingconfig::getfrequencypenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig19getFrequencyPenaltyEv", false]], "tensorrt_llm::executor::samplingconfig::getlengthpenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig16getLengthPenaltyEv", false]], "tensorrt_llm::executor::samplingconfig::getminp (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig7getMinPEv", false]], "tensorrt_llm::executor::samplingconfig::getmintokens (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig12getMinTokensEv", false]], "tensorrt_llm::executor::samplingconfig::getnorepeatngramsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig20getNoRepeatNgramSizeEv", false]], "tensorrt_llm::executor::samplingconfig::getnumreturnbeams (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig17getNumReturnBeamsEv", false]], "tensorrt_llm::executor::samplingconfig::getnumreturnsequences (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig21getNumReturnSequencesEv", false]], "tensorrt_llm::executor::samplingconfig::getpresencepenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig18getPresencePenaltyEv", false]], "tensorrt_llm::executor::samplingconfig::getpromptignorelength (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig21getPromptIgnoreLengthEv", false]], "tensorrt_llm::executor::samplingconfig::getrepetitionpenalty (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig20getRepetitionPenaltyEv", false]], "tensorrt_llm::executor::samplingconfig::getseed (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig7getSeedEv", false]], "tensorrt_llm::executor::samplingconfig::gettemperature (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig14getTemperatureEv", false]], "tensorrt_llm::executor::samplingconfig::gettopk (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig7getTopKEv", false]], "tensorrt_llm::executor::samplingconfig::gettopp (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig7getTopPEv", false]], "tensorrt_llm::executor::samplingconfig::gettoppdecay (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig12getTopPDecayEv", false]], "tensorrt_llm::executor::samplingconfig::gettoppmin (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig10getTopPMinEv", false]], "tensorrt_llm::executor::samplingconfig::gettoppresetids (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfig15getTopPResetIdsEv", false]], "tensorrt_llm::executor::samplingconfig::mbeamsearchdiversityrate (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig24mBeamSearchDiversityRateE", false]], "tensorrt_llm::executor::samplingconfig::mbeamwidth (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig10mBeamWidthE", false]], "tensorrt_llm::executor::samplingconfig::mbeamwidtharray (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig15mBeamWidthArrayE", false]], "tensorrt_llm::executor::samplingconfig::mearlystopping (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig14mEarlyStoppingE", false]], "tensorrt_llm::executor::samplingconfig::mfrequencypenalty (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig17mFrequencyPenaltyE", false]], "tensorrt_llm::executor::samplingconfig::mlengthpenalty (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig14mLengthPenaltyE", false]], "tensorrt_llm::executor::samplingconfig::mminp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig5mMinPE", false]], "tensorrt_llm::executor::samplingconfig::mmintokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig10mMinTokensE", false]], "tensorrt_llm::executor::samplingconfig::mnorepeatngramsize (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig18mNoRepeatNgramSizeE", false]], "tensorrt_llm::executor::samplingconfig::mnumreturnbeams (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig15mNumReturnBeamsE", false]], "tensorrt_llm::executor::samplingconfig::mnumreturnsequences (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig19mNumReturnSequencesE", false]], "tensorrt_llm::executor::samplingconfig::mpresencepenalty (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig16mPresencePenaltyE", false]], "tensorrt_llm::executor::samplingconfig::mpromptignorelength (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig19mPromptIgnoreLengthE", false]], "tensorrt_llm::executor::samplingconfig::mrepetitionpenalty (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig18mRepetitionPenaltyE", false]], "tensorrt_llm::executor::samplingconfig::mseed (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig5mSeedE", false]], "tensorrt_llm::executor::samplingconfig::mtemperature (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig12mTemperatureE", false]], "tensorrt_llm::executor::samplingconfig::mtopk (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig5mTopKE", false]], "tensorrt_llm::executor::samplingconfig::mtopp (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig5mTopPE", false]], "tensorrt_llm::executor::samplingconfig::mtoppdecay (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig10mTopPDecayE", false]], "tensorrt_llm::executor::samplingconfig::mtoppmin (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig8mTopPMinE", false]], "tensorrt_llm::executor::samplingconfig::mtoppresetids (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig13mTopPResetIdsE", false]], "tensorrt_llm::executor::samplingconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor14SamplingConfigeqERK14SamplingConfig", false]], "tensorrt_llm::executor::samplingconfig::samplingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig14SamplingConfigE10SizeType32RKNSt8optionalI10SizeType32EERKNSt8optionalI9FloatTypeEERKNSt8optionalI9FloatTypeEERKNSt8optionalI11TokenIdTypeEERKNSt8optionalI9FloatTypeEERKNSt8optionalI14RandomSeedTypeEERKNSt8optionalI9FloatTypeEERKNSt8optionalI10SizeType32EERKNSt8optionalI9FloatTypeEERKNSt8optionalI9FloatTypeEERKNSt8optionalI9FloatTypeEERKNSt8optionalI9FloatTypeEERKNSt8optionalI10SizeType32EERKNSt8optionalI9FloatTypeEERKNSt8optionalI10SizeType32EERKNSt8optionalI10SizeType32EERKNSt8optionalI10SizeType32EERKNSt8optionalI9FloatTypeEERKNSt8optionalINSt6vectorI10SizeType32EEEE", false]], "tensorrt_llm::executor::samplingconfig::setbeamsearchdiversityrate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig26setBeamSearchDiversityRateERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::setbeamwidth (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig12setBeamWidthE10SizeType32", false]], "tensorrt_llm::executor::samplingconfig::setbeamwidtharray (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig17setBeamWidthArrayERKNSt8optionalINSt6vectorI10SizeType32EEEE", false]], "tensorrt_llm::executor::samplingconfig::setearlystopping (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig16setEarlyStoppingERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::setfrequencypenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig19setFrequencyPenaltyERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::setlengthpenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig16setLengthPenaltyERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::setminp (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig7setMinPERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::setmintokens (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig12setMinTokensERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::setnorepeatngramsize (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig20setNoRepeatNgramSizeERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::setnumreturnsequences (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig21setNumReturnSequencesERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::setpresencepenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig18setPresencePenaltyERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::setpromptignorelength (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig21setPromptIgnoreLengthERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::setrepetitionpenalty (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig20setRepetitionPenaltyERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::setseed (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig7setSeedERKNSt8optionalI14RandomSeedTypeEE", false]], "tensorrt_llm::executor::samplingconfig::settemperature (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig14setTemperatureERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::settopk (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig7setTopKERKNSt8optionalI10SizeType32EE", false]], "tensorrt_llm::executor::samplingconfig::settopp (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig7setTopPERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::settoppdecay (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig12setTopPDecayERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::settoppmin (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig10setTopPMinERKNSt8optionalI9FloatTypeEE", false]], "tensorrt_llm::executor::samplingconfig::settoppresetids (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig15setTopPResetIdsERKNSt8optionalI11TokenIdTypeEE", false]], "tensorrt_llm::executor::samplingconfig::updatenumreturnbeams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor14SamplingConfig20updateNumReturnBeamsEv", false]], "tensorrt_llm::executor::schedulerconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor15SchedulerConfigE", false]], "tensorrt_llm::executor::schedulerconfig::getcapacityschedulerpolicy (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15SchedulerConfig26getCapacitySchedulerPolicyEv", false]], "tensorrt_llm::executor::schedulerconfig::getcontextchunkingpolicy (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15SchedulerConfig24getContextChunkingPolicyEv", false]], "tensorrt_llm::executor::schedulerconfig::getdynamicbatchconfig (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15SchedulerConfig21getDynamicBatchConfigEv", false]], "tensorrt_llm::executor::schedulerconfig::mcapacityschedulerpolicy (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15SchedulerConfig24mCapacitySchedulerPolicyE", false]], "tensorrt_llm::executor::schedulerconfig::mcontextchunkingpolicy (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15SchedulerConfig22mContextChunkingPolicyE", false]], "tensorrt_llm::executor::schedulerconfig::mdynamicbatchconfig (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor15SchedulerConfig19mDynamicBatchConfigE", false]], "tensorrt_llm::executor::schedulerconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor15SchedulerConfigeqERK15SchedulerConfig", false]], "tensorrt_llm::executor::schedulerconfig::schedulerconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor15SchedulerConfig15SchedulerConfigE23CapacitySchedulerPolicyNSt8optionalI21ContextChunkingPolicyEENSt8optionalI18DynamicBatchConfigEE", false]], "tensorrt_llm::executor::serialization (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor13SerializationE", false]], "tensorrt_llm::executor::serialization::deserializeadditionalmodeloutput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization32deserializeAdditionalModelOutputERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeadditionaloutput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization27deserializeAdditionalOutputERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeagentstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization21deserializeAgentStateERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeblockkey (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization19deserializeBlockKeyERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializebool (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization15deserializeBoolERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializecachestate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization21deserializeCacheStateERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializecachetransceiverconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization33deserializeCacheTransceiverConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializecommstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization20deserializeCommStateERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializecontextphaseparams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializeContextPhaseParamsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializedatatransceiverstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization31deserializeDataTransceiverStateERNSt6vectorIcEE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization31deserializeDataTransceiverStateERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializedebugconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization22deserializeDebugConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializedecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization25deserializeDecodingConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializedecodingmode (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization23deserializeDecodingModeERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializedisservingrequeststats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization33deserializeDisServingRequestStatsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializedynamicbatchconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializeDynamicBatchConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeeagleconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization22deserializeEagleConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeexecutorconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization25deserializeExecutorConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeextendedruntimeperfknobconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization40deserializeExtendedRuntimePerfKnobConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeexternaldrafttokensconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization36deserializeExternalDraftTokensConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeguideddecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization31deserializeGuidedDecodingConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeguideddecodingparams (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization31deserializeGuidedDecodingParamsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeinflightbatchingstats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization32deserializeInflightBatchingStatsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeiterationstats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization25deserializeIterationStatsERNSt6vectorIcEE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization25deserializeIterationStatsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeiterationstatsvec (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization28deserializeIterationStatsVecERNSt6vectorIcEE", false]], "tensorrt_llm::executor::serialization::deserializekvcacheconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization24deserializeKvCacheConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcachecreateddata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializeKVCacheCreatedDataERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcacheevent (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization23deserializeKVCacheEventERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcacheeventdiff (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor13Serialization27deserializeKVCacheEventDiffE16KVCacheEventDiffI1TERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcacheevents (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization24deserializeKVCacheEventsERNSt6vectorIcEE", false]], "tensorrt_llm::executor::serialization::deserializekvcacheremoveddata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializeKVCacheRemovedDataERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcacheretentionconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization33deserializeKvCacheRetentionConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcachestats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization23deserializeKvCacheStatsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcachestoredblockdata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization33deserializeKVCacheStoredBlockDataERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcachestoreddata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization28deserializeKVCacheStoredDataERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializekvcacheupdateddata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializeKVCacheUpdatedDataERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializelookaheaddecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization34deserializeLookaheadDecodingConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeloraconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization21deserializeLoraConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializemodeltype (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization20deserializeModelTypeERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializemropeconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization22deserializeMropeConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializemultimodalinput (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization26deserializeMultimodalInputERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeorchestratorconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializeOrchestratorConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeoutputconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization23deserializeOutputConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeparallelconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization25deserializeParallelConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializepeftcacheconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization26deserializePeftCacheConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeprompttuningconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializePromptTuningConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializerequest (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization18deserializeRequestERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializerequestperfmetrics (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization29deserializeRequestPerfMetricsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializerequeststage (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization23deserializeRequestStageERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializerequeststats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization23deserializeRequestStatsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializerequeststatsperiteration (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization35deserializeRequestStatsPerIterationERNSt6vectorIcEE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization35deserializeRequestStatsPerIterationERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializerequeststatsperiterationvec (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization38deserializeRequestStatsPerIterationVecERNSt6vectorIcEE", false]], "tensorrt_llm::executor::serialization::deserializeresponse (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization19deserializeResponseERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeresponses (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization20deserializeResponsesERNSt6vectorIcEE", false]], "tensorrt_llm::executor::serialization::deserializeresult (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization17deserializeResultERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializesamplingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization25deserializeSamplingConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeschedulerconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization26deserializeSchedulerConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializesocketstate (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization22deserializeSocketStateERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializespecdecfastlogitsinfo (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization32deserializeSpecDecFastLogitsInfoERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializespecdecodingstats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization28deserializeSpecDecodingStatsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializespeculativedecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization36deserializeSpeculativeDecodingConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializestaticbatchingstats (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization30deserializeStaticBatchingStatsERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializestring (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization17deserializeStringERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializetensor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization17deserializeTensorERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializetimepoint (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization20deserializeTimePointERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializetokenrangeretentionconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization36deserializeTokenRangeRetentionConfigERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::deserializeuniquetoken (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor13Serialization22deserializeUniqueTokenERNSt7istreamE", false]], "tensorrt_llm::executor::serialization::serialize (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor13Serialization9serializeEvRK16KVCacheEventDiffI1TERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK10LoraConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK11DebugConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK11EagleConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK11MropeConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK12DecodingModeRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK12KVCacheEventRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK12KvCacheStatsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK12OutputConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK12RequestStageRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK12RequestStatsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK13KvCacheConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK14DecodingConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK14ExecutorConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK14IterationStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK14IterationStatsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK14ParallelConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK14SamplingConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK15MultimodalInputRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK15PeftCacheConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK15SchedulerConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK16AdditionalOutputRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK17KVCacheStoredDataRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK17SpecDecodingStatsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18ContextPhaseParamsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18DynamicBatchConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18KVCacheCreatedDataRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18KVCacheRemovedDataRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18KVCacheUpdatedDataRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18OrchestratorConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18PromptTuningConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK18RequestPerfMetricsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK19StaticBatchingStatsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK20DataTransceiverState", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK20DataTransceiverStateRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK20GuidedDecodingConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK20GuidedDecodingParamsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK21AdditionalModelOutputRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK21InflightBatchingStatsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK22CacheTransceiverConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK22DisServingRequestStatsRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK22KVCacheStoredBlockDataRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK22KvCacheRetentionConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK23LookaheadDecodingConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK24RequestStatsPerIteration", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK24RequestStatsPerIterationRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK25ExternalDraftTokensConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK25SpeculativeDecodingConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK29ExtendedRuntimePerfKnobConfigRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK33SpeculativeDecodingFastLogitsInfoRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK6ResultRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK6TensorRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK7RequestRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERK8ResponseRNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN12tensorrt_llm13batch_manager16kv_cache_manager8BlockKeyERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN12tensorrt_llm7runtime11UniqueTokenERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN18RequestPerfMetrics9TimePointERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN22KvCacheRetentionConfig25TokenRangeRetentionConfigERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN8kv_cache10AgentStateERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN8kv_cache10CacheStateERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN8kv_cache11SocketStateERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKN8kv_cache9CommStateERNSt7ostreamE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKNSt5dequeI12KVCacheEventEE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKNSt6vectorI14IterationStatsEE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKNSt6vectorI24RequestStatsPerIterationEE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization9serializeERKNSt6vectorI8ResponseEE", false]], "tensorrt_llm::executor::serialization::serializedsize (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor13Serialization14serializedSizeE6size_tRK16KVCacheEventDiffI1TE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK10LoraConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK11DebugConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK11EagleConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK11MropeConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK12DecodingMode", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK12KVCacheEvent", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK12KvCacheStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK12OutputConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK12RequestStage", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK12RequestStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK13KvCacheConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK14DecodingConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK14ExecutorConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK14IterationStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK14ParallelConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK14SamplingConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK15MultimodalInput", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK15PeftCacheConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK15SchedulerConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK16AdditionalOutput", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK17KVCacheStoredData", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK17SpecDecodingStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18ContextPhaseParams", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18DynamicBatchConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18KVCacheCreatedData", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18KVCacheRemovedData", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18KVCacheUpdatedData", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18OrchestratorConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18PromptTuningConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK18RequestPerfMetrics", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK19StaticBatchingStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK20DataTransceiverState", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK20GuidedDecodingConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK20GuidedDecodingParams", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK21AdditionalModelOutput", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK21InflightBatchingStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK22CacheTransceiverConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK22DisServingRequestStats", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK22KVCacheStoredBlockData", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK22KvCacheRetentionConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK23LookaheadDecodingConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK24RequestStatsPerIteration", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK25ExternalDraftTokensConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK25SpeculativeDecodingConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK29ExtendedRuntimePerfKnobConfig", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK33SpeculativeDecodingFastLogitsInfo", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK6Result", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK6Tensor", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK7Request", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERK8Response", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN12tensorrt_llm13batch_manager16kv_cache_manager8BlockKeyE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN12tensorrt_llm7runtime11UniqueTokenE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN18RequestPerfMetrics9TimePointE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN22KvCacheRetentionConfig25TokenRangeRetentionConfigE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN8kv_cache10AgentStateE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN8kv_cache10CacheStateE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN8kv_cache11SocketStateE", false], [0, "_CPPv4N12tensorrt_llm8executor13Serialization14serializedSizeERKN8kv_cache9CommStateE", false]], "tensorrt_llm::executor::shape (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor5ShapeE", false]], "tensorrt_llm::executor::shape::base (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor5Shape4BaseE", false]], "tensorrt_llm::executor::shape::dimtype64 (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor5Shape9DimType64E", false]], "tensorrt_llm::executor::shape::shape (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor5Shape5ShapeENSt16initializer_listI9DimType64EE", false], [0, "_CPPv4N12tensorrt_llm8executor5Shape5ShapeEPK9DimType64N4Base9size_typeE", false], [0, "_CPPv4N12tensorrt_llm8executor5Shape5ShapeEv", false]], "tensorrt_llm::executor::sizetype32 (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor10SizeType32E", false]], "tensorrt_llm::executor::sizetype64 (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor10SizeType64E", false]], "tensorrt_llm::executor::specdecodingstats (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor17SpecDecodingStatsE", false]], "tensorrt_llm::executor::specdecodingstats::acceptancelength (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17SpecDecodingStats16acceptanceLengthE", false]], "tensorrt_llm::executor::specdecodingstats::draftoverhead (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17SpecDecodingStats13draftOverheadE", false]], "tensorrt_llm::executor::specdecodingstats::iterlatencyms (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17SpecDecodingStats13iterLatencyMSE", false]], "tensorrt_llm::executor::specdecodingstats::numacceptedtokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17SpecDecodingStats17numAcceptedTokensE", false]], "tensorrt_llm::executor::specdecodingstats::numdrafttokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17SpecDecodingStats14numDraftTokensE", false]], "tensorrt_llm::executor::specdecodingstats::numrequestswithdrafttokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor17SpecDecodingStats26numRequestsWithDraftTokensE", false]], "tensorrt_llm::executor::speculativedecodingconfig (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor25SpeculativeDecodingConfigE", false]], "tensorrt_llm::executor::speculativedecodingconfig::fastlogits (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor25SpeculativeDecodingConfig10fastLogitsE", false]], "tensorrt_llm::executor::speculativedecodingconfig::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor25SpeculativeDecodingConfigeqERK25SpeculativeDecodingConfig", false]], "tensorrt_llm::executor::speculativedecodingconfig::speculativedecodingconfig (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor25SpeculativeDecodingConfig25SpeculativeDecodingConfigEb", false]], "tensorrt_llm::executor::speculativedecodingfastlogitsinfo (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor33SpeculativeDecodingFastLogitsInfoE", false]], "tensorrt_llm::executor::speculativedecodingfastlogitsinfo::draftparticipantid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor33SpeculativeDecodingFastLogitsInfo18draftParticipantIdE", false]], "tensorrt_llm::executor::speculativedecodingfastlogitsinfo::draftrequestid (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor33SpeculativeDecodingFastLogitsInfo14draftRequestIdE", false]], "tensorrt_llm::executor::speculativedecodingfastlogitsinfo::totensor (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor33SpeculativeDecodingFastLogitsInfo8toTensorEv", false]], "tensorrt_llm::executor::staticbatchingstats (c++ struct)": [[0, "_CPPv4N12tensorrt_llm8executor19StaticBatchingStatsE", false]], "tensorrt_llm::executor::staticbatchingstats::emptygenslots (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor19StaticBatchingStats13emptyGenSlotsE", false]], "tensorrt_llm::executor::staticbatchingstats::numcontextrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor19StaticBatchingStats18numContextRequestsE", false]], "tensorrt_llm::executor::staticbatchingstats::numctxtokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor19StaticBatchingStats12numCtxTokensE", false]], "tensorrt_llm::executor::staticbatchingstats::numgentokens (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor19StaticBatchingStats12numGenTokensE", false]], "tensorrt_llm::executor::staticbatchingstats::numscheduledrequests (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor19StaticBatchingStats20numScheduledRequestsE", false]], "tensorrt_llm::executor::streamptr (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor9StreamPtrE", false]], "tensorrt_llm::executor::tensor (c++ class)": [[0, "_CPPv4N12tensorrt_llm8executor6TensorE", false]], "tensorrt_llm::executor::tensor::copyto (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor6copyToENSt10shared_ptrI4ImplEE13CudaStreamPtr", false]], "tensorrt_llm::executor::tensor::copytocpu (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor9copyToCpuEN6Tensor13CudaStreamPtrE", false]], "tensorrt_llm::executor::tensor::copytogpu (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor9copyToGpuEN6Tensor13CudaStreamPtrE", false]], "tensorrt_llm::executor::tensor::copytomanaged (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor13copyToManagedEN6Tensor13CudaStreamPtrE", false]], "tensorrt_llm::executor::tensor::copytopinned (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor12copyToPinnedEN6Tensor13CudaStreamPtrE", false]], "tensorrt_llm::executor::tensor::copytopooledpinned (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor18copyToPooledPinnedEN6Tensor13CudaStreamPtrE", false]], "tensorrt_llm::executor::tensor::cpu (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor3cpuE6Tensor5Shape", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor3cpuE8DataType5Shape", false]], "tensorrt_llm::executor::tensor::cudastreamptr (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor13CudaStreamPtrE", false]], "tensorrt_llm::executor::tensor::detail::ofitensor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor6detail9ofITensorENSt10shared_ptrIN7runtime7ITensorEEE", false]], "tensorrt_llm::executor::tensor::detail::toitensor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor6detail9toITensorERK6Tensor", false]], "tensorrt_llm::executor::tensor::getdata (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor7getDataEv", false], [0, "_CPPv4NK12tensorrt_llm8executor6Tensor7getDataEv", false]], "tensorrt_llm::executor::tensor::getdatatype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor11getDataTypeEv", false]], "tensorrt_llm::executor::tensor::getmemorytype (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor13getMemoryTypeEv", false]], "tensorrt_llm::executor::tensor::getruntimetype (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor14getRuntimeTypeE8DataTypev", false]], "tensorrt_llm::executor::tensor::getshape (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor8getShapeEv", false]], "tensorrt_llm::executor::tensor::getsize (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor7getSizeEv", false]], "tensorrt_llm::executor::tensor::getsizeinbytes (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6Tensor14getSizeInBytesEv", false]], "tensorrt_llm::executor::tensor::gpu (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor3gpuE6Tensor13CudaStreamPtr5Shape", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor3gpuE8DataType13CudaStreamPtr5Shape", false]], "tensorrt_llm::executor::tensor::impl (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor4ImplE", false]], "tensorrt_llm::executor::tensor::managed (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor7managedE6Tensor5Shape", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor7managedE8DataType5Shape", false]], "tensorrt_llm::executor::tensor::mtensor (c++ member)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor7mTensorE", false]], "tensorrt_llm::executor::tensor::of (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor2ofE6TensorP1T5Shape", false], [0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor2ofE6TensorR1T", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor2ofE8DataTypePv5Shape", false]], "tensorrt_llm::executor::tensor::operator bool (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6TensorcvbEv", false]], "tensorrt_llm::executor::tensor::operator!= (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6TensorneERK6Tensor", false]], "tensorrt_llm::executor::tensor::operator= (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6TensoraSERK6Tensor", false], [0, "_CPPv4N12tensorrt_llm8executor6TensoraSERR6Tensor", false]], "tensorrt_llm::executor::tensor::operator== (c++ function)": [[0, "_CPPv4NK12tensorrt_llm8executor6TensoreqERK6Tensor", false]], "tensorrt_llm::executor::tensor::pinned (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor6pinnedE6Tensor5Shape", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor6pinnedE8DataType5Shape", false]], "tensorrt_llm::executor::tensor::pooledpinned (c++ function)": [[0, "_CPPv4I0EN12tensorrt_llm8executor6Tensor12pooledPinnedE6Tensor5Shape", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor12pooledPinnedE8DataType5Shape", false]], "tensorrt_llm::executor::tensor::setfrom (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor7setFromERK6Tensor13CudaStreamPtr", false]], "tensorrt_llm::executor::tensor::setzero (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor7setZeroE13CudaStreamPtr", false]], "tensorrt_llm::executor::tensor::tensor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6Tensor6TensorENSt10shared_ptrIN7runtime7ITensorEEE", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor6TensorERK6Tensor", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor6TensorERR6Tensor", false], [0, "_CPPv4N12tensorrt_llm8executor6Tensor6TensorEv", false]], "tensorrt_llm::executor::tensor::~tensor (c++ function)": [[0, "_CPPv4N12tensorrt_llm8executor6TensorD0Ev", false]], "tensorrt_llm::executor::tensorptr (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor9TensorPtrE", false]], "tensorrt_llm::executor::tokenidtype (c++ type)": [[0, "_CPPv4N12tensorrt_llm8executor11TokenIdTypeE", false]], "tensorrt_llm::executor::typetraits (c++ struct)": [[0, "_CPPv4I0_bEN12tensorrt_llm8executor10TypeTraitsE", false]], "tensorrt_llm::executor::typetraitstrtllm-serve", "trtllm-serve", "Deployment Guide for DeepSeek R1 on TensorRT LLM - Blackwell & Hopper Hardware", "Deployment Guide for GPT-OSS on TensorRT-LLM - Blackwell Hardware", "Deployment Guide for Llama3.3 70B on TensorRT LLM - Blackwell & Hopper Hardware", "Deployment Guide for Llama4 Scout 17B on TensorRT LLM - Blackwell & Hopper Hardware", "Deployment Guide for Qwen3 Next on TensorRT LLM - Blackwell & Hopper Hardware", "Model Recipes", "LLM API Change Guide", "Continuous Integration Overview", "Using Dev Containers", "Introduction to KV Cache Transmission", "Architecture Overview", "Performance Analysis", "TensorRT LLM Benchmarking", "Overview", "Curl Chat Client", "Curl Chat Client For Multimodal", "Curl Completion Client", "LLM Common Customizations", "Deepseek R1 Reasoning Parser", "Dynamo K8s Example", "Genai Perf Client", "Genai Perf Client For Multimodal", "LLM Examples Introduction", "How to Change KV Cache Behavior", "How to Change Block Priorities", "LLM Examples", "Generate text with guided decoding", "Generate text", "Generate text asynchronously", "Generate text in streaming", "Distributed LLM Generation", "KV Cache Connector", "KV Cache Offloading", "Control generated text using logits processor", "Run LLM-API with pytorch backend on Slurm", "Run trtllm-bench with pytorch backend on Slurm", "Run trtllm-serve with pytorch backend on Slurm", "Generate text with multiple LoRA adapters", "Runtime Configuration Examples", "Sampling Techniques Showcase", "Sparse Attention", "Speculative Decoding", "OpenAI Chat Client", "OpenAI Chat Client for Multimodal", "OpenAI Completion Client", "Openai Completion Client For Lora", "OpenAI Completion Client with JSON Schema", "Online Serving Examples", "Additional Outputs", "Multi-Head, Multi-Query, and Group-Query Attention", "Benchmarking with trtllm-bench", "Example Run Script", "Expert Configuration of LLM API", "Logging Level", "Incorporating auto_deploy into your own workflow", "AutoDeploy (Prototype)", "Support Matrix", "Checkpoint Loading", "Disaggregated Serving", "Feature Combination Matrix", "KV Cache System", "Long Sequences", "LoRA (Low-Rank Adaptation)", "Multimodal Support in TensorRT LLM", "Overlap Scheduler", "Paged Attention, IFB, and Request Scheduling", "Parallelism in TensorRT LLM", "Quantization", "Ray Orchestrator (Prototype)", "Sampling", "Speculative Decoding", "Torch Compile & Piecewise CUDA Graph", "Welcome to TensorRT LLM\u2019s Documentation!", "Building from Source Code on Linux", "Pre-built release container images on NGC", "Installation", "Installing on Linux via pip", "Disaggregated-Service (Prototype)", "Executor API", "Expert Parallelism in TensorRT-LLM", "Multi-Head, Multi-Query, and Group-Query Attention", "C++ GPT Runtime", "Graph Rewriting Module", "KV Cache Management: Pools, Blocks, and Events", "KV cache reuse", "Run gpt-2b + LoRA using Executor / cpp runtime", "Low-Precision-AllReduce", "<no title>", "Speculative Sampling", "Running With Weight Streaming to Reduce GPU Memory Consumption", "Adding a Model", "TensorRT LLM Checkpoint", "Model Definition", "TensorRT-LLM Model Weights Loader", "TensorRT-LLM Build Workflow", "Build the TensorRT LLM Docker Image", "Develop TensorRT LLM on Runpod", "Key Features", "Performance Analysis", "TensorRT-LLM Benchmarking", "Benchmarking Default Performance", "Deciding Model Sharding Strategy", "FP8 Quantization", "Performance Tuning Guide", "Prerequisite Knowledge", "Tuning Max Batch Size and Max Num Tokens", "Useful Build-Time Flags", "Useful Runtime Options", "Functionals", "Layers", "Models", "Plugin", "Quantization", "Runtime", "Memory Usage of TensorRT-LLM", "Multimodal Feature Support Matrix (PyTorch Backend)", "Numerical Precision", "Support Matrix", "Troubleshooting", "LLM API with TensorRT Engine", "PyTorch Backend", "LLM API Introduction", "API Reference", "Adding a New Model", "Supported Models", "Overview", "Quick Start Guide", "Release Notes", "Adding a New Model in PyTorch Backend", "Architecture Ovewiew", "Attention", "Benchmarking with trtllm-bench", "Example Run Script", "Expert Configuration of LLM API", "Logging Level", "Serving with trtllm-serve", "Incorporating auto_deploy into your own workflow", "AutoDeploy", "Support Matrix", "Checkpoint Loading", "LoRA (Low-Rank Adaptation)", "Overlap Scheduler", "Quantization", "Sampling", "KV Cache Manager", "Scheduler"], "titleterms": {"": [4, 7, 77, 100, 108], "0": [149, 155], "000": [4, 5], "0528": 2, "0rc0": 9, "1": [2, 8, 9, 16, 18, 19, 20, 34, 41, 85, 94, 101, 118, 120, 142, 149, 155, 167], "10": [4, 155], "100m": 4, "1024": 17, "11": 155, "12": [5, 155], "1200": 17, "120b": [9, 21], "13": 155, "13b": 5, "14": 155, "15": 155, "16": 155, "17": 155, "17b": 31, "18": 155, "180b": 3, "19": 155, "2": [2, 6, 8, 16, 18, 19, 20, 34, 85, 94, 101, 118, 142, 149, 155, 167], "20": 155, "21": 155, "235b": 41, "256": 17, "2b": 113, "3": [2, 12, 16, 17, 18, 20, 30, 34, 41, 85, 94, 98, 118, 120, 127, 142, 145, 167], "30b": 41, "4": [2, 4, 18, 20, 41, 85, 118, 167], "405b": [41, 120], "4096": 17, "4400": 17, "4x": 6, "5": [2, 18], "6": [2, 3, 18], "6x": 4, "7": [18, 155], "70b": [3, 6, 30, 41, 120, 127], "7x": 3, "8": 155, "8192": 17, "8b": 41, "9": 155, "A": [11, 15], "As": 106, "For": [43, 49, 73], "In": [77, 93, 106, 108], "It": [92, 169], "Not": 142, "One": [13, 101], "The": [11, 16, 77, 88, 93, 106, 144], "To": 128, "With": [117, 153], "_prepare_draft_request": 98, "_prepare_draft_token": 98, "_torchllm": 34, "a100": [3, 4], "a22b": 41, "a3b": 41, "about": [24, 26, 27, 116, 129, 153], "absorb": 15, "accept": [13, 14], "access": [30, 31, 123], "account": 124, "accuraci": [7, 14, 28, 29, 30, 31, 114], "achiev": [4, 5, 14], "acknowledg": [8, 10, 11, 12, 13, 14, 15, 16, 17, 20], "across": 88, "activ": [137, 142], "ad": [34, 118, 151, 156], "adapt": [40, 65, 90, 127, 168], "add": [12, 34], "addit": [76, 106], "additional_model_output": 76, "address": 20, "adp": [8, 15], "advanc": [78, 83, 90, 101, 153, 159, 165, 168], "after": 149, "agent": 37, "algorithm": [19, 114], "alibi": [77, 108], "alloc": 88, "allreduc": 114, "alltoal": 12, "also": 163, "altern": 18, "an": [11, 34, 111], "analysi": [8, 10, 39, 126], "announc": 155, "api": [27, 28, 29, 30, 31, 32, 34, 62, 80, 100, 106, 110, 117, 122, 128, 147, 149, 150, 154, 155, 157, 161], "appli": 10, "approach": 20, "arbitrari": 106, "architect": 153, "architectur": [11, 13, 34, 38, 94, 96, 98, 153, 157], "argument": [23, 34, 80, 161], "art": 153, "artifact": 18, "asynchron": 56, "asyncio": 45, "aten": 99, "attempt": 20, "attent": [13, 14, 15, 68, 77, 84, 88, 89, 93, 94, 108, 119, 133, 134, 135, 137, 158, 166], "attention_backend": [26, 28, 30, 31], "attentionbackend": [77, 158], "attentionmetadata": [77, 158], "auto": [19, 99], "auto_deploi": [82, 164], "autodeploi": [83, 165], "autodeploy_config": [78, 159], "autoregress": 13, "auxiliari": 20, "avoid": [20, 35, 128], "awq": [3, 119, 144], "b200": [2, 9, 13], "backend": [13, 17, 21, 28, 29, 62, 63, 64, 77, 84, 86, 99, 143, 145, 148, 152, 156, 158, 166], "background": [8, 10, 11, 13, 14, 19, 90, 91, 99, 168], "balanc": [8, 13, 16, 20, 94], "base": [9, 11, 14, 45], "basecheckpointload": [85, 167], "baseconfigload": [85, 167], "baselin": [8, 130], "baseweightload": [85, 167], "baseweightmapp": [85, 167], "basic": [14, 28, 29, 30, 31, 32, 53, 78, 85, 88, 90, 96, 159, 167, 168], "batch": [77, 93, 106, 108, 133], "beam": [77, 97, 106, 108], "befor": [40, 127, 128], "begin": 128, "behavior": [51, 88, 127], "bench": [22, 39, 63, 78, 90, 91, 98, 126, 128, 159, 168], "benchmark": [2, 7, 26, 28, 29, 30, 31, 32, 40, 41, 78, 127, 128, 159], "best": [2, 7, 34, 35, 94], "beta": 34, "between": 10, "bf16": 144, "bia": 108, "bind": [20, 101, 106, 120], "blackwel": [15, 28, 29, 30, 31, 32, 144], "block": [52, 111], "blockmanag": 111, "blog": 100, "boost": [40, 127], "boundari": [12, 13], "break": 99, "breakdown": 26, "budget": 6, "buffer": [77, 108, 130, 142], "buffermanag": 1, "build": [2, 18, 22, 23, 45, 101, 119, 122, 123, 124, 127, 128, 134], "build_and_run_ad": [80, 161], "built": [80, 85, 102, 161, 167], "c": [16, 101, 106, 109, 142], "cach": [2, 17, 37, 51, 59, 60, 77, 86, 88, 90, 93, 95, 98, 108, 111, 112, 119, 130, 135, 142, 168, 172], "cachecommun": 0, "callback": 10, "can": [112, 149, 153], "cannot": 149, "capabl": 153, "capac": 135, "captur": 10, "case": [19, 21, 133], "cast": 137, "caveat": [40, 127], "challeng": 10, "chang": [34, 51, 52, 117, 133, 155], "chat": [9, 27, 42, 43, 70, 71], "check": 9, "checkpoint": [85, 119, 167], "choos": 7, "chunk": [2, 77, 89, 93, 108, 133, 135], "ci": 35, "class": [34, 106], "classic": 110, "cli": [80, 100, 122, 128, 161], "client": [42, 43, 44, 48, 49, 70, 71, 72, 73, 74, 90, 168], "clock": [2, 40, 127], "clone": 18, "close": [3, 6], "cluster": 86, "cnn_dailymail": 24, "code": [28, 29, 30, 31, 32, 101], "collect": [16, 39, 126], "combin": [2, 10, 87], "come": 7, "command": 41, "commit": 34, "common": [1, 34, 45, 99], "commun": [13, 16, 20, 37, 94, 129, 149], "compat": [10, 91], "compil": [2, 12, 84, 99, 101, 120, 166], "complet": [9, 27, 44, 72, 73, 74], "complex": 19, "compon": [37, 85, 109, 148, 167], "compos": 36, "comprehens": 153, "comput": [10, 11], "concat": 12, "conclus": [8, 130, 133, 134], "config": [23, 85, 119, 167], "configur": [8, 9, 13, 16, 21, 28, 29, 30, 31, 32, 34, 36, 45, 66, 78, 80, 90, 106, 109, 113, 124, 130, 134, 151, 156, 159, 161, 163, 168], "connect": [37, 124], "connector": 59, "consider": 114, "constructor": 34, "consumpt": 117, "contain": [2, 9, 21, 26, 28, 29, 30, 31, 32, 36, 101, 102, 123, 154], "content": [2, 8, 10, 11, 12, 13, 14, 15, 16, 19, 20, 35, 37, 85, 90, 99, 131, 132, 151, 156, 167, 168], "context": [2, 8, 12, 77, 89, 93, 94, 106, 108, 133, 134, 135], "contigu": [77, 93, 108], "continu": 35, "control": [11, 61, 88, 106], "conv": 137, "convers": [19, 118, 122], "cooper": 10, "coordin": [8, 39, 126], "copi": [12, 20], "core": [11, 16, 78, 85, 151, 156, 159, 167], "cot": 11, "coverag": 27, "cp": 94, "cpp": 113, "cpu": [10, 20], "creat": [9, 21, 85, 124, 167], "cross": [77, 88, 108], "cuda": [10, 13, 38, 78, 99, 159], "cuda_graph_config": [26, 28, 29, 30, 31, 32], "cudaev": 1, "cudamemcpyasync": 20, "cudastream": 1, "curl": [42, 43, 44], "current": 99, "custom": [37, 45, 53, 85, 94, 99, 121, 167, 172, 173], "cutlass": 13, "cyclic": [77, 108], "data": [10, 15, 94], "dataset": [2, 8, 16, 17, 19, 22, 40, 41, 127, 128], "datatransceiverst": 0, "datatyp": 88, "deadlock": [10, 20], "debug": [39, 86, 105, 126, 146], "decid": 129, "decod": [9, 10, 14, 19, 23, 54, 69, 88, 97, 98, 106, 116, 142, 157], "decoderst": 1, "decodinginput": 1, "decodingoutput": 1, "decor": 110, "deep": [15, 80, 161], "deepseek": [2, 13, 14, 15, 17, 28, 41, 46], "default": [2, 13, 80, 127, 128, 161], "definit": [35, 120, 151, 156], "dens": [13, 94], "depend": [12, 13], "deploi": 154, "deploy": [28, 29, 30, 31, 32, 33, 83, 100, 165], "deprec": [34, 88], "dequant": 144, "descript": [39, 126], "design": 16, "detail": [34, 113, 144], "dev": 36, "develop": [15, 98, 99, 100, 124, 148], "diagram": 13, "differ": 106, "disabl": [35, 45, 88], "disaggreg": [17, 27, 86, 105, 116], "disaggregated_mpi_work": 27, "disaggserverutil": 0, "distribut": [8, 58], "dive": 15, "do": 153, "docker": [18, 21, 28, 29, 30, 31, 32, 36, 101, 123, 124, 149, 154], "dockerhub": [123, 124], "document": [100, 155], "dora": 113, "dot": [80, 161], "download": [2, 9, 18], "dp": 94, "dq": 144, "draft": [10, 98, 116], "dynamo": [17, 47, 86], "dynasor": 11, "e2": [16, 26, 28, 29, 30, 31, 146], "eagl": [14, 98, 116], "eagle3": [9, 14, 18], "eaglebuff": 1, "eaglemodul": 1, "effect": [16, 19], "embed": [77, 108, 137], "enabl": [2, 19, 39, 88, 94, 107, 112, 123, 126, 130, 134], "encapsul": 37, "end": [12, 20, 26, 28, 29, 30, 31], "endpoint": [27, 28, 29, 30, 31, 32], "engin": [119, 120, 127, 128, 147, 157], "enhanc": 155, "environ": [34, 86, 105], "ep": [15, 16, 94], "eplb": [16, 20, 94], "error": 146, "errorcod": 149, "etp": 13, "eval": 24, "evalu": [14, 16, 28, 29, 30, 31, 119], "event": 111, "everyth": 13, "evolut": 37, "exampl": [11, 24, 39, 47, 50, 53, 66, 75, 78, 79, 91, 106, 113, 119, 120, 121, 126, 127, 149, 159, 160], "except": 142, "exchang": [17, 86], "execut": 146, "executor": [0, 106, 113], "exist": 34, "exp": 12, "expand": 16, "expect": [2, 112], "experi": 8, "experiment": 19, "expert": [12, 13, 15, 16, 20, 28, 80, 94, 107, 161], "explicitdrafttokensbuff": 1, "explor": 2, "extens": 16, "extra": 21, "extra_llm_api_opt": 26, "face": 149, "factor": [77, 108, 119], "fail": 35, "failur": 99, "falcon": 3, "faq": [86, 105, 142], "fast": 35, "faster": 3, "featur": [2, 11, 39, 77, 83, 87, 94, 96, 100, 125, 126, 143, 148, 152, 153, 155, 165], "ffn": 94, "field": 34, "file": [80, 101, 161], "find": 35, "first": [4, 19, 26, 28, 29, 30, 31], "fix": 155, "flag": [134, 144], "flayerinfo": 110, "flight": [77, 93, 106, 108], "flow": [40, 127], "fmha": [12, 77, 108], "format": [2, 85, 90, 113, 167, 168], "formatt": 37, "fp16": [2, 144], "fp32": 144, "fp4": [12, 41, 153], "fp8": [2, 4, 12, 41, 77, 95, 108, 119, 130, 144, 153], "fraction": 135, "framework": 11, "free": 135, "free_gpu_memory_fract": 32, "from": [34, 83, 101, 149, 165], "full": [8, 101], "fulli": 121, "function": [110, 121, 136], "further": 20, "fuse": 12, "fuse_a_gemm": 13, "fusion": [12, 13, 99, 120, 130, 134], "futur": [11, 13, 14, 15, 17, 20, 45], "garbag": [39, 126], "gate": 130, "gb200": [9, 16], "gc": [39, 126], "gemm": [12, 13, 130, 134], "genai": [48, 49], "gener": [16, 45, 54, 55, 56, 57, 58, 61, 65, 77, 86, 97, 99, 105, 108, 149], "get": [2, 9, 78, 83, 100, 159, 165], "gil": [10, 39, 126], "got": 149, "gpqa_diamond": 24, "gpqa_extend": 24, "gpqa_main": 24, "gpt": [9, 21, 29, 109, 113], "gptdecod": 1, "gptdecoderbatch": 1, "gptjsonconfig": 1, "gptq": 144, "gpu": [2, 3, 10, 13, 15, 16, 40, 117, 120, 127, 135, 142, 153], "gqa": 88, "gram": 19, "grammar": 10, "graph": [10, 13, 38, 78, 99, 110, 159], "group": [13, 77, 108], "gsm8k": [16, 24], "guarante": 34, "guid": [10, 28, 29, 30, 31, 32, 33, 34, 54, 94, 97, 98, 99, 100, 106, 131, 148, 151, 154, 156], "h": [0, 1], "h100": [4, 5, 21], "h200": [2, 3, 5, 6, 21], "ha": 4, "hang": 149, "hardwar": [8, 28, 29, 30, 31, 32, 41, 95, 145], "hbm": 5, "head": [12, 77, 108], "header": 101, "health": 9, "heurist": 19, "hierarch": 34, "hierarchi": 111, "high": [16, 21, 110], "highlight": [19, 20], "hopper": [28, 30, 31, 32, 144], "host": [10, 16, 20, 88, 112, 149], "how": [2, 11, 13, 14, 15, 18, 51, 52, 88, 92, 94, 107, 112, 127, 129, 133, 169], "hub": 149, "hug": 149, "huggingfac": [85, 167], "i": [4, 88, 129, 142], "ibuff": 1, "id": 113, "ifb": 93, "igptdecoderbatch": 1, "imag": [18, 36, 101, 102, 123, 124], "implement": [8, 11, 13, 14, 16, 20, 34, 77, 118, 158], "import": 108, "improv": 116, "incorpor": [82, 164], "increas": 6, "indic": 100, "infer": [7, 11, 14, 16, 17, 21, 27, 106, 142, 154], "inform": [39, 110, 126], "infrastructur": 155, "initi": 20, "inplac": 99, "input": [77, 108, 149], "instal": [2, 103, 104, 146], "int4": [3, 144], "int8": [77, 108, 144], "integr": [10, 35, 84, 166], "inter": [20, 26, 28, 29, 30, 31], "interfac": [16, 172], "intern": 109, "interv": 20, "introduct": [11, 15, 16, 26, 28, 29, 30, 31, 32, 37, 50, 149, 151, 156, 172, 173], "invok": 149, "ipcnvlsmemori": 1, "ipcutil": 1, "ir": 99, "isl": [2, 17], "issu": [2, 20, 99, 142, 148, 149, 155], "itensor": 1, "iter": [39, 126], "itl": [26, 28, 29, 30, 31], "jenkin": 35, "json": 74, "json_mode_ev": 24, "k": [12, 93], "k8": 47, "kei": [13, 26, 28, 29, 30, 31, 37, 83, 94, 121, 124, 125, 129, 148, 152, 153, 155, 165], "kernel": [6, 10, 12, 13, 16, 20, 21, 94], "knob": 34, "knowledg": [99, 131, 132], "known": [99, 101, 142, 148, 155], "kv": [2, 17, 37, 51, 59, 60, 77, 86, 88, 93, 95, 98, 108, 111, 112, 119, 130, 135, 142, 172], "kv_cache_config": [26, 28, 30, 31, 32], "kv_cache_free_gpu_memory_fract": [28, 29, 30, 31], "kvcacheeventmanag": 111, "kvcachemanag": 157, "larg": 16, "latenc": [2, 6, 8, 13, 21, 22, 26, 28, 29, 30, 31, 127, 128, 130], "latest": [5, 153], "launch": [9, 12, 13, 18, 21, 26, 28, 29, 30, 31, 32, 39, 86, 126, 154], "layer": [13, 15, 137], "layernorm": 119, "layout": [17, 86, 121], "level": [13, 16, 37, 81, 94, 110, 157, 162], "librari": 37, "licens": [30, 31], "light": 8, "limit": [40, 88, 101, 116, 127, 155, 163], "linear": 137, "link": [95, 100, 101], "linux": [101, 104], "list": 11, "llama": [3, 6, 41, 120, 127, 130, 134], "llama2": 5, "llama3": 30, "llama4": [18, 31], "llm": [2, 4, 5, 7, 9, 10, 11, 12, 14, 16, 17, 18, 19, 20, 21, 28, 29, 30, 31, 32, 34, 40, 45, 50, 53, 58, 62, 80, 83, 91, 94, 95, 99, 100, 101, 107, 116, 119, 121, 122, 123, 124, 127, 128, 133, 142, 145, 147, 149, 153, 154, 155, 161, 165], "llmarg": 34, "lm": 12, "load": [8, 16, 20, 85, 94, 121, 151, 156, 167], "loader": [85, 121, 167], "local": [12, 149], "log": [18, 81, 162], "logic": [16, 37], "logit": [23, 61, 97, 106], "logprob": 76, "long": 89, "longbench_v2": 24, "lookahead": 116, "lookaheadbuff": 1, "lookaheadmodul": 1, "lora": [23, 40, 65, 73, 90, 113, 127, 168], "loracach": [1, 113], "loracachepagemanagerconfig": 1, "loramodul": 1, "low": [12, 21, 37, 90, 114, 127, 130, 168], "lower": 12, "machin": [16, 17], "major": 11, "make": [10, 119], "manag": [20, 34, 40, 90, 110, 111, 127, 168, 172], "map": [113, 127], "mapper": [85, 167], "mark": 106, "marker": [39, 126], "mask": 10, "match": 120, "mathemat": 8, "matrix": [28, 29, 83, 84, 87, 91, 95, 143, 144, 145, 152, 165, 166], "maverick": [18, 41], "max": [2, 21, 127, 133, 135], "max_batch_s": [28, 29, 30, 31, 32, 93], "max_num_token": [28, 29, 30, 31, 32, 93], "max_seq_len": [28, 29, 30, 31, 32, 93], "maximum": 135, "measur": [17, 41], "mechan": 8, "medusa": [116, 127], "medusamodul": 1, "memori": [2, 5, 20, 88, 112, 117, 135, 142], "memorycount": 1, "merg": [35, 80, 161], "method": [7, 34, 110], "methodologi": [17, 26], "metric": [26, 27, 28, 29, 30, 31], "migrat": 20, "min": 2, "miscellan": 16, "mix": 13, "mixtur": [94, 107], "mla": [2, 15], "mlp": [119, 130, 137], "mlperf": 4, "mm_embedding_serv": 27, "mmlu": 24, "mmmu": 24, "modal": [27, 40, 127, 145], "mode": [40, 127], "model": [2, 8, 9, 10, 13, 14, 28, 29, 30, 31, 32, 33, 40, 41, 83, 84, 91, 94, 95, 98, 100, 109, 116, 118, 120, 121, 127, 129, 130, 134, 138, 145, 146, 149, 151, 152, 153, 155, 156, 157, 165, 166], "modelconfig": 1, "modelopt": 95, "modif": 37, "modifi": 34, "modul": [14, 15, 94, 110, 113], "moe": [15, 20, 21, 28, 29, 94, 107], "moe_backend": 13, "moe_config": [26, 28, 29, 30, 31, 32], "moe_expert_parallel_s": [28, 29, 30, 31, 32], "monitor": 18, "more": [2, 6, 12, 39, 126], "motiv": [8, 10, 11, 16, 17, 19, 86, 94, 96], "mount": 36, "mpi_abort": 149, "mpi_comm_world": 149, "mqa": [15, 88], "mtp": [12, 13, 14, 20, 98], "much": 88, "multi": [13, 17, 19, 20, 27, 40, 77, 86, 90, 99, 108, 120, 127, 145, 149, 168], "multimod": [26, 27, 43, 49, 71, 91, 143, 152], "multipl": [65, 134], "multithread": 20, "mutex": 10, "n": 19, "name": [23, 34, 35, 121], "nativ": [12, 121], "nearli": 5, "nemo": [90, 168], "net": 149, "network": [12, 40, 127], "new": [6, 34, 77, 118, 151, 156, 158], "next": [7, 32, 154], "ngc": [18, 26, 102], "ngram": [98, 116], "node": [27, 120, 149], "non": [34, 127], "norm": [130, 134], "normal": 137, "notat": [80, 161], "note": [106, 108, 155], "nsight": [39, 126], "num": 133, "numa": 20, "numer": 144, "nvfp4": 144, "nvidia": [13, 15, 39, 126], "nvtx": [39, 126], "o": 142, "observ": 16, "obtain": 106, "off": 8, "offlin": [16, 95, 154], "offload": [60, 88, 112], "one": 16, "onli": [13, 21, 39, 99, 101, 126, 144, 149], "onlin": [16, 20, 75, 154], "op": 99, "openai": [21, 70, 71, 72, 73, 74, 91], "oper": 99, "optim": [8, 12, 13, 14, 15, 17, 20, 38, 77, 78, 86, 91, 99, 108, 134, 153, 159], "option": [2, 18, 21, 28, 29, 30, 31, 32, 76, 78, 101, 130, 134, 135, 159], "orchestr": 96, "osl": [2, 17], "oss": [9, 21, 29], "other": 127, "out": [2, 151, 156], "outlook": 37, "output": [26, 28, 29, 30, 31, 76, 106, 127], "over": [3, 16, 34], "overhead": 20, "overlap": [12, 17, 38, 86, 92, 169], "overload": 37, "overrid": 36, "overview": [12, 34, 35, 38, 41, 85, 94, 109, 119, 121, 122, 153, 167], "ovewiew": 157, "own": [82, 164, 173], "p": 112, "pack": [77, 108], "pad": [99, 108], "page": [20, 77, 93, 108, 111, 133, 134, 135], "parallel": [12, 13, 15, 16, 20, 28, 94, 107, 113, 127, 129, 134], "paramet": 109, "pareto": 8, "parser": 46, "part": [12, 16, 20, 118], "partial": 88, "pattern": [110, 120], "pdl": 12, "per": [26, 28, 29, 30, 31], "perf": [48, 49], "perform": [2, 4, 7, 8, 10, 12, 13, 16, 17, 18, 19, 20, 21, 28, 29, 30, 31, 32, 39, 78, 99, 112, 114, 116, 126, 128, 130, 131, 134, 153, 159], "persist": [40, 127], "phase": [12, 77, 108], "piecewis": 99, "pip": 104, "pipelin": [35, 94, 129, 134], "pitfal": 128, "plugin": [23, 120, 130, 134, 139], "pod": 124, "polici": [88, 135], "pool": [111, 137, 142], "popular": 33, "posit": [77, 108], "post": [35, 106], "postprocess": 121, "power": [40, 127], "pp": 94, "practic": [7, 34, 35, 94], "pre": [95, 102], "preced": [80, 161], "precis": [12, 13, 15, 84, 114, 144, 166], "predict": 20, "prefer": 34, "prefil": 93, "prepar": [2, 18, 40, 41, 119, 124, 127, 128], "prepare_dataset": 22, "prerequisit": [2, 9, 18, 21, 28, 29, 30, 31, 32, 101, 131, 132, 151, 156], "prevent": [20, 112], "principl": 34, "prioriti": 52, "probe": 11, "process": 20, "processor": [61, 97, 106], "product": 153, "profil": [13, 39, 126, 134], "programmat": [12, 13], "promot": 34, "prompt_logprob": 76, "prompttuningparam": 1, "properti": 88, "propos": 10, "prototyp": [83, 96, 105, 148], "provid": [6, 98], "push": [12, 13, 18], "py": [22, 35, 80, 161], "pyexecutor": 157, "python": [10, 16, 101, 106, 142], "pytorch": [12, 39, 40, 62, 63, 64, 83, 126, 127, 143, 145, 148, 152, 153, 156, 165], "q": [12, 144], "qkv": [77, 108], "quantiz": [7, 12, 40, 45, 90, 95, 119, 122, 127, 130, 140, 144, 168, 170], "quantmod": 144, "queri": [77, 108], "quick": [9, 33, 91, 95, 98, 100, 148, 149, 154, 163], "quickstart": 127, "quit": 149, "qwen": 17, "qwen3": [32, 41], "r1": [2, 13, 14, 15, 17, 28, 41, 46], "rab": 108, "race": 10, "rai": 96, "rank": [90, 119, 149, 168], "ratio": 8, "rawengin": 1, "re": [13, 99], "reason": 46, "receiv": 37, "recip": 33, "recommend": [28, 29, 30, 31, 32, 130, 134, 142], "recompil": 99, "record_signatur": 110, "redraft": 116, "reduc": [20, 117, 130, 134], "reduct": 12, "refer": [11, 78, 92, 94, 100, 118, 150, 159, 169], "regist": 118, "registr": [151, 156], "registri": 18, "reject": 10, "rel": 108, "relat": 110, "relax": [13, 14], "releas": [18, 102, 155], "replic": 94, "repositori": 18, "reproduc": [2, 13, 15, 16, 17, 41], "request": [9, 18, 21, 26, 88, 93, 106], "requir": [110, 114], "resourcemanag": 157, "respons": 106, "result": [2, 8, 39, 41, 106, 126, 128], "retent": 88, "rethink": 12, "retriev": 110, "return_context_logit": 76, "return_generation_logit": 76, "reus": [88, 112], "revisit": [93, 133], "rewind": 98, "rewrit": 110, "right": 7, "roadmap": [83, 96, 165], "robin": 8, "roll": [77, 108], "rope": [77, 108], "rotari": [77, 108], "round": 8, "router": 13, "routergemm": 13, "run": [2, 9, 14, 16, 21, 26, 28, 29, 30, 31, 32, 34, 39, 40, 41, 62, 63, 64, 79, 90, 91, 95, 113, 117, 126, 127, 128, 149, 154, 160, 168], "runpod": 124, "runtim": [1, 10, 15, 37, 38, 45, 66, 84, 101, 109, 113, 120, 135, 141, 142, 166], "runtimedefault": 1, "runtimeerror": 149, "salt": 88, "same": 6, "sampl": [9, 18, 21, 45, 67, 97, 98, 109, 116, 171], "samplingconfig": 1, "save": 128, "scaffold": 11, "scaffoldingllm": 11, "scale": [12, 16, 20, 77, 108, 119], "scatter": 134, "schedul": [8, 38, 92, 93, 133, 135, 157, 169, 173], "schema": [34, 74], "scout": [31, 41], "script": [79, 160], "seamless": [83, 165], "seamlessli": 10, "search": [77, 97, 108], "sec": 5, "second": [26, 28, 29, 30, 31], "secur": 88, "see": 163, "select": [21, 36], "send": 106, "sender": 37, "sequenc": 89, "serial": 0, "serv": [17, 21, 25, 26, 27, 39, 64, 75, 86, 90, 91, 98, 116, 126, 154, 163, 168], "server": [9, 17, 18, 21, 27, 28, 29, 30, 31, 32, 86, 90, 91, 106, 168], "servic": [26, 105], "session": 149, "set": [8, 26, 28, 29, 30, 31, 32, 40, 78, 127, 129, 159], "setup": 19, "sever": 12, "shard": 129, "share": 12, "shoot": 121, "show": [28, 29, 30, 31, 32], "showcas": 67, "singl": [3, 90, 149, 168], "situat": 112, "size": [88, 133, 135, 142], "slide": [77, 89, 108], "slurm": [27, 53, 62, 63, 64, 86, 149], "smart": 13, "smoothquant": 144, "softwar": 145, "sol": 8, "sota": 7, "sourc": 101, "spars": [12, 13, 68], "specif": [33, 39, 126], "specul": [9, 10, 14, 19, 23, 69, 88, 98, 116], "speculativedecodingmod": 1, "speculativedecodingmodul": 1, "speed": [7, 8, 11, 19], "speedup": 14, "ssh": [123, 124], "stabil": 34, "stage": 35, "start": [9, 18, 26, 27, 33, 78, 83, 90, 91, 98, 100, 148, 149, 154, 159, 163, 165, 168], "state": [10, 153], "statist": 16, "statu": [20, 99], "step": [2, 16, 17, 18, 28, 29, 30, 31, 32, 85, 101, 118, 151, 154, 156, 167], "stop": 18, "strategi": [8, 13, 15, 94, 129], "stream": [13, 20, 57, 99, 117], "streamingllm": [77, 108], "structur": [12, 106], "studi": [14, 16, 17, 19, 133], "style": 45, "subcommand": 127, "summari": [8, 127, 130, 134], "support": [12, 14, 17, 20, 28, 29, 40, 83, 84, 86, 90, 91, 95, 101, 120, 121, 127, 143, 144, 145, 152, 153, 165, 166, 168], "swiglu": 130, "syntax": [24, 27], "synthet": 17, "system": [13, 39, 88, 126], "tabl": [2, 8, 10, 11, 12, 13, 14, 15, 16, 19, 20, 35, 37, 85, 90, 99, 100, 131, 132, 151, 156, 167, 168], "tag": [18, 102], "target": [10, 98, 116], "technic": 144, "techniqu": [7, 11, 67], "templat": 124, "tensor": [0, 12, 77, 94, 106, 107, 108, 110, 113, 129, 142], "tensor_parallel_s": [28, 29, 30, 31, 32], "tensorrt": [2, 4, 5, 7, 9, 10, 11, 12, 14, 16, 17, 18, 19, 20, 21, 28, 29, 30, 31, 32, 40, 83, 91, 94, 95, 99, 100, 101, 107, 116, 119, 120, 121, 122, 123, 124, 127, 128, 133, 142, 145, 147, 153, 155, 165], "test": [18, 21, 28, 29, 30, 31, 32, 34, 35, 146], "test_to_stage_map": 35, "text": [54, 55, 56, 57, 61, 65], "theoret": 8, "think": 129, "thought": 16, "thrash": 20, "throughput": [2, 6, 8, 15, 21, 22, 26, 28, 29, 30, 31, 40, 41, 127, 128], "time": [11, 26, 28, 29, 30, 31, 134, 142], "tip": [18, 21, 28, 29, 30, 31, 32, 78, 99, 128, 146, 149, 159, 163], "tlb": 20, "tllmlogger": 1, "tok": 4, "token": [4, 5, 8, 10, 20, 26, 28, 29, 30, 31, 45, 133, 135], "token_norm_dist": 22, "token_unif_dist": 22, "tool": 122, "top": 157, "topic": 101, "topologi": 114, "torch": [12, 99], "torchllmarg": 34, "total": [26, 28, 29, 30, 31], "tp": [8, 26, 28, 29, 30, 31, 94], "tpot": [26, 28, 29, 30, 31], "trace": [10, 99], "trade": 8, "tradeoff": [92, 169], "transceiv": 37, "transfer": 37, "transferag": 0, "transform": [17, 86], "translat": [16, 17, 19, 121], "transmiss": 37, "tree": [14, 116, 151, 156], "trigger": [35, 111], "triton": [17, 21, 106], "troubl": 121, "troubleshoot": [10, 18, 21, 28, 29, 30, 31, 32, 86, 105, 128, 146, 149], "trt": 7, "trtllm": [13, 17, 22, 23, 24, 25, 26, 27, 39, 63, 64, 78, 86, 90, 91, 98, 126, 128, 154, 159, 163, 168], "trtllmattent": 77, "trust_remote_cod": [28, 29, 30, 31, 32], "ttft": [26, 28, 29, 30, 31], "tune": [2, 18, 99, 112, 131, 133], "turn": 19, "two": [10, 98], "type": [0, 34, 111], "understand": [133, 142], "unit": [35, 146], "unnecessari": 35, "up": [3, 6, 7, 11, 19, 26], "updat": [20, 34, 155], "upload": [123, 124], "upper": 37, "us": [12, 18, 21, 35, 36, 61, 85, 110, 113, 116, 134, 135, 142, 149, 167], "usag": [24, 35, 78, 83, 85, 86, 90, 92, 95, 96, 97, 98, 99, 114, 142, 159, 165, 167, 168, 169], "user": [98, 130], "v": [5, 8, 12, 107], "valid": [34, 40, 127], "vanilla": 14, "variabl": [34, 41, 86, 105], "verif": [13, 98], "verifi": [28, 29, 30, 31, 118], "via": [91, 104, 128, 163], "virtualmemori": 1, "visual": [39, 126], "volum": 36, "vote": 11, "w4a16": 144, "w8a16": 144, "w8a8": 144, "wa": 149, "wai": 18, "wait": 8, "waiv": 35, "weekli": 18, "weight": [15, 20, 85, 117, 118, 119, 120, 121, 142, 144, 151, 156, 167], "welcom": 100, "what": [4, 7, 111, 153], "when": [13, 85, 110, 149, 167], "wide": [28, 94], "width": 106, "window": [77, 88, 89, 108, 135], "windowblockmanag": 111, "wip": 2, "within": 6, "without": 101, "wo": 12, "work": [11, 13, 14, 15, 17, 20, 92, 127, 169], "worker": 11, "workflow": [34, 37, 39, 40, 82, 110, 121, 122, 126, 127, 164], "workload": 13, "world": 109, "worldconfig": 1, "write": 118, "xqa": [6, 77, 108], "yaml": [28, 29, 30, 31, 32, 78, 80, 90, 159, 161, 163, 168], "you": [128, 153], "your": [18, 82, 164, 173]}})
\ No newline at end of file
diff --git a/latest/torch/adding_new_model.html b/latest/torch/adding_new_model.html
index ddddb3d692..2bfcc88e1e 100644
--- a/latest/torch/adding_new_model.html
+++ b/latest/torch/adding_new_model.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -766,10 +774,12 @@
+
@@ -781,12 +791,7 @@ For example, the Flashinfer metadata fills Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -834,9 +839,9 @@ For example, the Flashinfer metadata fills
diff --git a/latest/torch/auto_deploy/advanced/benchmarking_with_trtllm_bench.html b/latest/torch/auto_deploy/advanced/benchmarking_with_trtllm_bench.html
index c2e5812e97..b8e41777cc 100644
--- a/latest/torch/auto_deploy/advanced/benchmarking_with_trtllm_bench.html
+++ b/latest/torch/auto_deploy/advanced/benchmarking_with_trtllm_bench.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -697,10 +705,12 @@
+
+
@@ -722,12 +732,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -775,9 +780,9 @@
diff --git a/latest/torch/auto_deploy/advanced/example_run.html b/latest/torch/auto_deploy/advanced/example_run.html
index 88bfad4391..df704ddd10 100644
--- a/latest/torch/auto_deploy/advanced/example_run.html
+++ b/latest/torch/auto_deploy/advanced/example_run.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -613,10 +621,12 @@ python build_and_run_ad.py
+
+
@@ -638,12 +648,7 @@ python build_and_run_ad.py Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -691,9 +696,9 @@ python build_and_run_ad.py
diff --git a/latest/torch/auto_deploy/advanced/expert_configurations.html b/latest/torch/auto_deploy/advanced/expert_configurations.html
index f9850fcbde..8db50417e1 100644
--- a/latest/torch/auto_deploy/advanced/expert_configurations.html
+++ b/latest/torch/auto_deploy/advanced/expert_configurations.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -720,10 +728,12 @@ python build_and_run_ad.py
+
+
@@ -745,12 +755,7 @@ python build_and_run_ad.py Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -798,9 +803,9 @@ python build_and_run_ad.py
diff --git a/latest/torch/auto_deploy/advanced/logging.html b/latest/torch/auto_deploy/advanced/logging.html
index 7e0a37c714..9ae181034b 100644
--- a/latest/torch/auto_deploy/advanced/logging.html
+++ b/latest/torch/auto_deploy/advanced/logging.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -546,10 +554,12 @@ decreasing verbosity;
+
+
@@ -571,12 +581,7 @@ decreasing verbosity;
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -624,9 +629,9 @@ decreasing verbosity;
diff --git a/latest/torch/auto_deploy/advanced/serving_with_trtllm_serve.html b/latest/torch/auto_deploy/advanced/serving_with_trtllm_serve.html
index 57ec734557..bfb40ebf64 100644
--- a/latest/torch/auto_deploy/advanced/serving_with_trtllm_serve.html
+++ b/latest/torch/auto_deploy/advanced/serving_with_trtllm_serve.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -644,10 +652,12 @@
+
+
@@ -669,12 +679,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -722,9 +727,9 @@
diff --git a/latest/torch/auto_deploy/advanced/workflow.html b/latest/torch/auto_deploy/advanced/workflow.html
index 040c965e03..b6eae2ecf1 100644
--- a/latest/torch/auto_deploy/advanced/workflow.html
+++ b/latest/torch/auto_deploy/advanced/workflow.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -564,10 +572,12 @@
+
+
@@ -589,12 +599,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -642,9 +647,9 @@
diff --git a/latest/torch/auto_deploy/auto-deploy.html b/latest/torch/auto_deploy/auto-deploy.html
index 7e5c3803a2..65bdd5ce80 100644
--- a/latest/torch/auto_deploy/auto-deploy.html
+++ b/latest/torch/auto_deploy/auto-deploy.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -637,10 +645,12 @@ We welcome community contributions, see
+
@@ -662,12 +672,7 @@ We welcome community contributions, see Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -715,9 +720,9 @@ We welcome community contributions, see
diff --git a/latest/torch/auto_deploy/support_matrix.html b/latest/torch/auto_deploy/support_matrix.html
index e763fa1b2d..dc1d5ce03e 100644
--- a/latest/torch/auto_deploy/support_matrix.html
+++ b/latest/torch/auto_deploy/support_matrix.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -719,10 +727,12 @@ In addition, the following models have been officially validated using the defau
+
+
@@ -744,12 +754,7 @@ In addition, the following models have been officially validated using the defau
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -797,9 +802,9 @@ In addition, the following models have been officially validated using the defau
diff --git a/latest/torch/features/checkpoint_loading.html b/latest/torch/features/checkpoint_loading.html
index 7ce455785d..a9f2e4e964 100644
--- a/latest/torch/features/checkpoint_loading.html
+++ b/latest/torch/features/checkpoint_loading.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -904,10 +912,12 @@ If the format shares components with an existing framework (such as HF), you onl
+
+
@@ -929,12 +939,7 @@ If the format shares components with an existing framework (such as HF), you onl
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -982,9 +987,9 @@ If the format shares components with an existing framework (such as HF), you onl
diff --git a/latest/torch/features/lora.html b/latest/torch/features/lora.html
index 6879260eda..df48a2455d 100644
--- a/latest/torch/features/lora.html
+++ b/latest/torch/features/lora.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -820,10 +828,12 @@
+
+
@@ -845,12 +855,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -898,9 +903,9 @@
diff --git a/latest/torch/features/overlap_scheduler.html b/latest/torch/features/overlap_scheduler.html
index 0af3632f9d..11dd4ed3e4 100644
--- a/latest/torch/features/overlap_scheduler.html
+++ b/latest/torch/features/overlap_scheduler.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -583,10 +591,12 @@
+
+
@@ -608,12 +618,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -661,9 +666,9 @@
diff --git a/latest/torch/features/quantization.html b/latest/torch/features/quantization.html
index 1166e1b46e..883eca2168 100644
--- a/latest/torch/features/quantization.html
+++ b/latest/torch/features/quantization.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -549,10 +557,12 @@ scripts/huggingface_example.sh --model
+
+
@@ -574,12 +584,7 @@ scripts/huggingface_example.sh --model
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -627,9 +632,9 @@ scripts/huggingface_example.sh --model
diff --git a/latest/torch/features/sampling.html b/latest/torch/features/sampling.html
index 9afcbb40f1..93213bb9c4 100644
--- a/latest/torch/features/sampling.html
+++ b/latest/torch/features/sampling.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -550,10 +558,12 @@
+
+
@@ -575,12 +585,7 @@
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -628,9 +633,9 @@
diff --git a/latest/torch/kv_cache_manager.html b/latest/torch/kv_cache_manager.html
index 8b3ef94e47..c75705ce42 100644
--- a/latest/torch/kv_cache_manager.html
+++ b/latest/torch/kv_cache_manager.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -625,10 +633,12 @@ Then, test it to ensure the
+
+
@@ -650,12 +660,7 @@ Then, test it to ensure the Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -703,9 +708,9 @@ Then, test it to ensure the
diff --git a/latest/torch/scheduler.html b/latest/torch/scheduler.html
index 3b2a16ae14..4469d3e461 100644
--- a/latest/torch/scheduler.html
+++ b/latest/torch/scheduler.html
@@ -32,7 +32,7 @@
-
+
@@ -44,6 +44,8 @@
+
+
@@ -59,20 +61,24 @@
+
+
+
-
+
+
@@ -80,6 +86,8 @@
+
+
@@ -650,10 +658,12 @@ In the create_pytor
+
+
@@ -675,12 +685,7 @@ In the create_pytor
- Manage My Privacy
- |
-
-
-
- Do Not Sell or Share My Data
+ Your Privacy Choices
|
@@ -728,9 +733,9 @@ In the create_pytor