mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
Fix rotary_emb param in NemotronH attention (#3646)
Signed-off-by: Luis Vega <vegaluisjose@users.noreply.github.com>
This commit is contained in:
parent
950cadf2bd
commit
a06bff5052
@ -77,7 +77,6 @@ class TransformerLayer(Attention):
|
||||
num_key_value_heads=config.num_key_value_heads,
|
||||
max_position_embeddings=config.max_position_embeddings,
|
||||
bias=config.attention_bias,
|
||||
rotary_emb=None,
|
||||
pos_embd_params=None,
|
||||
layer_idx=layer_idx,
|
||||
dtype=config.torch_dtype,
|
||||
|
||||
Loading…
Reference in New Issue
Block a user