mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
[https://nvbugs/5501820][fix] Add requirements for numba-cuda version to WAR mem corruption (#7992) (#8414)
Signed-off-by: Pengbo Wang <221450789+pengbowang-nv@users.noreply.github.com>
This commit is contained in:
parent
14d0f5d683
commit
8ce2dc5cb7
@ -76,3 +76,5 @@ blobfile
|
||||
openai-harmony==0.0.4
|
||||
nvidia-cutlass-dsl==4.2.1; python_version >= "3.10"
|
||||
numexpr<2.14.0 # WAR for attempted use of nonexistent numpy.typing
|
||||
# numba-cuda # <For CUDA 12.9>
|
||||
numba-cuda>=0.19.0 # WAR for nvbugs/5501820
|
||||
|
||||
@ -22,8 +22,6 @@ MPI.pickle.__init__(
|
||||
|
||||
# needed since we reuse the mpi executor pool, first test running will leak a thread
|
||||
pytestmark = pytest.mark.threadleak(enabled=False)
|
||||
pytest.skip(reason="https://nvbugspro.nvidia.com/bug/5501820",
|
||||
allow_module_level=True)
|
||||
|
||||
|
||||
def rms_norm(x: torch.Tensor, weight: torch.Tensor = None, eps: float = 1e-6):
|
||||
|
||||
Loading…
Reference in New Issue
Block a user