remove waivers and cleanup

Signed-off-by: Xiwen Yu <13230610+VALLIS-NERIA@users.noreply.github.com>
This commit is contained in:
Xiwen Yu 2025-09-08 10:24:52 +08:00
parent 77657de972
commit d42201e235
7 changed files with 3 additions and 62 deletions

View File

@ -1,49 +0,0 @@
# ARCH=$(uname -m)
# if [ $ARCH == "x86_64" ]; then
# wget https://urm.nvidia.com/artifactory/sw-gpu-cuda-installer-generic-local/packaging/r13.0/cuda_nvrtc/linux-x86_64/13.0.48/cuda-nvrtc-dev-13-0_13.0.48-1_amd64.deb && \
# dpkg -i cuda-nvrtc-dev-13-0_13.0.48-1_amd64.deb && \
# rm cuda-nvrtc-dev-13-0_13.0.48-1_amd64.deb
# wget https://github.com/Kitware/CMake/releases/download/v4.0.3/cmake-4.0.3-linux-x86_64.sh && \
# bash cmake-4.0.3-linux-x86_64.sh --skip-license --prefix=/usr/local/cmake --exclude-subdir
# apt update
# apt install -y libstdc++-14-dev
# elif [ $ARCH == "aarch64" ]; then
# # to be moved to docker/common/ scripts
# wget https://urm.nvidia.com/artifactory/sw-gpu-cuda-installer-generic-local/packaging/r13.0/cuda_nvrtc/linux-sbsa/13.0.48/cuda-nvrtc-dev-13-0_13.0.48-1_arm64.deb && \
# dpkg -i cuda-nvrtc-dev-13-0_13.0.48-1_arm64.deb && \
# rm cuda-nvrtc-dev-13-0_13.0.48-1_arm64.deb
# wget https://github.com/Kitware/CMake/releases/download/v4.0.3/cmake-4.0.3-linux-aarch64.sh && \
# bash cmake-4.0.3-linux-aarch64.sh --skip-license --prefix=/usr/local/cmake --exclude-subdir
# apt update
# # fix LLVM build
# apt install -y libstdc++-14-dev
# # wait for https://github.com/NVIDIA/TensorRT-LLM/pull/6588
# pip install deep_gemm@git+https://github.com/VALLIS-NERIA/DeepGEMM.git@97d97a20c2ecd53a248ab64242219d780cf822b8 --no-build-isolation
# else
# echo "Unsupported architecture: $ARCH"
# exit 1
# fi
# # wait for new triton to be published
# cd /usr/local/lib/python3.12/dist-packages/ && \
# ls -la | grep pytorch_triton && \
# mv pytorch_triton-3.3.1+gitc8757738.dist-info triton-3.3.1+gitc8757738.dist-info && \
# cd triton-3.3.1+gitc8757738.dist-info && \
# echo "Current directory: $(pwd)" && \
# echo "Files in directory:" && \
# ls -la && \
# sed -i 's/^Name: pytorch-triton/Name: triton/' METADATA && \
# sed -i 's|pytorch_triton-3.3.1+gitc8757738.dist-info/|triton-3.3.1+gitc8757738.dist-info/|g' RECORD && \
# echo "METADATA after update:" && \
# grep "^Name:" METADATA
# # pip install git+https://github.com/triton-lang/triton.git@main

View File

@ -75,9 +75,6 @@ RUN pip3 uninstall -y opencv && \
rm -rf /usr/local/lib/python3*/dist-packages/cv2/ && \
pip3 install opencv-python-headless --force-reinstall --no-deps --no-cache-dir
# COPY bringup_fix.sh bringup_fix.sh
# RUN bash ./bringup_fix.sh && rm bringup_fix.sh
# WARs against security issues inherited from pytorch:25.06
# * https://github.com/advisories/GHSA-8qvm-5x2c-j2w7
RUN pip3 install --upgrade --no-cache-dir \

View File

@ -372,9 +372,9 @@ def preparation(pipeline, testFilter, globalVars)
stage("Setup Environment") {
setupPipelineEnvironment(pipeline, testFilter, globalVars)
}
// stage("Merge Test Waive List") {
// mergeWaiveList(pipeline, globalVars)
// }
stage("Merge Test Waive List") {
mergeWaiveList(pipeline, globalVars)
}
})
}

View File

@ -41,7 +41,6 @@ def _check_ad_config(experiment_config: ExperimentConfig, llm_args: LlmArgs):
)
@pytest.mark.skip("https://nvbugswb.nvidia.com/NVBugs5/redir.aspx?url=/5443039")
@pytest.mark.parametrize(
"experiment_config",
[

View File

@ -799,7 +799,6 @@ def trtllm_bench_unified_comparison(
)
@pytest.mark.skip("https://nvbugswb.nvidia.com/NVBugs5/redir.aspx?url=/5443039")
def test_trtllm_bench(llm_root): # noqa: F811
model_path_or_name, model_name, model_path = tiny_llama_details()

View File

@ -184,7 +184,6 @@ class RoPEModel(torch.nn.Module):
],
)
@torch.inference_mode()
@pytest.mark.skip("https://nvbugswb.nvidia.com/NVBugs5/redir.aspx?url=/5443039")
def test_rope_variants(
transformation,
variant,

View File

@ -2,7 +2,6 @@ import re
import unittest
from copy import deepcopy
import pytest
import torch
from _torch.helpers import create_mock_engine
from parameterized import parameterized
@ -10,7 +9,6 @@ from test_modeling_llama import Scenario, reduce_llama_config
from transformers import MllamaConfig
from transformers import \
MllamaForConditionalGeneration as HFMllamaForConditionalGeneration
from utils.util import getSMVersion
import tensorrt_llm
from tensorrt_llm._torch.attention_backend.utils import get_attention_backend
@ -291,8 +289,6 @@ class TestMLlama(unittest.TestCase):
"""
Compare output to HF
"""
if scenario.backend == "FLASHINFER" and getSMVersion() == 120:
pytest.skip("https://nvbugspro.nvidia.com/bug/5458945")
backend = scenario.backend
metadata_cls = get_attention_backend(backend).Metadata