mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-23 12:12:39 +08:00
remove waivers and cleanup
Signed-off-by: Xiwen Yu <13230610+VALLIS-NERIA@users.noreply.github.com>
This commit is contained in:
parent
77657de972
commit
d42201e235
@ -1,49 +0,0 @@
|
||||
# ARCH=$(uname -m)
|
||||
# if [ $ARCH == "x86_64" ]; then
|
||||
|
||||
# wget https://urm.nvidia.com/artifactory/sw-gpu-cuda-installer-generic-local/packaging/r13.0/cuda_nvrtc/linux-x86_64/13.0.48/cuda-nvrtc-dev-13-0_13.0.48-1_amd64.deb && \
|
||||
# dpkg -i cuda-nvrtc-dev-13-0_13.0.48-1_amd64.deb && \
|
||||
# rm cuda-nvrtc-dev-13-0_13.0.48-1_amd64.deb
|
||||
|
||||
# wget https://github.com/Kitware/CMake/releases/download/v4.0.3/cmake-4.0.3-linux-x86_64.sh && \
|
||||
# bash cmake-4.0.3-linux-x86_64.sh --skip-license --prefix=/usr/local/cmake --exclude-subdir
|
||||
|
||||
# apt update
|
||||
# apt install -y libstdc++-14-dev
|
||||
|
||||
# elif [ $ARCH == "aarch64" ]; then
|
||||
|
||||
# # to be moved to docker/common/ scripts
|
||||
# wget https://urm.nvidia.com/artifactory/sw-gpu-cuda-installer-generic-local/packaging/r13.0/cuda_nvrtc/linux-sbsa/13.0.48/cuda-nvrtc-dev-13-0_13.0.48-1_arm64.deb && \
|
||||
# dpkg -i cuda-nvrtc-dev-13-0_13.0.48-1_arm64.deb && \
|
||||
# rm cuda-nvrtc-dev-13-0_13.0.48-1_arm64.deb
|
||||
|
||||
# wget https://github.com/Kitware/CMake/releases/download/v4.0.3/cmake-4.0.3-linux-aarch64.sh && \
|
||||
# bash cmake-4.0.3-linux-aarch64.sh --skip-license --prefix=/usr/local/cmake --exclude-subdir
|
||||
|
||||
# apt update
|
||||
# # fix LLVM build
|
||||
# apt install -y libstdc++-14-dev
|
||||
|
||||
# # wait for https://github.com/NVIDIA/TensorRT-LLM/pull/6588
|
||||
# pip install deep_gemm@git+https://github.com/VALLIS-NERIA/DeepGEMM.git@97d97a20c2ecd53a248ab64242219d780cf822b8 --no-build-isolation
|
||||
|
||||
# else
|
||||
# echo "Unsupported architecture: $ARCH"
|
||||
# exit 1
|
||||
# fi
|
||||
|
||||
# # wait for new triton to be published
|
||||
# cd /usr/local/lib/python3.12/dist-packages/ && \
|
||||
# ls -la | grep pytorch_triton && \
|
||||
# mv pytorch_triton-3.3.1+gitc8757738.dist-info triton-3.3.1+gitc8757738.dist-info && \
|
||||
# cd triton-3.3.1+gitc8757738.dist-info && \
|
||||
# echo "Current directory: $(pwd)" && \
|
||||
# echo "Files in directory:" && \
|
||||
# ls -la && \
|
||||
# sed -i 's/^Name: pytorch-triton/Name: triton/' METADATA && \
|
||||
# sed -i 's|pytorch_triton-3.3.1+gitc8757738.dist-info/|triton-3.3.1+gitc8757738.dist-info/|g' RECORD && \
|
||||
# echo "METADATA after update:" && \
|
||||
# grep "^Name:" METADATA
|
||||
|
||||
# # pip install git+https://github.com/triton-lang/triton.git@main
|
||||
@ -75,9 +75,6 @@ RUN pip3 uninstall -y opencv && \
|
||||
rm -rf /usr/local/lib/python3*/dist-packages/cv2/ && \
|
||||
pip3 install opencv-python-headless --force-reinstall --no-deps --no-cache-dir
|
||||
|
||||
# COPY bringup_fix.sh bringup_fix.sh
|
||||
# RUN bash ./bringup_fix.sh && rm bringup_fix.sh
|
||||
|
||||
# WARs against security issues inherited from pytorch:25.06
|
||||
# * https://github.com/advisories/GHSA-8qvm-5x2c-j2w7
|
||||
RUN pip3 install --upgrade --no-cache-dir \
|
||||
|
||||
@ -372,9 +372,9 @@ def preparation(pipeline, testFilter, globalVars)
|
||||
stage("Setup Environment") {
|
||||
setupPipelineEnvironment(pipeline, testFilter, globalVars)
|
||||
}
|
||||
// stage("Merge Test Waive List") {
|
||||
// mergeWaiveList(pipeline, globalVars)
|
||||
// }
|
||||
stage("Merge Test Waive List") {
|
||||
mergeWaiveList(pipeline, globalVars)
|
||||
}
|
||||
})
|
||||
}
|
||||
|
||||
|
||||
@ -41,7 +41,6 @@ def _check_ad_config(experiment_config: ExperimentConfig, llm_args: LlmArgs):
|
||||
)
|
||||
|
||||
|
||||
@pytest.mark.skip("https://nvbugswb.nvidia.com/NVBugs5/redir.aspx?url=/5443039")
|
||||
@pytest.mark.parametrize(
|
||||
"experiment_config",
|
||||
[
|
||||
|
||||
@ -799,7 +799,6 @@ def trtllm_bench_unified_comparison(
|
||||
)
|
||||
|
||||
|
||||
@pytest.mark.skip("https://nvbugswb.nvidia.com/NVBugs5/redir.aspx?url=/5443039")
|
||||
def test_trtllm_bench(llm_root): # noqa: F811
|
||||
model_path_or_name, model_name, model_path = tiny_llama_details()
|
||||
|
||||
|
||||
@ -184,7 +184,6 @@ class RoPEModel(torch.nn.Module):
|
||||
],
|
||||
)
|
||||
@torch.inference_mode()
|
||||
@pytest.mark.skip("https://nvbugswb.nvidia.com/NVBugs5/redir.aspx?url=/5443039")
|
||||
def test_rope_variants(
|
||||
transformation,
|
||||
variant,
|
||||
|
||||
@ -2,7 +2,6 @@ import re
|
||||
import unittest
|
||||
from copy import deepcopy
|
||||
|
||||
import pytest
|
||||
import torch
|
||||
from _torch.helpers import create_mock_engine
|
||||
from parameterized import parameterized
|
||||
@ -10,7 +9,6 @@ from test_modeling_llama import Scenario, reduce_llama_config
|
||||
from transformers import MllamaConfig
|
||||
from transformers import \
|
||||
MllamaForConditionalGeneration as HFMllamaForConditionalGeneration
|
||||
from utils.util import getSMVersion
|
||||
|
||||
import tensorrt_llm
|
||||
from tensorrt_llm._torch.attention_backend.utils import get_attention_backend
|
||||
@ -291,8 +289,6 @@ class TestMLlama(unittest.TestCase):
|
||||
"""
|
||||
Compare output to HF
|
||||
"""
|
||||
if scenario.backend == "FLASHINFER" and getSMVersion() == 120:
|
||||
pytest.skip("https://nvbugspro.nvidia.com/bug/5458945")
|
||||
backend = scenario.backend
|
||||
metadata_cls = get_attention_backend(backend).Metadata
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user