From 069ad68d3c8316f807d3dfd0c58350fa3dcf331e Mon Sep 17 00:00:00 2001 From: Frida Hou <201670829+Fridah-nv@users.noreply.github.com> Date: Fri, 16 Jan 2026 13:24:37 -0800 Subject: [PATCH] [None][fix] AutoDeploy: skip mxfp4_moe test unless on Hopper (#10729) Signed-off-by: Fridah-nv <201670829+Fridah-nv@users.noreply.github.com> --- .../unit/multigpu/custom_ops/test_mxfp4_moe_ep.py | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/tests/unittest/_torch/auto_deploy/unit/multigpu/custom_ops/test_mxfp4_moe_ep.py b/tests/unittest/_torch/auto_deploy/unit/multigpu/custom_ops/test_mxfp4_moe_ep.py index 211053a299..a3e9b3254c 100644 --- a/tests/unittest/_torch/auto_deploy/unit/multigpu/custom_ops/test_mxfp4_moe_ep.py +++ b/tests/unittest/_torch/auto_deploy/unit/multigpu/custom_ops/test_mxfp4_moe_ep.py @@ -4,6 +4,7 @@ import pytest import torch import torch.distributed as dist from _dist_test_utils import get_device_counts +from utils.util import getSMVersion from tensorrt_llm._torch.auto_deploy.custom_ops.fused_moe.mxfp4_moe import ( IS_TRITON_KERNELS_AVAILABLE, @@ -109,6 +110,10 @@ def _run_mxfp4_mlp_ep_dtype_test(num_experts: int, topk: int, rank: int, world_s torch.testing.assert_close(part_out, ref_out, rtol=5e-2, atol=5e-2, equal_nan=True) +@pytest.mark.skipif( + getSMVersion() != 90, + reason="triton_mxfp4_moe is only supported in Hopper architecture", +) @pytest.mark.skipif( not IS_TRITON_KERNELS_AVAILABLE, reason="triton_kernels unavailable",