From 6fd765f89e589d0fc0c30c94d848a4c3db134ff4 Mon Sep 17 00:00:00 2001 From: Jiagan Cheng Date: Wed, 27 Aug 2025 20:25:11 -0700 Subject: [PATCH] [None][fix] fix trtllm moe backend error when running gptoss on b300 Signed-off-by: Jiagan Cheng --- cpp/tensorrt_llm/thop/mxFp4BlockScaleMoe.cpp | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/cpp/tensorrt_llm/thop/mxFp4BlockScaleMoe.cpp b/cpp/tensorrt_llm/thop/mxFp4BlockScaleMoe.cpp index 39459c8da4..8c959d054e 100644 --- a/cpp/tensorrt_llm/thop/mxFp4BlockScaleMoe.cpp +++ b/cpp/tensorrt_llm/thop/mxFp4BlockScaleMoe.cpp @@ -46,8 +46,8 @@ torch::Tensor dtype_mxe2m1_block_scale_moe_runner(torch::Tensor const& routing_l std::optional const routed_scaling_factor, int64_t const tile_tokens_dim, int64_t const routing_method_type, btg::Dtype const dtype, MoeRunnerType& moe_runner, int64_t moeConfigIndex) { - auto const sm = tensorrt_llm::common::getSMVersion(); - TORCH_CHECK(sm == 100, "Only SM100 is supported by FP4 block scale MOE"); + auto const sm = tensorrt_llm::common::getSMFamily(); + TORCH_CHECK(sm == 100, "Only SM100f is supported by MXFP4 block scale MOE"); TORCH_CHECK(tile_tokens_dim == 8 || tile_tokens_dim == 16 || tile_tokens_dim == 32 || tile_tokens_dim == 64, "tile_tokens_dim must be 8, 16, 32, 64"); TORCH_CHECK(routing_logits.scalar_type() == at::ScalarType::Float