mirror of
https://github.com/NVIDIA/TensorRT-LLM.git
synced 2026-01-14 06:27:45 +08:00
[Qwen3] chore: fix bug of fused_moe on tp > 1 (#4093)
* fix bug of fused_moe on tp > 1 Signed-off-by: bhsueh <11360707+byshiue@users.noreply.github.com> * refine codes Signed-off-by: bhsueh <11360707+byshiue@users.noreply.github.com> --------- Signed-off-by: bhsueh <11360707+byshiue@users.noreply.github.com>
This commit is contained in:
parent
c28b90984f
commit
f670a036df
@ -821,10 +821,7 @@ class FusedMoE(nn.Module):
|
||||
final_hidden_states = final_hidden_states[0]
|
||||
|
||||
if not self.enable_alltoall:
|
||||
if self.reduce_results and self.parallel_size > 1:
|
||||
return self.all_reduce(final_hidden_states)
|
||||
else:
|
||||
return final_hidden_states
|
||||
return final_hidden_states
|
||||
else:
|
||||
return self.alltoall_combine(final_hidden_states, alltoall_info,
|
||||
token_count)
|
||||
|
||||
Loading…
Reference in New Issue
Block a user