Skip to content

Commit 7bdc6f4

Browse files
authored
fix glm all_reduce tp group (#4188)
1 parent bba279c commit 7bdc6f4

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

fastdeploy/model_executor/models/glm4_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -160,7 +160,7 @@ def forward(self, x):
160160
out = out + shared_experts_out
161161
# We do to TP all reduce after the sum of experts.
162162
if self.tensor_parallel_size > 1:
163-
tensor_model_parallel_all_reduce(out)
163+
tensor_model_parallel_all_reduce(out, self.tp_group)
164164
return out
165165

166166

0 commit comments

Comments
 (0)