Skip to content

Commit eb27499

Browse files
committed
cherry pick #1755 from v0.9.1-dev
Signed-off-by: wangli <wangli858794774@gmail.com>
1 parent 7613ae3 commit eb27499

File tree

1 file changed

+7
-8
lines changed

1 file changed

+7
-8
lines changed

vllm_ascend/ops/fused_moe.py

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1384,16 +1384,15 @@ def forward(self,
13841384
if isinstance(e_hidden_states, tuple):
13851385
e_hidden_states, shared_hidden_states = e_hidden_states
13861386

1387-
if (tp_size > 1 and fused_moe_state not in [
1388-
FusedMoEState.AllGather, FusedMoEState.AllGatherEP,
1389-
FusedMoEState.NaiveMulticast
1390-
] and not replace_allreduce):
1391-
dist.all_gather(list(chunk_hidden_states), e_hidden_states,
1392-
self.tp_group)
1393-
final_hidden_states = torch.cat(chunk_hidden_states, dim=0)
1387+
if fused_moe_state != FusedMoEState.AllGather:
1388+
if tp_size > 1:
1389+
dist.all_gather(list(chunk_hidden_states), e_hidden_states,
1390+
self.tp_group)
1391+
final_hidden_states = torch.cat(chunk_hidden_states, dim=0)
1392+
else:
1393+
final_hidden_states = e_hidden_states
13941394
if num_tokens < tp_size:
13951395
final_hidden_states = final_hidden_states[:num_tokens]
1396-
dispose_tensor(e_hidden_states)
13971396
elif self.dp_size > 1:
13981397
if fused_moe_state == FusedMoEState.NaiveMulticast:
13991398
start = 0 if self.dp_rank == 0 else cu_tokens_across_dp_cpu[

0 commit comments

Comments
 (0)