We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 61e879e commit d2a792eCopy full SHA for d2a792e
vllm_ascend/ops/common_fused_moe.py
@@ -452,7 +452,6 @@ def forward(
452
flashcomm_v1_enabled = forward_context.flashcomm_v1_enabled
453
if flashcomm_v1_enabled:
454
hidden_states = torch.ops.vllm.maybe_all_gather_and_maybe_unpad(hidden_states, True)
455
- router_logits = torch.ops.vllm.maybe_all_gather_and_maybe_unpad(router_logits, True)
456
shared_out = self._shared_experts(hidden_states)
457
458
# NOTE: This is exactly the opposite of `maybe_all_reduce_tensor_model_parallel`
0 commit comments