Skip to content

Conversation

wangxiyuan
Copy link
Collaborator

@wangxiyuan wangxiyuan commented Sep 14, 2025

  • Fix world size bug in model_runner to make sure ep>16 runs with MC2
  • enable e2e test for vl

Co-Authored-By: whx-sjtu 2952154980@qq.com
Co-Authored-By: Icey 1790571317@qq.com

Copy link

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

@wangxiyuan
Copy link
Collaborator Author

rebase #2897 to test with the newest CI

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request addresses a bug in the selection of the MoE communication method within the model runner, specifically for scenarios involving expert parallelism on Ascend A2 hardware. The change correctly replaces world_size with world_size_across_dp in the condition to enable the 'mc2' communication method. This ensures that the decision is based on the world size relevant to expert parallelism, not the total world size which might include data parallelism dimensions. The accompanying test has been updated to reflect this logic change. The fix appears correct and is well-targeted to resolve the issue described.

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
@wangxiyuan
Copy link
Collaborator Author

also merged this one #2886

@wangxiyuan wangxiyuan added ready read for review ready-for-test start test by label for PR labels Sep 14, 2025
@wangxiyuan wangxiyuan merged commit 382c29f into vllm-project:main Sep 14, 2025
44 checks passed
offline893 pushed a commit to offline893/vllm-ascend that referenced this pull request Sep 16, 2025
- Fix world size bug in model_runner to make sure ep>16 runs with MC2
- enable e2e test for vl

Co-Authored-By: whx-sjtu <2952154980@qq.com>
Co-Authored-By: Icey <1790571317@qq.com>
- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@3e903b6

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: offline0806 <z00858301@china.huawei.com>
chopper0126 pushed a commit to chopper0126/vllm-ascend that referenced this pull request Sep 26, 2025
- Fix world size bug in model_runner to make sure ep>16 runs with MC2 
- enable e2e test for vl

Co-Authored-By: whx-sjtu <2952154980@qq.com>
Co-Authored-By: Icey <1790571317@qq.com>
- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@3e903b6

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
module:tests ready read for review ready-for-test start test by label for PR
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants