Skip to content

Conversation

@Meihan-chen
Copy link
Contributor

@Meihan-chen Meihan-chen commented Nov 5, 2025

What this PR does / why we need it?

Add Qwen3-Omni-30B-A3B-Thinking Tutorials

Does this PR introduce any user-facing change?

No

How was this patch tested?

@github-actions
Copy link

github-actions bot commented Nov 5, 2025

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

@github-actions github-actions bot added the documentation Improvements or additions to documentation label Nov 5, 2025
@Meihan-chen Meihan-chen changed the title [DOC] Add Qwen3-Omni-30B-A3B-Thinking Tutorials [Doc] Add Qwen3-Omni-30B-A3B-Thinking Tutorials Nov 5, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This PR adds a new tutorial for running Qwen3-Omni-30B-A3B-Thinking on multiple NPUs. The documentation is comprehensive, but I've found a few critical issues in the provided code snippets that would prevent users from successfully running the examples. Specifically, there's an incorrect package name in a pip install command and an incorrect model path in the offline inference script. There is also a typo in the filename of the new document. Please address these issues to ensure the tutorial is accurate and easy to follow.


```bash
# If you already have transformers installed, please update transformer version >= 4.57.0.dev0
# pip install transformer -U
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The package name for the Hugging Face transformers library is transformers (plural), not transformer. The current command will fail because the package transformer does not exist on PyPI.

Suggested change
# pip install transformer -U
# pip install transformers -U



def main():
MODEL_PATH = "/Qwen/Qwen3-Omni-30B-A3B-Thinking"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The MODEL_PATH is set to an absolute path "/Qwen/Qwen3-Omni-30B-A3B-Thinking". This is inconsistent with other tutorials and the online inference command in this same file, which use the model identifier directly. Using an absolute path might cause model loading to fail if the model is not present at that exact location in the container. It's better to use the model identifier and let vLLM handle the download and caching, especially since VLLM_USE_MODELSCOPE=True is set.

Suggested change
MODEL_PATH = "/Qwen/Qwen3-Omni-30B-A3B-Thinking"
MODEL_PATH = "Qwen/Qwen3-Omni-30B-A3B-Thinking"

@@ -0,0 +1,192 @@
# Multi-NPU (Qwen3-Omni-30B-A3B-Thinking)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

There's a typo in the filename. It should be multi_npu_qwen3_omni_30B_A3B_Thinking.md instead of multi_npu_qwen3_omni_30B_A3B_Thingking.md. Please correct the filename for consistency and clarity.

@wangxiyuan
Copy link
Collaborator

please following the change 5f08e07

We're working on tutorial refactor now.

@Meihan-chen Meihan-chen force-pushed the qwen3-omni-doc branch 4 times, most recently from 7b0e7ff to 6824773 Compare November 5, 2025 01:30
Signed-off-by: Meihan-chen <jcccx.cmh@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants