Skip to content

Conversation

sunbaosong
Copy link
Contributor

@sunbaosong sunbaosong commented Apr 29, 2025

What this PR does / why we need it?

Optimize NPU memory usage. #723

vllm v0.8.4.rc2 and DeepSeek R1 can only support a model length of 16K. When attempting to run with a model length of 32K, an "Out of Memory" (OOM) error will occur.

Does this PR introduce any user-facing change?

No

How was this patch tested?

CI passed

Signed-off-by: sunbaosong <13793883820@163.com>
@sunbaosong sunbaosong marked this pull request as ready for review April 29, 2025 14:44
@sunbaosong sunbaosong changed the title support 32K model len on deepseek r1 W8A8 [Performance] support 32K model len on deepseek r1 W8A8 Apr 29, 2025
@sunbaosong sunbaosong changed the title [Performance] support 32K model len on deepseek r1 W8A8 support 32K model len on deepseek r1 W8A8 Apr 29, 2025
@Yikun Yikun merged commit d6bfae8 into vllm-project:main May 6, 2025
14 checks passed
@Yikun
Copy link
Collaborator

Yikun commented May 6, 2025

Thanks, merged. Will be included in v0.8.5rc1.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants