Skip to content

[Perf] Add new npu_fused_infer_attention_score op to improve perfomance in splitfuse cases and resolve long-seq mask problems #5962

[Perf] Add new npu_fused_infer_attention_score op to improve perfomance in splitfuse cases and resolve long-seq mask problems

[Perf] Add new npu_fused_infer_attention_score op to improve perfomance in splitfuse cases and resolve long-seq mask problems #5962

Triggered via pull request September 20, 2025 04:44
@AngazennAngazenn
synchronize #2962
Status Success
Total duration 12s
Artifacts

format_pr_body.yaml

on: pull_request_target
update vLLM version
9s
update vLLM version
Fit to window
Zoom out
Zoom in

Annotations

1 warning
update vLLM version
The `python-version` input is not set. The version of Python currently in `PATH` will be used.