Skip to content

[Perf] Add new npu_fused_infer_attention_score op to improve perfomance in splitfuse cases and resolve long-seq mask problems #571

[Perf] Add new npu_fused_infer_attention_score op to improve perfomance in splitfuse cases and resolve long-seq mask problems

[Perf] Add new npu_fused_infer_attention_score op to improve perfomance in splitfuse cases and resolve long-seq mask problems #571

Triggered via pull request September 20, 2025 04:16
Status Skipped
Total duration 1s
Artifacts

vllm_ascend_test_full.yaml

on: pull_request
changes
0s
changes
Matrix: multicard e2e test - full
Matrix: singlecard e2e test - full
Fit to window
Zoom out
Zoom in