Skip to content

Commit 9f6157d

Browse files
author
unknown
committed
fix fc1
1 parent f3b44b4 commit 9f6157d

File tree

3 files changed

+0
-251
lines changed

3 files changed

+0
-251
lines changed

1.patch

Lines changed: 0 additions & 247 deletions
This file was deleted.

vllm_ascend/ops/layernorm.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,6 @@ def forward(
6363
self.variance_epsilon)
6464
return x
6565

66-
cnt = 0
6766
class AscendRMSNorm(RMSNorm):
6867
def __init__(
6968
self,
@@ -85,9 +84,7 @@ def forward_oot(
8584
import torch_npu
8685
from vllm_ascend.utils import is_310p
8786
if residual is not None:
88-
global cnt
8987
residual = torch.ops.vllm.maybe_chunk_residual(x, residual)
90-
cnt = cnt + 1
9188
assert x.size(0) == residual.size(0)
9289
if is_310p():
9390
orig_dtype = residual.dtype

vllm_ascend/ops/linear.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -409,7 +409,6 @@ def __init__(
409409
else:
410410
comm_group = get_tp_group()
411411
self.forward_type = "normal_tp"
412-
self.prefix = prefix
413412
self.comm_group = comm_group
414413
# TODO: check for disable_tp
415414
self.tp_rank = comm_group.rank_in_group

0 commit comments

Comments
 (0)