Skip to content

Commit 3145121

Browse files
committed
fix ut
Signed-off-by: wangli <wangli858794774@gmail.com>
1 parent cba0337 commit 3145121

File tree

3 files changed

+10
-0
lines changed

3 files changed

+10
-0
lines changed

tests/ut/ops/test_fused_ops.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -231,6 +231,9 @@ def apply(self, hidden_states: torch.Tensor,
231231
expert_weights: torch.Tensor) -> torch.Tensor:
232232
pass
233233

234+
def get_fused_moe_quant_config(self, layer: torch.nn.Module):
235+
pass
236+
234237

235238
class TestAscendFusedMoe:
236239

tests/ut/torchair/ops/test_torchair_fused_moe.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -197,6 +197,9 @@ def apply(self, hidden_states: torch.Tensor,
197197
expert_weights: torch.Tensor) -> torch.Tensor:
198198
pass
199199

200+
def get_fused_moe_quant_config(self, layer: torch.nn.Module):
201+
pass
202+
200203

201204
class TestTorchairAscendFusedMoe:
202205

vllm_ascend/quantization/quant_config.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -343,6 +343,10 @@ def process_weights_after_loading(self, layer: torch.nn.Module) -> None:
343343
if hasattr(self.quant_method, "process_weights_after_loading"):
344344
self.quant_method.process_weights_after_loading(layer)
345345

346+
def get_fused_moe_quant_config(self, layer: torch.nn.Module):
347+
# TODO: implement this function
348+
pass
349+
346350

347351
class AscendEmbeddingMethod(AscendLinearMethod):
348352
"""Embedding method for Ascend quantization.

0 commit comments

Comments
 (0)