Skip to content

Commit fccb4d0

Browse files
committed
lint
Signed-off-by: jiahanc <173873397+jiahanc@users.noreply.github.com>
1 parent be4dc1a commit fccb4d0

File tree

3 files changed

+1
-17
lines changed

3 files changed

+1
-17
lines changed

vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -38,11 +38,7 @@ def flashinfer_fused_moe_blockscale_fp8(
3838
topk_group = topk_group if topk_group is not None else 0
3939
assert top_k <= global_num_experts
4040
assert top_k <= 10
41-
# assert topk_group <= 4
42-
# assert global_num_experts > num_expert_group
43-
# assert global_num_experts % num_expert_group == 0
4441
assert global_num_experts % 4 == 0
45-
# assert top_k < (topk_group * global_num_experts / num_expert_group)
4642
assert block_shape == [128, 128]
4743
# Routing kernel expects #experts <= #threads 512
4844
assert global_num_experts <= 512

vllm/model_executor/layers/quantization/fp8.py

Lines changed: 1 addition & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1214,15 +1214,10 @@ def apply(
12141214
assert activation == "silu", (
12151215
f"Expected 'silu' activation but got {activation}"
12161216
)
1217-
# assert scoring_func == "sigmoid", (
1218-
# f"Expected 'sigmoid' scoring func but got {scoring_func}"
1219-
# )
1217+
12201218
if self.block_quant:
12211219
import vllm.model_executor.layers.fused_moe.flashinfer_trtllm_moe # noqa: E501, F401
12221220

1223-
# assert (
1224-
# renormalize and use_grouped_topk and custom_routing_function is None
1225-
# )
12261221
e_score_correction_bias = (
12271222
e_score_correction_bias.to(x.dtype)
12281223
if e_score_correction_bias is not None
@@ -1296,10 +1291,6 @@ def apply(
12961291
# can override fused_experts or cutlass but not rocm or marlin.
12971292
#
12981293
topk_weights, topk_ids, zero_expert_result = select_result
1299-
# if (topk_ids.shape[0] <100):
1300-
# print("=== MoE Routing Results ===")
1301-
# print(f"topk_ids: {topk_ids}")
1302-
# print(f"topk_weights: {topk_weights}")
13031294
if self.rocm_aiter_moe_enabled:
13041295
from vllm.model_executor.layers.fused_moe.rocm_aiter_fused_moe import ( # noqa: E501
13051296
rocm_aiter_fused_experts,

vllm/model_executor/models/qwen3_next.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -206,9 +206,6 @@ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
206206
>>>>>>> 9d88f1762 (update work)
207207

208208
if self.shared_expert is not None:
209-
# if ("model.layers.0." in self.prefix_print or "model.layers.1." in self.prefix_print or "model.layers.47." in self.prefix_print):
210-
# print(f"shared_expert: {final_hidden_states[0]}")
211-
# print(f"routed_expert: {final_hidden_states[1]}")
212209
final_hidden_states = final_hidden_states[0] + final_hidden_states[1]
213210

214211
if self.is_sequence_parallel:

0 commit comments

Comments
 (0)