File tree Expand file tree Collapse file tree 2 files changed +6
-5
lines changed Expand file tree Collapse file tree 2 files changed +6
-5
lines changed Original file line number Diff line number Diff line change @@ -49,7 +49,7 @@ def __init__(self, vllm_config):
49
49
"expert_map_record_path" ,
50
50
None ) # Provide path to export expert map
51
51
self .init_redundancy_expert = additional_config .get (
52
- "init_redundancy_expert" ,0 )
52
+ "init_redundancy_expert" , 0 )
53
53
self .dynamic_eplb = additional_config .get ("dynamic_eplb" , False )
54
54
self .num_iterations_eplb_update = additional_config .get (
55
55
"num_iterations_eplb_update" , 400 )
Original file line number Diff line number Diff line change 37
37
FusedMoE , UnquantizedFusedMoEMethod , determine_expert_map )
38
38
from vllm .model_executor .layers .quantization .base_config import \
39
39
QuantizationConfig
40
- from vllm_ascend .eplb .core .eplb_utils import (
41
- determine_default_expert_map ,
42
- determine_default_log2phy_map )
40
+
43
41
from vllm_ascend .ascend_config import get_ascend_config
44
42
from vllm_ascend .ascend_forward_context import FusedMoEState
45
43
from vllm_ascend .distributed .parallel_state import get_mc2_group
44
+ from vllm_ascend .eplb .core .eplb_utils import (determine_default_expert_map ,
45
+ determine_default_log2phy_map )
46
46
from vllm_ascend .ops .expert_load_balancer import ExpertLoadBalancer
47
47
from vllm_ascend .ops .sequence_parallel import MetadataForPadding
48
48
from vllm_ascend .quantization .quant_config import AscendFusedMoEMethod
@@ -1080,7 +1080,8 @@ def __init__(
1080
1080
assert self .quant_method is not None
1081
1081
1082
1082
self .moe_load = None
1083
- local_num_experts = (torch .sum (self .expert_map != - 1 ) if self .expert_map is not None else num_experts )
1083
+ local_num_experts = (torch .sum (self .expert_map != - 1 )
1084
+ if self .expert_map is not None else num_experts )
1084
1085
if self .dynamic_eplb :
1085
1086
self .moe_load = torch .zeros (local_num_experts , dtype = torch .int64 )
1086
1087
You can’t perform that action at this time.
0 commit comments