Skip to content

Commit 53f1507

Browse files
remove all redundant_enable
Signed-off-by: wanghanqingLYT <wanghanqing3@huawei.com>
1 parent 66c34c2 commit 53f1507

File tree

3 files changed

+10
-26
lines changed

3 files changed

+10
-26
lines changed

vllm_ascend/eplb/core/eplb_device_transfer_loader.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -96,7 +96,7 @@ def asyn_expert_weight_transfer(self, reqs):
9696

9797
self.state = ExpertWeightUpdateState.TRANSFERRING
9898

99-
def update_expert_map_and_weight(self, reqs, redundant_enable):
99+
def update_expert_map_and_weight(self, reqs):
100100
# Only after send/recv tasks have been luanched, expert_map and weight can be updated
101101
if self.state != ExpertWeightUpdateState.TRANSFERRING:
102102
return
@@ -113,9 +113,8 @@ def update_expert_map_and_weight(self, reqs, redundant_enable):
113113
self.updated_expert_map)
114114

115115
# update log2phy_map
116-
if redundant_enable:
117-
self.eplb_adaptor.do_update_log2phy_map(self.layer_id,
118-
self.updated_log2phy_map)
116+
self.eplb_adaptor.do_update_log2phy_map(self.layer_id,
117+
self.updated_log2phy_map)
119118

120119
# update expert weight
121120
buffer_tensor_id = 0

vllm_ascend/eplb/core/eplb_worker.py

Lines changed: 4 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -31,18 +31,13 @@
3131

3232
class EplbWorker:
3333

34-
def __init__(self,
35-
shared_dict,
36-
policy_type,
37-
enable_d2d: bool = True,
38-
redundant_enable=0):
34+
def __init__(self, shared_dict, policy_type, enable_d2d: bool = True):
3935
self.policy_type = policy_type
4036
self.policy = PolicyFactory.generate_policy(policy_type,
4137
DynamicConfig())
4238
self.shared_dict = shared_dict
4339
self.old_expert_maps = None
4440
self.enable_d2d = enable_d2d
45-
self.redundant_enable = redundant_enable
4641
self.rank_id = dist.get_rank()
4742

4843
def do_update(self):
@@ -381,11 +376,8 @@ def pack_update_info(self, update_info_generator):
381376

382377
maps.append(new_expert_map[self.rank_id].numpy().tolist())
383378

384-
if self.redundant_enable:
385-
log2phy_map = generate_log2phy_map(new_expert_map)
386-
log2phy_all.append(log2phy_map[self.rank_id].numpy().tolist())
387-
else:
388-
log2phy_all.append([])
379+
log2phy_map = generate_log2phy_map(new_expert_map)
380+
log2phy_all.append(log2phy_map[self.rank_id].numpy().tolist())
389381

390382
layer_ids.append(layer_id)
391383

@@ -398,7 +390,6 @@ def __init__(self,
398390
shared_dict,
399391
planner_q,
400392
block_update_q,
401-
redundant_enable,
402393
policy_type: int = 0,
403394
enable_d2d: bool = True):
404395
"""
@@ -412,11 +403,10 @@ def __init__(self,
412403
self.enable_d2d = enable_d2d
413404
self.planner_q = planner_q
414405
self.block_update_q = block_update_q
415-
self.redundant_enable = redundant_enable
416406

417407
# Create EplbWorker instance
418408
self.worker = EplbWorker(self.shared_dict, self.policy_type,
419-
self.enable_d2d, self.redundant_enable)
409+
self.enable_d2d)
420410

421411
def worker_process(self, planner_q, block_update_q):
422412
"""

vllm_ascend/eplb/eplb_updator.py

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,6 @@ def init_eplb(self, expert_map_path):
4545
self.rank_id = dist.get_rank()
4646
self.num_expert_load_gather = 10
4747
self.periodic_load_gather = True
48-
self.redundant_enable = (expert_map_path is not None)
4948
self.num_iterations_eplb_update: torch.int64 = self.ascend_config.num_iterations_eplb_update
5049
self.expert_map_path = expert_map_path
5150

@@ -83,7 +82,6 @@ def init_eplb(self, expert_map_path):
8382
self.eplb = EplbProcess(shared_dict=self.shared_dict,
8483
planner_q=self.planner_block_queue,
8584
block_update_q=self.block_update_queue,
86-
redundant_enable=self.redundant_enable,
8785
policy_type=1,
8886
enable_d2d=True)
8987

@@ -132,10 +130,8 @@ def forward_before(self):
132130
if self.update_expert_weight_flag():
133131
(expert_send_info, expert_recv_info, updated_expert_map,
134132
log2phy_map, layer_id) = self.update_info_all.pop(0)
135-
if self.redundant_enable:
136-
log2phy_map_this_rank = torch.from_numpy(
137-
numpy.array(log2phy_map))
138-
self.eplb_loader.set_log2phy_map(log2phy_map_this_rank)
133+
log2phy_map_this_rank = torch.from_numpy(numpy.array(log2phy_map))
134+
self.eplb_loader.set_log2phy_map(log2phy_map_this_rank)
139135
updated_expert_map_this_rank = torch.from_numpy(
140136
numpy.array(updated_expert_map))
141137
# logger.info(f"check update info, layer = {layer_id}, send = {expert_send_info_this_rank}, recv = {expert_recv_info_this_rank}")
@@ -159,8 +155,7 @@ def forward_end(self):
159155
self.wakeup_eplb_worker()
160156

161157
if self.update_expert_weight_flag():
162-
self.eplb_loader.update_expert_map_and_weight(
163-
self.reqs, self.redundant_enable)
158+
self.eplb_loader.update_expert_map_and_weight(self.reqs)
164159

165160
self.update_iteration()
166161

0 commit comments

Comments
 (0)