Skip to content

Commit e11e95a

Browse files
Update fused_moe.py
1 parent 265ffe0 commit e11e95a

File tree

1 file changed

+0
-2
lines changed

1 file changed

+0
-2
lines changed

vllm_ascend/ops/fused_moe/fused_moe.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -194,8 +194,6 @@ def __init__(self, *args, **kwargs):
194194
self.expert_load_balancer = ExpertLoadBalancer(
195195
self.expert_map_path, num_experts)
196196
self.expert_load_balancer.check_expert_map_tensor()
197-
# self.global_redundant_expert_num = (
198-
# self.expert_load_balancer.get_global_redundant_expert_num())
199197
self.global_num_experts = num_experts + self.global_redundant_expert_num
200198
try:
201199
self.local_num_experts, self.expert_map = (

0 commit comments

Comments
 (0)