Skip to content

Commit 669f47c

Browse files
committed
fixes
Signed-off-by: Bill Nell <[email protected]>
1 parent 30dfc50 commit 669f47c

File tree

1 file changed

+2
-35
lines changed
  • vllm/model_executor/layers/fused_moe

1 file changed

+2
-35
lines changed

vllm/model_executor/layers/fused_moe/layer.py

Lines changed: 2 additions & 35 deletions
Original file line numberDiff line numberDiff line change
@@ -490,7 +490,7 @@ def __init__(
490490
return_expert_mask=self.rocm_aiter_fmoe_enabled,
491491
)
492492
self.local_num_experts = local_num_experts
493-
self.register_buffer("expert_map", expert_map)
493+
self.register_buffer("_expert_map", expert_map)
494494
self.register_buffer("expert_mask", expert_mask)
495495
self._maybe_init_expert_routing_tables()
496496
logger.info_once(
@@ -827,7 +827,7 @@ def update_expert_map(self):
827827
return_expert_mask=self.rocm_aiter_fmoe_enabled,
828828
)
829829
self.local_num_experts = local_num_experts
830-
self.register_buffer("expert_map", expert_map)
830+
self.register_buffer("_expert_map", expert_map)
831831
self.register_buffer("expert_mask", expert_mask)
832832
self._maybe_init_expert_routing_tables()
833833
if self.aiter_fmoe_shared_expert_enabled:
@@ -1704,22 +1704,6 @@ def process_chunk(chunk_start, chunk_end, skip_result_store=False):
17041704
layer=self,
17051705
x=staged_hidden_states,
17061706
router_logits=staged_router_logits,
1707-
top_k=self.top_k,
1708-
renormalize=self.renormalize,
1709-
use_grouped_topk=self.use_grouped_topk,
1710-
global_num_experts=self.global_num_experts,
1711-
expert_map=self.expert_map,
1712-
topk_group=self.topk_group,
1713-
num_expert_group=self.num_expert_group,
1714-
custom_routing_function=self.custom_routing_function,
1715-
scoring_func=self.scoring_func,
1716-
routed_scaling_factor=self.routed_scaling_factor,
1717-
e_score_correction_bias=self.e_score_correction_bias,
1718-
activation=self.activation,
1719-
enable_eplb=self.enable_eplb,
1720-
expert_load_view=self.expert_load_view,
1721-
logical_to_physical_map=self.logical_to_physical_map,
1722-
logical_replica_count=self.logical_replica_count,
17231707
)
17241708

17251709
if has_separate_shared_experts:
@@ -1871,23 +1855,6 @@ def forward_impl(
18711855
if do_naive_dispatch_combine
18721856
else hidden_states,
18731857
router_logits=router_logits,
1874-
top_k=self.top_k,
1875-
renormalize=self.renormalize,
1876-
use_grouped_topk=self.use_grouped_topk,
1877-
global_num_experts=self.global_num_experts,
1878-
expert_map=self.expert_map,
1879-
topk_group=self.topk_group,
1880-
num_expert_group=self.num_expert_group,
1881-
custom_routing_function=self.custom_routing_function,
1882-
scoring_func=self.scoring_func,
1883-
routed_scaling_factor=self.routed_scaling_factor,
1884-
e_score_correction_bias=self.e_score_correction_bias,
1885-
activation=self.activation,
1886-
apply_router_weight_on_input=self.apply_router_weight_on_input,
1887-
enable_eplb=self.enable_eplb,
1888-
expert_load_view=self.expert_load_view,
1889-
logical_to_physical_map=self.logical_to_physical_map,
1890-
logical_replica_count=self.logical_replica_count,
18911858
)
18921859

18931860
if has_separate_shared_experts:

0 commit comments

Comments
 (0)