@@ -490,7 +490,7 @@ def __init__(
490490 return_expert_mask = self .rocm_aiter_fmoe_enabled ,
491491 )
492492 self .local_num_experts = local_num_experts
493- self .register_buffer ("expert_map " , expert_map )
493+ self .register_buffer ("_expert_map " , expert_map )
494494 self .register_buffer ("expert_mask" , expert_mask )
495495 self ._maybe_init_expert_routing_tables ()
496496 logger .info_once (
@@ -827,7 +827,7 @@ def update_expert_map(self):
827827 return_expert_mask = self .rocm_aiter_fmoe_enabled ,
828828 )
829829 self .local_num_experts = local_num_experts
830- self .register_buffer ("expert_map " , expert_map )
830+ self .register_buffer ("_expert_map " , expert_map )
831831 self .register_buffer ("expert_mask" , expert_mask )
832832 self ._maybe_init_expert_routing_tables ()
833833 if self .aiter_fmoe_shared_expert_enabled :
@@ -1704,22 +1704,6 @@ def process_chunk(chunk_start, chunk_end, skip_result_store=False):
17041704 layer = self ,
17051705 x = staged_hidden_states ,
17061706 router_logits = staged_router_logits ,
1707- top_k = self .top_k ,
1708- renormalize = self .renormalize ,
1709- use_grouped_topk = self .use_grouped_topk ,
1710- global_num_experts = self .global_num_experts ,
1711- expert_map = self .expert_map ,
1712- topk_group = self .topk_group ,
1713- num_expert_group = self .num_expert_group ,
1714- custom_routing_function = self .custom_routing_function ,
1715- scoring_func = self .scoring_func ,
1716- routed_scaling_factor = self .routed_scaling_factor ,
1717- e_score_correction_bias = self .e_score_correction_bias ,
1718- activation = self .activation ,
1719- enable_eplb = self .enable_eplb ,
1720- expert_load_view = self .expert_load_view ,
1721- logical_to_physical_map = self .logical_to_physical_map ,
1722- logical_replica_count = self .logical_replica_count ,
17231707 )
17241708
17251709 if has_separate_shared_experts :
@@ -1871,23 +1855,6 @@ def forward_impl(
18711855 if do_naive_dispatch_combine
18721856 else hidden_states ,
18731857 router_logits = router_logits ,
1874- top_k = self .top_k ,
1875- renormalize = self .renormalize ,
1876- use_grouped_topk = self .use_grouped_topk ,
1877- global_num_experts = self .global_num_experts ,
1878- expert_map = self .expert_map ,
1879- topk_group = self .topk_group ,
1880- num_expert_group = self .num_expert_group ,
1881- custom_routing_function = self .custom_routing_function ,
1882- scoring_func = self .scoring_func ,
1883- routed_scaling_factor = self .routed_scaling_factor ,
1884- e_score_correction_bias = self .e_score_correction_bias ,
1885- activation = self .activation ,
1886- apply_router_weight_on_input = self .apply_router_weight_on_input ,
1887- enable_eplb = self .enable_eplb ,
1888- expert_load_view = self .expert_load_view ,
1889- logical_to_physical_map = self .logical_to_physical_map ,
1890- logical_replica_count = self .logical_replica_count ,
18911858 )
18921859
18931860 if has_separate_shared_experts :
0 commit comments