@@ -439,7 +439,7 @@ def apply(
439439 apply_router_weight_on_input : bool = False ,
440440 activation : str = "silu" ,
441441 enable_eplb : bool = False ,
442- eplb_record_metrics : bool = False ,
442+ eplb_static : bool = False ,
443443 expert_load_view : torch .Tensor | None = None ,
444444 logical_to_physical_map : torch .Tensor | None = None ,
445445 logical_replica_count : torch .Tensor | None = None ,
@@ -1019,7 +1019,7 @@ def apply(
10191019 apply_router_weight_on_input : bool = False ,
10201020 activation : str = "silu" ,
10211021 enable_eplb : bool = False ,
1022- eplb_record_metrics : bool = False ,
1022+ eplb_static : bool = False ,
10231023 expert_load_view : torch .Tensor | None = None ,
10241024 logical_to_physical_map : torch .Tensor | None = None ,
10251025 logical_replica_count : torch .Tensor | None = None ,
@@ -1288,7 +1288,7 @@ def apply(
12881288 apply_router_weight_on_input : bool = False ,
12891289 activation : str = "silu" ,
12901290 enable_eplb : bool = False ,
1291- eplb_record_metrics : bool = False ,
1291+ eplb_static : bool = False ,
12921292 expert_load_view : torch .Tensor | None = None ,
12931293 logical_to_physical_map : torch .Tensor | None = None ,
12941294 logical_replica_count : torch .Tensor | None = None ,
@@ -1650,7 +1650,7 @@ def apply(
16501650 apply_router_weight_on_input : bool = False ,
16511651 activation : str = "silu" ,
16521652 enable_eplb : bool = False ,
1653- eplb_record_metrics : bool = False ,
1653+ eplb_static : bool = False ,
16541654 expert_load_view : torch .Tensor | None = None ,
16551655 logical_to_physical_map : torch .Tensor | None = None ,
16561656 logical_replica_count : torch .Tensor | None = None ,
@@ -1914,7 +1914,7 @@ def apply(
19141914 apply_router_weight_on_input : bool = False ,
19151915 activation : str = "silu" ,
19161916 enable_eplb : bool = False ,
1917- eplb_record_metrics : bool = False ,
1917+ eplb_static : bool = False ,
19181918 expert_load_view : torch .Tensor | None = None ,
19191919 logical_to_physical_map : torch .Tensor | None = None ,
19201920 logical_replica_count : torch .Tensor | None = None ,
@@ -2238,7 +2238,7 @@ def apply(
22382238 apply_router_weight_on_input : bool = False ,
22392239 activation : str = "silu" ,
22402240 enable_eplb : bool = False ,
2241- eplb_record_metrics : bool = False ,
2241+ eplb_static : bool = False ,
22422242 expert_load_view : torch .Tensor | None = None ,
22432243 logical_to_physical_map : torch .Tensor | None = None ,
22442244 logical_replica_count : torch .Tensor | None = None ,
0 commit comments