Skip to content

Commit 2ebf603

Browse files
committed
don't double eplb prefix
Signed-off-by: Patryk Saffer <[email protected]>
1 parent 64975d7 commit 2ebf603

File tree

5 files changed

+15
-15
lines changed

5 files changed

+15
-15
lines changed

tests/v1/e2e/test_eplb_offline.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ def test_eplb_model(
4848
)
4949

5050
# Save EPLB statistics to disk
51-
eplb_config_save = EPLBConfig(eplb_save_dir="/tmp")
51+
eplb_config_save = EPLBConfig(save_dir="/tmp")
5252
llm = LLM(eplb_config=eplb_config_save, **llm_args)
5353
llm.generate(test_prompts, sampling_config)
5454
del llm
@@ -57,7 +57,7 @@ def test_eplb_model(
5757

5858
# Load EPLB statistics from disk
5959
eplb_config_load = EPLBConfig(
60-
eplb_load_path="/tmp/global_expert_load_window_i0.safetensors"
60+
load_path="/tmp/global_expert_load_window_i0.safetensors"
6161
)
6262
llm = LLM(eplb_config=eplb_config_load, **llm_args)
6363
llm.generate(test_prompts, sampling_config)

vllm/config/parallel.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -62,15 +62,15 @@ class EPLBConfig:
6262
This is turned off by default since it will cause communication overhead.
6363
"""
6464

65-
eplb_save_dir: Path | None = None
65+
save_dir: Path | None = None
6666
"""Directory to save expert load balance metrics."""
67-
eplb_load_path: Path | None = None
67+
load_path: Path | None = None
6868
"""Path to load expert load balance metrics."""
6969

7070
@property
71-
def eplb_record_metrics(self) -> bool:
72-
return self.eplb_save_dir is not None or (
73-
self.eplb_save_dir is None and self.eplb_load_path is None
71+
def record_metrics(self) -> bool:
72+
return self.save_dir is not None or (
73+
self.save_dir is None and self.load_path is None
7474
)
7575

7676

vllm/distributed/eplb/eplb_state.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -151,11 +151,11 @@ class EplbState:
151151
See:
152152
https://github.com/vllm-project/vllm/pull/22167#pullrequestreview-3086143856
153153
"""
154-
eplb_load_path: Path | None = None
154+
load_path: Path | None = None
155155
"""
156156
Path for loading eplb initial state.
157157
"""
158-
eplb_save_dir: Path | None = None
158+
save_dir: Path | None = None
159159
"""
160160
Path where eplb states will be saved.
161161
"""
@@ -298,7 +298,7 @@ def build(
298298
)
299299

300300
eplb_load_path = parallel_config.eplb_config.eplb_load_path
301-
eplb_save_dir = parallel_config.eplb_config.eplb_save_dir
301+
eplb_save_dir = parallel_config.eplb_config.save_dir
302302

303303
eplb_step_interval = parallel_config.eplb_config.step_interval
304304
if eplb_load_path is not None or eplb_save_dir is not None:
@@ -498,8 +498,8 @@ def rearrange(
498498
time_start = time.perf_counter()
499499
logger.info("Rearranging experts %s...", "(profile)" if is_profile else "")
500500

501-
if self.eplb_load_path is not None and self.expert_rearrangement_step == 0:
502-
global_expert_load_window = load_eplb_state(self.eplb_load_path).to(
501+
if self.load_path is not None and self.expert_rearrangement_step == 0:
502+
global_expert_load_window = load_eplb_state(self.load_path).to(
503503
self.physical_to_logical_map.device
504504
)
505505
elif global_expert_load is None:

vllm/engine/arg_utils.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -400,8 +400,8 @@ class EngineArgs:
400400
eplb_window_size: int = EPLBConfig.window_size
401401
eplb_step_interval: int = EPLBConfig.step_interval
402402
eplb_log_balancedness: bool = EPLBConfig.log_balancedness
403-
eplb_save_dir: Path | None = EPLBConfig.eplb_save_dir
404-
eplb_load_path: Path | None = EPLBConfig.eplb_load_path
403+
eplb_save_dir: Path | None = EPLBConfig.save_dir
404+
eplb_load_path: Path | None = EPLBConfig.load_path
405405
max_parallel_loading_workers: int | None = (
406406
ParallelConfig.max_parallel_loading_workers
407407
)

vllm/model_executor/models/transformers/moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -236,7 +236,7 @@ def forward(self, *args, **kwargs):
236236
# Expert parallel load balancing kwargs
237237
enable_eplb = self.parallel_config.enable_eplb
238238
num_redundant_experts = self.parallel_config.eplb_config.num_redundant_experts
239-
eplb_record_metrics = self.parallel_config.eplb_config.eplb_record_metrics
239+
eplb_record_metrics = self.parallel_config.eplb_config.record_metrics
240240

241241
# MixtureOfExperts mixin settings
242242
ep_size = get_ep_group().world_size

0 commit comments

Comments
 (0)