Skip to content

Commit 428feb1

Browse files
committed
Allow vllm to disable log stats
1 parent 9a895c4 commit 428feb1

File tree

2 files changed

+18
-15
lines changed

2 files changed

+18
-15
lines changed

lmms_eval/models/chat/vllm.py

Lines changed: 15 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -96,20 +96,21 @@ def generate_until(self, requests) -> List[str]:
9696
res.extend(response_text)
9797
pbar.update(len(batch_requests))
9898

99-
metrics = self.get_format_metrics()
100-
total_tokens = metrics["generation_tokens"]
101-
avg_speed = total_tokens / e2e_latency if e2e_latency > 0 else 0
102-
metric_dict = {
103-
"total_tokens": total_tokens,
104-
"e2e_latency": e2e_latency,
105-
"avg_speed": avg_speed,
106-
"additional_metrics": {
107-
"ttft": metrics["ttft"],
108-
"tpot": metrics["tpot"],
109-
"rank": self.rank,
110-
},
111-
}
112-
log_metrics(**metric_dict)
99+
if not self.disable_log_stats:
100+
metrics = self.get_format_metrics()
101+
total_tokens = metrics["generation_tokens"]
102+
avg_speed = total_tokens / e2e_latency if e2e_latency > 0 else 0
103+
metric_dict = {
104+
"total_tokens": total_tokens,
105+
"e2e_latency": e2e_latency,
106+
"avg_speed": avg_speed,
107+
"additional_metrics": {
108+
"ttft": metrics["ttft"],
109+
"tpot": metrics["tpot"],
110+
"rank": self.rank,
111+
},
112+
}
113+
log_metrics(**metric_dict)
113114

114115
pbar.close()
115116
return res

lmms_eval/models/simple/vllm.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -155,6 +155,7 @@ def __init__(
155155
trust_remote_code: Optional[bool] = True,
156156
chat_template: Optional[str] = None,
157157
min_image_pixels: int = 28, # minimum image dimension, required for Qwen 2/2.5-VL models
158+
disable_log_stats: bool = False,
158159
**kwargs,
159160
) -> None:
160161
super().__init__()
@@ -216,10 +217,11 @@ def __init__(
216217
tensor_parallel_size=tensor_parallel_size,
217218
gpu_memory_utilization=gpu_memory_utilization,
218219
trust_remote_code=trust_remote_code,
219-
disable_log_stats=False,
220+
disable_log_stats=disable_log_stats,
220221
seed=1,
221222
**kwargs,
222223
)
224+
self.disable_log_stats = disable_log_stats
223225

224226
self.device = self.accelerator.device
225227
self.batch_size_per_gpu = int(batch_size)

0 commit comments

Comments
 (0)