Skip to content

Commit 78b68ab

Browse files
authored
[VLLM] Allows for max tokens to be set in model config file (#547)
* commit * commit * Update src/lighteval/main_vllm.py * commit * change doc * change doc * change doc * allow max new token to be set in model config file
1 parent da119e8 commit 78b68ab

File tree

2 files changed

+4
-2
lines changed

2 files changed

+4
-2
lines changed

examples/model_configs/vllm_model_config.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,5 +10,5 @@ model:
1010
top_k: -1
1111
min_p: 0.0
1212
top_p: 0.9
13-
max_new_tokens: 100
13+
max_new_tokens: 256
1414
stop_tokens: ["<EOS>", "<PAD>"]

src/lighteval/models/vllm/vllm_model.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -316,7 +316,9 @@ def _generate(
316316
sampling_params = self.sampling_params.clone() or SamplingParams()
317317
if generate:
318318
sampling_params.n = num_samples
319-
sampling_params.max_tokens = max_new_tokens
319+
sampling_params.max_tokens = (
320+
max_new_tokens if sampling_params.max_tokens is None else sampling_params.max_tokens
321+
)
320322
sampling_params.stop = stop_tokens
321323
sampling_params.logprobs = 1 if returns_logits else 0
322324

0 commit comments

Comments
 (0)