File tree Expand file tree Collapse file tree 1 file changed +6
-2
lines changed
vllm/model_executor/models Expand file tree Collapse file tree 1 file changed +6
-2
lines changed Original file line number Diff line number Diff line change @@ -467,11 +467,15 @@ class Llama4ForCausalLM(LlamaForCausalLM):
467
467
}
468
468
469
469
def __init__ (self , * , vllm_config : VllmConfig , prefix : str = "" ):
470
- # Update temperature tuning config from generation config
470
+ # update temperature tuning config from generation config
471
471
gen_config = vllm_config .model_config .try_get_generation_config ()
472
472
gen_config .update (vllm_config .model_config .override_generation_config )
473
+ # enable temperature tuning by default when max_model_len > 32K
474
+ default_attn_temperature_tuning = \
475
+ vllm_config .model_config .max_model_len > 32768
473
476
vllm_config .model_config .hf_config .attn_temperature_tuning \
474
- = gen_config .get ("attn_temperature_tuning" , False )
477
+ = gen_config .get (
478
+ "attn_temperature_tuning" , default_attn_temperature_tuning )
475
479
476
480
super ().__init__ (vllm_config = vllm_config ,
477
481
prefix = prefix ,
You can’t perform that action at this time.
0 commit comments