@@ -287,7 +287,8 @@ def llm_export(args: ExportArguments) -> None:
287287 'Skipping the conversion process.' )
288288 else :
289289 from swift .llm .megatron import MegatronArguments , convert_hf_to_megatron , patch_megatron
290- model , tokenizer = get_model_tokenizer (args .model_type , torch .float32 , {'device_map' : 'auto' })
290+ model , tokenizer = get_model_tokenizer (
291+ args .model_type , torch .float32 , {'device_map' : 'auto' }, model_id_or_path = args .model_id_or_path )
291292 res = MegatronArguments .load_megatron_config (tokenizer .model_dir )
292293 res ['model_type' ] = args .model_type
293294 res ['target_tensor_model_parallel_size' ] = args .tp
@@ -311,7 +312,8 @@ def llm_export(args: ExportArguments) -> None:
311312 'Skipping the conversion process.' )
312313 else :
313314 from swift .llm .megatron import MegatronArguments , convert_megatron_to_hf , patch_megatron
314- hf_model , tokenizer = get_model_tokenizer (args .model_type , torch .float32 , {'device_map' : 'auto' })
315+ hf_model , tokenizer = get_model_tokenizer (
316+ args .model_type , torch .float32 , {'device_map' : 'auto' }, model_id_or_path = args .model_id_or_path )
315317 res = MegatronArguments .load_megatron_config (tokenizer .model_dir )
316318 res ['model_type' ] = args .model_type
317319 res ['target_tensor_model_parallel_size' ] = args .tp
0 commit comments