-
Notifications
You must be signed in to change notification settings - Fork 32
Open
Description
In compress_llama.sh:
python utils/LoRA.py --prune_model --data_path yahma/alpaca-cleaned --output_dir $FINE_TUNE_PATH/first_half --lora_target_modules q_u_proj,k_u_proj,v_u_proj,o_u_proj,gate_u_proj,down_u_proj,up_u_proj --lora_r 8 --num_epochs 3 --learning_rate 1e-4 --batch_size 64
shouldn't --prune_model be provided with a HF/local model path here?
I have tried to add jeffwan_llama_7b_hf_whitening_only_0.8.pt as the --prune_model. but I got the following error:
File "/data/yusenp/SVD-VGGT/SVD-LLM/component/svd_llama.py", line 192, in forward
raise ValueError(
ValueError: Attention mask should be of size (4, 1, 360, 360), but is torch.Size([4, 1, 2048, 2048])
Can anyone give more clarification/update/correction of compress_llama.sh about this? I realy appreciate it!
Metadata
Metadata
Assignees
Labels
No labels