@@ -796,7 +796,7 @@ def fix_qwen_inplace_bug(model) -> None:
796796 * args , ** kwargs ).clone ()
797797 else :
798798 __old_forward = first_drop .forward
799- first_drop .forwad = lambda * args , ** kwargs : __old_forward (
799+ first_drop .forward = lambda * args , ** kwargs : __old_forward (
800800 * args , ** kwargs ).clone ()
801801 first_drop .__old_forward = __old_forward
802802
@@ -882,7 +882,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
882882 'qwen/Qwen-1_8B-Chat-Int8' ,
883883 LoRATM .qwen ,
884884 TemplateType .chatml ,
885- requires = ['auto_gptq>=0.4.2 ' ],
885+ requires = ['auto_gptq>=0.5 ' ],
886886 torch_dtype = torch .float16 ,
887887 function_kwargs = {'bits' : 8 },
888888 support_flash_attn = True )
@@ -891,7 +891,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
891891 'qwen/Qwen-1_8B-Chat-Int4' ,
892892 LoRATM .qwen ,
893893 TemplateType .chatml ,
894- requires = ['auto_gptq>=0.4.2 ' ],
894+ requires = ['auto_gptq>=0.5 ' ],
895895 torch_dtype = torch .float16 ,
896896 function_kwargs = {'bits' : 4 },
897897 support_flash_attn = True )
@@ -900,7 +900,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
900900 'qwen/Qwen-72B-Chat-Int8' ,
901901 LoRATM .qwen ,
902902 TemplateType .chatml ,
903- requires = ['auto_gptq>=0.4.2 ' ],
903+ requires = ['auto_gptq>=0.5 ' ],
904904 torch_dtype = torch .float16 ,
905905 function_kwargs = {'bits' : 8 },
906906 support_flash_attn = True )
@@ -909,7 +909,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
909909 'qwen/Qwen-72B-Chat-Int4' ,
910910 LoRATM .qwen ,
911911 TemplateType .chatml ,
912- requires = ['auto_gptq>=0.4.2 ' ],
912+ requires = ['auto_gptq>=0.5 ' ],
913913 torch_dtype = torch .float16 ,
914914 function_kwargs = {'bits' : 4 },
915915 support_flash_attn = True )
@@ -918,7 +918,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
918918 'TongyiFinance/Tongyi-Finance-14B-Chat-Int4' ,
919919 LoRATM .qwen ,
920920 TemplateType .chatml ,
921- requires = ['auto_gptq>=0.4.2 ' ],
921+ requires = ['auto_gptq>=0.5 ' ],
922922 torch_dtype = torch .float16 ,
923923 function_kwargs = {'bits' : 4 },
924924 support_flash_attn = True )
@@ -927,7 +927,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
927927 'qwen/Qwen-VL-Chat-Int4' ,
928928 LoRATM .qwen ,
929929 TemplateType .chatml ,
930- requires = ['auto_gptq>=0.4.2 ' ],
930+ requires = ['auto_gptq>=0.5 ' ],
931931 torch_dtype = torch .float16 ,
932932 support_flash_attn = True ,
933933 function_kwargs = {
@@ -939,7 +939,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
939939 'qwen/Qwen-14B-Chat-Int8' ,
940940 LoRATM .qwen ,
941941 TemplateType .chatml ,
942- requires = ['auto_gptq>=0.4.2 ' ],
942+ requires = ['auto_gptq>=0.5 ' ],
943943 torch_dtype = torch .float16 ,
944944 function_kwargs = {'bits' : 8 },
945945 support_flash_attn = True )
@@ -948,7 +948,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
948948 'qwen/Qwen-7B-Chat-Int8' ,
949949 LoRATM .qwen ,
950950 TemplateType .chatml ,
951- requires = ['auto_gptq>=0.4.2 ' ],
951+ requires = ['auto_gptq>=0.5 ' ],
952952 torch_dtype = torch .float16 ,
953953 function_kwargs = {'bits' : 8 },
954954 support_flash_attn = True )
@@ -957,7 +957,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
957957 'qwen/Qwen-14B-Chat-Int4' ,
958958 LoRATM .qwen ,
959959 TemplateType .chatml ,
960- requires = ['auto_gptq>=0.4.2 ' ],
960+ requires = ['auto_gptq>=0.5 ' ],
961961 torch_dtype = torch .float16 ,
962962 function_kwargs = {'bits' : 4 },
963963 support_flash_attn = True )
@@ -966,7 +966,7 @@ def get_model_tokenizer_qwen_audio(model_dir: str,
966966 'qwen/Qwen-7B-Chat-Int4' ,
967967 LoRATM .qwen ,
968968 TemplateType .chatml ,
969- requires = ['auto_gptq>=0.4.2 ' ],
969+ requires = ['auto_gptq>=0.5 ' ],
970970 torch_dtype = torch .float16 ,
971971 function_kwargs = {'bits' : 4 },
972972 support_flash_attn = True )
0 commit comments