@@ -58,24 +58,24 @@ def enable_attention_prefill_prefix(model_name, model):
5858
5959def enable_attention_prefill_context (model_name , model ):
6060 if "llama" in args .model :
61- from src .ape_llama import enable_llama_attention_prefill_context
61+ from ape .ape_llama import enable_llama_attention_prefill_context
6262 enable_llama_attention_prefill_context (model )
6363 elif "mistral" in model_name :
64- from src .ape_mistral import enable_mistral_attention_prefill_context
64+ from ape .ape_mistral import enable_mistral_attention_prefill_context
6565 enable_mistral_attention_prefill_context (model )
6666 elif "gemma" in model_name :
67- from src .ape_gemma import enable_gemma_attention_prefill_context
67+ from ape .ape_gemma import enable_gemma_attention_prefill_context
6868 enable_gemma_attention_prefill_context (model )
6969
7070def enable_attention_prefill_query (model_name , model , temperature , scale ):
7171 if "llama" in args .model :
72- from src .ape_llama import enable_llama_attention_prefill_query
72+ from ape .ape_llama import enable_llama_attention_prefill_query
7373 enable_llama_attention_prefill_query (model , temperature , scale )
7474 elif "mistral" in model_name :
75- from src .ape_mistral import enable_mistral_attention_prefill_query
75+ from ape .ape_mistral import enable_mistral_attention_prefill_query
7676 enable_mistral_attention_prefill_query (model , temperature , scale )
7777 elif "gemma" in model_name :
78- from src .ape_gemma import enable_gemma_attention_prefill_query
78+ from ape .ape_gemma import enable_gemma_attention_prefill_query
7979 enable_gemma_attention_prefill_query (model , temperature , scale )
8080
8181def seed_everything (seed ):
0 commit comments