You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
help="Use sdpa_with_kv_cache custom op in LLava text model.",
298
299
)
300
+
parser.add_argument(
301
+
"--max-context-len",
302
+
default=768,
303
+
type=int,
304
+
help="Maximum context length for the text model.",
305
+
)
299
306
parser.add_argument(
300
307
"--max-seq-len",
301
308
default=768,
@@ -325,12 +332,13 @@ def main():
325
332
llm_config=create_llava_config_from_args(args)
326
333
327
334
logging.info(
328
-
f"Exporting Llava model to ExecuTorch with sdpa_with_kv_cache: {llm_config.model.use_sdpa_with_kv_cache}, max_seq_len: {llm_config.export.max_seq_length}"
335
+
f"Exporting Llava model to ExecuTorch with sdpa_with_kv_cache: {llm_config.model.use_sdpa_with_kv_cache}, max_seq_len: {llm_config.export.max_seq_length}, max_context_len: {llm_config.export.max_context_length}"
0 commit comments