We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 8f43620 commit 7d9f385Copy full SHA for 7d9f385
test/modules/model/TinyLlamaWithFusedAttention/model.py
@@ -1,5 +1,5 @@
1
import torch
2
-
+from tico.config.v1 import CompileConfigV1
3
from tico.serialize.operators.adapters.llama_rmsnorm import patched_llama_rmsnorm
4
from tico.serialize.operators.adapters.onert.llama_attention import (
5
llama_attention_forward_adapter,
@@ -76,3 +76,6 @@ def get_example_inputs(self):
76
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
77
print(generated_text)
78
return captured_input, {}
79
+
80
+ def get_compile_config(self):
81
+ return CompileConfigV1(convert_single_batch_lhs_const_bmm_to_fc=True)
0 commit comments