File tree Expand file tree Collapse file tree 2 files changed +4
-4
lines changed
test/modules/model/LlamaDecoderLayerWithKVCacheAndFusedAttention Expand file tree Collapse file tree 2 files changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -126,8 +126,8 @@ def forward_adapter(
126126):
127127 # past_key_value is a dict with key_cache and value_cache.
128128 # It needs to be decomposed for tico and circle which does not know dict.
129- key_cache = past_key_value .key_cache
130- value_cache = past_key_value .value_cache
129+ key_cache = past_key_value .key_cache # type: ignore[union-attr]
130+ value_cache = past_key_value .value_cache # type: ignore[union-attr]
131131 return (
132132 torch .ops .circle .attention .llama (
133133 hidden_states ,
Original file line number Diff line number Diff line change @@ -128,8 +128,8 @@ def forward_adapter(
128128):
129129 # past_key_value is a dict with key_cache and value_cache.
130130 # It needs to be decomposed for tico and circle which does not know dict.
131- key_cache = past_key_value .key_cache
132- value_cache = past_key_value .value_cache
131+ key_cache = past_key_value .key_cache # type: ignore[union-attr]
132+ value_cache = past_key_value .value_cache # type: ignore[union-attr]
133133 return (
134134 torch .ops .circle .attention .llama (
135135 hidden_states ,
You can’t perform that action at this time.
0 commit comments