Skip to content

Commit b25e845

Browse files
authored
Upgrade transformers to 4.42.3 for ipex models (#812)
* upgrade ipex model to transformers 4.42.3 * fix tests
1 parent 439d61f commit b25e845

File tree

4 files changed

+8
-7
lines changed

4 files changed

+8
-7
lines changed

.github/workflows/test_ipex.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ jobs:
2121
fail-fast: false
2222
matrix:
2323
python-version: [3.8, 3.9]
24-
transformers-version: [4.39.0, 4.41.2]
24+
transformers-version: [4.39.0, 4.42.3]
2525
os: [ubuntu-latest]
2626

2727
runs-on: ${{ matrix.os }}

optimum/exporters/ipex/model_patcher.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@
3434

3535
# Please also update in the setup.py and .github/workflows/test_ipex.yml if you change the transformers version
3636
_TRANSFORMERS_MIN_VERSION = "4.39.0"
37-
_TRANSFORMERS_MAX_VERSION = "4.41.2"
37+
_TRANSFORMERS_MAX_VERSION = "4.42.3"
3838

3939
_IPEX_EXPORTED_GENERATION_TASKS = ("text-generation",)
4040

optimum/intel/ipex/modeling_base.py

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -134,7 +134,6 @@ class IPEXModel(OptimizedModel):
134134
base_model_prefix = "ipex_model"
135135
main_input_name = "input_ids"
136136
output_name = "last_hidden_state"
137-
_supports_cache_class = False
138137

139138
def __init__(
140139
self,
@@ -428,6 +427,8 @@ def forward(
428427
class IPEXModelForCausalLM(IPEXModel, GenerationMixin):
429428
auto_model_class = AutoModelForCausalLM
430429
export_feature = "text-generation"
430+
_supports_cache_class = False
431+
_is_stateful = False
431432

432433
def __init__(
433434
self,
@@ -476,8 +477,8 @@ def __init__(
476477
else:
477478
self._reorder_cache = self.model_cls._reorder_cache.__get__(self)
478479

479-
if is_transformers_version(">=", "4.38.0") and model_type in {"llama", "phi", "persimmon"}:
480-
self.prepare_inputs_for_generation = _prepare_inputs_for_generation_for_llama
480+
if is_transformers_version(">=", "4.38.0") and model_type in {"llama", "phi", "persimmon", "mistral"}:
481+
self.prepare_inputs_for_generation = _ipex_prepare_inputs_for_generation
481482
else:
482483
self.prepare_inputs_for_generation = self.model_cls.prepare_inputs_for_generation.__get__(self)
483484

@@ -613,7 +614,7 @@ def generate(self, *args, **kwargs):
613614
return super().generate(*args, **kwargs)
614615

615616

616-
def _prepare_inputs_for_generation_for_llama(
617+
def _ipex_prepare_inputs_for_generation(
617618
input_ids, past_key_values=None, attention_mask=None, inputs_embeds=None, **kwargs
618619
):
619620
from transformers.cache_utils import Cache

setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,7 @@
6262
"neural-compressor": ["neural-compressor>=2.2.0", "onnxruntime<1.15.0", "accelerate"],
6363
"openvino": ["openvino>=2023.3", "nncf>=2.11.0", "openvino-tokenizers[transformers]"],
6464
"nncf": ["nncf>=2.11.0"],
65-
"ipex": ["intel-extension-for-pytorch", "transformers>=4.39.0,<=4.41.2"],
65+
"ipex": ["intel-extension-for-pytorch", "transformers>=4.39.0,<=4.42.3"],
6666
"diffusers": ["diffusers"],
6767
"quality": QUALITY_REQUIRE,
6868
"tests": TESTS_REQUIRE,

0 commit comments

Comments
 (0)