Skip to content

Commit 6858586

Browse files
authored
Update transformers to 4.41.2 (#583)
* updated transformers lib to 4.41.2 * fix all versions ranges * fix _seen_tokens * downgrade numpy * seq_len fix
1 parent e268c99 commit 6858586

File tree

5 files changed

+9
-11
lines changed

5 files changed

+9
-11
lines changed

setup.cfg

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ install_requires =
3737
accelerate>=0.27.2
3838
huggingface-hub>=0.11.1,<1.0.0
3939
tokenizers>=0.13.3
40-
transformers==4.38.2 # if you change this, please also change version assert in petals/__init__.py
40+
transformers==4.41.2 # if you change this, please also change version assert in petals/__init__.py
4141
speedtest-cli==2.1.3
4242
pydantic>=1.10,<2.0 # 2.0 is incompatible with hivemind yet
4343
hivemind==1.1.10.post2
@@ -50,6 +50,7 @@ install_requires =
5050
peft==0.5.0
5151
safetensors>=0.3.1
5252
Dijkstar>=2.6.0
53+
numpy<2
5354

5455
[options.extras_require]
5556
dev =

src/petals/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,8 +22,8 @@
2222

2323
if not os.getenv("PETALS_IGNORE_DEPENDENCY_VERSION"):
2424
assert (
25-
version.parse("4.38.2") <= version.parse(transformers.__version__) < version.parse("4.39.0")
26-
), "Please install a proper transformers version: pip install transformers>=4.37.1,<4.39.0"
25+
version.parse("4.41.2") <= version.parse(transformers.__version__) < version.parse("4.42.0")
26+
), "Please install a proper transformers version: pip install transformers>=4.41.2,<4.42.0"
2727

2828

2929
def _override_bfloat16_mode_default():

src/petals/client/remote_generation.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,20 +22,20 @@ class RemotePastKeyValues(Cache):
2222

2323
def __init__(self) -> None:
2424
super().__init__()
25-
self.seen_tokens = 0
25+
self._seen_tokens = 0
2626
self.hypo_ids: Optional[torch.LongTensor] = None
2727

2828
def __getitem__(self, _index: int) -> List[torch.Tensor]:
2929
return [DUMMY] # For compatibility with BloomForCausalLM.prepare_inputs_for_generation()
3030

3131
def get_seq_length(self, layer_idx: Optional[int] = 0) -> int:
32-
return self.seen_tokens
32+
return self._seen_tokens
3333

3434
def get_max_length(self) -> Optional[int]:
3535
return None
3636

3737
def update_seen(self, new_seen: int) -> None:
38-
self.seen_tokens += new_seen
38+
self._seen_tokens += new_seen
3939

4040
def reorder_cache(self, beam_idx):
4141
raise NotImplementedError("Beam search reordering is not implemented yet")

src/petals/models/bloom/model.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -131,7 +131,7 @@ def prepare_inputs_for_generation(
131131
if past_key_values is not None:
132132
if isinstance(past_key_values, Cache):
133133
cache_length = past_key_values.get_seq_length()
134-
past_length = past_key_values.seen_tokens
134+
past_length = past_key_values._seen_tokens
135135
max_cache_length = past_key_values.get_max_length()
136136
else:
137137
cache_length = past_length = past_key_values[0][0].shape[2]

src/petals/models/llama/block.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -87,10 +87,7 @@ def forward(
8787
key_states = key_states.view(bsz, q_len, self.num_key_value_heads, self.head_dim).transpose(1, 2)
8888
value_states = value_states.view(bsz, q_len, self.num_key_value_heads, self.head_dim).transpose(1, 2)
8989

90-
kv_seq_len = key_states.shape[-2]
91-
if past_key_value is not None:
92-
kv_seq_len += past_key_value[0].shape[-2]
93-
cos, sin = self.rotary_emb(value_states, position_ids, seq_len=kv_seq_len)
90+
cos, sin = self.rotary_emb(value_states, position_ids)
9491
cos, sin = cos.unsqueeze(1), sin.unsqueeze(1)
9592

9693
if q_len == 1 and torch.is_inference_mode_enabled() and hidden_states.device.type == "cuda":

0 commit comments

Comments
 (0)