Skip to content

Commit 59891e0

Browse files
NathanHBthomwolf
andauthored
fixing nanotron (#21)
* fix * Update src/lighteval/models/nanotron_model.py Co-authored-by: Thomas Wolf <[email protected]> --------- Co-authored-by: Thomas Wolf <[email protected]>
1 parent b409d95 commit 59891e0

File tree

2 files changed

+1
-2
lines changed

2 files changed

+1
-2
lines changed

src/lighteval/main_nanotron.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -99,7 +99,6 @@ def main(
9999
parallel_config=lighteval_config.parallelism,
100100
lighteval_config=lighteval_config,
101101
batch_size=lighteval_config.batch_size,
102-
cache_dir=os.environ.get("HF_HOME", "/scratch"),
103102
debug_one_layer_model=False,
104103
model_class=model_cls,
105104
env_config=env_config,

src/lighteval/models/nanotron_model.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1116,7 +1116,7 @@ def greedy_until(
11161116
# automatic (variable) batch size detection for vectorization
11171117
# pull longest context sample from request
11181118
for request in requests:
1119-
request.stop_sequence = request.stop_sequence + (self.tokenizer.eos_token,)
1119+
request.stop_sequence = list(request.stop_sequence) + [self.tokenizer.eos_token]
11201120
request.tokenized_context = self.tok_encode(request.context)
11211121

11221122
dataset = GenerativeTaskDatasetNanotron(requests=requests, dataset_splits=dataset_splits)

0 commit comments

Comments
 (0)