Skip to content

Commit 27f8462

Browse files
authored
[CI] Fix (#8793)
* fix ci
1 parent 326cf82 commit 27f8462

File tree

9 files changed

+98
-214
lines changed

9 files changed

+98
-214
lines changed

legacy/examples/benchmark/clue/mrc/run_cmrc2018.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -233,7 +233,7 @@ def run(args):
233233
set_seed(args)
234234

235235
train_examples, dev_examples, test_examples = load_dataset(
236-
"clue", "cmrc2018", split=["train", "validation", "test"]
236+
"clue", "cmrc2018", split=["train", "validation", "test"], trust_remote_code=True
237237
)
238238

239239
column_names = train_examples.column_names

legacy/examples/information_extraction/msra_ner/train.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -88,9 +88,9 @@ def do_train(args):
8888

8989
# Create dataset, tokenizer and dataloader.
9090
if args.dataset == "peoples_daily_ner":
91-
raw_datasets = load_dataset(args.dataset)
91+
raw_datasets = load_dataset(args.dataset, trust_remote_code=True)
9292
else:
93-
raw_datasets = load_dataset(args.dataset)
93+
raw_datasets = load_dataset(args.dataset, trust_remote_code=True)
9494

9595
AutoForTokenClassification, AutoTokenizer = MODEL_CLASSES[args.model_type]
9696
tokenizer = AutoTokenizer.from_pretrained(args.model_name_or_path)

legacy/examples/machine_reading_comprehension/SQuAD/run_squad.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -230,11 +230,11 @@ def run(args):
230230
tokenizer = tokenizer_class.from_pretrained(args.model_name_or_path)
231231

232232
if args.version_2_with_negative:
233-
train_examples = load_dataset("squad_v2", split="train")
234-
dev_examples = load_dataset("squad_v2", split="validation")
233+
train_examples = load_dataset("squad_v2", split="train", trust_remote_code=True)
234+
dev_examples = load_dataset("squad_v2", split="validation", trust_remote_code=True)
235235
else:
236-
train_examples = load_dataset("squad", split="train")
237-
dev_examples = load_dataset("squad", split="validation")
236+
train_examples = load_dataset("squad", split="train", trust_remote_code=True)
237+
dev_examples = load_dataset("squad", split="validation", trust_remote_code=True)
238238
set_seed(args)
239239
if rank == 0:
240240
if os.path.exists(args.model_name_or_path):

legacy/model_zoo/bert/run_glue_trainer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,7 @@ def do_train():
9292

9393
sentence1_key, sentence2_key = task_to_keys[model_args.task_name]
9494

95-
train_ds = load_dataset("glue", model_args.task_name, split="train")
95+
train_ds = load_dataset("glue", model_args.task_name, split="train", trust_remote_code=True)
9696
columns = train_ds.column_names
9797
is_regression = model_args.task_name == "stsb"
9898
label_list = None

legacy/model_zoo/bert/static_ipu/run_squad.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -208,9 +208,9 @@ def load_squad_dataset(args):
208208
tokenizer = BertTokenizer.from_pretrained("bert-base-uncased")
209209
features_fn = prepare_train_features if args.is_training else prepare_validation_features
210210
if args.is_training:
211-
raw_dataset = load_dataset("squad", split="train")
211+
raw_dataset = load_dataset("squad", split="train", trust_remote_code=True)
212212
else:
213-
raw_dataset = load_dataset("squad", split="validation")
213+
raw_dataset = load_dataset("squad", split="validation", trust_remote_code=True)
214214
column_names = raw_dataset.column_names
215215
dataset = raw_dataset.map(
216216
partial(features_fn, tokenizer=tokenizer, args=args), batched=True, remove_columns=column_names, num_proc=4

legacy/model_zoo/ernie-1.0/finetune/run_qa.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -87,7 +87,10 @@ def main():
8787

8888
dataset_config = data_args.dataset.split(" ")
8989
raw_datasets = load_dataset(
90-
dataset_config[0], None if len(dataset_config) <= 1 else dataset_config[1], cache_dir=model_args.cache_dir
90+
dataset_config[0],
91+
None if len(dataset_config) <= 1 else dataset_config[1],
92+
cache_dir=model_args.cache_dir,
93+
trust_remote_code=True,
9194
)
9295

9396
label_list = getattr(raw_datasets["train"], "label_list", None)

paddlenlp/datasets/dataset.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -51,9 +51,9 @@ def load_from_ppnlp(path, *args, **kwargs):
5151
new_path = os.path.split(path)[-1]
5252
new_path = os.path.join(ppnlp_path, "hf_datasets", new_path + ".py")
5353
if os.path.exists(new_path):
54-
return origin_load_dataset(new_path, *args, **kwargs)
54+
return origin_load_dataset(new_path, trust_remote_code=True, *args, **kwargs)
5555
else:
56-
return origin_load_dataset(path, *args, **kwargs)
56+
return origin_load_dataset(path, trust_remote_code=True, *args, **kwargs)
5757

5858

5959
datasets.load_dataset = load_from_ppnlp

0 commit comments

Comments
 (0)