Skip to content

Commit d9f1b76

Browse files
authored
[GPT-3] fix gpt3 config (#7006)
* fix gpt3 config * update * update * update
1 parent 4a7665d commit d9f1b76

File tree

2 files changed

+8
-2
lines changed

2 files changed

+8
-2
lines changed

paddlenlp/transformers/gpt/configuration.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -89,7 +89,7 @@
8989
"vocab_size": 50304,
9090
"hidden_size": 5120,
9191
"num_hidden_layers": 40,
92-
"num_attention_heads": 128,
92+
"num_attention_heads": 40,
9393
"intermediate_size": 20480,
9494
"hidden_act": "gelu",
9595
"hidden_dropout_prob": 0.1,
@@ -102,7 +102,7 @@
102102
},
103103
"gpt3-6.7B-en": { # 6.7B
104104
"vocab_size": 50304,
105-
"hidden_size": 1024,
105+
"hidden_size": 4096,
106106
"num_hidden_layers": 32,
107107
"num_attention_heads": 32,
108108
"intermediate_size": 16384,

paddlenlp/transformers/gpt/tokenizer.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,10 @@
3535
PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES = {
3636
"gpt-cpm-large-cn": 1024,
3737
"gpt-cpm-small-cn-distill": 1024,
38+
"gpt3-175B-en": 1024,
39+
"gpt3-89B-en": 1024,
3840
"gpt3-13B-en": 1024,
41+
"gpt3-6.7B-en": 1024,
3942
"gpt3-1.3B-en": 1024,
4043
"gpt2-xl-en": 1024,
4144
"gpt2-large-en": 1024,
@@ -344,6 +347,7 @@ class GPTTokenizer(PretrainedTokenizer):
344347
"gpt3-175B-en": gpt_vocab_link,
345348
"gpt3-89B-en": gpt_vocab_link,
346349
"gpt3-13B-en": gpt_vocab_link,
350+
"gpt3-6.7B-en": gpt_vocab_link,
347351
"gpt3-1.3B-en": gpt_vocab_link,
348352
"gpt2-xl-en": gpt_vocab_link,
349353
"gpt2-large-en": gpt_vocab_link,
@@ -355,6 +359,7 @@ class GPTTokenizer(PretrainedTokenizer):
355359
"gpt3-175B-en": gpt_merges_link,
356360
"gpt3-89B-en": gpt_merges_link,
357361
"gpt3-13B-en": gpt_merges_link,
362+
"gpt3-6.7B-en": gpt_merges_link,
358363
"gpt3-1.3B-en": gpt_merges_link,
359364
"gpt2-xl-en": gpt_merges_link,
360365
"gpt2-large-en": gpt_merges_link,
@@ -367,6 +372,7 @@ class GPTTokenizer(PretrainedTokenizer):
367372
"gpt3-175B-en": {},
368373
"gpt3-89B-en": {},
369374
"gpt3-13B-en": {},
375+
"gpt3-6.7B-en": {},
370376
"gpt3-1.3B-en": {},
371377
"gpt2-xl-en": {},
372378
"gpt2-large-en": {},

0 commit comments

Comments
 (0)