|
16 | 16 |
|
17 | 17 | backbone_presets = { |
18 | 18 | "albert_base_en_uncased": { |
| 19 | + "metadata": { |
| 20 | + "description": ( |
| 21 | + "12-layer ALBERT model where all input is lowercased. " |
| 22 | + "Trained on English Wikipedia + BooksCorpus." |
| 23 | + ), |
| 24 | + "params": 11683584, |
| 25 | + "official_name": "ALBERT", |
| 26 | + "path": "albert", |
| 27 | + }, |
19 | 28 | "config": { |
20 | 29 | "vocabulary_size": 30000, |
21 | 30 | "num_layers": 12, |
|
30 | 39 | "num_segments": 2, |
31 | 40 | }, |
32 | 41 | "preprocessor_config": {}, |
33 | | - "description": ( |
34 | | - "Base size of ALBERT where all input is lowercased. " |
35 | | - "Trained on English Wikipedia + BooksCorpus." |
36 | | - ), |
37 | 42 | "weights_url": "https://storage.googleapis.com/keras-nlp/models/albert_base_en_uncased/v1/model.h5", |
38 | 43 | "weights_hash": "b83ccf3418dd84adc569324183176813", |
39 | 44 | "spm_proto_url": "https://storage.googleapis.com/keras-nlp/models/albert_base_en_uncased/v1/vocab.spm", |
40 | 45 | "spm_proto_hash": "73e62ff8e90f951f24c8b907913039a5", |
41 | 46 | }, |
42 | 47 | "albert_large_en_uncased": { |
| 48 | + "metadata": { |
| 49 | + "description": ( |
| 50 | + "24-layer ALBERT model where all input is lowercased. " |
| 51 | + "Trained on English Wikipedia + BooksCorpus." |
| 52 | + ), |
| 53 | + "params": 17683968, |
| 54 | + "official_name": "ALBERT", |
| 55 | + "path": "albert", |
| 56 | + }, |
43 | 57 | "config": { |
44 | 58 | "vocabulary_size": 30000, |
45 | 59 | "num_layers": 24, |
|
54 | 68 | "num_segments": 2, |
55 | 69 | }, |
56 | 70 | "preprocessor_config": {}, |
57 | | - "description": ( |
58 | | - "Large size of ALBERT where all input is lowercased. " |
59 | | - "Trained on English Wikipedia + BooksCorpus." |
60 | | - ), |
61 | 71 | "weights_url": "https://storage.googleapis.com/keras-nlp/models/albert_large_en_uncased/v1/model.h5", |
62 | 72 | "weights_hash": "c7754804efb245f06dd6e7ced32e082c", |
63 | 73 | "spm_proto_url": "https://storage.googleapis.com/keras-nlp/models/albert_large_en_uncased/v1/vocab.spm", |
64 | 74 | "spm_proto_hash": "73e62ff8e90f951f24c8b907913039a5", |
65 | 75 | }, |
66 | 76 | "albert_extra_large_en_uncased": { |
| 77 | + "metadata": { |
| 78 | + "description": ( |
| 79 | + "24-layer ALBERT model where all input is lowercased. " |
| 80 | + "Trained on English Wikipedia + BooksCorpus." |
| 81 | + ), |
| 82 | + "params": 58724864, |
| 83 | + "official_name": "ALBERT", |
| 84 | + "path": "albert", |
| 85 | + }, |
67 | 86 | "config": { |
68 | 87 | "vocabulary_size": 30000, |
69 | 88 | "num_layers": 24, |
|
78 | 97 | "num_segments": 2, |
79 | 98 | }, |
80 | 99 | "preprocessor_config": {}, |
81 | | - "description": ( |
82 | | - "Extra Large size of ALBERT where all input is lowercased. " |
83 | | - "Trained on English Wikipedia + BooksCorpus." |
84 | | - ), |
85 | 100 | "weights_url": "https://storage.googleapis.com/keras-nlp/models/albert_extra_large_en_uncased/v1/model.h5", |
86 | 101 | "weights_hash": "713209be8aadfa614fd79f18c9aeb16d", |
87 | 102 | "spm_proto_url": "https://storage.googleapis.com/keras-nlp/models/albert_extra_large_en_uncased/v1/vocab.spm", |
88 | 103 | "spm_proto_hash": "73e62ff8e90f951f24c8b907913039a5", |
89 | 104 | }, |
90 | 105 | "albert_extra_extra_large_en_uncased": { |
| 106 | + "metadata": { |
| 107 | + "description": ( |
| 108 | + "12-layer ALBERT model where all input is lowercased. " |
| 109 | + "Trained on English Wikipedia + BooksCorpus." |
| 110 | + ), |
| 111 | + "params": 222595584, |
| 112 | + "official_name": "ALBERT", |
| 113 | + "path": "albert", |
| 114 | + }, |
91 | 115 | "config": { |
92 | 116 | "vocabulary_size": 30000, |
93 | 117 | "num_layers": 12, |
|
102 | 126 | "num_segments": 2, |
103 | 127 | }, |
104 | 128 | "preprocessor_config": {}, |
105 | | - "description": ( |
106 | | - "Extra Large size of ALBERT where all input is lowercased. " |
107 | | - "Trained on English Wikipedia + BooksCorpus." |
108 | | - ), |
109 | 129 | "weights_url": "https://storage.googleapis.com/keras-nlp/models/albert_extra_extra_large_en_uncased/v1/model.h5", |
110 | 130 | "weights_hash": "a835177b692fb6a82139f94c66db2f22", |
111 | 131 | "spm_proto_url": "https://storage.googleapis.com/keras-nlp/models/albert_extra_extra_large_en_uncased/v1/vocab.spm", |
|
0 commit comments