Skip to content

Commit abc4fc0

Browse files
saberkuntensorflower-gardener
authored andcommitted
Publish new albert checkpoints and put tf-hub modules in the same format in pretrained_models.md.
PiperOrigin-RevId: 397672201
1 parent 418903b commit abc4fc0

File tree

2 files changed

+26
-0
lines changed

2 files changed

+26
-0
lines changed

official/nlp/albert/README.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,8 @@
11
# ALBERT (ALBERT: A Lite BERT for Self-supervised Learning of Language Representations)
22

3+
**WARNING**: We are on the way to deprecate this directory.
4+
We will add documentation in `nlp/docs` to use the new code in `nlp/modeling`.
5+
36
The academic paper which describes ALBERT in detail and provides full results on
47
a number of tasks can be found here: https://arxiv.org/abs/1909.11942.
58

official/nlp/docs/pretrained_models.md

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -70,3 +70,26 @@ Model | Configuration | Training Data
7070
BERT-base talking heads + ggelu | uncased_L-12_H-768_A-12 | Wiki + Books | [talkheads_ggelu_base](https://tfhub.dev/tensorflow/talkheads_ggelu_bert_en_base/1) | BERT-base trained with [talking heads attention](https://arxiv.org/abs/2003.02436) and [gated GeLU](https://arxiv.org/abs/2002.05202).
7171
BERT-large talking heads + ggelu | uncased_L-24_H-1024_A-16 | Wiki + Books | [talkheads_ggelu_large](https://tfhub.dev/tensorflow/talkheads_ggelu_bert_en_large/1) | BERT-large trained with [talking heads attention](https://arxiv.org/abs/2003.02436) and [gated GeLU](https://arxiv.org/abs/2002.05202).
7272
LAMBERT-large uncased English | uncased_L-24_H-1024_A-16 | Wiki + Books | [lambert](https://tfhub.dev/tensorflow/lambert_en_uncased_L-24_H-1024_A-16/1) | BERT trained with LAMB and techniques from RoBERTa.
73+
74+
## ALBERT
75+
76+
The academic paper that describes ALBERT in detail and provides full results on
77+
a number of tasks can be found here: https://arxiv.org/abs/1909.11942.
78+
79+
We released both checkpoints and tf.hub modules as the pretrained models for
80+
fine-tuning. They are TF 2.x compatible and are converted from the ALBERT v2
81+
checkpoints released in the TF 1.x official ALBERT repository
82+
[google-research/albert](https://github.com/google-research/albert)
83+
in order to be consistent with the ALBERT paper.
84+
85+
Our current released checkpoints are exactly the same as the TF 1.x official
86+
ALBERT repository.
87+
88+
### Checkpoints
89+
90+
Model | Training Data | Checkpoint & Vocabulary | TF-HUB SavedModels
91+
---------------------------------------- | ------------: | ----------------------: | ------:
92+
ALBERT-base English | Wiki + Books | [`ALBERT Base`](https://storage.googleapis.com/tf_model_garden/nlp/albert/albert_base.tar.gz) | https://tfhub.dev/tensorflow/albert_en_base/3
93+
ALBERT-large English | Wiki + Books | [`ALBERT Large`](https://storage.googleapis.com/tf_model_garden/nlp/albert/albert_large.tar.gz) | https://tfhub.dev/tensorflow/albert_en_large/3
94+
ALBERT-xlarge English | Wiki + Books | [`ALBERT XLarge`](https://storage.googleapis.com/tf_model_garden/nlp/albert/albert_xlarge.tar.gz) | https://tfhub.dev/tensorflow/albert_en_xlarge/3
95+
ALBERT-xxlarge English | Wiki + Books | [`ALBERT XXLarge`](https://storage.googleapis.com/tf_model_garden/nlp/albert/albert_xxlarge.tar.gz) | https://tfhub.dev/tensorflow/albert_en_xxlarge/3

0 commit comments

Comments
 (0)