diff --git a/README.rst b/README.rst index 39148cf..7f98173 100644 --- a/README.rst +++ b/README.rst @@ -75,7 +75,7 @@ BERT in `bert-for-tf2` is implemented as a Keras layer. You could instantiate it .. code:: python - from bert import BertModelLayer + from bert.model import BertModelLayer l_bert = BertModelLayer(**BertModelLayer.Params( vocab_size = 16000, # embedding params @@ -102,11 +102,13 @@ or by using the ``bert_config.json`` from a `pre-trained google model`_: .. code:: python import bert + from bert.model import BertModelLayer + from bert import loader model_dir = ".models/uncased_L-12_H-768_A-12" - bert_params = bert.params_from_pretrained_ckpt(model_dir) - l_bert = bert.BertModelLayer.from_params(bert_params, name="bert") + bert_params = loader.params_from_pretrained_ckpt(model_dir) + l_bert = BertModelLayer.from_params(bert_params, name="bert") now you can use the BERT layer in your Keras model like this: @@ -141,10 +143,10 @@ can be loaded in the BERT layer: .. code:: python - import bert + from bert import loader bert_ckpt_file = os.path.join(model_dir, "bert_model.ckpt") - bert.load_stock_weights(l_bert, bert_ckpt_file) + loader.load_stock_weights(l_bert, bert_ckpt_file) **N.B.** see `tests/test_bert_activations.py`_ for a complete example. @@ -154,16 +156,19 @@ FAQ .. code:: python + from bert.model import BertModelLayer + from bert import loader + model_name = "uncased_L-12_H-768_A-12" - model_dir = bert.fetch_google_bert_model(model_name, ".models") + model_dir = loader.fetch_google_bert_model(model_name, ".models") model_ckpt = os.path.join(model_dir, "bert_model.ckpt") - bert_params = bert.params_from_pretrained_ckpt(model_dir) - l_bert = bert.BertModelLayer.from_params(bert_params, name="bert") + bert_params = loader.params_from_pretrained_ckpt(model_dir) + l_bert = BertModelLayer.from_params(bert_params, name="bert") # use in Keras Model here, and call model.build() - bert.load_bert_weights(l_bert, model_ckpt) # should be called after model.build() + loader.load_bert_weights(l_bert, model_ckpt) # should be called after model.build() 2. How to use ALBERT with the `google-research/ALBERT`_ pre-trained weights (fetching from TFHub)? @@ -171,14 +176,17 @@ see `tests/nonci/test_load_pretrained_weights.py