Skip to content

Commit 5528afb

Browse files
authored
Added link to XLM-RoBERTa pre-trained model
1 parent c9c0d3f commit 5528afb

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -179,3 +179,4 @@ This resource was created in a semi-automatic way, by extracting the words and t
179179
- [Multilingual BERT](https://github.com/google-research/bert/blob/master/multilingual.md) - BERT (Bidirectional Encoder Representations from Transformers) is a model for generating contextual word representations. Multilingual cased model provided by Google supports 104 languages including Polish.
180180
- [Universal Sentence Encoder](https://tfhub.dev/google/universal-sentence-encoder-multilingual-large/1) - USE (Universal Sentence Encoder) generates sentence level langauge representations. Pre-trained multilingual model supports 16 langauges (Arabic, Chinese-simplified, Chinese-traditional, English, French, German, Italian, Japanese, Korean, Dutch, Polish, Portuguese, Spanish, Thai, Turkish, Russian).
181181
- [LASER Language-Agnostic SEntence Representations](https://github.com/facebookresearch/LASER) - A multilingual sentence encoder by Facebook Research, supporting 93 languages.
182+
- [XLM-RoBERTa](https://github.com/pytorch/fairseq/tree/master/examples/xlmr) - Cross lingual sentence encoder trained on 2.5 terabytes of data from CommonCrawl and Wikipedia. Supports 100 languages including Polish. See [Unsupervised Cross-lingual Representation Learning at Scale](https://arxiv.org/pdf/1911.02116.pdf) for details.

0 commit comments

Comments
 (0)