You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-66Lines changed: 1 addition & 66 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -44,7 +44,7 @@ Like SimCSE, [ConGen: Unsupervised Control and Generalization Distillation For S
44
44
45
45
[SCT: An Efficient Self-Supervised Cross-View Training For Sentence Embedding](https://github.com/mrpeerat/SCT) is another unsupervised technique to train a sentence embedding model. It is very similar to ConGen in its knowledge distillation methodology, but also supports self-supervised training procedure without a teacher model. The original paper proposes back-translation as its data augmentation technique, but we implemented single-word deletion and found it to perform better than our backtranslated corpus. We used the [official SCT implementation](https://github.com/mrpeerat/SCT) which was written on top of the Sentence Transformers library.
46
46
47
-
## Models
47
+
## Pretrained Models
48
48
49
49
| Model | #params | Base/Student Model | Teacher Model | Train Dataset | Supervised |
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
253
-
author = "Reimers, Nils and Gurevych, Iryna",
254
-
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
255
-
month = "11",
256
-
year = "2019",
257
-
publisher = "Association for Computational Linguistics",
258
-
url = "https://arxiv.org/abs/1908.10084",
259
-
}
260
-
```
261
-
262
-
```bibtex
263
-
@inproceedings{gao2021simcse,
264
-
title={{SimCSE}: Simple Contrastive Learning of Sentence Embeddings},
265
-
author={Gao, Tianyu and Yao, Xingcheng and Chen, Danqi},
266
-
booktitle={Empirical Methods in Natural Language Processing (EMNLP)},
267
-
year={2021}
268
-
}
269
-
```
270
-
271
-
```bibtex
272
-
@inproceedings{limkonchotiwat-etal-2022-congen,
273
-
title = "{ConGen}: Unsupervised Control and Generalization Distillation For Sentence Representation",
274
-
author = "Limkonchotiwat, Peerat and
275
-
Ponwitayarat, Wuttikorn and
276
-
Lowphansirikul, Lalita and
277
-
Udomcharoenchaikit, Can and
278
-
Chuangsuwanich, Ekapol and
279
-
Nutanong, Sarana",
280
-
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2022",
281
-
year = "2022",
282
-
publisher = "Association for Computational Linguistics",
283
-
}
284
-
```
285
-
286
-
```bibtex
287
-
@article{10.1162/tacl_a_00620,
288
-
author = {Limkonchotiwat, Peerat and Ponwitayarat, Wuttikorn and Lowphansirikul, Lalita and Udomcharoenchaikit, Can and Chuangsuwanich, Ekapol and Nutanong, Sarana},
289
-
title = "{An Efficient Self-Supervised Cross-View Training For Sentence Embedding}",
290
-
journal = {Transactions of the Association for Computational Linguistics},
Copy file name to clipboardExpand all lines: docs/index.md
+1-66Lines changed: 1 addition & 66 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -44,7 +44,7 @@ Like SimCSE, [ConGen: Unsupervised Control and Generalization Distillation For S
44
44
45
45
[SCT: An Efficient Self-Supervised Cross-View Training For Sentence Embedding](https://github.com/mrpeerat/SCT) is another unsupervised technique to train a sentence embedding model. It is very similar to ConGen in its knowledge distillation methodology, but also supports self-supervised training procedure without a teacher model. The original paper proposes back-translation as its data augmentation technique, but we implemented single-word deletion and found it to perform better than our backtranslated corpus. We used the [official SCT implementation](https://github.com/mrpeerat/SCT) which was written on top of the Sentence Transformers library.
46
46
47
-
## Models
47
+
## Pretrained Models
48
48
49
49
| Model | #params | Base/Student Model | Teacher Model | Train Dataset | Supervised |
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
253
-
author = "Reimers, Nils and Gurevych, Iryna",
254
-
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
255
-
month = "11",
256
-
year = "2019",
257
-
publisher = "Association for Computational Linguistics",
258
-
url = "https://arxiv.org/abs/1908.10084",
259
-
}
260
-
```
261
-
262
-
```bibtex
263
-
@inproceedings{gao2021simcse,
264
-
title={{SimCSE}: Simple Contrastive Learning of Sentence Embeddings},
265
-
author={Gao, Tianyu and Yao, Xingcheng and Chen, Danqi},
266
-
booktitle={Empirical Methods in Natural Language Processing (EMNLP)},
267
-
year={2021}
268
-
}
269
-
```
270
-
271
-
```bibtex
272
-
@inproceedings{limkonchotiwat-etal-2022-congen,
273
-
title = "{ConGen}: Unsupervised Control and Generalization Distillation For Sentence Representation",
274
-
author = "Limkonchotiwat, Peerat and
275
-
Ponwitayarat, Wuttikorn and
276
-
Lowphansirikul, Lalita and
277
-
Udomcharoenchaikit, Can and
278
-
Chuangsuwanich, Ekapol and
279
-
Nutanong, Sarana",
280
-
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2022",
281
-
year = "2022",
282
-
publisher = "Association for Computational Linguistics",
283
-
}
284
-
```
285
-
286
-
```bibtex
287
-
@article{10.1162/tacl_a_00620,
288
-
author = {Limkonchotiwat, Peerat and Ponwitayarat, Wuttikorn and Lowphansirikul, Lalita and Udomcharoenchaikit, Can and Chuangsuwanich, Ekapol and Nutanong, Sarana},
289
-
title = "{An Efficient Self-Supervised Cross-View Training For Sentence Embedding}",
290
-
journal = {Transactions of the Association for Computational Linguistics},
0 commit comments