Skip to content

Commit 7d99dec

Browse files
committed
Fix occurrences of David Adelani & co-authors
1 parent 6d3e96c commit 7d99dec

15 files changed

+120
-94
lines changed

data/xml/2020.emnlp.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3118,7 +3118,7 @@
31183118
<paper id="204">
31193119
<title>Transfer Learning and Distant Supervision for Multilingual Transformer Models: A Study on <fixed-case>A</fixed-case>frican Languages</title>
31203120
<author><first>Michael A.</first><last>Hedderich</last></author>
3121-
<author><first>David</first><last>Adelani</last></author>
3121+
<author><first>David I.</first><last>Adelani</last></author>
31223122
<author><first>Dawei</first><last>Zhu</last></author>
31233123
<author><first>Jesujoba</first><last>Alabi</last></author>
31243124
<author><first>Udia</first><last>Markus</last></author>

data/xml/2020.lrec.xml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4123,9 +4123,9 @@
41234123
</paper>
41244124
<paper id="335">
41254125
<title>Massive vs. Curated Embeddings for Low-Resourced Languages: the Case of <fixed-case>Y</fixed-case>orùbá and <fixed-case>T</fixed-case>wi</title>
4126-
<author><first>Jesujoba</first><last>Alabi</last></author>
4126+
<author><first>Jesujoba O.</first><last>Alabi</last></author>
41274127
<author><first>Kwabena</first><last>Amponsah-Kaakyire</last></author>
4128-
<author><first>David</first><last>Adelani</last></author>
4128+
<author><first>David I.</first><last>Adelani</last></author>
41294129
<author><first>Cristina</first><last>España-Bonet</last></author>
41304130
<pages>2754–2762</pages>
41314131
<abstract>The success of several architectures to learn semantic representations from unannotated text and the availability of these kind of texts in online multilingual resources such as Wikipedia has facilitated the massive and automatic creation of resources for multiple languages. The evaluation of such resources is usually done for the high-resourced languages, where one has a smorgasbord of tasks and test sets to evaluate on. For low-resourced languages, the evaluation is more difficult and normally ignored, with the hope that the impressive capability of deep learning architectures to learn (multilingual) representations in the high-resourced setting holds in the low-resourced setting too. In this paper we focus on two African languages, Yorùbá and Twi, and compare the word embeddings obtained in this way, with word embeddings obtained from curated corpora and a language-dependent processing. We analyse the noise in the publicly available corpora, collect high quality and noisy data for the two languages and quantify the improvements that depend not only on the amount of data but on the quality too. We also use different architectures that learn word representations both from surface forms and characters to further exploit all the available information which showed to be important for these languages. For the evaluation, we manually translate the wordsim-353 word pairs dataset from English into Yorùbá and Twi. We extend the analysis to contextual word embeddings and evaluate multilingual BERT on a named entity recognition task. For this, we annotate with named entities the Global Voices corpus for Yorùbá. As output of the work, we provide corpora, embeddings and the test suits for both languages.</abstract>

data/xml/2021.emnlp.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10761,7 +10761,7 @@
1076110761
</paper>
1076210762
<paper id="684">
1076310763
<title>Preventing Author Profiling through Zero-Shot Multilingual Back-Translation</title>
10764-
<author><first>David</first><last>Adelani</last></author>
10764+
<author><first>David Ifeoluwa</first><last>Adelani</last></author>
1076510765
<author><first>Miaoran</first><last>Zhang</last></author>
1076610766
<author><first>Xiaoyu</first><last>Shen</last></author>
1076710767
<author><first>Ali</first><last>Davody</last></author>

data/xml/2021.mtsummit.xml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -76,9 +76,9 @@
7676
</paper>
7777
<paper id="6">
7878
<title>The Effect of Domain and Diacritics in <fixed-case>Y</fixed-case>oruba–<fixed-case>E</fixed-case>nglish Neural Machine Translation</title>
79-
<author><first>David</first><last>Adelani</last></author>
79+
<author><first>David Ifeoluwa</first><last>Adelani</last></author>
8080
<author><first>Dana</first><last>Ruiter</last></author>
81-
<author><first>Jesujoba</first><last>Alabi</last></author>
81+
<author><first>Jesujoba O.</first><last>Alabi</last></author>
8282
<author><first>Damilola</first><last>Adebonojo</last></author>
8383
<author><first>Adesina</first><last>Ayeni</last></author>
8484
<author><first>Mofe</first><last>Adeyemi</last></author>

data/xml/2022.emnlp.xml

Lines changed: 14 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -3912,15 +3912,15 @@
39123912
</paper>
39133913
<paper id="298">
39143914
<title><fixed-case>M</fixed-case>asakha<fixed-case>NER</fixed-case> 2.0: <fixed-case>A</fixed-case>frica-centric Transfer Learning for Named Entity Recognition</title>
3915-
<author><first>David</first><last>Adelani</last><affiliation>University College London</affiliation></author>
3915+
<author><first>David Ifeoluwa</first><last>Adelani</last><affiliation>University College London</affiliation></author>
39163916
<author><first>Graham</first><last>Neubig</last><affiliation>Carnegie Mellon University</affiliation></author>
39173917
<author><first>Sebastian</first><last>Ruder</last><affiliation>Google</affiliation></author>
39183918
<author><first>Shruti</first><last>Rijhwani</last><affiliation>Carnegie Mellon University</affiliation></author>
39193919
<author><first>Michael</first><last>Beukman</last><affiliation>University of the Witwatersrand</affiliation></author>
39203920
<author><first>Chester</first><last>Palen-Michel</last><affiliation>Brandeis University</affiliation></author>
39213921
<author><first>Constantine</first><last>Lignos</last><affiliation>Brandeis University</affiliation></author>
3922-
<author><first>Jesujoba</first><last>Alabi</last><affiliation>Saarland University</affiliation></author>
3923-
<author><first>Shamsuddeen</first><last>Muhammad</last><affiliation>Bayero University, Kano</affiliation></author>
3922+
<author><first>Jesujoba O.</first><last>Alabi</last><affiliation>Saarland University</affiliation></author>
3923+
<author><first>Shamsuddeen H.</first><last>Muhammad</last><affiliation>Bayero University, Kano</affiliation></author>
39243924
<author><first>Peter</first><last>Nabende</last><affiliation>Makerere University</affiliation></author>
39253925
<author><first>Cheikh M. Bamba</first><last>Dione</last><affiliation>University of Bergen</affiliation></author>
39263926
<author><first>Andiswa</first><last>Bukula</last><affiliation>SADiLaR</affiliation></author>
@@ -3942,11 +3942,21 @@
39423942
<author><first>Allahsera Auguste</first><last>Tapo</last><affiliation>Rochester Institute of Technology</affiliation></author>
39433943
<author><first>Tebogo</first><last>Macucwa</last><affiliation>University of Pretoria, Masakhane</affiliation></author>
39443944
<author><first>Vukosi</first><last>Marivate</last><affiliation>Department of Computer Science, University of Pretoria</affiliation></author>
3945-
<author><first>Mboning Tchiaze</first><last>Elvis</last><affiliation>NTeALan</affiliation></author>
3945+
<author><first>Elvis</first><last>Mboning</last><affiliation>NTeALan</affiliation></author>
39463946
<author><first>Tajuddeen</first><last>Gwadabe</last><affiliation>University of Chinese Academy of Science</affiliation></author>
39473947
<author><first>Tosin</first><last>Adewumi</last><affiliation>Luleå University of Technology</affiliation></author>
39483948
<author><first>Orevaoghene</first><last>Ahia</last><affiliation>University of Washington</affiliation></author>
39493949
<author><first>Joyce</first><last>Nakatumba-Nabende</last><affiliation>Makerere University</affiliation></author>
3950+
<author><first>Neo L.</first><last>Mokono</last></author>
3951+
<author><first>Ignatius</first><last>Ezeani</last></author>
3952+
<author><first>Chiamaka</first><last>Chukwuneke</last></author>
3953+
<author><first>Mofetoluwa</first><last>Adeyemi</last></author>
3954+
<author><first>Gilles Q.</first><last>Hacheme</last></author>
3955+
<author><first>Idris</first><last>Abdulmumim</last></author>
3956+
<author><first>Odunayo</first><last>Ogundepo</last></author>
3957+
<author><first>Oreen</first><last>Yousuf</last></author>
3958+
<author><first>Tatiana</first><last>Moteu Ngoli</last></author>
3959+
<author><first>Dietrich</first><last>Klakow</last></author>
39503960
<pages>4488-4508</pages>
39513961
<abstract>African languages are spoken by over a billion people, but they are under-represented in NLP research and development. Multiple challenges exist, including the limited availability of annotated training and evaluation datasets as well as the lack of understanding of which settings, languages, and recently proposed methods like cross-lingual transfer will be effective. In this paper, we aim to move towards solutions for these challenges, focusing on the task of named entity recognition (NER). We present the creation of the largest to-date human-annotated NER dataset for 20 African languages. We study the behaviour of state-of-the-art cross-lingual transfer methods in an Africa-centric setting, empirically demonstrating that the choice of source transfer language significantly affects performance. While much previous work defaults to using English as the source language, our results show that choosing the best transfer language improves zero-shot F1 scores by an average of 14% over 20 languages as compared to using English.</abstract>
39523962
<url hash="51dbe768">2022.emnlp-main.298</url>

data/xml/2022.findings.xml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -106,13 +106,13 @@
106106
</paper>
107107
<paper id="6">
108108
<title>Pre-Trained Multilingual Sequence-to-Sequence Models: A Hope for Low-Resource Language Translation?</title>
109-
<author><first>En-Shiun</first><last>Lee</last></author>
109+
<author><first>En-Shiun Annie</first><last>Lee</last></author>
110110
<author><first>Sarubi</first><last>Thillainathan</last></author>
111111
<author><first>Shravan</first><last>Nayak</last></author>
112112
<author><first>Surangika</first><last>Ranathunga</last></author>
113-
<author><first>David</first><last>Adelani</last></author>
113+
<author><first>David Ifeoluwa</first><last>Adelani</last></author>
114114
<author><first>Ruisi</first><last>Su</last></author>
115-
<author><first>Arya</first><last>McCarthy</last></author>
115+
<author><first>Arya D.</first><last>McCarthy</last></author>
116116
<pages>58-67</pages>
117117
<abstract>What can pre-trained multilingual sequence-to-sequence models like mBART contribute to translating low-resource languages? We conduct a thorough empirical experiment in 10 languages to ascertain this, considering five factors: (1) the amount of fine-tuning data, (2) the noise in the fine-tuning data, (3) the amount of pre-training data in the model, (4) the impact of domain mismatch, and (5) language typology. In addition to yielding several heuristics, the experiments form a framework for evaluating the data sensitivities of machine translation systems. While mBART is robust to domain differences, its translations for unseen and typologically distant languages remain below 3.0 BLEU. In answer to our title’s question, mBART is not a low-resource panacea; we therefore encourage shifting the emphasis from new models to new data.</abstract>
118118
<url hash="57ee5031">2022.findings-acl.6</url>

data/xml/2022.insights.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -114,7 +114,7 @@
114114
<author><first>Dawei</first><last>Zhu</last></author>
115115
<author><first>Michael A.</first><last>Hedderich</last></author>
116116
<author><first>Fangzhou</first><last>Zhai</last></author>
117-
<author><first>David</first><last>Adelani</last></author>
117+
<author><first>David Ifeoluwa</first><last>Adelani</last></author>
118118
<author><first>Dietrich</first><last>Klakow</last></author>
119119
<pages>62-67</pages>
120120
<abstract>Incorrect labels in training data occur when human annotators make mistakes or when the data is generated via weak or distant supervision. It has been shown that complex noise-handling techniques - by modeling, cleaning or filtering the noisy instances - are required to prevent models from fitting this label noise. However, we show in this work that, for text classification tasks with modern NLP models like BERT, over a variety of noise types, existing noise-handling methods do not always improve its performance, and may even deteriorate it, suggesting the need for further investigation. We also back our observations with a comprehensive analysis.</abstract>

data/xml/2022.naacl.xml

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -3574,8 +3574,8 @@
35743574
</paper>
35753575
<paper id="223">
35763576
<title>A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for <fixed-case>A</fixed-case>frican News Translation</title>
3577-
<author><first>David</first><last>Adelani</last></author>
3578-
<author><first>Jesujoba</first><last>Alabi</last></author>
3577+
<author><first>David Ifeoluwa</first><last>Adelani</last></author>
3578+
<author><first>Jesujoba Oluwadara</first><last>Alabi</last></author>
35793579
<author><first>Angela</first><last>Fan</last></author>
35803580
<author><first>Julia</first><last>Kreutzer</last></author>
35813581
<author><first>Xiaoyu</first><last>Shen</last></author>
@@ -3590,27 +3590,27 @@
35903590
<author><first>Chris</first><last>Emezue</last></author>
35913591
<author><first>Colin</first><last>Leong</last></author>
35923592
<author><first>Michael</first><last>Beukman</last></author>
3593-
<author><first>Shamsuddeen</first><last>Muhammad</last></author>
3594-
<author><first>Guyo</first><last>Jarso</last></author>
3593+
<author><first>Shamsuddeen H.</first><last>Muhammad</last></author>
3594+
<author><first>Guyo D.</first><last>Jarso</last></author>
35953595
<author><first>Oreen</first><last>Yousuf</last></author>
3596-
<author><first>Andre</first><last>Niyongabo Rubungo</last></author>
3596+
<author><first>Andre N.</first><last>Niyongabo Rubungo</last></author>
35973597
<author><first>Gilles</first><last>Hacheme</last></author>
35983598
<author><first>Eric Peter</first><last>Wairagala</last></author>
35993599
<author><first>Muhammad Umair</first><last>Nasir</last></author>
3600-
<author><first>Benjamin</first><last>Ajibade</last></author>
3601-
<author><first>Tunde</first><last>Ajayi</last></author>
3602-
<author><first>Yvonne</first><last>Gitau</last></author>
3600+
<author><first>Benjamin A.</first><last>Ajibade</last></author>
3601+
<author><first>Tunde Oluwaseyi</first><last>Ajayi</last></author>
3602+
<author><first>Yvonne Wambui</first><last>Gitau</last></author>
36033603
<author><first>Jade</first><last>Abbott</last></author>
36043604
<author><first>Mohamed</first><last>Ahmed</last></author>
36053605
<author><first>Millicent</first><last>Ochieng</last></author>
36063606
<author><first>Anuoluwapo</first><last>Aremu</last></author>
36073607
<author><first>Perez</first><last>Ogayo</last></author>
36083608
<author><first>Jonathan</first><last>Mukiibi</last></author>
36093609
<author><first>Fatoumata</first><last>Ouoba Kabore</last></author>
3610-
<author><first>Godson</first><last>Kalipe</last></author>
3610+
<author><first>Godson Koffi</first><last>Kalipe</last></author>
36113611
<author><first>Derguene</first><last>Mbaye</last></author>
36123612
<author><first>Allahsera Auguste</first><last>Tapo</last></author>
3613-
<author><first>Victoire</first><last>Memdjokam Koagne</last></author>
3613+
<author><first>Victoire M.</first><last>Memdjokam Koagne</last></author>
36143614
<author><first>Edwin</first><last>Munkoh-Buabeng</last></author>
36153615
<author><first>Valencia</first><last>Wagner</last></author>
36163616
<author><first>Idris</first><last>Abdulmumin</last></author>
@@ -7076,8 +7076,8 @@
70767076
<title><fixed-case>MCSE</fixed-case>: <fixed-case>M</fixed-case>ultimodal Contrastive Learning of Sentence Embeddings</title>
70777077
<author><first>Miaoran</first><last>Zhang</last></author>
70787078
<author><first>Marius</first><last>Mosbach</last></author>
7079-
<author><first>David</first><last>Adelani</last></author>
7080-
<author><first>Michael</first><last>Hedderich</last></author>
7079+
<author><first>David Ifeoluwa</first><last>Adelani</last></author>
7080+
<author><first>Michael A.</first><last>Hedderich</last></author>
70817081
<author><first>Dietrich</first><last>Klakow</last></author>
70827082
<pages>5959-5969</pages>
70837083
<abstract>Learning semantically meaningful sentence embeddings is an open problem in natural language processing. In this work, we propose a sentence embedding learning approach that exploits both visual and textual information via a multimodal contrastive objective. Through experiments on a variety of semantic textual similarity tasks, we demonstrate that our approach consistently improves the performance across various datasets and pre-trained encoders. In particular, combining a small amount of multimodal data with a large text-only corpus, we improve the state-of-the-art average Spearman’s correlation by 1.7%. By analyzing the properties of the textual embedding space, we show that our model excels in aligning semantically similar sentences, providing an explanation for its improved performance.</abstract>

data/xml/2022.wmt.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -982,7 +982,7 @@
982982
</paper>
983983
<paper id="72">
984984
<title>Findings of the <fixed-case>WMT</fixed-case>’22 Shared Task on Large-Scale Machine Translation Evaluation for <fixed-case>A</fixed-case>frican Languages</title>
985-
<author><first>David</first><last>Adelani</last><affiliation>University College London</affiliation></author>
985+
<author><first>David Ifeoluwa</first><last>Adelani</last><affiliation>University College London</affiliation></author>
986986
<author><first>Md Mahfuz Ibn</first><last>Alam</last><affiliation>George Mason University</affiliation></author>
987987
<author><first>Antonios</first><last>Anastasopoulos</last><affiliation>George Mason University</affiliation></author>
988988
<author><first>Akshita</first><last>Bhagia</last><affiliation>Ai2</affiliation></author>

data/xml/2023.c3nlp.xml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
<booktitle>Proceedings of the First Workshop on Cross-Cultural Considerations in NLP (C3NLP)</booktitle>
66
<editor><first>Sunipa</first><last>Dev</last></editor>
77
<editor><first>Vinodkumar</first><last>Prabhakaran</last></editor>
8-
<editor><first>David</first><last>Adelani</last></editor>
8+
<editor><first>David Ifeoluwa</first><last>Adelani</last></editor>
99
<editor><first>Dirk</first><last>Hovy</last></editor>
1010
<editor><first>Luciana</first><last>Benotti</last></editor>
1111
<publisher>Association for Computational Linguistics</publisher>
@@ -22,8 +22,8 @@
2222
<paper id="1">
2323
<title>Varepsilon kú mask: Integrating <fixed-case>Y</fixed-case>orùbá cultural greetings into machine translation</title>
2424
<author><first>Idris</first><last>Akinade</last><affiliation>University of Ibadan</affiliation></author>
25-
<author><first>Jesujoba</first><last>Alabi</last><affiliation>Saarland University</affiliation></author>
26-
<author><first>David</first><last>Adelani</last><affiliation>University College London</affiliation></author>
25+
<author><first>Jesujoba O.</first><last>Alabi</last><affiliation>Saarland University</affiliation></author>
26+
<author><first>David Ifeoluwa</first><last>Adelani</last><affiliation>University College London</affiliation></author>
2727
<author><first>Clement</first><last>Odoje</last><affiliation>University of Ibadan</affiliation></author>
2828
<author><first>Dietrich</first><last>Klakow</last><affiliation>Saarland University</affiliation></author>
2929
<pages>1-7</pages>

0 commit comments

Comments
 (0)