davidpomerenke commited on
Commit
34983b9
·
verified ·
1 Parent(s): 5519510

Upload from GitHub Actions: Add citations

Browse files
Files changed (1) hide show
  1. bibliography.bib +138 -4
bibliography.bib CHANGED
@@ -8,6 +8,23 @@
8
  file = {/Users/david/Zotero/storage/AU759RXC/nllb-200-high-quality-machine-translation.html}
9
  }
10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  @misc{adelaniSIB200SimpleInclusive2024,
12
  title = {{{SIB-200}}: {{A Simple}}, {{Inclusive}}, and {{Big Evaluation Dataset}} for {{Topic Classification}} in 200+ {{Languages}} and {{Dialects}}},
13
  shorttitle = {{{SIB-200}}},
@@ -47,6 +64,23 @@
47
  file = {/Users/david/Zotero/storage/Q8A3WGUG/Ahuja et al. - 2024 - MEGAVERSE Benchmarking Large Language Models Across Languages, Modalities, Models and Tasks.pdf;/Users/david/Zotero/storage/ZHA8FR3E/2311.html}
48
  }
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  @misc{bapnaBuildingMachineTranslation2022,
51
  title = {Building {{Machine Translation Systems}} for the {{Next Thousand Languages}}},
52
  author = {Bapna, Ankur and Caswell, Isaac and Kreutzer, Julia and Firat, Orhan and van Esch, Daan and Siddhant, Aditya and Niu, Mengmeng and Baljekar, Pallavi and Garcia, Xavier and Macherey, Wolfgang and Breiner, Theresa and Axelrod, Vera and Riesa, Jason and Cao, Yuan and Chen, Mia Xu and Macherey, Klaus and Krikun, Maxim and Wang, Pidong and Gutkin, Alexander and Shah, Apurva and Huang, Yanping and Chen, Zhifeng and Wu, Yonghui and Hughes, Macduff},
@@ -63,7 +97,7 @@
63
  file = {/Users/david/Zotero/storage/YCW6FWWE/Bapna et al. - 2022 - Building Machine Translation Systems for the Next Thousand Languages.pdf;/Users/david/Zotero/storage/EL7PA6YJ/2205.html}
64
  }
65
 
66
- @article{costa-jussaScalingNeuralMachine2024,
67
  title = {Scaling Neural Machine Translation to 200 Languages},
68
  author = {{Costa-juss{\`a}}, Marta R. and Cross, James and {\c C}elebi, Onur and Elbayad, Maha and Heafield, Kenneth and Heffernan, Kevin and Kalbassi, Elahe and Lam, Janice and Licht, Daniel and Maillard, Jean and Sun, Anna and Wang, Skyler and Wenzek, Guillaume and Youngblood, Al and Akula, Bapi and Barrault, Loic and Gonzalez, Gabriel Mejia and Hansanti, Prangthip and Hoffman, John and Jarrett, Semarley and Sadagopan, Kaushik Ram and Rowe, Dirk and Spruit, Shannon and Tran, Chau and Andrews, Pierre and Ayan, Necip Fazil and Bhosale, Shruti and Edunov, Sergey and Fan, Angela and Gao, Cynthia and Goswami, Vedanuj and Guzm{\'a}n, Francisco and Koehn, Philipp and Mourachko, Alexandre and Ropers, Christophe and Saleem, Safiyyah and Schwenk, Holger and Wang, Jeff and {NLLB Team}},
69
  year = {2024},
@@ -75,12 +109,11 @@
75
  publisher = {Nature Publishing Group},
76
  issn = {1476-4687},
77
  doi = {10.1038/s41586-024-07335-x},
78
- urldate = {2024-11-02},
79
  abstract = {The development of neural techniques has opened up new avenues for research in machine translation. Today, neural machine translation (NMT) systems can leverage highly multilingual capacities and even perform zero-shot translation, delivering promising results in terms of language coverage and quality. However, scaling quality NMT requires large volumes of parallel bilingual data, which are not equally available for the 7,000+ languages in the world1. Focusing on improving the translation qualities of a relatively small group of high-resource languages comes at the expense of directing research attention to low-resource languages, exacerbating digital inequities in the long run. To break this pattern, here we introduce No Language Left Behind---a single massively multilingual model that leverages transfer learning across languages. We developed a conditional computational model based on the Sparsely Gated Mixture of Experts architecture2--7, which we trained on data obtained with new mining techniques tailored for low-resource languages. Furthermore, we devised multiple architectural and training improvements to counteract overfitting while training on thousands of tasks. We evaluated the performance of our model over 40,000 translation directions using tools created specifically for this purpose---an automatic benchmark (FLORES-200), a human evaluation metric (XSTS) and a toxicity detector that covers every language in our model. Compared with the previous state-of-the-art models, our model achieves an average of 44\% improvement in translation quality as measured by BLEU. By demonstrating how to scale NMT to 200 languages and making all contributions in this effort freely available for non-commercial use, our work lays important groundwork for the development of a universal translation system.},
80
  copyright = {2024 Meta},
81
  langid = {english},
82
- keywords = {dataset,model,n=200},
83
- file = {/Users/david/Zotero/storage/R7AB5BR3/Costa-jussà et al. - 2024 - Scaling neural machine translation to 200 languages.pdf}
84
  }
85
 
86
  @book{eberhard2024ethnologue,
@@ -107,6 +140,22 @@
107
  file = {/Users/david/Zotero/storage/E286EDPU/Federmann et al. - 2022 - NTREX-128 – News Test References for MT Evaluation of 128 Languages.pdf}
108
  }
109
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
110
  @misc{gurgurovLowREmRepositoryWord2024,
111
  title = {{{LowREm}}: {{A Repository}} of {{Word Embeddings}} for 87 {{Low-Resource Languages Enhanced}} with {{Multilingual Graph Knowledge}}},
112
  shorttitle = {{{LowREm}}},
@@ -154,6 +203,23 @@
154
  file = {/Users/david/Zotero/storage/SMRV9HE2/language.html}
155
  }
156
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
157
  @misc{Lanfrica,
158
  title = {Lanfrica},
159
  urldate = {2024-11-02},
@@ -180,6 +246,15 @@
180
  file = {/Users/david/Zotero/storage/6BYYZ7V2/Maillard et al. - 2023 - Small Data, Big Impact Leveraging Minimal Data for Effective Machine Translation.pdf}
181
  }
182
 
 
 
 
 
 
 
 
 
 
183
  @inproceedings{nekotoParticipatoryResearchLowresourced2020,
184
  title = {Participatory {{Research}} for {{Low-resourced Machine Translation}}: {{A Case Study}} in {{African Languages}}},
185
  shorttitle = {Participatory {{Research}} for {{Low-resourced Machine Translation}}},
@@ -198,6 +273,16 @@
198
  file = {/Users/david/Zotero/storage/JJ2S8CT3/Nekoto et al. - 2020 - Participatory Research for Low-resourced Machine Translation A Case Study in African Languages.pdf}
199
  }
200
 
 
 
 
 
 
 
 
 
 
 
201
  @misc{OpenlanguagedataFlores2024,
202
  title = {Openlanguagedata/Flores},
203
  year = {2024},
@@ -243,6 +328,22 @@
243
  file = {/Users/david/Zotero/storage/BWFBTAZ9/Robinson et al. - 2023 - ChatGPT MT Competitive for High- (but Not Low-) Resource Languages.pdf}
244
  }
245
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
246
  @misc{siminyuAI4DAfricanLanguage2021,
247
  title = {{{AI4D}} -- {{African Language Program}}},
248
  author = {Siminyu, Kathleen and Kalipe, Godson and Orlic, Davor and Abbott, Jade and Marivate, Vukosi and Freshia, Sackey and Sibal, Prateek and Neupane, Bhanu and Adelani, David I. and Taylor, Amelia and ALI, Jamiil Toure and Degila, Kevin and Balogoun, Momboladji and DIOP, Thierno Ibrahima and David, Davis and Fourati, Chayma and Haddad, Hatem and Naski, Malek},
@@ -259,6 +360,23 @@
259
  file = {/Users/david/Zotero/storage/VU6IFENR/Siminyu et al. - 2021 - AI4D -- African Language Program.pdf;/Users/david/Zotero/storage/7TV2PS8J/2104.html}
260
  }
261
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
262
  @misc{Tatoeba,
263
  title = {Tatoeba},
264
  urldate = {2024-11-03},
@@ -281,3 +399,19 @@
281
  keywords = {dataset,model,n=200},
282
  file = {/Users/david/Zotero/storage/GHWEGFFS/Team et al. - 2022 - No Language Left Behind Scaling Human-Centered Machine Translation.pdf;/Users/david/Zotero/storage/SZK3CP9C/2207.html}
283
  }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  file = {/Users/david/Zotero/storage/AU759RXC/nllb-200-high-quality-machine-translation.html}
9
  }
10
 
11
+ @misc{adelaniIrokoBenchNewBenchmark2025,
12
+ title = {{{IrokoBench}}: {{A New Benchmark}} for {{African Languages}} in the {{Age}} of {{Large Language Models}}},
13
+ shorttitle = {{{IrokoBench}}},
14
+ author = {Adelani, David Ifeoluwa and Ojo, Jessica and Azime, Israel Abebe and Zhuang, Jian Yun and Alabi, Jesujoba O. and He, Xuanli and Ochieng, Millicent and Hooker, Sara and Bukula, Andiswa and Lee, En-Shiun Annie and Chukwuneke, Chiamaka and Buzaaba, Happy and Sibanda, Blessing and Kalipe, Godson and Mukiibi, Jonathan and Kabongo, Salomon and Yuehgoh, Foutse and Setaka, Mmasibidi and Ndolela, Lolwethu and Odu, Nkiruka and Mabuya, Rooweither and Muhammad, Shamsuddeen Hassan and Osei, Salomey and Samb, Sokhar and Guge, Tadesse Kebede and Sherman, Tombekai Vangoni and Stenetorp, Pontus},
15
+ year = {2025},
16
+ month = jan,
17
+ number = {arXiv:2406.03368},
18
+ eprint = {2406.03368},
19
+ primaryclass = {cs},
20
+ publisher = {arXiv},
21
+ doi = {10.48550/arXiv.2406.03368},
22
+ urldate = {2025-05-28},
23
+ abstract = {Despite the widespread adoption of Large language models (LLMs), their remarkable capabilities remain limited to a few high-resource languages. Additionally, many low-resource languages ({\textbackslash}eg African languages) are often evaluated only on basic text classification tasks due to the lack of appropriate or comprehensive benchmarks outside of high-resource languages. In this paper, we introduce IrokoBench -- a human-translated benchmark dataset for 17 typologically-diverse low-resource African languages covering three tasks: natural language inference{\textasciitilde}(AfriXNLI), mathematical reasoning{\textasciitilde}(AfriMGSM), and multi-choice knowledge-based question answering{\textasciitilde}(AfriMMLU). We use IrokoBench to evaluate zero-shot, few-shot, and translate-test settings{\textasciitilde}(where test sets are translated into English) across 10 open and six proprietary LLMs. Our evaluation reveals a significant performance gap between high-resource languages{\textasciitilde}(such as English and French) and low-resource African languages. We observe a significant performance gap between open and proprietary models, with the highest performing open model, Gemma 2 27B only at 63{\textbackslash}\% of the best-performing proprietary model GPT-4o performance. In addition, machine translating the test set to English before evaluation helped to close the gap for larger models that are English-centric, such as Gemma 2 27B and LLaMa 3.1 70B. These findings suggest that more efforts are needed to develop and adapt LLMs for African languages.},
24
+ archiveprefix = {arXiv},
25
+ file = {/Users/david/Zotero/storage/9X39YMAR/Adelani et al. - 2025 - IrokoBench A New Benchmark for African Languages in the Age of Large Language Models.pdf;/Users/david/Zotero/storage/L69AFMRS/2406.html}
26
+ }
27
+
28
  @misc{adelaniSIB200SimpleInclusive2024,
29
  title = {{{SIB-200}}: {{A Simple}}, {{Inclusive}}, and {{Big Evaluation Dataset}} for {{Topic Classification}} in 200+ {{Languages}} and {{Dialects}}},
30
  shorttitle = {{{SIB-200}}},
 
64
  file = {/Users/david/Zotero/storage/Q8A3WGUG/Ahuja et al. - 2024 - MEGAVERSE Benchmarking Large Language Models Across Languages, Modalities, Models and Tasks.pdf;/Users/david/Zotero/storage/ZHA8FR3E/2311.html}
65
  }
66
 
67
+ @misc{aliExpandingFLORESBenchmark2024,
68
+ title = {Expanding {{FLORES}}+ {{Benchmark}} for More {{Low-Resource Settings}}: {{Portuguese-Emakhuwa Machine Translation Evaluation}}},
69
+ shorttitle = {Expanding {{FLORES}}+ {{Benchmark}} for More {{Low-Resource Settings}}},
70
+ author = {Ali, Felermino D. M. Antonio and Cardoso, Henrique Lopes and {Sousa-Silva}, Rui},
71
+ year = {2024},
72
+ month = aug,
73
+ number = {arXiv:2408.11457},
74
+ eprint = {2408.11457},
75
+ primaryclass = {cs},
76
+ publisher = {arXiv},
77
+ doi = {10.48550/arXiv.2408.11457},
78
+ urldate = {2025-05-28},
79
+ abstract = {As part of the Open Language Data Initiative shared tasks, we have expanded the FLORES+ evaluation set to include Emakhuwa, a low-resource language widely spoken in Mozambique. We translated the dev and devtest sets from Portuguese into Emakhuwa, and we detail the translation process and quality assurance measures used. Our methodology involved various quality checks, including post-editing and adequacy assessments. The resulting datasets consist of multiple reference sentences for each source. We present baseline results from training a Neural Machine Translation system and fine-tuning existing multilingual translation models. Our findings suggest that spelling inconsistencies remain a challenge in Emakhuwa. Additionally, the baseline models underperformed on this evaluation set, underscoring the necessity for further research to enhance machine translation quality for Emakhuwa. The data is publicly available at https://huggingface.co/datasets/LIACC/Emakhuwa-FLORES.},
80
+ archiveprefix = {arXiv},
81
+ file = {/Users/david/Zotero/storage/THELPPXB/Ali et al. - 2024 - Expanding FLORES+ Benchmark for more Low-Resource Settings Portuguese-Emakhuwa Machine Translation.pdf;/Users/david/Zotero/storage/NWMT7ZHL/2408.html}
82
+ }
83
+
84
  @misc{bapnaBuildingMachineTranslation2022,
85
  title = {Building {{Machine Translation Systems}} for the {{Next Thousand Languages}}},
86
  author = {Bapna, Ankur and Caswell, Isaac and Kreutzer, Julia and Firat, Orhan and van Esch, Daan and Siddhant, Aditya and Niu, Mengmeng and Baljekar, Pallavi and Garcia, Xavier and Macherey, Wolfgang and Breiner, Theresa and Axelrod, Vera and Riesa, Jason and Cao, Yuan and Chen, Mia Xu and Macherey, Klaus and Krikun, Maxim and Wang, Pidong and Gutkin, Alexander and Shah, Apurva and Huang, Yanping and Chen, Zhifeng and Wu, Yonghui and Hughes, Macduff},
 
97
  file = {/Users/david/Zotero/storage/YCW6FWWE/Bapna et al. - 2022 - Building Machine Translation Systems for the Next Thousand Languages.pdf;/Users/david/Zotero/storage/EL7PA6YJ/2205.html}
98
  }
99
 
100
+ @article{costa-jussaScalingNeuralMachine2024a,
101
  title = {Scaling Neural Machine Translation to 200 Languages},
102
  author = {{Costa-juss{\`a}}, Marta R. and Cross, James and {\c C}elebi, Onur and Elbayad, Maha and Heafield, Kenneth and Heffernan, Kevin and Kalbassi, Elahe and Lam, Janice and Licht, Daniel and Maillard, Jean and Sun, Anna and Wang, Skyler and Wenzek, Guillaume and Youngblood, Al and Akula, Bapi and Barrault, Loic and Gonzalez, Gabriel Mejia and Hansanti, Prangthip and Hoffman, John and Jarrett, Semarley and Sadagopan, Kaushik Ram and Rowe, Dirk and Spruit, Shannon and Tran, Chau and Andrews, Pierre and Ayan, Necip Fazil and Bhosale, Shruti and Edunov, Sergey and Fan, Angela and Gao, Cynthia and Goswami, Vedanuj and Guzm{\'a}n, Francisco and Koehn, Philipp and Mourachko, Alexandre and Ropers, Christophe and Saleem, Safiyyah and Schwenk, Holger and Wang, Jeff and {NLLB Team}},
103
  year = {2024},
 
109
  publisher = {Nature Publishing Group},
110
  issn = {1476-4687},
111
  doi = {10.1038/s41586-024-07335-x},
112
+ urldate = {2025-05-28},
113
  abstract = {The development of neural techniques has opened up new avenues for research in machine translation. Today, neural machine translation (NMT) systems can leverage highly multilingual capacities and even perform zero-shot translation, delivering promising results in terms of language coverage and quality. However, scaling quality NMT requires large volumes of parallel bilingual data, which are not equally available for the 7,000+ languages in the world1. Focusing on improving the translation qualities of a relatively small group of high-resource languages comes at the expense of directing research attention to low-resource languages, exacerbating digital inequities in the long run. To break this pattern, here we introduce No Language Left Behind---a single massively multilingual model that leverages transfer learning across languages. We developed a conditional computational model based on the Sparsely Gated Mixture of Experts architecture2--7, which we trained on data obtained with new mining techniques tailored for low-resource languages. Furthermore, we devised multiple architectural and training improvements to counteract overfitting while training on thousands of tasks. We evaluated the performance of our model over 40,000 translation directions using tools created specifically for this purpose---an automatic benchmark (FLORES-200), a human evaluation metric (XSTS) and a toxicity detector that covers every language in our model. Compared with the previous state-of-the-art models, our model achieves an average of 44\% improvement in translation quality as measured by BLEU. By demonstrating how to scale NMT to 200 languages and making all contributions in this effort freely available for non-commercial use, our work lays important groundwork for the development of a universal translation system.},
114
  copyright = {2024 Meta},
115
  langid = {english},
116
+ file = {/Users/david/Zotero/storage/6RWFDDH5/Costa-jussà et al. - 2024 - Scaling neural machine translation to 200 languages.pdf}
 
117
  }
118
 
119
  @book{eberhard2024ethnologue,
 
140
  file = {/Users/david/Zotero/storage/E286EDPU/Federmann et al. - 2022 - NTREX-128 – News Test References for MT Evaluation of 128 Languages.pdf}
141
  }
142
 
143
+ @inproceedings{gordeevFLORESTranslationMachine2024,
144
+ title = {{{FLORES}}+ {{Translation}} and {{Machine Translation Evaluation}} for the {{Erzya Language}}},
145
+ booktitle = {Proceedings of the {{Ninth Conference}} on {{Machine Translation}}},
146
+ author = {Gordeev, Isai and Kuldin, Sergey and Dale, David},
147
+ editor = {Haddow, Barry and Kocmi, Tom and Koehn, Philipp and Monz, Christof},
148
+ year = {2024},
149
+ month = nov,
150
+ pages = {614--623},
151
+ publisher = {Association for Computational Linguistics},
152
+ address = {Miami, Florida, USA},
153
+ doi = {10.18653/v1/2024.wmt-1.49},
154
+ urldate = {2025-05-28},
155
+ abstract = {This paper introduces a translation of the FLORES+ dataset into the endangered Erzya language, with the goal of evaluating machine translation between this language and any of the other 200 languages already included into FLORES+. This translation was carried out as a part of the Open Language Data shared task at WMT24. We also present a benchmark of existing translation models bases on this dataset and a new translation model that achieves the state-of-the-art quality of translation into Erzya from Russian and English.},
156
+ file = {/Users/david/Zotero/storage/KHZ84I94/Gordeev et al. - 2024 - FLORES+ Translation and Machine Translation Evaluation for the Erzya Language.pdf}
157
+ }
158
+
159
  @misc{gurgurovLowREmRepositoryWord2024,
160
  title = {{{LowREm}}: {{A Repository}} of {{Word Embeddings}} for 87 {{Low-Resource Languages Enhanced}} with {{Multilingual Graph Knowledge}}},
161
  shorttitle = {{{LowREm}}},
 
203
  file = {/Users/david/Zotero/storage/SMRV9HE2/language.html}
204
  }
205
 
206
+ @misc{laiOkapiInstructiontunedLarge2023,
207
+ title = {Okapi: {{Instruction-tuned Large Language Models}} in {{Multiple Languages}} with {{Reinforcement Learning}} from {{Human Feedback}}},
208
+ shorttitle = {Okapi},
209
+ author = {Lai, Viet Dac and Nguyen, Chien Van and Ngo, Nghia Trung and Nguyen, Thuat and Dernoncourt, Franck and Rossi, Ryan A. and Nguyen, Thien Huu},
210
+ year = {2023},
211
+ month = aug,
212
+ number = {arXiv:2307.16039},
213
+ eprint = {2307.16039},
214
+ primaryclass = {cs},
215
+ publisher = {arXiv},
216
+ doi = {10.48550/arXiv.2307.16039},
217
+ urldate = {2025-05-28},
218
+ abstract = {A key technology for the development of large language models (LLMs) involves instruction tuning that helps align the models' responses with human expectations to realize impressive learning abilities. Two major approaches for instruction tuning characterize supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF), which are currently applied to produce the best commercial LLMs (e.g., ChatGPT). To improve the accessibility of LLMs for research and development efforts, various instruction-tuned open-source LLMs have also been introduced recently, e.g., Alpaca, Vicuna, to name a few. However, existing open-source LLMs have only been instruction-tuned for English and a few popular languages, thus hindering their impacts and accessibility to many other languages in the world. Among a few very recent work to explore instruction tuning for LLMs in multiple languages, SFT has been used as the only approach to instruction-tune LLMs for multiple languages. This has left a significant gap for fine-tuned LLMs based on RLHF in diverse languages and raised important questions on how RLHF can boost the performance of multilingual instruction tuning. To overcome this issue, we present Okapi, the first system with instruction-tuned LLMs based on RLHF for multiple languages. Okapi introduces instruction and response-ranked data in 26 diverse languages to facilitate the experiments and development of future multilingual LLM research. We also present benchmark datasets to enable the evaluation of generative LLMs in multiple languages. Our experiments demonstrate the advantages of RLHF for multilingual instruction over SFT for different base models and datasets. Our framework and resources are released at https://github.com/nlp-uoregon/Okapi.},
219
+ archiveprefix = {arXiv},
220
+ file = {/Users/david/Zotero/storage/2GB79E4F/Lai et al. - 2023 - Okapi Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning fro.pdf;/Users/david/Zotero/storage/VZXBL7F2/2307.html}
221
+ }
222
+
223
  @misc{Lanfrica,
224
  title = {Lanfrica},
225
  urldate = {2024-11-02},
 
246
  file = {/Users/david/Zotero/storage/6BYYZ7V2/Maillard et al. - 2023 - Small Data, Big Impact Leveraging Minimal Data for Effective Machine Translation.pdf}
247
  }
248
 
249
+ @misc{MasakhaneAfrimmluDatasets2024,
250
+ title = {Masakhane/Afrimmlu {$\cdot$} {{Datasets}} at {{Hugging Face}}},
251
+ year = {2024},
252
+ month = jun,
253
+ urldate = {2025-05-28},
254
+ abstract = {We're on a journey to advance and democratize artificial intelligence through open source and open science.},
255
+ howpublished = {https://huggingface.co/datasets/masakhane/afrimmlu}
256
+ }
257
+
258
  @inproceedings{nekotoParticipatoryResearchLowresourced2020,
259
  title = {Participatory {{Research}} for {{Low-resourced Machine Translation}}: {{A Case Study}} in {{African Languages}}},
260
  shorttitle = {Participatory {{Research}} for {{Low-resourced Machine Translation}}},
 
273
  file = {/Users/david/Zotero/storage/JJ2S8CT3/Nekoto et al. - 2020 - Participatory Research for Low-resourced Machine Translation A Case Study in African Languages.pdf}
274
  }
275
 
276
+ @misc{OpenaiMMMLUDatasets2024,
277
+ title = {Openai/{{MMMLU}} {$\cdot$} {{Datasets}} at {{Hugging Face}}},
278
+ year = {2024},
279
+ month = oct,
280
+ urldate = {2025-05-28},
281
+ abstract = {We're on a journey to advance and democratize artificial intelligence through open source and open science.},
282
+ howpublished = {https://huggingface.co/datasets/openai/MMMLU},
283
+ file = {/Users/david/Zotero/storage/LPEN8SFL/MMMLU.html}
284
+ }
285
+
286
  @misc{OpenlanguagedataFlores2024,
287
  title = {Openlanguagedata/Flores},
288
  year = {2024},
 
328
  file = {/Users/david/Zotero/storage/BWFBTAZ9/Robinson et al. - 2023 - ChatGPT MT Competitive for High- (but Not Low-) Resource Languages.pdf}
329
  }
330
 
331
+ @misc{shiLanguageModelsAre2022,
332
+ title = {Language {{Models}} Are {{Multilingual Chain-of-Thought Reasoners}}},
333
+ author = {Shi, Freda and Suzgun, Mirac and Freitag, Markus and Wang, Xuezhi and Srivats, Suraj and Vosoughi, Soroush and Chung, Hyung Won and Tay, Yi and Ruder, Sebastian and Zhou, Denny and Das, Dipanjan and Wei, Jason},
334
+ year = {2022},
335
+ month = oct,
336
+ number = {arXiv:2210.03057},
337
+ eprint = {2210.03057},
338
+ primaryclass = {cs},
339
+ publisher = {arXiv},
340
+ doi = {10.48550/arXiv.2210.03057},
341
+ urldate = {2025-05-28},
342
+ abstract = {We evaluate the reasoning abilities of large language models in multilingual settings. We introduce the Multilingual Grade School Math (MGSM) benchmark, by manually translating 250 grade-school math problems from the GSM8K dataset (Cobbe et al., 2021) into ten typologically diverse languages. We find that the ability to solve MGSM problems via chain-of-thought prompting emerges with increasing model scale, and that models have strikingly strong multilingual reasoning abilities, even in underrepresented languages such as Bengali and Swahili. Finally, we show that the multilingual reasoning abilities of language models extend to other tasks such as commonsense reasoning and word-in-context semantic judgment. The MGSM benchmark is publicly available at https://github.com/google-research/url-nlp.},
343
+ archiveprefix = {arXiv},
344
+ file = {/Users/david/Zotero/storage/3W9ATYCI/Shi et al. - 2022 - Language Models are Multilingual Chain-of-Thought Reasoners.pdf;/Users/david/Zotero/storage/4HED3DGQ/2210.html}
345
+ }
346
+
347
  @misc{siminyuAI4DAfricanLanguage2021,
348
  title = {{{AI4D}} -- {{African Language Program}}},
349
  author = {Siminyu, Kathleen and Kalipe, Godson and Orlic, Davor and Abbott, Jade and Marivate, Vukosi and Freshia, Sackey and Sibal, Prateek and Neupane, Bhanu and Adelani, David I. and Taylor, Amelia and ALI, Jamiil Toure and Degila, Kevin and Balogoun, Momboladji and DIOP, Thierno Ibrahima and David, Davis and Fourati, Chayma and Haddad, Hatem and Naski, Malek},
 
360
  file = {/Users/david/Zotero/storage/VU6IFENR/Siminyu et al. - 2021 - AI4D -- African Language Program.pdf;/Users/david/Zotero/storage/7TV2PS8J/2104.html}
361
  }
362
 
363
+ @misc{singhGlobalMMLUUnderstanding2025,
364
+ title = {Global {{MMLU}}: {{Understanding}} and {{Addressing Cultural}} and {{Linguistic Biases}} in {{Multilingual Evaluation}}},
365
+ shorttitle = {Global {{MMLU}}},
366
+ author = {Singh, Shivalika and Romanou, Angelika and Fourrier, Cl{\'e}mentine and Adelani, David I. and Ngui, Jian Gang and {Vila-Suero}, Daniel and Limkonchotiwat, Peerat and Marchisio, Kelly and Leong, Wei Qi and Susanto, Yosephine and Ng, Raymond and Longpre, Shayne and Ko, Wei-Yin and Ruder, Sebastian and Smith, Madeline and Bosselut, Antoine and Oh, Alice and Martins, Andre F. T. and Choshen, Leshem and Ippolito, Daphne and Ferrante, Enzo and Fadaee, Marzieh and Ermis, Beyza and Hooker, Sara},
367
+ year = {2025},
368
+ month = feb,
369
+ number = {arXiv:2412.03304},
370
+ eprint = {2412.03304},
371
+ primaryclass = {cs},
372
+ publisher = {arXiv},
373
+ doi = {10.48550/arXiv.2412.03304},
374
+ urldate = {2025-05-28},
375
+ abstract = {Cultural biases in multilingual datasets pose significant challenges for their effectiveness as global benchmarks. These biases stem not only from differences in language but also from the cultural knowledge required to interpret questions, reducing the practical utility of translated datasets like MMLU. Furthermore, translation often introduces artefacts that can distort the meaning or clarity of questions in the target language. A common practice in multilingual evaluation is to rely on machine-translated evaluation sets, but simply translating a dataset is insufficient to address these challenges. In this work, we trace the impact of both of these issues on multilingual evaluations and ensuing model performances. Our large-scale evaluation of state-of-the-art open and proprietary models illustrates that progress on MMLU depends heavily on learning Western-centric concepts, with 28\% of all questions requiring culturally sensitive knowledge. Moreover, for questions requiring geographic knowledge, an astounding 84.9\% focus on either North American or European regions. Rankings of model evaluations change depending on whether they are evaluated on the full portion or the subset of questions annotated as culturally sensitive, showing the distortion to model rankings when blindly relying on translated MMLU. We release Global MMLU, an improved MMLU with evaluation coverage across 42 languages -- with improved overall quality by engaging with compensated professional and community annotators to verify translation quality while also rigorously evaluating cultural biases present in the original dataset. This comprehensive Global MMLU set also includes designated subsets labeled as culturally sensitive and culturally agnostic to allow for more holistic, complete evaluation.},
376
+ archiveprefix = {arXiv},
377
+ file = {/Users/david/Zotero/storage/23LXUKPU/Singh et al. - 2025 - Global MMLU Understanding and Addressing Cultural and Linguistic Biases in Multilingual Evaluation.pdf;/Users/david/Zotero/storage/563VJ87N/2412.html}
378
+ }
379
+
380
  @misc{Tatoeba,
381
  title = {Tatoeba},
382
  urldate = {2024-11-03},
 
399
  keywords = {dataset,model,n=200},
400
  file = {/Users/david/Zotero/storage/GHWEGFFS/Team et al. - 2022 - No Language Left Behind Scaling Human-Centered Machine Translation.pdf;/Users/david/Zotero/storage/SZK3CP9C/2207.html}
401
  }
402
+
403
+ @misc{thellmannMultilingualLLMEvaluation2024,
404
+ title = {Towards {{Multilingual LLM Evaluation}} for {{European Languages}}},
405
+ author = {Thellmann, Klaudia and Stadler, Bernhard and Fromm, Michael and Buschhoff, Jasper Schulze and Jude, Alex and Barth, Fabio and Leveling, Johannes and {Flores-Herr}, Nicolas and K{\"o}hler, Joachim and J{\"a}kel, Ren{\'e} and Ali, Mehdi},
406
+ year = {2024},
407
+ month = oct,
408
+ number = {arXiv:2410.08928},
409
+ eprint = {2410.08928},
410
+ primaryclass = {cs},
411
+ publisher = {arXiv},
412
+ doi = {10.48550/arXiv.2410.08928},
413
+ urldate = {2025-05-28},
414
+ abstract = {The rise of Large Language Models (LLMs) has revolutionized natural language processing across numerous languages and tasks. However, evaluating LLM performance in a consistent and meaningful way across multiple European languages remains challenging, especially due to the scarcity of language-parallel multilingual benchmarks. We introduce a multilingual evaluation approach tailored for European languages. We employ translated versions of five widely-used benchmarks to assess the capabilities of 40 LLMs across 21 European languages. Our contributions include examining the effectiveness of translated benchmarks, assessing the impact of different translation services, and offering a multilingual evaluation framework for LLMs that includes newly created datasets: EU20-MMLU, EU20-HellaSwag, EU20-ARC, EU20-TruthfulQA, and EU20-GSM8K. The benchmarks and results are made publicly available to encourage further research in multilingual LLM evaluation.},
415
+ archiveprefix = {arXiv},
416
+ file = {/Users/david/Zotero/storage/DCEINCKD/Thellmann et al. - 2024 - Towards Multilingual LLM Evaluation for European Languages.pdf;/Users/david/Zotero/storage/EAQS33RW/2410.html}
417
+ }