Loading...
ProverbEval: Exploring LLM Evaluation Challenges for Low-resource Language Understanding
Azime, Israel Abebe ; Tonja, Atnafu Lambebo ; Belay, Tadesse Destaw ; Chanie, Yonas ; Balcha, Bontu Fufa ; Abadi, Negasi Haile ; Ademtew, Henok Biadglign ; Nerea, Mulubrhan Abebe ; Yadeta, Debela Desalegn ; Geremew, Derartu Dagne ... show 4 more
Azime, Israel Abebe
Tonja, Atnafu Lambebo
Belay, Tadesse Destaw
Chanie, Yonas
Balcha, Bontu Fufa
Abadi, Negasi Haile
Ademtew, Henok Biadglign
Nerea, Mulubrhan Abebe
Yadeta, Debela Desalegn
Geremew, Derartu Dagne
Files
Supervisor
Department
Natural Language Processing
Embargo End Date
Type
Conference proceeding
Date
License
Language
Collections
Research Projects
Organizational Units
Journal Issue
Abstract
With the rapid development of evaluation datasets to assess LLMs understanding across a wide range of subjects and domains, identifying a suitable language understanding benchmark has become increasingly challenging. In this work, we explore LLM evaluation challenges for low-resource language understanding and introduce ProverbEval, LLM evaluation benchmark for low-resource languages, focusing on low-resource language understanding in culture-specific scenarios. We benchmark various LLMs and explore factors that create variability in the benchmarking process. We observed performance variances of up to 50%, depending on the order in which answer choices were presented in multiple-choice tasks. Native language proverb descriptions significantly improve tasks such as proverb generation, contributing to improved outcomes. Additionally, monolingual evaluations consistently outperformed their cross-lingual counterparts
in generation tasks. We argue that special attention must be given to the order of choices, the choice of prompt language, task variability, and generation tasks when creating LLM evaluation benchmarks
Citation
Source
Conference
Nations of the Americas Chapter of the Association for Computational Linguistics (NAACL) 2025
Keywords
Subjects
Source
Nations of the Americas Chapter of the Association for Computational Linguistics (NAACL) 2025
Publisher
Association for Computational Linguistics (ACL)
