Tackling Language Modelling Bias in Support of Linguistic Diversity

Open Access
Authors
  • G. Bella
  • P. Helm ORCID logo
  • G. Koch
  • F. Giunchiglia
Publication date 2024
Book title ACM FAccT '24
Book subtitle Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency : June 3rd-6th 2024, Rio de Janeiro, Brazil
ISBN (electronic)
  • 9798400704505
Event 2024 ACM Conference on Fairness, Accountability, and Transparency, FAccT 2024
Pages (from-to) 562-572
Number of pages 11
Publisher New York: The Association for Computing Machinery
Organisations
  • Faculty of Humanities (FGw) - Amsterdam Institute for Humanities Research (AIHR) - Amsterdam School for Cultural Analysis (ASCA)
Abstract
Current AI-based language technologies—language models, machine translation systems, multilingual dictionaries and corpora—are known to focus on the world’s 2–3% most widely spoken languages. Research efforts of the past decade have attempted to expand this coverage to ‘under-resourced languages.’ The goal of our paper is to bring attention to a corollary phenomenon that we call language modelling bias: multilingual language processing systems often exhibit a hardwired, yet usually involuntary and hidden representational preference towards certain languages. We define language modelling bias as uneven per-language performance under similar test conditions. We show that bias stems not only from technology but also from ethically problematic research and development methodologies that disregard the needs of language communities. Moving towards diversity-aware alternatives, we present an initiative that aims at reducing language modelling bias within lexical resources through both technology design and methodology, based on an eye-level collaboration with local communities.
Document type Conference contribution
Language English
Published at https://doi.org/10.1145/3630106.3658925
Downloads
3630106.3658925 (Final published version)
Permalink to this page
Back