Loading…

Diversity and language technology: how language modeling bias causes epistemic injustice

It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought...

Full description

Saved in:
Bibliographic Details
Published in:Ethics and information technology 2024-03, Vol.26 (1), p.8, Article 8
Main Authors: Helm, Paula, Bella, Gábor, Koch, Gertraud, Giunchiglia, Fausto
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3
container_end_page
container_issue 1
container_start_page 8
container_title Ethics and information technology
container_volume 26
creator Helm, Paula
Bella, Gábor
Koch, Gertraud
Giunchiglia, Fausto
description It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems.
doi_str_mv 10.1007/s10676-023-09742-6
format article
fullrecord <record><control><sourceid>proquest_hal_p</sourceid><recordid>TN_cdi_hal_primary_oai_HAL_hal_04421595v1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2918809984</sourcerecordid><originalsourceid>FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3</originalsourceid><addsrcrecordid>eNp9kMtOwzAQRS0EEqXwA6wisWIR8CN2bHZVeRSpEhtYIsuxndRVmpQ4AfVv8i35MlzCY8dqRjP3Xl0dAM4RvEIQptceQZayGGISQ5EmOGYHYIJoimOeEHEYdsJ5jARNj8GJ92sIIU1ROgGvt-7dNt61u0hVJipVVXSqsFFr9aqqy7rY3USr-mPofz5Dv6mNLV1VDH3mlI-06rz1Q2-3zrd24_TQu2rd-dZpewqOclV6e_Y9p-Dl_u55voiXTw-P89ky1iThbWw1F0xxKDhiKlEGakYsoRRqo2FKszyHBlOGTY6MwZlhJFxySnSCMpErS6bgcsxdqVJuG7dRzU7WysnFbCn3N5gkGFFB31HQXozabVO_dda3cl13TRXqSSwQDy1EYDYFeFTppva-sflvLIJyj1yOyGVALr-QSxZMZDT5IK4K2_xF_-P6BCc6iYk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2918809984</pqid></control><display><type>article</type><title>Diversity and language technology: how language modeling bias causes epistemic injustice</title><source>Springer Nature</source><creator>Helm, Paula ; Bella, Gábor ; Koch, Gertraud ; Giunchiglia, Fausto</creator><creatorcontrib>Helm, Paula ; Bella, Gábor ; Koch, Gertraud ; Giunchiglia, Fausto</creatorcontrib><description>It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems.</description><identifier>ISSN: 1388-1957</identifier><identifier>EISSN: 1572-8439</identifier><identifier>DOI: 10.1007/s10676-023-09742-6</identifier><language>eng</language><publisher>Dordrecht: Springer Netherlands</publisher><subject>Artificial Intelligence ; Bias ; Computation and Language ; Computer Science ; Computers and Society ; Ethics ; History, Philosophy and Sociology of Sciences ; Humanities and Social Sciences ; Innovation/Technology Management ; Language ; Languages ; Large language models ; Library Science ; Machine translation ; Management of Computing and Information Systems ; Modelling ; Original Paper ; User Interfaces and Human Computer Interaction</subject><ispartof>Ethics and information technology, 2024-03, Vol.26 (1), p.8, Article 8</ispartof><rights>The Author(s) 2024</rights><rights>The Author(s) 2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>Attribution</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3</cites><orcidid>0000-0002-2719-9721 ; 0000-0002-3868-1740</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,780,784,885,27924,27925</link.rule.ids><backlink>$$Uhttps://hal.science/hal-04421595$$DView record in HAL$$Hfree_for_read</backlink></links><search><creatorcontrib>Helm, Paula</creatorcontrib><creatorcontrib>Bella, Gábor</creatorcontrib><creatorcontrib>Koch, Gertraud</creatorcontrib><creatorcontrib>Giunchiglia, Fausto</creatorcontrib><title>Diversity and language technology: how language modeling bias causes epistemic injustice</title><title>Ethics and information technology</title><addtitle>Ethics Inf Technol</addtitle><description>It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems.</description><subject>Artificial Intelligence</subject><subject>Bias</subject><subject>Computation and Language</subject><subject>Computer Science</subject><subject>Computers and Society</subject><subject>Ethics</subject><subject>History, Philosophy and Sociology of Sciences</subject><subject>Humanities and Social Sciences</subject><subject>Innovation/Technology Management</subject><subject>Language</subject><subject>Languages</subject><subject>Large language models</subject><subject>Library Science</subject><subject>Machine translation</subject><subject>Management of Computing and Information Systems</subject><subject>Modelling</subject><subject>Original Paper</subject><subject>User Interfaces and Human Computer Interaction</subject><issn>1388-1957</issn><issn>1572-8439</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kMtOwzAQRS0EEqXwA6wisWIR8CN2bHZVeRSpEhtYIsuxndRVmpQ4AfVv8i35MlzCY8dqRjP3Xl0dAM4RvEIQptceQZayGGISQ5EmOGYHYIJoimOeEHEYdsJ5jARNj8GJ92sIIU1ROgGvt-7dNt61u0hVJipVVXSqsFFr9aqqy7rY3USr-mPofz5Dv6mNLV1VDH3mlI-06rz1Q2-3zrd24_TQu2rd-dZpewqOclV6e_Y9p-Dl_u55voiXTw-P89ky1iThbWw1F0xxKDhiKlEGakYsoRRqo2FKszyHBlOGTY6MwZlhJFxySnSCMpErS6bgcsxdqVJuG7dRzU7WysnFbCn3N5gkGFFB31HQXozabVO_dda3cl13TRXqSSwQDy1EYDYFeFTppva-sflvLIJyj1yOyGVALr-QSxZMZDT5IK4K2_xF_-P6BCc6iYk</recordid><startdate>20240301</startdate><enddate>20240301</enddate><creator>Helm, Paula</creator><creator>Bella, Gábor</creator><creator>Koch, Gertraud</creator><creator>Giunchiglia, Fausto</creator><general>Springer Netherlands</general><general>Springer Nature B.V</general><general>Springer Verlag</general><scope>C6C</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>1XC</scope><scope>BXJBU</scope><scope>IHQJB</scope><scope>VOOES</scope><orcidid>https://orcid.org/0000-0002-2719-9721</orcidid><orcidid>https://orcid.org/0000-0002-3868-1740</orcidid></search><sort><creationdate>20240301</creationdate><title>Diversity and language technology: how language modeling bias causes epistemic injustice</title><author>Helm, Paula ; Bella, Gábor ; Koch, Gertraud ; Giunchiglia, Fausto</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Artificial Intelligence</topic><topic>Bias</topic><topic>Computation and Language</topic><topic>Computer Science</topic><topic>Computers and Society</topic><topic>Ethics</topic><topic>History, Philosophy and Sociology of Sciences</topic><topic>Humanities and Social Sciences</topic><topic>Innovation/Technology Management</topic><topic>Language</topic><topic>Languages</topic><topic>Large language models</topic><topic>Library Science</topic><topic>Machine translation</topic><topic>Management of Computing and Information Systems</topic><topic>Modelling</topic><topic>Original Paper</topic><topic>User Interfaces and Human Computer Interaction</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Helm, Paula</creatorcontrib><creatorcontrib>Bella, Gábor</creatorcontrib><creatorcontrib>Koch, Gertraud</creatorcontrib><creatorcontrib>Giunchiglia, Fausto</creatorcontrib><collection>SpringerOpen</collection><collection>CrossRef</collection><collection>Hyper Article en Ligne (HAL)</collection><collection>HAL-SHS: Archive ouverte en Sciences de l'Homme et de la Société</collection><collection>HAL-SHS: Archive ouverte en Sciences de l'Homme et de la Société (Open Access)</collection><collection>Hyper Article en Ligne (HAL) (Open Access)</collection><jtitle>Ethics and information technology</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Helm, Paula</au><au>Bella, Gábor</au><au>Koch, Gertraud</au><au>Giunchiglia, Fausto</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Diversity and language technology: how language modeling bias causes epistemic injustice</atitle><jtitle>Ethics and information technology</jtitle><stitle>Ethics Inf Technol</stitle><date>2024-03-01</date><risdate>2024</risdate><volume>26</volume><issue>1</issue><spage>8</spage><pages>8-</pages><artnum>8</artnum><issn>1388-1957</issn><eissn>1572-8439</eissn><abstract>It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems.</abstract><cop>Dordrecht</cop><pub>Springer Netherlands</pub><doi>10.1007/s10676-023-09742-6</doi><orcidid>https://orcid.org/0000-0002-2719-9721</orcidid><orcidid>https://orcid.org/0000-0002-3868-1740</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1388-1957
ispartof Ethics and information technology, 2024-03, Vol.26 (1), p.8, Article 8
issn 1388-1957
1572-8439
language eng
recordid cdi_hal_primary_oai_HAL_hal_04421595v1
source Springer Nature
subjects Artificial Intelligence
Bias
Computation and Language
Computer Science
Computers and Society
Ethics
History, Philosophy and Sociology of Sciences
Humanities and Social Sciences
Innovation/Technology Management
Language
Languages
Large language models
Library Science
Machine translation
Management of Computing and Information Systems
Modelling
Original Paper
User Interfaces and Human Computer Interaction
title Diversity and language technology: how language modeling bias causes epistemic injustice
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-30T19%3A52%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_hal_p&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Diversity%20and%20language%20technology:%20how%C2%A0language%C2%A0modeling%C2%A0bias%20causes%C2%A0epistemic%C2%A0injustice&rft.jtitle=Ethics%20and%20information%20technology&rft.au=Helm,%20Paula&rft.date=2024-03-01&rft.volume=26&rft.issue=1&rft.spage=8&rft.pages=8-&rft.artnum=8&rft.issn=1388-1957&rft.eissn=1572-8439&rft_id=info:doi/10.1007/s10676-023-09742-6&rft_dat=%3Cproquest_hal_p%3E2918809984%3C/proquest_hal_p%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2918809984&rft_id=info:pmid/&rfr_iscdi=true