Loading…
Diversity and language technology: how language modeling bias causes epistemic injustice
It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought...
Saved in:
Published in: | Ethics and information technology 2024-03, Vol.26 (1), p.8, Article 8 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3 |
container_end_page | |
container_issue | 1 |
container_start_page | 8 |
container_title | Ethics and information technology |
container_volume | 26 |
creator | Helm, Paula Bella, Gábor Koch, Gertraud Giunchiglia, Fausto |
description | It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems. |
doi_str_mv | 10.1007/s10676-023-09742-6 |
format | article |
fullrecord | <record><control><sourceid>proquest_hal_p</sourceid><recordid>TN_cdi_hal_primary_oai_HAL_hal_04421595v1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2918809984</sourcerecordid><originalsourceid>FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3</originalsourceid><addsrcrecordid>eNp9kMtOwzAQRS0EEqXwA6wisWIR8CN2bHZVeRSpEhtYIsuxndRVmpQ4AfVv8i35MlzCY8dqRjP3Xl0dAM4RvEIQptceQZayGGISQ5EmOGYHYIJoimOeEHEYdsJ5jARNj8GJ92sIIU1ROgGvt-7dNt61u0hVJipVVXSqsFFr9aqqy7rY3USr-mPofz5Dv6mNLV1VDH3mlI-06rz1Q2-3zrd24_TQu2rd-dZpewqOclV6e_Y9p-Dl_u55voiXTw-P89ky1iThbWw1F0xxKDhiKlEGakYsoRRqo2FKszyHBlOGTY6MwZlhJFxySnSCMpErS6bgcsxdqVJuG7dRzU7WysnFbCn3N5gkGFFB31HQXozabVO_dda3cl13TRXqSSwQDy1EYDYFeFTppva-sflvLIJyj1yOyGVALr-QSxZMZDT5IK4K2_xF_-P6BCc6iYk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2918809984</pqid></control><display><type>article</type><title>Diversity and language technology: how language modeling bias causes epistemic injustice</title><source>Springer Nature</source><creator>Helm, Paula ; Bella, Gábor ; Koch, Gertraud ; Giunchiglia, Fausto</creator><creatorcontrib>Helm, Paula ; Bella, Gábor ; Koch, Gertraud ; Giunchiglia, Fausto</creatorcontrib><description>It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems.</description><identifier>ISSN: 1388-1957</identifier><identifier>EISSN: 1572-8439</identifier><identifier>DOI: 10.1007/s10676-023-09742-6</identifier><language>eng</language><publisher>Dordrecht: Springer Netherlands</publisher><subject>Artificial Intelligence ; Bias ; Computation and Language ; Computer Science ; Computers and Society ; Ethics ; History, Philosophy and Sociology of Sciences ; Humanities and Social Sciences ; Innovation/Technology Management ; Language ; Languages ; Large language models ; Library Science ; Machine translation ; Management of Computing and Information Systems ; Modelling ; Original Paper ; User Interfaces and Human Computer Interaction</subject><ispartof>Ethics and information technology, 2024-03, Vol.26 (1), p.8, Article 8</ispartof><rights>The Author(s) 2024</rights><rights>The Author(s) 2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>Attribution</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3</cites><orcidid>0000-0002-2719-9721 ; 0000-0002-3868-1740</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,780,784,885,27924,27925</link.rule.ids><backlink>$$Uhttps://hal.science/hal-04421595$$DView record in HAL$$Hfree_for_read</backlink></links><search><creatorcontrib>Helm, Paula</creatorcontrib><creatorcontrib>Bella, Gábor</creatorcontrib><creatorcontrib>Koch, Gertraud</creatorcontrib><creatorcontrib>Giunchiglia, Fausto</creatorcontrib><title>Diversity and language technology: how language modeling bias causes epistemic injustice</title><title>Ethics and information technology</title><addtitle>Ethics Inf Technol</addtitle><description>It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems.</description><subject>Artificial Intelligence</subject><subject>Bias</subject><subject>Computation and Language</subject><subject>Computer Science</subject><subject>Computers and Society</subject><subject>Ethics</subject><subject>History, Philosophy and Sociology of Sciences</subject><subject>Humanities and Social Sciences</subject><subject>Innovation/Technology Management</subject><subject>Language</subject><subject>Languages</subject><subject>Large language models</subject><subject>Library Science</subject><subject>Machine translation</subject><subject>Management of Computing and Information Systems</subject><subject>Modelling</subject><subject>Original Paper</subject><subject>User Interfaces and Human Computer Interaction</subject><issn>1388-1957</issn><issn>1572-8439</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kMtOwzAQRS0EEqXwA6wisWIR8CN2bHZVeRSpEhtYIsuxndRVmpQ4AfVv8i35MlzCY8dqRjP3Xl0dAM4RvEIQptceQZayGGISQ5EmOGYHYIJoimOeEHEYdsJ5jARNj8GJ92sIIU1ROgGvt-7dNt61u0hVJipVVXSqsFFr9aqqy7rY3USr-mPofz5Dv6mNLV1VDH3mlI-06rz1Q2-3zrd24_TQu2rd-dZpewqOclV6e_Y9p-Dl_u55voiXTw-P89ky1iThbWw1F0xxKDhiKlEGakYsoRRqo2FKszyHBlOGTY6MwZlhJFxySnSCMpErS6bgcsxdqVJuG7dRzU7WysnFbCn3N5gkGFFB31HQXozabVO_dda3cl13TRXqSSwQDy1EYDYFeFTppva-sflvLIJyj1yOyGVALr-QSxZMZDT5IK4K2_xF_-P6BCc6iYk</recordid><startdate>20240301</startdate><enddate>20240301</enddate><creator>Helm, Paula</creator><creator>Bella, Gábor</creator><creator>Koch, Gertraud</creator><creator>Giunchiglia, Fausto</creator><general>Springer Netherlands</general><general>Springer Nature B.V</general><general>Springer Verlag</general><scope>C6C</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>1XC</scope><scope>BXJBU</scope><scope>IHQJB</scope><scope>VOOES</scope><orcidid>https://orcid.org/0000-0002-2719-9721</orcidid><orcidid>https://orcid.org/0000-0002-3868-1740</orcidid></search><sort><creationdate>20240301</creationdate><title>Diversity and language technology: how language modeling bias causes epistemic injustice</title><author>Helm, Paula ; Bella, Gábor ; Koch, Gertraud ; Giunchiglia, Fausto</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Artificial Intelligence</topic><topic>Bias</topic><topic>Computation and Language</topic><topic>Computer Science</topic><topic>Computers and Society</topic><topic>Ethics</topic><topic>History, Philosophy and Sociology of Sciences</topic><topic>Humanities and Social Sciences</topic><topic>Innovation/Technology Management</topic><topic>Language</topic><topic>Languages</topic><topic>Large language models</topic><topic>Library Science</topic><topic>Machine translation</topic><topic>Management of Computing and Information Systems</topic><topic>Modelling</topic><topic>Original Paper</topic><topic>User Interfaces and Human Computer Interaction</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Helm, Paula</creatorcontrib><creatorcontrib>Bella, Gábor</creatorcontrib><creatorcontrib>Koch, Gertraud</creatorcontrib><creatorcontrib>Giunchiglia, Fausto</creatorcontrib><collection>SpringerOpen</collection><collection>CrossRef</collection><collection>Hyper Article en Ligne (HAL)</collection><collection>HAL-SHS: Archive ouverte en Sciences de l'Homme et de la Société</collection><collection>HAL-SHS: Archive ouverte en Sciences de l'Homme et de la Société (Open Access)</collection><collection>Hyper Article en Ligne (HAL) (Open Access)</collection><jtitle>Ethics and information technology</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Helm, Paula</au><au>Bella, Gábor</au><au>Koch, Gertraud</au><au>Giunchiglia, Fausto</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Diversity and language technology: how language modeling bias causes epistemic injustice</atitle><jtitle>Ethics and information technology</jtitle><stitle>Ethics Inf Technol</stitle><date>2024-03-01</date><risdate>2024</risdate><volume>26</volume><issue>1</issue><spage>8</spage><pages>8-</pages><artnum>8</artnum><issn>1388-1957</issn><eissn>1572-8439</eissn><abstract>It is well known that AI-based language technology—large language models, machine translation systems, multilingual dictionaries, and corpora—is currently limited to three percent of the world’s most widely spoken, financially and politically backed languages. In response, recent efforts have sought to address the “digital language divide” by extending the reach of large language models to “underserved languages.” We show how some of these efforts tend to produce flawed solutions that adhere to a hard-wired representational preference for certain languages, which we call language modeling bias. Language modeling bias is a specific and under-studied form of linguistic bias were language technology by design favors certain languages, dialects, or sociolects with respect to others. We show that language modeling bias can result in systems that, while being precise regarding languages and cultures of dominant powers, are limited in the expression of socio-culturally relevant notions of other communities. We further argue that at the root of this problem lies a systematic tendency of technology developer communities to apply a simplistic understanding of diversity which does not do justice to the more profound differences that languages, and ultimately the communities that speak them, embody. Drawing on the concept of epistemic injustice, we point to the broader ethico-political implications and show how it can lead not only to a disregard for valuable aspects of diversity but also to an under-representation of the needs of marginalized language communities. Finally, we present an alternative socio-technical approach that is designed to tackle some of the analyzed problems.</abstract><cop>Dordrecht</cop><pub>Springer Netherlands</pub><doi>10.1007/s10676-023-09742-6</doi><orcidid>https://orcid.org/0000-0002-2719-9721</orcidid><orcidid>https://orcid.org/0000-0002-3868-1740</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1388-1957 |
ispartof | Ethics and information technology, 2024-03, Vol.26 (1), p.8, Article 8 |
issn | 1388-1957 1572-8439 |
language | eng |
recordid | cdi_hal_primary_oai_HAL_hal_04421595v1 |
source | Springer Nature |
subjects | Artificial Intelligence Bias Computation and Language Computer Science Computers and Society Ethics History, Philosophy and Sociology of Sciences Humanities and Social Sciences Innovation/Technology Management Language Languages Large language models Library Science Machine translation Management of Computing and Information Systems Modelling Original Paper User Interfaces and Human Computer Interaction |
title | Diversity and language technology: how language modeling bias causes epistemic injustice |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-30T19%3A52%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_hal_p&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Diversity%20and%20language%20technology:%20how%C2%A0language%C2%A0modeling%C2%A0bias%20causes%C2%A0epistemic%C2%A0injustice&rft.jtitle=Ethics%20and%20information%20technology&rft.au=Helm,%20Paula&rft.date=2024-03-01&rft.volume=26&rft.issue=1&rft.spage=8&rft.pages=8-&rft.artnum=8&rft.issn=1388-1957&rft.eissn=1572-8439&rft_id=info:doi/10.1007/s10676-023-09742-6&rft_dat=%3Cproquest_hal_p%3E2918809984%3C/proquest_hal_p%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c348t-ec896a809816a4ad0c63e3550cdc075bff0d2562df1dd2bd63ff0f53c41b9fae3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2918809984&rft_id=info:pmid/&rfr_iscdi=true |