Loading…
FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders
Pretrained text encoders, such as BERT, have been applied increasingly in various natural language processing (NLP) tasks, and have recently demonstrated significant performance gains. However, recent studies have demonstrated the existence of social bias in these pretrained NLP models. Although pri...
Saved in:
Published in: | arXiv.org 2021-03 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Cheng, Pengyu Weituo Hao Yuan, Siyang Si, Shijing Lawrence, Carin |
description | Pretrained text encoders, such as BERT, have been applied increasingly in various natural language processing (NLP) tasks, and have recently demonstrated significant performance gains. However, recent studies have demonstrated the existence of social bias in these pretrained NLP models. Although prior works have made progress on word-level debiasing, improved sentence-level fairness of pretrained encoders still lacks exploration. In this paper, we proposed the first neural debiasing method for a pretrained sentence encoder, which transforms the pretrained encoder outputs into debiased representations via a fair filter (FairFil) network. To learn the FairFil, we introduce a contrastive learning framework that not only minimizes the correlation between filtered embeddings and bias words but also preserves rich semantic information of the original sentences. On real-world datasets, our FairFil effectively reduces the bias degree of pretrained text encoders, while continuously showing desirable performance on downstream tasks. Moreover, our post-hoc method does not require any retraining of the text encoders, further enlarging FairFil's application space. |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2500710097</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2500710097</sourcerecordid><originalsourceid>FETCH-proquest_journals_25007100973</originalsourceid><addsrcrecordid>eNqNyrEOgjAUQNHGxESi_MNLnElKa0VdEaKDxoGdVHloCWn1tRg_XwY_wOkO90xYJKRMk81KiBmLve8452KdCaVkxI6lNlSafge5s4G0D-aNcMaBdA97vBrtjb3DCcPDNdA6ggvh6IzFBir8BCjszTVIfsGmre49xr_O2bIsqvyQPMm9BvSh7txAdly1UJxnKefbTP6nvk0yO_A</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2500710097</pqid></control><display><type>article</type><title>FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders</title><source>Publicly Available Content Database</source><creator>Cheng, Pengyu ; Weituo Hao ; Yuan, Siyang ; Si, Shijing ; Lawrence, Carin</creator><creatorcontrib>Cheng, Pengyu ; Weituo Hao ; Yuan, Siyang ; Si, Shijing ; Lawrence, Carin</creatorcontrib><description>Pretrained text encoders, such as BERT, have been applied increasingly in various natural language processing (NLP) tasks, and have recently demonstrated significant performance gains. However, recent studies have demonstrated the existence of social bias in these pretrained NLP models. Although prior works have made progress on word-level debiasing, improved sentence-level fairness of pretrained encoders still lacks exploration. In this paper, we proposed the first neural debiasing method for a pretrained sentence encoder, which transforms the pretrained encoder outputs into debiased representations via a fair filter (FairFil) network. To learn the FairFil, we introduce a contrastive learning framework that not only minimizes the correlation between filtered embeddings and bias words but also preserves rich semantic information of the original sentences. On real-world datasets, our FairFil effectively reduces the bias degree of pretrained text encoders, while continuously showing desirable performance on downstream tasks. Moreover, our post-hoc method does not require any retraining of the text encoders, further enlarging FairFil's application space.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Bias ; Coders ; Human bias ; Natural language processing ; Retraining ; Sentences</subject><ispartof>arXiv.org, 2021-03</ispartof><rights>2021. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2500710097?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Cheng, Pengyu</creatorcontrib><creatorcontrib>Weituo Hao</creatorcontrib><creatorcontrib>Yuan, Siyang</creatorcontrib><creatorcontrib>Si, Shijing</creatorcontrib><creatorcontrib>Lawrence, Carin</creatorcontrib><title>FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders</title><title>arXiv.org</title><description>Pretrained text encoders, such as BERT, have been applied increasingly in various natural language processing (NLP) tasks, and have recently demonstrated significant performance gains. However, recent studies have demonstrated the existence of social bias in these pretrained NLP models. Although prior works have made progress on word-level debiasing, improved sentence-level fairness of pretrained encoders still lacks exploration. In this paper, we proposed the first neural debiasing method for a pretrained sentence encoder, which transforms the pretrained encoder outputs into debiased representations via a fair filter (FairFil) network. To learn the FairFil, we introduce a contrastive learning framework that not only minimizes the correlation between filtered embeddings and bias words but also preserves rich semantic information of the original sentences. On real-world datasets, our FairFil effectively reduces the bias degree of pretrained text encoders, while continuously showing desirable performance on downstream tasks. Moreover, our post-hoc method does not require any retraining of the text encoders, further enlarging FairFil's application space.</description><subject>Bias</subject><subject>Coders</subject><subject>Human bias</subject><subject>Natural language processing</subject><subject>Retraining</subject><subject>Sentences</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNyrEOgjAUQNHGxESi_MNLnElKa0VdEaKDxoGdVHloCWn1tRg_XwY_wOkO90xYJKRMk81KiBmLve8452KdCaVkxI6lNlSafge5s4G0D-aNcMaBdA97vBrtjb3DCcPDNdA6ggvh6IzFBir8BCjszTVIfsGmre49xr_O2bIsqvyQPMm9BvSh7txAdly1UJxnKefbTP6nvk0yO_A</recordid><startdate>20210311</startdate><enddate>20210311</enddate><creator>Cheng, Pengyu</creator><creator>Weituo Hao</creator><creator>Yuan, Siyang</creator><creator>Si, Shijing</creator><creator>Lawrence, Carin</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20210311</creationdate><title>FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders</title><author>Cheng, Pengyu ; Weituo Hao ; Yuan, Siyang ; Si, Shijing ; Lawrence, Carin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_25007100973</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Bias</topic><topic>Coders</topic><topic>Human bias</topic><topic>Natural language processing</topic><topic>Retraining</topic><topic>Sentences</topic><toplevel>online_resources</toplevel><creatorcontrib>Cheng, Pengyu</creatorcontrib><creatorcontrib>Weituo Hao</creatorcontrib><creatorcontrib>Yuan, Siyang</creatorcontrib><creatorcontrib>Si, Shijing</creatorcontrib><creatorcontrib>Lawrence, Carin</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Cheng, Pengyu</au><au>Weituo Hao</au><au>Yuan, Siyang</au><au>Si, Shijing</au><au>Lawrence, Carin</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders</atitle><jtitle>arXiv.org</jtitle><date>2021-03-11</date><risdate>2021</risdate><eissn>2331-8422</eissn><abstract>Pretrained text encoders, such as BERT, have been applied increasingly in various natural language processing (NLP) tasks, and have recently demonstrated significant performance gains. However, recent studies have demonstrated the existence of social bias in these pretrained NLP models. Although prior works have made progress on word-level debiasing, improved sentence-level fairness of pretrained encoders still lacks exploration. In this paper, we proposed the first neural debiasing method for a pretrained sentence encoder, which transforms the pretrained encoder outputs into debiased representations via a fair filter (FairFil) network. To learn the FairFil, we introduce a contrastive learning framework that not only minimizes the correlation between filtered embeddings and bias words but also preserves rich semantic information of the original sentences. On real-world datasets, our FairFil effectively reduces the bias degree of pretrained text encoders, while continuously showing desirable performance on downstream tasks. Moreover, our post-hoc method does not require any retraining of the text encoders, further enlarging FairFil's application space.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2021-03 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2500710097 |
source | Publicly Available Content Database |
subjects | Bias Coders Human bias Natural language processing Retraining Sentences |
title | FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T12%3A53%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=FairFil:%20Contrastive%20Neural%20Debiasing%20Method%20for%20Pretrained%20Text%20Encoders&rft.jtitle=arXiv.org&rft.au=Cheng,%20Pengyu&rft.date=2021-03-11&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2500710097%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_25007100973%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2500710097&rft_id=info:pmid/&rfr_iscdi=true |