Loading…

Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure

Sample weighting is widely used in deep learning. A large number of weighting methods essentially utilize the learning difficulty of training samples to calculate their weights. In this study, this scheme is called difficulty-based weighting. Two important issues arise when explaining this scheme. F...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2023-01
Main Authors: Zhou, Xiaoling, Wu, Ou, Zhu, Weiyao, Liang, Ziyang
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Zhou, Xiaoling
Wu, Ou
Zhu, Weiyao
Liang, Ziyang
description Sample weighting is widely used in deep learning. A large number of weighting methods essentially utilize the learning difficulty of training samples to calculate their weights. In this study, this scheme is called difficulty-based weighting. Two important issues arise when explaining this scheme. First, a unified difficulty measure that can be theoretically guaranteed for training samples does not exist. The learning difficulties of the samples are determined by multiple factors including noise level, imbalance degree, margin, and uncertainty. Nevertheless, existing measures only consider a single factor or in part, but not in their entirety. Second, a comprehensive theoretical explanation is lacking with respect to demonstrating why difficulty-based weighting schemes are effective in deep learning. In this study, we theoretically prove that the generalization error of a sample can be used as a universal difficulty measure. Furthermore, we provide formal theoretical justifications on the role of difficulty-based weighting for deep learning, consequently revealing its positive influences on both the optimization dynamics and generalization performance of deep models, which is instructive to existing weighting schemes.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2765220838</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2765220838</sourcerecordid><originalsourceid>FETCH-proquest_journals_27652208383</originalsourceid><addsrcrecordid>eNqNi7sKwjAUQIMgWLT_EHAuxBv72H3g4lSLY4n2tk2Jac1D8e-t4ODodIZzzoQEwPkqytYAMxJa2zHGIEkhjnlA8kJXaKwTupK6oVtZ1_LqlXtFF2Gxorm4DQrpGWXTuk_xlK6lghZaPsZPqJ-FHlFYb3BBprVQFsMv52S53502h2gw_d2jdWXXe6NHVUKaxAAs4xn_r3oDdkpAhg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2765220838</pqid></control><display><type>article</type><title>Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><creator>Zhou, Xiaoling ; Wu, Ou ; Zhu, Weiyao ; Liang, Ziyang</creator><creatorcontrib>Zhou, Xiaoling ; Wu, Ou ; Zhu, Weiyao ; Liang, Ziyang</creatorcontrib><description>Sample weighting is widely used in deep learning. A large number of weighting methods essentially utilize the learning difficulty of training samples to calculate their weights. In this study, this scheme is called difficulty-based weighting. Two important issues arise when explaining this scheme. First, a unified difficulty measure that can be theoretically guaranteed for training samples does not exist. The learning difficulties of the samples are determined by multiple factors including noise level, imbalance degree, margin, and uncertainty. Nevertheless, existing measures only consider a single factor or in part, but not in their entirety. Second, a comprehensive theoretical explanation is lacking with respect to demonstrating why difficulty-based weighting schemes are effective in deep learning. In this study, we theoretically prove that the generalization error of a sample can be used as a universal difficulty measure. Furthermore, we provide formal theoretical justifications on the role of difficulty-based weighting for deep learning, consequently revealing its positive influences on both the optimization dynamics and generalization performance of deep models, which is instructive to existing weighting schemes.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Deep learning ; Noise levels ; Optimization ; Training ; Weighting methods</subject><ispartof>arXiv.org, 2023-01</ispartof><rights>2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2765220838?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>776,780,25732,36991,44569</link.rule.ids></links><search><creatorcontrib>Zhou, Xiaoling</creatorcontrib><creatorcontrib>Wu, Ou</creatorcontrib><creatorcontrib>Zhu, Weiyao</creatorcontrib><creatorcontrib>Liang, Ziyang</creatorcontrib><title>Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure</title><title>arXiv.org</title><description>Sample weighting is widely used in deep learning. A large number of weighting methods essentially utilize the learning difficulty of training samples to calculate their weights. In this study, this scheme is called difficulty-based weighting. Two important issues arise when explaining this scheme. First, a unified difficulty measure that can be theoretically guaranteed for training samples does not exist. The learning difficulties of the samples are determined by multiple factors including noise level, imbalance degree, margin, and uncertainty. Nevertheless, existing measures only consider a single factor or in part, but not in their entirety. Second, a comprehensive theoretical explanation is lacking with respect to demonstrating why difficulty-based weighting schemes are effective in deep learning. In this study, we theoretically prove that the generalization error of a sample can be used as a universal difficulty measure. Furthermore, we provide formal theoretical justifications on the role of difficulty-based weighting for deep learning, consequently revealing its positive influences on both the optimization dynamics and generalization performance of deep models, which is instructive to existing weighting schemes.</description><subject>Deep learning</subject><subject>Noise levels</subject><subject>Optimization</subject><subject>Training</subject><subject>Weighting methods</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNi7sKwjAUQIMgWLT_EHAuxBv72H3g4lSLY4n2tk2Jac1D8e-t4ODodIZzzoQEwPkqytYAMxJa2zHGIEkhjnlA8kJXaKwTupK6oVtZ1_LqlXtFF2Gxorm4DQrpGWXTuk_xlK6lghZaPsZPqJ-FHlFYb3BBprVQFsMv52S53502h2gw_d2jdWXXe6NHVUKaxAAs4xn_r3oDdkpAhg</recordid><startdate>20230112</startdate><enddate>20230112</enddate><creator>Zhou, Xiaoling</creator><creator>Wu, Ou</creator><creator>Zhu, Weiyao</creator><creator>Liang, Ziyang</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20230112</creationdate><title>Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure</title><author>Zhou, Xiaoling ; Wu, Ou ; Zhu, Weiyao ; Liang, Ziyang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_27652208383</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Deep learning</topic><topic>Noise levels</topic><topic>Optimization</topic><topic>Training</topic><topic>Weighting methods</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhou, Xiaoling</creatorcontrib><creatorcontrib>Wu, Ou</creatorcontrib><creatorcontrib>Zhu, Weiyao</creatorcontrib><creatorcontrib>Liang, Ziyang</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhou, Xiaoling</au><au>Wu, Ou</au><au>Zhu, Weiyao</au><au>Liang, Ziyang</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure</atitle><jtitle>arXiv.org</jtitle><date>2023-01-12</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Sample weighting is widely used in deep learning. A large number of weighting methods essentially utilize the learning difficulty of training samples to calculate their weights. In this study, this scheme is called difficulty-based weighting. Two important issues arise when explaining this scheme. First, a unified difficulty measure that can be theoretically guaranteed for training samples does not exist. The learning difficulties of the samples are determined by multiple factors including noise level, imbalance degree, margin, and uncertainty. Nevertheless, existing measures only consider a single factor or in part, but not in their entirety. Second, a comprehensive theoretical explanation is lacking with respect to demonstrating why difficulty-based weighting schemes are effective in deep learning. In this study, we theoretically prove that the generalization error of a sample can be used as a universal difficulty measure. Furthermore, we provide formal theoretical justifications on the role of difficulty-based weighting for deep learning, consequently revealing its positive influences on both the optimization dynamics and generalization performance of deep models, which is instructive to existing weighting schemes.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-01
issn 2331-8422
language eng
recordid cdi_proquest_journals_2765220838
source Publicly Available Content Database (Proquest) (PQ_SDU_P3)
subjects Deep learning
Noise levels
Optimization
Training
Weighting methods
title Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-21T14%3A49%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Understanding%20Difficulty-based%20Sample%20Weighting%20with%20a%20Universal%20Difficulty%20Measure&rft.jtitle=arXiv.org&rft.au=Zhou,%20Xiaoling&rft.date=2023-01-12&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2765220838%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_27652208383%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2765220838&rft_id=info:pmid/&rfr_iscdi=true