Loading…
Preview-based Category Contrastive Learning for Knowledge Distillation
Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level featu...
Saved in:
Published in: | arXiv.org 2024-10 |
---|---|
Main Authors: | , , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Ding, Muhe Wu, Jianlong Xue Dong Li, Xiaojie Qin, Pengda Gan, Tian Nie, Liqiang |
description | Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods. |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3118928508</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3118928508</sourcerecordid><originalsourceid>FETCH-proquest_journals_31189285083</originalsourceid><addsrcrecordid>eNqNyrEKwjAUQNEgCBbtPwScC2liNc7VIujg4F4ifS0pIU9f0hb_3g5-gNMdzl2wRCqVZ3on5YqlIfRCCLk_yKJQCavuBKOFKXuaAA0vTYQO6cNL9JFMiHYEfgND3vqOt0j86nFy0HTAT3Zm50y06Dds2RoXIP11zbbV-VFeshfhe4AQ6x4H8jPVKs_1UepCaPXf9QX2ZzvK</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3118928508</pqid></control><display><type>article</type><title>Preview-based Category Contrastive Learning for Knowledge Distillation</title><source>ProQuest - Publicly Available Content Database</source><creator>Ding, Muhe ; Wu, Jianlong ; Xue Dong ; Li, Xiaojie ; Qin, Pengda ; Gan, Tian ; Nie, Liqiang</creator><creatorcontrib>Ding, Muhe ; Wu, Jianlong ; Xue Dong ; Li, Xiaojie ; Qin, Pengda ; Gan, Tian ; Nie, Liqiang</creatorcontrib><description>Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Knowledge representation ; Machine learning ; Performance enhancement</subject><ispartof>arXiv.org, 2024-10</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/3118928508?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Ding, Muhe</creatorcontrib><creatorcontrib>Wu, Jianlong</creatorcontrib><creatorcontrib>Xue Dong</creatorcontrib><creatorcontrib>Li, Xiaojie</creatorcontrib><creatorcontrib>Qin, Pengda</creatorcontrib><creatorcontrib>Gan, Tian</creatorcontrib><creatorcontrib>Nie, Liqiang</creatorcontrib><title>Preview-based Category Contrastive Learning for Knowledge Distillation</title><title>arXiv.org</title><description>Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods.</description><subject>Algorithms</subject><subject>Knowledge representation</subject><subject>Machine learning</subject><subject>Performance enhancement</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNyrEKwjAUQNEgCBbtPwScC2liNc7VIujg4F4ifS0pIU9f0hb_3g5-gNMdzl2wRCqVZ3on5YqlIfRCCLk_yKJQCavuBKOFKXuaAA0vTYQO6cNL9JFMiHYEfgND3vqOt0j86nFy0HTAT3Zm50y06Dds2RoXIP11zbbV-VFeshfhe4AQ6x4H8jPVKs_1UepCaPXf9QX2ZzvK</recordid><startdate>20241018</startdate><enddate>20241018</enddate><creator>Ding, Muhe</creator><creator>Wu, Jianlong</creator><creator>Xue Dong</creator><creator>Li, Xiaojie</creator><creator>Qin, Pengda</creator><creator>Gan, Tian</creator><creator>Nie, Liqiang</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241018</creationdate><title>Preview-based Category Contrastive Learning for Knowledge Distillation</title><author>Ding, Muhe ; Wu, Jianlong ; Xue Dong ; Li, Xiaojie ; Qin, Pengda ; Gan, Tian ; Nie, Liqiang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_31189285083</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Knowledge representation</topic><topic>Machine learning</topic><topic>Performance enhancement</topic><toplevel>online_resources</toplevel><creatorcontrib>Ding, Muhe</creatorcontrib><creatorcontrib>Wu, Jianlong</creatorcontrib><creatorcontrib>Xue Dong</creatorcontrib><creatorcontrib>Li, Xiaojie</creatorcontrib><creatorcontrib>Qin, Pengda</creatorcontrib><creatorcontrib>Gan, Tian</creatorcontrib><creatorcontrib>Nie, Liqiang</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>ProQuest - Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ding, Muhe</au><au>Wu, Jianlong</au><au>Xue Dong</au><au>Li, Xiaojie</au><au>Qin, Pengda</au><au>Gan, Tian</au><au>Nie, Liqiang</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Preview-based Category Contrastive Learning for Knowledge Distillation</atitle><jtitle>arXiv.org</jtitle><date>2024-10-18</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-10 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_3118928508 |
source | ProQuest - Publicly Available Content Database |
subjects | Algorithms Knowledge representation Machine learning Performance enhancement |
title | Preview-based Category Contrastive Learning for Knowledge Distillation |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T04%3A50%3A44IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Preview-based%20Category%20Contrastive%20Learning%20for%20Knowledge%20Distillation&rft.jtitle=arXiv.org&rft.au=Ding,%20Muhe&rft.date=2024-10-18&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3118928508%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_31189285083%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3118928508&rft_id=info:pmid/&rfr_iscdi=true |