Loading…

Preview-based Category Contrastive Learning for Knowledge Distillation

Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level featu...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2024-10
Main Authors: Ding, Muhe, Wu, Jianlong, Xue Dong, Li, Xiaojie, Qin, Pengda, Gan, Tian, Nie, Liqiang
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Ding, Muhe
Wu, Jianlong
Xue Dong
Li, Xiaojie
Qin, Pengda
Gan, Tian
Nie, Liqiang
description Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3118928508</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3118928508</sourcerecordid><originalsourceid>FETCH-proquest_journals_31189285083</originalsourceid><addsrcrecordid>eNqNyrEKwjAUQNEgCBbtPwScC2liNc7VIujg4F4ifS0pIU9f0hb_3g5-gNMdzl2wRCqVZ3on5YqlIfRCCLk_yKJQCavuBKOFKXuaAA0vTYQO6cNL9JFMiHYEfgND3vqOt0j86nFy0HTAT3Zm50y06Dds2RoXIP11zbbV-VFeshfhe4AQ6x4H8jPVKs_1UepCaPXf9QX2ZzvK</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3118928508</pqid></control><display><type>article</type><title>Preview-based Category Contrastive Learning for Knowledge Distillation</title><source>ProQuest - Publicly Available Content Database</source><creator>Ding, Muhe ; Wu, Jianlong ; Xue Dong ; Li, Xiaojie ; Qin, Pengda ; Gan, Tian ; Nie, Liqiang</creator><creatorcontrib>Ding, Muhe ; Wu, Jianlong ; Xue Dong ; Li, Xiaojie ; Qin, Pengda ; Gan, Tian ; Nie, Liqiang</creatorcontrib><description>Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Knowledge representation ; Machine learning ; Performance enhancement</subject><ispartof>arXiv.org, 2024-10</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/3118928508?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Ding, Muhe</creatorcontrib><creatorcontrib>Wu, Jianlong</creatorcontrib><creatorcontrib>Xue Dong</creatorcontrib><creatorcontrib>Li, Xiaojie</creatorcontrib><creatorcontrib>Qin, Pengda</creatorcontrib><creatorcontrib>Gan, Tian</creatorcontrib><creatorcontrib>Nie, Liqiang</creatorcontrib><title>Preview-based Category Contrastive Learning for Knowledge Distillation</title><title>arXiv.org</title><description>Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods.</description><subject>Algorithms</subject><subject>Knowledge representation</subject><subject>Machine learning</subject><subject>Performance enhancement</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNyrEKwjAUQNEgCBbtPwScC2liNc7VIujg4F4ifS0pIU9f0hb_3g5-gNMdzl2wRCqVZ3on5YqlIfRCCLk_yKJQCavuBKOFKXuaAA0vTYQO6cNL9JFMiHYEfgND3vqOt0j86nFy0HTAT3Zm50y06Dds2RoXIP11zbbV-VFeshfhe4AQ6x4H8jPVKs_1UepCaPXf9QX2ZzvK</recordid><startdate>20241018</startdate><enddate>20241018</enddate><creator>Ding, Muhe</creator><creator>Wu, Jianlong</creator><creator>Xue Dong</creator><creator>Li, Xiaojie</creator><creator>Qin, Pengda</creator><creator>Gan, Tian</creator><creator>Nie, Liqiang</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241018</creationdate><title>Preview-based Category Contrastive Learning for Knowledge Distillation</title><author>Ding, Muhe ; Wu, Jianlong ; Xue Dong ; Li, Xiaojie ; Qin, Pengda ; Gan, Tian ; Nie, Liqiang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_31189285083</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Knowledge representation</topic><topic>Machine learning</topic><topic>Performance enhancement</topic><toplevel>online_resources</toplevel><creatorcontrib>Ding, Muhe</creatorcontrib><creatorcontrib>Wu, Jianlong</creatorcontrib><creatorcontrib>Xue Dong</creatorcontrib><creatorcontrib>Li, Xiaojie</creatorcontrib><creatorcontrib>Qin, Pengda</creatorcontrib><creatorcontrib>Gan, Tian</creatorcontrib><creatorcontrib>Nie, Liqiang</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>ProQuest - Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ding, Muhe</au><au>Wu, Jianlong</au><au>Xue Dong</au><au>Li, Xiaojie</au><au>Qin, Pengda</au><au>Gan, Tian</au><au>Nie, Liqiang</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Preview-based Category Contrastive Learning for Knowledge Distillation</atitle><jtitle>arXiv.org</jtitle><date>2024-10-18</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Knowledge distillation is a mainstream algorithm in model compression by transferring knowledge from the larger model (teacher) to the smaller model (student) to improve the performance of student. Despite many efforts, existing methods mainly investigate the consistency between instance-level feature representation or prediction, which neglects the category-level information and the difficulty of each sample, leading to undesirable performance. To address these issues, we propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD). It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers in a contrastive learning fashion, which can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories, contributing to discriminative category centers and better classification results. Besides, we introduce a novel preview strategy to dynamically determine how much the student should learn from each sample according to their difficulty. Different from existing methods that treat all samples equally and curriculum learning that simply filters out hard samples, our method assigns a small weight for hard instances as a preview to better guide the student training. Extensive experiments on several challenging datasets, including CIFAR-100 and ImageNet, demonstrate the superiority over state-of-the-art methods.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-10
issn 2331-8422
language eng
recordid cdi_proquest_journals_3118928508
source ProQuest - Publicly Available Content Database
subjects Algorithms
Knowledge representation
Machine learning
Performance enhancement
title Preview-based Category Contrastive Learning for Knowledge Distillation
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T04%3A50%3A44IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Preview-based%20Category%20Contrastive%20Learning%20for%20Knowledge%20Distillation&rft.jtitle=arXiv.org&rft.au=Ding,%20Muhe&rft.date=2024-10-18&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3118928508%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_31189285083%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3118928508&rft_id=info:pmid/&rfr_iscdi=true