Loading…
APM: Adaptive parameter multiplexing for class incremental learning
In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and ad...
Saved in:
Published in: | Expert systems with applications 2024-12, Vol.258, p.125135, Article 125135 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483 |
container_end_page | |
container_issue | |
container_start_page | 125135 |
container_title | Expert systems with applications |
container_volume | 258 |
creator | Gao, Jinghan Xie, Tao Li, Ruifeng Wang, Ke Zhao, Lijun |
description | In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field.
•Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity |
doi_str_mv | 10.1016/j.eswa.2024.125135 |
format | article |
fullrecord | <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_eswa_2024_125135</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417424020025</els_id><sourcerecordid>S0957417424020025</sourcerecordid><originalsourceid>FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483</originalsourceid><addsrcrecordid>eNp9j7tOxDAQRV2AxLLwA1T-gQSPY29iRBOteEmLoIDacpwJcuQ8ZIcF_p5EoaYY3WJ0ru4h5ApYCgx2122K8cuknHGRApeQyROyYUrmiYBcnJHzGFvGIGcs35B9-fp8Q8vajJM7Ih1NMB1OGGj36Sc3evx2_QdthkCtNzFS19uAHfaT8dSjCf38viCnjfERL_9yS97v7972j8nh5eFpXx4SCwVMiRFVpbiUKGvDGIqcVwIzAVgUMgNlmx1vKpvPl6GwDZdzSFQKVD2Dosi2hK-9NgwxBmz0GFxnwo8Gphd13epFXS_qelWfodsVwnnZ0WHQ0TrsLdYuoJ10Pbj_8F-LtmUJ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>APM: Adaptive parameter multiplexing for class incremental learning</title><source>ScienceDirect Journals</source><creator>Gao, Jinghan ; Xie, Tao ; Li, Ruifeng ; Wang, Ke ; Zhao, Lijun</creator><creatorcontrib>Gao, Jinghan ; Xie, Tao ; Li, Ruifeng ; Wang, Ke ; Zhao, Lijun</creatorcontrib><description>In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field.
•Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity respectively.•Regularization loss cuts storage and boosts incremental classification.•Experiments show APM outperforms all current algorithms.</description><identifier>ISSN: 0957-4174</identifier><identifier>DOI: 10.1016/j.eswa.2024.125135</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Catastrophic forgetting ; Class incremental learning ; Deep learning ; Image classification ; Incremental learning</subject><ispartof>Expert systems with applications, 2024-12, Vol.258, p.125135, Article 125135</ispartof><rights>2024 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483</cites><orcidid>0000-0002-0062-5225 ; 0000-0001-9315-5180 ; 0000-0002-1383-7745 ; 0000-0003-1034-4221 ; 0000-0002-5615-0847</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27922,27923</link.rule.ids></links><search><creatorcontrib>Gao, Jinghan</creatorcontrib><creatorcontrib>Xie, Tao</creatorcontrib><creatorcontrib>Li, Ruifeng</creatorcontrib><creatorcontrib>Wang, Ke</creatorcontrib><creatorcontrib>Zhao, Lijun</creatorcontrib><title>APM: Adaptive parameter multiplexing for class incremental learning</title><title>Expert systems with applications</title><description>In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field.
•Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity respectively.•Regularization loss cuts storage and boosts incremental classification.•Experiments show APM outperforms all current algorithms.</description><subject>Catastrophic forgetting</subject><subject>Class incremental learning</subject><subject>Deep learning</subject><subject>Image classification</subject><subject>Incremental learning</subject><issn>0957-4174</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9j7tOxDAQRV2AxLLwA1T-gQSPY29iRBOteEmLoIDacpwJcuQ8ZIcF_p5EoaYY3WJ0ru4h5ApYCgx2122K8cuknHGRApeQyROyYUrmiYBcnJHzGFvGIGcs35B9-fp8Q8vajJM7Ih1NMB1OGGj36Sc3evx2_QdthkCtNzFS19uAHfaT8dSjCf38viCnjfERL_9yS97v7972j8nh5eFpXx4SCwVMiRFVpbiUKGvDGIqcVwIzAVgUMgNlmx1vKpvPl6GwDZdzSFQKVD2Dosi2hK-9NgwxBmz0GFxnwo8Gphd13epFXS_qelWfodsVwnnZ0WHQ0TrsLdYuoJ10Pbj_8F-LtmUJ</recordid><startdate>20241215</startdate><enddate>20241215</enddate><creator>Gao, Jinghan</creator><creator>Xie, Tao</creator><creator>Li, Ruifeng</creator><creator>Wang, Ke</creator><creator>Zhao, Lijun</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-0062-5225</orcidid><orcidid>https://orcid.org/0000-0001-9315-5180</orcidid><orcidid>https://orcid.org/0000-0002-1383-7745</orcidid><orcidid>https://orcid.org/0000-0003-1034-4221</orcidid><orcidid>https://orcid.org/0000-0002-5615-0847</orcidid></search><sort><creationdate>20241215</creationdate><title>APM: Adaptive parameter multiplexing for class incremental learning</title><author>Gao, Jinghan ; Xie, Tao ; Li, Ruifeng ; Wang, Ke ; Zhao, Lijun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Catastrophic forgetting</topic><topic>Class incremental learning</topic><topic>Deep learning</topic><topic>Image classification</topic><topic>Incremental learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Gao, Jinghan</creatorcontrib><creatorcontrib>Xie, Tao</creatorcontrib><creatorcontrib>Li, Ruifeng</creatorcontrib><creatorcontrib>Wang, Ke</creatorcontrib><creatorcontrib>Zhao, Lijun</creatorcontrib><collection>CrossRef</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Gao, Jinghan</au><au>Xie, Tao</au><au>Li, Ruifeng</au><au>Wang, Ke</au><au>Zhao, Lijun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>APM: Adaptive parameter multiplexing for class incremental learning</atitle><jtitle>Expert systems with applications</jtitle><date>2024-12-15</date><risdate>2024</risdate><volume>258</volume><spage>125135</spage><pages>125135-</pages><artnum>125135</artnum><issn>0957-4174</issn><abstract>In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field.
•Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity respectively.•Regularization loss cuts storage and boosts incremental classification.•Experiments show APM outperforms all current algorithms.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2024.125135</doi><orcidid>https://orcid.org/0000-0002-0062-5225</orcidid><orcidid>https://orcid.org/0000-0001-9315-5180</orcidid><orcidid>https://orcid.org/0000-0002-1383-7745</orcidid><orcidid>https://orcid.org/0000-0003-1034-4221</orcidid><orcidid>https://orcid.org/0000-0002-5615-0847</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0957-4174 |
ispartof | Expert systems with applications, 2024-12, Vol.258, p.125135, Article 125135 |
issn | 0957-4174 |
language | eng |
recordid | cdi_crossref_primary_10_1016_j_eswa_2024_125135 |
source | ScienceDirect Journals |
subjects | Catastrophic forgetting Class incremental learning Deep learning Image classification Incremental learning |
title | APM: Adaptive parameter multiplexing for class incremental learning |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T23%3A37%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=APM:%20Adaptive%20parameter%20multiplexing%20for%20class%20incremental%20learning&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Gao,%20Jinghan&rft.date=2024-12-15&rft.volume=258&rft.spage=125135&rft.pages=125135-&rft.artnum=125135&rft.issn=0957-4174&rft_id=info:doi/10.1016/j.eswa.2024.125135&rft_dat=%3Celsevier_cross%3ES0957417424020025%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |