Loading…

APM: Adaptive parameter multiplexing for class incremental learning

In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and ad...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications 2024-12, Vol.258, p.125135, Article 125135
Main Authors: Gao, Jinghan, Xie, Tao, Li, Ruifeng, Wang, Ke, Zhao, Lijun
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483
container_end_page
container_issue
container_start_page 125135
container_title Expert systems with applications
container_volume 258
creator Gao, Jinghan
Xie, Tao
Li, Ruifeng
Wang, Ke
Zhao, Lijun
description In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field. •Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity
doi_str_mv 10.1016/j.eswa.2024.125135
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_eswa_2024_125135</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417424020025</els_id><sourcerecordid>S0957417424020025</sourcerecordid><originalsourceid>FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483</originalsourceid><addsrcrecordid>eNp9j7tOxDAQRV2AxLLwA1T-gQSPY29iRBOteEmLoIDacpwJcuQ8ZIcF_p5EoaYY3WJ0ru4h5ApYCgx2122K8cuknHGRApeQyROyYUrmiYBcnJHzGFvGIGcs35B9-fp8Q8vajJM7Ih1NMB1OGGj36Sc3evx2_QdthkCtNzFS19uAHfaT8dSjCf38viCnjfERL_9yS97v7972j8nh5eFpXx4SCwVMiRFVpbiUKGvDGIqcVwIzAVgUMgNlmx1vKpvPl6GwDZdzSFQKVD2Dosi2hK-9NgwxBmz0GFxnwo8Gphd13epFXS_qelWfodsVwnnZ0WHQ0TrsLdYuoJ10Pbj_8F-LtmUJ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>APM: Adaptive parameter multiplexing for class incremental learning</title><source>ScienceDirect Journals</source><creator>Gao, Jinghan ; Xie, Tao ; Li, Ruifeng ; Wang, Ke ; Zhao, Lijun</creator><creatorcontrib>Gao, Jinghan ; Xie, Tao ; Li, Ruifeng ; Wang, Ke ; Zhao, Lijun</creatorcontrib><description>In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field. •Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity respectively.•Regularization loss cuts storage and boosts incremental classification.•Experiments show APM outperforms all current algorithms.</description><identifier>ISSN: 0957-4174</identifier><identifier>DOI: 10.1016/j.eswa.2024.125135</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Catastrophic forgetting ; Class incremental learning ; Deep learning ; Image classification ; Incremental learning</subject><ispartof>Expert systems with applications, 2024-12, Vol.258, p.125135, Article 125135</ispartof><rights>2024 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483</cites><orcidid>0000-0002-0062-5225 ; 0000-0001-9315-5180 ; 0000-0002-1383-7745 ; 0000-0003-1034-4221 ; 0000-0002-5615-0847</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27922,27923</link.rule.ids></links><search><creatorcontrib>Gao, Jinghan</creatorcontrib><creatorcontrib>Xie, Tao</creatorcontrib><creatorcontrib>Li, Ruifeng</creatorcontrib><creatorcontrib>Wang, Ke</creatorcontrib><creatorcontrib>Zhao, Lijun</creatorcontrib><title>APM: Adaptive parameter multiplexing for class incremental learning</title><title>Expert systems with applications</title><description>In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field. •Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity respectively.•Regularization loss cuts storage and boosts incremental classification.•Experiments show APM outperforms all current algorithms.</description><subject>Catastrophic forgetting</subject><subject>Class incremental learning</subject><subject>Deep learning</subject><subject>Image classification</subject><subject>Incremental learning</subject><issn>0957-4174</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9j7tOxDAQRV2AxLLwA1T-gQSPY29iRBOteEmLoIDacpwJcuQ8ZIcF_p5EoaYY3WJ0ru4h5ApYCgx2122K8cuknHGRApeQyROyYUrmiYBcnJHzGFvGIGcs35B9-fp8Q8vajJM7Ih1NMB1OGGj36Sc3evx2_QdthkCtNzFS19uAHfaT8dSjCf38viCnjfERL_9yS97v7972j8nh5eFpXx4SCwVMiRFVpbiUKGvDGIqcVwIzAVgUMgNlmx1vKpvPl6GwDZdzSFQKVD2Dosi2hK-9NgwxBmz0GFxnwo8Gphd13epFXS_qelWfodsVwnnZ0WHQ0TrsLdYuoJ10Pbj_8F-LtmUJ</recordid><startdate>20241215</startdate><enddate>20241215</enddate><creator>Gao, Jinghan</creator><creator>Xie, Tao</creator><creator>Li, Ruifeng</creator><creator>Wang, Ke</creator><creator>Zhao, Lijun</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-0062-5225</orcidid><orcidid>https://orcid.org/0000-0001-9315-5180</orcidid><orcidid>https://orcid.org/0000-0002-1383-7745</orcidid><orcidid>https://orcid.org/0000-0003-1034-4221</orcidid><orcidid>https://orcid.org/0000-0002-5615-0847</orcidid></search><sort><creationdate>20241215</creationdate><title>APM: Adaptive parameter multiplexing for class incremental learning</title><author>Gao, Jinghan ; Xie, Tao ; Li, Ruifeng ; Wang, Ke ; Zhao, Lijun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Catastrophic forgetting</topic><topic>Class incremental learning</topic><topic>Deep learning</topic><topic>Image classification</topic><topic>Incremental learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Gao, Jinghan</creatorcontrib><creatorcontrib>Xie, Tao</creatorcontrib><creatorcontrib>Li, Ruifeng</creatorcontrib><creatorcontrib>Wang, Ke</creatorcontrib><creatorcontrib>Zhao, Lijun</creatorcontrib><collection>CrossRef</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Gao, Jinghan</au><au>Xie, Tao</au><au>Li, Ruifeng</au><au>Wang, Ke</au><au>Zhao, Lijun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>APM: Adaptive parameter multiplexing for class incremental learning</atitle><jtitle>Expert systems with applications</jtitle><date>2024-12-15</date><risdate>2024</risdate><volume>258</volume><spage>125135</spage><pages>125135-</pages><artnum>125135</artnum><issn>0957-4174</issn><abstract>In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field. •Introducing novel adaptive parameter multiplexing for class incremental learning.•Effectively mitigating catastrophic forgetting.•Parameter multiplexing and increment enhance stability and plasticity respectively.•Regularization loss cuts storage and boosts incremental classification.•Experiments show APM outperforms all current algorithms.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2024.125135</doi><orcidid>https://orcid.org/0000-0002-0062-5225</orcidid><orcidid>https://orcid.org/0000-0001-9315-5180</orcidid><orcidid>https://orcid.org/0000-0002-1383-7745</orcidid><orcidid>https://orcid.org/0000-0003-1034-4221</orcidid><orcidid>https://orcid.org/0000-0002-5615-0847</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0957-4174
ispartof Expert systems with applications, 2024-12, Vol.258, p.125135, Article 125135
issn 0957-4174
language eng
recordid cdi_crossref_primary_10_1016_j_eswa_2024_125135
source ScienceDirect Journals
subjects Catastrophic forgetting
Class incremental learning
Deep learning
Image classification
Incremental learning
title APM: Adaptive parameter multiplexing for class incremental learning
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T23%3A37%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=APM:%20Adaptive%20parameter%20multiplexing%20for%20class%20incremental%20learning&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Gao,%20Jinghan&rft.date=2024-12-15&rft.volume=258&rft.spage=125135&rft.pages=125135-&rft.artnum=125135&rft.issn=0957-4174&rft_id=info:doi/10.1016/j.eswa.2024.125135&rft_dat=%3Celsevier_cross%3ES0957417424020025%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c181t-a4bb9255e5da00e472b4e341e885319cf62fbc7fbc3e4cf253e45e9919dbb9483%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true