Loading…

ABC-Norm Regularization for Fine-Grained and Long-Tailed Image Classification

Image classification for real-world applications often involves complicated data distributions such as fine-grained and long-tailed. To address the two challenging issues simultaneously, we propose a new regularization technique that yields an adversarial loss to strengthen the model learning. Speci...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on image processing 2023-01, Vol.32, p.1-1
Main Authors: Hsu, Yen-Chi, Hong, Cheng-Yao, Lee, Ming-Sui, Geiger, Davi, Liu, Tyng-Luh
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c348t-eea8faeb0b5a39fbd5bb36ea3a5aa89f833f58d1769d3018aeb4153bd2051bce3
cites cdi_FETCH-LOGICAL-c348t-eea8faeb0b5a39fbd5bb36ea3a5aa89f833f58d1769d3018aeb4153bd2051bce3
container_end_page 1
container_issue
container_start_page 1
container_title IEEE transactions on image processing
container_volume 32
creator Hsu, Yen-Chi
Hong, Cheng-Yao
Lee, Ming-Sui
Geiger, Davi
Liu, Tyng-Luh
description Image classification for real-world applications often involves complicated data distributions such as fine-grained and long-tailed. To address the two challenging issues simultaneously, we propose a new regularization technique that yields an adversarial loss to strengthen the model learning. Specifically, for each training batch, we construct an adaptive batch prediction (ABP) matrix and establish its corresponding adaptive batch confusion norm (ABC-Norm). The ABP matrix is a composition of two parts, including an adaptive component to class-wise encode the imbalanced data distribution, and the other component to batch-wise assess the softmax predictions. The ABC-Norm leads to a norm-based regularization loss, which can be theoretically shown to be an upper bound for an objective function closely related to rank minimization. By coupling with the conventional cross-entropy loss, the ABC-Norm regularization could introduce adaptive classification confusion and thus trigger adversarial learning to improve the effectiveness of model learning. Different from most of state-of-the-art techniques in solving either fine-grained or long-tailed problems, our method is characterized with its simple and efficient design, and most distinctively, provides a unified solution. In the experiments, we compare ABC-Norm with relevant techniques and demonstrate its efficacy on several benchmark datasets, including (CUB-LT, iNaturalist2018); (CUB, CAR, AIR); and (ImageNet-LT), which respectively correspond to the real-world, fine-grained, and long-tailed scenarios.
doi_str_mv 10.1109/TIP.2023.3273455
format article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmed_primary_37432822</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10179261</ieee_id><sourcerecordid>2836295898</sourcerecordid><originalsourceid>FETCH-LOGICAL-c348t-eea8faeb0b5a39fbd5bb36ea3a5aa89f833f58d1769d3018aeb4153bd2051bce3</originalsourceid><addsrcrecordid>eNpdkE1LxDAQhoMoft89iBS8eOmaZJo2Oeri6sL6gaznMm2nS6QfmmwP-uvNuquIp5mB530ZHsZOBB8Jwc3lfPo0klzCCGQGiVJbbF-YRMScJ3I77FxlcSYSs8cOvH_lXCRKpLtsD7IEpJZyn91fXY_jh9610TMthgad_cSl7buo7l00sR3Ftw7DqCLsqmjWd4t4jrYJ97TFBUXjBr23tS2_U0dsp8bG0_FmHrKXyc18fBfPHm-n46tZXEKilzER6hqp4IVCMHVRqaKAlBBQIWpTa4Ba6UpkqamACx3QRCgoKsmVKEqCQ3ax7n1z_ftAfpm31pfUNNhRP_hcakilUdrogJ7_Q1_7wXXhuxUV5HDgPFB8TZWu995Rnb8526L7yAXPV6rzoDpfqc43qkPkbFM8FC1Vv4EftwE4XQOWiP70iczIVMAXC4GBYQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2837140300</pqid></control><display><type>article</type><title>ABC-Norm Regularization for Fine-Grained and Long-Tailed Image Classification</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Hsu, Yen-Chi ; Hong, Cheng-Yao ; Lee, Ming-Sui ; Geiger, Davi ; Liu, Tyng-Luh</creator><creatorcontrib>Hsu, Yen-Chi ; Hong, Cheng-Yao ; Lee, Ming-Sui ; Geiger, Davi ; Liu, Tyng-Luh</creatorcontrib><description>Image classification for real-world applications often involves complicated data distributions such as fine-grained and long-tailed. To address the two challenging issues simultaneously, we propose a new regularization technique that yields an adversarial loss to strengthen the model learning. Specifically, for each training batch, we construct an adaptive batch prediction (ABP) matrix and establish its corresponding adaptive batch confusion norm (ABC-Norm). The ABP matrix is a composition of two parts, including an adaptive component to class-wise encode the imbalanced data distribution, and the other component to batch-wise assess the softmax predictions. The ABC-Norm leads to a norm-based regularization loss, which can be theoretically shown to be an upper bound for an objective function closely related to rank minimization. By coupling with the conventional cross-entropy loss, the ABC-Norm regularization could introduce adaptive classification confusion and thus trigger adversarial learning to improve the effectiveness of model learning. Different from most of state-of-the-art techniques in solving either fine-grained or long-tailed problems, our method is characterized with its simple and efficient design, and most distinctively, provides a unified solution. In the experiments, we compare ABC-Norm with relevant techniques and demonstrate its efficacy on several benchmark datasets, including (CUB-LT, iNaturalist2018); (CUB, CAR, AIR); and (ImageNet-LT), which respectively correspond to the real-world, fine-grained, and long-tailed scenarios.</description><identifier>ISSN: 1057-7149</identifier><identifier>EISSN: 1941-0042</identifier><identifier>DOI: 10.1109/TIP.2023.3273455</identifier><identifier>PMID: 37432822</identifier><identifier>CODEN: IIPRE4</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Adaptation models ; Classification ; Confusion ; Data models ; deep neural network ; fine-grained ; Image classification ; Learning ; long-tailed ; Predictive models ; Regularization ; Tail ; Task analysis ; Training ; Upper bounds ; Visualization</subject><ispartof>IEEE transactions on image processing, 2023-01, Vol.32, p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c348t-eea8faeb0b5a39fbd5bb36ea3a5aa89f833f58d1769d3018aeb4153bd2051bce3</citedby><cites>FETCH-LOGICAL-c348t-eea8faeb0b5a39fbd5bb36ea3a5aa89f833f58d1769d3018aeb4153bd2051bce3</cites><orcidid>0000-0003-0694-4400 ; 0000-0002-6699-6694 ; 0000-0001-8804-4582 ; 0000-0002-8366-5213 ; 0000-0003-2138-3517</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10179261$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37432822$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Hsu, Yen-Chi</creatorcontrib><creatorcontrib>Hong, Cheng-Yao</creatorcontrib><creatorcontrib>Lee, Ming-Sui</creatorcontrib><creatorcontrib>Geiger, Davi</creatorcontrib><creatorcontrib>Liu, Tyng-Luh</creatorcontrib><title>ABC-Norm Regularization for Fine-Grained and Long-Tailed Image Classification</title><title>IEEE transactions on image processing</title><addtitle>TIP</addtitle><addtitle>IEEE Trans Image Process</addtitle><description>Image classification for real-world applications often involves complicated data distributions such as fine-grained and long-tailed. To address the two challenging issues simultaneously, we propose a new regularization technique that yields an adversarial loss to strengthen the model learning. Specifically, for each training batch, we construct an adaptive batch prediction (ABP) matrix and establish its corresponding adaptive batch confusion norm (ABC-Norm). The ABP matrix is a composition of two parts, including an adaptive component to class-wise encode the imbalanced data distribution, and the other component to batch-wise assess the softmax predictions. The ABC-Norm leads to a norm-based regularization loss, which can be theoretically shown to be an upper bound for an objective function closely related to rank minimization. By coupling with the conventional cross-entropy loss, the ABC-Norm regularization could introduce adaptive classification confusion and thus trigger adversarial learning to improve the effectiveness of model learning. Different from most of state-of-the-art techniques in solving either fine-grained or long-tailed problems, our method is characterized with its simple and efficient design, and most distinctively, provides a unified solution. In the experiments, we compare ABC-Norm with relevant techniques and demonstrate its efficacy on several benchmark datasets, including (CUB-LT, iNaturalist2018); (CUB, CAR, AIR); and (ImageNet-LT), which respectively correspond to the real-world, fine-grained, and long-tailed scenarios.</description><subject>Adaptation models</subject><subject>Classification</subject><subject>Confusion</subject><subject>Data models</subject><subject>deep neural network</subject><subject>fine-grained</subject><subject>Image classification</subject><subject>Learning</subject><subject>long-tailed</subject><subject>Predictive models</subject><subject>Regularization</subject><subject>Tail</subject><subject>Task analysis</subject><subject>Training</subject><subject>Upper bounds</subject><subject>Visualization</subject><issn>1057-7149</issn><issn>1941-0042</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpdkE1LxDAQhoMoft89iBS8eOmaZJo2Oeri6sL6gaznMm2nS6QfmmwP-uvNuquIp5mB530ZHsZOBB8Jwc3lfPo0klzCCGQGiVJbbF-YRMScJ3I77FxlcSYSs8cOvH_lXCRKpLtsD7IEpJZyn91fXY_jh9610TMthgad_cSl7buo7l00sR3Ftw7DqCLsqmjWd4t4jrYJ97TFBUXjBr23tS2_U0dsp8bG0_FmHrKXyc18fBfPHm-n46tZXEKilzER6hqp4IVCMHVRqaKAlBBQIWpTa4Ba6UpkqamACx3QRCgoKsmVKEqCQ3ax7n1z_ftAfpm31pfUNNhRP_hcakilUdrogJ7_Q1_7wXXhuxUV5HDgPFB8TZWu995Rnb8526L7yAXPV6rzoDpfqc43qkPkbFM8FC1Vv4EftwE4XQOWiP70iczIVMAXC4GBYQ</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Hsu, Yen-Chi</creator><creator>Hong, Cheng-Yao</creator><creator>Lee, Ming-Sui</creator><creator>Geiger, Davi</creator><creator>Liu, Tyng-Luh</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-0694-4400</orcidid><orcidid>https://orcid.org/0000-0002-6699-6694</orcidid><orcidid>https://orcid.org/0000-0001-8804-4582</orcidid><orcidid>https://orcid.org/0000-0002-8366-5213</orcidid><orcidid>https://orcid.org/0000-0003-2138-3517</orcidid></search><sort><creationdate>20230101</creationdate><title>ABC-Norm Regularization for Fine-Grained and Long-Tailed Image Classification</title><author>Hsu, Yen-Chi ; Hong, Cheng-Yao ; Lee, Ming-Sui ; Geiger, Davi ; Liu, Tyng-Luh</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c348t-eea8faeb0b5a39fbd5bb36ea3a5aa89f833f58d1769d3018aeb4153bd2051bce3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Adaptation models</topic><topic>Classification</topic><topic>Confusion</topic><topic>Data models</topic><topic>deep neural network</topic><topic>fine-grained</topic><topic>Image classification</topic><topic>Learning</topic><topic>long-tailed</topic><topic>Predictive models</topic><topic>Regularization</topic><topic>Tail</topic><topic>Task analysis</topic><topic>Training</topic><topic>Upper bounds</topic><topic>Visualization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hsu, Yen-Chi</creatorcontrib><creatorcontrib>Hong, Cheng-Yao</creatorcontrib><creatorcontrib>Lee, Ming-Sui</creatorcontrib><creatorcontrib>Geiger, Davi</creatorcontrib><creatorcontrib>Liu, Tyng-Luh</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEL</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transactions on image processing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hsu, Yen-Chi</au><au>Hong, Cheng-Yao</au><au>Lee, Ming-Sui</au><au>Geiger, Davi</au><au>Liu, Tyng-Luh</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>ABC-Norm Regularization for Fine-Grained and Long-Tailed Image Classification</atitle><jtitle>IEEE transactions on image processing</jtitle><stitle>TIP</stitle><addtitle>IEEE Trans Image Process</addtitle><date>2023-01-01</date><risdate>2023</risdate><volume>32</volume><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>1057-7149</issn><eissn>1941-0042</eissn><coden>IIPRE4</coden><abstract>Image classification for real-world applications often involves complicated data distributions such as fine-grained and long-tailed. To address the two challenging issues simultaneously, we propose a new regularization technique that yields an adversarial loss to strengthen the model learning. Specifically, for each training batch, we construct an adaptive batch prediction (ABP) matrix and establish its corresponding adaptive batch confusion norm (ABC-Norm). The ABP matrix is a composition of two parts, including an adaptive component to class-wise encode the imbalanced data distribution, and the other component to batch-wise assess the softmax predictions. The ABC-Norm leads to a norm-based regularization loss, which can be theoretically shown to be an upper bound for an objective function closely related to rank minimization. By coupling with the conventional cross-entropy loss, the ABC-Norm regularization could introduce adaptive classification confusion and thus trigger adversarial learning to improve the effectiveness of model learning. Different from most of state-of-the-art techniques in solving either fine-grained or long-tailed problems, our method is characterized with its simple and efficient design, and most distinctively, provides a unified solution. In the experiments, we compare ABC-Norm with relevant techniques and demonstrate its efficacy on several benchmark datasets, including (CUB-LT, iNaturalist2018); (CUB, CAR, AIR); and (ImageNet-LT), which respectively correspond to the real-world, fine-grained, and long-tailed scenarios.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>37432822</pmid><doi>10.1109/TIP.2023.3273455</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0003-0694-4400</orcidid><orcidid>https://orcid.org/0000-0002-6699-6694</orcidid><orcidid>https://orcid.org/0000-0001-8804-4582</orcidid><orcidid>https://orcid.org/0000-0002-8366-5213</orcidid><orcidid>https://orcid.org/0000-0003-2138-3517</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1057-7149
ispartof IEEE transactions on image processing, 2023-01, Vol.32, p.1-1
issn 1057-7149
1941-0042
language eng
recordid cdi_pubmed_primary_37432822
source IEEE Electronic Library (IEL) Journals
subjects Adaptation models
Classification
Confusion
Data models
deep neural network
fine-grained
Image classification
Learning
long-tailed
Predictive models
Regularization
Tail
Task analysis
Training
Upper bounds
Visualization
title ABC-Norm Regularization for Fine-Grained and Long-Tailed Image Classification
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T12%3A56%3A41IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=ABC-Norm%20Regularization%20for%20Fine-Grained%20and%20Long-Tailed%20Image%20Classification&rft.jtitle=IEEE%20transactions%20on%20image%20processing&rft.au=Hsu,%20Yen-Chi&rft.date=2023-01-01&rft.volume=32&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=1057-7149&rft.eissn=1941-0042&rft.coden=IIPRE4&rft_id=info:doi/10.1109/TIP.2023.3273455&rft_dat=%3Cproquest_pubme%3E2836295898%3C/proquest_pubme%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c348t-eea8faeb0b5a39fbd5bb36ea3a5aa89f833f58d1769d3018aeb4153bd2051bce3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2837140300&rft_id=info:pmid/37432822&rft_ieee_id=10179261&rfr_iscdi=true