Loading…
Semantic Multigranularity Feature Learning for High-Resolution Remote Sensing Image Scene Classification
High-resolution remote sensing image scene classification is a challenging visual task due to the large intravariance and small intervariance between the categories. To accurately recognize the scene categories, it is essential to learn discriminative features from both global and local critical reg...
Saved in:
Published in: | Applied sciences 2021-10, Vol.11 (19), p.9204 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c322t-4d3b289190a830cba831d2f287563ad83e95fe8ca6b8e0386d93ad0a6da5cc3d3 |
container_end_page | |
container_issue | 19 |
container_start_page | 9204 |
container_title | Applied sciences |
container_volume | 11 |
creator | Ma, Xinyi Xiao, Zhifeng Yun, Hong-sik Lee, Seung-Jun |
description | High-resolution remote sensing image scene classification is a challenging visual task due to the large intravariance and small intervariance between the categories. To accurately recognize the scene categories, it is essential to learn discriminative features from both global and local critical regions. Recent efforts focus on how to encourage the network to learn multigranularity features with the destruction of the spatial information on the input image at different scales, which leads to meaningless edges that are harmful to training. In this study, we propose a novel method named Semantic Multigranularity Feature Learning Network (SMGFL-Net) for remote sensing image scene classification. The core idea is to learn both global and multigranularity local features from rearranged intermediate feature maps, thus, eliminating the meaningless edges. These features are then fused for the final prediction. Our proposed framework is compared with a collection of state-of-the-art (SOTA) methods on two fine-grained remote sensing image scene datasets, including the NWPU-RESISC45 and Aerial Image Datasets (AID). We justify several design choices, including the branch granularities, fusion strategies, pooling operations, and necessity of feature map rearrangement through a comparative study. Moreover, the overall performance results show that SMGFL-Net consistently outperforms other peer methods in classification accuracy, and the superiority is more apparent with less training data, demonstrating the efficacy of feature learning of our approach. |
doi_str_mv | 10.3390/app11199204 |
format | article |
fullrecord | <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_2a41437c57394308b3926dac86427aa6</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><doaj_id>oai_doaj_org_article_2a41437c57394308b3926dac86427aa6</doaj_id><sourcerecordid>2580964963</sourcerecordid><originalsourceid>FETCH-LOGICAL-c322t-4d3b289190a830cba831d2f287563ad83e95fe8ca6b8e0386d93ad0a6da5cc3d3</originalsourceid><addsrcrecordid>eNpNUU1LxDAQLaKg6J78AwWPUk0ybZocZfFjYUXw4xxm02k3S7dZk_bgvzfrijiH-Xg83szwsuySsxsAzW5xt-Ocay1YeZSdCVbLAkpeH__rT7NZjBuWQnNQnJ1l6zfa4jA6mz9P_ei6gMPUY3DjV_5AOE6B8iVhGNzQ5a0P-ZPr1sUrRd9Po_ND_kpbP1L-RkPcUxZb7NJkaaB83mOMrnUW98yL7KTFPtLst55nHw_37_OnYvnyuJjfLQsLQoxF2cBKKM01QwXMrlLmjWiFqisJ2CggXbWkLMqVIgZKNjrBDGWDlbXQwHm2OOg2HjdmF9wWw5fx6MwP4ENnMKR_ezICS15CbasadAlMrUCLpGOVLEWNKJPW1UFrF_znRHE0Gz-FIZ1vRKWYlqWWkFjXB5YNPsZA7d9WzszeGfPPGfgGKLeAxA</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2580964963</pqid></control><display><type>article</type><title>Semantic Multigranularity Feature Learning for High-Resolution Remote Sensing Image Scene Classification</title><source>Publicly Available Content Database</source><creator>Ma, Xinyi ; Xiao, Zhifeng ; Yun, Hong-sik ; Lee, Seung-Jun</creator><creatorcontrib>Ma, Xinyi ; Xiao, Zhifeng ; Yun, Hong-sik ; Lee, Seung-Jun</creatorcontrib><description>High-resolution remote sensing image scene classification is a challenging visual task due to the large intravariance and small intervariance between the categories. To accurately recognize the scene categories, it is essential to learn discriminative features from both global and local critical regions. Recent efforts focus on how to encourage the network to learn multigranularity features with the destruction of the spatial information on the input image at different scales, which leads to meaningless edges that are harmful to training. In this study, we propose a novel method named Semantic Multigranularity Feature Learning Network (SMGFL-Net) for remote sensing image scene classification. The core idea is to learn both global and multigranularity local features from rearranged intermediate feature maps, thus, eliminating the meaningless edges. These features are then fused for the final prediction. Our proposed framework is compared with a collection of state-of-the-art (SOTA) methods on two fine-grained remote sensing image scene datasets, including the NWPU-RESISC45 and Aerial Image Datasets (AID). We justify several design choices, including the branch granularities, fusion strategies, pooling operations, and necessity of feature map rearrangement through a comparative study. Moreover, the overall performance results show that SMGFL-Net consistently outperforms other peer methods in classification accuracy, and the superiority is more apparent with less training data, demonstrating the efficacy of feature learning of our approach.</description><identifier>ISSN: 2076-3417</identifier><identifier>EISSN: 2076-3417</identifier><identifier>DOI: 10.3390/app11199204</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Classification ; Comparative studies ; Cooperative learning ; Datasets ; Feature maps ; fine-grained ; High resolution ; Image classification ; Learning ; Localization ; Methods ; multigranularity ; Neural networks ; Remote sensing ; scene classification ; Semantics ; Spatial data ; Visual tasks</subject><ispartof>Applied sciences, 2021-10, Vol.11 (19), p.9204</ispartof><rights>2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c322t-4d3b289190a830cba831d2f287563ad83e95fe8ca6b8e0386d93ad0a6da5cc3d3</cites><orcidid>0000-0002-7230-376X ; 0000-0003-3327-8108 ; 0000-0003-4350-990X ; 0000-0002-2104-9423</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2580964963/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2580964963?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,25753,27924,27925,37012,44590,75126</link.rule.ids></links><search><creatorcontrib>Ma, Xinyi</creatorcontrib><creatorcontrib>Xiao, Zhifeng</creatorcontrib><creatorcontrib>Yun, Hong-sik</creatorcontrib><creatorcontrib>Lee, Seung-Jun</creatorcontrib><title>Semantic Multigranularity Feature Learning for High-Resolution Remote Sensing Image Scene Classification</title><title>Applied sciences</title><description>High-resolution remote sensing image scene classification is a challenging visual task due to the large intravariance and small intervariance between the categories. To accurately recognize the scene categories, it is essential to learn discriminative features from both global and local critical regions. Recent efforts focus on how to encourage the network to learn multigranularity features with the destruction of the spatial information on the input image at different scales, which leads to meaningless edges that are harmful to training. In this study, we propose a novel method named Semantic Multigranularity Feature Learning Network (SMGFL-Net) for remote sensing image scene classification. The core idea is to learn both global and multigranularity local features from rearranged intermediate feature maps, thus, eliminating the meaningless edges. These features are then fused for the final prediction. Our proposed framework is compared with a collection of state-of-the-art (SOTA) methods on two fine-grained remote sensing image scene datasets, including the NWPU-RESISC45 and Aerial Image Datasets (AID). We justify several design choices, including the branch granularities, fusion strategies, pooling operations, and necessity of feature map rearrangement through a comparative study. Moreover, the overall performance results show that SMGFL-Net consistently outperforms other peer methods in classification accuracy, and the superiority is more apparent with less training data, demonstrating the efficacy of feature learning of our approach.</description><subject>Classification</subject><subject>Comparative studies</subject><subject>Cooperative learning</subject><subject>Datasets</subject><subject>Feature maps</subject><subject>fine-grained</subject><subject>High resolution</subject><subject>Image classification</subject><subject>Learning</subject><subject>Localization</subject><subject>Methods</subject><subject>multigranularity</subject><subject>Neural networks</subject><subject>Remote sensing</subject><subject>scene classification</subject><subject>Semantics</subject><subject>Spatial data</subject><subject>Visual tasks</subject><issn>2076-3417</issn><issn>2076-3417</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNpNUU1LxDAQLaKg6J78AwWPUk0ybZocZfFjYUXw4xxm02k3S7dZk_bgvzfrijiH-Xg83szwsuySsxsAzW5xt-Ocay1YeZSdCVbLAkpeH__rT7NZjBuWQnNQnJ1l6zfa4jA6mz9P_ei6gMPUY3DjV_5AOE6B8iVhGNzQ5a0P-ZPr1sUrRd9Po_ND_kpbP1L-RkPcUxZb7NJkaaB83mOMrnUW98yL7KTFPtLst55nHw_37_OnYvnyuJjfLQsLQoxF2cBKKM01QwXMrlLmjWiFqisJ2CggXbWkLMqVIgZKNjrBDGWDlbXQwHm2OOg2HjdmF9wWw5fx6MwP4ENnMKR_ezICS15CbasadAlMrUCLpGOVLEWNKJPW1UFrF_znRHE0Gz-FIZ1vRKWYlqWWkFjXB5YNPsZA7d9WzszeGfPPGfgGKLeAxA</recordid><startdate>20211001</startdate><enddate>20211001</enddate><creator>Ma, Xinyi</creator><creator>Xiao, Zhifeng</creator><creator>Yun, Hong-sik</creator><creator>Lee, Seung-Jun</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-7230-376X</orcidid><orcidid>https://orcid.org/0000-0003-3327-8108</orcidid><orcidid>https://orcid.org/0000-0003-4350-990X</orcidid><orcidid>https://orcid.org/0000-0002-2104-9423</orcidid></search><sort><creationdate>20211001</creationdate><title>Semantic Multigranularity Feature Learning for High-Resolution Remote Sensing Image Scene Classification</title><author>Ma, Xinyi ; Xiao, Zhifeng ; Yun, Hong-sik ; Lee, Seung-Jun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c322t-4d3b289190a830cba831d2f287563ad83e95fe8ca6b8e0386d93ad0a6da5cc3d3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Classification</topic><topic>Comparative studies</topic><topic>Cooperative learning</topic><topic>Datasets</topic><topic>Feature maps</topic><topic>fine-grained</topic><topic>High resolution</topic><topic>Image classification</topic><topic>Learning</topic><topic>Localization</topic><topic>Methods</topic><topic>multigranularity</topic><topic>Neural networks</topic><topic>Remote sensing</topic><topic>scene classification</topic><topic>Semantics</topic><topic>Spatial data</topic><topic>Visual tasks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ma, Xinyi</creatorcontrib><creatorcontrib>Xiao, Zhifeng</creatorcontrib><creatorcontrib>Yun, Hong-sik</creatorcontrib><creatorcontrib>Lee, Seung-Jun</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Applied sciences</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ma, Xinyi</au><au>Xiao, Zhifeng</au><au>Yun, Hong-sik</au><au>Lee, Seung-Jun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Semantic Multigranularity Feature Learning for High-Resolution Remote Sensing Image Scene Classification</atitle><jtitle>Applied sciences</jtitle><date>2021-10-01</date><risdate>2021</risdate><volume>11</volume><issue>19</issue><spage>9204</spage><pages>9204-</pages><issn>2076-3417</issn><eissn>2076-3417</eissn><abstract>High-resolution remote sensing image scene classification is a challenging visual task due to the large intravariance and small intervariance between the categories. To accurately recognize the scene categories, it is essential to learn discriminative features from both global and local critical regions. Recent efforts focus on how to encourage the network to learn multigranularity features with the destruction of the spatial information on the input image at different scales, which leads to meaningless edges that are harmful to training. In this study, we propose a novel method named Semantic Multigranularity Feature Learning Network (SMGFL-Net) for remote sensing image scene classification. The core idea is to learn both global and multigranularity local features from rearranged intermediate feature maps, thus, eliminating the meaningless edges. These features are then fused for the final prediction. Our proposed framework is compared with a collection of state-of-the-art (SOTA) methods on two fine-grained remote sensing image scene datasets, including the NWPU-RESISC45 and Aerial Image Datasets (AID). We justify several design choices, including the branch granularities, fusion strategies, pooling operations, and necessity of feature map rearrangement through a comparative study. Moreover, the overall performance results show that SMGFL-Net consistently outperforms other peer methods in classification accuracy, and the superiority is more apparent with less training data, demonstrating the efficacy of feature learning of our approach.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/app11199204</doi><orcidid>https://orcid.org/0000-0002-7230-376X</orcidid><orcidid>https://orcid.org/0000-0003-3327-8108</orcidid><orcidid>https://orcid.org/0000-0003-4350-990X</orcidid><orcidid>https://orcid.org/0000-0002-2104-9423</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2076-3417 |
ispartof | Applied sciences, 2021-10, Vol.11 (19), p.9204 |
issn | 2076-3417 2076-3417 |
language | eng |
recordid | cdi_doaj_primary_oai_doaj_org_article_2a41437c57394308b3926dac86427aa6 |
source | Publicly Available Content Database |
subjects | Classification Comparative studies Cooperative learning Datasets Feature maps fine-grained High resolution Image classification Learning Localization Methods multigranularity Neural networks Remote sensing scene classification Semantics Spatial data Visual tasks |
title | Semantic Multigranularity Feature Learning for High-Resolution Remote Sensing Image Scene Classification |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T05%3A58%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Semantic%20Multigranularity%20Feature%20Learning%20for%20High-Resolution%20Remote%20Sensing%20Image%20Scene%20Classification&rft.jtitle=Applied%20sciences&rft.au=Ma,%20Xinyi&rft.date=2021-10-01&rft.volume=11&rft.issue=19&rft.spage=9204&rft.pages=9204-&rft.issn=2076-3417&rft.eissn=2076-3417&rft_id=info:doi/10.3390/app11199204&rft_dat=%3Cproquest_doaj_%3E2580964963%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c322t-4d3b289190a830cba831d2f287563ad83e95fe8ca6b8e0386d93ad0a6da5cc3d3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2580964963&rft_id=info:pmid/&rfr_iscdi=true |