Loading…
ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer
Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing pair...
Saved in:
Published in: | arXiv.org 2024-06 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Chen, Hongruixuan Lan, Cuiling Song, Jian Broni-Bediako, Clifford Xia, Junshi Yokoya, Naoto |
description | Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer |
doi_str_mv | 10.48550/arxiv.2310.02674 |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2873070278</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2873070278</sourcerecordid><originalsourceid>FETCH-LOGICAL-a528-8fe8f957c4cb844eefcb62473f4929b79aad99146be4c58360b1e29ab9a7a8b33</originalsourceid><addsrcrecordid>eNotjVtLwzAYhoMgOOZ-gHcBrzPTJG0S76S6A1Qquvvxpf3aZWztTNuhF_536-Hqhff0EHIT8bkycczvIHz481zI0eAi0eqCTISUETNKiCsy67o95z-BiGM5IV-52y_acMRwTzOE0Pimphk0JUvbMwaa7qCpsaOL0B7pC_iAJc3fnukj9EDHGs1PvS_gQFe-3rFX7NrD0Pu2oesj1Bg-6dkDHRlY9Gw5-HKcbwI0XfXLvCaXFRw6nP3rlGwWT5t0xbJ8uU4fMgaxMMxUaCob60IVziiFWBUuEUrLSllhnbYApbWRShyqIjYy4S5CYcFZ0GCclFNy-3d7Cu37gF2_3bdDaEbiVhgtueZCG_kNx3BgFw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2873070278</pqid></control><display><type>article</type><title>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</title><source>Publicly Available Content Database</source><creator>Chen, Hongruixuan ; Lan, Cuiling ; Song, Jian ; Broni-Bediako, Clifford ; Xia, Junshi ; Yokoya, Naoto</creator><creatorcontrib>Chen, Hongruixuan ; Lan, Cuiling ; Song, Jian ; Broni-Bediako, Clifford ; Xia, Junshi ; Yokoya, Naoto</creatorcontrib><description>Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2310.02674</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Change detection ; Coders ; Data sources ; Decoders ; Digital mapping ; High resolution ; Image analysis ; Image resolution ; Land cover ; Modules ; Semantics</subject><ispartof>arXiv.org, 2024-06</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by-nc-sa/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2873070278?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25751,27923,37010,44588</link.rule.ids></links><search><creatorcontrib>Chen, Hongruixuan</creatorcontrib><creatorcontrib>Lan, Cuiling</creatorcontrib><creatorcontrib>Song, Jian</creatorcontrib><creatorcontrib>Broni-Bediako, Clifford</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Yokoya, Naoto</creatorcontrib><title>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</title><title>arXiv.org</title><description>Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer</description><subject>Change detection</subject><subject>Coders</subject><subject>Data sources</subject><subject>Decoders</subject><subject>Digital mapping</subject><subject>High resolution</subject><subject>Image analysis</subject><subject>Image resolution</subject><subject>Land cover</subject><subject>Modules</subject><subject>Semantics</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNotjVtLwzAYhoMgOOZ-gHcBrzPTJG0S76S6A1Qquvvxpf3aZWztTNuhF_536-Hqhff0EHIT8bkycczvIHz481zI0eAi0eqCTISUETNKiCsy67o95z-BiGM5IV-52y_acMRwTzOE0Pimphk0JUvbMwaa7qCpsaOL0B7pC_iAJc3fnukj9EDHGs1PvS_gQFe-3rFX7NrD0Pu2oesj1Bg-6dkDHRlY9Gw5-HKcbwI0XfXLvCaXFRw6nP3rlGwWT5t0xbJ8uU4fMgaxMMxUaCob60IVziiFWBUuEUrLSllhnbYApbWRShyqIjYy4S5CYcFZ0GCclFNy-3d7Cu37gF2_3bdDaEbiVhgtueZCG_kNx3BgFw</recordid><startdate>20240626</startdate><enddate>20240626</enddate><creator>Chen, Hongruixuan</creator><creator>Lan, Cuiling</creator><creator>Song, Jian</creator><creator>Broni-Bediako, Clifford</creator><creator>Xia, Junshi</creator><creator>Yokoya, Naoto</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240626</creationdate><title>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</title><author>Chen, Hongruixuan ; Lan, Cuiling ; Song, Jian ; Broni-Bediako, Clifford ; Xia, Junshi ; Yokoya, Naoto</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a528-8fe8f957c4cb844eefcb62473f4929b79aad99146be4c58360b1e29ab9a7a8b33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Change detection</topic><topic>Coders</topic><topic>Data sources</topic><topic>Decoders</topic><topic>Digital mapping</topic><topic>High resolution</topic><topic>Image analysis</topic><topic>Image resolution</topic><topic>Land cover</topic><topic>Modules</topic><topic>Semantics</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Hongruixuan</creatorcontrib><creatorcontrib>Lan, Cuiling</creatorcontrib><creatorcontrib>Song, Jian</creatorcontrib><creatorcontrib>Broni-Bediako, Clifford</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Yokoya, Naoto</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chen, Hongruixuan</au><au>Lan, Cuiling</au><au>Song, Jian</au><au>Broni-Bediako, Clifford</au><au>Xia, Junshi</au><au>Yokoya, Naoto</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</atitle><jtitle>arXiv.org</jtitle><date>2024-06-26</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2310.02674</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-06 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2873070278 |
source | Publicly Available Content Database |
subjects | Change detection Coders Data sources Decoders Digital mapping High resolution Image analysis Image resolution Land cover Modules Semantics |
title | ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T10%3A11%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=ObjFormer:%20Learning%20Land-Cover%20Changes%20From%20Paired%20OSM%20Data%20and%20Optical%20High-Resolution%20Imagery%20via%20Object-Guided%20Transformer&rft.jtitle=arXiv.org&rft.au=Chen,%20Hongruixuan&rft.date=2024-06-26&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2310.02674&rft_dat=%3Cproquest%3E2873070278%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a528-8fe8f957c4cb844eefcb62473f4929b79aad99146be4c58360b1e29ab9a7a8b33%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2873070278&rft_id=info:pmid/&rfr_iscdi=true |