Loading…

ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer

Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing pair...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2024-06
Main Authors: Chen, Hongruixuan, Lan, Cuiling, Song, Jian, Broni-Bediako, Clifford, Xia, Junshi, Yokoya, Naoto
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Chen, Hongruixuan
Lan, Cuiling
Song, Jian
Broni-Bediako, Clifford
Xia, Junshi
Yokoya, Naoto
description Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer
doi_str_mv 10.48550/arxiv.2310.02674
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2873070278</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2873070278</sourcerecordid><originalsourceid>FETCH-LOGICAL-a528-8fe8f957c4cb844eefcb62473f4929b79aad99146be4c58360b1e29ab9a7a8b33</originalsourceid><addsrcrecordid>eNotjVtLwzAYhoMgOOZ-gHcBrzPTJG0S76S6A1Qquvvxpf3aZWztTNuhF_536-Hqhff0EHIT8bkycczvIHz481zI0eAi0eqCTISUETNKiCsy67o95z-BiGM5IV-52y_acMRwTzOE0Pimphk0JUvbMwaa7qCpsaOL0B7pC_iAJc3fnukj9EDHGs1PvS_gQFe-3rFX7NrD0Pu2oesj1Bg-6dkDHRlY9Gw5-HKcbwI0XfXLvCaXFRw6nP3rlGwWT5t0xbJ8uU4fMgaxMMxUaCob60IVziiFWBUuEUrLSllhnbYApbWRShyqIjYy4S5CYcFZ0GCclFNy-3d7Cu37gF2_3bdDaEbiVhgtueZCG_kNx3BgFw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2873070278</pqid></control><display><type>article</type><title>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</title><source>Publicly Available Content Database</source><creator>Chen, Hongruixuan ; Lan, Cuiling ; Song, Jian ; Broni-Bediako, Clifford ; Xia, Junshi ; Yokoya, Naoto</creator><creatorcontrib>Chen, Hongruixuan ; Lan, Cuiling ; Song, Jian ; Broni-Bediako, Clifford ; Xia, Junshi ; Yokoya, Naoto</creatorcontrib><description>Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2310.02674</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Change detection ; Coders ; Data sources ; Decoders ; Digital mapping ; High resolution ; Image analysis ; Image resolution ; Land cover ; Modules ; Semantics</subject><ispartof>arXiv.org, 2024-06</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by-nc-sa/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2873070278?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25751,27923,37010,44588</link.rule.ids></links><search><creatorcontrib>Chen, Hongruixuan</creatorcontrib><creatorcontrib>Lan, Cuiling</creatorcontrib><creatorcontrib>Song, Jian</creatorcontrib><creatorcontrib>Broni-Bediako, Clifford</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Yokoya, Naoto</creatorcontrib><title>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</title><title>arXiv.org</title><description>Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer</description><subject>Change detection</subject><subject>Coders</subject><subject>Data sources</subject><subject>Decoders</subject><subject>Digital mapping</subject><subject>High resolution</subject><subject>Image analysis</subject><subject>Image resolution</subject><subject>Land cover</subject><subject>Modules</subject><subject>Semantics</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNotjVtLwzAYhoMgOOZ-gHcBrzPTJG0S76S6A1Qquvvxpf3aZWztTNuhF_536-Hqhff0EHIT8bkycczvIHz481zI0eAi0eqCTISUETNKiCsy67o95z-BiGM5IV-52y_acMRwTzOE0Pimphk0JUvbMwaa7qCpsaOL0B7pC_iAJc3fnukj9EDHGs1PvS_gQFe-3rFX7NrD0Pu2oesj1Bg-6dkDHRlY9Gw5-HKcbwI0XfXLvCaXFRw6nP3rlGwWT5t0xbJ8uU4fMgaxMMxUaCob60IVziiFWBUuEUrLSllhnbYApbWRShyqIjYy4S5CYcFZ0GCclFNy-3d7Cu37gF2_3bdDaEbiVhgtueZCG_kNx3BgFw</recordid><startdate>20240626</startdate><enddate>20240626</enddate><creator>Chen, Hongruixuan</creator><creator>Lan, Cuiling</creator><creator>Song, Jian</creator><creator>Broni-Bediako, Clifford</creator><creator>Xia, Junshi</creator><creator>Yokoya, Naoto</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240626</creationdate><title>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</title><author>Chen, Hongruixuan ; Lan, Cuiling ; Song, Jian ; Broni-Bediako, Clifford ; Xia, Junshi ; Yokoya, Naoto</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a528-8fe8f957c4cb844eefcb62473f4929b79aad99146be4c58360b1e29ab9a7a8b33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Change detection</topic><topic>Coders</topic><topic>Data sources</topic><topic>Decoders</topic><topic>Digital mapping</topic><topic>High resolution</topic><topic>Image analysis</topic><topic>Image resolution</topic><topic>Land cover</topic><topic>Modules</topic><topic>Semantics</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Hongruixuan</creatorcontrib><creatorcontrib>Lan, Cuiling</creatorcontrib><creatorcontrib>Song, Jian</creatorcontrib><creatorcontrib>Broni-Bediako, Clifford</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Yokoya, Naoto</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chen, Hongruixuan</au><au>Lan, Cuiling</au><au>Song, Jian</au><au>Broni-Bediako, Clifford</au><au>Xia, Junshi</au><au>Yokoya, Naoto</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer</atitle><jtitle>arXiv.org</jtitle><date>2024-06-26</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Optical high-resolution imagery and OSM data are two important data sources of change detection (CD). Previous related studies focus on utilizing the information in OSM data to aid the CD on optical high-resolution images. This paper pioneers the direct detection of land-cover changes utilizing paired OSM data and optical imagery, thereby expanding the scope of CD tasks. To this end, we propose an object-guided Transformer (ObjFormer) by naturally combining the object-based image analysis (OBIA) technique with the advanced vision Transformer architecture. This combination can significantly reduce the computational overhead in the self-attention module without adding extra parameters or layers. ObjFormer has a hierarchical pseudo-siamese encoder consisting of object-guided self-attention modules that extracts multi-level heterogeneous features from OSM data and optical images; a decoder consisting of object-guided cross-attention modules can recover land-cover changes from the extracted heterogeneous features. Beyond basic binary change detection, this paper raises a new semi-supervised semantic change detection task that does not require any manually annotated land-cover labels to train semantic change detectors. Two lightweight semantic decoders are added to ObjFormer to accomplish this task efficiently. A converse cross-entropy loss is designed to fully utilize negative samples, contributing to the great performance improvement in this task. A large-scale benchmark dataset called OpenMapCD containing 1,287 samples covering 40 regions on six continents is constructed to conduct detailed experiments. The results show the effectiveness of our methods in this new kind of CD task. Additionally, case studies in Japanese cities demonstrate the framework's generalizability and practical potential. The OpenMapCD and source code are available in https://github.com/ChenHongruixuan/ObjFormer</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2310.02674</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-06
issn 2331-8422
language eng
recordid cdi_proquest_journals_2873070278
source Publicly Available Content Database
subjects Change detection
Coders
Data sources
Decoders
Digital mapping
High resolution
Image analysis
Image resolution
Land cover
Modules
Semantics
title ObjFormer: Learning Land-Cover Changes From Paired OSM Data and Optical High-Resolution Imagery via Object-Guided Transformer
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T10%3A11%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=ObjFormer:%20Learning%20Land-Cover%20Changes%20From%20Paired%20OSM%20Data%20and%20Optical%20High-Resolution%20Imagery%20via%20Object-Guided%20Transformer&rft.jtitle=arXiv.org&rft.au=Chen,%20Hongruixuan&rft.date=2024-06-26&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2310.02674&rft_dat=%3Cproquest%3E2873070278%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a528-8fe8f957c4cb844eefcb62473f4929b79aad99146be4c58360b1e29ab9a7a8b33%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2873070278&rft_id=info:pmid/&rfr_iscdi=true