Loading…

A Biologically Interpretable Two-Stage Deep Neural Network (BIT-DNN) for Vegetation Recognition From Hyperspectral Imagery

Spectral-spatial-based deep learning models have recently proven to be effective in hyper-spectral image (HSI) classification for various earth monitoring applications such as land cover classification and agricultural monitoring. However, due to the nature of "black-box" model representat...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on geoscience and remote sensing 2022, Vol.60, p.1-20
Main Authors: Shi, Yue, Han, Liangxiu, Huang, Wenjiang, Chang, Sheng, Dong, Yingying, Dancey, Darren, Han, Lianghao
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c336t-4cc4a8cb319feb7c00aadec7b3e31dc913b372e2d3119f6d21fa56a4bb09af103
cites cdi_FETCH-LOGICAL-c336t-4cc4a8cb319feb7c00aadec7b3e31dc913b372e2d3119f6d21fa56a4bb09af103
container_end_page 20
container_issue
container_start_page 1
container_title IEEE transactions on geoscience and remote sensing
container_volume 60
creator Shi, Yue
Han, Liangxiu
Huang, Wenjiang
Chang, Sheng
Dong, Yingying
Dancey, Darren
Han, Lianghao
description Spectral-spatial-based deep learning models have recently proven to be effective in hyper-spectral image (HSI) classification for various earth monitoring applications such as land cover classification and agricultural monitoring. However, due to the nature of "black-box" model representation, how to explain and interpret the learning process and the model decision, especially for vegetation classification, remains an open challenge. This study proposes a novel interpretable deep learning model-a biologically interpretable two-stage deep neural network (BIT-DNN), by incorporating the prior-knowledge (i.e., biophysical and biochemical attributes and their hierarchical structures of target entities)-based spectral-spatial feature transformation into the proposed framework, capable of achieving both high accuracy and interpretability on HSI-based classification tasks. The proposed model introduces a two-stage feature learning process: in the first stage, an enhanced interpretable feature block extracts the low-level spectral features associated with the biophysical and biochemical attributes of target entities; and in the second stage, an interpretable capsule block extracts and encapsulates the high-level joint spectral-spatial features representing the hierarchical structure of biophysical and biochemical attributes of these target entities, which provides the model an improved performance on classification and intrinsic interpretability with reduced computational complexity. We have tested and evaluated the model using four real HSI data sets for four separate tasks (i.e., plant species classification, land cover classification, urban scene recognition, and crop disease recognition tasks). The proposed model has been compared with five state-of-the-art deep learning models. The results demonstrate that the proposed model has competitive advantages in terms of both classification accuracy and model interpretability, especially for vegetation classification.
doi_str_mv 10.1109/TGRS.2021.3058782
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2607877821</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9362293</ieee_id><sourcerecordid>2607877821</sourcerecordid><originalsourceid>FETCH-LOGICAL-c336t-4cc4a8cb319feb7c00aadec7b3e31dc913b372e2d3119f6d21fa56a4bb09af103</originalsourceid><addsrcrecordid>eNo9kEtLw0AUhQdRsFZ_gLgZcKOL1LkzeS5rax9QKrTRbZhMbkJqmomTlBJ_vYktrs5dnO9c-Ai5BzYCYMFLON9sR5xxGAnm-J7PL8gAHMe3mGvbl2TAIHAt7gf8mtzU9Y4xsB3wBuRnTF9zXegsV7IoWrosGzSVwUbGBdLwqK1tIzOkU8SKrvFgZNFFc9Tmiz69LkNrul4_01Qb-olZRzW5LukGlc7K_O-eGb2ni7ZCU1eomp5f7rtF096Sq1QWNd6dc0g-Zm_hZGGt3ufLyXhlKSHcxrKVsqWvYgFBirGnGJMyQeXFAgUkKgARC48jTwR0DTfhkErHlXYcs0CmwMSQPJ52K6O_D1g30U4fTNm9jLjLPN_rbEHXglNLGV3XBtOoMvlemjYCFvWKo15x1CuOzoo75uHE5Ij43w-Ey3kgxC9ZaHj1</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2607877821</pqid></control><display><type>article</type><title>A Biologically Interpretable Two-Stage Deep Neural Network (BIT-DNN) for Vegetation Recognition From Hyperspectral Imagery</title><source>IEEE Xplore (Online service)</source><creator>Shi, Yue ; Han, Liangxiu ; Huang, Wenjiang ; Chang, Sheng ; Dong, Yingying ; Dancey, Darren ; Han, Lianghao</creator><creatorcontrib>Shi, Yue ; Han, Liangxiu ; Huang, Wenjiang ; Chang, Sheng ; Dong, Yingying ; Dancey, Darren ; Han, Lianghao</creatorcontrib><description>Spectral-spatial-based deep learning models have recently proven to be effective in hyper-spectral image (HSI) classification for various earth monitoring applications such as land cover classification and agricultural monitoring. However, due to the nature of "black-box" model representation, how to explain and interpret the learning process and the model decision, especially for vegetation classification, remains an open challenge. This study proposes a novel interpretable deep learning model-a biologically interpretable two-stage deep neural network (BIT-DNN), by incorporating the prior-knowledge (i.e., biophysical and biochemical attributes and their hierarchical structures of target entities)-based spectral-spatial feature transformation into the proposed framework, capable of achieving both high accuracy and interpretability on HSI-based classification tasks. The proposed model introduces a two-stage feature learning process: in the first stage, an enhanced interpretable feature block extracts the low-level spectral features associated with the biophysical and biochemical attributes of target entities; and in the second stage, an interpretable capsule block extracts and encapsulates the high-level joint spectral-spatial features representing the hierarchical structure of biophysical and biochemical attributes of these target entities, which provides the model an improved performance on classification and intrinsic interpretability with reduced computational complexity. We have tested and evaluated the model using four real HSI data sets for four separate tasks (i.e., plant species classification, land cover classification, urban scene recognition, and crop disease recognition tasks). The proposed model has been compared with five state-of-the-art deep learning models. The results demonstrate that the proposed model has competitive advantages in terms of both classification accuracy and model interpretability, especially for vegetation classification.</description><identifier>ISSN: 0196-2892</identifier><identifier>EISSN: 1558-0644</identifier><identifier>DOI: 10.1109/TGRS.2021.3058782</identifier><identifier>CODEN: IGRSD2</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Accuracy ; Artificial neural networks ; Biochemistry ; Biological system modeling ; Biology ; Classification ; Computer applications ; Crop diseases ; Data models ; Deep learning ; Feature extraction ; hyper-spectral images (HSIs) ; Hyperspectral imaging ; Image classification ; Imagery ; interpretability ; Land cover ; Machine learning ; Model accuracy ; Monitoring ; Neural networks ; Plant diseases ; Plant species ; Recognition ; Spatial discrimination learning ; Species classification ; Spectra ; Structural hierarchy ; Vegetation ; Vegetation mapping</subject><ispartof>IEEE transactions on geoscience and remote sensing, 2022, Vol.60, p.1-20</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c336t-4cc4a8cb319feb7c00aadec7b3e31dc913b372e2d3119f6d21fa56a4bb09af103</citedby><cites>FETCH-LOGICAL-c336t-4cc4a8cb319feb7c00aadec7b3e31dc913b372e2d3119f6d21fa56a4bb09af103</cites><orcidid>0000-0003-2491-7473 ; 0000-0001-8672-1017 ; 0000-0002-2865-5020 ; 0000-0001-7870-7047 ; 0000-0001-8424-6996 ; 0000-0001-7251-8958</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9362293$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,4009,27902,27903,27904,54775</link.rule.ids></links><search><creatorcontrib>Shi, Yue</creatorcontrib><creatorcontrib>Han, Liangxiu</creatorcontrib><creatorcontrib>Huang, Wenjiang</creatorcontrib><creatorcontrib>Chang, Sheng</creatorcontrib><creatorcontrib>Dong, Yingying</creatorcontrib><creatorcontrib>Dancey, Darren</creatorcontrib><creatorcontrib>Han, Lianghao</creatorcontrib><title>A Biologically Interpretable Two-Stage Deep Neural Network (BIT-DNN) for Vegetation Recognition From Hyperspectral Imagery</title><title>IEEE transactions on geoscience and remote sensing</title><addtitle>TGRS</addtitle><description>Spectral-spatial-based deep learning models have recently proven to be effective in hyper-spectral image (HSI) classification for various earth monitoring applications such as land cover classification and agricultural monitoring. However, due to the nature of "black-box" model representation, how to explain and interpret the learning process and the model decision, especially for vegetation classification, remains an open challenge. This study proposes a novel interpretable deep learning model-a biologically interpretable two-stage deep neural network (BIT-DNN), by incorporating the prior-knowledge (i.e., biophysical and biochemical attributes and their hierarchical structures of target entities)-based spectral-spatial feature transformation into the proposed framework, capable of achieving both high accuracy and interpretability on HSI-based classification tasks. The proposed model introduces a two-stage feature learning process: in the first stage, an enhanced interpretable feature block extracts the low-level spectral features associated with the biophysical and biochemical attributes of target entities; and in the second stage, an interpretable capsule block extracts and encapsulates the high-level joint spectral-spatial features representing the hierarchical structure of biophysical and biochemical attributes of these target entities, which provides the model an improved performance on classification and intrinsic interpretability with reduced computational complexity. We have tested and evaluated the model using four real HSI data sets for four separate tasks (i.e., plant species classification, land cover classification, urban scene recognition, and crop disease recognition tasks). The proposed model has been compared with five state-of-the-art deep learning models. The results demonstrate that the proposed model has competitive advantages in terms of both classification accuracy and model interpretability, especially for vegetation classification.</description><subject>Accuracy</subject><subject>Artificial neural networks</subject><subject>Biochemistry</subject><subject>Biological system modeling</subject><subject>Biology</subject><subject>Classification</subject><subject>Computer applications</subject><subject>Crop diseases</subject><subject>Data models</subject><subject>Deep learning</subject><subject>Feature extraction</subject><subject>hyper-spectral images (HSIs)</subject><subject>Hyperspectral imaging</subject><subject>Image classification</subject><subject>Imagery</subject><subject>interpretability</subject><subject>Land cover</subject><subject>Machine learning</subject><subject>Model accuracy</subject><subject>Monitoring</subject><subject>Neural networks</subject><subject>Plant diseases</subject><subject>Plant species</subject><subject>Recognition</subject><subject>Spatial discrimination learning</subject><subject>Species classification</subject><subject>Spectra</subject><subject>Structural hierarchy</subject><subject>Vegetation</subject><subject>Vegetation mapping</subject><issn>0196-2892</issn><issn>1558-0644</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><recordid>eNo9kEtLw0AUhQdRsFZ_gLgZcKOL1LkzeS5rax9QKrTRbZhMbkJqmomTlBJ_vYktrs5dnO9c-Ai5BzYCYMFLON9sR5xxGAnm-J7PL8gAHMe3mGvbl2TAIHAt7gf8mtzU9Y4xsB3wBuRnTF9zXegsV7IoWrosGzSVwUbGBdLwqK1tIzOkU8SKrvFgZNFFc9Tmiz69LkNrul4_01Qb-olZRzW5LukGlc7K_O-eGb2ni7ZCU1eomp5f7rtF096Sq1QWNd6dc0g-Zm_hZGGt3ufLyXhlKSHcxrKVsqWvYgFBirGnGJMyQeXFAgUkKgARC48jTwR0DTfhkErHlXYcs0CmwMSQPJ52K6O_D1g30U4fTNm9jLjLPN_rbEHXglNLGV3XBtOoMvlemjYCFvWKo15x1CuOzoo75uHE5Ij43w-Ey3kgxC9ZaHj1</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Shi, Yue</creator><creator>Han, Liangxiu</creator><creator>Huang, Wenjiang</creator><creator>Chang, Sheng</creator><creator>Dong, Yingying</creator><creator>Dancey, Darren</creator><creator>Han, Lianghao</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7UA</scope><scope>8FD</scope><scope>C1K</scope><scope>F1W</scope><scope>FR3</scope><scope>H8D</scope><scope>H96</scope><scope>KR7</scope><scope>L.G</scope><scope>L7M</scope><orcidid>https://orcid.org/0000-0003-2491-7473</orcidid><orcidid>https://orcid.org/0000-0001-8672-1017</orcidid><orcidid>https://orcid.org/0000-0002-2865-5020</orcidid><orcidid>https://orcid.org/0000-0001-7870-7047</orcidid><orcidid>https://orcid.org/0000-0001-8424-6996</orcidid><orcidid>https://orcid.org/0000-0001-7251-8958</orcidid></search><sort><creationdate>2022</creationdate><title>A Biologically Interpretable Two-Stage Deep Neural Network (BIT-DNN) for Vegetation Recognition From Hyperspectral Imagery</title><author>Shi, Yue ; Han, Liangxiu ; Huang, Wenjiang ; Chang, Sheng ; Dong, Yingying ; Dancey, Darren ; Han, Lianghao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c336t-4cc4a8cb319feb7c00aadec7b3e31dc913b372e2d3119f6d21fa56a4bb09af103</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Accuracy</topic><topic>Artificial neural networks</topic><topic>Biochemistry</topic><topic>Biological system modeling</topic><topic>Biology</topic><topic>Classification</topic><topic>Computer applications</topic><topic>Crop diseases</topic><topic>Data models</topic><topic>Deep learning</topic><topic>Feature extraction</topic><topic>hyper-spectral images (HSIs)</topic><topic>Hyperspectral imaging</topic><topic>Image classification</topic><topic>Imagery</topic><topic>interpretability</topic><topic>Land cover</topic><topic>Machine learning</topic><topic>Model accuracy</topic><topic>Monitoring</topic><topic>Neural networks</topic><topic>Plant diseases</topic><topic>Plant species</topic><topic>Recognition</topic><topic>Spatial discrimination learning</topic><topic>Species classification</topic><topic>Spectra</topic><topic>Structural hierarchy</topic><topic>Vegetation</topic><topic>Vegetation mapping</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Shi, Yue</creatorcontrib><creatorcontrib>Han, Liangxiu</creatorcontrib><creatorcontrib>Huang, Wenjiang</creatorcontrib><creatorcontrib>Chang, Sheng</creatorcontrib><creatorcontrib>Dong, Yingying</creatorcontrib><creatorcontrib>Dancey, Darren</creatorcontrib><creatorcontrib>Han, Lianghao</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE/IET Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Water Resources Abstracts</collection><collection>Technology Research Database</collection><collection>Environmental Sciences and Pollution Management</collection><collection>ASFA: Aquatic Sciences and Fisheries Abstracts</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Aquatic Science &amp; Fisheries Abstracts (ASFA) 2: Ocean Technology, Policy &amp; Non-Living Resources</collection><collection>Civil Engineering Abstracts</collection><collection>Aquatic Science &amp; Fisheries Abstracts (ASFA) Professional</collection><collection>Advanced Technologies Database with Aerospace</collection><jtitle>IEEE transactions on geoscience and remote sensing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Shi, Yue</au><au>Han, Liangxiu</au><au>Huang, Wenjiang</au><au>Chang, Sheng</au><au>Dong, Yingying</au><au>Dancey, Darren</au><au>Han, Lianghao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Biologically Interpretable Two-Stage Deep Neural Network (BIT-DNN) for Vegetation Recognition From Hyperspectral Imagery</atitle><jtitle>IEEE transactions on geoscience and remote sensing</jtitle><stitle>TGRS</stitle><date>2022</date><risdate>2022</risdate><volume>60</volume><spage>1</spage><epage>20</epage><pages>1-20</pages><issn>0196-2892</issn><eissn>1558-0644</eissn><coden>IGRSD2</coden><abstract>Spectral-spatial-based deep learning models have recently proven to be effective in hyper-spectral image (HSI) classification for various earth monitoring applications such as land cover classification and agricultural monitoring. However, due to the nature of "black-box" model representation, how to explain and interpret the learning process and the model decision, especially for vegetation classification, remains an open challenge. This study proposes a novel interpretable deep learning model-a biologically interpretable two-stage deep neural network (BIT-DNN), by incorporating the prior-knowledge (i.e., biophysical and biochemical attributes and their hierarchical structures of target entities)-based spectral-spatial feature transformation into the proposed framework, capable of achieving both high accuracy and interpretability on HSI-based classification tasks. The proposed model introduces a two-stage feature learning process: in the first stage, an enhanced interpretable feature block extracts the low-level spectral features associated with the biophysical and biochemical attributes of target entities; and in the second stage, an interpretable capsule block extracts and encapsulates the high-level joint spectral-spatial features representing the hierarchical structure of biophysical and biochemical attributes of these target entities, which provides the model an improved performance on classification and intrinsic interpretability with reduced computational complexity. We have tested and evaluated the model using four real HSI data sets for four separate tasks (i.e., plant species classification, land cover classification, urban scene recognition, and crop disease recognition tasks). The proposed model has been compared with five state-of-the-art deep learning models. The results demonstrate that the proposed model has competitive advantages in terms of both classification accuracy and model interpretability, especially for vegetation classification.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TGRS.2021.3058782</doi><tpages>20</tpages><orcidid>https://orcid.org/0000-0003-2491-7473</orcidid><orcidid>https://orcid.org/0000-0001-8672-1017</orcidid><orcidid>https://orcid.org/0000-0002-2865-5020</orcidid><orcidid>https://orcid.org/0000-0001-7870-7047</orcidid><orcidid>https://orcid.org/0000-0001-8424-6996</orcidid><orcidid>https://orcid.org/0000-0001-7251-8958</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0196-2892
ispartof IEEE transactions on geoscience and remote sensing, 2022, Vol.60, p.1-20
issn 0196-2892
1558-0644
language eng
recordid cdi_proquest_journals_2607877821
source IEEE Xplore (Online service)
subjects Accuracy
Artificial neural networks
Biochemistry
Biological system modeling
Biology
Classification
Computer applications
Crop diseases
Data models
Deep learning
Feature extraction
hyper-spectral images (HSIs)
Hyperspectral imaging
Image classification
Imagery
interpretability
Land cover
Machine learning
Model accuracy
Monitoring
Neural networks
Plant diseases
Plant species
Recognition
Spatial discrimination learning
Species classification
Spectra
Structural hierarchy
Vegetation
Vegetation mapping
title A Biologically Interpretable Two-Stage Deep Neural Network (BIT-DNN) for Vegetation Recognition From Hyperspectral Imagery
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T05%3A57%3A47IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Biologically%20Interpretable%20Two-Stage%20Deep%20Neural%20Network%20(BIT-DNN)%20for%20Vegetation%20Recognition%20From%20Hyperspectral%20Imagery&rft.jtitle=IEEE%20transactions%20on%20geoscience%20and%20remote%20sensing&rft.au=Shi,%20Yue&rft.date=2022&rft.volume=60&rft.spage=1&rft.epage=20&rft.pages=1-20&rft.issn=0196-2892&rft.eissn=1558-0644&rft.coden=IGRSD2&rft_id=info:doi/10.1109/TGRS.2021.3058782&rft_dat=%3Cproquest_cross%3E2607877821%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c336t-4cc4a8cb319feb7c00aadec7b3e31dc913b372e2d3119f6d21fa56a4bb09af103%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2607877821&rft_id=info:pmid/&rft_ieee_id=9362293&rfr_iscdi=true