Loading…

A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage

Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences....

Full description

Saved in:
Bibliographic Details
Published in:Computer-aided civil and infrastructure engineering 2022-09, Vol.37 (11), p.1387-1401
Main Authors: Chun, Pang‐Jo, Yamane, Tatsuro, Maemura, Yu
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3
cites cdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3
container_end_page 1401
container_issue 11
container_start_page 1387
container_title Computer-aided civil and infrastructure engineering
container_volume 37
creator Chun, Pang‐Jo
Yamane, Tatsuro
Maemura, Yu
description Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI.
doi_str_mv 10.1111/mice.12793
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2726034907</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2726034907</sourcerecordid><originalsourceid>FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</originalsourceid><addsrcrecordid>eNp9kM9Og0AQxonRxFq9-ASbeDOh7h_YhWPTVG2i8aJnMixDSwMs7oLam4_gM_okLuLZucwk8_u-yXxBcMnogvm6aSqNC8ZVKo6CGYukChMp1bGfaSrCVCbqNDhzbk99RZGYBcOSFIgdqRFsW7Xb78-vHBwWpGpgi0RD11dmXJAG-50pSG8IDL1poK801PWBbLFFC71nTdNZ3GHrqjck-NHV0MKodsSUJLdV4Q0LGH3Pg5MSaocXf30evNyun1f34cPT3Wa1fAh1RIUIhU7ipMwhp0AxxjxSsRRM0gjKRMYJA15GMQCCSjjNFeU85hplmrJS5qhLMQ-uJt_OmtcBXZ_tzWBbfzLjiksqopQqT11PlLbGOYtl1ln_vj1kjGZjrNkYa_Ybq4fZBL9XNR7-IbPHzWo9aX4AI6B9Wg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2726034907</pqid></control><display><type>article</type><title>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</title><source>Wiley</source><creator>Chun, Pang‐Jo ; Yamane, Tatsuro ; Maemura, Yu</creator><creatorcontrib>Chun, Pang‐Jo ; Yamane, Tatsuro ; Maemura, Yu</creatorcontrib><description>Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI.</description><identifier>ISSN: 1093-9687</identifier><identifier>EISSN: 1467-8667</identifier><identifier>DOI: 10.1111/mice.12793</identifier><language>eng</language><publisher>Hoboken: Wiley Subscription Services, Inc</publisher><subject>Algorithms ; Artificial intelligence ; Bridge inspection ; Damage ; Deep learning ; Engineers ; Machine learning ; Technical information</subject><ispartof>Computer-aided civil and infrastructure engineering, 2022-09, Vol.37 (11), p.1387-1401</ispartof><rights>2021 The Authors. published by Wiley Periodicals LLC on behalf of Editor.</rights><rights>2021. This article is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</citedby><cites>FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Chun, Pang‐Jo</creatorcontrib><creatorcontrib>Yamane, Tatsuro</creatorcontrib><creatorcontrib>Maemura, Yu</creatorcontrib><title>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</title><title>Computer-aided civil and infrastructure engineering</title><description>Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI.</description><subject>Algorithms</subject><subject>Artificial intelligence</subject><subject>Bridge inspection</subject><subject>Damage</subject><subject>Deep learning</subject><subject>Engineers</subject><subject>Machine learning</subject><subject>Technical information</subject><issn>1093-9687</issn><issn>1467-8667</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>24P</sourceid><recordid>eNp9kM9Og0AQxonRxFq9-ASbeDOh7h_YhWPTVG2i8aJnMixDSwMs7oLam4_gM_okLuLZucwk8_u-yXxBcMnogvm6aSqNC8ZVKo6CGYukChMp1bGfaSrCVCbqNDhzbk99RZGYBcOSFIgdqRFsW7Xb78-vHBwWpGpgi0RD11dmXJAG-50pSG8IDL1poK801PWBbLFFC71nTdNZ3GHrqjck-NHV0MKodsSUJLdV4Q0LGH3Pg5MSaocXf30evNyun1f34cPT3Wa1fAh1RIUIhU7ipMwhp0AxxjxSsRRM0gjKRMYJA15GMQCCSjjNFeU85hplmrJS5qhLMQ-uJt_OmtcBXZ_tzWBbfzLjiksqopQqT11PlLbGOYtl1ln_vj1kjGZjrNkYa_Ybq4fZBL9XNR7-IbPHzWo9aX4AI6B9Wg</recordid><startdate>202209</startdate><enddate>202209</enddate><creator>Chun, Pang‐Jo</creator><creator>Yamane, Tatsuro</creator><creator>Maemura, Yu</creator><general>Wiley Subscription Services, Inc</general><scope>24P</scope><scope>WIN</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>202209</creationdate><title>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</title><author>Chun, Pang‐Jo ; Yamane, Tatsuro ; Maemura, Yu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Algorithms</topic><topic>Artificial intelligence</topic><topic>Bridge inspection</topic><topic>Damage</topic><topic>Deep learning</topic><topic>Engineers</topic><topic>Machine learning</topic><topic>Technical information</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Chun, Pang‐Jo</creatorcontrib><creatorcontrib>Yamane, Tatsuro</creatorcontrib><creatorcontrib>Maemura, Yu</creatorcontrib><collection>Wiley Online Library Open Access</collection><collection>Wiley-Blackwell Backfiles (Open access)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Computer-aided civil and infrastructure engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chun, Pang‐Jo</au><au>Yamane, Tatsuro</au><au>Maemura, Yu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</atitle><jtitle>Computer-aided civil and infrastructure engineering</jtitle><date>2022-09</date><risdate>2022</risdate><volume>37</volume><issue>11</issue><spage>1387</spage><epage>1401</epage><pages>1387-1401</pages><issn>1093-9687</issn><eissn>1467-8667</eissn><abstract>Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI.</abstract><cop>Hoboken</cop><pub>Wiley Subscription Services, Inc</pub><doi>10.1111/mice.12793</doi><tpages>15</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1093-9687
ispartof Computer-aided civil and infrastructure engineering, 2022-09, Vol.37 (11), p.1387-1401
issn 1093-9687
1467-8667
language eng
recordid cdi_proquest_journals_2726034907
source Wiley
subjects Algorithms
Artificial intelligence
Bridge inspection
Damage
Deep learning
Engineers
Machine learning
Technical information
title A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T15%3A03%3A27IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20deep%20learning%E2%80%90based%20image%20captioning%20method%20to%20automatically%20generate%20comprehensive%20explanations%20of%20bridge%20damage&rft.jtitle=Computer-aided%20civil%20and%20infrastructure%20engineering&rft.au=Chun,%20Pang%E2%80%90Jo&rft.date=2022-09&rft.volume=37&rft.issue=11&rft.spage=1387&rft.epage=1401&rft.pages=1387-1401&rft.issn=1093-9687&rft.eissn=1467-8667&rft_id=info:doi/10.1111/mice.12793&rft_dat=%3Cproquest_cross%3E2726034907%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2726034907&rft_id=info:pmid/&rfr_iscdi=true