Loading…
A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage
Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences....
Saved in:
Published in: | Computer-aided civil and infrastructure engineering 2022-09, Vol.37 (11), p.1387-1401 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3 |
---|---|
cites | cdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3 |
container_end_page | 1401 |
container_issue | 11 |
container_start_page | 1387 |
container_title | Computer-aided civil and infrastructure engineering |
container_volume | 37 |
creator | Chun, Pang‐Jo Yamane, Tatsuro Maemura, Yu |
description | Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI. |
doi_str_mv | 10.1111/mice.12793 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2726034907</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2726034907</sourcerecordid><originalsourceid>FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</originalsourceid><addsrcrecordid>eNp9kM9Og0AQxonRxFq9-ASbeDOh7h_YhWPTVG2i8aJnMixDSwMs7oLam4_gM_okLuLZucwk8_u-yXxBcMnogvm6aSqNC8ZVKo6CGYukChMp1bGfaSrCVCbqNDhzbk99RZGYBcOSFIgdqRFsW7Xb78-vHBwWpGpgi0RD11dmXJAG-50pSG8IDL1poK801PWBbLFFC71nTdNZ3GHrqjck-NHV0MKodsSUJLdV4Q0LGH3Pg5MSaocXf30evNyun1f34cPT3Wa1fAh1RIUIhU7ipMwhp0AxxjxSsRRM0gjKRMYJA15GMQCCSjjNFeU85hplmrJS5qhLMQ-uJt_OmtcBXZ_tzWBbfzLjiksqopQqT11PlLbGOYtl1ln_vj1kjGZjrNkYa_Ybq4fZBL9XNR7-IbPHzWo9aX4AI6B9Wg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2726034907</pqid></control><display><type>article</type><title>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</title><source>Wiley</source><creator>Chun, Pang‐Jo ; Yamane, Tatsuro ; Maemura, Yu</creator><creatorcontrib>Chun, Pang‐Jo ; Yamane, Tatsuro ; Maemura, Yu</creatorcontrib><description>Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI.</description><identifier>ISSN: 1093-9687</identifier><identifier>EISSN: 1467-8667</identifier><identifier>DOI: 10.1111/mice.12793</identifier><language>eng</language><publisher>Hoboken: Wiley Subscription Services, Inc</publisher><subject>Algorithms ; Artificial intelligence ; Bridge inspection ; Damage ; Deep learning ; Engineers ; Machine learning ; Technical information</subject><ispartof>Computer-aided civil and infrastructure engineering, 2022-09, Vol.37 (11), p.1387-1401</ispartof><rights>2021 The Authors. published by Wiley Periodicals LLC on behalf of Editor.</rights><rights>2021. This article is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</citedby><cites>FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Chun, Pang‐Jo</creatorcontrib><creatorcontrib>Yamane, Tatsuro</creatorcontrib><creatorcontrib>Maemura, Yu</creatorcontrib><title>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</title><title>Computer-aided civil and infrastructure engineering</title><description>Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI.</description><subject>Algorithms</subject><subject>Artificial intelligence</subject><subject>Bridge inspection</subject><subject>Damage</subject><subject>Deep learning</subject><subject>Engineers</subject><subject>Machine learning</subject><subject>Technical information</subject><issn>1093-9687</issn><issn>1467-8667</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>24P</sourceid><recordid>eNp9kM9Og0AQxonRxFq9-ASbeDOh7h_YhWPTVG2i8aJnMixDSwMs7oLam4_gM_okLuLZucwk8_u-yXxBcMnogvm6aSqNC8ZVKo6CGYukChMp1bGfaSrCVCbqNDhzbk99RZGYBcOSFIgdqRFsW7Xb78-vHBwWpGpgi0RD11dmXJAG-50pSG8IDL1poK801PWBbLFFC71nTdNZ3GHrqjck-NHV0MKodsSUJLdV4Q0LGH3Pg5MSaocXf30evNyun1f34cPT3Wa1fAh1RIUIhU7ipMwhp0AxxjxSsRRM0gjKRMYJA15GMQCCSjjNFeU85hplmrJS5qhLMQ-uJt_OmtcBXZ_tzWBbfzLjiksqopQqT11PlLbGOYtl1ln_vj1kjGZjrNkYa_Ybq4fZBL9XNR7-IbPHzWo9aX4AI6B9Wg</recordid><startdate>202209</startdate><enddate>202209</enddate><creator>Chun, Pang‐Jo</creator><creator>Yamane, Tatsuro</creator><creator>Maemura, Yu</creator><general>Wiley Subscription Services, Inc</general><scope>24P</scope><scope>WIN</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>202209</creationdate><title>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</title><author>Chun, Pang‐Jo ; Yamane, Tatsuro ; Maemura, Yu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Algorithms</topic><topic>Artificial intelligence</topic><topic>Bridge inspection</topic><topic>Damage</topic><topic>Deep learning</topic><topic>Engineers</topic><topic>Machine learning</topic><topic>Technical information</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Chun, Pang‐Jo</creatorcontrib><creatorcontrib>Yamane, Tatsuro</creatorcontrib><creatorcontrib>Maemura, Yu</creatorcontrib><collection>Wiley Online Library Open Access</collection><collection>Wiley-Blackwell Backfiles (Open access)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Computer-aided civil and infrastructure engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chun, Pang‐Jo</au><au>Yamane, Tatsuro</au><au>Maemura, Yu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage</atitle><jtitle>Computer-aided civil and infrastructure engineering</jtitle><date>2022-09</date><risdate>2022</risdate><volume>37</volume><issue>11</issue><spage>1387</spage><epage>1401</epage><pages>1387-1401</pages><issn>1093-9687</issn><eissn>1467-8667</eissn><abstract>Photographs of bridges can reveal considerable technical information such as the part of the structure that is damaged and the type of damage. Maintenance and inspection engineers can benefit greatly from a technology that can automatically extract and express such information in readable sentences. This is possibly the first study on developing a deep learning model that can generate sentences describing the damage condition of a bridge from images through an image captioning method. Our study shows that by introducing an attention mechanism into the deep learning model, highly accurate descriptive sentences can be generated. In addition, often multiple forms of damage can be observed in the images of bridges; hence, our algorithm is adapted to output multiple sentences to provide a comprehensive interpretation of complex images. In our dataset, the scores of Bilingual Evaluation Understudy (BLEU)‐1 to BLEU‐4 were 0.782, 0.749, 0.711, and 0.693, respectively, and the percentage of correctly output explanatory sentences is 69.3%. All of these results are better than the model without the attention mechanism. The developed method makes it possible to provide user‐friendly, text‐based explanations of bridge damage in images, making it easier for engineers with relatively little experience and even administrative staff without extensive technical expertise to understand images of bridge damage. Future research in this field is expected to lead to the unification of field expertise with artificial intelligence (AI), which will be the foundation of the evolutionary development of bridge inspection AI.</abstract><cop>Hoboken</cop><pub>Wiley Subscription Services, Inc</pub><doi>10.1111/mice.12793</doi><tpages>15</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1093-9687 |
ispartof | Computer-aided civil and infrastructure engineering, 2022-09, Vol.37 (11), p.1387-1401 |
issn | 1093-9687 1467-8667 |
language | eng |
recordid | cdi_proquest_journals_2726034907 |
source | Wiley |
subjects | Algorithms Artificial intelligence Bridge inspection Damage Deep learning Engineers Machine learning Technical information |
title | A deep learning‐based image captioning method to automatically generate comprehensive explanations of bridge damage |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T15%3A03%3A27IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20deep%20learning%E2%80%90based%20image%20captioning%20method%20to%20automatically%20generate%20comprehensive%20explanations%20of%20bridge%20damage&rft.jtitle=Computer-aided%20civil%20and%20infrastructure%20engineering&rft.au=Chun,%20Pang%E2%80%90Jo&rft.date=2022-09&rft.volume=37&rft.issue=11&rft.spage=1387&rft.epage=1401&rft.pages=1387-1401&rft.issn=1093-9687&rft.eissn=1467-8667&rft_id=info:doi/10.1111/mice.12793&rft_dat=%3Cproquest_cross%3E2726034907%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c4033-3c858fbab0a0e5eb475631604af86581a2f45aaea7820b702252ce6991f6becf3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2726034907&rft_id=info:pmid/&rfr_iscdi=true |