Loading…

A visual SLAM-based lightweight multi-modal semantic framework for an intelligent substation robot

Visual simultaneous localisation and mapping (vSLAM) has shown considerable promise in positioning and navigating across a variety of indoor and outdoor settings, significantly enhancing the mobility of robots employed in industrial and everyday services. Nonetheless, the prevalent reliance of vSLAM...

Full description

Saved in:
Bibliographic Details
Published in:Robotica 2024-07, Vol.42 (7), p.2169-2183
Main Authors: Li, Shaohu, Gu, Jason, Li, Zhijun, Li, Shaofeng, Guo, Bixiang, Gao, Shangbing, Zhao, Feng, Yang, Yuwei, Li, Guoxin, Dong, Lanfang
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c269t-5883f1b5d027b897d54c1776fc5a05c5f308a480a877b3eb85f2e4eeb00f6d063
container_end_page 2183
container_issue 7
container_start_page 2169
container_title Robotica
container_volume 42
creator Li, Shaohu
Gu, Jason
Li, Zhijun
Li, Shaofeng
Guo, Bixiang
Gao, Shangbing
Zhao, Feng
Yang, Yuwei
Li, Guoxin
Dong, Lanfang
description Visual simultaneous localisation and mapping (vSLAM) has shown considerable promise in positioning and navigating across a variety of indoor and outdoor settings, significantly enhancing the mobility of robots employed in industrial and everyday services. Nonetheless, the prevalent reliance of vSLAM technology on the assumption of static environments has led to suboptimal performance in practical implementations, particularly in unstructured and dynamically noisy environments such as substations. Despite advancements in mitigating the influence of dynamic objects through the integration of geometric and semantic information, existing approaches have struggled to strike an equilibrium between performance and real-time responsiveness. This study introduces a lightweight, multi-modal semantic framework predicated on vSLAM, designed to enable intelligent robots to adeptly navigate the dynamic environments characteristic of substations. The framework notably enhances vSLAM performance by mitigating the impact of dynamic objects through a synergistic combination of object detection and instance segmentation techniques. Initially, an enhanced lightweight instance segmentation network is deployed to ensure both the real-time responsiveness and accuracy of the algorithm. Subsequently, the algorithm’s performance is further refined by amalgamating the outcomes of detection and segmentation processes. With a commitment to maximising performance, the framework also ensures the algorithm’s real-time capability. Assessments conducted on public datasets and through empirical experiments have demonstrated that the proposed method markedly improves both the accuracy and real-time performance of vSLAM in dynamic environments.
doi_str_mv 10.1017/S0263574724000511
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3114770742</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><cupid>10_1017_S0263574724000511</cupid><sourcerecordid>3114770742</sourcerecordid><originalsourceid>FETCH-LOGICAL-c269t-5883f1b5d027b897d54c1776fc5a05c5f308a480a877b3eb85f2e4eeb00f6d063</originalsourceid><addsrcrecordid>eNp1kEtPwzAQhC0EEqXwA7hZ4hxYx88eq4qXVMShcI7sxC4uSVxsB8S_J1UrcUBcdg8z36x2ELokcE2AyJsVlIJyyWTJAIATcoQmhIlZoYRQx2iyk4udforOUtoAEEqYnCAzx58-DbrFq-X8qTA62Qa3fv2Wv-xu4m5osy-60IyWZDvdZ19jF3Vnv0J8xy5ErHvs-2zbEbN9xmkwKevsQ49jMCGfoxOn22QvDnuKXu9uXxYPxfL5_nExXxZ1KWa54EpRRwxvoJRGzWTDWU2kFK7mGnjNHQWlmQKtpDTUGsVdaZm1BsCJBgSdoqt97jaGj8GmXG3CEPvxZEXJ-KwEycrRRfauOoaUonXVNvpOx--KQLWrsvpT5cjQA6M7E32ztr_R_1M_w-h1_w</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3114770742</pqid></control><display><type>article</type><title>A visual SLAM-based lightweight multi-modal semantic framework for an intelligent substation robot</title><source>Cambridge Journals Online</source><creator>Li, Shaohu ; Gu, Jason ; Li, Zhijun ; Li, Shaofeng ; Guo, Bixiang ; Gao, Shangbing ; Zhao, Feng ; Yang, Yuwei ; Li, Guoxin ; Dong, Lanfang</creator><creatorcontrib>Li, Shaohu ; Gu, Jason ; Li, Zhijun ; Li, Shaofeng ; Guo, Bixiang ; Gao, Shangbing ; Zhao, Feng ; Yang, Yuwei ; Li, Guoxin ; Dong, Lanfang</creatorcontrib><description>Visual simultaneous localisation and mapping (vSLAM) has shown considerable promise in positioning and navigating across a variety of indoor and outdoor settings, significantly enhancing the mobility of robots employed in industrial and everyday services. Nonetheless, the prevalent reliance of vSLAM technology on the assumption of static environments has led to suboptimal performance in practical implementations, particularly in unstructured and dynamically noisy environments such as substations. Despite advancements in mitigating the influence of dynamic objects through the integration of geometric and semantic information, existing approaches have struggled to strike an equilibrium between performance and real-time responsiveness. This study introduces a lightweight, multi-modal semantic framework predicated on vSLAM, designed to enable intelligent robots to adeptly navigate the dynamic environments characteristic of substations. The framework notably enhances vSLAM performance by mitigating the impact of dynamic objects through a synergistic combination of object detection and instance segmentation techniques. Initially, an enhanced lightweight instance segmentation network is deployed to ensure both the real-time responsiveness and accuracy of the algorithm. Subsequently, the algorithm’s performance is further refined by amalgamating the outcomes of detection and segmentation processes. With a commitment to maximising performance, the framework also ensures the algorithm’s real-time capability. Assessments conducted on public datasets and through empirical experiments have demonstrated that the proposed method markedly improves both the accuracy and real-time performance of vSLAM in dynamic environments.</description><identifier>ISSN: 0263-5747</identifier><identifier>EISSN: 1469-8668</identifier><identifier>DOI: 10.1017/S0263574724000511</identifier><language>eng</language><publisher>Cambridge, UK: Cambridge University Press</publisher><subject>Accuracy ; Algorithms ; Cameras ; Deep learning ; Geometry ; Instance segmentation ; Intelligence in Robotics ; Lasers ; Lightweight ; Localization ; Mapping ; Methods ; Neural networks ; Object recognition ; Real time ; Robotics ; Robots ; Semantics ; Simultaneous localization and mapping ; Substations ; Unstructured data ; Weight reduction</subject><ispartof>Robotica, 2024-07, Vol.42 (7), p.2169-2183</ispartof><rights>The Author(s), 2024. Published by Cambridge University Press</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c269t-5883f1b5d027b897d54c1776fc5a05c5f308a480a877b3eb85f2e4eeb00f6d063</cites><orcidid>0009-0008-8495-9468 ; 0000-0001-8384-0002</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.cambridge.org/core/product/identifier/S0263574724000511/type/journal_article$$EHTML$$P50$$Gcambridge$$H</linktohtml><link.rule.ids>314,780,784,27923,27924,72831</link.rule.ids></links><search><creatorcontrib>Li, Shaohu</creatorcontrib><creatorcontrib>Gu, Jason</creatorcontrib><creatorcontrib>Li, Zhijun</creatorcontrib><creatorcontrib>Li, Shaofeng</creatorcontrib><creatorcontrib>Guo, Bixiang</creatorcontrib><creatorcontrib>Gao, Shangbing</creatorcontrib><creatorcontrib>Zhao, Feng</creatorcontrib><creatorcontrib>Yang, Yuwei</creatorcontrib><creatorcontrib>Li, Guoxin</creatorcontrib><creatorcontrib>Dong, Lanfang</creatorcontrib><title>A visual SLAM-based lightweight multi-modal semantic framework for an intelligent substation robot</title><title>Robotica</title><addtitle>Robotica</addtitle><description>Visual simultaneous localisation and mapping (vSLAM) has shown considerable promise in positioning and navigating across a variety of indoor and outdoor settings, significantly enhancing the mobility of robots employed in industrial and everyday services. Nonetheless, the prevalent reliance of vSLAM technology on the assumption of static environments has led to suboptimal performance in practical implementations, particularly in unstructured and dynamically noisy environments such as substations. Despite advancements in mitigating the influence of dynamic objects through the integration of geometric and semantic information, existing approaches have struggled to strike an equilibrium between performance and real-time responsiveness. This study introduces a lightweight, multi-modal semantic framework predicated on vSLAM, designed to enable intelligent robots to adeptly navigate the dynamic environments characteristic of substations. The framework notably enhances vSLAM performance by mitigating the impact of dynamic objects through a synergistic combination of object detection and instance segmentation techniques. Initially, an enhanced lightweight instance segmentation network is deployed to ensure both the real-time responsiveness and accuracy of the algorithm. Subsequently, the algorithm’s performance is further refined by amalgamating the outcomes of detection and segmentation processes. With a commitment to maximising performance, the framework also ensures the algorithm’s real-time capability. Assessments conducted on public datasets and through empirical experiments have demonstrated that the proposed method markedly improves both the accuracy and real-time performance of vSLAM in dynamic environments.</description><subject>Accuracy</subject><subject>Algorithms</subject><subject>Cameras</subject><subject>Deep learning</subject><subject>Geometry</subject><subject>Instance segmentation</subject><subject>Intelligence in Robotics</subject><subject>Lasers</subject><subject>Lightweight</subject><subject>Localization</subject><subject>Mapping</subject><subject>Methods</subject><subject>Neural networks</subject><subject>Object recognition</subject><subject>Real time</subject><subject>Robotics</subject><subject>Robots</subject><subject>Semantics</subject><subject>Simultaneous localization and mapping</subject><subject>Substations</subject><subject>Unstructured data</subject><subject>Weight reduction</subject><issn>0263-5747</issn><issn>1469-8668</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp1kEtPwzAQhC0EEqXwA7hZ4hxYx88eq4qXVMShcI7sxC4uSVxsB8S_J1UrcUBcdg8z36x2ELokcE2AyJsVlIJyyWTJAIATcoQmhIlZoYRQx2iyk4udforOUtoAEEqYnCAzx58-DbrFq-X8qTA62Qa3fv2Wv-xu4m5osy-60IyWZDvdZ19jF3Vnv0J8xy5ErHvs-2zbEbN9xmkwKevsQ49jMCGfoxOn22QvDnuKXu9uXxYPxfL5_nExXxZ1KWa54EpRRwxvoJRGzWTDWU2kFK7mGnjNHQWlmQKtpDTUGsVdaZm1BsCJBgSdoqt97jaGj8GmXG3CEPvxZEXJ-KwEycrRRfauOoaUonXVNvpOx--KQLWrsvpT5cjQA6M7E32ztr_R_1M_w-h1_w</recordid><startdate>20240701</startdate><enddate>20240701</enddate><creator>Li, Shaohu</creator><creator>Gu, Jason</creator><creator>Li, Zhijun</creator><creator>Li, Shaofeng</creator><creator>Guo, Bixiang</creator><creator>Gao, Shangbing</creator><creator>Zhao, Feng</creator><creator>Yang, Yuwei</creator><creator>Li, Guoxin</creator><creator>Dong, Lanfang</creator><general>Cambridge University Press</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7TB</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0009-0008-8495-9468</orcidid><orcidid>https://orcid.org/0000-0001-8384-0002</orcidid></search><sort><creationdate>20240701</creationdate><title>A visual SLAM-based lightweight multi-modal semantic framework for an intelligent substation robot</title><author>Li, Shaohu ; Gu, Jason ; Li, Zhijun ; Li, Shaofeng ; Guo, Bixiang ; Gao, Shangbing ; Zhao, Feng ; Yang, Yuwei ; Li, Guoxin ; Dong, Lanfang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c269t-5883f1b5d027b897d54c1776fc5a05c5f308a480a877b3eb85f2e4eeb00f6d063</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>Algorithms</topic><topic>Cameras</topic><topic>Deep learning</topic><topic>Geometry</topic><topic>Instance segmentation</topic><topic>Intelligence in Robotics</topic><topic>Lasers</topic><topic>Lightweight</topic><topic>Localization</topic><topic>Mapping</topic><topic>Methods</topic><topic>Neural networks</topic><topic>Object recognition</topic><topic>Real time</topic><topic>Robotics</topic><topic>Robots</topic><topic>Semantics</topic><topic>Simultaneous localization and mapping</topic><topic>Substations</topic><topic>Unstructured data</topic><topic>Weight reduction</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Li, Shaohu</creatorcontrib><creatorcontrib>Gu, Jason</creatorcontrib><creatorcontrib>Li, Zhijun</creatorcontrib><creatorcontrib>Li, Shaofeng</creatorcontrib><creatorcontrib>Guo, Bixiang</creatorcontrib><creatorcontrib>Gao, Shangbing</creatorcontrib><creatorcontrib>Zhao, Feng</creatorcontrib><creatorcontrib>Yang, Yuwei</creatorcontrib><creatorcontrib>Li, Guoxin</creatorcontrib><creatorcontrib>Dong, Lanfang</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Robotica</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Shaohu</au><au>Gu, Jason</au><au>Li, Zhijun</au><au>Li, Shaofeng</au><au>Guo, Bixiang</au><au>Gao, Shangbing</au><au>Zhao, Feng</au><au>Yang, Yuwei</au><au>Li, Guoxin</au><au>Dong, Lanfang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A visual SLAM-based lightweight multi-modal semantic framework for an intelligent substation robot</atitle><jtitle>Robotica</jtitle><addtitle>Robotica</addtitle><date>2024-07-01</date><risdate>2024</risdate><volume>42</volume><issue>7</issue><spage>2169</spage><epage>2183</epage><pages>2169-2183</pages><issn>0263-5747</issn><eissn>1469-8668</eissn><abstract>Visual simultaneous localisation and mapping (vSLAM) has shown considerable promise in positioning and navigating across a variety of indoor and outdoor settings, significantly enhancing the mobility of robots employed in industrial and everyday services. Nonetheless, the prevalent reliance of vSLAM technology on the assumption of static environments has led to suboptimal performance in practical implementations, particularly in unstructured and dynamically noisy environments such as substations. Despite advancements in mitigating the influence of dynamic objects through the integration of geometric and semantic information, existing approaches have struggled to strike an equilibrium between performance and real-time responsiveness. This study introduces a lightweight, multi-modal semantic framework predicated on vSLAM, designed to enable intelligent robots to adeptly navigate the dynamic environments characteristic of substations. The framework notably enhances vSLAM performance by mitigating the impact of dynamic objects through a synergistic combination of object detection and instance segmentation techniques. Initially, an enhanced lightweight instance segmentation network is deployed to ensure both the real-time responsiveness and accuracy of the algorithm. Subsequently, the algorithm’s performance is further refined by amalgamating the outcomes of detection and segmentation processes. With a commitment to maximising performance, the framework also ensures the algorithm’s real-time capability. Assessments conducted on public datasets and through empirical experiments have demonstrated that the proposed method markedly improves both the accuracy and real-time performance of vSLAM in dynamic environments.</abstract><cop>Cambridge, UK</cop><pub>Cambridge University Press</pub><doi>10.1017/S0263574724000511</doi><tpages>15</tpages><orcidid>https://orcid.org/0009-0008-8495-9468</orcidid><orcidid>https://orcid.org/0000-0001-8384-0002</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0263-5747
ispartof Robotica, 2024-07, Vol.42 (7), p.2169-2183
issn 0263-5747
1469-8668
language eng
recordid cdi_proquest_journals_3114770742
source Cambridge Journals Online
subjects Accuracy
Algorithms
Cameras
Deep learning
Geometry
Instance segmentation
Intelligence in Robotics
Lasers
Lightweight
Localization
Mapping
Methods
Neural networks
Object recognition
Real time
Robotics
Robots
Semantics
Simultaneous localization and mapping
Substations
Unstructured data
Weight reduction
title A visual SLAM-based lightweight multi-modal semantic framework for an intelligent substation robot
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T11%3A16%3A14IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20visual%20SLAM-based%20lightweight%20multi-modal%20semantic%20framework%20for%20an%20intelligent%20substation%20robot&rft.jtitle=Robotica&rft.au=Li,%20Shaohu&rft.date=2024-07-01&rft.volume=42&rft.issue=7&rft.spage=2169&rft.epage=2183&rft.pages=2169-2183&rft.issn=0263-5747&rft.eissn=1469-8668&rft_id=info:doi/10.1017/S0263574724000511&rft_dat=%3Cproquest_cross%3E3114770742%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c269t-5883f1b5d027b897d54c1776fc5a05c5f308a480a877b3eb85f2e4eeb00f6d063%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3114770742&rft_id=info:pmid/&rft_cupid=10_1017_S0263574724000511&rfr_iscdi=true