Loading…

LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation

The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on t...

Full description

Saved in:
Bibliographic Details
Published in:Computational and structural biotechnology journal 2024-12, Vol.24, p.213-224
Main Authors: Tong, Le, Li, Tianjiu, Zhang, Qian, Zhang, Qin, Zhu, Renchaoli, Du, Wei, Hu, Pengwei
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603
cites cdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603
container_end_page 224
container_issue
container_start_page 213
container_title Computational and structural biotechnology journal
container_volume 24
creator Tong, Le
Li, Tianjiu
Zhang, Qian
Zhang, Qin
Zhu, Renchaoli
Du, Wei
Hu, Pengwei
description The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements. •A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach.
doi_str_mv 10.1016/j.csbj.2024.03.003
format article
fullrecord <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_24ff6196ab994c96a17a781d636d7ab0</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S2001037024000564</els_id><doaj_id>oai_doaj_org_article_24ff6196ab994c96a17a781d636d7ab0</doaj_id><sourcerecordid>3033007860</sourcerecordid><originalsourceid>FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</originalsourceid><addsrcrecordid>eNp9kU1v1DAQhi0EolXpH-CAcuRAgj82toOQUFXxUWkFly3iZtnOeOs0iYvt3Yp_j9MtVXvBB8_YnnlmPC9CrwluCCb8_dDYZIaGYrpqMGswZs_QMcWY1JgJ_PyRf4ROUxpwWZLwjuGX6IjJVlDC5TH6tfY__ab-DvlDdVZdLk49-mt4V41-e5VvYdmrTdRzciFOEKsZ8m2I11U5VhGyn_VY7SElGKsE2wnmrLMP8yv0wukxwem9PUGXXz5vzr_V6x9fL87P1rVtKc1157iThhLRCdBUGiOJpUZrLo3kjlLXOSGMcURwYZx1hgFhtG35ykrRcsxO0MWB2wc9qJvoJx3_qKC9ursIcat0zN6OoOjKOU46rk3XrWyxRGghSc8Z74U2C-vTgXWzMxP0tvwl6vEJ9OnL7K_UNuwVwZ0UUopCeHtPiOH3DlJWk08WxlHPEHZJMcwYxkLeNU4PoTaGlCK4hzoEq0ViNahFYrVIrDBTReKS9OZxhw8p_wQtAR8PAVBmvvcQVbIeZgu9j2BzGYr_H_8v7cq38A</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3033007860</pqid></control><display><type>article</type><title>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</title><source>ScienceDirect Journals</source><source>PubMed Central</source><creator>Tong, Le ; Li, Tianjiu ; Zhang, Qian ; Zhang, Qin ; Zhu, Renchaoli ; Du, Wei ; Hu, Pengwei</creator><creatorcontrib>Tong, Le ; Li, Tianjiu ; Zhang, Qian ; Zhang, Qin ; Zhu, Renchaoli ; Du, Wei ; Hu, Pengwei</creatorcontrib><description>The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements. •A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach.</description><identifier>ISSN: 2001-0370</identifier><identifier>EISSN: 2001-0370</identifier><identifier>DOI: 10.1016/j.csbj.2024.03.003</identifier><identifier>PMID: 38572168</identifier><language>eng</language><publisher>Netherlands: Elsevier B.V</publisher><subject>Joint loss ; Lightweight ; Retinal vessel segmentation ; Transformer</subject><ispartof>Computational and structural biotechnology journal, 2024-12, Vol.24, p.213-224</ispartof><rights>2024 The Author(s)</rights><rights>2024 The Author(s).</rights><rights>2024 The Author(s) 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</citedby><cites>FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</cites><orcidid>0009-0002-0864-2211 ; 0000-0002-4861-4787</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC10987887/pdf/$$EPDF$$P50$$Gpubmedcentral$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S2001037024000564$$EHTML$$P50$$Gelsevier$$Hfree_for_read</linktohtml><link.rule.ids>230,314,727,780,784,885,3549,27924,27925,45780,53791,53793</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/38572168$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Tong, Le</creatorcontrib><creatorcontrib>Li, Tianjiu</creatorcontrib><creatorcontrib>Zhang, Qian</creatorcontrib><creatorcontrib>Zhang, Qin</creatorcontrib><creatorcontrib>Zhu, Renchaoli</creatorcontrib><creatorcontrib>Du, Wei</creatorcontrib><creatorcontrib>Hu, Pengwei</creatorcontrib><title>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</title><title>Computational and structural biotechnology journal</title><addtitle>Comput Struct Biotechnol J</addtitle><description>The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements. •A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach.</description><subject>Joint loss</subject><subject>Lightweight</subject><subject>Retinal vessel segmentation</subject><subject>Transformer</subject><issn>2001-0370</issn><issn>2001-0370</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>DOA</sourceid><recordid>eNp9kU1v1DAQhi0EolXpH-CAcuRAgj82toOQUFXxUWkFly3iZtnOeOs0iYvt3Yp_j9MtVXvBB8_YnnlmPC9CrwluCCb8_dDYZIaGYrpqMGswZs_QMcWY1JgJ_PyRf4ROUxpwWZLwjuGX6IjJVlDC5TH6tfY__ab-DvlDdVZdLk49-mt4V41-e5VvYdmrTdRzciFOEKsZ8m2I11U5VhGyn_VY7SElGKsE2wnmrLMP8yv0wukxwem9PUGXXz5vzr_V6x9fL87P1rVtKc1157iThhLRCdBUGiOJpUZrLo3kjlLXOSGMcURwYZx1hgFhtG35ykrRcsxO0MWB2wc9qJvoJx3_qKC9ursIcat0zN6OoOjKOU46rk3XrWyxRGghSc8Z74U2C-vTgXWzMxP0tvwl6vEJ9OnL7K_UNuwVwZ0UUopCeHtPiOH3DlJWk08WxlHPEHZJMcwYxkLeNU4PoTaGlCK4hzoEq0ViNahFYrVIrDBTReKS9OZxhw8p_wQtAR8PAVBmvvcQVbIeZgu9j2BzGYr_H_8v7cq38A</recordid><startdate>20241201</startdate><enddate>20241201</enddate><creator>Tong, Le</creator><creator>Li, Tianjiu</creator><creator>Zhang, Qian</creator><creator>Zhang, Qin</creator><creator>Zhu, Renchaoli</creator><creator>Du, Wei</creator><creator>Hu, Pengwei</creator><general>Elsevier B.V</general><general>Research Network of Computational and Structural Biotechnology</general><general>Elsevier</general><scope>6I.</scope><scope>AAFTH</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>5PM</scope><scope>DOA</scope><orcidid>https://orcid.org/0009-0002-0864-2211</orcidid><orcidid>https://orcid.org/0000-0002-4861-4787</orcidid></search><sort><creationdate>20241201</creationdate><title>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</title><author>Tong, Le ; Li, Tianjiu ; Zhang, Qian ; Zhang, Qin ; Zhu, Renchaoli ; Du, Wei ; Hu, Pengwei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Joint loss</topic><topic>Lightweight</topic><topic>Retinal vessel segmentation</topic><topic>Transformer</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Tong, Le</creatorcontrib><creatorcontrib>Li, Tianjiu</creatorcontrib><creatorcontrib>Zhang, Qian</creatorcontrib><creatorcontrib>Zhang, Qin</creatorcontrib><creatorcontrib>Zhu, Renchaoli</creatorcontrib><creatorcontrib>Du, Wei</creatorcontrib><creatorcontrib>Hu, Pengwei</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Computational and structural biotechnology journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Tong, Le</au><au>Li, Tianjiu</au><au>Zhang, Qian</au><au>Zhang, Qin</au><au>Zhu, Renchaoli</au><au>Du, Wei</au><au>Hu, Pengwei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</atitle><jtitle>Computational and structural biotechnology journal</jtitle><addtitle>Comput Struct Biotechnol J</addtitle><date>2024-12-01</date><risdate>2024</risdate><volume>24</volume><spage>213</spage><epage>224</epage><pages>213-224</pages><issn>2001-0370</issn><eissn>2001-0370</eissn><abstract>The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements. •A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach.</abstract><cop>Netherlands</cop><pub>Elsevier B.V</pub><pmid>38572168</pmid><doi>10.1016/j.csbj.2024.03.003</doi><tpages>12</tpages><orcidid>https://orcid.org/0009-0002-0864-2211</orcidid><orcidid>https://orcid.org/0000-0002-4861-4787</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2001-0370
ispartof Computational and structural biotechnology journal, 2024-12, Vol.24, p.213-224
issn 2001-0370
2001-0370
language eng
recordid cdi_doaj_primary_oai_doaj_org_article_24ff6196ab994c96a17a781d636d7ab0
source ScienceDirect Journals; PubMed Central
subjects Joint loss
Lightweight
Retinal vessel segmentation
Transformer
title LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T12%3A00%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=LiViT-Net:%20A%20U-Net-like,%20lightweight%20Transformer%20network%20for%20retinal%20vessel%20segmentation&rft.jtitle=Computational%20and%20structural%20biotechnology%20journal&rft.au=Tong,%20Le&rft.date=2024-12-01&rft.volume=24&rft.spage=213&rft.epage=224&rft.pages=213-224&rft.issn=2001-0370&rft.eissn=2001-0370&rft_id=info:doi/10.1016/j.csbj.2024.03.003&rft_dat=%3Cproquest_doaj_%3E3033007860%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3033007860&rft_id=info:pmid/38572168&rfr_iscdi=true