Loading…
LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation
The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on t...
Saved in:
Published in: | Computational and structural biotechnology journal 2024-12, Vol.24, p.213-224 |
---|---|
Main Authors: | , , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603 |
---|---|
cites | cdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603 |
container_end_page | 224 |
container_issue | |
container_start_page | 213 |
container_title | Computational and structural biotechnology journal |
container_volume | 24 |
creator | Tong, Le Li, Tianjiu Zhang, Qian Zhang, Qin Zhu, Renchaoli Du, Wei Hu, Pengwei |
description | The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements.
•A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach. |
doi_str_mv | 10.1016/j.csbj.2024.03.003 |
format | article |
fullrecord | <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_24ff6196ab994c96a17a781d636d7ab0</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S2001037024000564</els_id><doaj_id>oai_doaj_org_article_24ff6196ab994c96a17a781d636d7ab0</doaj_id><sourcerecordid>3033007860</sourcerecordid><originalsourceid>FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</originalsourceid><addsrcrecordid>eNp9kU1v1DAQhi0EolXpH-CAcuRAgj82toOQUFXxUWkFly3iZtnOeOs0iYvt3Yp_j9MtVXvBB8_YnnlmPC9CrwluCCb8_dDYZIaGYrpqMGswZs_QMcWY1JgJ_PyRf4ROUxpwWZLwjuGX6IjJVlDC5TH6tfY__ab-DvlDdVZdLk49-mt4V41-e5VvYdmrTdRzciFOEKsZ8m2I11U5VhGyn_VY7SElGKsE2wnmrLMP8yv0wukxwem9PUGXXz5vzr_V6x9fL87P1rVtKc1157iThhLRCdBUGiOJpUZrLo3kjlLXOSGMcURwYZx1hgFhtG35ykrRcsxO0MWB2wc9qJvoJx3_qKC9ursIcat0zN6OoOjKOU46rk3XrWyxRGghSc8Z74U2C-vTgXWzMxP0tvwl6vEJ9OnL7K_UNuwVwZ0UUopCeHtPiOH3DlJWk08WxlHPEHZJMcwYxkLeNU4PoTaGlCK4hzoEq0ViNahFYrVIrDBTReKS9OZxhw8p_wQtAR8PAVBmvvcQVbIeZgu9j2BzGYr_H_8v7cq38A</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3033007860</pqid></control><display><type>article</type><title>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</title><source>ScienceDirect Journals</source><source>PubMed Central</source><creator>Tong, Le ; Li, Tianjiu ; Zhang, Qian ; Zhang, Qin ; Zhu, Renchaoli ; Du, Wei ; Hu, Pengwei</creator><creatorcontrib>Tong, Le ; Li, Tianjiu ; Zhang, Qian ; Zhang, Qin ; Zhu, Renchaoli ; Du, Wei ; Hu, Pengwei</creatorcontrib><description>The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements.
•A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach.</description><identifier>ISSN: 2001-0370</identifier><identifier>EISSN: 2001-0370</identifier><identifier>DOI: 10.1016/j.csbj.2024.03.003</identifier><identifier>PMID: 38572168</identifier><language>eng</language><publisher>Netherlands: Elsevier B.V</publisher><subject>Joint loss ; Lightweight ; Retinal vessel segmentation ; Transformer</subject><ispartof>Computational and structural biotechnology journal, 2024-12, Vol.24, p.213-224</ispartof><rights>2024 The Author(s)</rights><rights>2024 The Author(s).</rights><rights>2024 The Author(s) 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</citedby><cites>FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</cites><orcidid>0009-0002-0864-2211 ; 0000-0002-4861-4787</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC10987887/pdf/$$EPDF$$P50$$Gpubmedcentral$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S2001037024000564$$EHTML$$P50$$Gelsevier$$Hfree_for_read</linktohtml><link.rule.ids>230,314,727,780,784,885,3549,27924,27925,45780,53791,53793</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/38572168$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Tong, Le</creatorcontrib><creatorcontrib>Li, Tianjiu</creatorcontrib><creatorcontrib>Zhang, Qian</creatorcontrib><creatorcontrib>Zhang, Qin</creatorcontrib><creatorcontrib>Zhu, Renchaoli</creatorcontrib><creatorcontrib>Du, Wei</creatorcontrib><creatorcontrib>Hu, Pengwei</creatorcontrib><title>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</title><title>Computational and structural biotechnology journal</title><addtitle>Comput Struct Biotechnol J</addtitle><description>The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements.
•A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach.</description><subject>Joint loss</subject><subject>Lightweight</subject><subject>Retinal vessel segmentation</subject><subject>Transformer</subject><issn>2001-0370</issn><issn>2001-0370</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>DOA</sourceid><recordid>eNp9kU1v1DAQhi0EolXpH-CAcuRAgj82toOQUFXxUWkFly3iZtnOeOs0iYvt3Yp_j9MtVXvBB8_YnnlmPC9CrwluCCb8_dDYZIaGYrpqMGswZs_QMcWY1JgJ_PyRf4ROUxpwWZLwjuGX6IjJVlDC5TH6tfY__ab-DvlDdVZdLk49-mt4V41-e5VvYdmrTdRzciFOEKsZ8m2I11U5VhGyn_VY7SElGKsE2wnmrLMP8yv0wukxwem9PUGXXz5vzr_V6x9fL87P1rVtKc1157iThhLRCdBUGiOJpUZrLo3kjlLXOSGMcURwYZx1hgFhtG35ykrRcsxO0MWB2wc9qJvoJx3_qKC9ursIcat0zN6OoOjKOU46rk3XrWyxRGghSc8Z74U2C-vTgXWzMxP0tvwl6vEJ9OnL7K_UNuwVwZ0UUopCeHtPiOH3DlJWk08WxlHPEHZJMcwYxkLeNU4PoTaGlCK4hzoEq0ViNahFYrVIrDBTReKS9OZxhw8p_wQtAR8PAVBmvvcQVbIeZgu9j2BzGYr_H_8v7cq38A</recordid><startdate>20241201</startdate><enddate>20241201</enddate><creator>Tong, Le</creator><creator>Li, Tianjiu</creator><creator>Zhang, Qian</creator><creator>Zhang, Qin</creator><creator>Zhu, Renchaoli</creator><creator>Du, Wei</creator><creator>Hu, Pengwei</creator><general>Elsevier B.V</general><general>Research Network of Computational and Structural Biotechnology</general><general>Elsevier</general><scope>6I.</scope><scope>AAFTH</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>5PM</scope><scope>DOA</scope><orcidid>https://orcid.org/0009-0002-0864-2211</orcidid><orcidid>https://orcid.org/0000-0002-4861-4787</orcidid></search><sort><creationdate>20241201</creationdate><title>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</title><author>Tong, Le ; Li, Tianjiu ; Zhang, Qian ; Zhang, Qin ; Zhu, Renchaoli ; Du, Wei ; Hu, Pengwei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Joint loss</topic><topic>Lightweight</topic><topic>Retinal vessel segmentation</topic><topic>Transformer</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Tong, Le</creatorcontrib><creatorcontrib>Li, Tianjiu</creatorcontrib><creatorcontrib>Zhang, Qian</creatorcontrib><creatorcontrib>Zhang, Qin</creatorcontrib><creatorcontrib>Zhu, Renchaoli</creatorcontrib><creatorcontrib>Du, Wei</creatorcontrib><creatorcontrib>Hu, Pengwei</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Computational and structural biotechnology journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Tong, Le</au><au>Li, Tianjiu</au><au>Zhang, Qian</au><au>Zhang, Qin</au><au>Zhu, Renchaoli</au><au>Du, Wei</au><au>Hu, Pengwei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation</atitle><jtitle>Computational and structural biotechnology journal</jtitle><addtitle>Comput Struct Biotechnol J</addtitle><date>2024-12-01</date><risdate>2024</risdate><volume>24</volume><spage>213</spage><epage>224</epage><pages>213-224</pages><issn>2001-0370</issn><eissn>2001-0370</eissn><abstract>The intricate task of precisely segmenting retinal vessels from images, which is critical for diagnosing various eye diseases, presents significant challenges for models due to factors such as scale variation, complex anatomical patterns, low contrast, and limitations in training data. Building on these challenges, we offer novel contributions spanning model architecture, loss function design, robustness, and real-time efficacy. To comprehensively address these challenges, a new U-Net-like, lightweight Transformer network for retinal vessel segmentation is presented. By integrating MobileViT+ and a novel local representation in the encoder, our design emphasizes lightweight processing while capturing intricate image structures, enhancing vessel edge precision. A novel joint loss is designed, leveraging the characteristics of weighted cross-entropy and Dice loss to effectively guide the model through the task's challenges, such as foreground-background imbalance and intricate vascular structures. Exhaustive experiments were performed on three prominent retinal image databases. The results underscore the robustness and generalizability of the proposed LiViT-Net, which outperforms other methods in complex scenarios, especially in intricate environments with fine vessels or vessel edges. Importantly, optimized for efficiency, LiViT-Net excels on devices with constrained computational power, as evidenced by its fast performance. To demonstrate the model proposed in this study, a freely accessible and interactive website was established (https://hz-t3.matpool.com:28765?token=aQjYR4hqMI), revealing real-time performance with no login requirements.
•A lightweight Transformer network is developed for retinal vessel segmentation.•In the MobileViT+ block, parallel convolutions enhance local representation, improving ViT's bias and interpatch relations.•A remapped, weighted joint loss mechanism is introduced to address pixel imbalances.•Extensive tests on DRIVE, CHASEDB1 and HRF datasets demonstrate the robustness and computational efficiency of our approach.</abstract><cop>Netherlands</cop><pub>Elsevier B.V</pub><pmid>38572168</pmid><doi>10.1016/j.csbj.2024.03.003</doi><tpages>12</tpages><orcidid>https://orcid.org/0009-0002-0864-2211</orcidid><orcidid>https://orcid.org/0000-0002-4861-4787</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2001-0370 |
ispartof | Computational and structural biotechnology journal, 2024-12, Vol.24, p.213-224 |
issn | 2001-0370 2001-0370 |
language | eng |
recordid | cdi_doaj_primary_oai_doaj_org_article_24ff6196ab994c96a17a781d636d7ab0 |
source | ScienceDirect Journals; PubMed Central |
subjects | Joint loss Lightweight Retinal vessel segmentation Transformer |
title | LiViT-Net: A U-Net-like, lightweight Transformer network for retinal vessel segmentation |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T12%3A00%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=LiViT-Net:%20A%20U-Net-like,%20lightweight%20Transformer%20network%20for%20retinal%20vessel%20segmentation&rft.jtitle=Computational%20and%20structural%20biotechnology%20journal&rft.au=Tong,%20Le&rft.date=2024-12-01&rft.volume=24&rft.spage=213&rft.epage=224&rft.pages=213-224&rft.issn=2001-0370&rft.eissn=2001-0370&rft_id=info:doi/10.1016/j.csbj.2024.03.003&rft_dat=%3Cproquest_doaj_%3E3033007860%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c522t-9f6f8b21797ea28bb81c2baa68b86f22f9f77bbf1767bfcfb3e1325564c875603%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3033007860&rft_id=info:pmid/38572168&rfr_iscdi=true |