Loading…

A vision transformer-based framework for knowledge transfer from multi-modal to mono-modal lymphoma subtyping models

Determining lymphoma subtypes is a crucial step for better patient treatment targeting to potentially increase their survival chances. In this context, the existing gold standard diagnosis method, which relies on gene expression technology, is highly expensive and time-consuming, making it less acce...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2024-05
Main Authors: Guetarni, Bilel, Windal, Feryal, Halim Benhabiles, Petit, Marianne, Dubois, Romain, Leteurtre, Emmanuelle, Collard, Dominique
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Guetarni, Bilel
Windal, Feryal
Halim Benhabiles
Petit, Marianne
Dubois, Romain
Leteurtre, Emmanuelle
Collard, Dominique
description Determining lymphoma subtypes is a crucial step for better patient treatment targeting to potentially increase their survival chances. In this context, the existing gold standard diagnosis method, which relies on gene expression technology, is highly expensive and time-consuming, making it less accessibility. Although alternative diagnosis methods based on IHC (immunohistochemistry) technologies exist (recommended by the WHO), they still suffer from similar limitations and are less accurate. Whole Slide Image (WSI) analysis using deep learning models has shown promising potential for cancer diagnosis, that could offer cost-effective and faster alternatives to existing methods. In this work, we propose a vision transformer-based framework for distinguishing DLBCL (Diffuse Large B-Cell Lymphoma) cancer subtypes from high-resolution WSIs. To this end, we introduce a multi-modal architecture to train a classifier model from various WSI modalities. We then leverage this model through a knowledge distillation process to efficiently guide the learning of a mono-modal classifier. Our experimental study conducted on a lymphoma dataset of 157 patients shows the promising performance of our mono-modal classification model, outperforming six recent state-of-the-art methods. In addition, the power-law curve, estimated on our experimental data, suggests that with more training data from a reasonable number of additional patients, our model could achieve competitive diagnosis accuracy with IHC technologies. Furthermore, the efficiency of our framework is confirmed through an additional experimental study on an external breast cancer dataset (BCI dataset).
doi_str_mv 10.48550/arxiv.2308.01328
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2845948718</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2845948718</sourcerecordid><originalsourceid>FETCH-LOGICAL-a1298-84cac12283d1bb5c2811806fc0a55cfdedb928580f969dd907ce36d562879af13</originalsourceid><addsrcrecordid>eNotTslqwzAUFIVCQ5oP6E3Qs1PpybKfjiF0g0Av7TnIlpQ6sSxXspPm72uoT8MszAwhD5ytc5SSPen425zXIBiuGReAN2QBQvAMc4A7skrpyBiDogQpxYIMG3puUhM6OkTdJReitzGrdLKGuqi9vYR4opNMT124tNYc7Jy0cQoET_3YDk3mg9EtHQL1oQsza6--_w5e0zRWw7VvusPkGtume3LrdJvsasYl-Xp5_ty-ZbuP1_ftZpdpDgqnx7WuOQAKw6tK1oCcIytczbSUtTPWVApQInOqUMYoVtZWFEYWgKXSjoslefzv7WP4GW0a9scwxm6a3APmUuVYchR_iMpfbw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2845948718</pqid></control><display><type>article</type><title>A vision transformer-based framework for knowledge transfer from multi-modal to mono-modal lymphoma subtyping models</title><source>ProQuest - Publicly Available Content Database</source><creator>Guetarni, Bilel ; Windal, Feryal ; Halim Benhabiles ; Petit, Marianne ; Dubois, Romain ; Leteurtre, Emmanuelle ; Collard, Dominique</creator><creatorcontrib>Guetarni, Bilel ; Windal, Feryal ; Halim Benhabiles ; Petit, Marianne ; Dubois, Romain ; Leteurtre, Emmanuelle ; Collard, Dominique</creatorcontrib><description>Determining lymphoma subtypes is a crucial step for better patient treatment targeting to potentially increase their survival chances. In this context, the existing gold standard diagnosis method, which relies on gene expression technology, is highly expensive and time-consuming, making it less accessibility. Although alternative diagnosis methods based on IHC (immunohistochemistry) technologies exist (recommended by the WHO), they still suffer from similar limitations and are less accurate. Whole Slide Image (WSI) analysis using deep learning models has shown promising potential for cancer diagnosis, that could offer cost-effective and faster alternatives to existing methods. In this work, we propose a vision transformer-based framework for distinguishing DLBCL (Diffuse Large B-Cell Lymphoma) cancer subtypes from high-resolution WSIs. To this end, we introduce a multi-modal architecture to train a classifier model from various WSI modalities. We then leverage this model through a knowledge distillation process to efficiently guide the learning of a mono-modal classifier. Our experimental study conducted on a lymphoma dataset of 157 patients shows the promising performance of our mono-modal classification model, outperforming six recent state-of-the-art methods. In addition, the power-law curve, estimated on our experimental data, suggests that with more training data from a reasonable number of additional patients, our model could achieve competitive diagnosis accuracy with IHC technologies. Furthermore, the efficiency of our framework is confirmed through an additional experimental study on an external breast cancer dataset (BCI dataset).</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2308.01328</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Cancer ; Classification ; Classifiers ; Deep learning ; Diagnosis ; Distillation ; Gene expression ; Knowledge management ; Lymphoma ; Machine learning</subject><ispartof>arXiv.org, 2024-05</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2845948718?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,27925,37012,44590</link.rule.ids></links><search><creatorcontrib>Guetarni, Bilel</creatorcontrib><creatorcontrib>Windal, Feryal</creatorcontrib><creatorcontrib>Halim Benhabiles</creatorcontrib><creatorcontrib>Petit, Marianne</creatorcontrib><creatorcontrib>Dubois, Romain</creatorcontrib><creatorcontrib>Leteurtre, Emmanuelle</creatorcontrib><creatorcontrib>Collard, Dominique</creatorcontrib><title>A vision transformer-based framework for knowledge transfer from multi-modal to mono-modal lymphoma subtyping models</title><title>arXiv.org</title><description>Determining lymphoma subtypes is a crucial step for better patient treatment targeting to potentially increase their survival chances. In this context, the existing gold standard diagnosis method, which relies on gene expression technology, is highly expensive and time-consuming, making it less accessibility. Although alternative diagnosis methods based on IHC (immunohistochemistry) technologies exist (recommended by the WHO), they still suffer from similar limitations and are less accurate. Whole Slide Image (WSI) analysis using deep learning models has shown promising potential for cancer diagnosis, that could offer cost-effective and faster alternatives to existing methods. In this work, we propose a vision transformer-based framework for distinguishing DLBCL (Diffuse Large B-Cell Lymphoma) cancer subtypes from high-resolution WSIs. To this end, we introduce a multi-modal architecture to train a classifier model from various WSI modalities. We then leverage this model through a knowledge distillation process to efficiently guide the learning of a mono-modal classifier. Our experimental study conducted on a lymphoma dataset of 157 patients shows the promising performance of our mono-modal classification model, outperforming six recent state-of-the-art methods. In addition, the power-law curve, estimated on our experimental data, suggests that with more training data from a reasonable number of additional patients, our model could achieve competitive diagnosis accuracy with IHC technologies. Furthermore, the efficiency of our framework is confirmed through an additional experimental study on an external breast cancer dataset (BCI dataset).</description><subject>Cancer</subject><subject>Classification</subject><subject>Classifiers</subject><subject>Deep learning</subject><subject>Diagnosis</subject><subject>Distillation</subject><subject>Gene expression</subject><subject>Knowledge management</subject><subject>Lymphoma</subject><subject>Machine learning</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNotTslqwzAUFIVCQ5oP6E3Qs1PpybKfjiF0g0Av7TnIlpQ6sSxXspPm72uoT8MszAwhD5ytc5SSPen425zXIBiuGReAN2QBQvAMc4A7skrpyBiDogQpxYIMG3puUhM6OkTdJReitzGrdLKGuqi9vYR4opNMT124tNYc7Jy0cQoET_3YDk3mg9EtHQL1oQsza6--_w5e0zRWw7VvusPkGtume3LrdJvsasYl-Xp5_ty-ZbuP1_ftZpdpDgqnx7WuOQAKw6tK1oCcIytczbSUtTPWVApQInOqUMYoVtZWFEYWgKXSjoslefzv7WP4GW0a9scwxm6a3APmUuVYchR_iMpfbw</recordid><startdate>20240529</startdate><enddate>20240529</enddate><creator>Guetarni, Bilel</creator><creator>Windal, Feryal</creator><creator>Halim Benhabiles</creator><creator>Petit, Marianne</creator><creator>Dubois, Romain</creator><creator>Leteurtre, Emmanuelle</creator><creator>Collard, Dominique</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240529</creationdate><title>A vision transformer-based framework for knowledge transfer from multi-modal to mono-modal lymphoma subtyping models</title><author>Guetarni, Bilel ; Windal, Feryal ; Halim Benhabiles ; Petit, Marianne ; Dubois, Romain ; Leteurtre, Emmanuelle ; Collard, Dominique</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a1298-84cac12283d1bb5c2811806fc0a55cfdedb928580f969dd907ce36d562879af13</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Cancer</topic><topic>Classification</topic><topic>Classifiers</topic><topic>Deep learning</topic><topic>Diagnosis</topic><topic>Distillation</topic><topic>Gene expression</topic><topic>Knowledge management</topic><topic>Lymphoma</topic><topic>Machine learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Guetarni, Bilel</creatorcontrib><creatorcontrib>Windal, Feryal</creatorcontrib><creatorcontrib>Halim Benhabiles</creatorcontrib><creatorcontrib>Petit, Marianne</creatorcontrib><creatorcontrib>Dubois, Romain</creatorcontrib><creatorcontrib>Leteurtre, Emmanuelle</creatorcontrib><creatorcontrib>Collard, Dominique</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>ProQuest - Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Guetarni, Bilel</au><au>Windal, Feryal</au><au>Halim Benhabiles</au><au>Petit, Marianne</au><au>Dubois, Romain</au><au>Leteurtre, Emmanuelle</au><au>Collard, Dominique</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>A vision transformer-based framework for knowledge transfer from multi-modal to mono-modal lymphoma subtyping models</atitle><jtitle>arXiv.org</jtitle><date>2024-05-29</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Determining lymphoma subtypes is a crucial step for better patient treatment targeting to potentially increase their survival chances. In this context, the existing gold standard diagnosis method, which relies on gene expression technology, is highly expensive and time-consuming, making it less accessibility. Although alternative diagnosis methods based on IHC (immunohistochemistry) technologies exist (recommended by the WHO), they still suffer from similar limitations and are less accurate. Whole Slide Image (WSI) analysis using deep learning models has shown promising potential for cancer diagnosis, that could offer cost-effective and faster alternatives to existing methods. In this work, we propose a vision transformer-based framework for distinguishing DLBCL (Diffuse Large B-Cell Lymphoma) cancer subtypes from high-resolution WSIs. To this end, we introduce a multi-modal architecture to train a classifier model from various WSI modalities. We then leverage this model through a knowledge distillation process to efficiently guide the learning of a mono-modal classifier. Our experimental study conducted on a lymphoma dataset of 157 patients shows the promising performance of our mono-modal classification model, outperforming six recent state-of-the-art methods. In addition, the power-law curve, estimated on our experimental data, suggests that with more training data from a reasonable number of additional patients, our model could achieve competitive diagnosis accuracy with IHC technologies. Furthermore, the efficiency of our framework is confirmed through an additional experimental study on an external breast cancer dataset (BCI dataset).</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2308.01328</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-05
issn 2331-8422
language eng
recordid cdi_proquest_journals_2845948718
source ProQuest - Publicly Available Content Database
subjects Cancer
Classification
Classifiers
Deep learning
Diagnosis
Distillation
Gene expression
Knowledge management
Lymphoma
Machine learning
title A vision transformer-based framework for knowledge transfer from multi-modal to mono-modal lymphoma subtyping models
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T13%3A32%3A56IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=A%20vision%20transformer-based%20framework%20for%20knowledge%20transfer%20from%20multi-modal%20to%20mono-modal%20lymphoma%20subtyping%20models&rft.jtitle=arXiv.org&rft.au=Guetarni,%20Bilel&rft.date=2024-05-29&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2308.01328&rft_dat=%3Cproquest%3E2845948718%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a1298-84cac12283d1bb5c2811806fc0a55cfdedb928580f969dd907ce36d562879af13%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2845948718&rft_id=info:pmid/&rfr_iscdi=true