Loading…

Assessing parameter efficient methods for pre-trained language model in annotating scRNA-seq data

Annotating cell types of single-cell RNA sequencing (scRNA-seq) data is crucial for studying cellular heterogeneity in the tumor microenvironment. Recently, large-scale pre-trained language models (PLMs) have achieved significant progress in cell-type annotation of scRNA-seq data. This approach effe...

Full description

Saved in:
Bibliographic Details
Published in:Methods (San Diego, Calif.) Calif.), 2024-08, Vol.228, p.12-21
Main Authors: Xia, Yucheng, Liu, Yuhang, Li, Tianhao, He, Sihan, Chang, Hong, Wang, Yaqing, Zhang, Yongqing, Ge, Wenyi
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Annotating cell types of single-cell RNA sequencing (scRNA-seq) data is crucial for studying cellular heterogeneity in the tumor microenvironment. Recently, large-scale pre-trained language models (PLMs) have achieved significant progress in cell-type annotation of scRNA-seq data. This approach effectively addresses previous methods' shortcomings in performance and generalization. However, fine-tuning PLMs for different downstream tasks demands considerable computational resources, rendering it impractical. Hence, a new research branch introduces parameter-efficient fine-tuning (PEFT). This involves optimizing a few parameters while leaving the majority unchanged, leading to substantial reductions in computational expenses. Here, we utilize scBERT, a large-scale pre-trained model, to explore the capabilities of three PEFT methods in scRNA-seq cell type annotation. Extensive benchmark studies across several datasets demonstrate the superior applicability of PEFT methods. Furthermore, downstream analysis using models obtained through PEFT showcases their utility in novel cell type discovery and model interpretability for potential marker genes. Our findings underscore the considerable potential of PEFT in PLM-based cell type annotation, presenting novel perspectives for the analysis of scRNA-seq data. •Employ scBERT to investigate three parameter-efficient fine-tuning (PEFT) methods.•Investigation of the PEFT methods in discovering new cell types.•Exploration of the interpretability of models fine-tuned using PEFT methods.
ISSN:1046-2023
1095-9130
DOI:10.1016/j.ymeth.2024.05.007