Loading…

MICD: More intra-class diversity in few-shot text classification with many classes

Few-shot learning has gained much interest and achieved remarkable performance in handling limited data scenarios. However, existing few-shot text classification methods typically aim at classifying a limited number of classes, usually ranging from 5 to 10, posing a challenge for many real-world tas...

Full description

Saved in:
Bibliographic Details
Published in:Knowledge-based systems 2025-01, Vol.309, p.112851, Article 112851
Main Authors: Jang, Gwangseon, Jeong, Hyeon Ji, Yi, Mun Yong
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Few-shot learning has gained much interest and achieved remarkable performance in handling limited data scenarios. However, existing few-shot text classification methods typically aim at classifying a limited number of classes, usually ranging from 5 to 10, posing a challenge for many real-world tasks that require few-shot text classification for many classes. Few-shot text classification for many classes has rarely been studied and it is a challenging problem. Distinguishing differences among many classes is more challenging than distinguishing differences among small classes. To address this issue, we propose a new few-shot text classification model for many classes called MICD (More Intra-Class Diversity in few-shot text classification with many classes). Our model comprises two crucial components: Intra-Class Diversity Contrastive Learning (ICDCL) and Intra-Class Augmentation (ICA). ICDCL trains an encoder to enhance feature discriminability by maintaining both intra-class diversity and inter-class specificity, effectively improving generalization performance, even when data is limited. ICA addresses data scarcity by selecting diverse support samples and applying intra-class mix-up, enabling robust generalization to out-of-distribution data—an essential consideration in many-class few-shot learning scenarios. Experimental results on four real datasets show that MICD provides significant performance improvement over the other state-of-the-art approaches.
ISSN:0950-7051
DOI:10.1016/j.knosys.2024.112851