Loading…

A multi‐label social short text classification method based on contrastive learning and improved ml‐KNN

Short texts on social platforms often have the problems of diverse categories and semantic sparsity, making it challenging to identify the diverse intentions of users. To address this issue, this article proposes a multi‐label social short text classification method (IML‐CL) based on contrastive lea...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems 2024-07, Vol.41 (7), p.n/a
Main Authors: Tian, Gang, Wang, Jiachang, Wang, Rui, Zhao, Guangxin, He, Cheng
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Short texts on social platforms often have the problems of diverse categories and semantic sparsity, making it challenging to identify the diverse intentions of users. To address this issue, this article proposes a multi‐label social short text classification method (IML‐CL) based on contrastive learning and improved ml‐KNN. First, a contrastive learning approach is employed to train a multi‐label text classification model. This approach improves semantic sparsity by leveraging the knowledge from the existing samples to enrich the feature representation of short texts. Simultaneously, an improved ml‐KNN algorithm is developed to enhance the accuracy of label prediction. This algorithm utilizes a two‐layer nearest neighbor rule and introduces a penalty function and weight optimization. Next, the model generates the feature representation for the test sample and predicts its label. Additionally, the improved ml‐KNN algorithm retrieves neighbors of the test sample and uses their label information for prediction. Finally, the two predictions are combined to obtain the final prediction, which accurately identifies the user's intention. The experimental results demonstrate that, on the dataset constructed in this article, the IML‐CL method effectively boosts the performance of the baseline model.
ISSN:0266-4720
1468-0394
DOI:10.1111/exsy.13547