Loading…

A fragmented neural network ensemble method and its application to image classification

In recent years, deep neural networks have evolved rapidly in engineering technology, with models becoming larger and deeper. However, for most companies, developing large models is extremely costly and highly risky. Researchers usually focus on the performance of the model, neglecting its cost and...

Full description

Saved in:
Bibliographic Details
Published in:Scientific reports 2024-01, Vol.14 (1), p.2291-2291, Article 2291
Main Authors: Zhang, Xu, Liu, Shuai, Wang, Xueli, Li, Yumei
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In recent years, deep neural networks have evolved rapidly in engineering technology, with models becoming larger and deeper. However, for most companies, developing large models is extremely costly and highly risky. Researchers usually focus on the performance of the model, neglecting its cost and accessibility. In fact, most regular business scenarios do not require high-level AI. A simple and inexpensive modeling method for fulfilling certain demands for practical applications of AI is needed. In this paper, a Fragmented neural network method is proposed. Inspired by the random forest algorithm, both the samples and features are randomly sampled on image data. Images are randomly split into smaller pieces. Weak neural networks are trained using these fragmented images, and many weak neural networks are then ensembled to build a strong neural network by voting. In this way, sufficient accuracy is achieved while reducing the complexity and data volume of each base learner, enabling mass production through parallel and distributed computing. By conducting experiments on the MNIST and CIFAR10 datasets, we build a model pool using FNN, CNN, DenseNet, and ResNet as the basic network structure. We find that the accuracy of the ensemble weak network is significantly higher than that of each base learner. Meanwhile, the accuracy of the ensemble network is highly dependent on the performance of each base learner. The accuracy of the ensemble network is comparable to or even exceeds that of the full model and has better robustness. Unlike other similar studies, we do not pursue SOTA models. Instead, we achieved results close to the full model with a smaller number of parameters and amount of data.
ISSN:2045-2322
2045-2322
DOI:10.1038/s41598-024-52945-0