Loading…

Predicting and explaining performance and diversity of neural network architecture for semantic segmentation

This paper proposes searching for network architectures which achieve similar performance while promoting diversity, in order to facilitate ensembling. We explain prediction performance and diversity of various network sizes and activation functions applied to semantic segmentation of the CityScapes...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications 2023-03, Vol.214, p.119101, Article 119101
Main Authors: Graham-Knight, John Brandon, Bond, Corey, Najjaran, Homayoun, Lucet, Yves, Lasserre, Patricia
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This paper proposes searching for network architectures which achieve similar performance while promoting diversity, in order to facilitate ensembling. We explain prediction performance and diversity of various network sizes and activation functions applied to semantic segmentation of the CityScapes dataset. We show that both performance and diversity can be predicted from neural network architecture using explainable boosting machines. A grid search of 144 models is performed, and many of the models exhibit no significant difference in mean performance within a 95% confidence interval. Notably, we find the best performing models have varied network architecture parameters. The explanations for performance largely agree with the accepted wisdom of the machine learning community, which shows that the method is extracting information of value. We find that diversity between models can be achieved by varying network size. Moreover, homogeneous network sizes generally show positive correlation in predictions, and larger models tend to converge to similar solutions. These explanations provide a better understanding of the effects of network parameters to deep learning practitioners; they could also be used in place of naïve search methods or a model pool to inform growing an ensemble. •Confirm explainable boosting machines predict performance of network architectures.•Show pairwise diversity of neural networks is predictable similar to performance.•Explain network architecture parameter effect on diversity.•Show that diversity can be promoted through architectural differences.•Show that diversity through architecture can maintain underlying model performance.
ISSN:0957-4174
1873-6793
DOI:10.1016/j.eswa.2022.119101