Loading…
Improving disentanglement in variational auto-encoders via feature imbalance-informed dimension weighting
Using Variational Auto-Encoder (VAE) to learn disentangled representation holds great promise. But there is a feature imbalance in the learning process of VAEs, and the model usually concentrates on the learning of some dimensions at the expense of others. Instead of attempting to rectify it, we exp...
Saved in:
Published in: | Knowledge-based systems 2024-07, Vol.296, p.111818, Article 111818 |
---|---|
Main Authors: | , , , , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Using Variational Auto-Encoder (VAE) to learn disentangled representation holds great promise. But there is a feature imbalance in the learning process of VAEs, and the model usually concentrates on the learning of some dimensions at the expense of others. Instead of attempting to rectify it, we exploit feature imbalance to propose a Dimension Weighting method that can boost the disentanglement effect of VAE-based models. In order to conduct disentanglement learning under a fixed size of the latent space, the intrinsic dimension of raw data is estimated by a Dimension Number Estimator and set as the latent space size. Then, leveraging the feature imbalance, a Dimension Importance Evaluator is constructed to separate each dimension in the latent variable into important, unimportant, and general dimensions. By exerting different learning pressures on specific dimensions, we further optimize the variational lower bound of model and retrain it, thus promoting the disentanglement of important dimensions. The experiments on four benchmark datasets show that the Dimension Weighting can further improve the disentangling effect without compromising model performance, and in approximately 80% of cases, the results in disentangling metric evaluation experiments achieve better disentangling scores than the original models. This reveals that not all dimensions of the latent variables are equally influential. By focusing on the crucial dimensions within the latent representation, the model can achieve better performance. |
---|---|
ISSN: | 0950-7051 1872-7409 |
DOI: | 10.1016/j.knosys.2024.111818 |