Loading…
Personalized facial beauty assessment: a meta-learning approach
Automatic facial beauty assessment has recently attracted a growing interest and achieved impressive results. However, despite the obvious subjectivity of beauty perception, most studies are addressed to predict generic or universal beauty and only few works investigate an individual’s preferences i...
Saved in:
Published in: | The Visual computer 2023-03, Vol.39 (3), p.1095-1107 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Automatic facial beauty assessment has recently attracted a growing interest and achieved impressive results. However, despite the obvious subjectivity of beauty perception, most studies are addressed to predict generic or universal beauty and only few works investigate an individual’s preferences in facial attractiveness. Unlike universal beauty assessment, an effective personalized method is required to produce a reasonable accuracy on a small amount of training images as the number of annotated samples from an individual is limited in real-world applications. In this work, a novel personalized facial beauty assessment approach based on meta-learning is introduced. First of all, beauty preferences shared by an extensive number of individuals are learnt during meta-training. Then, the model is adapted to a new individual with a few rated image samples in the meta-testing phase. The experiments are conducted on a facial beauty dataset that includes faces of various ethnic, gender, age groups and rated by hundreds of volunteers with different social and cultural backgrounds. The results demonstrate that the proposed method is capable of effectively learning personal beauty preferences from a limited number of annotated images and outperforms the facial beauty prediction state-of-the-art on quantitative comparisons. |
---|---|
ISSN: | 0178-2789 1432-2315 |
DOI: | 10.1007/s00371-021-02387-w |