Loading…

A Unified Model for Style Classification and Emotional Response Analysis

The emergence of Convolutional Neural Networks (CNNs) and Vision Transformers (ViT) has markedly transformed the field of image classification and analysis, especially within the realm of computer vision. This advancement has significantly impacted various sectors, including medical diagnostics and...

Full description

Saved in:
Bibliographic Details
Published in:IEEE access 2024, Vol.12, p.91770-91779
Main Author: Yin, Chu-Ze
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The emergence of Convolutional Neural Networks (CNNs) and Vision Transformers (ViT) has markedly transformed the field of image classification and analysis, especially within the realm of computer vision. This advancement has significantly impacted various sectors, including medical diagnostics and autonomous driving, while also fostering novel intersections with artistic exploration. Despite these advancements, the challenge of seamlessly integrating art style classification with emotion prediction remains. The complex interplay between an artwork's style and the emotional reactions it triggers requires a refined methodology to accurately encapsulate this dynamic relationship. Addressing this challenge, our study presents a Unified Model for Art Style and Emotion Prediction (ASE), which adopts a multi-task learning approach. This model is structured around three main elements: Artwork Style Classification, Emotion Prediction for viewers of art, and a Task-Specific Attention Module. By incorporating a pre-trained image encoder alongside a task-specific attention mechanism, our framework facilitates the concurrent processing of multiple tasks, while honing in on specialized feature representations. The efficacy of our model is validated through the Artemis dataset, demonstrating its proficiency in both precise art style classification and the identification of emotional responses. This highlights its capability to navigate the complex relationships present within artworks effectively.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2024.3419851