Loading…

TENet: Accurate light-field salient object detection with a transformer embedding network

Current light-field salient object detection methods have difficulty in accurately distinguishing objects from complex backgrounds. In this paper, we believe that this problem can be mitigated by optimizing feature fusion and enlarging receptive field, and thus propose a novel transformer embedding...

Full description

Saved in:
Bibliographic Details
Published in:Image and vision computing 2023-01, Vol.129, p.104595, Article 104595
Main Authors: Wang, Xingzheng, Chen, Songwei, Wei, Guoyao, Liu, Jiehao
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Current light-field salient object detection methods have difficulty in accurately distinguishing objects from complex backgrounds. In this paper, we believe that this problem can be mitigated by optimizing feature fusion and enlarging receptive field, and thus propose a novel transformer embedding network named TENet. The main idea of the network is to (1) selectively aggregate multi-features for fuller feature fusion; (2) integrate the Transformer for larger receptive field, so as to accurately identify salient objects. For the former, firstly, a multi-modal feature fusion module (MMFF) is designed to mine the different contributions of multi-modal features (i.e., all-in-focus image features and focal stack features). Then, a multi-level feature fusion module (MLFF) is developed to iteratively select and fuse complementary cues from multi-level features. For the latter, we integrate the Transformer for the first time and propose a transformer-based feature enhancement module (TFE), to provide a wider receptive field for each pixel of high-level features. To validate our idea, we comprehensively evaluate the performance of our TENet on three challenging datasets. Experimental results show that our method outperforms the state-of-the-art method, e.g., the detection accuracy is improved by 28.1%, 20.3%, and 14.9% in MAE metric, respectively.
ISSN:0262-8856
1872-8138
DOI:10.1016/j.imavis.2022.104595