Loading…
PCTNet: depth estimation from single structured light image with a parallel CNN-transformer network
Recent approaches based on convolutional neural networks significantly improve the performance of structured light image depth estimation in fringe projection and speckle projection 3D measurement. However, it remains challenging to simultaneously preserve the global structure and local details of o...
Saved in:
Published in: | Measurement science & technology 2023-08, Vol.34 (8), p.85402 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Recent approaches based on convolutional neural networks significantly improve the performance of structured light image depth estimation in fringe projection and speckle projection 3D measurement. However, it remains challenging to simultaneously preserve the global structure and local details of objects for the structured light images in complex scenes. In this paper, we design a parallel CNN-transformer network (PCTNet), which consists of a CNN branch, a transformer branch, a bidirectional feature fusion module (BFFM), and a cross-feature multi-scale fusion module (CFMS). The BFFM and CFMS modules are proposed to fuse local and global features of the double branches in order to achieve better depth estimation. Comprehensive experiments are conducted to evaluate our model on four structured light datasets, i.e. our established simulated fringe and speckle structured light datasets, and public real fringe and speckle structured light datasets. Experiments demonstrate that the proposed PCTNet is an effective architecture, achieving state-of-the-art performance in both qualitative and quantitative evaluation. |
---|---|
ISSN: | 0957-0233 1361-6501 |
DOI: | 10.1088/1361-6501/acd136 |