Loading…

DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction

As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolu...

Full description

Saved in:
Bibliographic Details
Published in:Sensors (Basel, Switzerland) Switzerland), 2022-03, Vol.22 (5), p.1914
Main Authors: Ibrahem, Hatem, Salem, Ahmed, Kang, Hyun-Soo
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53
cites cdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53
container_end_page
container_issue 5
container_start_page 1914
container_title Sensors (Basel, Switzerland)
container_volume 22
creator Ibrahem, Hatem
Salem, Ahmed
Kang, Hyun-Soo
description As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time.
doi_str_mv 10.3390/s22051914
format article
fullrecord <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_e6aad392684140fdb240b0ecb9af05fa</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><doaj_id>oai_doaj_org_article_e6aad392684140fdb240b0ecb9af05fa</doaj_id><sourcerecordid>2638712936</sourcerecordid><originalsourceid>FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</originalsourceid><addsrcrecordid>eNpdkc1u1DAUhS1ERcvAghdAkdiUhcF_cWwWSGhaYKRKSMx0w8ZynJtpRkk8tZ1KvD2epozarmyf8-n46F6E3lHyiXNNPkfGSEk1FS_QGRVMYJWFl4_up-h1jDtCGOdcvUKnvGQVJZKeoT8XmzW-gH26-VL8BtvjTTdAse7GbQ94NdgtFPducRlTN9jU-bG4jtmeZZw8Xu-tg2Jml36MKUzuwL1BJ63tI7x9OBfo-vvlZvkTX_36sVp-u8JOSJ0wcFeVgtVOKSG007bVUjDdMkeUo1zVLclvy6ECJ5paSgKalVaIknLeQskXaDXnNt7uzD7kmuGv8bYz94IPW2ND6lwPBqS1DddMKkEFaZuaCVITcHX-lZStzVlf56z9VA_QOBhTsP2T0KfO2N2Yrb8zKk9fy0OZ84eA4G8niMkMXXTQ93YEP0XDJFcVZZrLjH54hu78FMY8qgNVVaqSmVugjzPlgo8xQHssQ4k5bN8ct5_Z94_bH8n_6-b_AEFfqJ8</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2637787629</pqid></control><display><type>article</type><title>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</title><source>Publicly Available Content Database</source><source>PubMed Central</source><creator>Ibrahem, Hatem ; Salem, Ahmed ; Kang, Hyun-Soo</creator><creatorcontrib>Ibrahem, Hatem ; Salem, Ahmed ; Kang, Hyun-Soo</creatorcontrib><description>As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time.</description><identifier>ISSN: 1424-8220</identifier><identifier>EISSN: 1424-8220</identifier><identifier>DOI: 10.3390/s22051914</identifier><identifier>PMID: 35271061</identifier><language>eng</language><publisher>Switzerland: MDPI AG</publisher><subject>3-D graphics ; Accuracy ; Algorithms ; Autonomous vehicles ; Computer architecture ; Construction ; convolutional neural networks ; depth estimation ; Embedded systems ; Experiments ; Frames per second ; High resolution ; Methods ; Neural networks ; Real time ; real-time processing</subject><ispartof>Sensors (Basel, Switzerland), 2022-03, Vol.22 (5), p.1914</ispartof><rights>2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>2022 by the authors. 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</citedby><cites>FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</cites><orcidid>0000-0001-8722-3300 ; 0000-0002-4682-0368 ; 0000-0002-4333-2852</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2637787629/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2637787629?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>230,314,723,776,780,881,25733,27903,27904,36991,36992,44569,53769,53771,74872</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/35271061$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Ibrahem, Hatem</creatorcontrib><creatorcontrib>Salem, Ahmed</creatorcontrib><creatorcontrib>Kang, Hyun-Soo</creatorcontrib><title>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</title><title>Sensors (Basel, Switzerland)</title><addtitle>Sensors (Basel)</addtitle><description>As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time.</description><subject>3-D graphics</subject><subject>Accuracy</subject><subject>Algorithms</subject><subject>Autonomous vehicles</subject><subject>Computer architecture</subject><subject>Construction</subject><subject>convolutional neural networks</subject><subject>depth estimation</subject><subject>Embedded systems</subject><subject>Experiments</subject><subject>Frames per second</subject><subject>High resolution</subject><subject>Methods</subject><subject>Neural networks</subject><subject>Real time</subject><subject>real-time processing</subject><issn>1424-8220</issn><issn>1424-8220</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNpdkc1u1DAUhS1ERcvAghdAkdiUhcF_cWwWSGhaYKRKSMx0w8ZynJtpRkk8tZ1KvD2epozarmyf8-n46F6E3lHyiXNNPkfGSEk1FS_QGRVMYJWFl4_up-h1jDtCGOdcvUKnvGQVJZKeoT8XmzW-gH26-VL8BtvjTTdAse7GbQ94NdgtFPducRlTN9jU-bG4jtmeZZw8Xu-tg2Jml36MKUzuwL1BJ63tI7x9OBfo-vvlZvkTX_36sVp-u8JOSJ0wcFeVgtVOKSG007bVUjDdMkeUo1zVLclvy6ECJ5paSgKalVaIknLeQskXaDXnNt7uzD7kmuGv8bYz94IPW2ND6lwPBqS1DddMKkEFaZuaCVITcHX-lZStzVlf56z9VA_QOBhTsP2T0KfO2N2Yrb8zKk9fy0OZ84eA4G8niMkMXXTQ93YEP0XDJFcVZZrLjH54hu78FMY8qgNVVaqSmVugjzPlgo8xQHssQ4k5bN8ct5_Z94_bH8n_6-b_AEFfqJ8</recordid><startdate>20220301</startdate><enddate>20220301</enddate><creator>Ibrahem, Hatem</creator><creator>Salem, Ahmed</creator><creator>Kang, Hyun-Soo</creator><general>MDPI AG</general><general>MDPI</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7X7</scope><scope>7XB</scope><scope>88E</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>K9.</scope><scope>M0S</scope><scope>M1P</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>7X8</scope><scope>5PM</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-8722-3300</orcidid><orcidid>https://orcid.org/0000-0002-4682-0368</orcidid><orcidid>https://orcid.org/0000-0002-4333-2852</orcidid></search><sort><creationdate>20220301</creationdate><title>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</title><author>Ibrahem, Hatem ; Salem, Ahmed ; Kang, Hyun-Soo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>3-D graphics</topic><topic>Accuracy</topic><topic>Algorithms</topic><topic>Autonomous vehicles</topic><topic>Computer architecture</topic><topic>Construction</topic><topic>convolutional neural networks</topic><topic>depth estimation</topic><topic>Embedded systems</topic><topic>Experiments</topic><topic>Frames per second</topic><topic>High resolution</topic><topic>Methods</topic><topic>Neural networks</topic><topic>Real time</topic><topic>real-time processing</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ibrahem, Hatem</creatorcontrib><creatorcontrib>Salem, Ahmed</creatorcontrib><creatorcontrib>Kang, Hyun-Soo</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Health &amp; Medical Collection</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Medical Database (Alumni Edition)</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><collection>Health &amp; Medical Collection (Alumni Edition)</collection><collection>Medical Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Sensors (Basel, Switzerland)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ibrahem, Hatem</au><au>Salem, Ahmed</au><au>Kang, Hyun-Soo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</atitle><jtitle>Sensors (Basel, Switzerland)</jtitle><addtitle>Sensors (Basel)</addtitle><date>2022-03-01</date><risdate>2022</risdate><volume>22</volume><issue>5</issue><spage>1914</spage><pages>1914-</pages><issn>1424-8220</issn><eissn>1424-8220</eissn><abstract>As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time.</abstract><cop>Switzerland</cop><pub>MDPI AG</pub><pmid>35271061</pmid><doi>10.3390/s22051914</doi><orcidid>https://orcid.org/0000-0001-8722-3300</orcidid><orcidid>https://orcid.org/0000-0002-4682-0368</orcidid><orcidid>https://orcid.org/0000-0002-4333-2852</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1424-8220
ispartof Sensors (Basel, Switzerland), 2022-03, Vol.22 (5), p.1914
issn 1424-8220
1424-8220
language eng
recordid cdi_doaj_primary_oai_doaj_org_article_e6aad392684140fdb240b0ecb9af05fa
source Publicly Available Content Database; PubMed Central
subjects 3-D graphics
Accuracy
Algorithms
Autonomous vehicles
Computer architecture
Construction
convolutional neural networks
depth estimation
Embedded systems
Experiments
Frames per second
High resolution
Methods
Neural networks
Real time
real-time processing
title DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-27T18%3A29%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DTS-Depth:%20Real-Time%20Single-Image%20Depth%20Estimation%20Using%20Depth-to-Space%20Image%20Construction&rft.jtitle=Sensors%20(Basel,%20Switzerland)&rft.au=Ibrahem,%20Hatem&rft.date=2022-03-01&rft.volume=22&rft.issue=5&rft.spage=1914&rft.pages=1914-&rft.issn=1424-8220&rft.eissn=1424-8220&rft_id=info:doi/10.3390/s22051914&rft_dat=%3Cproquest_doaj_%3E2638712936%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2637787629&rft_id=info:pmid/35271061&rfr_iscdi=true