Loading…
DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction
As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolu...
Saved in:
Published in: | Sensors (Basel, Switzerland) Switzerland), 2022-03, Vol.22 (5), p.1914 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53 |
---|---|
cites | cdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53 |
container_end_page | |
container_issue | 5 |
container_start_page | 1914 |
container_title | Sensors (Basel, Switzerland) |
container_volume | 22 |
creator | Ibrahem, Hatem Salem, Ahmed Kang, Hyun-Soo |
description | As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time. |
doi_str_mv | 10.3390/s22051914 |
format | article |
fullrecord | <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_doaj_primary_oai_doaj_org_article_e6aad392684140fdb240b0ecb9af05fa</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><doaj_id>oai_doaj_org_article_e6aad392684140fdb240b0ecb9af05fa</doaj_id><sourcerecordid>2638712936</sourcerecordid><originalsourceid>FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</originalsourceid><addsrcrecordid>eNpdkc1u1DAUhS1ERcvAghdAkdiUhcF_cWwWSGhaYKRKSMx0w8ZynJtpRkk8tZ1KvD2epozarmyf8-n46F6E3lHyiXNNPkfGSEk1FS_QGRVMYJWFl4_up-h1jDtCGOdcvUKnvGQVJZKeoT8XmzW-gH26-VL8BtvjTTdAse7GbQ94NdgtFPducRlTN9jU-bG4jtmeZZw8Xu-tg2Jml36MKUzuwL1BJ63tI7x9OBfo-vvlZvkTX_36sVp-u8JOSJ0wcFeVgtVOKSG007bVUjDdMkeUo1zVLclvy6ECJ5paSgKalVaIknLeQskXaDXnNt7uzD7kmuGv8bYz94IPW2ND6lwPBqS1DddMKkEFaZuaCVITcHX-lZStzVlf56z9VA_QOBhTsP2T0KfO2N2Yrb8zKk9fy0OZ84eA4G8niMkMXXTQ93YEP0XDJFcVZZrLjH54hu78FMY8qgNVVaqSmVugjzPlgo8xQHssQ4k5bN8ct5_Z94_bH8n_6-b_AEFfqJ8</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2637787629</pqid></control><display><type>article</type><title>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</title><source>Publicly Available Content Database</source><source>PubMed Central</source><creator>Ibrahem, Hatem ; Salem, Ahmed ; Kang, Hyun-Soo</creator><creatorcontrib>Ibrahem, Hatem ; Salem, Ahmed ; Kang, Hyun-Soo</creatorcontrib><description>As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time.</description><identifier>ISSN: 1424-8220</identifier><identifier>EISSN: 1424-8220</identifier><identifier>DOI: 10.3390/s22051914</identifier><identifier>PMID: 35271061</identifier><language>eng</language><publisher>Switzerland: MDPI AG</publisher><subject>3-D graphics ; Accuracy ; Algorithms ; Autonomous vehicles ; Computer architecture ; Construction ; convolutional neural networks ; depth estimation ; Embedded systems ; Experiments ; Frames per second ; High resolution ; Methods ; Neural networks ; Real time ; real-time processing</subject><ispartof>Sensors (Basel, Switzerland), 2022-03, Vol.22 (5), p.1914</ispartof><rights>2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>2022 by the authors. 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</citedby><cites>FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</cites><orcidid>0000-0001-8722-3300 ; 0000-0002-4682-0368 ; 0000-0002-4333-2852</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2637787629/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2637787629?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>230,314,723,776,780,881,25733,27903,27904,36991,36992,44569,53769,53771,74872</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/35271061$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Ibrahem, Hatem</creatorcontrib><creatorcontrib>Salem, Ahmed</creatorcontrib><creatorcontrib>Kang, Hyun-Soo</creatorcontrib><title>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</title><title>Sensors (Basel, Switzerland)</title><addtitle>Sensors (Basel)</addtitle><description>As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time.</description><subject>3-D graphics</subject><subject>Accuracy</subject><subject>Algorithms</subject><subject>Autonomous vehicles</subject><subject>Computer architecture</subject><subject>Construction</subject><subject>convolutional neural networks</subject><subject>depth estimation</subject><subject>Embedded systems</subject><subject>Experiments</subject><subject>Frames per second</subject><subject>High resolution</subject><subject>Methods</subject><subject>Neural networks</subject><subject>Real time</subject><subject>real-time processing</subject><issn>1424-8220</issn><issn>1424-8220</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><sourceid>DOA</sourceid><recordid>eNpdkc1u1DAUhS1ERcvAghdAkdiUhcF_cWwWSGhaYKRKSMx0w8ZynJtpRkk8tZ1KvD2epozarmyf8-n46F6E3lHyiXNNPkfGSEk1FS_QGRVMYJWFl4_up-h1jDtCGOdcvUKnvGQVJZKeoT8XmzW-gH26-VL8BtvjTTdAse7GbQ94NdgtFPducRlTN9jU-bG4jtmeZZw8Xu-tg2Jml36MKUzuwL1BJ63tI7x9OBfo-vvlZvkTX_36sVp-u8JOSJ0wcFeVgtVOKSG007bVUjDdMkeUo1zVLclvy6ECJ5paSgKalVaIknLeQskXaDXnNt7uzD7kmuGv8bYz94IPW2ND6lwPBqS1DddMKkEFaZuaCVITcHX-lZStzVlf56z9VA_QOBhTsP2T0KfO2N2Yrb8zKk9fy0OZ84eA4G8niMkMXXTQ93YEP0XDJFcVZZrLjH54hu78FMY8qgNVVaqSmVugjzPlgo8xQHssQ4k5bN8ct5_Z94_bH8n_6-b_AEFfqJ8</recordid><startdate>20220301</startdate><enddate>20220301</enddate><creator>Ibrahem, Hatem</creator><creator>Salem, Ahmed</creator><creator>Kang, Hyun-Soo</creator><general>MDPI AG</general><general>MDPI</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7X7</scope><scope>7XB</scope><scope>88E</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>K9.</scope><scope>M0S</scope><scope>M1P</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>7X8</scope><scope>5PM</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-8722-3300</orcidid><orcidid>https://orcid.org/0000-0002-4682-0368</orcidid><orcidid>https://orcid.org/0000-0002-4333-2852</orcidid></search><sort><creationdate>20220301</creationdate><title>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</title><author>Ibrahem, Hatem ; Salem, Ahmed ; Kang, Hyun-Soo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>3-D graphics</topic><topic>Accuracy</topic><topic>Algorithms</topic><topic>Autonomous vehicles</topic><topic>Computer architecture</topic><topic>Construction</topic><topic>convolutional neural networks</topic><topic>depth estimation</topic><topic>Embedded systems</topic><topic>Experiments</topic><topic>Frames per second</topic><topic>High resolution</topic><topic>Methods</topic><topic>Neural networks</topic><topic>Real time</topic><topic>real-time processing</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ibrahem, Hatem</creatorcontrib><creatorcontrib>Salem, Ahmed</creatorcontrib><creatorcontrib>Kang, Hyun-Soo</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Health & Medical Collection</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Medical Database (Alumni Edition)</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Health & Medical Complete (Alumni)</collection><collection>Health & Medical Collection (Alumni Edition)</collection><collection>Medical Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>Sensors (Basel, Switzerland)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ibrahem, Hatem</au><au>Salem, Ahmed</au><au>Kang, Hyun-Soo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction</atitle><jtitle>Sensors (Basel, Switzerland)</jtitle><addtitle>Sensors (Basel)</addtitle><date>2022-03-01</date><risdate>2022</risdate><volume>22</volume><issue>5</issue><spage>1914</spage><pages>1914-</pages><issn>1424-8220</issn><eissn>1424-8220</eissn><abstract>As most of the recent high-resolution depth-estimation algorithms are computationally so expensive that they cannot work in real time, the common solution is using a low-resolution input image to reduce the computational complexity. We propose a different approach, an efficient and real-time convolutional neural network-based depth-estimation algorithm using a single high-resolution image as the input. The proposed method efficiently constructs a high-resolution depth map using a small encoding architecture and eliminates the need for a decoder, which is typically used in the encoder-decoder architectures employed for depth estimation. The proposed algorithm adopts a modified MobileNetV2 architecture, which is a lightweight architecture, to estimate the depth information through the depth-to-space image construction, which is generally employed in image super-resolution. As a result, it realizes fast frame processing and can predict a high-accuracy depth in real time. We train and test our method on the challenging KITTI, Cityscapes, and NYUV2 depth datasets. The proposed method achieves low relative absolute error (0.028 for KITTI, 0.167 for CITYSCAPES, and 0.069 for NYUV2) while working at speed reaching 48 frames per second on a GPU and 20 frames per second on a CPU for high-resolution test images. We compare our method with the state-of-the-art methods on depth estimation, showing that our method outperforms those methods. However, the architecture is less complex and works in real time.</abstract><cop>Switzerland</cop><pub>MDPI AG</pub><pmid>35271061</pmid><doi>10.3390/s22051914</doi><orcidid>https://orcid.org/0000-0001-8722-3300</orcidid><orcidid>https://orcid.org/0000-0002-4682-0368</orcidid><orcidid>https://orcid.org/0000-0002-4333-2852</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1424-8220 |
ispartof | Sensors (Basel, Switzerland), 2022-03, Vol.22 (5), p.1914 |
issn | 1424-8220 1424-8220 |
language | eng |
recordid | cdi_doaj_primary_oai_doaj_org_article_e6aad392684140fdb240b0ecb9af05fa |
source | Publicly Available Content Database; PubMed Central |
subjects | 3-D graphics Accuracy Algorithms Autonomous vehicles Computer architecture Construction convolutional neural networks depth estimation Embedded systems Experiments Frames per second High resolution Methods Neural networks Real time real-time processing |
title | DTS-Depth: Real-Time Single-Image Depth Estimation Using Depth-to-Space Image Construction |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-27T18%3A29%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DTS-Depth:%20Real-Time%20Single-Image%20Depth%20Estimation%20Using%20Depth-to-Space%20Image%20Construction&rft.jtitle=Sensors%20(Basel,%20Switzerland)&rft.au=Ibrahem,%20Hatem&rft.date=2022-03-01&rft.volume=22&rft.issue=5&rft.spage=1914&rft.pages=1914-&rft.issn=1424-8220&rft.eissn=1424-8220&rft_id=info:doi/10.3390/s22051914&rft_dat=%3Cproquest_doaj_%3E2638712936%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c469t-e3c7542bc88449c9af96429f2c08c138bf0642a3e7ec4db660e925a445133fe53%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2637787629&rft_id=info:pmid/35271061&rfr_iscdi=true |