Loading…

Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty

Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measur...

Full description

Saved in:
Bibliographic Details
Main Authors: Joshi, Bhaskar, Kapur, Dhruv, Kandath, Harikumar
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 284
container_issue
container_start_page 278
container_title
container_volume
creator Joshi, Bhaskar
Kapur, Dhruv
Kandath, Harikumar
description Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment.
doi_str_mv 10.1109/ICARA60736.2024.10553074
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_10553074</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10553074</ieee_id><sourcerecordid>10553074</sourcerecordid><originalsourceid>FETCH-LOGICAL-i119t-b92c17d75f771af31b0be7734afed90922c6ec296ad68715f13ecc821c3973973</originalsourceid><addsrcrecordid>eNo1UFtLwzAYjYLgmPsHPuQPtObS5Fse67xMqAyq9XWk6dcR2dKRVGH_3soUDpyHc4FzCKGc5Zwzc_eyKutSM5A6F0wUOWdKSQbFBVkYMEupmDSFKNQlmQnQkAEU6posUvpkjEkxqVrNyO7NH7JxyGq0e_qAeKQ1-tAP0eEBw0grtDH4sKP3NmFHN20ardsjLb8H39ngkE5e2pQfiTahw0hf0aaveA43kx5H68N4uiFXvd0nXPzxnDRPj--rdVZtnqclVeY5N2PWGuE4dKB6AG57yVvWIoAsbI-dYUYIp9EJo22nl8BVzyU6txTcSQO_mJPbc69HxO0x-oONp-3_N_IHfDFZXw</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</title><source>IEEE Xplore All Conference Series</source><creator>Joshi, Bhaskar ; Kapur, Dhruv ; Kandath, Harikumar</creator><creatorcontrib>Joshi, Bhaskar ; Kapur, Dhruv ; Kandath, Harikumar</creatorcontrib><description>Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment.</description><identifier>EISSN: 2767-7745</identifier><identifier>EISBN: 9798350394245</identifier><identifier>DOI: 10.1109/ICARA60736.2024.10553074</identifier><language>eng</language><publisher>IEEE</publisher><subject>Autonomous aerial vehicles ; Autonomous navigation ; Deep reinforcement learning ; measurement noise ; Measurement uncertainty ; Navigation ; Noise ; obstacle avoidance ; proximal policy optimization ; Training ; Trajectory ; unmanned aerial vehicle</subject><ispartof>2024 10th International Conference on Automation, Robotics and Applications (ICARA), 2024, p.278-284</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10553074$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,27902,54530,54907</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10553074$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Joshi, Bhaskar</creatorcontrib><creatorcontrib>Kapur, Dhruv</creatorcontrib><creatorcontrib>Kandath, Harikumar</creatorcontrib><title>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</title><title>2024 10th International Conference on Automation, Robotics and Applications (ICARA)</title><addtitle>ICARA</addtitle><description>Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment.</description><subject>Autonomous aerial vehicles</subject><subject>Autonomous navigation</subject><subject>Deep reinforcement learning</subject><subject>measurement noise</subject><subject>Measurement uncertainty</subject><subject>Navigation</subject><subject>Noise</subject><subject>obstacle avoidance</subject><subject>proximal policy optimization</subject><subject>Training</subject><subject>Trajectory</subject><subject>unmanned aerial vehicle</subject><issn>2767-7745</issn><isbn>9798350394245</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2024</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNo1UFtLwzAYjYLgmPsHPuQPtObS5Fse67xMqAyq9XWk6dcR2dKRVGH_3soUDpyHc4FzCKGc5Zwzc_eyKutSM5A6F0wUOWdKSQbFBVkYMEupmDSFKNQlmQnQkAEU6posUvpkjEkxqVrNyO7NH7JxyGq0e_qAeKQ1-tAP0eEBw0grtDH4sKP3NmFHN20ardsjLb8H39ngkE5e2pQfiTahw0hf0aaveA43kx5H68N4uiFXvd0nXPzxnDRPj--rdVZtnqclVeY5N2PWGuE4dKB6AG57yVvWIoAsbI-dYUYIp9EJo22nl8BVzyU6txTcSQO_mJPbc69HxO0x-oONp-3_N_IHfDFZXw</recordid><startdate>20240222</startdate><enddate>20240222</enddate><creator>Joshi, Bhaskar</creator><creator>Kapur, Dhruv</creator><creator>Kandath, Harikumar</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>20240222</creationdate><title>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</title><author>Joshi, Bhaskar ; Kapur, Dhruv ; Kandath, Harikumar</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i119t-b92c17d75f771af31b0be7734afed90922c6ec296ad68715f13ecc821c3973973</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Autonomous aerial vehicles</topic><topic>Autonomous navigation</topic><topic>Deep reinforcement learning</topic><topic>measurement noise</topic><topic>Measurement uncertainty</topic><topic>Navigation</topic><topic>Noise</topic><topic>obstacle avoidance</topic><topic>proximal policy optimization</topic><topic>Training</topic><topic>Trajectory</topic><topic>unmanned aerial vehicle</topic><toplevel>online_resources</toplevel><creatorcontrib>Joshi, Bhaskar</creatorcontrib><creatorcontrib>Kapur, Dhruv</creatorcontrib><creatorcontrib>Kandath, Harikumar</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Joshi, Bhaskar</au><au>Kapur, Dhruv</au><au>Kandath, Harikumar</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</atitle><btitle>2024 10th International Conference on Automation, Robotics and Applications (ICARA)</btitle><stitle>ICARA</stitle><date>2024-02-22</date><risdate>2024</risdate><spage>278</spage><epage>284</epage><pages>278-284</pages><eissn>2767-7745</eissn><eisbn>9798350394245</eisbn><abstract>Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment.</abstract><pub>IEEE</pub><doi>10.1109/ICARA60736.2024.10553074</doi><tpages>7</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2767-7745
ispartof 2024 10th International Conference on Automation, Robotics and Applications (ICARA), 2024, p.278-284
issn 2767-7745
language eng
recordid cdi_ieee_primary_10553074
source IEEE Xplore All Conference Series
subjects Autonomous aerial vehicles
Autonomous navigation
Deep reinforcement learning
measurement noise
Measurement uncertainty
Navigation
Noise
obstacle avoidance
proximal policy optimization
Training
Trajectory
unmanned aerial vehicle
title Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-12T17%3A45%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Sim-to-Real%20Deep%20Reinforcement%20Learning%20Based%20Obstacle%20Avoidance%20for%20UAVs%20Under%20Measurement%20Uncertainty&rft.btitle=2024%2010th%20International%20Conference%20on%20Automation,%20Robotics%20and%20Applications%20(ICARA)&rft.au=Joshi,%20Bhaskar&rft.date=2024-02-22&rft.spage=278&rft.epage=284&rft.pages=278-284&rft.eissn=2767-7745&rft_id=info:doi/10.1109/ICARA60736.2024.10553074&rft.eisbn=9798350394245&rft_dat=%3Cieee_CHZPO%3E10553074%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i119t-b92c17d75f771af31b0be7734afed90922c6ec296ad68715f13ecc821c3973973%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10553074&rfr_iscdi=true