Loading…
Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty
Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measur...
Saved in:
Main Authors: | , , |
---|---|
Format: | Conference Proceeding |
Language: | English |
Subjects: | |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | 284 |
container_issue | |
container_start_page | 278 |
container_title | |
container_volume | |
creator | Joshi, Bhaskar Kapur, Dhruv Kandath, Harikumar |
description | Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment. |
doi_str_mv | 10.1109/ICARA60736.2024.10553074 |
format | conference_proceeding |
fullrecord | <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_10553074</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10553074</ieee_id><sourcerecordid>10553074</sourcerecordid><originalsourceid>FETCH-LOGICAL-i119t-b92c17d75f771af31b0be7734afed90922c6ec296ad68715f13ecc821c3973973</originalsourceid><addsrcrecordid>eNo1UFtLwzAYjYLgmPsHPuQPtObS5Fse67xMqAyq9XWk6dcR2dKRVGH_3soUDpyHc4FzCKGc5Zwzc_eyKutSM5A6F0wUOWdKSQbFBVkYMEupmDSFKNQlmQnQkAEU6posUvpkjEkxqVrNyO7NH7JxyGq0e_qAeKQ1-tAP0eEBw0grtDH4sKP3NmFHN20ardsjLb8H39ngkE5e2pQfiTahw0hf0aaveA43kx5H68N4uiFXvd0nXPzxnDRPj--rdVZtnqclVeY5N2PWGuE4dKB6AG57yVvWIoAsbI-dYUYIp9EJo22nl8BVzyU6txTcSQO_mJPbc69HxO0x-oONp-3_N_IHfDFZXw</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</title><source>IEEE Xplore All Conference Series</source><creator>Joshi, Bhaskar ; Kapur, Dhruv ; Kandath, Harikumar</creator><creatorcontrib>Joshi, Bhaskar ; Kapur, Dhruv ; Kandath, Harikumar</creatorcontrib><description>Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment.</description><identifier>EISSN: 2767-7745</identifier><identifier>EISBN: 9798350394245</identifier><identifier>DOI: 10.1109/ICARA60736.2024.10553074</identifier><language>eng</language><publisher>IEEE</publisher><subject>Autonomous aerial vehicles ; Autonomous navigation ; Deep reinforcement learning ; measurement noise ; Measurement uncertainty ; Navigation ; Noise ; obstacle avoidance ; proximal policy optimization ; Training ; Trajectory ; unmanned aerial vehicle</subject><ispartof>2024 10th International Conference on Automation, Robotics and Applications (ICARA), 2024, p.278-284</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10553074$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,27902,54530,54907</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10553074$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Joshi, Bhaskar</creatorcontrib><creatorcontrib>Kapur, Dhruv</creatorcontrib><creatorcontrib>Kandath, Harikumar</creatorcontrib><title>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</title><title>2024 10th International Conference on Automation, Robotics and Applications (ICARA)</title><addtitle>ICARA</addtitle><description>Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment.</description><subject>Autonomous aerial vehicles</subject><subject>Autonomous navigation</subject><subject>Deep reinforcement learning</subject><subject>measurement noise</subject><subject>Measurement uncertainty</subject><subject>Navigation</subject><subject>Noise</subject><subject>obstacle avoidance</subject><subject>proximal policy optimization</subject><subject>Training</subject><subject>Trajectory</subject><subject>unmanned aerial vehicle</subject><issn>2767-7745</issn><isbn>9798350394245</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2024</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNo1UFtLwzAYjYLgmPsHPuQPtObS5Fse67xMqAyq9XWk6dcR2dKRVGH_3soUDpyHc4FzCKGc5Zwzc_eyKutSM5A6F0wUOWdKSQbFBVkYMEupmDSFKNQlmQnQkAEU6posUvpkjEkxqVrNyO7NH7JxyGq0e_qAeKQ1-tAP0eEBw0grtDH4sKP3NmFHN20ardsjLb8H39ngkE5e2pQfiTahw0hf0aaveA43kx5H68N4uiFXvd0nXPzxnDRPj--rdVZtnqclVeY5N2PWGuE4dKB6AG57yVvWIoAsbI-dYUYIp9EJo22nl8BVzyU6txTcSQO_mJPbc69HxO0x-oONp-3_N_IHfDFZXw</recordid><startdate>20240222</startdate><enddate>20240222</enddate><creator>Joshi, Bhaskar</creator><creator>Kapur, Dhruv</creator><creator>Kandath, Harikumar</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>20240222</creationdate><title>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</title><author>Joshi, Bhaskar ; Kapur, Dhruv ; Kandath, Harikumar</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i119t-b92c17d75f771af31b0be7734afed90922c6ec296ad68715f13ecc821c3973973</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Autonomous aerial vehicles</topic><topic>Autonomous navigation</topic><topic>Deep reinforcement learning</topic><topic>measurement noise</topic><topic>Measurement uncertainty</topic><topic>Navigation</topic><topic>Noise</topic><topic>obstacle avoidance</topic><topic>proximal policy optimization</topic><topic>Training</topic><topic>Trajectory</topic><topic>unmanned aerial vehicle</topic><toplevel>online_resources</toplevel><creatorcontrib>Joshi, Bhaskar</creatorcontrib><creatorcontrib>Kapur, Dhruv</creatorcontrib><creatorcontrib>Kandath, Harikumar</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Joshi, Bhaskar</au><au>Kapur, Dhruv</au><au>Kandath, Harikumar</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty</atitle><btitle>2024 10th International Conference on Automation, Robotics and Applications (ICARA)</btitle><stitle>ICARA</stitle><date>2024-02-22</date><risdate>2024</risdate><spage>278</spage><epage>284</epage><pages>278-284</pages><eissn>2767-7745</eissn><eisbn>9798350394245</eisbn><abstract>Deep Reinforcement Learning is quickly becoming a popular method for training autonomous Unmanned Aerial Vehicles (UAVs). Our work analyzes the effects of measurement uncertainty on the performance of Deep Reinforcement Learning (DRL) based waypoint navigation and obstacle avoidance for UAVs. Measurement uncertainty originates from noise in the sensors used for localization and detecting obstacles. Measurement uncertainty/noise is considered to follow a Gaussian probability distribution with unknown non-zero mean and variance. We evaluate the performance of a DRL agent, trained using the Proximal Policy Optimization (PPO) algorithm in an environment with continuous state and action spaces. The environment is randomized with different numbers of obstacles for each simulation episode in the presence of varying degrees of noise, to capture the effects of realistic sensor measurements. Denoising techniques like the low pass filter and Kalman filter improve performance in the presence of unbiased noise. Moreover, we show that artificially injecting noise into the measurements during evaluation actually improves performance in certain scenarios. Extensive training and testing of the DRL agent under various UAV navigation scenarios are performed in the PyBullet physics simulator. To evaluate the practical validity of our method, we port the policy trained in simulation onto a real UAV without any further modifications and verify the results in a real-world environment.</abstract><pub>IEEE</pub><doi>10.1109/ICARA60736.2024.10553074</doi><tpages>7</tpages></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | EISSN: 2767-7745 |
ispartof | 2024 10th International Conference on Automation, Robotics and Applications (ICARA), 2024, p.278-284 |
issn | 2767-7745 |
language | eng |
recordid | cdi_ieee_primary_10553074 |
source | IEEE Xplore All Conference Series |
subjects | Autonomous aerial vehicles Autonomous navigation Deep reinforcement learning measurement noise Measurement uncertainty Navigation Noise obstacle avoidance proximal policy optimization Training Trajectory unmanned aerial vehicle |
title | Sim-to-Real Deep Reinforcement Learning Based Obstacle Avoidance for UAVs Under Measurement Uncertainty |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-12T17%3A45%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Sim-to-Real%20Deep%20Reinforcement%20Learning%20Based%20Obstacle%20Avoidance%20for%20UAVs%20Under%20Measurement%20Uncertainty&rft.btitle=2024%2010th%20International%20Conference%20on%20Automation,%20Robotics%20and%20Applications%20(ICARA)&rft.au=Joshi,%20Bhaskar&rft.date=2024-02-22&rft.spage=278&rft.epage=284&rft.pages=278-284&rft.eissn=2767-7745&rft_id=info:doi/10.1109/ICARA60736.2024.10553074&rft.eisbn=9798350394245&rft_dat=%3Cieee_CHZPO%3E10553074%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i119t-b92c17d75f771af31b0be7734afed90922c6ec296ad68715f13ecc821c3973973%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10553074&rfr_iscdi=true |