Loading…
Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function
This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and...
Saved in:
Published in: | IEEE access 2023, Vol.11, p.47681-47689 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023 |
---|---|
cites | cdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023 |
container_end_page | 47689 |
container_issue | |
container_start_page | 47681 |
container_title | IEEE access |
container_volume | 11 |
creator | Simhon, Or Karni, Zohar Berman, Sigal Zarrouk, David |
description | This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY ^{\mathrm {TM}} ). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video). |
doi_str_mv | 10.1109/ACCESS.2023.3274675 |
format | article |
fullrecord | <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_proquest_journals_2815684432</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10121761</ieee_id><doaj_id>oai_doaj_org_article_ae6a3d702797448a81a50111389812b3</doaj_id><sourcerecordid>2815684432</sourcerecordid><originalsourceid>FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</originalsourceid><addsrcrecordid>eNpNkU9vEzEQxVcIJKrSTwAHS5w3-N_au8cSUqgUFKml6tGatWdTh8QO9qaodz44XrZC9cXWm3m_seZV1XtGF4zR7tPlcrm6vV1wysVCcC2Vbl5VZ5yprhaNUK9fvN9WFznvaDltkRp9Vv3ZPGKy8eDDlmz6PILdYyb3fnwgQG7QxjD47SlBv0dyE_s4krs89X5BPJa6D0NMFg8YRrJGSGGqfYaMjsRQCN_RPkDwFvbkPqaf9Spg2j4V429Ijlydgh19DO-qNwPsM1483-fV3dXqx_Jbvd58vV5ermsraTfWQzso2XHb2AE0U041gE47qVrOLKqiSU6ptZ1zjex7xqW1ysHQaUE1K-s5r65nrouwM8fkD5CeTARv_gkxbQ2k0ZcVGCg84TTlutNSttAyaChjTLRdy3gvCuvjzDqm-OuEeTS7eEqhfN_wljWqlVJME8XcZVPMOeHwfyqjZkrPzOmZKT3znF5xfZhdHhFfOBhnWjHxF2q-leA</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2815684432</pqid></control><display><type>article</type><title>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</title><source>IEEE Xplore Open Access Journals</source><creator>Simhon, Or ; Karni, Zohar ; Berman, Sigal ; Zarrouk, David</creator><creatorcontrib>Simhon, Or ; Karni, Zohar ; Berman, Sigal ; Zarrouk, David</creatorcontrib><description>This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY<inline-formula> <tex-math notation="LaTeX">^{\mathrm {TM}} </tex-math></inline-formula>). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2023.3274675</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Barriers ; Deep learning ; Energy ; Energy requirements ; Friction ; Legged locomotion ; Obstacle negotiation ; Reconfigurable devices ; reconfigurable robot ; Reconfiguration ; Reinforcement learning ; reinforcement learning (RL) ; reward shaping ; Robots ; Torque ; Wheels</subject><ispartof>IEEE access, 2023, Vol.11, p.47681-47689</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</citedby><cites>FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</cites><orcidid>0000-0003-3106-1861 ; 0000-0001-7717-7259</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10121761$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,4024,27633,27923,27924,27925,54933</link.rule.ids></links><search><creatorcontrib>Simhon, Or</creatorcontrib><creatorcontrib>Karni, Zohar</creatorcontrib><creatorcontrib>Berman, Sigal</creatorcontrib><creatorcontrib>Zarrouk, David</creatorcontrib><title>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</title><title>IEEE access</title><addtitle>Access</addtitle><description>This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY<inline-formula> <tex-math notation="LaTeX">^{\mathrm {TM}} </tex-math></inline-formula>). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).</description><subject>Barriers</subject><subject>Deep learning</subject><subject>Energy</subject><subject>Energy requirements</subject><subject>Friction</subject><subject>Legged locomotion</subject><subject>Obstacle negotiation</subject><subject>Reconfigurable devices</subject><subject>reconfigurable robot</subject><subject>Reconfiguration</subject><subject>Reinforcement learning</subject><subject>reinforcement learning (RL)</subject><subject>reward shaping</subject><subject>Robots</subject><subject>Torque</subject><subject>Wheels</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>DOA</sourceid><recordid>eNpNkU9vEzEQxVcIJKrSTwAHS5w3-N_au8cSUqgUFKml6tGatWdTh8QO9qaodz44XrZC9cXWm3m_seZV1XtGF4zR7tPlcrm6vV1wysVCcC2Vbl5VZ5yprhaNUK9fvN9WFznvaDltkRp9Vv3ZPGKy8eDDlmz6PILdYyb3fnwgQG7QxjD47SlBv0dyE_s4krs89X5BPJa6D0NMFg8YRrJGSGGqfYaMjsRQCN_RPkDwFvbkPqaf9Spg2j4V429Ijlydgh19DO-qNwPsM1483-fV3dXqx_Jbvd58vV5ermsraTfWQzso2XHb2AE0U041gE47qVrOLKqiSU6ptZ1zjex7xqW1ysHQaUE1K-s5r65nrouwM8fkD5CeTARv_gkxbQ2k0ZcVGCg84TTlutNSttAyaChjTLRdy3gvCuvjzDqm-OuEeTS7eEqhfN_wljWqlVJME8XcZVPMOeHwfyqjZkrPzOmZKT3znF5xfZhdHhFfOBhnWjHxF2q-leA</recordid><startdate>2023</startdate><enddate>2023</enddate><creator>Simhon, Or</creator><creator>Karni, Zohar</creator><creator>Berman, Sigal</creator><creator>Zarrouk, David</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0003-3106-1861</orcidid><orcidid>https://orcid.org/0000-0001-7717-7259</orcidid></search><sort><creationdate>2023</creationdate><title>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</title><author>Simhon, Or ; Karni, Zohar ; Berman, Sigal ; Zarrouk, David</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Barriers</topic><topic>Deep learning</topic><topic>Energy</topic><topic>Energy requirements</topic><topic>Friction</topic><topic>Legged locomotion</topic><topic>Obstacle negotiation</topic><topic>Reconfigurable devices</topic><topic>reconfigurable robot</topic><topic>Reconfiguration</topic><topic>Reinforcement learning</topic><topic>reinforcement learning (RL)</topic><topic>reward shaping</topic><topic>Robots</topic><topic>Torque</topic><topic>Wheels</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Simhon, Or</creatorcontrib><creatorcontrib>Karni, Zohar</creatorcontrib><creatorcontrib>Berman, Sigal</creatorcontrib><creatorcontrib>Zarrouk, David</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) Online</collection><collection>IEL</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Simhon, Or</au><au>Karni, Zohar</au><au>Berman, Sigal</au><au>Zarrouk, David</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2023</date><risdate>2023</risdate><volume>11</volume><spage>47681</spage><epage>47689</epage><pages>47681-47689</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY<inline-formula> <tex-math notation="LaTeX">^{\mathrm {TM}} </tex-math></inline-formula>). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2023.3274675</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0003-3106-1861</orcidid><orcidid>https://orcid.org/0000-0001-7717-7259</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2023, Vol.11, p.47681-47689 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_proquest_journals_2815684432 |
source | IEEE Xplore Open Access Journals |
subjects | Barriers Deep learning Energy Energy requirements Friction Legged locomotion Obstacle negotiation Reconfigurable devices reconfigurable robot Reconfiguration Reinforcement learning reinforcement learning (RL) reward shaping Robots Torque Wheels |
title | Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T14%3A20%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Overcoming%20Obstacles%20With%20a%20Reconfigurable%20Robot%20Using%20Deep%20Reinforcement%20Learning%20Based%20on%20a%20Mechanical%20Work-Energy%20Reward%20Function&rft.jtitle=IEEE%20access&rft.au=Simhon,%20Or&rft.date=2023&rft.volume=11&rft.spage=47681&rft.epage=47689&rft.pages=47681-47689&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2023.3274675&rft_dat=%3Cproquest_doaj_%3E2815684432%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2815684432&rft_id=info:pmid/&rft_ieee_id=10121761&rfr_iscdi=true |