Loading…

Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function

This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and...

Full description

Saved in:
Bibliographic Details
Published in:IEEE access 2023, Vol.11, p.47681-47689
Main Authors: Simhon, Or, Karni, Zohar, Berman, Sigal, Zarrouk, David
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023
cites cdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023
container_end_page 47689
container_issue
container_start_page 47681
container_title IEEE access
container_volume 11
creator Simhon, Or
Karni, Zohar
Berman, Sigal
Zarrouk, David
description This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY ^{\mathrm {TM}} ). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).
doi_str_mv 10.1109/ACCESS.2023.3274675
format article
fullrecord <record><control><sourceid>proquest_doaj_</sourceid><recordid>TN_cdi_proquest_journals_2815684432</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10121761</ieee_id><doaj_id>oai_doaj_org_article_ae6a3d702797448a81a50111389812b3</doaj_id><sourcerecordid>2815684432</sourcerecordid><originalsourceid>FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</originalsourceid><addsrcrecordid>eNpNkU9vEzEQxVcIJKrSTwAHS5w3-N_au8cSUqgUFKml6tGatWdTh8QO9qaodz44XrZC9cXWm3m_seZV1XtGF4zR7tPlcrm6vV1wysVCcC2Vbl5VZ5yprhaNUK9fvN9WFznvaDltkRp9Vv3ZPGKy8eDDlmz6PILdYyb3fnwgQG7QxjD47SlBv0dyE_s4krs89X5BPJa6D0NMFg8YRrJGSGGqfYaMjsRQCN_RPkDwFvbkPqaf9Spg2j4V429Ijlydgh19DO-qNwPsM1483-fV3dXqx_Jbvd58vV5ermsraTfWQzso2XHb2AE0U041gE47qVrOLKqiSU6ptZ1zjex7xqW1ysHQaUE1K-s5r65nrouwM8fkD5CeTARv_gkxbQ2k0ZcVGCg84TTlutNSttAyaChjTLRdy3gvCuvjzDqm-OuEeTS7eEqhfN_wljWqlVJME8XcZVPMOeHwfyqjZkrPzOmZKT3znF5xfZhdHhFfOBhnWjHxF2q-leA</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2815684432</pqid></control><display><type>article</type><title>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</title><source>IEEE Xplore Open Access Journals</source><creator>Simhon, Or ; Karni, Zohar ; Berman, Sigal ; Zarrouk, David</creator><creatorcontrib>Simhon, Or ; Karni, Zohar ; Berman, Sigal ; Zarrouk, David</creatorcontrib><description>This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY&lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;^{\mathrm {TM}} &lt;/tex-math&gt;&lt;/inline-formula&gt;). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2023.3274675</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Barriers ; Deep learning ; Energy ; Energy requirements ; Friction ; Legged locomotion ; Obstacle negotiation ; Reconfigurable devices ; reconfigurable robot ; Reconfiguration ; Reinforcement learning ; reinforcement learning (RL) ; reward shaping ; Robots ; Torque ; Wheels</subject><ispartof>IEEE access, 2023, Vol.11, p.47681-47689</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</citedby><cites>FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</cites><orcidid>0000-0003-3106-1861 ; 0000-0001-7717-7259</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10121761$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,4024,27633,27923,27924,27925,54933</link.rule.ids></links><search><creatorcontrib>Simhon, Or</creatorcontrib><creatorcontrib>Karni, Zohar</creatorcontrib><creatorcontrib>Berman, Sigal</creatorcontrib><creatorcontrib>Zarrouk, David</creatorcontrib><title>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</title><title>IEEE access</title><addtitle>Access</addtitle><description>This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY&lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;^{\mathrm {TM}} &lt;/tex-math&gt;&lt;/inline-formula&gt;). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).</description><subject>Barriers</subject><subject>Deep learning</subject><subject>Energy</subject><subject>Energy requirements</subject><subject>Friction</subject><subject>Legged locomotion</subject><subject>Obstacle negotiation</subject><subject>Reconfigurable devices</subject><subject>reconfigurable robot</subject><subject>Reconfiguration</subject><subject>Reinforcement learning</subject><subject>reinforcement learning (RL)</subject><subject>reward shaping</subject><subject>Robots</subject><subject>Torque</subject><subject>Wheels</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>DOA</sourceid><recordid>eNpNkU9vEzEQxVcIJKrSTwAHS5w3-N_au8cSUqgUFKml6tGatWdTh8QO9qaodz44XrZC9cXWm3m_seZV1XtGF4zR7tPlcrm6vV1wysVCcC2Vbl5VZ5yprhaNUK9fvN9WFznvaDltkRp9Vv3ZPGKy8eDDlmz6PILdYyb3fnwgQG7QxjD47SlBv0dyE_s4krs89X5BPJa6D0NMFg8YRrJGSGGqfYaMjsRQCN_RPkDwFvbkPqaf9Spg2j4V429Ijlydgh19DO-qNwPsM1483-fV3dXqx_Jbvd58vV5ermsraTfWQzso2XHb2AE0U041gE47qVrOLKqiSU6ptZ1zjex7xqW1ysHQaUE1K-s5r65nrouwM8fkD5CeTARv_gkxbQ2k0ZcVGCg84TTlutNSttAyaChjTLRdy3gvCuvjzDqm-OuEeTS7eEqhfN_wljWqlVJME8XcZVPMOeHwfyqjZkrPzOmZKT3znF5xfZhdHhFfOBhnWjHxF2q-leA</recordid><startdate>2023</startdate><enddate>2023</enddate><creator>Simhon, Or</creator><creator>Karni, Zohar</creator><creator>Berman, Sigal</creator><creator>Zarrouk, David</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0003-3106-1861</orcidid><orcidid>https://orcid.org/0000-0001-7717-7259</orcidid></search><sort><creationdate>2023</creationdate><title>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</title><author>Simhon, Or ; Karni, Zohar ; Berman, Sigal ; Zarrouk, David</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Barriers</topic><topic>Deep learning</topic><topic>Energy</topic><topic>Energy requirements</topic><topic>Friction</topic><topic>Legged locomotion</topic><topic>Obstacle negotiation</topic><topic>Reconfigurable devices</topic><topic>reconfigurable robot</topic><topic>Reconfiguration</topic><topic>Reinforcement learning</topic><topic>reinforcement learning (RL)</topic><topic>reward shaping</topic><topic>Robots</topic><topic>Torque</topic><topic>Wheels</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Simhon, Or</creatorcontrib><creatorcontrib>Karni, Zohar</creatorcontrib><creatorcontrib>Berman, Sigal</creatorcontrib><creatorcontrib>Zarrouk, David</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) Online</collection><collection>IEL</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Simhon, Or</au><au>Karni, Zohar</au><au>Berman, Sigal</au><au>Zarrouk, David</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2023</date><risdate>2023</risdate><volume>11</volume><spage>47681</spage><epage>47689</epage><pages>47681-47689</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY&lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;^{\mathrm {TM}} &lt;/tex-math&gt;&lt;/inline-formula&gt;). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2023.3274675</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0003-3106-1861</orcidid><orcidid>https://orcid.org/0000-0001-7717-7259</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2023, Vol.11, p.47681-47689
issn 2169-3536
2169-3536
language eng
recordid cdi_proquest_journals_2815684432
source IEEE Xplore Open Access Journals
subjects Barriers
Deep learning
Energy
Energy requirements
Friction
Legged locomotion
Obstacle negotiation
Reconfigurable devices
reconfigurable robot
Reconfiguration
Reinforcement learning
reinforcement learning (RL)
reward shaping
Robots
Torque
Wheels
title Overcoming Obstacles With a Reconfigurable Robot Using Deep Reinforcement Learning Based on a Mechanical Work-Energy Reward Function
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T14%3A20%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_doaj_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Overcoming%20Obstacles%20With%20a%20Reconfigurable%20Robot%20Using%20Deep%20Reinforcement%20Learning%20Based%20on%20a%20Mechanical%20Work-Energy%20Reward%20Function&rft.jtitle=IEEE%20access&rft.au=Simhon,%20Or&rft.date=2023&rft.volume=11&rft.spage=47681&rft.epage=47689&rft.pages=47681-47689&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2023.3274675&rft_dat=%3Cproquest_doaj_%3E2815684432%3C/proquest_doaj_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c409t-f8f6492c5cfa716d65aed7d46821ce6a714200cc9dd54bb124cc6daf973071023%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2815684432&rft_id=info:pmid/&rft_ieee_id=10121761&rfr_iscdi=true