Loading…

Symmetry Considerations for Learning Task Symmetric Robot Policies

Symmetry is a fundamental aspect of many real-world robotic tasks. However, current deep reinforcement learning (DRL) approaches can seldom harness and exploit symmetry effectively. Often, the learned behaviors fail to achieve the desired transformation invariances and suffer from motion artifacts....

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2024-03
Main Authors: Mittal, Mayank, Rudin, Nikita, Klemm, Victor, Allshire, Arthur, Hutter, Marco
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Mittal, Mayank
Rudin, Nikita
Klemm, Victor
Allshire, Arthur
Hutter, Marco
description Symmetry is a fundamental aspect of many real-world robotic tasks. However, current deep reinforcement learning (DRL) approaches can seldom harness and exploit symmetry effectively. Often, the learned behaviors fail to achieve the desired transformation invariances and suffer from motion artifacts. For instance, a quadruped may exhibit different gaits when commanded to move forward or backward, even though it is symmetrical about its torso. This issue becomes further pronounced in high-dimensional or complex environments, where DRL methods are prone to local optima and fail to explore regions of the state space equally. Past methods on encouraging symmetry for robotic tasks have studied this topic mainly in a single-task setting, where symmetry usually refers to symmetry in the motion, such as the gait patterns. In this paper, we revisit this topic for goal-conditioned tasks in robotics, where symmetry lies mainly in task execution and not necessarily in the learned motions themselves. In particular, we investigate two approaches to incorporate symmetry invariance into DRL -- data augmentation and mirror loss function. We provide a theoretical foundation for using augmented samples in an on-policy setting. Based on this, we show that the corresponding approach achieves faster convergence and improves the learned behaviors in various challenging robotic tasks, from climbing boxes with a quadruped to dexterous manipulation.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2952801167</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2952801167</sourcerecordid><originalsourceid>FETCH-proquest_journals_29528011673</originalsourceid><addsrcrecordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mRwCq7MzU0tKapUcM7PK85MSS1KLMkEshTS8osUfFITi_Iy89IVQhKLsxWgKjOTFYLyk_JLFALyczKTM1OLeRhY0xJzilN5oTQ3g7Kba4izh25BUX5haWpxSXxWfmlRHlAq3sjS1MjCwNDQzNyYOFUA83Y6LQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2952801167</pqid></control><display><type>article</type><title>Symmetry Considerations for Learning Task Symmetric Robot Policies</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><creator>Mittal, Mayank ; Rudin, Nikita ; Klemm, Victor ; Allshire, Arthur ; Hutter, Marco</creator><creatorcontrib>Mittal, Mayank ; Rudin, Nikita ; Klemm, Victor ; Allshire, Arthur ; Hutter, Marco</creatorcontrib><description>Symmetry is a fundamental aspect of many real-world robotic tasks. However, current deep reinforcement learning (DRL) approaches can seldom harness and exploit symmetry effectively. Often, the learned behaviors fail to achieve the desired transformation invariances and suffer from motion artifacts. For instance, a quadruped may exhibit different gaits when commanded to move forward or backward, even though it is symmetrical about its torso. This issue becomes further pronounced in high-dimensional or complex environments, where DRL methods are prone to local optima and fail to explore regions of the state space equally. Past methods on encouraging symmetry for robotic tasks have studied this topic mainly in a single-task setting, where symmetry usually refers to symmetry in the motion, such as the gait patterns. In this paper, we revisit this topic for goal-conditioned tasks in robotics, where symmetry lies mainly in task execution and not necessarily in the learned motions themselves. In particular, we investigate two approaches to incorporate symmetry invariance into DRL -- data augmentation and mirror loss function. We provide a theoretical foundation for using augmented samples in an on-policy setting. Based on this, we show that the corresponding approach achieves faster convergence and improves the learned behaviors in various challenging robotic tasks, from climbing boxes with a quadruped to dexterous manipulation.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Cognitive tasks ; Data augmentation ; Deep learning ; Gait ; Robot learning ; Robotics ; Symmetry ; Torso</subject><ispartof>arXiv.org, 2024-03</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2952801167?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Mittal, Mayank</creatorcontrib><creatorcontrib>Rudin, Nikita</creatorcontrib><creatorcontrib>Klemm, Victor</creatorcontrib><creatorcontrib>Allshire, Arthur</creatorcontrib><creatorcontrib>Hutter, Marco</creatorcontrib><title>Symmetry Considerations for Learning Task Symmetric Robot Policies</title><title>arXiv.org</title><description>Symmetry is a fundamental aspect of many real-world robotic tasks. However, current deep reinforcement learning (DRL) approaches can seldom harness and exploit symmetry effectively. Often, the learned behaviors fail to achieve the desired transformation invariances and suffer from motion artifacts. For instance, a quadruped may exhibit different gaits when commanded to move forward or backward, even though it is symmetrical about its torso. This issue becomes further pronounced in high-dimensional or complex environments, where DRL methods are prone to local optima and fail to explore regions of the state space equally. Past methods on encouraging symmetry for robotic tasks have studied this topic mainly in a single-task setting, where symmetry usually refers to symmetry in the motion, such as the gait patterns. In this paper, we revisit this topic for goal-conditioned tasks in robotics, where symmetry lies mainly in task execution and not necessarily in the learned motions themselves. In particular, we investigate two approaches to incorporate symmetry invariance into DRL -- data augmentation and mirror loss function. We provide a theoretical foundation for using augmented samples in an on-policy setting. Based on this, we show that the corresponding approach achieves faster convergence and improves the learned behaviors in various challenging robotic tasks, from climbing boxes with a quadruped to dexterous manipulation.</description><subject>Cognitive tasks</subject><subject>Data augmentation</subject><subject>Deep learning</subject><subject>Gait</subject><subject>Robot learning</subject><subject>Robotics</subject><subject>Symmetry</subject><subject>Torso</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mRwCq7MzU0tKapUcM7PK85MSS1KLMkEshTS8osUfFITi_Iy89IVQhKLsxWgKjOTFYLyk_JLFALyczKTM1OLeRhY0xJzilN5oTQ3g7Kba4izh25BUX5haWpxSXxWfmlRHlAq3sjS1MjCwNDQzNyYOFUA83Y6LQ</recordid><startdate>20240307</startdate><enddate>20240307</enddate><creator>Mittal, Mayank</creator><creator>Rudin, Nikita</creator><creator>Klemm, Victor</creator><creator>Allshire, Arthur</creator><creator>Hutter, Marco</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240307</creationdate><title>Symmetry Considerations for Learning Task Symmetric Robot Policies</title><author>Mittal, Mayank ; Rudin, Nikita ; Klemm, Victor ; Allshire, Arthur ; Hutter, Marco</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_29528011673</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Cognitive tasks</topic><topic>Data augmentation</topic><topic>Deep learning</topic><topic>Gait</topic><topic>Robot learning</topic><topic>Robotics</topic><topic>Symmetry</topic><topic>Torso</topic><toplevel>online_resources</toplevel><creatorcontrib>Mittal, Mayank</creatorcontrib><creatorcontrib>Rudin, Nikita</creatorcontrib><creatorcontrib>Klemm, Victor</creatorcontrib><creatorcontrib>Allshire, Arthur</creatorcontrib><creatorcontrib>Hutter, Marco</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>ProQuest Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Mittal, Mayank</au><au>Rudin, Nikita</au><au>Klemm, Victor</au><au>Allshire, Arthur</au><au>Hutter, Marco</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Symmetry Considerations for Learning Task Symmetric Robot Policies</atitle><jtitle>arXiv.org</jtitle><date>2024-03-07</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Symmetry is a fundamental aspect of many real-world robotic tasks. However, current deep reinforcement learning (DRL) approaches can seldom harness and exploit symmetry effectively. Often, the learned behaviors fail to achieve the desired transformation invariances and suffer from motion artifacts. For instance, a quadruped may exhibit different gaits when commanded to move forward or backward, even though it is symmetrical about its torso. This issue becomes further pronounced in high-dimensional or complex environments, where DRL methods are prone to local optima and fail to explore regions of the state space equally. Past methods on encouraging symmetry for robotic tasks have studied this topic mainly in a single-task setting, where symmetry usually refers to symmetry in the motion, such as the gait patterns. In this paper, we revisit this topic for goal-conditioned tasks in robotics, where symmetry lies mainly in task execution and not necessarily in the learned motions themselves. In particular, we investigate two approaches to incorporate symmetry invariance into DRL -- data augmentation and mirror loss function. We provide a theoretical foundation for using augmented samples in an on-policy setting. Based on this, we show that the corresponding approach achieves faster convergence and improves the learned behaviors in various challenging robotic tasks, from climbing boxes with a quadruped to dexterous manipulation.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-03
issn 2331-8422
language eng
recordid cdi_proquest_journals_2952801167
source Publicly Available Content Database (Proquest) (PQ_SDU_P3)
subjects Cognitive tasks
Data augmentation
Deep learning
Gait
Robot learning
Robotics
Symmetry
Torso
title Symmetry Considerations for Learning Task Symmetric Robot Policies
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T05%3A33%3A26IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Symmetry%20Considerations%20for%20Learning%20Task%20Symmetric%20Robot%20Policies&rft.jtitle=arXiv.org&rft.au=Mittal,%20Mayank&rft.date=2024-03-07&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2952801167%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_29528011673%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2952801167&rft_id=info:pmid/&rfr_iscdi=true