Loading…
A Reinforcement Learning-based Adaptive Time-Delay Control and Its Application to Robot Manipulators
This study proposes an innovative reinforcement learning-based time-delay control (RL-TDC) scheme to provide more intelligent, timely, and aggressive control efforts than the existing simple-structured adaptive time-delay controls (ATDCs) that are well-known for achieving good tracking performances...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Conference Proceeding |
Language: | English |
Subjects: | |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | 2729 |
container_issue | |
container_start_page | 2722 |
container_title | |
container_volume | |
creator | Baek, Seungmin Baek, Jongchan Choi, Jinsuk Han, Soohee |
description | This study proposes an innovative reinforcement learning-based time-delay control (RL-TDC) scheme to provide more intelligent, timely, and aggressive control efforts than the existing simple-structured adaptive time-delay controls (ATDCs) that are well-known for achieving good tracking performances in practical applications. The proposed control scheme adopts a state-of-the-art RL algorithm called soft actor critic (SAC) with which the inertia gain matrix of the time-delay control is adjusted toward maximizing the expected return obtained from tracking errors over all the future time periods. By learning the dynamics of the robot manipulator with a data-driven approach, and capturing its intractable and complicated phenomena, the proposed RL-TDC is trained to effectively suppress the inherent time delay estimation (TDE) errors arising from time delay control, thereby ensuring the best tracking performance within the given control capacity limits. As expected, it is demonstrated through simulation with a robot manipulator that the proposed RL-TDC avoids conservative small control actions when large ones are required, for maximizing the tracking performance. It is observed that the stability condition is fully exploited to provide more effective control actions. |
doi_str_mv | 10.23919/ACC53348.2022.9867835 |
format | conference_proceeding |
fullrecord | <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_9867835</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9867835</ieee_id><sourcerecordid>9867835</sourcerecordid><originalsourceid>FETCH-LOGICAL-i133t-510c265647f54e5d041fee5d64012fc4b7e4d41c4db8762de608036ae6a0a4c23</originalsourceid><addsrcrecordid>eNot0N1KwzAYgOEoCM65KxAkN9CZ_yaHpf4NKsKYx-Nr81UibVLaKOzuFdzRc_YevITcc7YV0nH3UNW1llLZrWBCbJ01pZX6gmxcabkxWmnujLwkKyFLW2hr-DW5WZYvxrhzhq2Ir-geQ-zT3OGIMdMGYY4hfhYtLOhp5WHK4QfpIYxYPOIAJ1qnmOc0UIie7vJCq2kaQgc5pEhzovvUpkzfIIbpe4Cc5uWWXPUwLLg5uyYfz0-H-rVo3l92ddUUgUuZC81ZJ4w2quy1Qu2Z4j3-aRTjou9UW6LyinfKt7Y0wqNhlkkDaICB6oRck7v_bkDE4zSHEebT8TxF_gK7AFcs</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>A Reinforcement Learning-based Adaptive Time-Delay Control and Its Application to Robot Manipulators</title><source>IEEE Xplore All Conference Series</source><creator>Baek, Seungmin ; Baek, Jongchan ; Choi, Jinsuk ; Han, Soohee</creator><creatorcontrib>Baek, Seungmin ; Baek, Jongchan ; Choi, Jinsuk ; Han, Soohee</creatorcontrib><description>This study proposes an innovative reinforcement learning-based time-delay control (RL-TDC) scheme to provide more intelligent, timely, and aggressive control efforts than the existing simple-structured adaptive time-delay controls (ATDCs) that are well-known for achieving good tracking performances in practical applications. The proposed control scheme adopts a state-of-the-art RL algorithm called soft actor critic (SAC) with which the inertia gain matrix of the time-delay control is adjusted toward maximizing the expected return obtained from tracking errors over all the future time periods. By learning the dynamics of the robot manipulator with a data-driven approach, and capturing its intractable and complicated phenomena, the proposed RL-TDC is trained to effectively suppress the inherent time delay estimation (TDE) errors arising from time delay control, thereby ensuring the best tracking performance within the given control capacity limits. As expected, it is demonstrated through simulation with a robot manipulator that the proposed RL-TDC avoids conservative small control actions when large ones are required, for maximizing the tracking performance. It is observed that the stability condition is fully exploited to provide more effective control actions.</description><identifier>EISSN: 2378-5861</identifier><identifier>EISBN: 9781665451963</identifier><identifier>EISBN: 1665451963</identifier><identifier>DOI: 10.23919/ACC53348.2022.9867835</identifier><language>eng</language><publisher>American Automatic Control Council</publisher><subject>Adaptation models ; Delay effects ; Estimation ; Heuristic algorithms ; Manipulator dynamics ; Reliability ; Stability analysis</subject><ispartof>2022 American Control Conference (ACC), 2022, p.2722-2729</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9867835$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,27925,54555,54932</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9867835$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Baek, Seungmin</creatorcontrib><creatorcontrib>Baek, Jongchan</creatorcontrib><creatorcontrib>Choi, Jinsuk</creatorcontrib><creatorcontrib>Han, Soohee</creatorcontrib><title>A Reinforcement Learning-based Adaptive Time-Delay Control and Its Application to Robot Manipulators</title><title>2022 American Control Conference (ACC)</title><addtitle>ACC</addtitle><description>This study proposes an innovative reinforcement learning-based time-delay control (RL-TDC) scheme to provide more intelligent, timely, and aggressive control efforts than the existing simple-structured adaptive time-delay controls (ATDCs) that are well-known for achieving good tracking performances in practical applications. The proposed control scheme adopts a state-of-the-art RL algorithm called soft actor critic (SAC) with which the inertia gain matrix of the time-delay control is adjusted toward maximizing the expected return obtained from tracking errors over all the future time periods. By learning the dynamics of the robot manipulator with a data-driven approach, and capturing its intractable and complicated phenomena, the proposed RL-TDC is trained to effectively suppress the inherent time delay estimation (TDE) errors arising from time delay control, thereby ensuring the best tracking performance within the given control capacity limits. As expected, it is demonstrated through simulation with a robot manipulator that the proposed RL-TDC avoids conservative small control actions when large ones are required, for maximizing the tracking performance. It is observed that the stability condition is fully exploited to provide more effective control actions.</description><subject>Adaptation models</subject><subject>Delay effects</subject><subject>Estimation</subject><subject>Heuristic algorithms</subject><subject>Manipulator dynamics</subject><subject>Reliability</subject><subject>Stability analysis</subject><issn>2378-5861</issn><isbn>9781665451963</isbn><isbn>1665451963</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2022</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNot0N1KwzAYgOEoCM65KxAkN9CZ_yaHpf4NKsKYx-Nr81UibVLaKOzuFdzRc_YevITcc7YV0nH3UNW1llLZrWBCbJ01pZX6gmxcabkxWmnujLwkKyFLW2hr-DW5WZYvxrhzhq2Ir-geQ-zT3OGIMdMGYY4hfhYtLOhp5WHK4QfpIYxYPOIAJ1qnmOc0UIie7vJCq2kaQgc5pEhzovvUpkzfIIbpe4Cc5uWWXPUwLLg5uyYfz0-H-rVo3l92ddUUgUuZC81ZJ4w2quy1Qu2Z4j3-aRTjou9UW6LyinfKt7Y0wqNhlkkDaICB6oRck7v_bkDE4zSHEebT8TxF_gK7AFcs</recordid><startdate>20220608</startdate><enddate>20220608</enddate><creator>Baek, Seungmin</creator><creator>Baek, Jongchan</creator><creator>Choi, Jinsuk</creator><creator>Han, Soohee</creator><general>American Automatic Control Council</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>20220608</creationdate><title>A Reinforcement Learning-based Adaptive Time-Delay Control and Its Application to Robot Manipulators</title><author>Baek, Seungmin ; Baek, Jongchan ; Choi, Jinsuk ; Han, Soohee</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i133t-510c265647f54e5d041fee5d64012fc4b7e4d41c4db8762de608036ae6a0a4c23</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Adaptation models</topic><topic>Delay effects</topic><topic>Estimation</topic><topic>Heuristic algorithms</topic><topic>Manipulator dynamics</topic><topic>Reliability</topic><topic>Stability analysis</topic><toplevel>online_resources</toplevel><creatorcontrib>Baek, Seungmin</creatorcontrib><creatorcontrib>Baek, Jongchan</creatorcontrib><creatorcontrib>Choi, Jinsuk</creatorcontrib><creatorcontrib>Han, Soohee</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Baek, Seungmin</au><au>Baek, Jongchan</au><au>Choi, Jinsuk</au><au>Han, Soohee</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>A Reinforcement Learning-based Adaptive Time-Delay Control and Its Application to Robot Manipulators</atitle><btitle>2022 American Control Conference (ACC)</btitle><stitle>ACC</stitle><date>2022-06-08</date><risdate>2022</risdate><spage>2722</spage><epage>2729</epage><pages>2722-2729</pages><eissn>2378-5861</eissn><eisbn>9781665451963</eisbn><eisbn>1665451963</eisbn><abstract>This study proposes an innovative reinforcement learning-based time-delay control (RL-TDC) scheme to provide more intelligent, timely, and aggressive control efforts than the existing simple-structured adaptive time-delay controls (ATDCs) that are well-known for achieving good tracking performances in practical applications. The proposed control scheme adopts a state-of-the-art RL algorithm called soft actor critic (SAC) with which the inertia gain matrix of the time-delay control is adjusted toward maximizing the expected return obtained from tracking errors over all the future time periods. By learning the dynamics of the robot manipulator with a data-driven approach, and capturing its intractable and complicated phenomena, the proposed RL-TDC is trained to effectively suppress the inherent time delay estimation (TDE) errors arising from time delay control, thereby ensuring the best tracking performance within the given control capacity limits. As expected, it is demonstrated through simulation with a robot manipulator that the proposed RL-TDC avoids conservative small control actions when large ones are required, for maximizing the tracking performance. It is observed that the stability condition is fully exploited to provide more effective control actions.</abstract><pub>American Automatic Control Council</pub><doi>10.23919/ACC53348.2022.9867835</doi><tpages>8</tpages></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | EISSN: 2378-5861 |
ispartof | 2022 American Control Conference (ACC), 2022, p.2722-2729 |
issn | 2378-5861 |
language | eng |
recordid | cdi_ieee_primary_9867835 |
source | IEEE Xplore All Conference Series |
subjects | Adaptation models Delay effects Estimation Heuristic algorithms Manipulator dynamics Reliability Stability analysis |
title | A Reinforcement Learning-based Adaptive Time-Delay Control and Its Application to Robot Manipulators |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T17%3A42%3A53IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=A%20Reinforcement%20Learning-based%20Adaptive%20Time-Delay%20Control%20and%20Its%20Application%20to%20Robot%20Manipulators&rft.btitle=2022%20American%20Control%20Conference%20(ACC)&rft.au=Baek,%20Seungmin&rft.date=2022-06-08&rft.spage=2722&rft.epage=2729&rft.pages=2722-2729&rft.eissn=2378-5861&rft_id=info:doi/10.23919/ACC53348.2022.9867835&rft.eisbn=9781665451963&rft.eisbn_list=1665451963&rft_dat=%3Cieee_CHZPO%3E9867835%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i133t-510c265647f54e5d041fee5d64012fc4b7e4d41c4db8762de608036ae6a0a4c23%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=9867835&rfr_iscdi=true |