Loading…

Approximating a deep reinforcement learning docking agent using linear model trees

Deep reinforcement learning has led to numerous notable results in robotics. However, deep neural networks (DNNs) are unintuitive, which makes it difficult to understand their predictions and strongly limits their potential for real-world applications due to economic, safety, and assurance reasons....

Full description

Saved in:
Bibliographic Details
Main Authors: Gjærum, Vilde Benoni, Rørvik, Ella-Lovise H, Lekkas, Anastasios M
Format: Article
Language:English
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Gjærum, Vilde Benoni
Rørvik, Ella-Lovise H
Lekkas, Anastasios M
description Deep reinforcement learning has led to numerous notable results in robotics. However, deep neural networks (DNNs) are unintuitive, which makes it difficult to understand their predictions and strongly limits their potential for real-world applications due to economic, safety, and assurance reasons. To remedy this problem, a number of explainable AI methods have been presented, such as SHAP and LIME, but these can be either be too costly to be used in real-time robotic applications or provide only local explanations. In this paper, the main contribution is the use of a linear model tree (LMT) to approximate a DNN policy, originally trained via proximal policy optimization(PPO), for an autonomous surface vehicle with five control inputs performing a docking operation. The two main benefits of the proposed approach are: a) LMTs are transparent which makes it possible to associate directly the outputs (control actions, in our case) with specific values of the input features, b) LMTs are computationally efficient and can provide information in real-time. In our simulations, the opaque DNN policy controls the vehicle and the LMT runs in parallel to provide explanations in the form of feature attributions. Our results indicate that LMTs can be a useful component within digital assurance frameworks for autonomous ships.
format article
fullrecord <record><control><sourceid>cristin_3HK</sourceid><recordid>TN_cdi_cristin_nora_11250_2838850</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>11250_2838850</sourcerecordid><originalsourceid>FETCH-cristin_nora_11250_28388503</originalsourceid><addsrcrecordid>eNrjZAhyLCgoyq_IzE0sycxLV0hUSElNLVAoSs3MS8svSk7NTc0rUchJTSzKA8mm5Cdng1Wlg4RLi0HsnMw8oLRCbn5Kao5CSVFqajEPA2taYk5xKi-U5mZQdHMNcfbQTS7KLAbaEp-XX5QYb2hoZGoQb2RhbGFhamBMjBoAQFc4Dw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Approximating a deep reinforcement learning docking agent using linear model trees</title><source>NORA - Norwegian Open Research Archives</source><creator>Gjærum, Vilde Benoni ; Rørvik, Ella-Lovise H ; Lekkas, Anastasios M</creator><creatorcontrib>Gjærum, Vilde Benoni ; Rørvik, Ella-Lovise H ; Lekkas, Anastasios M</creatorcontrib><description>Deep reinforcement learning has led to numerous notable results in robotics. However, deep neural networks (DNNs) are unintuitive, which makes it difficult to understand their predictions and strongly limits their potential for real-world applications due to economic, safety, and assurance reasons. To remedy this problem, a number of explainable AI methods have been presented, such as SHAP and LIME, but these can be either be too costly to be used in real-time robotic applications or provide only local explanations. In this paper, the main contribution is the use of a linear model tree (LMT) to approximate a DNN policy, originally trained via proximal policy optimization(PPO), for an autonomous surface vehicle with five control inputs performing a docking operation. The two main benefits of the proposed approach are: a) LMTs are transparent which makes it possible to associate directly the outputs (control actions, in our case) with specific values of the input features, b) LMTs are computationally efficient and can provide information in real-time. In our simulations, the opaque DNN policy controls the vehicle and the LMT runs in parallel to provide explanations in the form of feature attributions. Our results indicate that LMTs can be a useful component within digital assurance frameworks for autonomous ships.</description><language>eng</language><publisher>Institute of Electrical and Electronics Engineers (IEEE)</publisher><creationdate>2021</creationdate><rights>info:eu-repo/semantics/openAccess</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,780,885,26567</link.rule.ids><linktorsrc>$$Uhttp://hdl.handle.net/11250/2838850$$EView_record_in_NORA$$FView_record_in_$$GNORA$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>Gjærum, Vilde Benoni</creatorcontrib><creatorcontrib>Rørvik, Ella-Lovise H</creatorcontrib><creatorcontrib>Lekkas, Anastasios M</creatorcontrib><title>Approximating a deep reinforcement learning docking agent using linear model trees</title><description>Deep reinforcement learning has led to numerous notable results in robotics. However, deep neural networks (DNNs) are unintuitive, which makes it difficult to understand their predictions and strongly limits their potential for real-world applications due to economic, safety, and assurance reasons. To remedy this problem, a number of explainable AI methods have been presented, such as SHAP and LIME, but these can be either be too costly to be used in real-time robotic applications or provide only local explanations. In this paper, the main contribution is the use of a linear model tree (LMT) to approximate a DNN policy, originally trained via proximal policy optimization(PPO), for an autonomous surface vehicle with five control inputs performing a docking operation. The two main benefits of the proposed approach are: a) LMTs are transparent which makes it possible to associate directly the outputs (control actions, in our case) with specific values of the input features, b) LMTs are computationally efficient and can provide information in real-time. In our simulations, the opaque DNN policy controls the vehicle and the LMT runs in parallel to provide explanations in the form of feature attributions. Our results indicate that LMTs can be a useful component within digital assurance frameworks for autonomous ships.</description><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>3HK</sourceid><recordid>eNrjZAhyLCgoyq_IzE0sycxLV0hUSElNLVAoSs3MS8svSk7NTc0rUchJTSzKA8mm5Cdng1Wlg4RLi0HsnMw8oLRCbn5Kao5CSVFqajEPA2taYk5xKi-U5mZQdHMNcfbQTS7KLAbaEp-XX5QYb2hoZGoQb2RhbGFhamBMjBoAQFc4Dw</recordid><startdate>2021</startdate><enddate>2021</enddate><creator>Gjærum, Vilde Benoni</creator><creator>Rørvik, Ella-Lovise H</creator><creator>Lekkas, Anastasios M</creator><general>Institute of Electrical and Electronics Engineers (IEEE)</general><scope>3HK</scope></search><sort><creationdate>2021</creationdate><title>Approximating a deep reinforcement learning docking agent using linear model trees</title><author>Gjærum, Vilde Benoni ; Rørvik, Ella-Lovise H ; Lekkas, Anastasios M</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-cristin_nora_11250_28388503</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><toplevel>online_resources</toplevel><creatorcontrib>Gjærum, Vilde Benoni</creatorcontrib><creatorcontrib>Rørvik, Ella-Lovise H</creatorcontrib><creatorcontrib>Lekkas, Anastasios M</creatorcontrib><collection>NORA - Norwegian Open Research Archives</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Gjærum, Vilde Benoni</au><au>Rørvik, Ella-Lovise H</au><au>Lekkas, Anastasios M</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Approximating a deep reinforcement learning docking agent using linear model trees</atitle><date>2021</date><risdate>2021</risdate><abstract>Deep reinforcement learning has led to numerous notable results in robotics. However, deep neural networks (DNNs) are unintuitive, which makes it difficult to understand their predictions and strongly limits their potential for real-world applications due to economic, safety, and assurance reasons. To remedy this problem, a number of explainable AI methods have been presented, such as SHAP and LIME, but these can be either be too costly to be used in real-time robotic applications or provide only local explanations. In this paper, the main contribution is the use of a linear model tree (LMT) to approximate a DNN policy, originally trained via proximal policy optimization(PPO), for an autonomous surface vehicle with five control inputs performing a docking operation. The two main benefits of the proposed approach are: a) LMTs are transparent which makes it possible to associate directly the outputs (control actions, in our case) with specific values of the input features, b) LMTs are computationally efficient and can provide information in real-time. In our simulations, the opaque DNN policy controls the vehicle and the LMT runs in parallel to provide explanations in the form of feature attributions. Our results indicate that LMTs can be a useful component within digital assurance frameworks for autonomous ships.</abstract><pub>Institute of Electrical and Electronics Engineers (IEEE)</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng
recordid cdi_cristin_nora_11250_2838850
source NORA - Norwegian Open Research Archives
title Approximating a deep reinforcement learning docking agent using linear model trees
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T14%3A47%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-cristin_3HK&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Approximating%20a%20deep%20reinforcement%20learning%20docking%20agent%20using%20linear%20model%20trees&rft.au=Gj%C3%A6rum,%20Vilde%20Benoni&rft.date=2021&rft_id=info:doi/&rft_dat=%3Ccristin_3HK%3E11250_2838850%3C/cristin_3HK%3E%3Cgrp_id%3Ecdi_FETCH-cristin_nora_11250_28388503%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true