Loading…

Learning Symbolic Rules over Abstract Meaning Representations for Textual Reinforcement Learning

Text-based reinforcement learning agents have predominantly been neural network-based models with embeddings-based representation, learning uninterpretable policies that often do not generalize well to unseen games. On the other hand, neuro-symbolic methods, specifically those that leverage an inter...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2023-07
Main Authors: Chaudhury, Subhajit, Swaminathan, Sarathkrishna, Kimura, Daiki, Sen, Prithviraj, Murugesan, Keerthiram, Uceda-Sosa, Rosario, Tatsubori, Michiaki, Fokoue, Achille, Kapanipathi, Pavan, Munawar, Asim, Gray, Alexander
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Chaudhury, Subhajit
Swaminathan, Sarathkrishna
Kimura, Daiki
Sen, Prithviraj
Murugesan, Keerthiram
Uceda-Sosa, Rosario
Tatsubori, Michiaki
Fokoue, Achille
Kapanipathi, Pavan
Munawar, Asim
Gray, Alexander
description Text-based reinforcement learning agents have predominantly been neural network-based models with embeddings-based representation, learning uninterpretable policies that often do not generalize well to unseen games. On the other hand, neuro-symbolic methods, specifically those that leverage an intermediate formal representation, are gaining significant attention in language understanding tasks. This is because of their advantages ranging from inherent interpretability, the lesser requirement of training data, and being generalizable in scenarios with unseen data. Therefore, in this paper, we propose a modular, NEuro-Symbolic Textual Agent (NESTA) that combines a generic semantic parser with a rule induction system to learn abstract interpretable rules as policies. Our experiments on established text-based game benchmarks show that the proposed NESTA method outperforms deep reinforcement learning-based techniques by achieving better generalization to unseen test games and learning from fewer training interactions.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2834345290</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2834345290</sourcerecordid><originalsourceid>FETCH-proquest_journals_28343452903</originalsourceid><addsrcrecordid>eNqNjL0KwjAURoMgWLTvcMG5EJNW6yiiOOhS3WtabiUlTWp-RN_eIro7fRzO4RuRiHG-SPKUsQmJnWsppWy5YlnGI3I9orBa6hucX11llKyhCAodmAda2FTOW1F7OKH4RAX2Fh1qL7w02kFjLFzw6YNQg5N64Bq7wcPvd0bGjVAO4-9OyXy_u2wPSW_NPaDzZWuC1YMqWc5TnmZsTfl_1RtDEkWW</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2834345290</pqid></control><display><type>article</type><title>Learning Symbolic Rules over Abstract Meaning Representations for Textual Reinforcement Learning</title><source>Publicly Available Content Database</source><creator>Chaudhury, Subhajit ; Swaminathan, Sarathkrishna ; Kimura, Daiki ; Sen, Prithviraj ; Murugesan, Keerthiram ; Uceda-Sosa, Rosario ; Tatsubori, Michiaki ; Fokoue, Achille ; Kapanipathi, Pavan ; Munawar, Asim ; Gray, Alexander</creator><creatorcontrib>Chaudhury, Subhajit ; Swaminathan, Sarathkrishna ; Kimura, Daiki ; Sen, Prithviraj ; Murugesan, Keerthiram ; Uceda-Sosa, Rosario ; Tatsubori, Michiaki ; Fokoue, Achille ; Kapanipathi, Pavan ; Munawar, Asim ; Gray, Alexander</creatorcontrib><description>Text-based reinforcement learning agents have predominantly been neural network-based models with embeddings-based representation, learning uninterpretable policies that often do not generalize well to unseen games. On the other hand, neuro-symbolic methods, specifically those that leverage an intermediate formal representation, are gaining significant attention in language understanding tasks. This is because of their advantages ranging from inherent interpretability, the lesser requirement of training data, and being generalizable in scenarios with unseen data. Therefore, in this paper, we propose a modular, NEuro-Symbolic Textual Agent (NESTA) that combines a generic semantic parser with a rule induction system to learn abstract interpretable rules as policies. Our experiments on established text-based game benchmarks show that the proposed NESTA method outperforms deep reinforcement learning-based techniques by achieving better generalization to unseen test games and learning from fewer training interactions.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Deep learning ; Games ; Machine learning ; Neural networks ; Policies ; Representations ; Rule induction ; Training</subject><ispartof>arXiv.org, 2023-07</ispartof><rights>2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2834345290?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Chaudhury, Subhajit</creatorcontrib><creatorcontrib>Swaminathan, Sarathkrishna</creatorcontrib><creatorcontrib>Kimura, Daiki</creatorcontrib><creatorcontrib>Sen, Prithviraj</creatorcontrib><creatorcontrib>Murugesan, Keerthiram</creatorcontrib><creatorcontrib>Uceda-Sosa, Rosario</creatorcontrib><creatorcontrib>Tatsubori, Michiaki</creatorcontrib><creatorcontrib>Fokoue, Achille</creatorcontrib><creatorcontrib>Kapanipathi, Pavan</creatorcontrib><creatorcontrib>Munawar, Asim</creatorcontrib><creatorcontrib>Gray, Alexander</creatorcontrib><title>Learning Symbolic Rules over Abstract Meaning Representations for Textual Reinforcement Learning</title><title>arXiv.org</title><description>Text-based reinforcement learning agents have predominantly been neural network-based models with embeddings-based representation, learning uninterpretable policies that often do not generalize well to unseen games. On the other hand, neuro-symbolic methods, specifically those that leverage an intermediate formal representation, are gaining significant attention in language understanding tasks. This is because of their advantages ranging from inherent interpretability, the lesser requirement of training data, and being generalizable in scenarios with unseen data. Therefore, in this paper, we propose a modular, NEuro-Symbolic Textual Agent (NESTA) that combines a generic semantic parser with a rule induction system to learn abstract interpretable rules as policies. Our experiments on established text-based game benchmarks show that the proposed NESTA method outperforms deep reinforcement learning-based techniques by achieving better generalization to unseen test games and learning from fewer training interactions.</description><subject>Deep learning</subject><subject>Games</subject><subject>Machine learning</subject><subject>Neural networks</subject><subject>Policies</subject><subject>Representations</subject><subject>Rule induction</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNjL0KwjAURoMgWLTvcMG5EJNW6yiiOOhS3WtabiUlTWp-RN_eIro7fRzO4RuRiHG-SPKUsQmJnWsppWy5YlnGI3I9orBa6hucX11llKyhCAodmAda2FTOW1F7OKH4RAX2Fh1qL7w02kFjLFzw6YNQg5N64Bq7wcPvd0bGjVAO4-9OyXy_u2wPSW_NPaDzZWuC1YMqWc5TnmZsTfl_1RtDEkWW</recordid><startdate>20230705</startdate><enddate>20230705</enddate><creator>Chaudhury, Subhajit</creator><creator>Swaminathan, Sarathkrishna</creator><creator>Kimura, Daiki</creator><creator>Sen, Prithviraj</creator><creator>Murugesan, Keerthiram</creator><creator>Uceda-Sosa, Rosario</creator><creator>Tatsubori, Michiaki</creator><creator>Fokoue, Achille</creator><creator>Kapanipathi, Pavan</creator><creator>Munawar, Asim</creator><creator>Gray, Alexander</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20230705</creationdate><title>Learning Symbolic Rules over Abstract Meaning Representations for Textual Reinforcement Learning</title><author>Chaudhury, Subhajit ; Swaminathan, Sarathkrishna ; Kimura, Daiki ; Sen, Prithviraj ; Murugesan, Keerthiram ; Uceda-Sosa, Rosario ; Tatsubori, Michiaki ; Fokoue, Achille ; Kapanipathi, Pavan ; Munawar, Asim ; Gray, Alexander</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_28343452903</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Deep learning</topic><topic>Games</topic><topic>Machine learning</topic><topic>Neural networks</topic><topic>Policies</topic><topic>Representations</topic><topic>Rule induction</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Chaudhury, Subhajit</creatorcontrib><creatorcontrib>Swaminathan, Sarathkrishna</creatorcontrib><creatorcontrib>Kimura, Daiki</creatorcontrib><creatorcontrib>Sen, Prithviraj</creatorcontrib><creatorcontrib>Murugesan, Keerthiram</creatorcontrib><creatorcontrib>Uceda-Sosa, Rosario</creatorcontrib><creatorcontrib>Tatsubori, Michiaki</creatorcontrib><creatorcontrib>Fokoue, Achille</creatorcontrib><creatorcontrib>Kapanipathi, Pavan</creatorcontrib><creatorcontrib>Munawar, Asim</creatorcontrib><creatorcontrib>Gray, Alexander</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chaudhury, Subhajit</au><au>Swaminathan, Sarathkrishna</au><au>Kimura, Daiki</au><au>Sen, Prithviraj</au><au>Murugesan, Keerthiram</au><au>Uceda-Sosa, Rosario</au><au>Tatsubori, Michiaki</au><au>Fokoue, Achille</au><au>Kapanipathi, Pavan</au><au>Munawar, Asim</au><au>Gray, Alexander</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Learning Symbolic Rules over Abstract Meaning Representations for Textual Reinforcement Learning</atitle><jtitle>arXiv.org</jtitle><date>2023-07-05</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Text-based reinforcement learning agents have predominantly been neural network-based models with embeddings-based representation, learning uninterpretable policies that often do not generalize well to unseen games. On the other hand, neuro-symbolic methods, specifically those that leverage an intermediate formal representation, are gaining significant attention in language understanding tasks. This is because of their advantages ranging from inherent interpretability, the lesser requirement of training data, and being generalizable in scenarios with unseen data. Therefore, in this paper, we propose a modular, NEuro-Symbolic Textual Agent (NESTA) that combines a generic semantic parser with a rule induction system to learn abstract interpretable rules as policies. Our experiments on established text-based game benchmarks show that the proposed NESTA method outperforms deep reinforcement learning-based techniques by achieving better generalization to unseen test games and learning from fewer training interactions.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-07
issn 2331-8422
language eng
recordid cdi_proquest_journals_2834345290
source Publicly Available Content Database
subjects Deep learning
Games
Machine learning
Neural networks
Policies
Representations
Rule induction
Training
title Learning Symbolic Rules over Abstract Meaning Representations for Textual Reinforcement Learning
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T23%3A13%3A44IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Learning%20Symbolic%20Rules%20over%20Abstract%20Meaning%20Representations%20for%20Textual%20Reinforcement%20Learning&rft.jtitle=arXiv.org&rft.au=Chaudhury,%20Subhajit&rft.date=2023-07-05&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2834345290%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_28343452903%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2834345290&rft_id=info:pmid/&rfr_iscdi=true