Loading…
Symbolic Representation of Recurrent Neural Network Dynamics
Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic hum...
Saved in:
Published in: | IEEE transaction on neural networks and learning systems 2012-10, Vol.23 (10), p.1649-1658 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3 |
---|---|
cites | cdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3 |
container_end_page | 1658 |
container_issue | 10 |
container_start_page | 1649 |
container_title | IEEE transaction on neural networks and learning systems |
container_volume | 23 |
creator | Huynh, T. Q. Reggia, J. A. |
description | Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods. |
doi_str_mv | 10.1109/TNNLS.2012.2210242 |
format | article |
fullrecord | <record><control><sourceid>proquest_pasca</sourceid><recordid>TN_cdi_pascalfrancis_primary_26446530</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>6269105</ieee_id><sourcerecordid>1523404873</sourcerecordid><originalsourceid>FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</originalsourceid><addsrcrecordid>eNqFkd9LG0EQx5dSqRL9B1oogSL4cnF2dvZHwBfRWoUQQVPo27G32YPT-xF37yj577sxaQRfnJcZZj4zzMyXsa8cJpzD9Hwxn88eJwgcJ4gckPATO0KuMENhzOd9rP8cspMYnyCZAqlo-oUdIhkwANMjdvG4boqurtz4wa-Cj77tbV917bgrU8YNIaTMeO6HYOvk-r9deB5fr1vbVC4es4PS1tGf7PyI_b75ubi6zWb3v-6uLmeZI059pqX2zpEkgqVagkDhCIyQ3DrrCqmMJi1cgbwwBUlLhdbESyy5LVMFSzFiZ9u5q9C9DD72eVNF5-vatr4bYs4lCgIyWnyMcqFIgjSY0B_v0KduCG06JOdAoLXCtNeI4ZZyoYsx-DJfhaqxYZ2gfKNE_qpEvlEi3ymRmr7vRg9F45f7lv9_T8DpDrDR2boMtnVVfOMUkZICEvdty1Xe-31ZoZpykOIfcsWW4Q</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1040776247</pqid></control><display><type>article</type><title>Symbolic Representation of Recurrent Neural Network Dynamics</title><source>IEEE Xplore (Online service)</source><creator>Huynh, T. Q. ; Reggia, J. A.</creator><creatorcontrib>Huynh, T. Q. ; Reggia, J. A.</creatorcontrib><description>Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.</description><identifier>ISSN: 2162-237X</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2012.2210242</identifier><identifier>PMID: 24808009</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>New York, NY: IEEE</publisher><subject>Algorithms ; Applied sciences ; Artificial intelligence ; Back propagation ; Backpropagation ; Computer science; control theory; systems ; Computer Simulation ; Connectionism. Neural networks ; Context ; Dynamical systems ; Encoding ; Exact sciences and technology ; Feedback ; Finite-state machines ; hidden layer representation ; Models, Statistical ; Neural networks ; Neural Networks (Computer) ; Pattern Recognition, Automated - methods ; penalty function ; Recurrent neural networks ; Studies ; Symbolism ; Training ; Vectors</subject><ispartof>IEEE transaction on neural networks and learning systems, 2012-10, Vol.23 (10), p.1649-1658</ispartof><rights>2014 INIST-CNRS</rights><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) Oct 2012</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</citedby><cites>FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/6269105$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&idt=26446530$$DView record in Pascal Francis$$Hfree_for_read</backlink><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/24808009$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Huynh, T. Q.</creatorcontrib><creatorcontrib>Reggia, J. A.</creatorcontrib><title>Symbolic Representation of Recurrent Neural Network Dynamics</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.</description><subject>Algorithms</subject><subject>Applied sciences</subject><subject>Artificial intelligence</subject><subject>Back propagation</subject><subject>Backpropagation</subject><subject>Computer science; control theory; systems</subject><subject>Computer Simulation</subject><subject>Connectionism. Neural networks</subject><subject>Context</subject><subject>Dynamical systems</subject><subject>Encoding</subject><subject>Exact sciences and technology</subject><subject>Feedback</subject><subject>Finite-state machines</subject><subject>hidden layer representation</subject><subject>Models, Statistical</subject><subject>Neural networks</subject><subject>Neural Networks (Computer)</subject><subject>Pattern Recognition, Automated - methods</subject><subject>penalty function</subject><subject>Recurrent neural networks</subject><subject>Studies</subject><subject>Symbolism</subject><subject>Training</subject><subject>Vectors</subject><issn>2162-237X</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2012</creationdate><recordtype>article</recordtype><recordid>eNqFkd9LG0EQx5dSqRL9B1oogSL4cnF2dvZHwBfRWoUQQVPo27G32YPT-xF37yj577sxaQRfnJcZZj4zzMyXsa8cJpzD9Hwxn88eJwgcJ4gckPATO0KuMENhzOd9rP8cspMYnyCZAqlo-oUdIhkwANMjdvG4boqurtz4wa-Cj77tbV917bgrU8YNIaTMeO6HYOvk-r9deB5fr1vbVC4es4PS1tGf7PyI_b75ubi6zWb3v-6uLmeZI059pqX2zpEkgqVagkDhCIyQ3DrrCqmMJi1cgbwwBUlLhdbESyy5LVMFSzFiZ9u5q9C9DD72eVNF5-vatr4bYs4lCgIyWnyMcqFIgjSY0B_v0KduCG06JOdAoLXCtNeI4ZZyoYsx-DJfhaqxYZ2gfKNE_qpEvlEi3ymRmr7vRg9F45f7lv9_T8DpDrDR2boMtnVVfOMUkZICEvdty1Xe-31ZoZpykOIfcsWW4Q</recordid><startdate>20121001</startdate><enddate>20121001</enddate><creator>Huynh, T. Q.</creator><creator>Reggia, J. A.</creator><general>IEEE</general><general>Institute of Electrical and Electronics Engineers</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>IQODW</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QP</scope><scope>7QQ</scope><scope>7QR</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7TK</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>P64</scope><scope>7X8</scope></search><sort><creationdate>20121001</creationdate><title>Symbolic Representation of Recurrent Neural Network Dynamics</title><author>Huynh, T. Q. ; Reggia, J. A.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2012</creationdate><topic>Algorithms</topic><topic>Applied sciences</topic><topic>Artificial intelligence</topic><topic>Back propagation</topic><topic>Backpropagation</topic><topic>Computer science; control theory; systems</topic><topic>Computer Simulation</topic><topic>Connectionism. Neural networks</topic><topic>Context</topic><topic>Dynamical systems</topic><topic>Encoding</topic><topic>Exact sciences and technology</topic><topic>Feedback</topic><topic>Finite-state machines</topic><topic>hidden layer representation</topic><topic>Models, Statistical</topic><topic>Neural networks</topic><topic>Neural Networks (Computer)</topic><topic>Pattern Recognition, Automated - methods</topic><topic>penalty function</topic><topic>Recurrent neural networks</topic><topic>Studies</topic><topic>Symbolism</topic><topic>Training</topic><topic>Vectors</topic><toplevel>online_resources</toplevel><creatorcontrib>Huynh, T. Q.</creatorcontrib><creatorcontrib>Reggia, J. A.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Xplore (Online service)</collection><collection>Pascal-Francis</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Calcium & Calcified Tissue Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology & Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Huynh, T. Q.</au><au>Reggia, J. A.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Symbolic Representation of Recurrent Neural Network Dynamics</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2012-10-01</date><risdate>2012</risdate><volume>23</volume><issue>10</issue><spage>1649</spage><epage>1658</epage><pages>1649-1658</pages><issn>2162-237X</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.</abstract><cop>New York, NY</cop><pub>IEEE</pub><pmid>24808009</pmid><doi>10.1109/TNNLS.2012.2210242</doi><tpages>10</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2162-237X |
ispartof | IEEE transaction on neural networks and learning systems, 2012-10, Vol.23 (10), p.1649-1658 |
issn | 2162-237X 2162-2388 |
language | eng |
recordid | cdi_pascalfrancis_primary_26446530 |
source | IEEE Xplore (Online service) |
subjects | Algorithms Applied sciences Artificial intelligence Back propagation Backpropagation Computer science control theory systems Computer Simulation Connectionism. Neural networks Context Dynamical systems Encoding Exact sciences and technology Feedback Finite-state machines hidden layer representation Models, Statistical Neural networks Neural Networks (Computer) Pattern Recognition, Automated - methods penalty function Recurrent neural networks Studies Symbolism Training Vectors |
title | Symbolic Representation of Recurrent Neural Network Dynamics |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T14%3A17%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pasca&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Symbolic%20Representation%20of%20Recurrent%20Neural%20Network%20Dynamics&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Huynh,%20T.%20Q.&rft.date=2012-10-01&rft.volume=23&rft.issue=10&rft.spage=1649&rft.epage=1658&rft.pages=1649-1658&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2012.2210242&rft_dat=%3Cproquest_pasca%3E1523404873%3C/proquest_pasca%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1040776247&rft_id=info:pmid/24808009&rft_ieee_id=6269105&rfr_iscdi=true |