Loading…

Symbolic Representation of Recurrent Neural Network Dynamics

Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic hum...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transaction on neural networks and learning systems 2012-10, Vol.23 (10), p.1649-1658
Main Authors: Huynh, T. Q., Reggia, J. A.
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3
cites cdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3
container_end_page 1658
container_issue 10
container_start_page 1649
container_title IEEE transaction on neural networks and learning systems
container_volume 23
creator Huynh, T. Q.
Reggia, J. A.
description Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.
doi_str_mv 10.1109/TNNLS.2012.2210242
format article
fullrecord <record><control><sourceid>proquest_pasca</sourceid><recordid>TN_cdi_pascalfrancis_primary_26446530</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>6269105</ieee_id><sourcerecordid>1523404873</sourcerecordid><originalsourceid>FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</originalsourceid><addsrcrecordid>eNqFkd9LG0EQx5dSqRL9B1oogSL4cnF2dvZHwBfRWoUQQVPo27G32YPT-xF37yj577sxaQRfnJcZZj4zzMyXsa8cJpzD9Hwxn88eJwgcJ4gckPATO0KuMENhzOd9rP8cspMYnyCZAqlo-oUdIhkwANMjdvG4boqurtz4wa-Cj77tbV917bgrU8YNIaTMeO6HYOvk-r9deB5fr1vbVC4es4PS1tGf7PyI_b75ubi6zWb3v-6uLmeZI059pqX2zpEkgqVagkDhCIyQ3DrrCqmMJi1cgbwwBUlLhdbESyy5LVMFSzFiZ9u5q9C9DD72eVNF5-vatr4bYs4lCgIyWnyMcqFIgjSY0B_v0KduCG06JOdAoLXCtNeI4ZZyoYsx-DJfhaqxYZ2gfKNE_qpEvlEi3ymRmr7vRg9F45f7lv9_T8DpDrDR2boMtnVVfOMUkZICEvdty1Xe-31ZoZpykOIfcsWW4Q</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1040776247</pqid></control><display><type>article</type><title>Symbolic Representation of Recurrent Neural Network Dynamics</title><source>IEEE Xplore (Online service)</source><creator>Huynh, T. Q. ; Reggia, J. A.</creator><creatorcontrib>Huynh, T. Q. ; Reggia, J. A.</creatorcontrib><description>Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.</description><identifier>ISSN: 2162-237X</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2012.2210242</identifier><identifier>PMID: 24808009</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>New York, NY: IEEE</publisher><subject>Algorithms ; Applied sciences ; Artificial intelligence ; Back propagation ; Backpropagation ; Computer science; control theory; systems ; Computer Simulation ; Connectionism. Neural networks ; Context ; Dynamical systems ; Encoding ; Exact sciences and technology ; Feedback ; Finite-state machines ; hidden layer representation ; Models, Statistical ; Neural networks ; Neural Networks (Computer) ; Pattern Recognition, Automated - methods ; penalty function ; Recurrent neural networks ; Studies ; Symbolism ; Training ; Vectors</subject><ispartof>IEEE transaction on neural networks and learning systems, 2012-10, Vol.23 (10), p.1649-1658</ispartof><rights>2014 INIST-CNRS</rights><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) Oct 2012</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</citedby><cites>FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/6269105$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&amp;idt=26446530$$DView record in Pascal Francis$$Hfree_for_read</backlink><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/24808009$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Huynh, T. Q.</creatorcontrib><creatorcontrib>Reggia, J. A.</creatorcontrib><title>Symbolic Representation of Recurrent Neural Network Dynamics</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.</description><subject>Algorithms</subject><subject>Applied sciences</subject><subject>Artificial intelligence</subject><subject>Back propagation</subject><subject>Backpropagation</subject><subject>Computer science; control theory; systems</subject><subject>Computer Simulation</subject><subject>Connectionism. Neural networks</subject><subject>Context</subject><subject>Dynamical systems</subject><subject>Encoding</subject><subject>Exact sciences and technology</subject><subject>Feedback</subject><subject>Finite-state machines</subject><subject>hidden layer representation</subject><subject>Models, Statistical</subject><subject>Neural networks</subject><subject>Neural Networks (Computer)</subject><subject>Pattern Recognition, Automated - methods</subject><subject>penalty function</subject><subject>Recurrent neural networks</subject><subject>Studies</subject><subject>Symbolism</subject><subject>Training</subject><subject>Vectors</subject><issn>2162-237X</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2012</creationdate><recordtype>article</recordtype><recordid>eNqFkd9LG0EQx5dSqRL9B1oogSL4cnF2dvZHwBfRWoUQQVPo27G32YPT-xF37yj577sxaQRfnJcZZj4zzMyXsa8cJpzD9Hwxn88eJwgcJ4gckPATO0KuMENhzOd9rP8cspMYnyCZAqlo-oUdIhkwANMjdvG4boqurtz4wa-Cj77tbV917bgrU8YNIaTMeO6HYOvk-r9deB5fr1vbVC4es4PS1tGf7PyI_b75ubi6zWb3v-6uLmeZI059pqX2zpEkgqVagkDhCIyQ3DrrCqmMJi1cgbwwBUlLhdbESyy5LVMFSzFiZ9u5q9C9DD72eVNF5-vatr4bYs4lCgIyWnyMcqFIgjSY0B_v0KduCG06JOdAoLXCtNeI4ZZyoYsx-DJfhaqxYZ2gfKNE_qpEvlEi3ymRmr7vRg9F45f7lv9_T8DpDrDR2boMtnVVfOMUkZICEvdty1Xe-31ZoZpykOIfcsWW4Q</recordid><startdate>20121001</startdate><enddate>20121001</enddate><creator>Huynh, T. Q.</creator><creator>Reggia, J. A.</creator><general>IEEE</general><general>Institute of Electrical and Electronics Engineers</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>IQODW</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QP</scope><scope>7QQ</scope><scope>7QR</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7TK</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>P64</scope><scope>7X8</scope></search><sort><creationdate>20121001</creationdate><title>Symbolic Representation of Recurrent Neural Network Dynamics</title><author>Huynh, T. Q. ; Reggia, J. A.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2012</creationdate><topic>Algorithms</topic><topic>Applied sciences</topic><topic>Artificial intelligence</topic><topic>Back propagation</topic><topic>Backpropagation</topic><topic>Computer science; control theory; systems</topic><topic>Computer Simulation</topic><topic>Connectionism. Neural networks</topic><topic>Context</topic><topic>Dynamical systems</topic><topic>Encoding</topic><topic>Exact sciences and technology</topic><topic>Feedback</topic><topic>Finite-state machines</topic><topic>hidden layer representation</topic><topic>Models, Statistical</topic><topic>Neural networks</topic><topic>Neural Networks (Computer)</topic><topic>Pattern Recognition, Automated - methods</topic><topic>penalty function</topic><topic>Recurrent neural networks</topic><topic>Studies</topic><topic>Symbolism</topic><topic>Training</topic><topic>Vectors</topic><toplevel>online_resources</toplevel><creatorcontrib>Huynh, T. Q.</creatorcontrib><creatorcontrib>Reggia, J. A.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Xplore (Online service)</collection><collection>Pascal-Francis</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Calcium &amp; Calcified Tissue Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Huynh, T. Q.</au><au>Reggia, J. A.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Symbolic Representation of Recurrent Neural Network Dynamics</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2012-10-01</date><risdate>2012</risdate><volume>23</volume><issue>10</issue><spage>1649</spage><epage>1658</epage><pages>1649-1658</pages><issn>2162-237X</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.</abstract><cop>New York, NY</cop><pub>IEEE</pub><pmid>24808009</pmid><doi>10.1109/TNNLS.2012.2210242</doi><tpages>10</tpages></addata></record>
fulltext fulltext
identifier ISSN: 2162-237X
ispartof IEEE transaction on neural networks and learning systems, 2012-10, Vol.23 (10), p.1649-1658
issn 2162-237X
2162-2388
language eng
recordid cdi_pascalfrancis_primary_26446530
source IEEE Xplore (Online service)
subjects Algorithms
Applied sciences
Artificial intelligence
Back propagation
Backpropagation
Computer science
control theory
systems
Computer Simulation
Connectionism. Neural networks
Context
Dynamical systems
Encoding
Exact sciences and technology
Feedback
Finite-state machines
hidden layer representation
Models, Statistical
Neural networks
Neural Networks (Computer)
Pattern Recognition, Automated - methods
penalty function
Recurrent neural networks
Studies
Symbolism
Training
Vectors
title Symbolic Representation of Recurrent Neural Network Dynamics
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T14%3A17%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pasca&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Symbolic%20Representation%20of%20Recurrent%20Neural%20Network%20Dynamics&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Huynh,%20T.%20Q.&rft.date=2012-10-01&rft.volume=23&rft.issue=10&rft.spage=1649&rft.epage=1658&rft.pages=1649-1658&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2012.2210242&rft_dat=%3Cproquest_pasca%3E1523404873%3C/proquest_pasca%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c414t-757ecc45440d6d0323c408351acacb5687473cb21b8b45a4b7741f2f1af7472f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1040776247&rft_id=info:pmid/24808009&rft_ieee_id=6269105&rfr_iscdi=true