Loading…
TD models of reward predictive responses in dopamine neurons
This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occ...
Saved in:
Published in: | Neural networks 2002-06, Vol.15 (4), p.523-533 |
---|---|
Main Author: | |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203 |
---|---|
cites | cdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203 |
container_end_page | 533 |
container_issue | 4 |
container_start_page | 523 |
container_title | Neural networks |
container_volume | 15 |
creator | Suri, Roland E. |
description | This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences.
Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes. |
doi_str_mv | 10.1016/S0893-6080(02)00046-1 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_72161180</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0893608002000461</els_id><sourcerecordid>72161180</sourcerecordid><originalsourceid>FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</originalsourceid><addsrcrecordid>eNqFkE1LxDAQhoMouq7-BKUn0UN1Jm2TFASR9RMWPKjn0CZTiGw_TLor_nuju-jR0wwvzzsDD2NHCOcIKC6eQZVZKkDBKfAzAMhFiltsgkqWKZeKb7PJL7LH9kN4i5BQebbL9pBnEgsoJ-zy5SZpe0uLkPRN4umj8jYZPFlnRreimISh7wKFxHWJ7YeqdR0lHS19TA_YTlMtAh1u5pS93t2-zB7S-dP94-x6nppCiTGVVV7muRUGa1HbojF5VTdFJmops5qXOWSiKOIqea3QUC5UgVDaqpSAmeKQTdnJ-u7g-_clhVG3LhhaLKqO-mXQkqNAVP-DqASgQhXBYg0a34fgqdGDd23lPzWC_varf_zqb3kauP7xqzH2jjcPlnVL9q-1ERqBqzUQjdLKkdfBOOpMFOrJjNr27p8XX3NqiEs</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>18601818</pqid></control><display><type>article</type><title>TD models of reward predictive responses in dopamine neurons</title><source>Elsevier</source><creator>Suri, Roland E.</creator><creatorcontrib>Suri, Roland E.</creatorcontrib><description>This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences.
Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/S0893-6080(02)00046-1</identifier><identifier>PMID: 12371509</identifier><language>eng</language><publisher>United States: Elsevier Ltd</publisher><subject>Animals ; Dopamine - physiology ; Forecasting - methods ; Humans ; Models, Biological ; Neuromodulation ; Neurons - physiology ; Planning ; Prediction ; Reinforcement ; Reward ; Sensorimotor ; Temporal difference</subject><ispartof>Neural networks, 2002-06, Vol.15 (4), p.523-533</ispartof><rights>2002 Elsevier Science Ltd</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</citedby><cites>FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/12371509$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Suri, Roland E.</creatorcontrib><title>TD models of reward predictive responses in dopamine neurons</title><title>Neural networks</title><addtitle>Neural Netw</addtitle><description>This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences.
Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes.</description><subject>Animals</subject><subject>Dopamine - physiology</subject><subject>Forecasting - methods</subject><subject>Humans</subject><subject>Models, Biological</subject><subject>Neuromodulation</subject><subject>Neurons - physiology</subject><subject>Planning</subject><subject>Prediction</subject><subject>Reinforcement</subject><subject>Reward</subject><subject>Sensorimotor</subject><subject>Temporal difference</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2002</creationdate><recordtype>article</recordtype><recordid>eNqFkE1LxDAQhoMouq7-BKUn0UN1Jm2TFASR9RMWPKjn0CZTiGw_TLor_nuju-jR0wwvzzsDD2NHCOcIKC6eQZVZKkDBKfAzAMhFiltsgkqWKZeKb7PJL7LH9kN4i5BQebbL9pBnEgsoJ-zy5SZpe0uLkPRN4umj8jYZPFlnRreimISh7wKFxHWJ7YeqdR0lHS19TA_YTlMtAh1u5pS93t2-zB7S-dP94-x6nppCiTGVVV7muRUGa1HbojF5VTdFJmops5qXOWSiKOIqea3QUC5UgVDaqpSAmeKQTdnJ-u7g-_clhVG3LhhaLKqO-mXQkqNAVP-DqASgQhXBYg0a34fgqdGDd23lPzWC_varf_zqb3kauP7xqzH2jjcPlnVL9q-1ERqBqzUQjdLKkdfBOOpMFOrJjNr27p8XX3NqiEs</recordid><startdate>20020601</startdate><enddate>20020601</enddate><creator>Suri, Roland E.</creator><general>Elsevier Ltd</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QO</scope><scope>7TK</scope><scope>8FD</scope><scope>FR3</scope><scope>P64</scope><scope>7X8</scope></search><sort><creationdate>20020601</creationdate><title>TD models of reward predictive responses in dopamine neurons</title><author>Suri, Roland E.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2002</creationdate><topic>Animals</topic><topic>Dopamine - physiology</topic><topic>Forecasting - methods</topic><topic>Humans</topic><topic>Models, Biological</topic><topic>Neuromodulation</topic><topic>Neurons - physiology</topic><topic>Planning</topic><topic>Prediction</topic><topic>Reinforcement</topic><topic>Reward</topic><topic>Sensorimotor</topic><topic>Temporal difference</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Suri, Roland E.</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Biotechnology Research Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Suri, Roland E.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>TD models of reward predictive responses in dopamine neurons</atitle><jtitle>Neural networks</jtitle><addtitle>Neural Netw</addtitle><date>2002-06-01</date><risdate>2002</risdate><volume>15</volume><issue>4</issue><spage>523</spage><epage>533</epage><pages>523-533</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences.
Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes.</abstract><cop>United States</cop><pub>Elsevier Ltd</pub><pmid>12371509</pmid><doi>10.1016/S0893-6080(02)00046-1</doi><tpages>11</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0893-6080 |
ispartof | Neural networks, 2002-06, Vol.15 (4), p.523-533 |
issn | 0893-6080 1879-2782 |
language | eng |
recordid | cdi_proquest_miscellaneous_72161180 |
source | Elsevier |
subjects | Animals Dopamine - physiology Forecasting - methods Humans Models, Biological Neuromodulation Neurons - physiology Planning Prediction Reinforcement Reward Sensorimotor Temporal difference |
title | TD models of reward predictive responses in dopamine neurons |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T16%3A09%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=TD%20models%20of%20reward%20predictive%20responses%20in%20dopamine%20neurons&rft.jtitle=Neural%20networks&rft.au=Suri,%20Roland%20E.&rft.date=2002-06-01&rft.volume=15&rft.issue=4&rft.spage=523&rft.epage=533&rft.pages=523-533&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/S0893-6080(02)00046-1&rft_dat=%3Cproquest_cross%3E72161180%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=18601818&rft_id=info:pmid/12371509&rfr_iscdi=true |