Loading…

TD models of reward predictive responses in dopamine neurons

This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occ...

Full description

Saved in:
Bibliographic Details
Published in:Neural networks 2002-06, Vol.15 (4), p.523-533
Main Author: Suri, Roland E.
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203
cites cdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203
container_end_page 533
container_issue 4
container_start_page 523
container_title Neural networks
container_volume 15
creator Suri, Roland E.
description This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences. Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes.
doi_str_mv 10.1016/S0893-6080(02)00046-1
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_72161180</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0893608002000461</els_id><sourcerecordid>72161180</sourcerecordid><originalsourceid>FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</originalsourceid><addsrcrecordid>eNqFkE1LxDAQhoMouq7-BKUn0UN1Jm2TFASR9RMWPKjn0CZTiGw_TLor_nuju-jR0wwvzzsDD2NHCOcIKC6eQZVZKkDBKfAzAMhFiltsgkqWKZeKb7PJL7LH9kN4i5BQebbL9pBnEgsoJ-zy5SZpe0uLkPRN4umj8jYZPFlnRreimISh7wKFxHWJ7YeqdR0lHS19TA_YTlMtAh1u5pS93t2-zB7S-dP94-x6nppCiTGVVV7muRUGa1HbojF5VTdFJmops5qXOWSiKOIqea3QUC5UgVDaqpSAmeKQTdnJ-u7g-_clhVG3LhhaLKqO-mXQkqNAVP-DqASgQhXBYg0a34fgqdGDd23lPzWC_varf_zqb3kauP7xqzH2jjcPlnVL9q-1ERqBqzUQjdLKkdfBOOpMFOrJjNr27p8XX3NqiEs</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>18601818</pqid></control><display><type>article</type><title>TD models of reward predictive responses in dopamine neurons</title><source>Elsevier</source><creator>Suri, Roland E.</creator><creatorcontrib>Suri, Roland E.</creatorcontrib><description>This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences. Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/S0893-6080(02)00046-1</identifier><identifier>PMID: 12371509</identifier><language>eng</language><publisher>United States: Elsevier Ltd</publisher><subject>Animals ; Dopamine - physiology ; Forecasting - methods ; Humans ; Models, Biological ; Neuromodulation ; Neurons - physiology ; Planning ; Prediction ; Reinforcement ; Reward ; Sensorimotor ; Temporal difference</subject><ispartof>Neural networks, 2002-06, Vol.15 (4), p.523-533</ispartof><rights>2002 Elsevier Science Ltd</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</citedby><cites>FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/12371509$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Suri, Roland E.</creatorcontrib><title>TD models of reward predictive responses in dopamine neurons</title><title>Neural networks</title><addtitle>Neural Netw</addtitle><description>This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences. Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes.</description><subject>Animals</subject><subject>Dopamine - physiology</subject><subject>Forecasting - methods</subject><subject>Humans</subject><subject>Models, Biological</subject><subject>Neuromodulation</subject><subject>Neurons - physiology</subject><subject>Planning</subject><subject>Prediction</subject><subject>Reinforcement</subject><subject>Reward</subject><subject>Sensorimotor</subject><subject>Temporal difference</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2002</creationdate><recordtype>article</recordtype><recordid>eNqFkE1LxDAQhoMouq7-BKUn0UN1Jm2TFASR9RMWPKjn0CZTiGw_TLor_nuju-jR0wwvzzsDD2NHCOcIKC6eQZVZKkDBKfAzAMhFiltsgkqWKZeKb7PJL7LH9kN4i5BQebbL9pBnEgsoJ-zy5SZpe0uLkPRN4umj8jYZPFlnRreimISh7wKFxHWJ7YeqdR0lHS19TA_YTlMtAh1u5pS93t2-zB7S-dP94-x6nppCiTGVVV7muRUGa1HbojF5VTdFJmops5qXOWSiKOIqea3QUC5UgVDaqpSAmeKQTdnJ-u7g-_clhVG3LhhaLKqO-mXQkqNAVP-DqASgQhXBYg0a34fgqdGDd23lPzWC_varf_zqb3kauP7xqzH2jjcPlnVL9q-1ERqBqzUQjdLKkdfBOOpMFOrJjNr27p8XX3NqiEs</recordid><startdate>20020601</startdate><enddate>20020601</enddate><creator>Suri, Roland E.</creator><general>Elsevier Ltd</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QO</scope><scope>7TK</scope><scope>8FD</scope><scope>FR3</scope><scope>P64</scope><scope>7X8</scope></search><sort><creationdate>20020601</creationdate><title>TD models of reward predictive responses in dopamine neurons</title><author>Suri, Roland E.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2002</creationdate><topic>Animals</topic><topic>Dopamine - physiology</topic><topic>Forecasting - methods</topic><topic>Humans</topic><topic>Models, Biological</topic><topic>Neuromodulation</topic><topic>Neurons - physiology</topic><topic>Planning</topic><topic>Prediction</topic><topic>Reinforcement</topic><topic>Reward</topic><topic>Sensorimotor</topic><topic>Temporal difference</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Suri, Roland E.</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Biotechnology Research Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Suri, Roland E.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>TD models of reward predictive responses in dopamine neurons</atitle><jtitle>Neural networks</jtitle><addtitle>Neural Netw</addtitle><date>2002-06-01</date><risdate>2002</risdate><volume>15</volume><issue>4</issue><spage>523</spage><epage>533</epage><pages>523-533</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>This article focuses on recent modeling studies of dopamine neuron activity and their influence on behavior. Activity of midbrain dopamine neurons is phasically increased by stimuli that increase the animal's reward expectation and is decreased below baseline levels when the reward fails to occur. These characteristics resemble the reward prediction error signal of the temporal difference (TD) model, which is a model of reinforcement learning. Computational modeling studies show that such a dopamine-like reward prediction error can serve as a powerful teaching signal for learning with delayed reinforcement, in particular for learning of motor sequences. Several lines of evidence suggest that dopamine is also involved in ‘cognitive’ processes that are not addressed by standard TD models. I propose the hypothesis that dopamine neuron activity is crucial for planning processes, also referred to as ‘goal-directed behavior’, which select actions by evaluating predictions about their motivational outcomes.</abstract><cop>United States</cop><pub>Elsevier Ltd</pub><pmid>12371509</pmid><doi>10.1016/S0893-6080(02)00046-1</doi><tpages>11</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0893-6080
ispartof Neural networks, 2002-06, Vol.15 (4), p.523-533
issn 0893-6080
1879-2782
language eng
recordid cdi_proquest_miscellaneous_72161180
source Elsevier
subjects Animals
Dopamine - physiology
Forecasting - methods
Humans
Models, Biological
Neuromodulation
Neurons - physiology
Planning
Prediction
Reinforcement
Reward
Sensorimotor
Temporal difference
title TD models of reward predictive responses in dopamine neurons
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T16%3A09%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=TD%20models%20of%20reward%20predictive%20responses%20in%20dopamine%20neurons&rft.jtitle=Neural%20networks&rft.au=Suri,%20Roland%20E.&rft.date=2002-06-01&rft.volume=15&rft.issue=4&rft.spage=523&rft.epage=533&rft.pages=523-533&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/S0893-6080(02)00046-1&rft_dat=%3Cproquest_cross%3E72161180%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c586t-7a4944d6c1b6bd5fc4abf536b773b2940365573b72b81ce4685109da970138203%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=18601818&rft_id=info:pmid/12371509&rfr_iscdi=true