| Keyword search (4,164 papers available) | ![]() |
"Reinforcement learning" Keyword-tagged Publications:
| Title | Authors | PubMed ID | |
|---|---|---|---|
| 1 | Disentangling prediction error and value in a formal test of dopamine s role in reinforcement learning | Usypchuk AA; Maes EJP; Lozzi M; Avramidis DK; Schoenbaum G; Esber GR; Gardner MPH; Iordanova MD; | 40738112 CSBN |
| 2 | Comprehensive review of reinforcement learning for medical ultrasound imaging | Elmekki H; Islam S; Alagha A; Sami H; Spilkin A; Zakeri E; Zanuttini AM; Bentahar J; Kadem L; Xie WF; Pibarot P; Mizouni R; Otrok H; Singh S; Mourad A; | 40567264 ENCS |
| 3 | Machine learning innovations in CPR: a comprehensive survey on enhanced resuscitation techniques | Islam S; Rjoub G; Elmekki H; Bentahar J; Pedrycz W; Cohen R; | 40336660 ENCS |
| 4 | Computational neuroscience across the lifespan: Promises and pitfalls | van den Bos W; Bruckner R; Nassar MR; Mata R; Eppinger B; | 29066078 PSYCHOLOGY |
| 5 | Does phasic dopamine release cause policy updates? | Carter F; Cossette MP; Trujillo-Pisanty I; Pallikaras V; Breton YA; Conover K; Caplan J; Solis P; Voisard J; Yaksich A; Shizgal P; | 38039083 PSYCHOLOGY |
| 6 | Nonlinear dynamic modeling and model-based AI-driven control of a magnetoactive soft continuum robot in a fluidic environment | Moezi SA; Sedaghati R; Rakheja S; | 37932207 ENCS |
| 7 | Sub-hourly measurement datasets from 6 real buildings: Energy use and indoor climate | Sartori I; Walnum HT; Skeie KS; Georges L; Knudsen MD; Bacher P; Candanedo J; Sigounis AM; Prakash AK; Pritoni M; Granderson J; Yang S; Wan MP; | 37153123 ENCS |
| 8 | Reinforcement learning for automatic quadrilateral mesh generation: A soft actor-critic approach | Pan J; Huang J; Cheng G; Zeng Y; | 36375347 ENCS |
| 9 | Trust-Augmented Deep Reinforcement Learning for Federated Learning Client Selection | Rjoub G; Wahab OA; Bentahar J; Cohen R; Bataineh AS; | 35875592 ENCS |
| 10 | Designing a hybrid reinforcement learning based algorithm with application in prediction of the COVID-19 pandemic in Quebec. | Khalilpourazari S, Hashemi Doulabi H | 33424076 ENCS |
| 11 | Cue-Evoked Dopamine Neuron Activity Helps Maintain but Does Not Encode Expected Value. | Mendoza JA, Lafferty CK, Yang AK, Britt JP | 31693885 CSBN |
| 12 | Metacontrol of decision-making strategies in human aging. | Bolenz F, Kool W, Reiter AM, Eppinger B | 31397670 PERFORM |
| 13 | Developmental Changes in Learning: Computational Mechanisms and Social Influences. | Bolenz F, Reiter AMF, Eppinger B | 29250006 PERFORM |
| Title: | Disentangling prediction error and value in a formal test of dopamine s role in reinforcement learning | ||||
| Authors: | Usypchuk AA, Maes EJP, Lozzi M, Avramidis DK, Schoenbaum G, Esber GR, Gardner MPH, Iordanova MD | ||||
| Link: | https://pubmed.ncbi.nlm.nih.gov/40738112/ | ||||
| DOI: | 10.1016/j.cub.2025.06.076 | ||||
| Publication: | Current biology : CB | ||||
| Keywords: | Rescorla-Wagner model; channelrhodopsin; error correction; mesolimbic; optogenetics; rodent; scalar value; temporal difference reinforcement learning; tyrosine hydrohylase; | ||||
| PMID: | 40738112 | Category: | Date Added: | 2025-07-31 | |
| Dept Affiliation: |
CSBN
1 Department of Psychology, Centre for Studies in Behavioural Neurobiology, Concordia University, Montreal, QC H4B 1R6, Canada. 2 NIDA Intramural Research Program, Baltimore, MD 21224, USA; Departments of Anatomy & Neurobiology and Psychiatry, University of Maryland School of Medicine, Baltimore, MD 21201, USA; Solomon H. Snyder Department of Neuroscience, the Johns Hopkins University, Baltimore, MD 21287, USA. 3 Department of Psychology, Centre for Studies in Behavioural Neurobiology, Concordia University, Montreal, QC H4B 1R6, Canada. Electronic address: mihaela.iordanova@concordia.ca. |
||||
Description: |
The discovery that midbrain dopamine (DA) transients can be mapped onto reward prediction errors (RPEs), the critical signal that drives learning, is a landmark in neuroscience. Causal support for the RPE hypothesis comes from studies showing that stimulating DA neurons can drive learning under conditions where it would not otherwise occur.1,2,3 However, such stimulation might also promote learning by adding reward value and indirectly inducing an RPE. This added value could support new learning even when it is insufficient to support instrumental behavior.4,5 Thus, these competing interpretations are challenging to disentangle and require direct comparison under matched conditions. We developed two computational models grounded in temporal difference reinforcement learning (TDRL)6,7,8 that dissociate the role of DA as an RPE versus a value signal. We validated our models by showing that they both predict learning (unblocking) when ventral tegmental area (VTA) DA stimulation occurs during expected reward delivery in a behavioral blocking design and confirmed this behaviorally. We then contrasted the models by delivering constant optogenetic stimulation during reward across both learning phases of blocking. The value model predicted blocking; the RPE model predicted unblocking. Behavioral results aligned with the latter. Moreover, the RPE model uniquely predicted that constant stimulation would unblock learning at higher frequencies (>20 Hz) when the artificial error alone drives learning. This, too, was confirmed experimentally. We demonstrate a principled computational and empirical dissociation between DA as an RPE versus a value signal. Our results advance understanding of how DA neuron stimulation drives learning. |



