1
|
Arumugam D, Ho MK, Goodman ND, Van Roy B. Bayesian Reinforcement Learning With Limited Cognitive Load. Open Mind (Camb) 2024; 8:395-438. [PMID: 38665544 PMCID: PMC11045037 DOI: 10.1162/opmi_a_00132] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2023] [Accepted: 02/16/2024] [Indexed: 04/28/2024] Open
Abstract
All biological and artificial agents must act given limits on their ability to acquire and process information. As such, a general theory of adaptive behavior should be able to account for the complex interactions between an agent's learning history, decisions, and capacity constraints. Recent work in computer science has begun to clarify the principles that shape these dynamics by bridging ideas from reinforcement learning, Bayesian decision-making, and rate-distortion theory. This body of work provides an account of capacity-limited Bayesian reinforcement learning, a unifying normative framework for modeling the effect of processing constraints on learning and action selection. Here, we provide an accessible review of recent algorithms and theoretical results in this setting, paying special attention to how these ideas can be applied to studying questions in the cognitive and behavioral sciences.
Collapse
Affiliation(s)
| | - Mark K. Ho
- Center for Data Science, New York University
| | - Noah D. Goodman
- Department of Computer Science, Stanford University
- Department of Psychology, Stanford University
| | - Benjamin Van Roy
- Department of Electrical Engineering, Stanford University
- Department of Management Science & Engineering, Stanford University
| |
Collapse
|
2
|
Eshel N, Touponse GC, Wang AR, Osterman AK, Shank AN, Groome AM, Taniguchi L, Cardozo Pinto DF, Tucciarone J, Bentzley BS, Malenka RC. Striatal dopamine integrates cost, benefit, and motivation. Neuron 2024; 112:500-514.e5. [PMID: 38016471 PMCID: PMC10922131 DOI: 10.1016/j.neuron.2023.10.038] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Revised: 07/06/2023] [Accepted: 10/26/2023] [Indexed: 11/30/2023]
Abstract
Striatal dopamine (DA) release has long been linked to reward processing, but it remains controversial whether DA release reflects costs or benefits and how these signals vary with motivation. Here, we measure DA release in the nucleus accumbens (NAc) and dorsolateral striatum (DLS) while independently varying costs and benefits and apply behavioral economic principles to determine a mouse's level of motivation. We reveal that DA release in both structures incorporates both reward magnitude and sunk cost. Surprisingly, motivation was inversely correlated with reward-evoked DA release. Furthermore, optogenetically evoked DA release was also heavily dependent on sunk cost. Our results reconcile previous disparate findings by demonstrating that striatal DA release simultaneously encodes cost, benefit, and motivation but in distinct manners over different timescales. Future work will be necessary to determine whether the reduction in phasic DA release in highly motivated animals is due to changes in tonic DA levels.
Collapse
Affiliation(s)
- Neir Eshel
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA.
| | - Gavin C Touponse
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Allan R Wang
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Amber K Osterman
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Amei N Shank
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Alexandra M Groome
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Lara Taniguchi
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Daniel F Cardozo Pinto
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Jason Tucciarone
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Brandon S Bentzley
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA
| | - Robert C Malenka
- Nancy Pritzker Laboratory, Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, CA, USA.
| |
Collapse
|
3
|
Pinto SR, Uchida N. Tonic dopamine and biases in value learning linked through a biologically inspired reinforcement learning model. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.11.10.566580. [PMID: 38014087 PMCID: PMC10680794 DOI: 10.1101/2023.11.10.566580] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/29/2023]
Abstract
A hallmark of various psychiatric disorders is biased future predictions. Here we examined the mechanisms for biased value learning using reinforcement learning models incorporating recent findings on synaptic plasticity and opponent circuit mechanisms in the basal ganglia. We show that variations in tonic dopamine can alter the balance between learning from positive and negative reward prediction errors, leading to biased value predictions. This bias arises from the sigmoidal shapes of the dose-occupancy curves and distinct affinities of D1- and D2-type dopamine receptors: changes in tonic dopamine differentially alters the slope of the dose-occupancy curves of these receptors, thus sensitivities, at baseline dopamine concentrations. We show that this mechanism can explain biased value learning in both mice and humans and may also contribute to symptoms observed in psychiatric disorders. Our model provides a foundation for understanding the basal ganglia circuit and underscores the significance of tonic dopamine in modulating learning processes.
Collapse
Affiliation(s)
- Sandra Romero Pinto
- Department of Molecular and Cellular Biology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
- Program in Speech and Hearing Bioscience and Technology, Division of Medical Sciences, Harvard Medical School, Boston, MA 02115, USA
| | - Naoshige Uchida
- Department of Molecular and Cellular Biology, Center for Brain Science, Harvard University, Cambridge, MA 02138, USA
| |
Collapse
|
4
|
Chakroun K, Wiehler A, Wagner B, Mathar D, Ganzer F, van Eimeren T, Sommer T, Peters J. Dopamine regulates decision thresholds in human reinforcement learning in males. Nat Commun 2023; 14:5369. [PMID: 37666865 PMCID: PMC10477234 DOI: 10.1038/s41467-023-41130-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2022] [Accepted: 08/22/2023] [Indexed: 09/06/2023] Open
Abstract
Dopamine fundamentally contributes to reinforcement learning, but recent accounts also suggest a contribution to specific action selection mechanisms and the regulation of response vigour. Here, we examine dopaminergic mechanisms underlying human reinforcement learning and action selection via a combined pharmacological neuroimaging approach in male human volunteers (n = 31, within-subjects; Placebo, 150 mg of the dopamine precursor L-dopa, 2 mg of the D2 receptor antagonist Haloperidol). We found little credible evidence for previously reported beneficial effects of L-dopa vs. Haloperidol on learning from gains and altered neural prediction error signals, which may be partly due to differences experimental design and/or drug dosages. Reinforcement learning drift diffusion models account for learning-related changes in accuracy and response times, and reveal consistent decision threshold reductions under both drugs, in line with the idea that lower dosages of D2 receptor antagonists increase striatal DA release via an autoreceptor-mediated feedback mechanism. These results are in line with the idea that dopamine regulates decision thresholds during reinforcement learning, and may help to bridge action selection and response vigor accounts of dopamine.
Collapse
Affiliation(s)
- Karima Chakroun
- Institute for Systems Neuroscience, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| | - Antonius Wiehler
- Motivation, Brain and Behavior Lab, Paris Brain Institute (ICM), Pitié-Salpêtrière Hospital, Paris, France
| | - Ben Wagner
- Chair of Cognitive Computational Neuroscience, Technical University Dresden, Dresden, Germany
| | - David Mathar
- Department of Psychology, Biological Psychology, University of Cologne, Cologne, Germany
| | - Florian Ganzer
- Integrated Psychiatry Winterthur, Winterthur, Switzerland
| | - Thilo van Eimeren
- Multimodal Neuroimaging Group, Department of Nuclear Medicine, University Medical Center Cologne, Cologne, Germany
| | - Tobias Sommer
- Institute for Systems Neuroscience, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| | - Jan Peters
- Institute for Systems Neuroscience, University Medical Center Hamburg-Eppendorf, Hamburg, Germany.
- Department of Psychology, Biological Psychology, University of Cologne, Cologne, Germany.
| |
Collapse
|
5
|
Undermatching Is a Consequence of Policy Compression. J Neurosci 2023; 43:447-457. [PMID: 36639891 PMCID: PMC9864556 DOI: 10.1523/jneurosci.1003-22.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2022] [Revised: 10/14/2022] [Accepted: 11/17/2022] [Indexed: 12/12/2022] Open
Abstract
The matching law describes the tendency of agents to match the ratio of choices allocated to the ratio of rewards received when choosing among multiple options (Herrnstein, 1961). Perfect matching, however, is infrequently observed. Instead, agents tend to undermatch or bias choices toward the poorer option. Overmatching, or the tendency to bias choices toward the richer option, is rarely observed. Despite the ubiquity of undermatching, it has received an inadequate normative justification. Here, we assume agents not only seek to maximize reward, but also seek to minimize cognitive cost, which we formalize as policy complexity (the mutual information between actions and states of the environment). Policy complexity measures the extent to which the policy of an agent is state dependent. Our theory states that capacity-constrained agents (i.e., agents that must compress their policies to reduce complexity) can only undermatch or perfectly match, but not overmatch, consistent with the empirical evidence. Moreover, using mouse behavioral data (male), we validate a novel prediction about which task conditions exaggerate undermatching. Finally, in patients with Parkinson's disease (male and female), we argue that a reduction in undermatching with higher dopamine levels is consistent with an increased policy complexity.SIGNIFICANCE STATEMENT The matching law describes the tendency of agents to match the ratio of choices allocated to different options to the ratio of reward received. For example, if option a yields twice as much reward as option b, matching states that agents will choose option a twice as much. However, agents typically undermatch: they choose the poorer option more frequently than expected. Here, we assume that agents seek to simultaneously maximize reward and minimize the complexity of their action policies. We show that this theory explains when and why undermatching occurs. Neurally, we show that policy complexity, and by extension undermatching, is controlled by tonic dopamine, consistent with other evidence that dopamine plays an important role in cognitive resource allocation.
Collapse
|
6
|
Chebolu S, Dayan P, Lloyd K. Vigilance, arousal, and acetylcholine: Optimal control of attention in a simple detection task. PLoS Comput Biol 2022; 18:e1010642. [PMID: 36315594 PMCID: PMC9648841 DOI: 10.1371/journal.pcbi.1010642] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2022] [Revised: 11/10/2022] [Accepted: 10/06/2022] [Indexed: 11/05/2022] Open
Abstract
Paying attention to particular aspects of the world or being more vigilant in general can be interpreted as forms of ‘internal’ action. Such arousal-related choices come with the benefit of increasing the quality and situational appropriateness of information acquisition and processing, but incur potentially expensive energetic and opportunity costs. One implementational route for these choices is widespread ascending neuromodulation, including by acetylcholine (ACh). The key computational question that elective attention poses for sensory processing is when it is worthwhile paying these costs, and this includes consideration of whether sufficient information has yet been collected to justify the higher signal-to-noise ratio afforded by greater attention and, particularly if a change in attentional state is more expensive than its maintenance, when states of heightened attention ought to persist. We offer a partially observable Markov decision-process treatment of optional attention in a detection task, and use it to provide a qualitative model of the results of studies using modern techniques to measure and manipulate ACh in rodents performing a similar task. Paying attention to a stimulus is costly, both in terms of energy and the lost opportunity to pay attention to something else. It is also beneficial, providing more information about its target. Thus, whether and when we pay more or less attention may best be considered as a choice of internal action that responds to this trade-off. Furthermore, measurements and manipulation of the neuromodulator acetylcholine have suggested that it is one of the instruments of attention, providing us with a window onto this choice. Here, we build an abstract model of a task in which an animal must look out for a brief visual stimulus that may or may not occur on each trial. We show that optimal attentional choices in the model depend on many factors, including how likely a signal is to occur across time, the balance between the improvement in information possible by paying greater attention and its increased cost, and whether there are also costs associated with switching between different attentional states. We also show that our model can qualitatively match results from experiments involving acetylcholine.
Collapse
Affiliation(s)
- Sahiti Chebolu
- Graduate Training Centre of Neuroscience, International Max Planck Research School, Tübingen, Germany
- Indian Institute of Science Education and Research Pune, India
| | - Peter Dayan
- Department for Computational Neuroscience, Max Planck Institute for Biological Cybernetics, Tübingen, Germany
- University of Tübingen, Tübingen, Germany
| | - Kevin Lloyd
- Department for Computational Neuroscience, Max Planck Institute for Biological Cybernetics, Tübingen, Germany
- * E-mail:
| |
Collapse
|
7
|
Hofmans L, Westbrook A, van den Bosch R, Booij J, Verkes RJ, Cools R. Effects of average reward rate on vigor as a function of individual variation in striatal dopamine. Psychopharmacology (Berl) 2022; 239:465-478. [PMID: 34735591 DOI: 10.1007/s00213-021-06017-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/30/2021] [Accepted: 10/15/2021] [Indexed: 11/24/2022]
Abstract
RATIONALE We constantly need to decide not only which actions to perform, but also how vigorously to perform them. In agreement with an earlier theoretical model, it has been shown that a significant portion of the variance in our action vigor can be explained by the average rate of rewards received for that action. Moreover, this invigorating effect of average reward rate was shown to vary with within-subject changes in dopamine, both in human individuals and experimental rodents. OBJECTIVES Here, we assessed whether individual differences in the effect of average reward rate on vigor are related to individual variation in a stable measure of striatal dopamine function in healthy, unmedicated participants. METHODS Forty-four participants performed a discrimination task to test the effect of average reward rate on response times to index vigor and completed an [18F]-DOPA PET scan to index striatal dopamine synthesis capacity. RESULTS We did not find an interaction between dopamine synthesis capacity and average reward rate across the entire group. However, a post hoc analysis revealed that participants with higher striatal dopamine synthesis capacity, particularly in the nucleus accumbens, exhibited a stronger invigorating effect of average reward rate among the 30 slowest participants. CONCLUSIONS Our findings provide converging evidence for a role of striatal dopamine in average reward rate signaling, thereby extending the current literature on the mechanistic link between average reward rate, vigor, and dopamine.
Collapse
Affiliation(s)
- Lieke Hofmans
- Donders Institute for Brain, Cognition & Behaviour, Radboud University, Nijmegen, The Netherlands. .,Department of Psychiatry, Radboudumc, Nijmegen, The Netherlands. .,Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands.
| | - Andrew Westbrook
- Donders Institute for Brain, Cognition & Behaviour, Radboud University, Nijmegen, The Netherlands.,Department of Psychiatry, Radboudumc, Nijmegen, The Netherlands.,Department of Cognitive, Linguistics and Psychological Sciences, Brown University, Providence, USA
| | - Ruben van den Bosch
- Donders Institute for Brain, Cognition & Behaviour, Radboud University, Nijmegen, The Netherlands.,Department of Psychiatry, Radboudumc, Nijmegen, The Netherlands
| | - Jan Booij
- Department of Medical Imaging, Radboudumc, Nijmegen, The Netherlands.,Department of Radiology and Nuclear Medicine, Amsterdam University Medical Centers, location Academic Medical Center, University of Amsterdam, Amsterdam, The Netherlands
| | - Robbert-Jan Verkes
- Department of Psychiatry, Radboudumc, Nijmegen, The Netherlands.,Forensic Psychiatric Centre Nijmegen, Pompestichting, Nijmegen, The Netherlands.,Department of Criminal Law, Law School, Radboud Universiteit, Nijmegen, The Netherlands
| | - Roshan Cools
- Donders Institute for Brain, Cognition & Behaviour, Radboud University, Nijmegen, The Netherlands.,Department of Psychiatry, Radboudumc, Nijmegen, The Netherlands
| |
Collapse
|
8
|
Dopamine firing plays a dual role in coding reward prediction errors and signaling motivation in a working memory task. Proc Natl Acad Sci U S A 2022; 119:2113311119. [PMID: 34992139 PMCID: PMC8764687 DOI: 10.1073/pnas.2113311119] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/29/2021] [Indexed: 11/21/2022] Open
Abstract
Little is known about how dopamine (DA) neuron firing rates behave in cognitively demanding decision-making tasks. Here, we investigated midbrain DA activity in monkeys performing a discrimination task in which the animal had to use working memory (WM) to report which of two sequentially applied vibrotactile stimuli had the higher frequency. We found that perception was altered by an internal bias, likely generated by deterioration of the representation of the first frequency during the WM period. This bias greatly controlled the DA phasic response during the two stimulation periods, confirming that DA reward prediction errors reflected stimulus perception. In contrast, tonic dopamine activity during WM was not affected by the bias and did not encode the stored frequency. More interestingly, both delay-period activity and phasic responses before the second stimulus negatively correlated with reaction times of the animals after the trial start cue and thus represented motivated behavior on a trial-by-trial basis. During WM, this motivation signal underwent a ramp-like increase. At the same time, motivation positively correlated with accuracy, especially in difficult trials, probably by decreasing the effect of the bias. Overall, our results indicate that DA activity, in addition to encoding reward prediction errors, could at the same time be involved in motivation and WM. In particular, the ramping activity during the delay period suggests a possible DA role in stabilizing sustained cortical activity, hypothetically by increasing the gain communicated to prefrontal neurons in a motivation-dependent way.
Collapse
|
9
|
Mikhael JG, Gershman SJ. Impulsivity and risk-seeking as Bayesian inference under dopaminergic control. Neuropsychopharmacology 2022; 47:465-476. [PMID: 34376813 PMCID: PMC8674258 DOI: 10.1038/s41386-021-01125-z] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/01/2020] [Revised: 07/17/2021] [Accepted: 07/21/2021] [Indexed: 02/07/2023]
Abstract
Bayesian models successfully account for several of dopamine (DA)'s effects on contextual calibration in interval timing and reward estimation. In these models, tonic levels of DA control the precision of stimulus encoding, which is weighed against contextual information when making decisions. When DA levels are high, the animal relies more heavily on the (highly precise) stimulus encoding, whereas when DA levels are low, the context affects decisions more strongly. Here, we extend this idea to intertemporal choice and probability discounting tasks. In intertemporal choice tasks, agents must choose between a small reward delivered soon and a large reward delivered later, whereas in probability discounting tasks, agents must choose between a small reward that is always delivered and a large reward that may be omitted with some probability. Beginning with the principle that animals will seek to maximize their reward rates, we show that the Bayesian model predicts a number of curious empirical findings in both tasks. First, the model predicts that higher DA levels should normally promote selection of the larger/later option, which is often taken to imply that DA decreases 'impulsivity,' and promote selection of the large/risky option, often taken to imply that DA increases 'risk-seeking.' However, if the temporal precision is sufficiently decreased, higher DA levels should have the opposite effect-promoting selection of the smaller/sooner option (higher impulsivity) and the small/safe option (lower risk-seeking). Second, high enough levels of DA can result in preference reversals. Third, selectively decreasing the temporal precision, without manipulating DA, should promote selection of the larger/later and large/risky options. Fourth, when a different post-reward delay is associated with each option, animals will not learn the option-delay contingencies, but this learning can be salvaged when the post-reward delays are made more salient. Finally, the Bayesian model predicts correlations among behavioral phenotypes: Animals that are better timers will also appear less impulsive.
Collapse
Affiliation(s)
- John G. Mikhael
- grid.38142.3c000000041936754XProgram in Neuroscience, Harvard Medical School, Boston, MA USA ,grid.38142.3c000000041936754XMD-PhD Program, Harvard Medical School, Boston, MA USA
| | - Samuel J. Gershman
- grid.38142.3c000000041936754XDepartment of Psychology and Center for Brain Science, Harvard University, Cambridge, MA USA ,grid.116068.80000 0001 2341 2786Center for Brains, Minds and Machines, Massachusetts Institute of Technology, Cambridge, MA USA
| |
Collapse
|
10
|
Westbrook A, Ghosh A, van den Bosch R, Määttä JI, Hofmans L, Cools R. Striatal dopamine synthesis capacity reflects smartphone social activity. iScience 2021; 24:102497. [PMID: 34113831 PMCID: PMC8170001 DOI: 10.1016/j.isci.2021.102497] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2020] [Revised: 03/01/2021] [Accepted: 04/28/2021] [Indexed: 01/15/2023] Open
Abstract
Striatal dopamine and smartphone behavior have both been linked with behavioral variability. Here, we leverage day-to-day logs of natural, unconstrained smartphone behavior and establish a correlation between a measure of smartphone social activity previously linked with behavioral variability and a measure of striatal dopamine synthesis capacity using [18F]-DOPA PET in (N = 22) healthy adult humans. Specifically, we find that a higher proportion of social app interactions correlates with lower dopamine synthesis capacity in the bilateral putamen. Permutation tests and penalized regressions provide evidence that this link between dopamine synthesis capacity and social versus non-social smartphone interactions is specific. These observations provide a key empirical grounding for current speculations about dopamine's role in digital social behavior. Putamen dopamine synthesis capacity correlates with smartphone social app use. The correlation parallels a prior link between social app use and motor variability. It is selective to social app use, controlling for multiple smartphone use factors.
Collapse
Affiliation(s)
- Andrew Westbrook
- Radboud University Medical Centre, Department of Psychiatry, Nijmegen 6525 GA, The Netherlands.,Radboud University, Donders Institute for Brain, Cognition and Behaviour, Nijmegen 6525 EN, The Netherlands.,Department of Cognitive, Linguistic & Psychological Sciences, Brown University, Providence, RI 02912, USA
| | - Arko Ghosh
- Institute of Psychology, Cognitive Psychology Unit, Leiden University, Leiden 2333 AK, The Netherlands
| | - Ruben van den Bosch
- Radboud University Medical Centre, Department of Psychiatry, Nijmegen 6525 GA, The Netherlands.,Radboud University, Donders Institute for Brain, Cognition and Behaviour, Nijmegen 6525 EN, The Netherlands
| | - Jessica I Määttä
- Radboud University Medical Centre, Department of Psychiatry, Nijmegen 6525 GA, The Netherlands.,Radboud University, Donders Institute for Brain, Cognition and Behaviour, Nijmegen 6525 EN, The Netherlands
| | - Lieke Hofmans
- Radboud University Medical Centre, Department of Psychiatry, Nijmegen 6525 GA, The Netherlands.,Radboud University, Donders Institute for Brain, Cognition and Behaviour, Nijmegen 6525 EN, The Netherlands
| | - Roshan Cools
- Radboud University Medical Centre, Department of Psychiatry, Nijmegen 6525 GA, The Netherlands.,Radboud University, Donders Institute for Brain, Cognition and Behaviour, Nijmegen 6525 EN, The Netherlands
| |
Collapse
|