1
|
Wojtak W, Coombes S, Avitabile D, Bicho E, Erlhagen W. A dynamic neural field model of continuous input integration. BIOLOGICAL CYBERNETICS 2021; 115:451-471. [PMID: 34417880 DOI: 10.1007/s00422-021-00893-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Accepted: 08/01/2021] [Indexed: 06/13/2023]
Abstract
The ability of neural systems to turn transient inputs into persistent changes in activity is thought to be a fundamental requirement for higher cognitive functions. In continuous attractor networks frequently used to model working memory or decision making tasks, the persistent activity settles to a stable pattern with the stereotyped shape of a "bump" independent of integration time or input strength. Here, we investigate a new bump attractor model in which the bump width and amplitude not only reflect qualitative and quantitative characteristics of a preceding input but also the continuous integration of evidence over longer timescales. The model is formalized by two coupled dynamic field equations of Amari-type which combine recurrent interactions mediated by a Mexican-hat connectivity with local feedback mechanisms that balance excitation and inhibition. We analyze the existence, stability and bifurcation structure of single and multi-bump solutions and discuss the relevance of their input dependence to modeling cognitive functions. We then systematically compare the pattern formation process of the two-field model with the classical Amari model. The results reveal that the balanced local feedback mechanisms facilitate the encoding and maintenance of multi-item memories. The existence of stable subthreshold bumps suggests that different to the Amari model, the suppression effect of neighboring bumps in the range of lateral competition may not lead to a complete loss of information. Moreover, bumps with larger amplitude are less vulnerable to noise-induced drifts and distance-dependent interaction effects resulting in more faithful memory representations over time.
Collapse
Affiliation(s)
- Weronika Wojtak
- Research Centre of Mathematics, University of Minho, Guimarães, Portugal.
- Research Centre Algoritmi, University of Minho, Guimarães, Portugal.
| | - Stephen Coombes
- Centre for Mathematical Medicine and Biology, School of Mathematical Sciences, University of Nottingham, Nottingham, UK
| | - Daniele Avitabile
- Department of Mathematics, Vrije Universiteit, Amsterdam, The Netherlands
- MathNeuro Team, Inria Sophia Antipolis Méditerranée Research Centre, Sophia Antipolis, France
| | - Estela Bicho
- Research Centre Algoritmi, University of Minho, Guimarães, Portugal
| | - Wolfram Erlhagen
- Research Centre of Mathematics, University of Minho, Guimarães, Portugal
| |
Collapse
|
2
|
Wojtak W, Ferreira F, Vicente P, Louro L, Bicho E, Erlhagen W. A neural integrator model for planning and value-based decision making of a robotics assistant. Neural Comput Appl 2021. [DOI: 10.1007/s00521-020-05224-8] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
3
|
Zschache J. Applying the matching law as micro-foundation of social phenomena. SOCIAL SCIENCE RESEARCH 2018; 73:189-206. [PMID: 29793686 DOI: 10.1016/j.ssresearch.2018.03.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2017] [Revised: 03/19/2018] [Accepted: 03/20/2018] [Indexed: 06/08/2023]
Abstract
Social phenomena are suggested to be explained by the matching law - an empirical regularity of individual behaviour. While a considerable amount of psychological research on this law exists, only a few sociological applications can be found. This paper points to the problems that come with its usage as micro-foundation of social behaviour and provides solutions. In particular, a model of melioration learning enables the derivation of social phenomena from the matching law. The proposed approach is illustrated by the application of the learning model to the volunteer's dilemma. In contrast to game-theoretical solutions, the matching law leads to more intuitive results in case of the asymmetric dilemma. The relationship between the matching law and utility maximisation is discussed by its integration into economic consumer theory.
Collapse
Affiliation(s)
- Johannes Zschache
- Institute of Sociology, Leipzig University, Beethovenstraße 15, 04107, Leipzig, Germany.
| |
Collapse
|
4
|
Abstract
Melioration learning is an empirically well-grounded model of reinforcement learning. By means of computer simulations, this paper derives predictions for several repeatedly played two-person games from this model. The results indicate a likely convergence to a pure Nash equilibrium of the game. If no pure equilibrium exists, the relative frequencies of choice may approach the predictions of the mixed Nash equilibrium. Yet in some games, no stable state is reached.
Collapse
Affiliation(s)
- Johannes Zschache
- Institute of Sociology, Leipzig University, Leipzig, Germany
- * E-mail:
| |
Collapse
|
5
|
Jensen G, Ward RD, Balsam PD. Information: theory, brain, and behavior. J Exp Anal Behav 2013; 100:408-31. [PMID: 24122456 DOI: 10.1002/jeab.49] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2013] [Accepted: 08/26/2013] [Indexed: 01/15/2023]
Abstract
In the 65 years since its formal specification, information theory has become an established statistical paradigm, providing powerful tools for quantifying probabilistic relationships. Behavior analysis has begun to adopt these tools as a novel means of measuring the interrelations between behavior, stimuli, and contingent outcomes. This approach holds great promise for making more precise determinations about the causes of behavior and the forms in which conditioning may be encoded by organisms. In addition to providing an introduction to the basics of information theory, we review some of the ways that information theory has informed the studies of Pavlovian conditioning, operant conditioning, and behavioral neuroscience. In addition to enriching each of these empirical domains, information theory has the potential to act as a common statistical framework by which results from different domains may be integrated, compared, and ultimately unified.
Collapse
|
6
|
Accuracy and response-time distributions for decision-making: linear perfect integrators versus nonlinear attractor-based neural circuits. J Comput Neurosci 2013; 35:261-94. [PMID: 23608921 PMCID: PMC3825033 DOI: 10.1007/s10827-013-0452-x] [Citation(s) in RCA: 29] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2012] [Revised: 03/25/2013] [Accepted: 03/27/2013] [Indexed: 12/31/2022]
Abstract
Animals choose actions based on imperfect, ambiguous data. “Noise” inherent in neural processing adds further variability to this already-noisy input signal. Mathematical analysis has suggested that the optimal apparatus (in terms of the speed/accuracy trade-off) for reaching decisions about such noisy inputs is perfect accumulation of the inputs by a temporal integrator. Thus, most highly cited models of neural circuitry underlying decision-making have been instantiations of a perfect integrator. Here, in accordance with a growing mathematical and empirical literature, we describe circumstances in which perfect integration is rendered suboptimal. In particular we highlight the impact of three biological constraints: (1) significant noise arising within the decision-making circuitry itself; (2) bounding of integration by maximal neural firing rates; and (3) time limitations on making a decision. Under conditions (1) and (2), an attractor system with stable attractor states can easily best an integrator when accuracy is more important than speed. Moreover, under conditions in which such stable attractor networks do not best the perfect integrator, a system with unstable initial states can do so if readout of the system’s final state is imperfect. Ubiquitously, an attractor system with a nonselective time-dependent input current is both more accurate and more robust to imprecise tuning of parameters than an integrator with such input. Given that neural responses that switch stochastically between discrete states can “masquerade” as integration in single-neuron and trial-averaged data, our results suggest that such networks should be considered as plausible alternatives to the integrator model.
Collapse
|
7
|
Bourjaily MA, Miller P. Dynamic afferent synapses to decision-making networks improve performance in tasks requiring stimulus associations and discriminations. J Neurophysiol 2012; 108:513-27. [PMID: 22457467 DOI: 10.1152/jn.00806.2011] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/22/2022] Open
Abstract
Animals must often make opposing responses to similar complex stimuli. Multiple sensory inputs from such stimuli combine to produce stimulus-specific patterns of neural activity. It is the differences between these activity patterns, even when small, that provide the basis for any differences in behavioral response. In the present study, we investigate three tasks with differing degrees of overlap in the inputs, each with just two response possibilities. We simulate behavioral output via winner-takes-all activity in one of two pools of neurons forming a biologically based decision-making layer. The decision-making layer receives inputs either in a direct stimulus-dependent manner or via an intervening recurrent network of neurons that form the associative layer, whose activity helps distinguish the stimuli of each task. We show that synaptic facilitation of synapses to the decision-making layer improves performance in these tasks, robustly increasing accuracy and speed of responses across multiple configurations of network inputs. Conversely, we find that synaptic depression worsens performance. In a linearly nonseparable task with exclusive-or logic, the benefit of synaptic facilitation lies in its superlinear transmission: effective synaptic strength increases with presynaptic firing rate, which enhances the already present superlinearity of presynaptic firing rate as a function of stimulus-dependent input. In linearly separable single-stimulus discrimination tasks, we find that facilitating synapses are always beneficial because synaptic facilitation always enhances any differences between inputs. Thus we predict that for optimal decision-making accuracy and speed, synapses from sensory or associative areas to decision-making or premotor areas should be facilitating.
Collapse
Affiliation(s)
- Mark A Bourjaily
- Neuroscience Program, Brandeis University, Waltham, MA 02454-9110, USA
| | | |
Collapse
|
8
|
Cortical attractor network dynamics with diluted connectivity. Brain Res 2011; 1434:212-25. [PMID: 21875702 DOI: 10.1016/j.brainres.2011.08.002] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2011] [Revised: 07/29/2011] [Accepted: 08/02/2011] [Indexed: 11/23/2022]
Abstract
The connectivity of the cerebral cortex is diluted, with the probability of excitatory connections between even nearby pyramidal cells rarely more than 0.1, and in the hippocampus 0.04. To investigate the extent to which this diluted connectivity affects the dynamics of attractor networks in the cerebral cortex, we simulated an integrate-and-fire attractor network taking decisions between competing inputs with diluted connectivity of 0.25 or 0.1, and with the same number of synaptic connections per neuron for the recurrent collateral synapses within an attractor population as for full connectivity. The results indicated that there was less spiking-related noise with the diluted connectivity in that the stability of the network when in the spontaneous state of firing increased, and the accuracy of the correct decisions increased. The decision times were a little slower with diluted than with complete connectivity. Given that the capacity of the network is set by the number of recurrent collateral synaptic connections per neuron, on which there is a biological limit, the findings indicate that the stability of cortical networks, and the accuracy of their correct decisions or memory recall operations, can be increased by utilizing diluted connectivity and correspondingly increasing the number of neurons in the network, with little impact on the speed of processing of the cortex. Thus diluted connectivity can decrease cortical spiking-related noise. In addition, we show that the Fano factor for the trial-to-trial variability of the neuronal firing decreases from the spontaneous firing state value when the attractor network makes a decision. This article is part of a Special Issue entitled "Neural Coding".
Collapse
|
9
|
Pfeiffer M, Nessler B, Douglas RJ, Maass W. Reward-modulated Hebbian learning of decision making. Neural Comput 2010; 22:1399-444. [PMID: 20141476 DOI: 10.1162/neco.2010.03-09-980] [Citation(s) in RCA: 29] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
We introduce a framework for decision making in which the learning of decision making is reduced to its simplest and biologically most plausible form: Hebbian learning on a linear neuron. We cast our Bayesian-Hebb learning rule as reinforcement learning in which certain decisions are rewarded and prove that each synaptic weight will on average converge exponentially fast to the log-odd of receiving a reward when its pre- and postsynaptic neurons are active. In our simple architecture, a particular action is selected from the set of candidate actions by a winner-take-all operation. The global reward assigned to this action then modulates the update of each synapse. Apart from this global reward signal, our reward-modulated Bayesian Hebb rule is a pure Hebb update that depends only on the coactivation of the pre- and postsynaptic neurons, not on the weighted sum of all presynaptic inputs to the postsynaptic neuron as in the perceptron learning rule or the Rescorla-Wagner rule. This simple approach to action-selection learning requires that information about sensory inputs be presented to the Bayesian decision stage in a suitably preprocessed form resulting from other adaptive processes (acting on a larger timescale) that detect salient dependencies among input features. Hence our proposed framework for fast learning of decisions also provides interesting new hypotheses regarding neural nodes and computational goals of cortical areas that provide input to the final decision stage.
Collapse
Affiliation(s)
- Michael Pfeiffer
- Institute for Theoretical Computer Science, Graz University of Technology, A-8010 Graz, Austria.
| | | | | | | |
Collapse
|
10
|
Abstract
What kind of strategies subjects follow in various behavioral circumstances has been a central issue in decision making. In particular, which behavioral strategy, maximizing or matching, is more fundamental to animal's decision behavior has been a matter of debate. Here, we prove that any algorithm to achieve the stationary condition for maximizing the average reward should lead to matching when it ignores the dependence of the expected outcome on subject's past choices. We may term this strategy of partial reward maximization “matching strategy”. Then, this strategy is applied to the case where the subject's decision system updates the information for making a decision. Such information includes subject's past actions or sensory stimuli, and the internal storage of this information is often called “state variables”. We demonstrate that the matching strategy provides an easy way to maximize reward when combined with the exploration of the state variables that correctly represent the crucial information for reward maximization. Our results reveal for the first time how a strategy to achieve matching behavior is beneficial to reward maximization, achieving a novel insight into the relationship between maximizing and matching.
Collapse
Affiliation(s)
- Yutaka Sakai
- Brain Science Institute, Tamagawa University, Machida, Tokyo, Japan
| | - Tomoki Fukai
- Laboratory for Neural Circuit Theory, Brain Science Institute, RIKEN, Wako, Saitama, Japan
- * E-mail:
| |
Collapse
|
11
|
Dankert JR, Olson B, Si J. Asynchronous decision making in a memorized paddle pressing task. J Neural Eng 2008; 5:363-73. [PMID: 18784389 DOI: 10.1088/1741-2560/5/4/001] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
This paper presents a method for asynchronous decision making using recorded neural data in a binary decision task. This is a demonstration of a technique for developing motor cortical neural prosthetics that do not rely on external cued timing information. The system presented in this paper uses support vector machines and leaky integrate-and-fire elements to predict directional paddle presses. In addition to the traditional metrics of accuracy, asynchronous systems must also optimize the time needed to make a decision. The system presented is able to predict paddle presses with a median accuracy of 88% and all decisions are made before the time of the actual paddle press. An alternative bit rate measure of performance is defined to show that the system proposed here is able to perform the task with the same efficiency as the rats.
Collapse
Affiliation(s)
- James R Dankert
- Department of Electrical Engineering, Arizona State University, Tempe, AZ 85287-5706, USA
| | | | | |
Collapse
|
12
|
Loewenstein Y. Robustness of learning that is based on covariance-driven synaptic plasticity. PLoS Comput Biol 2008; 4:e1000007. [PMID: 18369414 PMCID: PMC2265526 DOI: 10.1371/journal.pcbi.1000007] [Citation(s) in RCA: 26] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2007] [Accepted: 01/21/2008] [Indexed: 11/24/2022] Open
Abstract
It is widely believed that learning is due, at least in part, to long-lasting modifications of the strengths of synapses in the brain. Theoretical studies have shown that a family of synaptic plasticity rules, in which synaptic changes are driven by covariance, is particularly useful for many forms of learning, including associative memory, gradient estimation, and operant conditioning. Covariance-based plasticity is inherently sensitive. Even a slight mistuning of the parameters of a covariance-based plasticity rule is likely to result in substantial changes in synaptic efficacies. Therefore, the biological relevance of covariance-based plasticity models is questionable. Here, we study the effects of mistuning parameters of the plasticity rule in a decision making model in which synaptic plasticity is driven by the covariance of reward and neural activity. An exact covariance plasticity rule yields Herrnstein's matching law. We show that although the effect of slight mistuning of the plasticity rule on the synaptic efficacies is large, the behavioral effect is small. Thus, matching behavior is robust to mistuning of the parameters of the covariance-based plasticity rule. Furthermore, the mistuned covariance rule results in undermatching, which is consistent with experimentally observed behavior. These results substantiate the hypothesis that approximate covariance-based synaptic plasticity underlies operant conditioning. However, we show that the mistuning of the mean subtraction makes behavior sensitive to the mistuning of the properties of the decision making network. Thus, there is a tradeoff between the robustness of matching behavior to changes in the plasticity rule and its robustness to changes in the properties of the decision making network. It is widely believed that learning is due, at least in part, to modifications of synapses in the brain. The ability of a synapse to change its strength is called “synaptic plasticity,” and the rules governing these changes are a subject of intense research. Theoretical studies have shown that a particular family of synaptic plasticity rules, known as covariance rules, could underlie many forms of learning. While it is possible that a biological synapse would be able to approximately implement such abstract rules, it seems unlikely that this implementation would be exact. Covariance rules are inherently sensitive, and even a slight inaccuracy in their implementation is likely to result in substantial changes in synaptic strengths. Thus, the biological relevance of these rules remains questionable. Here we study the consequences of the mistuning of a covariance plasticity rule in the context of operant conditioning. In a previous study, we showed that an approximate phenomenological law of behavior called “the matching law” naturally emerges if synapses change according to the covariance rule. Here we show that although the effect of slight mistuning of the covariance rule on synaptic strengths is substantial, it leads to only small deviations from the matching law. Furthermore, these deviations are observed experimentally. Thus, our results support the hypothesis that covariance synaptic plasticity underlies operant conditioning.
Collapse
Affiliation(s)
- Yonatan Loewenstein
- Department of Neurobiology, Interdisciplinary Center for Neural Computation, Hebrew University, Jerusalem, Israel.
| |
Collapse
|
13
|
Okamoto H, Isomura Y, Takada M, Fukai T. Temporal integration by stochastic recurrent network dynamics with bimodal neurons. J Neurophysiol 2007; 97:3859-67. [PMID: 17392417 DOI: 10.1152/jn.01100.2006] [Citation(s) in RCA: 33] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Temporal integration of externally or internally driven information is required for a variety of cognitive processes. This computation is generally linked with graded rate changes in cortical neurons, which typically appear during a delay period of cognitive task in the prefrontal and other cortical areas. Here, we present a neural network model to produce graded (climbing or descending) neuronal activity. Model neurons are interconnected randomly by AMPA-receptor-mediated fast excitatory synapses and are subject to noisy background excitatory and inhibitory synaptic inputs. In each neuron, a prolonged afterdepolarizing potential follows every spike generation. Then, driven by an external input, the individual neurons display bimodal rate changes between a baseline state and an elevated firing state, with the latter being sustained by regenerated afterdepolarizing potentials. When the variance of background input and the uniform weight of recurrent synapses are adequately tuned, we show that stochastic noise and reverberating synaptic input organize these bimodal changes into a sequence that exhibits graded population activity with a nearly constant slope. To test the validity of the proposed mechanism, we analyzed the graded activity of anterior cingulate cortex neurons in monkeys performing delayed conditional Go/No-go discrimination tasks. The delay-period activities of cingulate neurons exhibited bimodal activity patterns and trial-to-trial variability that are similar to those predicted by the proposed model.
Collapse
Affiliation(s)
- Hiroshi Okamoto
- Laboratory for Neural Circuit Theory, RIKEN Brain Science Institute, Hirosawa 2-1, Wako, Saitama 351-0198, Japan
| | | | | | | |
Collapse
|
14
|
Kawato M, Samejima K. Efficient reinforcement learning: computational theories, neuroscience and robotics. Curr Opin Neurobiol 2007; 17:205-12. [PMID: 17374483 DOI: 10.1016/j.conb.2007.03.004] [Citation(s) in RCA: 44] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2007] [Accepted: 03/08/2007] [Indexed: 11/22/2022]
Abstract
Reinforcement learning algorithms have provided some of the most influential computational theories for behavioral learning that depends on reward and penalty. After briefly reviewing supporting experimental data, this paper tackles three difficult theoretical issues that remain to be explored. First, plain reinforcement learning is much too slow to be considered a plausible brain model. Second, although the temporal-difference error has an important role both in theory and in experiments, how to compute it remains an enigma. Third, function of all brain areas, including the cerebral cortex, cerebellum, brainstem and basal ganglia, seems to necessitate a new computational framework. Computational studies that emphasize meta-parameters, hierarchy, modularity and supervised learning to resolve these issues are reviewed here, together with the related experimental data.
Collapse
Affiliation(s)
- Mitsuo Kawato
- ATR Computational Neuroscience Laboratories, 2-2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-0288, Japan.
| | | |
Collapse
|