1
|
Kim JZ, Larsen B, Parkes L. Shaping dynamical neural computations using spatiotemporal constraints. Biochem Biophys Res Commun 2024; 728:150302. [PMID: 38968771 DOI: 10.1016/j.bbrc.2024.150302] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Revised: 03/21/2024] [Accepted: 04/11/2024] [Indexed: 07/07/2024]
Abstract
Dynamics play a critical role in computation. The principled evolution of states over time enables both biological and artificial networks to represent and integrate information to make decisions. In the past few decades, significant multidisciplinary progress has been made in bridging the gap between how we understand biological versus artificial computation, including how insights gained from one can translate to the other. Research has revealed that neurobiology is a key determinant of brain network architecture, which gives rise to spatiotemporally constrained patterns of activity that underlie computation. Here, we discuss how neural systems use dynamics for computation, and claim that the biological constraints that shape brain networks may be leveraged to improve the implementation of artificial neural networks. To formalize this discussion, we consider a natural artificial analog of the brain that has been used extensively to model neural computation: the recurrent neural network (RNN). In both the brain and the RNN, we emphasize the common computational substrate atop which dynamics occur-the connectivity between neurons-and we explore the unique computational advantages offered by biophysical constraints such as resource efficiency, spatial embedding, and neurodevelopment.
Collapse
Affiliation(s)
- Jason Z Kim
- Department of Physics, Cornell University, Ithaca, NY, 14853, USA.
| | - Bart Larsen
- Department of Pediatrics, Masonic Institute for the Developing Brain, University of Minnesota, USA
| | - Linden Parkes
- Department of Psychiatry, Rutgers University, Piscataway, NJ, 08854, USA.
| |
Collapse
|
2
|
Heins C, Millidge B, Da Costa L, Mann RP, Friston KJ, Couzin ID. Collective behavior from surprise minimization. Proc Natl Acad Sci U S A 2024; 121:e2320239121. [PMID: 38630721 PMCID: PMC11046639 DOI: 10.1073/pnas.2320239121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2023] [Accepted: 03/08/2024] [Indexed: 04/19/2024] Open
Abstract
Collective motion is ubiquitous in nature; groups of animals, such as fish, birds, and ungulates appear to move as a whole, exhibiting a rich behavioral repertoire that ranges from directed movement to milling to disordered swarming. Typically, such macroscopic patterns arise from decentralized, local interactions among constituent components (e.g., individual fish in a school). Preeminent models of this process describe individuals as self-propelled particles, subject to self-generated motion and "social forces" such as short-range repulsion and long-range attraction or alignment. However, organisms are not particles; they are probabilistic decision-makers. Here, we introduce an approach to modeling collective behavior based on active inference. This cognitive framework casts behavior as the consequence of a single imperative: to minimize surprise. We demonstrate that many empirically observed collective phenomena, including cohesion, milling, and directed motion, emerge naturally when considering behavior as driven by active Bayesian inference-without explicitly building behavioral rules or goals into individual agents. Furthermore, we show that active inference can recover and generalize the classical notion of social forces as agents attempt to suppress prediction errors that conflict with their expectations. By exploring the parameter space of the belief-based model, we reveal nontrivial relationships between the individual beliefs and group properties like polarization and the tendency to visit different collective states. We also explore how individual beliefs about uncertainty determine collective decision-making accuracy. Finally, we show how agents can update their generative model over time, resulting in groups that are collectively more sensitive to external fluctuations and encode information more robustly.
Collapse
Affiliation(s)
- Conor Heins
- Department of Collective Behaviour, Max Planck Institute of Animal Behavior, KonstanzD-78457, Germany
- Centre for the Advanced Study of Collective Behaviour, University of Konstanz, KonstanzD-78457, Germany
- Department of Biology, University of Konstanz, KonstanzD-78457, Germany
- VERSES Research Lab, Los Angeles, CA90016
| | - Beren Millidge
- Medical Research Council Brain Networks Dynamics Unit, University of Oxford, OxfordOX1 3TH, United Kingdom
| | - Lancelot Da Costa
- VERSES Research Lab, Los Angeles, CA90016
- Department of Mathematics, Imperial College London, LondonSW7 2AZ, United Kingdom
- Wellcome Centre for Human Neuroimaging, University College London, LondonWC1N 3AR, United Kingdom
| | - Richard P. Mann
- Department of Statistics, School of Mathematics, University of Leeds, LeedsLS2 9JT, United Kingdom
| | - Karl J. Friston
- VERSES Research Lab, Los Angeles, CA90016
- Wellcome Centre for Human Neuroimaging, University College London, LondonWC1N 3AR, United Kingdom
| | - Iain D. Couzin
- Department of Collective Behaviour, Max Planck Institute of Animal Behavior, KonstanzD-78457, Germany
- Centre for the Advanced Study of Collective Behaviour, University of Konstanz, KonstanzD-78457, Germany
- Department of Biology, University of Konstanz, KonstanzD-78457, Germany
| |
Collapse
|
3
|
Mograbi DC, Hall S, Arantes B, Huntley J. The cognitive neuroscience of self-awareness: Current framework, clinical implications, and future research directions. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2024; 15:e1670. [PMID: 38043919 DOI: 10.1002/wcs.1670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/20/2023] [Revised: 11/04/2023] [Accepted: 11/06/2023] [Indexed: 12/05/2023]
Abstract
Self-awareness, the ability to take oneself as the object of awareness, has been an enigma for our species, with different answers to this question being provided by religion, philosophy, and, more recently, science. The current review aims to discuss the neurocognitive mechanisms underlying self-awareness. The multidimensional nature of self-awareness will be explored, suggesting how it can be thought of as an emergent property observed in different cognitive complexity levels, within a predictive coding approach. A presentation of alterations of self-awareness in neuropsychiatric conditions will ground a discussion on alternative frameworks to understand this phenomenon, in health and psychopathology, with future research directions being indicated to fill current gaps in the literature. This article is categorized under: Philosophy > Consciousness Psychology > Brain Function and Dysfunction Neuroscience > Cognition.
Collapse
Affiliation(s)
- Daniel C Mograbi
- Department of Psychology, Pontifical Catholic University of Rio de Janeiro, Rio de Janeiro, Brazil
- Institute of Psychiatry, Psychology & Neuroscience, King's College London, London, UK
| | - Simon Hall
- Camden and Islington NHS Foundation Trust, London, UK
| | - Beatriz Arantes
- Department of Psychology, Pontifical Catholic University of Rio de Janeiro, Rio de Janeiro, Brazil
| | - Jonathan Huntley
- Division of Psychiatry, University College London, London, UK
- Faculty of Health and Life Sciences, University of Exeter, Exeter, UK
| |
Collapse
|
4
|
Sladky R, Kargl D, Haubensak W, Lamm C. An active inference perspective for the amygdala complex. Trends Cogn Sci 2024; 28:223-236. [PMID: 38103984 DOI: 10.1016/j.tics.2023.11.004] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2023] [Revised: 11/14/2023] [Accepted: 11/16/2023] [Indexed: 12/19/2023]
Abstract
The amygdala is a heterogeneous network of subcortical nuclei with central importance in cognitive and clinical neuroscience. Various experimental designs in human psychology and animal model research have mapped multiple conceptual frameworks (e.g., valence/salience and decision making) to ever more refined amygdala circuitry. However, these predominantly bottom up-driven accounts often rely on interpretations tailored to a specific phenomenon, thus preventing comprehensive and integrative theories. We argue here that an active inference model of amygdala function could unify these fractionated approaches into an overarching framework for clearer empirical predictions and mechanistic interpretations. This framework embeds top-down predictive models, informed by prior knowledge and belief updating, within a dynamical system distributed across amygdala circuits in which self-regulation is implemented by continuously tracking environmental and homeostatic demands.
Collapse
Affiliation(s)
- Ronald Sladky
- Social, Cognitive, and Affective Neuroscience Unit, Department of Cognition, Emotion, and Methods in Psychology, Faculty of Psychology, University of Vienna, Liebiggasse 5, 1010 Vienna, Austria; Vienna Cognitive Science Hub, University of Vienna, 1010 Vienna, Austria.
| | - Dominic Kargl
- Department of Neuronal Cell Biology, Center for Brain Research, Medical University of Vienna, Spitalgasse 4, 1090 Vienna, Austria
| | - Wulf Haubensak
- Department of Neuronal Cell Biology, Center for Brain Research, Medical University of Vienna, Spitalgasse 4, 1090 Vienna, Austria; Research Institute of Molecular Pathology (IMP), Vienna Biocenter (VBC), Campus Vienna Biocenter 1, 1030 Vienna, Austria
| | - Claus Lamm
- Social, Cognitive, and Affective Neuroscience Unit, Department of Cognition, Emotion, and Methods in Psychology, Faculty of Psychology, University of Vienna, Liebiggasse 5, 1010 Vienna, Austria; Vienna Cognitive Science Hub, University of Vienna, 1010 Vienna, Austria
| |
Collapse
|
5
|
Rao RPN, Gklezakos DC, Sathish V. Active Predictive Coding: A Unifying Neural Model for Active Perception, Compositional Learning, and Hierarchical Planning. Neural Comput 2023; 36:1-32. [PMID: 38052084 DOI: 10.1162/neco_a_01627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Accepted: 09/20/2023] [Indexed: 12/07/2023]
Abstract
There is growing interest in predictive coding as a model of how the brain learns through predictions and prediction errors. Predictive coding models have traditionally focused on sensory coding and perception. Here we introduce active predictive coding (APC) as a unifying model for perception, action, and cognition. The APC model addresses important open problems in cognitive science and AI, including (1) how we learn compositional representations (e.g., part-whole hierarchies for equivariant vision) and (2) how we solve large-scale planning problems, which are hard for traditional reinforcement learning, by composing complex state dynamics and abstract actions from simpler dynamics and primitive actions. By using hypernetworks, self-supervised learning, and reinforcement learning, APC learns hierarchical world models by combining task-invariant state transition networks and task-dependent policy networks at multiple abstraction levels. We illustrate the applicability of the APC model to active visual perception and hierarchical planning. Our results represent, to our knowledge, the first proof-of-concept demonstration of a unified approach to addressing the part-whole learning problem in vision, the nested reference frames learning problem in cognition, and the integrated state-action hierarchy learning problem in reinforcement learning.
Collapse
Affiliation(s)
- Rajesh P N Rao
- Paul G. Allen School of Computer Science and Engineering and Center for Neurotechnology, University of Washington, Seattle, WA 98195, U.S.A.
| | - Dimitrios C Gklezakos
- Paul G. Allen School of Computer Science and Engineering and Center for Neurotechnology, University of Washington, Seattle, WA 98195, U.S.A.
| | - Vishwas Sathish
- Paul G. Allen School of Computer Science and Engineering and Center for Neurotechnology, University of Washington, Seattle, WA 98195, U.S.A.
| |
Collapse
|
6
|
Op de Beeck H, Bracci S. Going after the bigger picture: Using high-capacity models to understand mind and brain. Behav Brain Sci 2023; 46:e404. [PMID: 38054291 DOI: 10.1017/s0140525x2300153x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
Deep neural networks (DNNs) provide a unique opportunity to move towards a generic modelling framework in psychology. The high representational capacity of these models combined with the possibility for further extensions has already allowed us to investigate the forest, namely the complex landscape of representations and processes that underlie human cognition, without forgetting about the trees, which include individual psychological phenomena.
Collapse
Affiliation(s)
| | - Stefania Bracci
- Center for Mind/Brain Sciences, University of Trento, Rovereto, Italy ://webapps.unitn.it/du/en/Persona/PER0076943/Curriculum
| |
Collapse
|
7
|
Kim JZ, Larsen B, Parkes L. Shaping dynamical neural computations using spatiotemporal constraints. ARXIV 2023:arXiv:2311.15572v1. [PMID: 38076517 PMCID: PMC10705584] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 12/20/2023]
Abstract
Dynamics play a critical role in computation. The principled evolution of states over time enables both biological and artificial networks to represent and integrate information to make decisions. In the past few decades, significant multidisciplinary progress has been made in bridging the gap between how we understand biological versus artificial computation, including how insights gained from one can translate to the other. Research has revealed that neurobiology is a key determinant of brain network architecture, which gives rise to spatiotemporally constrained patterns of activity that underlie computation. Here, we discuss how neural systems use dynamics for computation, and claim that the biological constraints that shape brain networks may be leveraged to improve the implementation of artificial neural networks. To formalize this discussion, we consider a natural artificial analog of the brain that has been used extensively to model neural computation: the recurrent neural network (RNN). In both the brain and the RNN, we emphasize the common computational substrate atop which dynamics occur-the connectivity between neurons-and we explore the unique computational advantages offered by biophysical constraints such as resource efficiency, spatial embedding, and neurodevelopment.
Collapse
Affiliation(s)
- Jason Z. Kim
- Department of Physics, Cornell University, Ithaca, NY 14853, USA
| | - Bart Larsen
- Department of Pediatrics, Masonic Institute for the Developing Brain, University of Minnesota
| | - Linden Parkes
- Department of Psychiatry, Rutgers University, Piscataway, NJ 08854, USA
| |
Collapse
|
8
|
Ryskin R, Nieuwland MS. Prediction during language comprehension: what is next? Trends Cogn Sci 2023; 27:1032-1052. [PMID: 37704456 DOI: 10.1016/j.tics.2023.08.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2022] [Revised: 08/03/2023] [Accepted: 08/04/2023] [Indexed: 09/15/2023]
Abstract
Prediction is often regarded as an integral aspect of incremental language comprehension, but little is known about the cognitive architectures and mechanisms that support it. We review studies showing that listeners and readers use all manner of contextual information to generate multifaceted predictions about upcoming input. The nature of these predictions may vary between individuals owing to differences in language experience, among other factors. We then turn to unresolved questions which may guide the search for the underlying mechanisms. (i) Is prediction essential to language processing or an optional strategy? (ii) Are predictions generated from within the language system or by domain-general processes? (iii) What is the relationship between prediction and memory? (iv) Does prediction in comprehension require simulation via the production system? We discuss promising directions for making progress in answering these questions and for developing a mechanistic understanding of prediction in language.
Collapse
Affiliation(s)
- Rachel Ryskin
- Department of Cognitive and Information Sciences, University of California Merced, 5200 Lake Road, Merced, CA 95343, USA.
| | - Mante S Nieuwland
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands; Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, The Netherlands
| |
Collapse
|
9
|
Pan Y, Wen Y, Jin J, Chen J. The interpersonal computational psychiatry of social coordination in schizophrenia. Lancet Psychiatry 2023; 10:801-808. [PMID: 37478889 DOI: 10.1016/s2215-0366(23)00146-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Revised: 04/13/2023] [Accepted: 04/13/2023] [Indexed: 07/23/2023]
Abstract
Impairments in social coordination form a core dimension of various psychiatric disorders, including schizophrenia. Advances in interpersonal and computational psychiatry support a major change in studying social coordination in schizophrenia. Although these developments provided novel perspectives to study how interpersonal activities shape coordination and to examine computational mechanisms, direct attempts to integrate the two methodologies have been sparse. Here, we propose an interpersonal computational framework that (1) leverages the active inference framework to model aberrant social coordination processes in schizophrenia and (2) incorporates dynamical system models to dissect intrapersonal and interpersonal synchronisation to inform a statistical model based on active inference. We discuss how this interpersonal computational psychiatry framework can elucidate the aberrant processes leading to psychopathology, with schizophrenia as an example, and highlight how it might aid clinical intervention and practice. Finally, we discuss challenges and opportunities for using the framework in studying social coordination impairments.
Collapse
Affiliation(s)
- Yafeng Pan
- Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou, China
| | - Yalan Wen
- Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou, China
| | - Jingwen Jin
- Department of Psychology, The University of Hong Kong, Hong Kong Special Administrative Region, China; State Key Laboratory of Brain and Cognitive Sciences, The University of Hong Kong, Hong Kong Special Administrative Region, China
| | - Ji Chen
- Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou, China; Department of Psychiatry, The Fourth Affiliated Hospital, Zhejiang University School of Medicine, Yiwu, Zhejiang, China.
| |
Collapse
|
10
|
Doerig A, Sommers RP, Seeliger K, Richards B, Ismael J, Lindsay GW, Kording KP, Konkle T, van Gerven MAJ, Kriegeskorte N, Kietzmann TC. The neuroconnectionist research programme. Nat Rev Neurosci 2023:10.1038/s41583-023-00705-w. [PMID: 37253949 DOI: 10.1038/s41583-023-00705-w] [Citation(s) in RCA: 14] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/21/2023] [Indexed: 06/01/2023]
Abstract
Artificial neural networks (ANNs) inspired by biology are beginning to be widely used to model behavioural and neural data, an approach we call 'neuroconnectionism'. ANNs have been not only lauded as the current best models of information processing in the brain but also criticized for failing to account for basic cognitive functions. In this Perspective article, we propose that arguing about the successes and failures of a restricted set of current ANNs is the wrong approach to assess the promise of neuroconnectionism for brain science. Instead, we take inspiration from the philosophy of science, and in particular from Lakatos, who showed that the core of a scientific research programme is often not directly falsifiable but should be assessed by its capacity to generate novel insights. Following this view, we present neuroconnectionism as a general research programme centred around ANNs as a computational language for expressing falsifiable theories about brain computation. We describe the core of the programme, the underlying computational framework and its tools for testing specific neuroscientific hypotheses and deriving novel understanding. Taking a longitudinal view, we review past and present neuroconnectionist projects and their responses to challenges and argue that the research programme is highly progressive, generating new and otherwise unreachable insights into the workings of the brain.
Collapse
Affiliation(s)
- Adrien Doerig
- Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany.
- Donders Institute for Brain, Cognition and Behaviour, Nijmegen, The Netherlands.
| | - Rowan P Sommers
- Department of Neurobiology of Language, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Katja Seeliger
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Blake Richards
- Department of Neurology and Neurosurgery, McGill University, Montréal, QC, Canada
- School of Computer Science, McGill University, Montréal, QC, Canada
- Mila, Montréal, QC, Canada
- Montréal Neurological Institute, Montréal, QC, Canada
- Learning in Machines and Brains Program, CIFAR, Toronto, ON, Canada
| | | | | | - Konrad P Kording
- Learning in Machines and Brains Program, CIFAR, Toronto, ON, Canada
- Bioengineering, Neuroscience, University of Pennsylvania, Pennsylvania, PA, USA
| | | | | | | | - Tim C Kietzmann
- Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany
| |
Collapse
|
11
|
Kirubeswaran OR, Storrs KR. Inconsistent illusory motion in predictive coding deep neural networks. Vision Res 2023; 206:108195. [PMID: 36801664 DOI: 10.1016/j.visres.2023.108195] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2022] [Revised: 01/31/2023] [Accepted: 01/31/2023] [Indexed: 02/19/2023]
Abstract
Why do we perceive illusory motion in some static images? Several accounts point to eye movements, response latencies to different image elements, or interactions between image patterns and motion energy detectors. Recently PredNet, a recurrent deep neural network (DNN) based on predictive coding principles, was reported to reproduce the "Rotating Snakes" illusion, suggesting a role for predictive coding. We begin by replicating this finding, then use a series of "in silico" psychophysics and electrophysiology experiments to examine whether PredNet behaves consistently with human observers and non-human primate neural data. A pretrained PredNet predicted illusory motion for all subcomponents of the Rotating Snakes pattern, consistent with human observers. However, we found no simple response delays in internal units, unlike evidence from electrophysiological data. PredNet's detection of motion in gradients seemed dependent on contrast, but depends predominantly on luminance in humans. Finally, we examined the robustness of the illusion across ten PredNets of identical architecture, retrained on the same video data. There was large variation across network instances in whether they reproduced the Rotating Snakes illusion, and what motion, if any, they predicted for simplified variants. Unlike human observers, no network predicted motion for greyscale variants of the Rotating Snakes pattern. Our results sound a cautionary note: even when a DNN successfully reproduces some idiosyncrasy of human vision, more detailed investigation can reveal inconsistencies between humans and the network, and between different instances of the same network. These inconsistencies suggest that predictive coding does not reliably give rise to human-like illusory motion.
Collapse
Affiliation(s)
| | - Katherine R Storrs
- Department of Experimental Psychology, Justus Liebig University Giessen, Germany; Centre for Mind, Brain and Behaviour (CMBB), University of Marburg and Justus Liebig University Giessen, Germany; School of Psychology, University of Auckland, New Zealand
| |
Collapse
|
12
|
Bakhtiari S. Energy efficiency as a normative account for predictive coding. PATTERNS (NEW YORK, N.Y.) 2022; 3:100661. [PMID: 38283565 PMCID: PMC10810825 DOI: 10.1016/j.patter.2022.100661] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
In this issue of Patterns, Ali et al. demonstrate that predictive coding emerges in an artificial neural network optimized to be energy efficient. The results offer an explanation for why brains may implement predictive coding.
Collapse
Affiliation(s)
- Shahab Bakhtiari
- Psychology Department, University of Montreal, Montreal, QC, Canada
- Mila (Quebec AI Institute), Montreal, QC, Canada
| |
Collapse
|