1
|
Gerstenberg T. Counterfactual simulation in causal cognition. Trends Cogn Sci 2024; 28:924-936. [PMID: 38777661 DOI: 10.1016/j.tics.2024.04.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2024] [Revised: 04/23/2024] [Accepted: 04/23/2024] [Indexed: 05/25/2024]
Abstract
How do people make causal judgments and assign responsibility? In this review article, I argue that counterfactual simulations are key. To simulate counterfactuals, we need three ingredients: a generative mental model of the world, the ability to perform interventions on that model, and the capacity to simulate the consequences of these interventions. The counterfactual simulation model (CSM) uses these ingredients to capture people's intuitive understanding of the physical and social world. In the physical domain, the CSM predicts people's causal judgments about dynamic collision events, complex situations that involve multiple causes, omissions as causes, and causes that sustain physical stability. In the social domain, the CSM predicts responsibility judgments in helping and hindering scenarios.
Collapse
Affiliation(s)
- Tobias Gerstenberg
- Stanford University, Department of Psychology, 450 Jane Stanford Way, Bldg 420, Stanford, CA 94305, USA.
| |
Collapse
|
2
|
Piantadosi ST, Muller DCY, Rule JS, Kaushik K, Gorenstein M, Leib ER, Sanford E. Why concepts are (probably) vectors. Trends Cogn Sci 2024; 28:844-856. [PMID: 39112125 DOI: 10.1016/j.tics.2024.06.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Revised: 06/18/2024] [Accepted: 06/27/2024] [Indexed: 09/06/2024]
Abstract
For decades, cognitive scientists have debated what kind of representation might characterize human concepts. Whatever the format of the representation, it must allow for the computation of varied properties, including similarities, features, categories, definitions, and relations. It must also support the development of theories, ad hoc categories, and knowledge of procedures. Here, we discuss why vector-based representations provide a compelling account that can meet all these needs while being plausibly encoded into neural architectures. This view has become especially promising with recent advances in both large language models and vector symbolic architectures. These innovations show how vectors can handle many properties traditionally thought to be out of reach for neural models, including compositionality, definitions, structures, and symbolic computational processes.
Collapse
Affiliation(s)
- Steven T Piantadosi
- Department of Psychology, University of California, Berkeley, CA, USA; Department of Neuroscience, University of California, Berkeley, CA, USA.
| | - Dyana C Y Muller
- Department of Neuroscience, University of California, Berkeley, CA, USA
| | - Joshua S Rule
- Department of Psychology, University of California, Berkeley, CA, USA
| | | | - Mark Gorenstein
- Department of Neuroscience, University of California, Berkeley, CA, USA
| | - Elena R Leib
- Department of Psychology, University of California, Berkeley, CA, USA
| | - Emily Sanford
- Department of Psychology, University of California, Berkeley, CA, USA
| |
Collapse
|
3
|
Rule JS, Piantadosi ST, Cropper A, Ellis K, Nye M, Tenenbaum JB. Symbolic metaprogram search improves learning efficiency and explains rule learning in humans. Nat Commun 2024; 15:6847. [PMID: 39127796 DOI: 10.1038/s41467-024-50966-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Accepted: 07/23/2024] [Indexed: 08/12/2024] Open
Abstract
Throughout their lives, humans seem to learn a variety of rules for things like applying category labels, following procedures, and explaining causal relationships. These rules are often algorithmically rich but are nonetheless acquired with minimal data and computation. Symbolic models based on program learning successfully explain rule-learning in many domains, but performance degrades quickly as program complexity increases. It remains unclear how to scale symbolic rule-learning methods to model human performance in challenging domains. Here we show that symbolic search over the space of metaprograms-programs that revise programs-dramatically improves learning efficiency. On a behavioral benchmark of 100 algorithmically rich rules, this approach fits human learning more accurately than alternative models while also using orders of magnitude less search. The computation required to match median human performance is consistent with conservative estimates of human thinking time. Our results suggest that metaprogram-like representations may help human learners to efficiently acquire rules.
Collapse
Affiliation(s)
- Joshua S Rule
- Psychology, University of California, Berkeley, Berkeley, CA, 94704, USA.
| | | | | | - Kevin Ellis
- Computer Science, Cornell University, Ithaca, NY, 14850, USA
| | - Maxwell Nye
- Adept AI Labs, San Francisco, CA, 94110, USA
| | - Joshua B Tenenbaum
- Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, 02139, USA
| |
Collapse
|
4
|
Bramley NR, Zhao B, Quillien T, Lucas CG. Local Search and the Evolution of World Models. Top Cogn Sci 2023. [PMID: 37850714 DOI: 10.1111/tops.12703] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Revised: 10/02/2023] [Accepted: 10/03/2023] [Indexed: 10/19/2023]
Abstract
An open question regarding how people develop their models of the world is how new candidates are generated for consideration out of infinitely many possibilities. We discuss the role that evolutionary mechanisms play in this process. Specifically, we argue that when it comes to developing a global world model, innovation is necessarily incremental, involving the generation and selection among random local mutations and recombinations of (parts of) one's current model. We argue that, by narrowing and guiding exploration, this feature of cognitive search is what allows human learners to discover better theories, without ever grappling directly with the problem of finding a "global optimum," or best possible world model. We suggest this aspect of cognitive processing works analogously to how blind variation and selection mechanisms drive biological evolution. We propose algorithms developed for program synthesis provide candidate mechanisms for how human minds might achieve this. We discuss objections and implications of this perspective, finally suggesting that a better process-level understanding of how humans incrementally explore compositional theory spaces can shed light on how we think, and provide explanatory traction on fundamental cognitive biases, including anchoring, probability matching, and confirmation bias.
Collapse
Affiliation(s)
| | - Bonan Zhao
- Department of Psychology, University of Edinburgh
| | - Tadeg Quillien
- Institute of Language, Cognition & Computation, Informatics University of Edinburgh
| | - Christopher G Lucas
- Institute of Language, Cognition & Computation, Informatics University of Edinburgh
| |
Collapse
|
5
|
Ellis K, Wong L, Nye M, Sablé-Meyer M, Cary L, Anaya Pozo L, Hewitt L, Solar-Lezama A, Tenenbaum JB. DreamCoder: growing generalizable, interpretable knowledge with wake-sleep Bayesian program learning. PHILOSOPHICAL TRANSACTIONS. SERIES A, MATHEMATICAL, PHYSICAL, AND ENGINEERING SCIENCES 2023; 381:20220050. [PMID: 37271169 DOI: 10.1098/rsta.2022.0050] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/31/2023] [Accepted: 03/23/2023] [Indexed: 06/06/2023]
Abstract
Expert problem-solving is driven by powerful languages for thinking about problems and their solutions. Acquiring expertise means learning these languages-systems of concepts, alongside the skills to use them. We present DreamCoder, a system that learns to solve problems by writing programs. It builds expertise by creating domain-specific programming languages for expressing domain concepts, together with neural networks to guide the search for programs within these languages. A 'wake-sleep' learning algorithm alternately extends the language with new symbolic abstractions and trains the neural network on imagined and replayed problems. DreamCoder solves both classic inductive programming tasks and creative tasks such as drawing pictures and building scenes. It rediscovers the basics of modern functional programming, vector algebra and classical physics, including Newton's and Coulomb's laws. Concepts are built compositionally from those learned earlier, yielding multilayered symbolic representations that are interpretable and transferrable to new tasks, while still growing scalably and flexibly with experience. This article is part of a discussion meeting issue 'Cognitive artificial intelligence'.
Collapse
|
6
|
Algorithms of adaptation in inductive inference. Cogn Psychol 2022; 137:101506. [PMID: 35872374 DOI: 10.1016/j.cogpsych.2022.101506] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2021] [Revised: 06/01/2022] [Accepted: 07/08/2022] [Indexed: 11/20/2022]
Abstract
We investigate the idea that human concept inference utilizes local adaptive search within a compositional mental theory space. To explore this, we study human judgments in a challenging task that involves actively gathering evidence about a symbolic rule governing the behavior of a simulated environment. Participants learn by performing mini-experiments before making generalizations and explicit guesses about a hidden rule. They then collect additional evidence themselves (Experiment 1) or observe evidence gathered by someone else (Experiment 2) before revising their own generalizations and guesses. In each case, we focus on the relationship between participants' initial and revised guesses about the hidden rule concept. We find an order effect whereby revised guesses are anchored to idiosyncratic elements of the earlier guess. To explain this pattern, we develop a family of process accounts that combine program induction ideas with local (MCMC-like) adaptation mechanisms. A particularly local variant of this adaptive account captures participants' hypothesis revisions better than a range of alternative explanations. We take this as suggestive that people deal with the inherent complexity of concept inference partly through use of local adaptive search in a latent compositional theory space.
Collapse
|
7
|
Piantadosi ST. The computational origin of representation. Minds Mach (Dordr) 2021; 31:1-58. [PMID: 34305318 PMCID: PMC8300595 DOI: 10.1007/s11023-020-09540-9] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2019] [Accepted: 08/29/2020] [Indexed: 01/29/2023]
Abstract
Each of our theories of mental representation provides some insight into how the mind works. However, these insights often seem incompatible, as the debates between symbolic, dynamical, emergentist, sub-symbolic, and grounded approaches to cognition attest. Mental representations-whatever they are-must share many features with each of our theories of representation, and yet there are few hypotheses about how a synthesis could be possible. Here, I develop a theory of the underpinnings of symbolic cognition that shows how sub-symbolic dynamics may give rise to higher-level cognitive representations of structures, systems of knowledge, and algorithmic processes. This theory implements a version of conceptual role semantics by positing an internal universal representation language in which learners may create mental models to capture dynamics they observe in the world. The theory formalizes one account of how truly novel conceptual content may arise, allowing us to explain how even elementary logical and computational operations may be learned from a more primitive basis. I provide an implementation that learns to represent a variety of structures, including logic, number, kinship trees, regular languages, context-free languages, domains of theories like magnetism, dominance hierarchies, list structures, quantification, and computational primitives like repetition, reversal, and recursion. This account is based on simple discrete dynamical processes that could be implemented in a variety of different physical or biological systems. In particular, I describe how the required dynamics can be directly implemented in a connectionist framework. The resulting theory provides an "assembly language" for cognition, where high-level theories of symbolic computation can be implemented in simple dynamics that themselves could be encoded in biologically plausible systems.
Collapse
|
8
|
Rule JS, Tenenbaum JB, Piantadosi ST. The Child as Hacker. Trends Cogn Sci 2020; 24:900-915. [PMID: 33012688 PMCID: PMC7673661 DOI: 10.1016/j.tics.2020.07.005] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2019] [Revised: 07/13/2020] [Accepted: 07/16/2020] [Indexed: 01/29/2023]
Abstract
The scope of human learning and development poses a radical challenge for cognitive science. We propose that developmental theories can address this challenge by adopting perspectives from computer science. Many of our best models treat learning as analogous to computer programming because symbolic programs provide the most compelling account of sophisticated mental representations. We specifically propose that children's learning is analogous to a particular style of programming called hacking, making code better along many dimensions through an open-ended set of goals and activities. By contrast to existing theories, which depend primarily on local search and simple metrics, this view highlights the many features of good mental representations and the multiple complementary processes children use to create them.
Collapse
Affiliation(s)
- Joshua S Rule
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA.
| | - Joshua B Tenenbaum
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Steven T Piantadosi
- Department of Psychology, University of California, Berkeley, Berkeley, CA, USA
| |
Collapse
|
9
|
Mahr JB. The dimensions of episodic simulation. Cognition 2020; 196:104085. [DOI: 10.1016/j.cognition.2019.104085] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2019] [Revised: 09/18/2019] [Accepted: 09/29/2019] [Indexed: 01/28/2023]
|
10
|
Bramley NR, Gerstenberg T, Tenenbaum JB, Gureckis TM. Intuitive experimentation in the physical world. Cogn Psychol 2018; 105:9-38. [DOI: 10.1016/j.cogpsych.2018.05.001] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2017] [Revised: 03/27/2018] [Accepted: 05/11/2018] [Indexed: 10/14/2022]
|
11
|
Ellis AW, Mast FW. Toward a Dynamic Probabilistic Model for Vestibular Cognition. Front Psychol 2017; 8:138. [PMID: 28203219 PMCID: PMC5285352 DOI: 10.3389/fpsyg.2017.00138] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2016] [Accepted: 01/19/2017] [Indexed: 11/16/2022] Open
Abstract
We suggest that research in vestibular cognition will benefit from the theoretical framework of probabilistic models. This will aid in developing an understanding of how interactions between high-level cognition and low-level sensory processing might occur. Many such interactions have been shown experimentally; however, to date, no attempt has been made to systematically explore vestibular cognition by using computational modeling. It is widely assumed that mental imagery and perception share at least in part neural circuitry, and it has been proposed that mental simulation is closely connected to the brain’s ability to make predictions. We claim that this connection has been disregarded in the vestibular domain, and we suggest ways in which future research may take this into consideration.
Collapse
Affiliation(s)
- Andrew W Ellis
- Department of Psychology, University of Bern Bern, Switzerland
| | - Fred W Mast
- Department of Psychology, University of Bern Bern, Switzerland
| |
Collapse
|
12
|
Edelman S. The minority report: some common assumptions to reconsider in the modelling of the brain and behaviour. J EXP THEOR ARTIF IN 2015. [DOI: 10.1080/0952813x.2015.1042534] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
13
|
|
14
|
Abstract
Causal knowledge plays a crucial role in human thought, but the nature of causal representation and inference remains a puzzle. Can human causal inference be captured by relations of probabilistic dependency, or does it draw on richer forms of representation? This article explores this question by reviewing research in reasoning, decision making, various forms of judgment, and attribution. We endorse causal Bayesian networks as the best normative framework and as a productive guide to theory building. However, it is incomplete as an account of causal thinking. On the basis of a range of experimental work, we identify three hallmarks of causal reasoning-the role of mechanism, narrative, and mental simulation-all of which go beyond mere probabilistic knowledge. We propose that the hallmarks are closely related. Mental simulations are representations over time of mechanisms. When multiple actors are involved, these simulations are aggregated into narratives.
Collapse
Affiliation(s)
- Steven A Sloman
- Department of Cognitive, Linguistic, and Psychological Sciences, Brown University, Providence, Rhode Island 02912;
| | | |
Collapse
|