1
|
Vu M, Singhal B, Zeng S, Li JS. Data-driven control of oscillator networks with population-level measurement. CHAOS (WOODBURY, N.Y.) 2024; 34:033138. [PMID: 38526979 DOI: 10.1063/5.0191851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/16/2023] [Accepted: 02/28/2024] [Indexed: 03/27/2024]
Abstract
Controlling complex networks of nonlinear limit-cycle oscillators is an important problem pertinent to various applications in engineering and natural sciences. While in recent years the control of oscillator populations with comprehensive biophysical models or simplified models, e.g., phase models, has seen notable advances, learning appropriate controls directly from data without prior model assumptions or pre-existing data remains a challenging and less developed area of research. In this paper, we address this problem by leveraging the network's current dynamics to iteratively learn an appropriate control online without constructing a global model of the system. We illustrate through a range of numerical simulations that the proposed technique can effectively regulate synchrony in various oscillator networks after a small number of trials using only one input and one noisy population-level output measurement. We provide a theoretical analysis of our approach, illustrate its robustness to system variations, and compare its performance with existing model-based and data-driven approaches.
Collapse
Affiliation(s)
- Minh Vu
- Department of Electrical and Systems Engineering, Washington University in St Louis, St Louis, Missouri 63130, USA
| | - Bharat Singhal
- Department of Electrical and Systems Engineering, Washington University in St Louis, St Louis, Missouri 63130, USA
| | - Shen Zeng
- Department of Electrical and Systems Engineering, Washington University in St Louis, St Louis, Missouri 63130, USA
| | - Jr-Shin Li
- Department of Electrical and Systems Engineering, Washington University in St Louis, St Louis, Missouri 63130, USA
| |
Collapse
|
2
|
Chen Z, Anglea T, Zhang Y, Wang Y. Optimal synchronization in pulse-coupled oscillator networks using reinforcement learning. PNAS NEXUS 2023; 2:pgad102. [PMID: 37077885 PMCID: PMC10109446 DOI: 10.1093/pnasnexus/pgad102] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/25/2022] [Revised: 03/10/2023] [Accepted: 03/16/2023] [Indexed: 03/29/2023]
Abstract
Spontaneous synchronization is ubiquitous in natural and man-made systems. It underlies emergent behaviors such as neuronal response modulation and is fundamental to the coordination of robot swarms and autonomous vehicle fleets. Due to its simplicity and physical interpretability, pulse-coupled oscillators has emerged as one of the standard models for synchronization. However, existing analytical results for this model assume ideal conditions, including homogeneous oscillator frequencies and negligible coupling delays, as well as strict requirements on the initial phase distribution and the network topology. Using reinforcement learning, we obtain an optimal pulse-interaction mechanism (encoded in phase response function) that optimizes the probability of synchronization even in the presence of nonideal conditions. For small oscillator heterogeneities and propagation delays, we propose a heuristic formula for highly effective phase response functions that can be applied to general networks and unrestricted initial phase distributions. This allows us to bypass the need to relearn the phase response function for every new network.
Collapse
Affiliation(s)
- Ziqin Chen
- Department of Electrical & Computer Engineering, Clemson University, Clemson, SC 29634, USA
| | - Timothy Anglea
- Department of Electrical & Computer Engineering, Clemson University, Clemson, SC 29634, USA
| | - Yuanzhao Zhang
- Santa Fe Institute, 1399 Hyde Park Road, Santa Fe, NM 87501, USA
| | - Yongqiang Wang
- Department of Electrical & Computer Engineering, Clemson University, Clemson, SC 29634, USA
| |
Collapse
|
3
|
Vu M, Singhal B, Zeng S, Li JS. Data-Driven Control of Neuronal Networks with Population-Level Measurement. RESEARCH SQUARE 2023:rs.3.rs-2600572. [PMID: 36993505 PMCID: PMC10055505 DOI: 10.21203/rs.3.rs-2600572/v1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Controlling complex networks of nonlinear neurons is an important problem pertinent to various applications in engineering and natural sciences. While in recent years the control of neural populations with comprehensive biophysical models or simplified models, e.g., phase models, has seen notable advances, learning appropriate controls directly from data without any model assumptions remains a challenging and less developed area of research. In this paper, we address this problem by leveraging the network's local dynamics to iteratively learn an appropriate control without constructing a global model of the system. The proposed technique can effectively regulate synchrony in a neuronal network using only one input and one noisy population-level output measurement. We provide a theoretical analysis of our approach and illustrate its robustness to system variations and its generalizability to accommodate various physical constraints, such as charge-balanced inputs.
Collapse
Affiliation(s)
- Minh Vu
- Department of Electrical and Systems Engineering, Washington University in St. Louis, MO, USA
| | - Bharat Singhal
- Department of Electrical and Systems Engineering, Washington University in St. Louis, MO, USA
| | - Shen Zeng
- Department of Electrical and Systems Engineering, Washington University in St. Louis, MO, USA
| | - Jr-Shin Li
- Department of Electrical and Systems Engineering, Washington University in St. Louis, MO, USA
| |
Collapse
|
4
|
Rueckauer B, van Gerven M. An in-silico framework for modeling optimal control of neural systems. Front Neurosci 2023; 17:1141884. [PMID: 36968496 PMCID: PMC10030734 DOI: 10.3389/fnins.2023.1141884] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2023] [Accepted: 02/20/2023] [Indexed: 03/10/2023] Open
Abstract
Introduction Brain-machine interfaces have reached an unprecedented capacity to measure and drive activity in the brain, allowing restoration of impaired sensory, cognitive or motor function. Classical control theory is pushed to its limit when aiming to design control laws that are suitable for large-scale, complex neural systems. This work proposes a scalable, data-driven, unified approach to study brain-machine-environment interaction using established tools from dynamical systems, optimal control theory, and deep learning. Methods To unify the methodology, we define the environment, neural system, and prosthesis in terms of differential equations with learnable parameters, which effectively reduce to recurrent neural networks in the discrete-time case. Drawing on tools from optimal control, we describe three ways to train the system: Direct optimization of an objective function, oracle-based learning, and reinforcement learning. These approaches are adapted to different assumptions about knowledge of system equations, linearity, differentiability, and observability. Results We apply the proposed framework to train an in-silico neural system to perform tasks in a linear and a nonlinear environment, namely particle stabilization and pole balancing. After training, this model is perturbed to simulate impairment of sensor and motor function. We show how a prosthetic controller can be trained to restore the behavior of the neural system under increasing levels of perturbation. Discussion We expect that the proposed framework will enable rapid and flexible synthesis of control algorithms for neural prostheses that reduce the need for in-vivo testing. We further highlight implications for sparse placement of prosthetic sensor and actuator components.
Collapse
Affiliation(s)
- Bodo Rueckauer
- Department of Artificial Intelligence, Donders Institute for Brain, Cognition and Behavior, Radboud University, Nijmegen, Netherlands
| | | |
Collapse
|
5
|
Matchen TD, Moehlis J. Leveraging deep learning to control neural oscillators. BIOLOGICAL CYBERNETICS 2021; 115:219-235. [PMID: 33909165 DOI: 10.1007/s00422-021-00874-w] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/17/2020] [Accepted: 04/10/2021] [Indexed: 06/12/2023]
Abstract
Modulation of the firing times of neural oscillators has long been an important control objective, with applications including Parkinson's disease, Tourette's syndrome, epilepsy, and learning. One common goal for such modulation is desynchronization, wherein two or more oscillators are stimulated to transition from firing in phase with each other to firing out of phase. The optimization of such stimuli has been well studied, but this typically relies on either a reduction of the dimensionality of the system or complete knowledge of the parameters and state of the system. This limits the applicability of results to real problems in neural control. Here, we present a trained artificial neural network capable of accurately estimating the effects of square-wave stimuli on neurons using minimal output information from the neuron. We then apply the results of this network to solve several related control problems in desynchronization, including desynchronizing pairs of neurons and achieving clustered subpopulations of neurons in the presence of coupling and noise.
Collapse
Affiliation(s)
- Timothy D Matchen
- Department of Mechanical Engineering, University of California, Santa Barbara, CA, 93106, USA.
| | - Jeff Moehlis
- Department of Mechanical Engineering, Program in Dynamical Neuroscience, University of California, Santa Barbara, CA, 93106, USA
| |
Collapse
|
6
|
Tafazoli S, MacDowell CJ, Che Z, Letai KC, Steinhardt CR, Buschman TJ. Learning to control the brain through adaptive closed-loop patterned stimulation. J Neural Eng 2020; 17:056007. [PMID: 32927437 DOI: 10.1088/1741-2552/abb860] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2022]
Abstract
OBJECTIVE Stimulation of neural activity is an important scientific and clinical tool, causally testing hypotheses and treating neurodegenerative and neuropsychiatric diseases. However, current stimulation approaches cannot flexibly control the pattern of activity in populations of neurons. To address this, we developed a model-free, adaptive, closed-loop stimulation (ACLS) system that learns to use multi-site electrical stimulation to control the pattern of activity of a population of neurons. APPROACH The ACLS system combined multi-electrode electrophysiological recordings with multi-site electrical stimulation to simultaneously record the activity of a population of 5-15 multiunit neurons and deliver spatially-patterned electrical stimulation across 4-16 sites. Using a closed-loop learning system, ACLS iteratively updated the pattern of stimulation to reduce the difference between the observed neural response and a specific target pattern of firing rates in the recorded multiunits. MAIN RESULTS In silico and in vivo experiments showed ACLS learns to produce specific patterns of neural activity (in ∼15 min) and was robust to noise and drift in neural responses. In visual cortex of awake mice, ACLS learned electrical stimulation patterns that produced responses similar to the natural response evoked by visual stimuli. Similar to how repetition of a visual stimulus causes an adaptation in the neural response, the response to electrical stimulation was adapted when it was preceded by the associated visual stimulus. SIGNIFICANCE Our results show an ACLS system that can learn, in real-time, to generate specific patterns of neural activity. This work provides a framework for using model-free closed-loop learning to control neural activity.
Collapse
Affiliation(s)
- Sina Tafazoli
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ 08540, United States of America. Lead contact and corresponding author
| | | | | | | | | | | |
Collapse
|
7
|
Yu YC, Narayanan V, Ching S, Li JS. Learning to Control Neurons using Aggregated Measurements. PROCEEDINGS OF THE ... AMERICAN CONTROL CONFERENCE. AMERICAN CONTROL CONFERENCE 2020; 2020:4028-4033. [PMID: 38009125 PMCID: PMC10676147 DOI: 10.23919/acc45564.2020.9147426] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/28/2023]
Abstract
Controlling a population of neurons with one or a few control signals is challenging due to the severely underactuated nature of the control system and the inherent nonlinear dynamics of the neurons that are typically unknown. Control strategies that incorporate deep neural networks and machine learning techniques directly use data to learn a sequence of control actions for targeted manipulation of a population of neurons. However, these learning strategies inherently assume that perfect feedback data from each neuron at every sampling instant are available, and do not scale gracefully as the number of neurons in the population increases. As a result, the learning models need to be retrained whenever such a change occurs. In this work, we propose a learning strategy to design a control sequence by using population-level aggregated measurements and incorporate reinforcement learning techniques to find a (bounded, piecewise constant) control policy that fulfills the given control task. We demonstrate the feasibility of the proposed approach using numerical experiments on a finite population of nonlinear dynamical systems and canonical phase models that are widely used in neuroscience.
Collapse
Affiliation(s)
- Yao-Chi Yu
- Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, 63130, USA
| | - Vignesh Narayanan
- Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, 63130, USA
| | - ShiNung Ching
- Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, 63130, USA
- Division of Biology and Biomedical Sciences, Washington University in St. Louis, St. Louis, MO, 63130, USA
| | - Jr-Shin Li
- Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, 63130, USA
- Division of Biology and Biomedical Sciences, Washington University in St. Louis, St. Louis, MO, 63130, USA
| |
Collapse
|
8
|
Mitchell BA, Lauharatanahirun N, Garcia JO, Wymbs N, Grafton S, Vettel JM, Petzold LR. A Minimum Free Energy Model of Motor Learning. Neural Comput 2019; 31:1945-1963. [PMID: 31393824 DOI: 10.1162/neco_a_01219] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Even highly trained behaviors demonstrate variability, which is correlated with performance on current and future tasks. An objective of motor learning that is general enough to explain these phenomena has not been precisely formulated. In this six-week longitudinal learning study, participants practiced a set of motor sequences each day, and neuroimaging data were collected on days 1, 14, 28, and 42 to capture the neural correlates of the learning process. In our analysis, we first modeled the underlying neural and behavioral dynamics during learning. Our results demonstrate that the densities of whole-brain response, task-active regional response, and behavioral performance evolve according to a Fokker-Planck equation during the acquisition of a motor skill. We show that this implies that the brain concurrently optimizes the entropy of a joint density over neural response and behavior (as measured by sampling over multiple trials and subjects) and the expected performance under this density; we call this formulation of learning minimum free energy learning (MFEL). This model provides an explanation as to how behavioral variability can be tuned while simultaneously improving performance during learning. We then develop a novel variant of inverse reinforcement learning to retrieve the cost function optimized by the brain during the learning process, as well as the parameter used to tune variability. We show that this population-level analysis can be used to derive a learning objective that each subject optimizes during his or her study. In this way, MFEL effectively acts as a unifying principle, allowing users to precisely formulate learning objectives and infer their structure.
Collapse
Affiliation(s)
- B A Mitchell
- Department of Computer Science, University of California, Santa Barbara, Santa Barbara, CA 931056, U.S.A.
| | - N Lauharatanahirun
- Human Research and Engineering Directorate, The CCDC Army Research Laboratory, Aberdeen Proving Ground, MD 21005, U.S.A., and Annenberg School for Communication, University of Pennsylvania, Philadelphia, PA 19104, U.S.A.
| | - J O Garcia
- Human Research and Engineering Directorate, The CCDC Army Research Laboratory, Aberdeen Proving Ground, MD 21005, U.S.A., and Department of Bioengineering, University of Pennsylvania, Philadelphia, PA 19104, U.S.A.
| | - N Wymbs
- Department of Physical Medicine and Rehabilitation, Johns Hopkins Medical Institution, Baltimore, MD 21205, U.S.A.
| | - S Grafton
- Department of Psychological Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA 931056, U.S.A.
| | - J M Vettel
- Department of Psychological Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA 931056, U.S.A.; Human Research and Engineering Directorate, The CCDC Army Research Laboratory, Aberdeen Proving Ground, MD 21005, U.S.A.; and Department of Bioengineering, University of Pennsylvania, Philadelphia, PA 19104, U.S.A.
| | - L R Petzold
- Department of Computer Science and Department of Mechanical Engineering, University of California, Santa Barbara, Santa Barbara, CA 931056, U.S.A.
| |
Collapse
|