1
|
Chi J, Shu J, Li M, Mudappathi R, Jin Y, Lewis F, Boon A, Qin X, Liu L, Gu H. Artificial Intelligence in Metabolomics: A Current Review. Trends Analyt Chem 2024; 178:117852. [PMID: 39071116 PMCID: PMC11271759 DOI: 10.1016/j.trac.2024.117852] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/30/2024]
Abstract
Metabolomics and artificial intelligence (AI) form a synergistic partnership. Metabolomics generates large datasets comprising hundreds to thousands of metabolites with complex relationships. AI, aiming to mimic human intelligence through computational modeling, possesses extraordinary capabilities for big data analysis. In this review, we provide a recent overview of the methodologies and applications of AI in metabolomics studies in the context of systems biology and human health. We first introduce the AI concept, history, and key algorithms for machine learning and deep learning, summarizing their strengths and weaknesses. We then discuss studies that have successfully used AI across different aspects of metabolomic analysis, including analytical detection, data preprocessing, biomarker discovery, predictive modeling, and multi-omics data integration. Lastly, we discuss the existing challenges and future perspectives in this rapidly evolving field. Despite limitations and challenges, the combination of metabolomics and AI holds great promises for revolutionary advancements in enhancing human health.
Collapse
Affiliation(s)
- Jinhua Chi
- College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
- Center for Translational Science, Florida International University, Port St. Lucie, FL 34987, USA
| | - Jingmin Shu
- College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
- Center for Personalized Diagnostics, Biodesign Institute, Arizona State University, Tempe, AZ 85281, USA
| | - Ming Li
- Phoenix VA Health Care System, Phoenix, AZ 85012, USA
- University of Arizona College of Medicine, Phoenix, AZ 85004, USA
| | - Rekha Mudappathi
- College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
- Center for Personalized Diagnostics, Biodesign Institute, Arizona State University, Tempe, AZ 85281, USA
| | - Yan Jin
- Center for Translational Science, Florida International University, Port St. Lucie, FL 34987, USA
| | - Freeman Lewis
- Center for Translational Science, Florida International University, Port St. Lucie, FL 34987, USA
| | - Alexandria Boon
- Center for Translational Science, Florida International University, Port St. Lucie, FL 34987, USA
| | - Xiaoyan Qin
- College of Liberal Arts and Sciences, Arizona State University, Tempe, AZ 85281, USA
| | - Li Liu
- College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
- Center for Personalized Diagnostics, Biodesign Institute, Arizona State University, Tempe, AZ 85281, USA
| | - Haiwei Gu
- College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
- Center for Translational Science, Florida International University, Port St. Lucie, FL 34987, USA
| |
Collapse
|
2
|
Kundu P, Beura S, Mondal S, Das AK, Ghosh A. Machine learning for the advancement of genome-scale metabolic modeling. Biotechnol Adv 2024; 74:108400. [PMID: 38944218 DOI: 10.1016/j.biotechadv.2024.108400] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Revised: 05/13/2024] [Accepted: 06/23/2024] [Indexed: 07/01/2024]
Abstract
Constraint-based modeling (CBM) has evolved as the core systems biology tool to map the interrelations between genotype, phenotype, and external environment. The recent advancement of high-throughput experimental approaches and multi-omics strategies has generated a plethora of new and precise information from wide-ranging biological domains. On the other hand, the continuously growing field of machine learning (ML) and its specialized branch of deep learning (DL) provide essential computational architectures for decoding complex and heterogeneous biological data. In recent years, both multi-omics and ML have assisted in the escalation of CBM. Condition-specific omics data, such as transcriptomics and proteomics, helped contextualize the model prediction while analyzing a particular phenotypic signature. At the same time, the advanced ML tools have eased the model reconstruction and analysis to increase the accuracy and prediction power. However, the development of these multi-disciplinary methodological frameworks mainly occurs independently, which limits the concatenation of biological knowledge from different domains. Hence, we have reviewed the potential of integrating multi-disciplinary tools and strategies from various fields, such as synthetic biology, CBM, omics, and ML, to explore the biochemical phenomenon beyond the conventional biological dogma. How the integrative knowledge of these intersected domains has improved bioengineering and biomedical applications has also been highlighted. We categorically explained the conventional genome-scale metabolic model (GEM) reconstruction tools and their improvement strategies through ML paradigms. Further, the crucial role of ML and DL in omics data restructuring for GEM development has also been briefly discussed. Finally, the case-study-based assessment of the state-of-the-art method for improving biomedical and metabolic engineering strategies has been elaborated. Therefore, this review demonstrates how integrating experimental and in silico strategies can help map the ever-expanding knowledge of biological systems driven by condition-specific cellular information. This multiview approach will elevate the application of ML-based CBM in the biomedical and bioengineering fields for the betterment of society and the environment.
Collapse
Affiliation(s)
- Pritam Kundu
- School School of Energy Science and Engineering, Indian Institute of Technology Kharagpur, West Bengal 721302, India
| | - Satyajit Beura
- Department of Bioscience and Biotechnology, Indian Institute of Technology, Kharagpur, West Bengal 721302, India
| | - Suman Mondal
- P.K. Sinha Centre for Bioenergy and Renewables, Indian Institute of Technology Kharagpur, West Bengal 721302, India
| | - Amit Kumar Das
- Department of Bioscience and Biotechnology, Indian Institute of Technology, Kharagpur, West Bengal 721302, India
| | - Amit Ghosh
- School School of Energy Science and Engineering, Indian Institute of Technology Kharagpur, West Bengal 721302, India; P.K. Sinha Centre for Bioenergy and Renewables, Indian Institute of Technology Kharagpur, West Bengal 721302, India.
| |
Collapse
|
3
|
Redinbaugh MJ, Saalmann YB. Contributions of Basal Ganglia Circuits to Perception, Attention, and Consciousness. J Cogn Neurosci 2024; 36:1620-1642. [PMID: 38695762 PMCID: PMC11223727 DOI: 10.1162/jocn_a_02177] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/04/2024]
Abstract
Research into ascending sensory pathways and cortical networks has generated detailed models of perception. These same cortical regions are strongly connected to subcortical structures, such as the basal ganglia (BG), which have been conceptualized as playing key roles in reinforcement learning and action selection. However, because the BG amasses experiential evidence from higher and lower levels of cortical hierarchies, as well as higher-order thalamus, it is well positioned to dynamically influence perception. Here, we review anatomical, functional, and clinical evidence to demonstrate how the BG can influence perceptual processing and conscious states. This depends on the integrative relationship between cortex, BG, and thalamus, which allows contributions to sensory gating, predictive processing, selective attention, and representation of the temporal structure of events.
Collapse
Affiliation(s)
| | - Yuri B Saalmann
- University of Wisconsin-Madison
- Wisconsin National Primate Research Center
| |
Collapse
|
4
|
Krauhausen I, Griggs S, McCulloch I, den Toonder JMJ, Gkoupidenis P, van de Burgt Y. Bio-inspired multimodal learning with organic neuromorphic electronics for behavioral conditioning in robotics. Nat Commun 2024; 15:4765. [PMID: 38834541 DOI: 10.1038/s41467-024-48881-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 05/13/2024] [Indexed: 06/06/2024] Open
Abstract
Biological systems interact directly with the environment and learn by receiving multimodal feedback via sensory stimuli that shape the formation of internal neuronal representations. Drawing inspiration from biological concepts such as exploration and sensory processing that eventually lead to behavioral conditioning, we present a robotic system handling objects through multimodal learning. A small-scale organic neuromorphic circuit locally integrates and adaptively processes multimodal sensory stimuli, enabling the robot to interact intelligently with its surroundings. The real-time handling of sensory stimuli via low-voltage organic neuromorphic devices with synaptic functionality forms multimodal associative connections that lead to behavioral conditioning, and thus the robot learns to avoid potentially dangerous objects. This work demonstrates that adaptive neuro-inspired circuitry with multifunctional organic materials, can accommodate locally efficient bio-inspired learning for advancing intelligent robotics.
Collapse
Affiliation(s)
- Imke Krauhausen
- Institute for Complex Molecular Systems, Eindhoven University of Technology, Eindhoven, The Netherlands
- Microsystems, Department of Mechanical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
- Max Planck Institute for Polymer Research, Mainz, Germany
| | - Sophie Griggs
- Department of Chemistry, University of Oxford, Oxford, UK
| | - Iain McCulloch
- Department of Chemistry, University of Oxford, Oxford, UK
| | - Jaap M J den Toonder
- Institute for Complex Molecular Systems, Eindhoven University of Technology, Eindhoven, The Netherlands
- Microsystems, Department of Mechanical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
| | | | - Yoeri van de Burgt
- Institute for Complex Molecular Systems, Eindhoven University of Technology, Eindhoven, The Netherlands.
- Microsystems, Department of Mechanical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands.
| |
Collapse
|
5
|
Ferdous S, Shihab IF, Chowdhury R, Reuel NF. Reinforcement learning-guided control strategies for CAR T-cell activation and expansion. Biotechnol Bioeng 2024. [PMID: 38812405 DOI: 10.1002/bit.28753] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Revised: 04/12/2024] [Accepted: 05/12/2024] [Indexed: 05/31/2024]
Abstract
Reinforcement learning (RL), a subset of machine learning (ML), could optimize and control biomanufacturing processes, such as improved production of therapeutic cells. Here, the process of CAR T-cell activation by antigen-presenting beads and their subsequent expansion is formulated in silico. The simulation is used as an environment to train RL-agents to dynamically control the number of beads in culture to maximize the population of robust effector cells at the end of the culture. We make periodic decisions of incremental bead addition or complete removal. The simulation is designed to operate in OpenAI Gym, enabling testing of different environments, cell types, RL-agent algorithms, and state inputs to the RL-agent. RL-agent training is demonstrated with three different algorithms (PPO, A2C, and DQN), each sampling three different state input types (tabular, image, mixed); PPO-tabular performs best for this simulation environment. Using this approach, training of the RL-agent on different cell types is demonstrated, resulting in unique control strategies for each type. Sensitivity to input-noise (sensor performance), number of control step interventions, and advantages of pre-trained RL-agents are also evaluated. Therefore, we present an RL framework to maximize the population of robust effector cells in CAR T-cell therapy production.
Collapse
Affiliation(s)
- Sakib Ferdous
- Department of Chemical and Biological Engineering, Iowa State University, Ames, Iowa, USA
| | | | - Ratul Chowdhury
- Department of Chemical and Biological Engineering, Iowa State University, Ames, Iowa, USA
| | - Nigel F Reuel
- Department of Chemical and Biological Engineering, Iowa State University, Ames, Iowa, USA
| |
Collapse
|
6
|
Vassiliadis P, Beanato E, Popa T, Windel F, Morishita T, Neufeld E, Duque J, Derosiere G, Wessel MJ, Hummel FC. Non-invasive stimulation of the human striatum disrupts reinforcement learning of motor skills. Nat Hum Behav 2024:10.1038/s41562-024-01901-z. [PMID: 38811696 DOI: 10.1038/s41562-024-01901-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2022] [Accepted: 04/23/2024] [Indexed: 05/31/2024]
Abstract
Reinforcement feedback can improve motor learning, but the underlying brain mechanisms remain underexplored. In particular, the causal contribution of specific patterns of oscillatory activity within the human striatum is unknown. To address this question, we exploited a recently developed non-invasive deep brain stimulation technique called transcranial temporal interference stimulation (tTIS) during reinforcement motor learning with concurrent neuroimaging, in a randomized, sham-controlled, double-blind study. Striatal tTIS applied at 80 Hz, but not at 20 Hz, abolished the benefits of reinforcement on motor learning. This effect was related to a selective modulation of neural activity within the striatum. Moreover, 80 Hz, but not 20 Hz, tTIS increased the neuromodulatory influence of the striatum on frontal areas involved in reinforcement motor learning. These results show that tTIS can non-invasively and selectively modulate a striatal mechanism involved in reinforcement learning, expanding our tools for the study of causal relationships between deep brain structures and human behaviour.
Collapse
Affiliation(s)
- Pierre Vassiliadis
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, École Polytechnique Fédérale de Lausanne (EPFL), Geneva, Switzerland
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, EPFL Valais, Clinique Romande de Réadaptation, Sion, Switzerland
- Institute of Neuroscience, Université Catholique de Louvain, Brussels, Belgium
| | - Elena Beanato
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, École Polytechnique Fédérale de Lausanne (EPFL), Geneva, Switzerland
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, EPFL Valais, Clinique Romande de Réadaptation, Sion, Switzerland
| | - Traian Popa
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, École Polytechnique Fédérale de Lausanne (EPFL), Geneva, Switzerland
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, EPFL Valais, Clinique Romande de Réadaptation, Sion, Switzerland
| | - Fabienne Windel
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, École Polytechnique Fédérale de Lausanne (EPFL), Geneva, Switzerland
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, EPFL Valais, Clinique Romande de Réadaptation, Sion, Switzerland
| | - Takuya Morishita
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, École Polytechnique Fédérale de Lausanne (EPFL), Geneva, Switzerland
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, EPFL Valais, Clinique Romande de Réadaptation, Sion, Switzerland
| | - Esra Neufeld
- Foundation for Research on Information Technologies in Society, Zurich, Switzerland
| | - Julie Duque
- Institute of Neuroscience, Université Catholique de Louvain, Brussels, Belgium
| | - Gerard Derosiere
- Institute of Neuroscience, Université Catholique de Louvain, Brussels, Belgium
- Lyon Neuroscience Research Center, Impact Team, Inserm U1028, CNRS UMR5292, Lyon 1 University, Bron, France
| | - Maximilian J Wessel
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, École Polytechnique Fédérale de Lausanne (EPFL), Geneva, Switzerland
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, EPFL Valais, Clinique Romande de Réadaptation, Sion, Switzerland
- Department of Neurology, University Hospital Würzburg, Würzburg, Germany
| | - Friedhelm C Hummel
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, École Polytechnique Fédérale de Lausanne (EPFL), Geneva, Switzerland.
- Defitech Chair of Clinical Neuroengineering, Neuro-X Institute, EPFL Valais, Clinique Romande de Réadaptation, Sion, Switzerland.
- Clinical Neuroscience, University of Geneva Medical School, Geneva, Switzerland.
| |
Collapse
|
7
|
Bose S, Banerjee S, Kumar S, Saha A, Nandy D, Hazra S. Review of applications of artificial intelligence (AI) methods in crop research. J Appl Genet 2024; 65:225-240. [PMID: 38216788 DOI: 10.1007/s13353-023-00826-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2023] [Revised: 12/23/2023] [Accepted: 12/26/2023] [Indexed: 01/14/2024]
Abstract
Sophisticated and modern crop improvement techniques can bridge the gap for feeding the ever-increasing population. Artificial intelligence (AI) refers to the simulation of human intelligence in machines, which refers to the application of computational algorithms, machine learning (ML) and deep learning (DL) techniques. This is aimed to generalise patterns and relationships from historical data, employing various mathematical optimisation techniques thus making prediction models for facilitating selection of superior genotypes. These techniques are less resource intensive and can solve the problem based on the analysis of large-scale phenotypic datasets. ML for genomic selection (GS) uses high-throughput genotyping technologies to gather genetic information on a large number of markers across the genome. The prediction of GS models is based on the mathematical relation between genotypic and phenotypic data from the training population. ML techniques have emerged as powerful tools for genome editing through analysing large-scale genomic data and facilitating the development of accurate prediction models. Precise phenotyping is a prerequisite to advance crop breeding for solving agricultural production-related issues. ML algorithms can solve this problem through generating predictive models, based on the analysis of large-scale phenotypic datasets. DL models also have the potential reliability of precise phenotyping. This review provides a comprehensive overview on various ML and DL models, their applications, potential to enhance the efficiency, specificity and safety towards advanced crop improvement protocols such as genomic selection, genome editing, along with phenotypic prediction to promote accelerated breeding.
Collapse
Affiliation(s)
- Suvojit Bose
- Department of Vegetables and Spice Crops, Uttar Banga Krishi Viswavidyalaya, Pundibari, Cooch Behar, 736165, West Bengal, India
| | | | - Soumya Kumar
- School of Agricultural Sciences, JIS University, Kolkata, 700109, West Bengal, India
| | - Akash Saha
- School of Agricultural Sciences, JIS University, Kolkata, 700109, West Bengal, India
| | - Debalina Nandy
- School of Agricultural Sciences, JIS University, Kolkata, 700109, West Bengal, India
| | - Soham Hazra
- Department of Agriculture, Brainware University, Barasat, 700125, West Bengal, India.
| |
Collapse
|
8
|
Nikolaidou A, Mougkogiannis P, Adamatzky A. Electroactive composite biofilms integrating Kombucha, Chlorella and synthetic proteinoid Proto-Brains. ROYAL SOCIETY OPEN SCIENCE 2024; 11:240238. [PMID: 39076784 PMCID: PMC11285679 DOI: 10.1098/rsos.240238] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 07/31/2024]
Abstract
In this study, we present electroactive biofilms made from a combination of Kombucha zoogleal mats and thermal proteinoids. These biofilms have potential applications in unconventional computing and robotic skin. Proteinoids are synthesized by thermally polymerizing amino acids, resulting in the formation of synthetic protocells that display electrical signalling similar to neurons. By incorporating proteinoids into Kombucha zoogleal cellulose mats, hydrogel biofilms can be created that have the ability to efficiently transfer charges, perform sensory transduction and undergo processing. We conducted a study on the memfractance and memristance behaviours of composite biofilms, showcasing their capacity to carry out unconventional computing operations. The porous nanostructure and electroactivity of the biofilm create a biocompatible interface that can be used to record and stimulate neuronal networks. In addition to in vitro neuronal interfaces, these soft electroactive biofilms show potential as components for bioinspired robotics, smart wearables, unconventional computing devices and adaptive biorobotic systems. Kombucha-proteinoids composite films are a highly customizable material that can be synthesized to suit specific needs. These films belong to a unique category of 'living' materials, as they have the ability to support cellular systems and improve bioelectronic functionality. This makes them an exciting prospect in various applications. Ongoing efforts are currently being directed towards enhancing the compositional tuning of conductivity, signal processing and integration within hybrid bioelectronic circuits.
Collapse
Affiliation(s)
- Anna Nikolaidou
- Unconventional Computing Laboratory, University of the West of England, Bristol, UK
- School of Architecture and Environment, University of the West of England, Bristol, UK
| | | | - Andrew Adamatzky
- Unconventional Computing Laboratory, University of the West of England, Bristol, UK
| |
Collapse
|
9
|
Hu M, Scott C. Toward the development of a molecular toolkit for the microbial remediation of per-and polyfluoroalkyl substances. Appl Environ Microbiol 2024; 90:e0015724. [PMID: 38477530 PMCID: PMC11022551 DOI: 10.1128/aem.00157-24] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/14/2024] Open
Abstract
Per- and polyfluoroalkyl substances (PFAS) are highly fluorinated synthetic organic compounds that have been used extensively in various industries owing to their unique properties. The PFAS family encompasses diverse classes, with only a fraction being commercially relevant. These substances are found in the environment, including in water sources, soil, and wildlife, leading to human exposure and fueling concerns about potential human health impacts. Although PFAS degradation is challenging, biodegradation offers a promising, eco-friendly solution. Biodegradation has been effective for a variety of organic contaminants but is yet to be successful for PFAS due to a paucity of identified microbial species capable of transforming these compounds. Recent studies have investigated PFAS biotransformation and fluoride release; however, the number of specific microorganisms and enzymes with demonstrable activity with PFAS remains limited. This review discusses enzymes that could be used in PFAS metabolism, including haloacid dehalogenases, reductive dehalogenases, cytochromes P450, alkane and butane monooxygenases, peroxidases, laccases, desulfonases, and the mechanisms of microbial resistance to intracellular fluoride. Finally, we emphasize the potential of enzyme and microbial engineering to advance PFAS degradation strategies and provide insights for future research in this field.
Collapse
Affiliation(s)
- Miao Hu
- CSIRO Environment, Black Mountain Science and Innovation Park, Canberra, ACT, Australia
| | - Colin Scott
- CSIRO Environment, Black Mountain Science and Innovation Park, Canberra, ACT, Australia
| |
Collapse
|
10
|
Dhoble S, Wu TH, Kenry. Decoding Nanomaterial-Biosystem Interactions through Machine Learning. Angew Chem Int Ed Engl 2024; 63:e202318380. [PMID: 38687554 DOI: 10.1002/anie.202318380] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2023] [Indexed: 05/02/2024]
Abstract
The interactions between biosystems and nanomaterials regulate most of their theranostic and nanomedicine applications. These nanomaterial-biosystem interactions are highly complex and influenced by a number of entangled factors, including but not limited to the physicochemical features of nanomaterials, the types and characteristics of the interacting biosystems, and the properties of the surrounding microenvironments. Over the years, different experimental approaches coupled with computational modeling have revealed important insights into these interactions, although many outstanding questions remain unanswered. The emergence of machine learning has provided a timely and unique opportunity to revisit nanomaterial-biosystem interactions and to further push the boundary of this field. This minireview highlights the development and use of machine learning to decode nanomaterial-biosystem interactions and provides our perspectives on the current challenges and potential opportunities in this field.
Collapse
Affiliation(s)
- Sagar Dhoble
- Department of Pharmacology and Toxicology, R. Ken Coit College of Pharmacy, University of Arizona, Tucson, AZ 85721, USA
| | - Tzu-Hsien Wu
- Department of Pharmacology and Toxicology, R. Ken Coit College of Pharmacy, University of Arizona, Tucson, AZ 85721, USA
| | - Kenry
- Department of Pharmacology and Toxicology, R. Ken Coit College of Pharmacy, University of Arizona, Tucson, AZ 85721, USA
- University of Arizona Cancer Center, University of Arizona, Tucson, AZ 85721, USA
- BIO5 Institute, University of Arizona, Tucson, AZ 85721, USA
| |
Collapse
|
11
|
Bernklau TW, Righetti B, Mehrke LS, Jacob SN. Striatal dopamine signals reflect perceived cue-action-outcome associations in mice. Nat Neurosci 2024; 27:747-757. [PMID: 38291283 PMCID: PMC11001585 DOI: 10.1038/s41593-023-01567-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2022] [Accepted: 12/21/2023] [Indexed: 02/01/2024]
Abstract
Striatal dopamine drives associative learning by acting as a teaching signal. Much work has focused on simple learning paradigms, including Pavlovian and instrumental learning. However, higher cognition requires that animals generate internal concepts of their environment, where sensory stimuli, actions and outcomes become flexibly associated. Here, we performed fiber photometry dopamine measurements across the striatum of male mice as they learned cue-action-outcome associations based on implicit and changing task rules. Reinforcement learning models of the behavioral and dopamine data showed that rule changes lead to adjustments of learned cue-action-outcome associations. After rule changes, mice discarded learned associations and reset outcome expectations. Cue- and outcome-triggered dopamine signals became uncoupled and dependent on the adopted behavioral strategy. As mice learned the new association, coupling between cue- and outcome-triggered dopamine signals and task performance re-emerged. Our results suggest that dopaminergic reward prediction errors reflect an agent's perceived locus of control.
Collapse
Affiliation(s)
- Tobias W Bernklau
- Translational Neurotechnology Laboratory, Department of Neurosurgery, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
- Graduate School of Systemic Neurosciences, Ludwig-Maximilians-University Munich, Munich, Germany
| | - Beatrice Righetti
- Translational Neurotechnology Laboratory, Department of Neurosurgery, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | - Leonie S Mehrke
- Translational Neurotechnology Laboratory, Department of Neurosurgery, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | - Simon N Jacob
- Translational Neurotechnology Laboratory, Department of Neurosurgery, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany.
| |
Collapse
|
12
|
Sandini G, Sciutti A, Morasso P. Artificial cognition vs. artificial intelligence for next-generation autonomous robotic agents. Front Comput Neurosci 2024; 18:1349408. [PMID: 38585280 PMCID: PMC10995397 DOI: 10.3389/fncom.2024.1349408] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Accepted: 02/20/2024] [Indexed: 04/09/2024] Open
Abstract
The trend in industrial/service robotics is to develop robots that can cooperate with people, interacting with them in an autonomous, safe and purposive way. These are the fundamental elements characterizing the fourth and the fifth industrial revolutions (4IR, 5IR): the crucial innovation is the adoption of intelligent technologies that can allow the development of cyber-physical systems, similar if not superior to humans. The common wisdom is that intelligence might be provided by AI (Artificial Intelligence), a claim that is supported more by media coverage and commercial interests than by solid scientific evidence. AI is currently conceived in a quite broad sense, encompassing LLMs and a lot of other things, without any unifying principle, but self-motivating for the success in various areas. The current view of AI robotics mostly follows a purely disembodied approach that is consistent with the old-fashioned, Cartesian mind-body dualism, reflected in the software-hardware distinction inherent to the von Neumann computing architecture. The working hypothesis of this position paper is that the road to the next generation of autonomous robotic agents with cognitive capabilities requires a fully brain-inspired, embodied cognitive approach that avoids the trap of mind-body dualism and aims at the full integration of Bodyware and Cogniware. We name this approach Artificial Cognition (ACo) and ground it in Cognitive Neuroscience. It is specifically focused on proactive knowledge acquisition based on bidirectional human-robot interaction: the practical advantage is to enhance generalization and explainability. Moreover, we believe that a brain-inspired network of interactions is necessary for allowing humans to cooperate with artificial cognitive agents, building a growing level of personal trust and reciprocal accountability: this is clearly missing, although actively sought, in current AI. The ACo approach is a work in progress that can take advantage of a number of research threads, some of them antecedent the early attempts to define AI concepts and methods. In the rest of the paper we will consider some of the building blocks that need to be re-visited in a unitary framework: the principles of developmental robotics, the methods of action representation with prospection capabilities, and the crucial role of social interaction.
Collapse
Affiliation(s)
| | | | - Pietro Morasso
- Italian Institute of Technology, Cognitive Architecture for Collaborative Technologies (CONTACT) and Robotics, Brain and Cognitive Sciences (RBCS) Research Units, Genoa, Italy
| |
Collapse
|
13
|
Colas JT, O’Doherty JP, Grafton ST. Active reinforcement learning versus action bias and hysteresis: control with a mixture of experts and nonexperts. PLoS Comput Biol 2024; 20:e1011950. [PMID: 38552190 PMCID: PMC10980507 DOI: 10.1371/journal.pcbi.1011950] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2023] [Accepted: 02/26/2024] [Indexed: 04/01/2024] Open
Abstract
Active reinforcement learning enables dynamic prediction and control, where one should not only maximize rewards but also minimize costs such as of inference, decisions, actions, and time. For an embodied agent such as a human, decisions are also shaped by physical aspects of actions. Beyond the effects of reward outcomes on learning processes, to what extent can modeling of behavior in a reinforcement-learning task be complicated by other sources of variance in sequential action choices? What of the effects of action bias (for actions per se) and action hysteresis determined by the history of actions chosen previously? The present study addressed these questions with incremental assembly of models for the sequential choice data from a task with hierarchical structure for additional complexity in learning. With systematic comparison and falsification of computational models, human choices were tested for signatures of parallel modules representing not only an enhanced form of generalized reinforcement learning but also action bias and hysteresis. We found evidence for substantial differences in bias and hysteresis across participants-even comparable in magnitude to the individual differences in learning. Individuals who did not learn well revealed the greatest biases, but those who did learn accurately were also significantly biased. The direction of hysteresis varied among individuals as repetition or, more commonly, alternation biases persisting from multiple previous actions. Considering that these actions were button presses with trivial motor demands, the idiosyncratic forces biasing sequences of action choices were robust enough to suggest ubiquity across individuals and across tasks requiring various actions. In light of how bias and hysteresis function as a heuristic for efficient control that adapts to uncertainty or low motivation by minimizing the cost of effort, these phenomena broaden the consilient theory of a mixture of experts to encompass a mixture of expert and nonexpert controllers of behavior.
Collapse
Affiliation(s)
- Jaron T. Colas
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, California, United States of America
- Division of the Humanities and Social Sciences, California Institute of Technology, Pasadena, California, United States of America
- Computation and Neural Systems Program, California Institute of Technology, Pasadena, California, United States of America
| | - John P. O’Doherty
- Division of the Humanities and Social Sciences, California Institute of Technology, Pasadena, California, United States of America
- Computation and Neural Systems Program, California Institute of Technology, Pasadena, California, United States of America
| | - Scott T. Grafton
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, California, United States of America
| |
Collapse
|
14
|
Kozielska M, Weissing FJ. A neural network model for the evolution of learning in changing environments. PLoS Comput Biol 2024; 20:e1011840. [PMID: 38289971 PMCID: PMC10857588 DOI: 10.1371/journal.pcbi.1011840] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Revised: 02/09/2024] [Accepted: 01/18/2024] [Indexed: 02/01/2024] Open
Abstract
Learning from past experience is an important adaptation and theoretical models may help to understand its evolution. Many of the existing models study simple phenotypes and do not consider the mechanisms underlying learning while the more complex neural network models often make biologically unrealistic assumptions and rarely consider evolutionary questions. Here, we present a novel way of modelling learning using small neural networks and a simple, biology-inspired learning algorithm. Learning affects only part of the network, and it is governed by the difference between expectations and reality. We use this model to study the evolution of learning under various environmental conditions and different scenarios for the trade-off between exploration (learning) and exploitation (foraging). Efficient learning readily evolves in our individual-based simulations. However, in line with previous studies, the evolution of learning is less likely in relatively constant environments, where genetic adaptation alone can lead to efficient foraging, or in short-lived organisms that cannot afford to spend much of their lifetime on exploration. Once learning does evolve, the characteristics of the learning strategy (i.e. the duration of the learning period and the learning rate) and the average performance after learning are surprisingly little affected by the frequency and/or magnitude of environmental change. In contrast, an organism's lifespan and the distribution of resources in the environment have a clear effect on the evolved learning strategy: a shorter lifespan or a broader resource distribution lead to fewer learning episodes and larger learning rates. Interestingly, a longer learning period does not always lead to better performance, indicating that the evolved neural networks differ in the effectiveness of learning. Overall, however, we show that a biologically inspired, yet relatively simple, learning mechanism can evolve to lead to an efficient adaptation in a changing environment.
Collapse
Affiliation(s)
- Magdalena Kozielska
- Groningen Institute for Evolutionary Life Sciences, University of Groningen, Groningen, The Netherlands
| | - Franz J. Weissing
- Groningen Institute for Evolutionary Life Sciences, University of Groningen, Groningen, The Netherlands
| |
Collapse
|
15
|
Seifert G, Sealander A, Marzen S, Levin M. From reinforcement learning to agency: Frameworks for understanding basal cognition. Biosystems 2024; 235:105107. [PMID: 38128873 DOI: 10.1016/j.biosystems.2023.105107] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Revised: 12/17/2023] [Accepted: 12/17/2023] [Indexed: 12/23/2023]
Abstract
Organisms play, explore, and mimic those around them. Is there a purpose to this behavior? Are organisms just behaving, or are they trying to achieve goals? We believe this is a false dichotomy. To that end, to understand organisms, we attempt to unify two approaches for understanding complex agents, whether evolved or engineered. We argue that formalisms describing multiscale competencies and goal-directedness in biology (e.g., TAME), and reinforcement learning (RL), can be combined in a symbiotic framework. While RL has been largely focused on higher-level organisms and robots of high complexity, TAME is naturally capable of describing lower-level organisms and minimal agents as well. We propose several novel questions that come from using RL/TAME to understand biology as well as ones that come from using biology to formulate new theory in AI. We hope that the research programs proposed in this piece shape future efforts to understand biological organisms and also future efforts to build artificial agents.
Collapse
Affiliation(s)
- Gabriella Seifert
- Department of Physics, University of Colorado, Boulder, CO 80309, USA; W. M. Keck Science Department, Pitzer, Scripps, and Claremont McKenna College, Claremont, CA 91711, USA
| | - Ava Sealander
- Department of Electrical Engineering, School of Engineering and Applied Sciences, Columbia University, New York, NY 10027, USA; W. M. Keck Science Department, Pitzer, Scripps, and Claremont McKenna College, Claremont, CA 91711, USA
| | - Sarah Marzen
- W. M. Keck Science Department, Pitzer, Scripps, and Claremont McKenna College, Claremont, CA 91711, USA.
| | - Michael Levin
- Department of Biology, Tufts University, Medford, MA 02155, USA; Allen Discovery Center at Tufts University, Medford, MA 02155, USA
| |
Collapse
|
16
|
Chafai N, Hayah I, Houaga I, Badaoui B. A review of machine learning models applied to genomic prediction in animal breeding. Front Genet 2023; 14:1150596. [PMID: 37745853 PMCID: PMC10516561 DOI: 10.3389/fgene.2023.1150596] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Accepted: 08/22/2023] [Indexed: 09/26/2023] Open
Abstract
The advent of modern genotyping technologies has revolutionized genomic selection in animal breeding. Large marker datasets have shown several drawbacks for traditional genomic prediction methods in terms of flexibility, accuracy, and computational power. Recently, the application of machine learning models in animal breeding has gained a lot of interest due to their tremendous flexibility and their ability to capture patterns in large noisy datasets. Here, we present a general overview of a handful of machine learning algorithms and their application in genomic prediction to provide a meta-picture of their performance in genomic estimated breeding values estimation, genotype imputation, and feature selection. Finally, we discuss a potential adoption of machine learning models in genomic prediction in developing countries. The results of the reviewed studies showed that machine learning models have indeed performed well in fitting large noisy data sets and modeling minor nonadditive effects in some of the studies. However, sometimes conventional methods outperformed machine learning models, which confirms that there's no universal method for genomic prediction. In summary, machine learning models have great potential for extracting patterns from single nucleotide polymorphism datasets. Nonetheless, the level of their adoption in animal breeding is still low due to data limitations, complex genetic interactions, a lack of standardization and reproducibility, and the lack of interpretability of machine learning models when trained with biological data. Consequently, there is no remarkable outperformance of machine learning methods compared to traditional methods in genomic prediction. Therefore, more research should be conducted to discover new insights that could enhance livestock breeding programs.
Collapse
Affiliation(s)
- Narjice Chafai
- Laboratory of Biodiversity, Ecology, and Genome, Department of Biology, Faculty of Sciences, Mohammed V University in Rabat, Rabat, Morocco
| | - Ichrak Hayah
- Laboratory of Biodiversity, Ecology, and Genome, Department of Biology, Faculty of Sciences, Mohammed V University in Rabat, Rabat, Morocco
| | - Isidore Houaga
- Centre for Tropical Livestock Genetics and Health, The Roslin Institute, Royal (Dick) School of Veterinary Medicine, The University of Edinburgh, Edinburgh, United Kingdom
- The Roslin Institute, Royal (Dick) School of Veterinary Studies, University of Edinburgh, Edinburgh, United Kingdom
| | - Bouabid Badaoui
- Laboratory of Biodiversity, Ecology, and Genome, Department of Biology, Faculty of Sciences, Mohammed V University in Rabat, Rabat, Morocco
- African Sustainable Agriculture Research Institute (ASARI), Mohammed VI Polytechnic University (UM6P), Laayoune, Morocco
| |
Collapse
|
17
|
Deng Y, Song D, Ni J, Qing H, Quan Z. Reward prediction error in learning-related behaviors. Front Neurosci 2023; 17:1171612. [PMID: 37662112 PMCID: PMC10471312 DOI: 10.3389/fnins.2023.1171612] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2023] [Accepted: 07/31/2023] [Indexed: 09/05/2023] Open
Abstract
Learning is a complex process, during which our opinions and decisions are easily changed due to unexpected information. But the neural mechanism underlying revision and correction during the learning process remains unclear. For decades, prediction error has been regarded as the core of changes to perception in learning, even driving the learning progress. In this article, we reviewed the concept of reward prediction error, and the encoding mechanism of dopaminergic neurons and the related neural circuities. We also discussed the relationship between reward prediction error and learning-related behaviors, including reversal learning. We then demonstrated the evidence of reward prediction error signals in several neurological diseases, including Parkinson's disease and addiction. These observations may help to better understand the regulatory mechanism of reward prediction error in learning-related behaviors.
Collapse
Affiliation(s)
- Yujun Deng
- Key Laboratory of Molecular Medicine and Biotherapy, School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Da Song
- Key Laboratory of Molecular Medicine and Biotherapy, School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Junjun Ni
- Key Laboratory of Molecular Medicine and Biotherapy, School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Hong Qing
- Key Laboratory of Molecular Medicine and Biotherapy, School of Life Science, Beijing Institute of Technology, Beijing, China
- Department of Biology, Shenzhen MSU-BIT University, Shenzhen, China
| | - Zhenzhen Quan
- Key Laboratory of Molecular Medicine and Biotherapy, School of Life Science, Beijing Institute of Technology, Beijing, China
| |
Collapse
|
18
|
Brahim Belhaouari S, Shakeel MB, Erbad A, Oflaz Z, Kassoul K. Bird's Eye View feature selection for high-dimensional data. Sci Rep 2023; 13:13303. [PMID: 37587137 PMCID: PMC10432524 DOI: 10.1038/s41598-023-39790-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Accepted: 07/31/2023] [Indexed: 08/18/2023] Open
Abstract
In machine learning, an informative dataset is crucial for accurate predictions. However, high dimensional data often contains irrelevant features, outliers, and noise, which can negatively impact model performance and consume computational resources. To tackle this challenge, the Bird's Eye View (BEV) feature selection technique is introduced. This approach is inspired by the natural world, where a bird searches for important features in a sparse dataset, similar to how a bird search for sustenance in a sprawling jungle. BEV incorporates elements of Evolutionary Algorithms with a Genetic Algorithm to maintain a population of top-performing agents, Dynamic Markov Chain to steer the movement of agents in the search space, and Reinforcement Learning to reward and penalize agents based on their progress. The proposed strategy in this paper leads to improved classification performance and a reduced number of features compared to conventional methods, as demonstrated by outperforming state-of-the-art feature selection techniques across multiple benchmark datasets.
Collapse
Affiliation(s)
- Samir Brahim Belhaouari
- Division of Information and Computing Technology, College of Science and Engineering, Hamad Bin Khalifa University, Doha, Qatar.
| | - Mohammed Bilal Shakeel
- Division of Information and Computing Technology, College of Science and Engineering, Hamad Bin Khalifa University, Doha, Qatar
| | - Aiman Erbad
- Division of Information and Computing Technology, College of Science and Engineering, Hamad Bin Khalifa University, Doha, Qatar
| | - Zarina Oflaz
- Department of Industrial Engineering, Faculty of Engineering and Natural Sciences, KTO Karatay University, Konya, Turkey
| | - Khelil Kassoul
- Geneva School of Economics and Management (GSEM), University of Geneva, 1211, Geneva, Switzerland.
| |
Collapse
|
19
|
Chapman M, Xu L, Lapeyrolerie M, Boettiger C. Bridging adaptive management and reinforcement learning for more robust decisions. Philos Trans R Soc Lond B Biol Sci 2023; 378:20220195. [PMID: 37246377 PMCID: PMC10225849 DOI: 10.1098/rstb.2022.0195] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2022] [Accepted: 02/01/2023] [Indexed: 05/30/2023] Open
Abstract
From out-competing grandmasters in chess to informing high-stakes healthcare decisions, emerging methods from artificial intelligence are increasingly capable of making complex and strategic decisions in diverse, high-dimensional and uncertain situations. But can these methods help us devise robust strategies for managing environmental systems under great uncertainty? Here we explore how reinforcement learning (RL), a subfield of artificial intelligence, approaches decision problems through a lens similar to adaptive environmental management: learning through experience to gradually improve decisions with updated knowledge. We review where RL holds promise for improving evidence-informed adaptive management decisions even when classical optimization methods are intractable and discuss technical and social issues that arise when applying RL to adaptive management problems in the environmental domain. Our synthesis suggests that environmental management and computer science can learn from one another about the practices, promises and perils of experience-based decision-making. This article is part of the theme issue 'Detecting and attributing the causes of biodiversity change: needs, gaps and solutions'.
Collapse
Affiliation(s)
- Melissa Chapman
- Department of Environmental Science, Policy, and Management, University of California, Berkeley, CA 94720, USA
| | - Lily Xu
- School of Engineering and Applied Sciences, Harvard University, Cambridge, MA 02138, USA
| | - Marcus Lapeyrolerie
- Department of Environmental Science, Policy, and Management, University of California, Berkeley, CA 94720, USA
| | - Carl Boettiger
- Department of Environmental Science, Policy, and Management, University of California, Berkeley, CA 94720, USA
| |
Collapse
|
20
|
Yang LWY, Ng WY, Lei X, Tan SCY, Wang Z, Yan M, Pargi MK, Zhang X, Lim JS, Gunasekeran DV, Tan FCP, Lee CE, Yeo KK, Tan HK, Ho HSS, Tan BWB, Wong TY, Kwek KYC, Goh RSM, Liu Y, Ting DSW. Development and testing of a multi-lingual Natural Language Processing-based deep learning system in 10 languages for COVID-19 pandemic crisis: A multi-center study. Front Public Health 2023; 11:1063466. [PMID: 36860378 PMCID: PMC9968846 DOI: 10.3389/fpubh.2023.1063466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2022] [Accepted: 01/26/2023] [Indexed: 02/16/2023] Open
Abstract
Purpose The COVID-19 pandemic has drastically disrupted global healthcare systems. With the higher demand for healthcare and misinformation related to COVID-19, there is a need to explore alternative models to improve communication. Artificial Intelligence (AI) and Natural Language Processing (NLP) have emerged as promising solutions to improve healthcare delivery. Chatbots could fill a pivotal role in the dissemination and easy accessibility of accurate information in a pandemic. In this study, we developed a multi-lingual NLP-based AI chatbot, DR-COVID, which responds accurately to open-ended, COVID-19 related questions. This was used to facilitate pandemic education and healthcare delivery. Methods First, we developed DR-COVID with an ensemble NLP model on the Telegram platform (https://t.me/drcovid_nlp_chatbot). Second, we evaluated various performance metrics. Third, we evaluated multi-lingual text-to-text translation to Chinese, Malay, Tamil, Filipino, Thai, Japanese, French, Spanish, and Portuguese. We utilized 2,728 training questions and 821 test questions in English. Primary outcome measurements were (A) overall and top 3 accuracies; (B) Area Under the Curve (AUC), precision, recall, and F1 score. Overall accuracy referred to a correct response for the top answer, whereas top 3 accuracy referred to an appropriate response for any one answer amongst the top 3 answers. AUC and its relevant matrices were obtained from the Receiver Operation Characteristics (ROC) curve. Secondary outcomes were (A) multi-lingual accuracy; (B) comparison to enterprise-grade chatbot systems. The sharing of training and testing datasets on an open-source platform will also contribute to existing data. Results Our NLP model, utilizing the ensemble architecture, achieved overall and top 3 accuracies of 0.838 [95% confidence interval (CI): 0.826-0.851] and 0.922 [95% CI: 0.913-0.932] respectively. For overall and top 3 results, AUC scores of 0.917 [95% CI: 0.911-0.925] and 0.960 [95% CI: 0.955-0.964] were achieved respectively. We achieved multi-linguicism with nine non-English languages, with Portuguese performing the best overall at 0.900. Lastly, DR-COVID generated answers more accurately and quickly than other chatbots, within 1.12-2.15 s across three devices tested. Conclusion DR-COVID is a clinically effective NLP-based conversational AI chatbot, and a promising solution for healthcare delivery in the pandemic era.
Collapse
Affiliation(s)
| | - Wei Yan Ng
- Singapore National Eye Center, Singapore Eye Research Institute, Singapore, Singapore,Duke-National University of Singapore Medical School, National University of Singapore, Singapore, Singapore
| | - Xiaofeng Lei
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore, Singapore
| | - Shaun Chern Yuan Tan
- Singapore National Eye Center, Singapore Eye Research Institute, Singapore, Singapore
| | - Zhaoran Wang
- Duke-National University of Singapore Medical School, National University of Singapore, Singapore, Singapore
| | - Ming Yan
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore, Singapore
| | - Mohan Kashyap Pargi
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore, Singapore
| | - Xiaoman Zhang
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore, Singapore
| | - Jane Sujuan Lim
- Singapore National Eye Center, Singapore Eye Research Institute, Singapore, Singapore
| | - Dinesh Visva Gunasekeran
- Singapore National Eye Center, Singapore Eye Research Institute, Singapore, Singapore,Duke-National University of Singapore Medical School, National University of Singapore, Singapore, Singapore
| | | | - Chen Ee Lee
- Division of Digital Strategy Office, Singapore Health Services, Singapore, Singapore
| | - Khung Keong Yeo
- Office of Innovation and Transformation, Singapore Health Services, Singapore, Singapore
| | - Hiang Khoon Tan
- Department of Head and Neck Surgery, Singapore General Hospital, Singapore, Singapore
| | - Henry Sun Sien Ho
- Duke-National University of Singapore Medical School, National University of Singapore, Singapore, Singapore,Department of Urology, Singapore General Hospital, Singapore, Singapore
| | - Benedict Wee Bor Tan
- Division of Digital Strategy Office, Singapore Health Services, Singapore, Singapore
| | - Tien Yin Wong
- Singapore National Eye Center, Singapore Eye Research Institute, Singapore, Singapore,Duke-National University of Singapore Medical School, National University of Singapore, Singapore, Singapore,Tsinghua Medicine, Tsinghua University, Beijing, China
| | | | - Rick Siow Mong Goh
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore, Singapore
| | - Yong Liu
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore, Singapore
| | - Daniel Shu Wei Ting
- Singapore National Eye Center, Singapore Eye Research Institute, Singapore, Singapore,Duke-National University of Singapore Medical School, National University of Singapore, Singapore, Singapore,*Correspondence: Daniel Shu Wei Ting ✉
| |
Collapse
|
21
|
Scott-Fordsmand JJ, Amorim MJB. Using Machine Learning to make nanomaterials sustainable. THE SCIENCE OF THE TOTAL ENVIRONMENT 2023; 859:160303. [PMID: 36410486 DOI: 10.1016/j.scitotenv.2022.160303] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Revised: 11/06/2022] [Accepted: 11/15/2022] [Indexed: 06/16/2023]
Abstract
Sustainable development is a key challenge for contemporary human societies; failure to achieve sustainability could threaten human survival. In this review article, we illustrate how Machine Learning (ML) could support more sustainable development, covering the basics of data gathering through each step of the Environmental Risk Assessment (ERA). The literature provides several examples showing how ML can be employed in most steps of a typical ERA.A key observation is that there are currently no clear guidance for using such autonomous technologies in ERAs or which standards/checks are required. Steering thus seems to be the most important task for supporting the use of ML in the ERA of nano- and smart-materials. Resources should be devoted to developing a strategy for implementing ML in ERA with a strong emphasis on data foundations, methodologies, and the related sensitivities/uncertainties. We should recognise historical errors and biases (e.g., in data) to avoid embedding them during ML programming.
Collapse
Affiliation(s)
| | - Mónica J B Amorim
- Department of Biology & CESAM, University of Aveiro, 3810-193 Aveiro, Portugal.
| |
Collapse
|
22
|
Burk DC, Averbeck BB. Environmental uncertainty and the advantage of impulsive choice strategies. PLoS Comput Biol 2023; 19:e1010873. [PMID: 36716320 PMCID: PMC9910799 DOI: 10.1371/journal.pcbi.1010873] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2022] [Revised: 02/09/2023] [Accepted: 01/15/2023] [Indexed: 02/01/2023] Open
Abstract
Choice impulsivity is characterized by the choice of immediate, smaller reward options over future, larger reward options, and is often thought to be associated with negative life outcomes. However, some environments make future rewards more uncertain, and in these environments impulsive choices can be beneficial. Here we examined the conditions under which impulsive vs. non-impulsive decision strategies would be advantageous. We used Markov Decision Processes (MDPs) to model three common decision-making tasks: Temporal Discounting, Information Sampling, and an Explore-Exploit task. We manipulated environmental variables to create circumstances where future outcomes were relatively uncertain. We then manipulated the discount factor of an MDP agent, which affects the value of immediate versus future rewards, to model impulsive and non-impulsive behavior. This allowed us to examine the performance of impulsive and non-impulsive agents in more or less predictable environments. In Temporal Discounting, we manipulated the transition probability to delayed rewards and found that the agent with the lower discount factor (i.e. the impulsive agent) collected more average reward than the agent with a higher discount factor (the non-impulsive agent) by selecting immediate reward options when the probability of receiving the future reward was low. In the Information Sampling task, we manipulated the amount of information obtained with each sample. When sampling led to small information gains, the impulsive MDP agent collected more average reward than the non-impulsive agent. Third, in the Explore-Exploit task, we manipulated the substitution rate for novel options. When the substitution rate was high, the impulsive agent again performed better than the non-impulsive agent, as it explored the novel options less and instead exploited options with known reward values. The results of these analyses show that impulsivity can be advantageous in environments that are unexpectedly uncertain.
Collapse
Affiliation(s)
- Diana C. Burk
- Laboratory of Neuropsychology, National Institute of Mental Health, National Institutes of Health, Bethesda, Maryland, United States of America
| | - Bruno B. Averbeck
- Laboratory of Neuropsychology, National Institute of Mental Health, National Institutes of Health, Bethesda, Maryland, United States of America
- * E-mail:
| |
Collapse
|
23
|
Modularity in Nervous Systems—a Key to Efficient Adaptivity for Deep Reinforcement Learning. Cognit Comput 2023. [DOI: 10.1007/s12559-022-10080-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/21/2023]
Abstract
AbstractModularity as observed in biological systems has proven valuable for guiding classical motor theories towards good answers about action selection and execution. New challenges arise when we turn to learning: Trying to scale current computational models, such as deep reinforcement learning (DRL), to action spaces, input dimensions, and time horizons seen in biological systems still faces severe obstacles unless vast amounts of training data are available. This leads to the question: does biological modularity also hold an important key for better answers to obtain efficient adaptivity for deep reinforcement learning? We review biological experimental work on modularity in biological motor control and link this with current examples of (deep) RL approaches. Analyzing outcomes of simulation studies, we show that these approaches benefit from forms of modularization as found in biological systems. We identify three different strands of modularity exhibited in biological control systems. Two of them—modularity in state (i) and in action (ii) spaces—appear as a consequence of local interconnectivity (as in reflexes) and are often modulated by higher levels in a control hierarchy. A third strand arises from chunking of action elements along a (iii) temporal dimension. Usually interacting in an overarching spatio-temporal hierarchy of the overall system, the three strands offer major “factors” decomposing the entire modularity structure. We conclude that modularity with its above strands can provide an effective prior for DRL approaches to speed up learning considerably and making learned controllers more robust and adaptive.
Collapse
|
24
|
Fan C, Yao L, Zhang J, Zhen Z, Wu X. Advanced Reinforcement Learning and Its Connections with Brain Neuroscience. RESEARCH (WASHINGTON, D.C.) 2023; 6:0064. [PMID: 36939448 PMCID: PMC10017102 DOI: 10.34133/research.0064] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/27/2022] [Accepted: 01/10/2023] [Indexed: 01/22/2023]
Abstract
In recent years, brain science and neuroscience have greatly propelled the innovation of computer science. In particular, knowledge from the neurobiology and neuropsychology of the brain revolutionized the development of reinforcement learning (RL) by providing novel interpretable mechanisms of how the brain achieves intelligent and efficient decision making. Triggered by this, there has been a boom in research about advanced RL algorithms that are built upon the inspirations of brain neuroscience. In this work, to further strengthen the bidirectional link between the 2 communities and especially promote the research on modern RL technology, we provide a comprehensive survey of recent advances in the area of brain-inspired/related RL algorithms. We start with basis theories of RL, and present a concise introduction to brain neuroscience related to RL. Then, we classify these advanced RL methodologies into 3 categories according to different connections of the brain, i.e., micro-neural activity, macro-brain structure, and cognitive function. Each category is further surveyed by presenting several modern RL algorithms along with their mathematical models, correlations with the brain, and open issues. Finally, we introduce several important applications of RL algorithms, followed by the discussions of challenges and opportunities for future research.
Collapse
Affiliation(s)
- Chaoqiong Fan
- School of Artificial Intelligence,
Beijing Normal University, Beijing, China
| | - Li Yao
- School of Artificial Intelligence,
Beijing Normal University, Beijing, China
| | - Jiacai Zhang
- School of Artificial Intelligence,
Beijing Normal University, Beijing, China
| | - Zonglei Zhen
- Faculty of Psychology,
Beijing Normal University, Beijing, China
| | - Xia Wu
- School of Artificial Intelligence,
Beijing Normal University, Beijing, China
- Address correspondence to:
| |
Collapse
|
25
|
Transferring policy of deep reinforcement learning from simulation to reality for robotics. NAT MACH INTELL 2022. [DOI: 10.1038/s42256-022-00573-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/16/2022]
|
26
|
Song Z, Wang Y, Qian P, Song S, Coenen F, Jiang Z, Su J. From deterministic to stochastic: an interpretable stochastic model-free reinforcement learning framework for portfolio optimization. APPL INTELL 2022; 53:15188-15203. [PMID: 36405345 PMCID: PMC9651127 DOI: 10.1007/s10489-022-04217-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/28/2022] [Indexed: 11/13/2022]
Abstract
As a fundamental problem in algorithmic trading, portfolio optimization aims to maximize the cumulative return by continuously investing in various financial derivatives within a given time period. Recent years have witnessed the transformation from traditional machine learning trading algorithms to reinforcement learning algorithms due to their superior nature of sequential decision making. However, the exponential growth of the imperfect and noisy financial data that is supposedly leveraged by the deterministic strategy in reinforcement learning, makes it increasingly challenging for one to continuously obtain a profitable portfolio. Thus, in this work, we first reconstruct several deterministic and stochastic reinforcement algorithms as benchmarks. On this basis, we introduce a risk-aware reward function to balance the risk and return. Importantly, we propose a novel interpretable stochastic reinforcement learning framework which tailors a stochastic policy parameterized by Gaussian Mixtures and a distributional critic realized by quantiles for the problem of portfolio optimization. In our experiment, the proposed algorithm demonstrates its superior performance on U.S. market stocks with a 63.1% annual rate of return while at the same time reducing the market value max drawdown by 10% when back-testing during the stock market crash around March 2020.
Collapse
Affiliation(s)
- Zitao Song
- Department of Mathematical Sciences, Xi’an Jiaotong-Liverpool University, Suzhou, China
| | - Yining Wang
- Department of Mathematical Sciences, Xi’an Jiaotong-Liverpool University, Suzhou, China
| | - Pin Qian
- Department of Computer Sciences, Xi’an Jiaotong-Liverpool University, Suzhou, China
| | - Sifan Song
- Department of Mathematical Sciences, Xi’an Jiaotong-Liverpool University, Suzhou, China
| | - Frans Coenen
- Department of Computer Sciences, University of Liverpool, Liverpool, UK
| | - Zhengyong Jiang
- Department of Mathematical Sciences, Xi’an Jiaotong-Liverpool University, Suzhou, China
| | - Jionglong Su
- School of AI and Advanced Computing, XJTLU Entrepreneur College (Taicang), Xi’an Jiaotong-Liverpool University, Suzhou, China
| |
Collapse
|
27
|
Ujfalussy BB, Orbán G. Sampling motion trajectories during hippocampal theta sequences. eLife 2022; 11:74058. [DOI: 10.7554/elife.74058] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2021] [Accepted: 09/28/2022] [Indexed: 11/06/2022] Open
Abstract
Efficient planning in complex environments requires that uncertainty associated with current inferences and possible consequences of forthcoming actions is represented. Representation of uncertainty has been established in sensory systems during simple perceptual decision making tasks but it remains unclear if complex cognitive computations such as planning and navigation are also supported by probabilistic neural representations. Here, we capitalized on gradually changing uncertainty along planned motion trajectories during hippocampal theta sequences to capture signatures of uncertainty representation in population responses. In contrast with prominent theories, we found no evidence of encoding parameters of probability distributions in the momentary population activity recorded in an open-field navigation task in rats. Instead, uncertainty was encoded sequentially by sampling motion trajectories randomly and efficiently in subsequent theta cycles from the distribution of potential trajectories. Our analysis is the first to demonstrate that the hippocampus is well equipped to contribute to optimal planning by representing uncertainty.
Collapse
Affiliation(s)
- Balazs B Ujfalussy
- Laboratory of Biological Computation, Institute of Experimental Medicine
- Laboratory of Neuronal Signalling, Institute of Experimental Medicine, Budapest
| | - Gergő Orbán
- Computational Systems Neuroscience Lab, Wigner Research Center for Physics, Budapest
| |
Collapse
|
28
|
Sawant Y, Kundu JN, Radhakrishnan VB, Sridharan D. A Midbrain Inspired Recurrent Neural Network Model for Robust Change Detection. J Neurosci 2022; 42:8262-8283. [PMID: 36123120 PMCID: PMC9653281 DOI: 10.1523/jneurosci.0164-22.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Revised: 07/26/2022] [Accepted: 07/30/2022] [Indexed: 11/21/2022] Open
Abstract
We present a biologically inspired recurrent neural network (RNN) that efficiently detects changes in natural images. The model features sparse, topographic connectivity (st-RNN), closely modeled on the circuit architecture of a "midbrain attention network." We deployed the st-RNN in a challenging change blindness task, in which changes must be detected in a discontinuous sequence of images. Compared with a conventional RNN, the st-RNN learned 9x faster and achieved state-of-the-art performance with 15x fewer connections. An analysis of low-dimensional dynamics revealed putative circuit mechanisms, including a critical role for a global inhibitory (GI) motif, for successful change detection. The model reproduced key experimental phenomena, including midbrain neurons' sensitivity to dynamic stimuli, neural signatures of stimulus competition, as well as hallmark behavioral effects of midbrain microstimulation. Finally, the model accurately predicted human gaze fixations in a change blindness experiment, surpassing state-of-the-art saliency-based methods. The st-RNN provides a novel deep learning model for linking neural computations underlying change detection with psychophysical mechanisms.SIGNIFICANCE STATEMENT For adaptive survival, our brains must be able to accurately and rapidly detect changing aspects of our visual world. We present a novel deep learning model, a sparse, topographic recurrent neural network (st-RNN), that mimics the neuroanatomy of an evolutionarily conserved "midbrain attention network." The st-RNN achieved robust change detection in challenging change blindness tasks, outperforming conventional RNN architectures. The model also reproduced hallmark experimental phenomena, both neural and behavioral, reported in seminal midbrain studies. Lastly, the st-RNN outperformed state-of-the-art models at predicting human gaze fixations in a laboratory change blindness experiment. Our deep learning model may provide important clues about key mechanisms by which the brain efficiently detects changes.
Collapse
Affiliation(s)
- Yash Sawant
- Centre for Neuroscience, Indian Institute of Science, Bangalore 560012, India
| | - Jogendra Nath Kundu
- Department of Computational and Data Sciences, Indian Institute of Science, Bangalore 560012, India
| | | | - Devarajan Sridharan
- Centre for Neuroscience, Indian Institute of Science, Bangalore 560012, India
- Department of Computer Science and Automation, Indian Institute of Science, Bangalore 560012, India
| |
Collapse
|
29
|
Ran M, Li J, Xie L. Reinforcement-Learning-Based Disturbance Rejection Control for Uncertain Nonlinear Systems. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:9621-9633. [PMID: 33729973 DOI: 10.1109/tcyb.2021.3060736] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
This article investigates the reinforcement-learning (RL)-based disturbance rejection control for uncertain nonlinear systems having nonsimple nominal models. An extended state observer (ESO) is first designed to estimate the system state and the total uncertainty, which represents the perturbation to the nominal system dynamics. Based on the output of the observer, the control compensates for the total uncertainty in real time, and simultaneously, online approximates the optimal policy for the compensated system using a simulation of experience-based RL technique. Rigorous theoretical analysis is given to show the practical convergence of the system state to the origin and the developed policy to the ideal optimal policy. It is worth mentioning that the widely used restrictive persistence of excitation (PE) condition is not required in the established framework. Simulation results are presented to illustrate the effectiveness of the proposed method.
Collapse
|
30
|
Namgung SD, Kim RM, Lim YC, Lee JW, Cho NH, Kim H, Huh JS, Rhee H, Nah S, Song MK, Kwon JY, Nam KT. Circularly polarized light-sensitive, hot electron transistor with chiral plasmonic nanoparticles. Nat Commun 2022; 13:5081. [PMID: 36038547 PMCID: PMC9424280 DOI: 10.1038/s41467-022-32721-2] [Citation(s) in RCA: 19] [Impact Index Per Article: 9.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2021] [Accepted: 08/12/2022] [Indexed: 11/09/2022] Open
Abstract
The quantitative detection of circularly polarized light (CPL) is necessary in next-generation optical communication carrying high-density information and in phase-controlled displays exhibiting volumetric imaging. In the current technology, multiple pixels of different wavelengths and polarizers are required, inevitably resulting in high loss and low detection efficiency. Here, we demonstrate a highly efficient CPL-detecting transistor composed of chiral plasmonic nanoparticles with a high Khun's dissymmetry (g-factor) of 0.2 and a high mobility conducting oxide of InGaZnO. The device successfully distinguished the circular polarization state and displayed an unprecedented photoresponsivity of over 1 A/W under visible CPL excitation. This observation is mainly attributed to the hot electron generation in chiral plasmonic nanoparticles and to the effective collection of hot electrons in the oxide semiconducting transistor. Such characteristics further contribute to opto-neuromorphic operation and the artificial nervous system based on the device successfully performs image classification work. We anticipate that our strategy will aid in the rational design and fabrication of a high-performance CPL detector and opto-neuromorphic operation with a chiral plasmonic structure depending on the wavelength and circular polarization state.
Collapse
Affiliation(s)
- Seok Daniel Namgung
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea.,Soft Foundry, Seoul National University, Seoul, 08826, Republic of Korea
| | - Ryeong Myeong Kim
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Yae-Chan Lim
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Jong Woo Lee
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Nam Heon Cho
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Hyeohn Kim
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Jin-Suk Huh
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea.,Soft Foundry, Seoul National University, Seoul, 08826, Republic of Korea
| | - Hanju Rhee
- Seoul Center, Korea Basic Science Institute, Seoul, 02841, Republic of Korea
| | - Sanghee Nah
- Seoul Center, Korea Basic Science Institute, Seoul, 02841, Republic of Korea
| | - Min-Kyu Song
- School of Integrated Technology, Yonsei University, Incheon, 21983, Republic of Korea
| | - Jang-Yeon Kwon
- School of Integrated Technology, Yonsei University, Incheon, 21983, Republic of Korea
| | - Ki Tae Nam
- Department of Materials Science and Engineering, Seoul National University, Seoul, 08826, Republic of Korea. .,Soft Foundry, Seoul National University, Seoul, 08826, Republic of Korea.
| |
Collapse
|
31
|
Dvoretskii S, Gong Z, Gupta A, Parent J, Alicea B. Braitenberg Vehicles as Developmental Neurosimulation. ARTIFICIAL LIFE 2022; 28:369-395. [PMID: 35881679 DOI: 10.1162/artl_a_00384] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Connecting brain and behavior is a longstanding issue in the areas of behavioral science, artificial intelligence, and neurobiology. As is standard among models of artificial and biological neural networks, an analogue of the fully mature brain is presented as a blank slate. However, this does not consider the realities of biological development and developmental learning. Our purpose is to model the development of an artificial organism that exhibits complex behaviors. We introduce three alternate approaches to demonstrate how developmental embodied agents can be implemented. The resulting developmental Braitenberg vehicles (dBVs) will generate behaviors ranging from stimulus responses to group behavior that resembles collective motion. We will situate this work in the domain of artificial brain networks along with broader themes such as embodied cognition, feedback, and emergence. Our perspective is exemplified by three software instantiations that demonstrate how a BV-genetic algorithm hybrid model, a multisensory Hebbian learning model, and multi-agent approaches can be used to approach BV development. We introduce use cases such as optimized spatial cognition (vehicle-genetic algorithm hybrid model), hinges connecting behavioral and neural models (multisensory Hebbian learning model), and cumulative classification (multi-agent approaches). In conclusion, we consider future applications of the developmental neurosimulation approach.
Collapse
Affiliation(s)
| | | | | | | | - Bradly Alicea
- Orthogonal Research and Education Laboratory
- OpenWorm Foundation.
| |
Collapse
|
32
|
Yin L, Li Y. Fuzzy vector reinforcement learning algorithm for generation control of power systems considering flywheel energy storage. Appl Soft Comput 2022. [DOI: 10.1016/j.asoc.2022.109149] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
33
|
Janssen M, LeWarne C, Burk D, Averbeck BB. Hierarchical Reinforcement Learning, Sequential Behavior, and the Dorsal Frontostriatal System. J Cogn Neurosci 2022; 34:1307-1325. [PMID: 35579977 PMCID: PMC9274316 DOI: 10.1162/jocn_a_01869] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
To effectively behave within ever-changing environments, biological agents must learn and act at varying hierarchical levels such that a complex task may be broken down into more tractable subtasks. Hierarchical reinforcement learning (HRL) is a computational framework that provides an understanding of this process by combining sequential actions into one temporally extended unit called an option. However, there are still open questions within the HRL framework, including how options are formed and how HRL mechanisms might be realized within the brain. In this review, we propose that the existing human motor sequence literature can aid in understanding both of these questions. We give specific emphasis to visuomotor sequence learning tasks such as the discrete sequence production task and the M × N (M steps × N sets) task to understand how hierarchical learning and behavior manifest across sequential action tasks as well as how the dorsal cortical-subcortical circuitry could support this kind of behavior. This review highlights how motor chunks within a motor sequence can function as HRL options. Furthermore, we aim to merge findings from motor sequence literature with reinforcement learning perspectives to inform experimental design in each respective subfield.
Collapse
Affiliation(s)
| | | | - Diana Burk
- National Institute of Mental Health, Bethesda, MD
| | | |
Collapse
|
34
|
Fuentes-Cortés LF, Flores-Tlacuahuac A, Nigam KDP. Machine Learning Algorithms Used in PSE Environments: A Didactic Approach and Critical Perspective. Ind Eng Chem Res 2022. [DOI: 10.1021/acs.iecr.2c00335] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
- Luis Fabián Fuentes-Cortés
- Departamento de Ingeniería Química, Tecnologico Nacional de México - Instituto Tecnológico de Celaya, Celaya, Guanajuato 38010, Mexico
| | - Antonio Flores-Tlacuahuac
- Tecnologico de Monterrey, Escuela de Ingeniería y Ciencias Ave. Eugenio Garza Sada 2501, Monterrey, N.L. 64849, Mexico
| | - Krishna D. P. Nigam
- Tecnologico de Monterrey, Escuela de Ingeniería y Ciencias Ave. Eugenio Garza Sada 2501, Monterrey, N.L. 64849, Mexico
- Department of Chemical Engineering, Indian Institute of Technology Delhi 600036, India
| |
Collapse
|
35
|
Almeida H, Tsang A, Diallo AB. Improving candidate Biosynthetic Gene Clusters in fungi through reinforcement learning. Bioinformatics 2022; 38:3984-3991. [PMID: 35762945 PMCID: PMC9364373 DOI: 10.1093/bioinformatics/btac420] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2021] [Revised: 05/17/2022] [Accepted: 06/23/2022] [Indexed: 12/24/2022] Open
Abstract
MOTIVATION Precise identification of Biosynthetic Gene Clusters (BGCs) is a challenging task. Performance of BGC discovery tools is limited by their capacity to accurately predict components belonging to candidate BGCs, often overestimating cluster boundaries. To support optimizing the composition and boundaries of candidate BGCs, we propose reinforcement learning approach relying on protein domains and functional annotations from expert curated BGCs. RESULTS The proposed reinforcement learning method aims to improve candidate BGCs obtained with state-of-the-art tools. It was evaluated on candidate BGCs obtained for two fungal genomes, Aspergillus niger and Aspergillus nidulans. The results highlight an improvement of the gene precision by above 15% for TOUCAN, fungiSMASH and DeepBGC; and cluster precision by above 25% for fungiSMASH and DeepBCG, allowing these tools to obtain almost perfect precision in cluster prediction. This can pave the way of optimizing current prediction of candidate BGCs in fungi, while minimizing the curation effort required by domain experts. AVAILABILITY AND IMPLEMENTATION https://github.com/bioinfoUQAM/RL-bgc-components. SUPPLEMENTARY INFORMATION Supplementary data are available at Bioinformatics online.
Collapse
Affiliation(s)
- Hayda Almeida
- Departement d’Informatique, UQAM, Montréal, QC H2X 3Y7, Canada,Centre for Structural and Functional Genomics, Concordia University, Montréal, QC H4B 1R6, Canada,Laboratoire d’Algèbre, de Combinatoire, et d’Informatique Mathématique (LACIM), UQAM, Montréal, QC H2X 3Y, Canada
| | - Adrian Tsang
- Departement d’Informatique, UQAM, Montréal, QC H2X 3Y7, Canada,Centre for Structural and Functional Genomics, Concordia University, Montréal, QC H4B 1R6, Canada
| | | |
Collapse
|
36
|
Wu J, Huang Z, Huang W, Lv C. Prioritized Experience-Based Reinforcement Learning With Human Guidance for Autonomous Driving. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; PP:855-869. [PMID: 35687630 DOI: 10.1109/tnnls.2022.3177685] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Reinforcement learning (RL) requires skillful definition and remarkable computational efforts to solve optimization and control problems, which could impair its prospect. Introducing human guidance into RL is a promising way to improve learning performance. In this article, a comprehensive human guidance-based RL framework is established. A novel prioritized experience replay mechanism that adapts to human guidance in the RL process is proposed to boost the efficiency and performance of the RL algorithm. To relieve the heavy workload on human participants, a behavior model is established based on an incremental online learning method to mimic human actions. We design two challenging autonomous driving tasks for evaluating the proposed algorithm. Experiments are conducted to access the training and testing performance and learning mechanism of the proposed algorithm. Comparative results against the state-of-the-art methods suggest the advantages of our algorithm in terms of learning efficiency, performance, and robustness.
Collapse
|
37
|
Suhaimi A, Lim AWH, Chia XW, Li C, Makino H. Representation learning in the artificial and biological neural networks underlying sensorimotor integration. SCIENCE ADVANCES 2022; 8:eabn0984. [PMID: 35658033 PMCID: PMC9166289 DOI: 10.1126/sciadv.abn0984] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/03/2021] [Accepted: 04/18/2022] [Indexed: 06/15/2023]
Abstract
The integration of deep learning and theories of reinforcement learning (RL) is a promising avenue to explore novel hypotheses on reward-based learning and decision-making in humans and other animals. Here, we trained deep RL agents and mice in the same sensorimotor task with high-dimensional state and action space and studied representation learning in their respective neural networks. Evaluation of thousands of neural network models with extensive hyperparameter search revealed that learning-dependent enrichment of state-value and policy representations of the task-performance-optimized deep RL agent closely resembled neural activity of the posterior parietal cortex (PPC). These representations were critical for the task performance in both systems. PPC neurons also exhibited representations of the internally defined subgoal, a feature of deep RL algorithms postulated to improve sample efficiency. Such striking resemblance between the artificial and biological networks and their functional convergence in sensorimotor integration offers new opportunities to better understand respective intelligent systems.
Collapse
|
38
|
Simultaneous Release of Silver Ions and 10–Undecenoic Acid from Silver Iron–Oxide Nanoparticles Impregnated Membranes. MEMBRANES 2022; 12:membranes12060557. [PMID: 35736264 PMCID: PMC9227798 DOI: 10.3390/membranes12060557] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/05/2022] [Revised: 05/01/2022] [Accepted: 05/23/2022] [Indexed: 01/27/2023]
Abstract
The bio-medical benefits of silver ions and 10–undecenoic acid in various chemical-pharmaceutical preparations are indisputable, thus justifying numerous research studies on delayed and/or controlled release. This paper presents the effect of the polymer matrix in the simultaneous release of silver ions and 10–undecenoic acid in an aqueous medium of controlled pH and ionic strength. The study took into consideration polymeric matrices consisting of cellulose acetate (CA) and polysulfone (PSf), which were impregnated with oxide nanoparticles containing silver and 10–undecenoic acid. The studied oxide nanoparticles are nanoparticles of iron and silver oxides obtained by an accessible electrochemical method. The obtained results show that silver can be released, simultaneously with 10–undecenoic acid, from an impregnated polymeric membrane, at concentrations that ensure the biocidal and fungicidal capacity. Concentrations of active substances can be controlled by choosing the polymer matrix or, in some cases, by changing the pH of the target medium. In the studied case, higher concentrations of silver ions are released from the polysulfone matrix, while higher concentrations of 10–undecenoic acid are released from the cellulose acetate matrix. The results of the study show that a correlation can be established between the two released target substances, which is dependent on the solubility of the organic compound in the aqueous medium and the interaction of this compound with the silver ions. The ability of 10–undecenoic acid to interact with the silver ion, both through the carboxyl and alkene groups, contributes to the increase in the content of the silver ions transported in the aqueous medium.
Collapse
|
39
|
Khoshkhou M, Montakhab A. Optimal reinforcement learning near the edge of a synchronization transition. Phys Rev E 2022; 105:044312. [PMID: 35590577 DOI: 10.1103/physreve.105.044312] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2021] [Accepted: 03/30/2022] [Indexed: 06/15/2023]
Abstract
Recent experimental and theoretical studies have indicated that the putative criticality of cortical dynamics may correspond to a synchronization phase transition. The critical dynamics near such a critical point needs further investigation specifically when compared to the critical behavior near the standard absorbing state phase transition. Since the phenomena of learning and self-organized criticality (SOC) at the edge of synchronization transition can emerge jointly in spiking neural networks due to the presence of spike-timing dependent plasticity (STDP), it is tempting to ask the following: what is the relationship between synchronization and learning in neural networks? Further, does learning benefit from SOC at the edge of synchronization transition? In this paper, we intend to address these important issues. Accordingly, we construct a biologically inspired model of a cognitive system which learns to perform stimulus-response tasks. We train this system using a reinforcement learning rule implemented through dopamine-modulated STDP. We find that the system exhibits a continuous transition from synchronous to asynchronous neural oscillations upon increasing the average axonal time delay. We characterize the learning performance of the system and observe that it is optimized near the synchronization transition. We also study neuronal avalanches in the system and provide evidence that optimized learning is achieved in a slightly supercritical state.
Collapse
Affiliation(s)
- Mahsa Khoshkhou
- Department of Physics, College of Sciences, Shiraz University, Shiraz 71946-84795, Iran
| | - Afshin Montakhab
- Department of Physics, College of Sciences, Shiraz University, Shiraz 71946-84795, Iran
| |
Collapse
|
40
|
Bloem B, Huda R, Amemori KI, Abate AS, Krishna G, Wilson AL, Carter CW, Sur M, Graybiel AM. Multiplexed action-outcome representation by striatal striosome-matrix compartments detected with a mouse cost-benefit foraging task. Nat Commun 2022; 13:1541. [PMID: 35318343 PMCID: PMC8941061 DOI: 10.1038/s41467-022-28983-5] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2021] [Accepted: 02/15/2022] [Indexed: 11/17/2022] Open
Abstract
Learning about positive and negative outcomes of actions is crucial for survival and underpinned by conserved circuits including the striatum. How associations between actions and outcomes are formed is not fully understood, particularly when the outcomes have mixed positive and negative features. We developed a novel foraging (‘bandit’) task requiring mice to maximize rewards while minimizing punishments. By 2-photon Ca++ imaging, we monitored activity of visually identified anterodorsal striatal striosomal and matrix neurons. We found that action-outcome associations for reward and punishment were encoded in parallel in partially overlapping populations. Single neurons could, for one action, encode outcomes of opposing valence. Striosome compartments consistently exhibited stronger representations of reinforcement outcomes than matrix, especially for high reward or punishment prediction errors. These findings demonstrate multiplexing of action-outcome contingencies by single identified striatal neurons and suggest that striosomal neurons are particularly important in action-outcome learning. The role that the striatum plays in tracking the association between actions and combinations of rewarding and aversive outcomes remains unclear. Here, by using both calcium imaging in mice and reinforcement learning models, the authors find that individual striatal neurons can encode associations between actions and multiple, sometimes conflicting, outcomes.
Collapse
Affiliation(s)
- Bernard Bloem
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Sinopia Biosciences, 600W Broadway, Suite 700, San Diego, CA, 92101, USA
| | - Rafiq Huda
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Picower Institute for Learning and Memory, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Department of Cell Biology and Neuroscience, WM Keck Center for Collaborative Neuroscience, Rutgers University, 604 Allison Rd, Piscataway, NJ, 08854, USA
| | - Ken-Ichi Amemori
- Institute for the Advanced Study of Human Biology, Kyoto University, Yoshida Konoe-cho, Sakyo-ku, Kyoto, 606-8501, Japan
| | - Alex S Abate
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA
| | - Gayathri Krishna
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA
| | - Anna L Wilson
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA
| | - Cody W Carter
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA
| | - Mriganka Sur
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.,Picower Institute for Learning and Memory, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA
| | - Ann M Graybiel
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA. .,Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, MA, 02139, USA.
| |
Collapse
|
41
|
Abstract
A hallmark of adaptation in humans and other animals is our ability to control how we think and behave across different settings. Research has characterized the various forms cognitive control can take-including enhancement of goal-relevant information, suppression of goal-irrelevant information, and overall inhibition of potential responses-and has identified computations and neural circuits that underpin this multitude of control types. Studies have also identified a wide range of situations that elicit adjustments in control allocation (e.g., those eliciting signals indicating an error or increased processing conflict), but the rules governing when a given situation will give rise to a given control adjustment remain poorly understood. Significant progress has recently been made on this front by casting the allocation of control as a decision-making problem. This approach has developed unifying and normative models that prescribe when and how a change in incentives and task demands will result in changes in a given form of control. Despite their successes, these models, and the experiments that have been developed to test them, have yet to face their greatest challenge: deciding how to select among the multiplicity of configurations that control can take at any given time. Here, we will lay out the complexities of the inverse problem inherent to cognitive control allocation, and their close parallels to inverse problems within motor control (e.g., choosing between redundant limb movements). We discuss existing solutions to motor control's inverse problems drawn from optimal control theory, which have proposed that effort costs act to regularize actions and transform motor planning into a well-posed problem. These same principles may help shed light on how our brains optimize over complex control configuration, while providing a new normative perspective on the origins of mental effort.
Collapse
|
42
|
|
43
|
Wang Z, Xu Y, Wang D, Yang J, Bao Z. Hierarchical deep reinforcement learning reveals a modular mechanism of cell movement. NAT MACH INTELL 2022; 4:73-83. [DOI: 10.1038/s42256-021-00431-x] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/30/2022]
|
44
|
Differential coding of goals and actions in ventral and dorsal corticostriatal circuits during goal-directed behavior. Cell Rep 2022; 38:110198. [PMID: 34986350 PMCID: PMC9608360 DOI: 10.1016/j.celrep.2021.110198] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2021] [Revised: 07/08/2021] [Accepted: 12/10/2021] [Indexed: 02/04/2023] Open
Abstract
Goal-directed behavior requires identifying objects in the environment that can satisfy internal needs and executing actions to obtain those objects. The current study examines ventral and dorsal corticostriatal circuits that support complementary aspects of goal-directed behavior. We analyze activity from the amygdala, ventral striatum, orbitofrontal cortex, and lateral prefrontal cortex (LPFC) while monkeys perform a three-armed bandit task. Information about chosen stimuli and their value is primarily encoded in the amygdala, ventral striatum, and orbitofrontal cortex, while the spatial information is primarily encoded in the LPFC. Before the options are presented, information about the to-be-chosen stimulus is represented in the amygdala, ventral striatum, and orbitofrontal cortex; at the time of choice, the information is passed to the LPFC to direct a saccade. Thus, learned value information specifying behavioral goals is maintained throughout the ventral corticostriatal circuit, and it is routed through the dorsal circuit at the time actions are selected.
Collapse
|
45
|
Averbeck B, O'Doherty JP. Reinforcement-learning in fronto-striatal circuits. Neuropsychopharmacology 2022; 47:147-162. [PMID: 34354249 PMCID: PMC8616931 DOI: 10.1038/s41386-021-01108-0] [Citation(s) in RCA: 34] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/04/2021] [Revised: 07/06/2021] [Accepted: 07/09/2021] [Indexed: 01/03/2023]
Abstract
We review the current state of knowledge on the computational and neural mechanisms of reinforcement-learning with a particular focus on fronto-striatal circuits. We divide the literature in this area into five broad research themes: the target of the learning-whether it be learning about the value of stimuli or about the value of actions; the nature and complexity of the algorithm used to drive the learning and inference process; how learned values get converted into choices and associated actions; the nature of state representations, and of other cognitive machinery that support the implementation of various reinforcement-learning operations. An emerging fifth area focuses on how the brain allocates or arbitrates control over different reinforcement-learning sub-systems or "experts". We will outline what is known about the role of the prefrontal cortex and striatum in implementing each of these functions. We then conclude by arguing that it will be necessary to build bridges from algorithmic level descriptions of computational reinforcement-learning to implementational level models to better understand how reinforcement-learning emerges from multiple distributed neural networks in the brain.
Collapse
Affiliation(s)
| | - John P O'Doherty
- Division of Humanities and Social Sciences, California Institute of Technology, Pasadena, CA, USA.
| |
Collapse
|
46
|
Hennig JA, Oby ER, Losey DM, Batista AP, Yu BM, Chase SM. How learning unfolds in the brain: toward an optimization view. Neuron 2021; 109:3720-3735. [PMID: 34648749 PMCID: PMC8639641 DOI: 10.1016/j.neuron.2021.09.005] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2021] [Revised: 08/25/2021] [Accepted: 09/02/2021] [Indexed: 12/17/2022]
Abstract
How do changes in the brain lead to learning? To answer this question, consider an artificial neural network (ANN), where learning proceeds by optimizing a given objective or cost function. This "optimization framework" may provide new insights into how the brain learns, as many idiosyncratic features of neural activity can be recapitulated by an ANN trained to perform the same task. Nevertheless, there are key features of how neural population activity changes throughout learning that cannot be readily explained in terms of optimization and are not typically features of ANNs. Here we detail three of these features: (1) the inflexibility of neural variability throughout learning, (2) the use of multiple learning processes even during simple tasks, and (3) the presence of large task-nonspecific activity changes. We propose that understanding the role of these features in the brain will be key to describing biological learning using an optimization framework.
Collapse
Affiliation(s)
- Jay A Hennig
- Neuroscience Institute, Carnegie Mellon University, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, Pittsburgh, PA, USA; Machine Learning Department, Carnegie Mellon University, Pittsburgh, PA, USA.
| | - Emily R Oby
- Center for the Neural Basis of Cognition, Pittsburgh, PA, USA; Department of Bioengineering, University of Pittsburgh, Pittsburgh, PA, USA
| | - Darby M Losey
- Neuroscience Institute, Carnegie Mellon University, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, Pittsburgh, PA, USA; Machine Learning Department, Carnegie Mellon University, Pittsburgh, PA, USA
| | - Aaron P Batista
- Center for the Neural Basis of Cognition, Pittsburgh, PA, USA; Department of Bioengineering, University of Pittsburgh, Pittsburgh, PA, USA
| | - Byron M Yu
- Neuroscience Institute, Carnegie Mellon University, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, Pittsburgh, PA, USA; Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, USA; Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, PA, USA
| | - Steven M Chase
- Neuroscience Institute, Carnegie Mellon University, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, Pittsburgh, PA, USA; Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, PA, USA
| |
Collapse
|
47
|
Prediction of genetic alteration of phospholipase C isozymes in brain disorders: Studies with deep learning. Adv Biol Regul 2021; 82:100833. [PMID: 34773889 DOI: 10.1016/j.jbior.2021.100833] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2021] [Revised: 10/15/2021] [Accepted: 10/19/2021] [Indexed: 11/22/2022]
Abstract
Genetic mutations leading to the development of various diseases, such as cancer, diabetes, and neurodegenerative disorders, can be attributed to multiple mechanisms and exposure to diverse environments. These disorders further increase gene mutation rates and affect the activity of translated proteins, both phenomena associated with cellular responses. Therefore, maintaining the integrity of genetic and epigenetic information is critical for disease suppression and prevention. With the advent of genome sequencing technologies, large-scale genomic data-based machine learning tools, including deep learning, have been used to predict and identify somatic inactivation or negative dominant expression of target genes in various diseases. Although deep learning studies have recently been highlighted for their ability to distinguish between the genetic information of diseases, conventional wisdom is also necessary to explain the correlation between genotype and phenotype. Herein, we summarize the current understanding of phosphoinositide-specific phospholipase C isozymes (PLCs) and an overview of their associations with genetic variation, as well as their emerging roles in several diseases. We also predicted and discussed new findings of cryptic PLC splice variants by deep learning and the clinical implications of the PLC genetic variations predicted using these tools.
Collapse
|
48
|
Schilling M, Melnik A, Ohl FW, Ritter HJ, Hammer B. Decentralized control and local information for robust and adaptive decentralized Deep Reinforcement Learning. Neural Netw 2021; 144:699-725. [PMID: 34673323 DOI: 10.1016/j.neunet.2021.09.017] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2021] [Revised: 09/13/2021] [Accepted: 09/21/2021] [Indexed: 12/18/2022]
Abstract
Decentralization is a central characteristic of biological motor control that allows for fast responses relying on local sensory information. In contrast, the current trend of Deep Reinforcement Learning (DRL) based approaches to motor control follows a centralized paradigm using a single, holistic controller that has to untangle the whole input information space. This motivates to ask whether decentralization as seen in biological control architectures might also be beneficial for embodied sensori-motor control systems when using DRL. To answer this question, we provide an analysis and comparison of eight control architectures for adaptive locomotion that were derived for a four-legged agent, but with their degree of decentralization varying systematically between the extremes of fully centralized and fully decentralized. Our comparison shows that learning speed is significantly enhanced in distributed architectures-while still reaching the same high performance level of centralized architectures-due to smaller search spaces and local costs providing more focused information for learning. Second, we find an increased robustness of the learning process in the decentralized cases-it is less demanding to hyperparameter selection and less prone to becoming trapped in poor local minima. Finally, when examining generalization to uneven terrains-not used during training-we find best performance for an intermediate architecture that is decentralized, but integrates only local information from both neighboring legs. Together, these findings demonstrate beneficial effects of distributing control into decentralized units and relying on local information. This appears as a promising approach towards more robust DRL and better generalization towards adaptive behavior.
Collapse
Affiliation(s)
- Malte Schilling
- Machine Learning Group, Bielefeld University, 33501 Bielefeld, Germany.
| | - Andrew Melnik
- Neuroinformatics Group, Bielefeld University, 33501 Bielefeld, Germany
| | - Frank W Ohl
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg, Germany; Institute of Biology, Otto-von-Guericke University, Magdeburg, Germany
| | - Helge J Ritter
- Neuroinformatics Group, Bielefeld University, 33501 Bielefeld, Germany
| | - Barbara Hammer
- Machine Learning Group, Bielefeld University, 33501 Bielefeld, Germany
| |
Collapse
|
49
|
Abstract
We report a possible solution for the long-standing problem of the biological function of swirling motion, when a group of animals orbits a common center of the group. We exploit the hypothesis that learning processes in the nervous system of animals may be modelled by reinforcement learning (RL) and apply it to explain the phenomenon. In contrast to hardly justified models of physical interactions between animals, we propose a small set of rules to be learned by the agents, which results in swirling. The rules are extremely simple and thus applicable to animals with very limited level of information processing. We demonstrate that swirling may be understood in terms of the escort behavior, when an individual animal tries to reside within a certain distance from the swarm center. Moreover, we reveal the biological function of swirling motion: a trained for swirling swarm is by orders of magnitude more resistant to external perturbations, than an untrained one. Using our approach we analyze another class of a coordinated motion of animals-a group locomotion in viscous fluid. On a model example we demonstrate that RL provides an optimal disposition of coherently moving animals with a minimal dissipation of energy.
Collapse
|
50
|
Balch JA, Delitto D, Tighe PJ, Zarrinpar A, Efron PA, Rashidi P, Upchurch GR, Bihorac A, Loftus TJ. Machine Learning Applications in Solid Organ Transplantation and Related Complications. Front Immunol 2021; 12:739728. [PMID: 34603324 PMCID: PMC8481939 DOI: 10.3389/fimmu.2021.739728] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2021] [Accepted: 08/25/2021] [Indexed: 11/13/2022] Open
Abstract
The complexity of transplant medicine pushes the boundaries of innate, human reasoning. From networks of immune modulators to dynamic pharmacokinetics to variable postoperative graft survival to equitable allocation of scarce organs, machine learning promises to inform clinical decision making by deciphering prodigious amounts of available data. This paper reviews current research describing how algorithms have the potential to augment clinical practice in solid organ transplantation. We provide a general introduction to different machine learning techniques, describing their strengths, limitations, and barriers to clinical implementation. We summarize emerging evidence that recent advances that allow machine learning algorithms to predict acute post-surgical and long-term outcomes, classify biopsy and radiographic data, augment pharmacologic decision making, and accurately represent the complexity of host immune response. Yet, many of these applications exist in pre-clinical form only, supported primarily by evidence of single-center, retrospective studies. Prospective investigation of these technologies has the potential to unlock the potential of machine learning to augment solid organ transplantation clinical care and health care delivery systems.
Collapse
Affiliation(s)
- Jeremy A Balch
- Department of Surgery, University of Florida Health, Gainesville, FL, United States
| | - Daniel Delitto
- Department of Surgery, Johns Hopkins University, Baltimore, MD, United States
| | - Patrick J Tighe
- Department of Anesthesiology, University of Florida Health, Gainesville, FL, United States.,Department of Orthopedics, University of Florida Health, Gainesville, FL, United States.,Department of Information Systems/Operations Management, University of Florida Health, Gainesville, FL, United States
| | - Ali Zarrinpar
- Department of Surgery, University of Florida Health, Gainesville, FL, United States
| | - Philip A Efron
- Department of Surgery, University of Florida Health, Gainesville, FL, United States
| | - Parisa Rashidi
- Department of Biomedical Engineering, University of Florida, Gainesville, FL, United States.,Department of Computer and Information Science and Engineering University of Florida, Gainesville, FL, United States.,Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL, United States.,Precision and Intelligent Systems in Medicine (PrismaP), University of Florida, Gainesville, FL, United States
| | - Gilbert R Upchurch
- Department of Surgery, University of Florida Health, Gainesville, FL, United States
| | - Azra Bihorac
- Precision and Intelligent Systems in Medicine (PrismaP), University of Florida, Gainesville, FL, United States.,Department of Medicine, University of Florida Health, Gainesville, FL, United States
| | - Tyler J Loftus
- Department of Surgery, University of Florida Health, Gainesville, FL, United States.,Precision and Intelligent Systems in Medicine (PrismaP), University of Florida, Gainesville, FL, United States
| |
Collapse
|