1
|
Bishof Z, Scheuerman J, Michael CJ. Closed-Loop Uncertainty: The Evaluation and Calibration of Uncertainty for Human-Machine Teams under Data Drift. Entropy (Basel) 2023; 25:1443. [PMID: 37895564 PMCID: PMC10606420 DOI: 10.3390/e25101443] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/31/2023] [Revised: 10/05/2023] [Accepted: 10/06/2023] [Indexed: 10/29/2023]
Abstract
Though an accurate measurement of entropy, or more generally uncertainty, is critical to the success of human-machine teams, the evaluation of the accuracy of such metrics as a probability of machine correctness is often aggregated and not assessed as an iterative control process. The entropy of the decisions made by human-machine teams may not be accurately measured under cold start or at times of data drift unless disagreements between the human and machine are immediately fed back to the classifier iteratively. In this study, we present a stochastic framework by which an uncertainty model may be evaluated iteratively as a probability of machine correctness. We target a novel problem, referred to as the threshold selection problem, which involves a user subjectively selecting the point at which a signal transitions to a low state. This problem is designed to be simple and replicable for human-machine experimentation while exhibiting properties of more complex applications. Finally, we explore the potential of incorporating feedback of machine correctness into a baseline naïve Bayes uncertainty model with a novel reinforcement learning approach. The approach refines a baseline uncertainty model by incorporating machine correctness at every iteration. Experiments are conducted over a large number of realizations to properly evaluate uncertainty at each iteration of the human-machine team. Results show that our novel approach, called closed-loop uncertainty, outperforms the baseline in every case, yielding about 45% improvement on average.
Collapse
Affiliation(s)
- Zachary Bishof
- U.S. Naval Research Laboratory, 1005 Balch Boulevard, Stennis Space Center, St. Louis, MS 39529, USA; (J.S.); (C.J.M.)
| | | | | |
Collapse
|
2
|
Pfeuffer N, Baum L, Stammer W, Abdel-Karim BM, Schramowski P, Bucher AM, Hügel C, Rohde G, Kersting K, Hinz O. Explanatory Interactive Machine Learning. Bus Inf Syst Eng 2023. [PMCID: PMC10119840 DOI: 10.1007/s12599-023-00806-x] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/15/2022] [Accepted: 01/17/2023] [Indexed: 11/22/2023]
Abstract
The most promising standard machine learning methods can deliver highly accurate classification results, often outperforming standard white-box methods. However, it is hardly possible for humans to fully understand the rationale behind the black-box results, and thus, these powerful methods hamper the creation of new knowledge on the part of humans and the broader acceptance of this technology. Explainable Artificial Intelligence attempts to overcome this problem by making the results more interpretable, while Interactive Machine Learning integrates humans into the process of insight discovery. The paper builds on recent successes in combining these two cutting-edge technologies and proposes how Explanatory Interactive Machine Learning (XIL) is embedded in a generalizable Action Design Research (ADR) process – called XIL-ADR. This approach can be used to analyze data, inspect models, and iteratively improve them. The paper shows the application of this process using the diagnosis of viral pneumonia, e.g., Covid-19, as an illustrative example. By these means, the paper also illustrates how XIL-ADR can help identify shortcomings of standard machine learning projects, gain new insights on the part of the human user, and thereby can help to unlock the full potential of AI-based systems for organizations and research.
Collapse
Affiliation(s)
- Nicolas Pfeuffer
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Lorenz Baum
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Wolfgang Stammer
- Machine Learning Group, Department of Computer Science, Technical University of Darmstadt, Darmstadt, Germany
| | - Benjamin M. Abdel-Karim
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Patrick Schramowski
- Machine Learning Group, Department of Computer Science, Technical University of Darmstadt, Darmstadt, Germany
| | - Andreas M. Bucher
- Diagnostic and Interventional Radiology, Center of Radiology, Hospital of the Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Christian Hügel
- Pneumology and Allergology, Center of Internal Medicine, Hospital of the Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Gernot Rohde
- Pneumology and Allergology, Center of Internal Medicine, Hospital of the Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Kristian Kersting
- Machine Learning Group, Department of Computer Science, Technical University of Darmstadt, Darmstadt, Germany
| | - Oliver Hinz
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| |
Collapse
|
3
|
Abstract
We present a meta-learning framework for interactive medical image registration. Our proposed framework comprises three components: a learning-based medical image registration algorithm, a form of user interaction that refines registration at inference, and a meta-learning protocol that learns a rapidly adaptable network initialization. This paper describes a specific algorithm that implements the registration, interaction and meta-learning protocol for our exemplar clinical application: registration of magnetic resonance (MR) imaging to interactively acquired, sparsely-sampled transrectal ultrasound (TRUS) images. Our approach obtains comparable registration error (4.26 mm) to the best-performing non-interactive learning-based 3D-to-3D method (3.97 mm) while requiring only a fraction of the data, and occurring in real-time during acquisition. Applying sparsely sampled data to non-interactive methods yields higher registration errors (6.26 mm), demonstrating the effectiveness of interactive MR-TRUS registration, which may be applied intraoperatively given the real-time nature of the adaptation process.
Collapse
Affiliation(s)
- Zachary M. C. Baum
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London W1W 7TS, U.K.,; UCL Centre for Medical Image Computing, University College London, London W1W 7TS, U.K
| | - Yipeng Hu
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London W1W 7TS, U.K.,; UCL Centre for Medical Image Computing, University College London, London W1W 7TS, U.K
| | - Dean C. Barratt
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London W1W 7TS, U.K.,; UCL Centre for Medical Image Computing, University College London, London W1W 7TS, U.K
| |
Collapse
|
4
|
Cheng F, Keller MS, Qu H, Gehlenborg N, Wang Q. Polyphony: an Interactive Transfer Learning Framework for Single-Cell Data Analysis. IEEE Trans Vis Comput Graph 2023; 29:591-601. [PMID: 36155452 PMCID: PMC10039961 DOI: 10.1109/tvcg.2022.3209408] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Reference-based cell-type annotation can significantly reduce time and effort in single-cell analysis by transferring labels from a previously-annotated dataset to a new dataset. However, label transfer by end-to-end computational methods is challenging due to the entanglement of technical (e.g., from different sequencing batches or techniques) and biological (e.g., from different cellular microenvironments) variations, only the first of which must be removed. To address this issue, we propose Polyphony, an interactive transfer learning (ITL) framework, to complement biologists' knowledge with advanced computational methods. Polyphony is motivated and guided by domain experts' needs for a controllable, interactive, and algorithm-assisted annotation process, identified through interviews with seven biologists. We introduce anchors, i.e., analogous cell populations across datasets, as a paradigm to explain the computational process and collect user feedback for model improvement. We further design a set of visualizations and interactions to empower users to add, delete, or modify anchors, resulting in refined cell type annotations. The effectiveness of this approach is demonstrated through quantitative experiments, two hypothetical use cases, and interviews with two biologists. The results show that our anchor-based ITL method takes advantage of both human and machine intelligence in annotating massive single-cell datasets.
Collapse
|
5
|
Teso S, Alkan Ö, Stammer W, Daly E. Leveraging explanations in interactive machine learning: An overview. Front Artif Intell 2023; 6:1066049. [PMID: 36909207 PMCID: PMC9995896 DOI: 10.3389/frai.2023.1066049] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2022] [Accepted: 02/01/2023] [Indexed: 02/25/2023] Open
Abstract
Explanations have gained an increasing level of interest in the AI and Machine Learning (ML) communities in order to improve model transparency and allow users to form a mental model of a trained ML model. However, explanations can go beyond this one way communication as a mechanism to elicit user control, because once users understand, they can then provide feedback. The goal of this paper is to present an overview of research where explanations are combined with interactive capabilities as a mean to learn new models from scratch and to edit and debug existing ones. To this end, we draw a conceptual map of the state-of-the-art, grouping relevant approaches based on their intended purpose and on how they structure the interaction, highlighting similarities and differences between them. We also discuss open research issues and outline possible directions forward, with the hope of spurring further research on this blooming research topic.
Collapse
Affiliation(s)
- Stefano Teso
- CIMeC and DISI, University of Trento, Trento, Italy
| | | | - Wolfgang Stammer
- Machine Learning Group, Department of Computer Science, TU Darmstadt, Darmstadt, Germany
| | | |
Collapse
|
6
|
Smith AG, Han E, Petersen J, Olsen NAF, Giese C, Athmann M, Dresbøll DB, Thorup‐Kristensen K. RootPainter: deep learning segmentation of biological images with corrective annotation. New Phytol 2022; 236:774-791. [PMID: 35851958 PMCID: PMC9804377 DOI: 10.1111/nph.18387] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/08/2022] [Accepted: 06/30/2022] [Indexed: 05/27/2023]
Abstract
Convolutional neural networks (CNNs) are a powerful tool for plant image analysis, but challenges remain in making them more accessible to researchers without a machine-learning background. We present RootPainter, an open-source graphical user interface based software tool for the rapid training of deep neural networks for use in biological image analysis. We evaluate RootPainter by training models for root length extraction from chicory (Cichorium intybus L.) roots in soil, biopore counting, and root nodule counting. We also compare dense annotations with corrective ones that are added during the training process based on the weaknesses of the current model. Five out of six times the models trained using RootPainter with corrective annotations created within 2 h produced measurements strongly correlating with manual measurements. Model accuracy had a significant correlation with annotation duration, indicating further improvements could be obtained with extended annotation. Our results show that a deep-learning model can be trained to a high accuracy for the three respective datasets of varying target objects, background, and image quality with < 2 h of annotation time. They indicate that, when using RootPainter, for many datasets it is possible to annotate, train, and complete data processing within 1 d.
Collapse
Affiliation(s)
- Abraham George Smith
- Department of Plant and Environmental ScienceUniversity of CopenhagenHøjbakkegårds Alle 13Tåstrup2630Denmark
- Department of Computer ScienceUniversity of CopenhagenUniversitetsparken 12100CopenhagenDenmark
| | - Eusun Han
- Department of Plant and Environmental ScienceUniversity of CopenhagenHøjbakkegårds Alle 13Tåstrup2630Denmark
- CSIRO Agriculture and FoodPO Box 1700CanberraACT2601Australia
| | - Jens Petersen
- Department of Computer ScienceUniversity of CopenhagenUniversitetsparken 12100CopenhagenDenmark
| | - Niels Alvin Faircloth Olsen
- Department of Plant and Environmental ScienceUniversity of CopenhagenHøjbakkegårds Alle 13Tåstrup2630Denmark
| | - Christian Giese
- Department of Agroecology and Organic FarmingUniversity of BonnRegina‐Pacis‐Weg 353113BonnGermany
| | - Miriam Athmann
- Department of Organic Farming and Plant ProductionUniversity of KasselNordbahnhofstr. 1aD‐37213WitzenhausenGermany
| | - Dorte Bodin Dresbøll
- Department of Plant and Environmental ScienceUniversity of CopenhagenHøjbakkegårds Alle 13Tåstrup2630Denmark
| | - Kristian Thorup‐Kristensen
- Department of Plant and Environmental ScienceUniversity of CopenhagenHøjbakkegårds Alle 13Tåstrup2630Denmark
| |
Collapse
|
7
|
Bender D, Licht DJ, Nataraj C. A Novel Embedded Feature Selection and Dimensionality Reduction Method for an SVM Type Classifier to Predict Periventricular Leukomalacia (PVL) in Neonates. Appl Sci (Basel) 2021; 11:11156. [PMID: 37885926 PMCID: PMC10601609 DOI: 10.3390/app112311156] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
Abstract
This paper is concerned with the prediction of the occurrence of periventricular leukomalacia (PVL) in neonates after heart surgery. Our prior work shows that the Support Vector Machine (SVM) classifier can be a powerful tool in predicting clinical outcomes of such complicated and uncommon diseases, even when the number of data samples is low. In the presented work, we first illustrate and discuss the shortcomings of the traditional automatic machine learning (aML) approach. Consequently, we describe our methodology for addressing these shortcomings, while utilizing the designed interactive ML (iML) algorithm. Finally, we conclude with a discussion of the developed method and the results obtained. In sum, by adding an additional (Genetic Algorithm) optimization step in the SVM learning framework, we were able to (a) reduce the dimensionality of an SVM model from 248 to 53 features, (b) increase generalization that was confirmed by a 100% accuracy assessed on an unseen testing set, and (c) improve the overall SVM model's performance from 65% to 100% testing accuracy, utilizing the proposed iML method.
Collapse
Affiliation(s)
- Dieter Bender
- Villanova Center for Analytics of Dynamic Systems, Villanova University, 800 Lancaster Ave, Villanova, PA 19085, USA
| | - Daniel J. Licht
- June and Steve Wolfson Laboratory for Clinical and Biomedical Optics, Children’s Hospital of Philadelphia, 324 S 34th St, Philadelphia, PA 19104, USA
| | - C. Nataraj
- Villanova Center for Analytics of Dynamic Systems, Villanova University, 800 Lancaster Ave, Villanova, PA 19085, USA
| |
Collapse
|
8
|
Abstract
In this paper, we provide an overview of the existing methods for integrating human advice into a reinforcement learning process. We first propose a taxonomy of the different forms of advice that can be provided to a learning agent. We then describe the methods that can be used for interpreting advice when its meaning is not determined beforehand. Finally, we review different approaches for integrating advice into the learning process.
Collapse
Affiliation(s)
- Anis Najar
- Laboratoire de Neurosciences Cognitives Computationnelles, INSERM U960, Paris, France
| | - Mohamed Chetouani
- Institute for Intelligent Systems and Robotics, Sorbonne Université, CNRS UMR 7222, Paris, France
| |
Collapse
|
9
|
Bernardo F, Zbyszyński M, Grierson M, Fiebrink R. Designing and Evaluating the Usability of a Machine Learning API for Rapid Prototyping Music Technology. Front Artif Intell 2021; 3:13. [PMID: 33733132 PMCID: PMC7861239 DOI: 10.3389/frai.2020.00013] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2019] [Accepted: 03/09/2020] [Indexed: 11/23/2022] Open
Abstract
To better support creative software developers and music technologists' needs, and to empower them as machine learning users and innovators, the usability of and developer experience with machine learning tools must be considered and better understood. We review background research on the design and evaluation of application programming interfaces (APIs), with a focus on the domain of machine learning for music technology software development. We present the design rationale for the RAPID-MIX API, an easy-to-use API for rapid prototyping with interactive machine learning, and a usability evaluation study with software developers of music technology. A cognitive dimensions questionnaire was designed and delivered to a group of 12 participants who used the RAPID-MIX API in their software projects, including people who developed systems for personal use and professionals developing software products for music and creative technology companies. The results from questionnaire indicate that participants found the RAPID-MIX API a machine learning API which is easy to learn and use, fun, and good for rapid prototyping with interactive machine learning. Based on these findings, we present an analysis and characterization of the RAPID-MIX API based on the cognitive dimensions framework, and discuss its design trade-offs and usability issues. We use these insights and our design experience to provide design recommendations for ML APIs for rapid prototyping of music technology. We conclude with a summary of the main insights, a discussion of the merits and challenges of the application of the CDs framework to the evaluation of machine learning APIs, and directions to future work which our research deems valuable.
Collapse
Affiliation(s)
- Francisco Bernardo
- EMuTe Lab, School of Media, Film and Music, University of Sussex, Brighton, United Kingdom.,EAVI, Department of Computing, Goldsmiths, University of London, London, United Kingdom
| | - Michael Zbyszyński
- EAVI, Department of Computing, Goldsmiths, University of London, London, United Kingdom
| | - Mick Grierson
- EAVI, Department of Computing, Goldsmiths, University of London, London, United Kingdom.,Creative Computing Institute, University of the Arts, London, United Kingdom
| | - Rebecca Fiebrink
- EAVI, Department of Computing, Goldsmiths, University of London, London, United Kingdom.,Creative Computing Institute, University of the Arts, London, United Kingdom
| |
Collapse
|
10
|
Maadi M, Akbarzadeh Khorshidi H, Aickelin U. A Review on Human-AI Interaction in Machine Learning and Insights for Medical Applications. Int J Environ Res Public Health 2021; 18:ijerph18042121. [PMID: 33671609 PMCID: PMC7926732 DOI: 10.3390/ijerph18042121] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/30/2020] [Revised: 02/08/2021] [Accepted: 02/12/2021] [Indexed: 11/19/2022]
Abstract
Objective: To provide a human–Artificial Intelligence (AI) interaction review for Machine Learning (ML) applications to inform how to best combine both human domain expertise and computational power of ML methods. The review focuses on the medical field, as the medical ML application literature highlights a special necessity of medical experts collaborating with ML approaches. Methods: A scoping literature review is performed on Scopus and Google Scholar using the terms “human in the loop”, “human in the loop machine learning”, and “interactive machine learning”. Peer-reviewed papers published from 2015 to 2020 are included in our review. Results: We design four questions to investigate and describe human–AI interaction in ML applications. These questions are “Why should humans be in the loop?”, “Where does human–AI interaction occur in the ML processes?”, “Who are the humans in the loop?”, and “How do humans interact with ML in Human-In-the-Loop ML (HILML)?”. To answer the first question, we describe three main reasons regarding the importance of human involvement in ML applications. To address the second question, human–AI interaction is investigated in three main algorithmic stages: 1. data producing and pre-processing; 2. ML modelling; and 3. ML evaluation and refinement. The importance of the expertise level of the humans in human–AI interaction is described to answer the third question. The number of human interactions in HILML is grouped into three categories to address the fourth question. We conclude the paper by offering a discussion on open opportunities for future research in HILML.
Collapse
|
11
|
Rhodes C, Allmendinger R, Climent R. New Interfaces and Approaches to Machine Learning When Classifying Gestures within Music. Entropy (Basel) 2020; 22:E1384. [PMID: 33297582 DOI: 10.3390/e22121384] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/14/2020] [Revised: 12/03/2020] [Accepted: 12/03/2020] [Indexed: 11/25/2022]
Abstract
Interactive music uses wearable sensors (i.e., gestural interfaces—GIs) and biometric datasets to reinvent traditional human–computer interaction and enhance music composition. In recent years, machine learning (ML) has been important for the artform. This is because ML helps process complex biometric datasets from GIs when predicting musical actions (termed performance gestures). ML allows musicians to create novel interactions with digital media. Wekinator is a popular ML software amongst artists, allowing users to train models through demonstration. It is built on the Waikato Environment for Knowledge Analysis (WEKA) framework, which is used to build supervised predictive models. Previous research has used biometric data from GIs to train specific ML models. However, previous research does not inform optimum ML model choice, within music, or compare model performance. Wekinator offers several ML models. Thus, we used Wekinator and the Myo armband GI and study three performance gestures for piano practice to solve this problem. Using these, we trained all models in Wekinator and investigated their accuracy, how gesture representation affects model accuracy and if optimisation can arise. Results show that neural networks are the strongest continuous classifiers, mapping behaviour differs amongst continuous models, optimisation can occur and gesture representation disparately affects model mapping behaviour; impacting music practice.
Collapse
|
12
|
Li H, Fang S, Mukhopadhyay S, Saykin AJ, Shen L. Interactive Machine Learning by Visualization: A Small Data Solution. Proc IEEE Int Conf Big Data 2018; 2018:3513-3521. [PMID: 31061990 DOI: 10.1109/bigdata.2018.8621952] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Machine learning algorithms and traditional data mining process usually require a large volume of data to train the algorithm-specific models, with little or no user feedback during the model building process. Such a "big data" based automatic learning strategy is sometimes unrealistic for applications where data collection or processing is very expensive or difficult, such as in clinical trials. Furthermore, expert knowledge can be very valuable in the model building process in some fields such as biomedical sciences. In this paper, we propose a new visual analytics approach to interactive machine learning and visual data mining. In this approach, multi-dimensional data visualization techniques are employed to facilitate user interactions with the machine learning and mining process. This allows dynamic user feedback in different forms, such as data selection, data labeling, and data correction, to enhance the efficiency of model building. In particular, this approach can significantly reduce the amount of data required for training an accurate model, and therefore can be highly impactful for applications where large amount of data is hard to obtain. The proposed approach is tested on two application problems: the handwriting recognition (classification) problem and the human cognitive score prediction (regression) problem. Both experiments show that visualization supported interactive machine learning and data mining can achieve the same accuracy as an automatic process can with much smaller training data sets.
Collapse
Affiliation(s)
- Huang Li
- Department of Computer & Information Science, Indiana University Purdue University Indianapolis
| | - Shiaofen Fang
- Department of Computer & Information Science, Indiana University Purdue University Indianapolis
| | - Snehasis Mukhopadhyay
- Department of Computer & Information Science, Indiana University Purdue University Indianapolis
| | | | - Li Shen
- University of Pennsylvania Perelman School of Medicine
| |
Collapse
|