1
|
Kumar V, Barik S, Aggarwal S, Kumar D, Raj V. The use of artificial intelligence for persons with disability: a bright and promising future ahead. Disabil Rehabil Assist Technol 2024; 19:2415-2417. [PMID: 38079540 DOI: 10.1080/17483107.2023.2288241] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2023] [Accepted: 11/18/2023] [Indexed: 08/03/2024]
Abstract
Artificial intelligence (AI) driven solutions have the potential to significantly impact individuals with disabilities by providing assistance in their daily activities and facilitating the acquisition of new abilities. The utilisation of AI technology in assisting individuals with disabilities has novel prospects for enhancing accessibility, fostering inclusivity throughout society, and enabling autonomous living, which would otherwise pose considerable challenges or remain unattainable. As the field of AI continues to progress, it holds the potential to facilitate the development of increasingly sophisticated and groundbreaking approaches to tackle the multifaceted obstacles encountered by individuals with disabilities. Consequently, AI has the capacity to foster greater inclusivity for this population.
Collapse
Affiliation(s)
- Vishal Kumar
- Department of Orthopedics, PGIMER, Chandigarh, India
| | - Sitanshu Barik
- Department of Orthopedics, All India Institute of Medical Sciences, Nagpur, India
| | | | - Deepak Kumar
- Department of Physical Medicine and Rehabilitation, All India Institute of Medical Sciences, Deoghar, India
| | - Vikash Raj
- Department of Orthopedics, All India Institute of Medical Sciences, Deoghar, India
| |
Collapse
|
2
|
Yu Y. Evaluation on interactive waiting experience design of mobile internet products based on machine learning. Sci Rep 2023; 13:16985. [PMID: 37813893 PMCID: PMC10562367 DOI: 10.1038/s41598-023-43405-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2023] [Accepted: 09/23/2023] [Indexed: 10/11/2023] Open
Abstract
In today's rapidly changing economy, efficient lifestyle has become the current situation of most mobile product users. With the development of performance tools and technologies, a fast lifestyle has brought more wealth and opportunities to users. The slow pace and fluctuating time are indirect income losses, which cause user anxiety to some extent. When the waiting time exceeds the user's waiting threshold, users would experience negative emotions, such as boredom, anxiety and anger, and product satisfaction would drop significantly. Therefore, by analyzing the uniqueness of mobile Internet products and the characteristics of users, this paper studied the reasons and influencing factors of product interactive waiting, and then used machine learning algorithm to analyze the cost function of interactive waiting experience. Finally, the corresponding interactive waiting experience design strategy was proposed. By comparison, the user experience after product interaction optimization design was 8.4% higher than that before product interaction optimization design, and the user frequency was also 14.7% higher after optimization design. In short, user experience plays an important role in product interaction design.
Collapse
Affiliation(s)
- Yi Yu
- Artistics, Krirk University, Bangkhen, Bangkok, 10220, Thailand.
| |
Collapse
|
3
|
Pfeuffer N, Baum L, Stammer W, Abdel-Karim BM, Schramowski P, Bucher AM, Hügel C, Rohde G, Kersting K, Hinz O. Explanatory Interactive Machine Learning. BUSINESS & INFORMATION SYSTEMS ENGINEERING 2023. [PMCID: PMC10119840 DOI: 10.1007/s12599-023-00806-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/15/2022] [Accepted: 01/17/2023] [Indexed: 11/22/2023]
Abstract
The most promising standard machine learning methods can deliver highly accurate classification results, often outperforming standard white-box methods. However, it is hardly possible for humans to fully understand the rationale behind the black-box results, and thus, these powerful methods hamper the creation of new knowledge on the part of humans and the broader acceptance of this technology. Explainable Artificial Intelligence attempts to overcome this problem by making the results more interpretable, while Interactive Machine Learning integrates humans into the process of insight discovery. The paper builds on recent successes in combining these two cutting-edge technologies and proposes how Explanatory Interactive Machine Learning (XIL) is embedded in a generalizable Action Design Research (ADR) process – called XIL-ADR. This approach can be used to analyze data, inspect models, and iteratively improve them. The paper shows the application of this process using the diagnosis of viral pneumonia, e.g., Covid-19, as an illustrative example. By these means, the paper also illustrates how XIL-ADR can help identify shortcomings of standard machine learning projects, gain new insights on the part of the human user, and thereby can help to unlock the full potential of AI-based systems for organizations and research.
Collapse
Affiliation(s)
- Nicolas Pfeuffer
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Lorenz Baum
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Wolfgang Stammer
- Machine Learning Group, Department of Computer Science, Technical University of Darmstadt, Darmstadt, Germany
| | - Benjamin M. Abdel-Karim
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Patrick Schramowski
- Machine Learning Group, Department of Computer Science, Technical University of Darmstadt, Darmstadt, Germany
| | - Andreas M. Bucher
- Diagnostic and Interventional Radiology, Center of Radiology, Hospital of the Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Christian Hügel
- Pneumology and Allergology, Center of Internal Medicine, Hospital of the Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Gernot Rohde
- Pneumology and Allergology, Center of Internal Medicine, Hospital of the Goethe University Frankfurt, Frankfurt am Main, Germany
| | - Kristian Kersting
- Machine Learning Group, Department of Computer Science, Technical University of Darmstadt, Darmstadt, Germany
| | - Oliver Hinz
- Information Systems and Information Management, Goethe University Frankfurt, Frankfurt am Main, Germany
| |
Collapse
|
4
|
Afzal S, Ghani S, Hittawe MM, Rashid SF, Knio OM, Hadwiger M, Hoteit I. Visualization and Visual Analytics Approaches for Image and Video Datasets: A Survey. ACM T INTERACT INTEL 2023. [DOI: 10.1145/3576935] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
Abstract
Image and video data analysis has become an increasingly important research area with applications in different domains such as security surveillance, healthcare, augmented and virtual reality, video and image editing, activity analysis and recognition, synthetic content generation, distance education, telepresence, remote sensing, sports analytics, art, non-photorealistic rendering, search engines, and social media. Recent advances in Artificial Intelligence (AI) and particularly deep learning have sparked new research challenges and led to significant advancements, especially in image and video analysis. These advancements have also resulted in significant research and development in other areas such as visualization and visual analytics, and have created new opportunities for future lines of research. In this survey paper, we present the current state of the art at the intersection of visualization and visual analytics, and image and video data analysis. We categorize the visualization papers included in our survey based on different taxonomies used in visualization and visual analytics research. We review these papers in terms of task requirements, tools, datasets, and application areas. We also discuss insights based on our survey results, trends and patterns, the current focus of visualization research, and opportunities for future research.
Collapse
Affiliation(s)
- Shehzad Afzal
- King Abdullah University of Science & Technology, Saudi Arabia
| | - Sohaib Ghani
- King Abdullah University of Science & Technology, Saudi Arabia
| | | | | | - Omar M Knio
- King Abdullah University of Science & Technology, Saudi Arabia
| | - Markus Hadwiger
- King Abdullah University of Science & Technology, Saudi Arabia
| | - Ibrahim Hoteit
- King Abdullah University of Science & Technology, Saudi Arabia
| |
Collapse
|
5
|
Yu Y, Kruyff D, Jiao J, Becker T, Behrisch M. PSEUDo: Interactive Pattern Search in Multivariate Time Series with Locality-Sensitive Hashing and Relevance Feedback. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2023; 29:33-42. [PMID: 36170404 DOI: 10.1109/tvcg.2022.3209431] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
We present PSEUDo, a visual pattern retrieval tool for multivariate time series. It aims to overcome the uneconomic (re-)training problem accompanying deep learning-based methods. Very high-dimensional time series emerge on an unprecedented scale due to increasing sensor usage and data storage. Visual pattern search is one of the most frequent tasks on time series. Automatic pattern retrieval methods often suffer from inefficient training data, a lack of ground truth labels, and a discrepancy between the similarity perceived by the algorithm and required by the user or the task. Our proposal is based on the query-aware locality-sensitive hashing technique to create a representation of multivariate time series windows. It features sub-linear training and inference time with respect to data dimensions. This performance gain allows an instantaneous relevance-feedback-driven adaption to converge to users' similarity notion. We demonstrate PSEUDo's performance in terms of accuracy, speed, steerability, and usability through quantitative benchmarks with representative time series retrieval methods and a case study. We find that PSEUDo detects patterns in high-dimensional time series efficiently, improves the result with relevance feedback through feature selection, and allows an understandable as well as user-friendly retrieval process.
Collapse
|
6
|
Wu J, Liu D, Guo Z, Wu Y. RASIPAM: Interactive Pattern Mining of Multivariate Event Sequences in Racket Sports. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2023; 29:940-950. [PMID: 36179006 DOI: 10.1109/tvcg.2022.3209452] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Experts in racket sports like tennis and badminton use tactical analysis to gain insight into competitors' playing styles. Many data-driven methods apply pattern mining to racket sports data - which is often recorded as multivariate event sequences - to uncover sports tactics. However, tactics obtained in this way are often inconsistent with those deduced by experts through their domain knowledge, which can be confusing to those experts. This work introduces RASIPAM, a RAcket-Sports Interactive PAttern Mining system, which allows experts to incorporate their knowledge into data mining algorithms to discover meaningful tactics interactively. RASIPAM consists of a constraint-based pattern mining algorithm that responds to the analysis demands of experts: Experts provide suggestions for finding tactics in intuitive written language, and these suggestions are translated into constraints to run the algorithm. RASIPAM further introduces a tailored visual interface that allows experts to compare the new tactics with the original ones and decide whether to apply a given adjustment. This interactive workflow iteratively progresses until experts are satisfied with all tactics. We conduct a quantitative experiment to show that our algorithm supports real-time interaction. Two case studies in tennis and in badminton respectively, each involving two domain experts, are conducted to show the effectiveness and usefulness of the system.
Collapse
|
7
|
Fröhlich P, Mirnig AG, Falcioni D, Schrammel J, Diamond L, Fischer I, Tscheligi M. Effects of reliability indicators on usage, acceptance and preference of predictive process management decision support systems. QUALITY AND USER EXPERIENCE 2022; 7:6. [PMID: 36092253 PMCID: PMC9442562 DOI: 10.1007/s41233-022-00053-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/22/2021] [Accepted: 08/03/2022] [Indexed: 11/10/2022]
Abstract
Despite the growing availability of data, simulation technologies, and predictive analytics, it is not yet clear whether and under which conditions users will trust Decision Support Systems (DSS). DSS are designed to support users in making more informed decisions in specialized tasks through more accurate predictions and recommendations. This mixed-methods user study contributes to the research on trust calibration by analyzing the potential effects of integrated reliability indication in DSS user interfaces for process management in first-time usage situations characterized by uncertainty. Ten experts specialized in digital tools for construction were asked to test and assess two versions of a DSS in a renovation project scenario. We found that while users stated that they need full access to all information to make their own decisions, reliability indication in DSS tends to make users more willing to make preliminary decisions, with users adapting their confidence and reliance to the indicated reliability. Reliability indication in DSS also increases subjective usefulness and system reliability. Based on these findings, it is recommended that for the design of reliability indication practitioners consider displaying a combination of reliability information at several granularity levels in DSS user interfaces, including visualizations, such as a traffic light system, and to also provide explanations for the reliability information. Further research directions towards achieving trustworthy decision support in complex environments are proposed.
Collapse
Affiliation(s)
- Peter Fröhlich
- Center for Technology Experience, AIT Austrian Institute of Technology, Giefinggasse 2, 1210 Vienna, Austria
| | - Alexander G. Mirnig
- Center for Technology Experience, AIT Austrian Institute of Technology, Giefinggasse 2, 1210 Vienna, Austria
- Center for Human-Computer Interaction, University of Salzburg, Jakob-Haringer-Straße 8, 5020 Salzburg, Austria
| | | | - Johann Schrammel
- Center for Technology Experience, AIT Austrian Institute of Technology, Giefinggasse 2, 1210 Vienna, Austria
| | - Lisa Diamond
- Center for Technology Experience, AIT Austrian Institute of Technology, Giefinggasse 2, 1210 Vienna, Austria
| | - Isabel Fischer
- Warwick Business School, University of Warwick, Scarman Rd., Coventry, CV4 7AL UK
| | - Manfred Tscheligi
- Center for Technology Experience, AIT Austrian Institute of Technology, Giefinggasse 2, 1210 Vienna, Austria
- Center for Human-Computer Interaction, University of Salzburg, Jakob-Haringer-Straße 8, 5020 Salzburg, Austria
| |
Collapse
|
8
|
Vajiac C, Chau DH, Olligschlaeger A, Mackenzie R, Nair P, Lee MC, Li Y, Park N, Rabbany R, Faloutsos C. TRAFFICVIS: Visualizing Organized Activity and Spatio-Temporal Patterns for Detecting and Labeling Human Trafficking. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2022; PP:1-10. [PMID: 36201417 DOI: 10.1109/tvcg.2022.3209403] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Law enforcement and domain experts can detect human trafficking (HT) in online escort websites by analyzing suspicious clusters of connected ads. How can we explain clustering results intuitively and interactively, visualizing potential evidence for experts to analyze? We present TRAFFICVIS, the first interface for cluster-level HT detection and labeling. Developed through months of participatory design with domain experts, TRAFFICVIS provides coordinated views in conjunction with carefully chosen backend algorithms to effectively show spatio-temporal and text patterns to a wide variety of anti-HT stakeholders. We build upon state-of-the-art text clustering algorithms by incorporating shared metadata as a signal of connected and possibly suspicious activity, then visualize the results. Domain experts can use TRAFFICVIS to label clusters as HT, or other, suspicious, but non-HT activity such as spam and scam, quickly creating labeled datasets to enable further HT research. Through domain expert feedback and a usage scenario, we demonstrate TRAFFICVIS's efficacy. The feedback was overwhelmingly positive, with repeated high praises for the usability and explainability of our tool, the latter being vital for indicting possible criminals.
Collapse
|
9
|
Constructing Explainable Classifiers from the Start—Enabling Human-in-the Loop Machine Learning. INFORMATION 2022. [DOI: 10.3390/info13100464] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Interactive machine learning (IML) enables the incorporation of human expertise because the human participates in the construction of the learned model. Moreover, with human-in-the-loop machine learning (HITL-ML), the human experts drive the learning, and they can steer the learning objective not only for accuracy but perhaps for characterisation and discrimination rules, where separating one class from others is the primary objective. Moreover, this interaction enables humans to explore and gain insights into the dataset as well as validate the learned models. Validation requires transparency and interpretable classifiers. The huge relevance of understandable classification has been recently emphasised for many applications under the banner of explainable artificial intelligence (XAI). We use parallel coordinates to deploy an IML system that enables the visualisation of decision tree classifiers but also the generation of interpretable splits beyond parallel axis splits. Moreover, we show that characterisation and discrimination rules are also well communicated using parallel coordinates. In particular, we report results from the largest usability study of a IML system, confirming the merits of our approach.
Collapse
|
10
|
Abstract
AbstractResearchers are defining new types of interactions between humans and machine learning algorithms generically called human-in-the-loop machine learning. Depending on who is in control of the learning process, we can identify: active learning, in which the system remains in control; interactive machine learning, in which there is a closer interaction between users and learning systems; and machine teaching, where human domain experts have control over the learning process. Aside from control, humans can also be involved in the learning process in other ways. In curriculum learning human domain experts try to impose some structure on the examples presented to improve the learning; in explainable AI the focus is on the ability of the model to explain to humans why a given solution was chosen. This collaboration between AI models and humans should not be limited only to the learning process; if we go further, we can see other terms that arise such as Usable and Useful AI. In this paper we review the state of the art of the techniques involved in the new forms of relationship between humans and ML algorithms. Our contribution is not merely listing the different approaches, but to provide definitions clarifying confusing, varied and sometimes contradictory terms; to elucidate and determine the boundaries between the different methods; and to correlate all the techniques searching for the connections and influences between them.
Collapse
|
11
|
Buschek D, Eiband M, Hussmann H. How to Support Users in Understanding Intelligent Systems? An Analysis and Conceptual Framework of User Questions Considering User Mindsets, Involvement and Knowledge Outcomes. ACM T INTERACT INTEL 2022. [DOI: 10.1145/3519264] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/01/2022]
Abstract
The opaque nature of many intelligent systems violates established usability principles and thus presents a challenge for human-computer interaction. Research in the field therefore highlights the need for transparency, scrutability, intelligibility, interpretability and explainability, among others. While all of these terms carry a vision of supporting users in understanding intelligent systems, the underlying notions and assumptions about users and their interaction with the system often remain unclear.
We review the literature in HCI through the lens of implied user questions to synthesise a conceptual framework integrating user mindsets, user involvement, and knowledge outcomes to reveal, differentiate and classify current notions in prior work. This framework aims to resolve conceptual ambiguity in the field and enables researchers to clarify their assumptions and become aware of those made in prior work. We further discuss related aspects such as stakeholders and trust, and also provide material to apply our framework in practice (e.g. ideation / design sessions). We thus hope to advance and structure the dialogue on supporting users in understanding intelligent systems.
Collapse
Affiliation(s)
- Daniel Buschek
- Department of Computer Science, University of Bayreuth, Germany
| | | | | |
Collapse
|
12
|
Van Berkel N, Opie J, Ahmad OF, Lovat L, Stoyanov D, Blandford A. Initial Responses to False Positives in AI-Supported Continuous Interactions: A Colonoscopy Case Study. ACM T INTERACT INTEL 2022. [DOI: 10.1145/3480247] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/01/2022]
Abstract
The use of artificial intelligence (AI) in clinical support systems is increasing. In this article, we focus on AI support for continuous interaction scenarios. A thorough understanding of end-user behaviour during these continuous human-AI interactions, in which user input is sustained over time and during which AI suggestions can appear at any time, is still missing. We present a controlled lab study involving 21 endoscopists and an AI colonoscopy support system. Using a custom-developed application and an off-the-shelf videogame controller, we record participants’ navigation behaviour and clinical assessment across 14 endoscopic videos. Each video is manually annotated to mimic an AI recommendation, being either true positive or false positive in nature. We find that time between AI recommendation and clinical assessment is significantly longer for incorrect assessments. Further, the type of medical content displayed significantly affects decision time. Finally, we discover that the participant’s clinical role plays a large part in the perception of clinical AI support systems. Our study presents a realistic assessment of the effects of imperfect and continuous AI support in a clinical scenario.
Collapse
Affiliation(s)
- Niels Van Berkel
- Aalborg University, Denmark and University College London, London, United Kingdom
| | - Jeremy Opie
- University College London, London, United Kingdom
| | | | - Laurence Lovat
- University College London Hospitals, London, United Kingdom
| | | | | |
Collapse
|
13
|
Meisinger N. Blue collar with tie: a human-centered reformulation of the ironies of automation. AI & SOCIETY 2022. [DOI: 10.1007/s00146-021-01320-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
AbstractWhen Lisanne Bainbridge wrote about counterintuitive consequences of the increasing human–machine interaction, she concentrated on the resulting issues for system performance, stability, and safety. Now, decades later, however, the automized work environment is substantially more pervasive, sophisticated, and interactive. Current advances in machine learning technologies reshape the value, meaning, and future of the human workforce. While the ‘human factor’ still challenges automation system architects, inconspicuously new ironic settings have evolved that only become distinctly evident from a human-centered perspective. This brief essay discusses the role of the human workforce in human–machine interaction as machine learning continues to improve, and it points to the counterintuitive insight that although the demand for blue-collar workers may decrease, exactly this labor class increasingly enters more privileged working domains and establishes itself henceforth as ‘blue collar with tie.’
Collapse
|
14
|
Eirich J, Bonart J, Jackle D, Sedlmair M, Schmid U, Fischbach K, Schreck T, Bernard J. IRVINE: A Design Study on Analyzing Correlation Patterns of Electrical Engines. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2022; 28:11-21. [PMID: 34587040 DOI: 10.1109/tvcg.2021.3114797] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
In this design study, we present IRVINE, a Visual Analytics (VA) system, which facilitates the analysis of acoustic data to detect and understand previously unknown errors in the manufacturing of electrical engines. In serial manufacturing processes, signatures from acoustic data provide valuable information on how the relationship between multiple produced engines serves to detect and understand previously unknown errors. To analyze such signatures, IRVINE leverages interactive clustering and data labeling techniques, allowing users to analyze clusters of engines with similar signatures, drill down to groups of engines, and select an engine of interest. Furthermore, IRVINE allows to assign labels to engines and clusters and annotate the cause of an error in the acoustic raw measurement of an engine. Since labels and annotations represent valuable knowledge, they are conserved in a knowledge database to be available for other stakeholders. We contribute a design study, where we developed IRVINE in four main iterations with engineers from a company in the automotive sector. To validate IRVINE, we conducted a field study with six domain experts. Our results suggest a high usability and usefulness of IRVINE as part of the improvement of a real-world manufacturing process. Specifically, with IRVINE domain experts were able to label and annotate produced electrical engines more than 30% faster.
Collapse
|
15
|
Mohseni S, Zarei N, Ragan ED. A Multidisciplinary Survey and Framework for Design and Evaluation of Explainable AI Systems. ACM T INTERACT INTEL 2021. [DOI: 10.1145/3387166] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Abstract
The need for interpretable and accountable intelligent systems grows along with the prevalence of
artificial intelligence
(
AI
) applications used in everyday life.
Explainable AI
(
XAI
) systems are intended to self-explain the reasoning behind system decisions and predictions. Researchers from different disciplines work together to define, design, and evaluate explainable systems. However, scholars from different disciplines focus on different objectives and fairly independent topics of XAI research, which poses challenges for identifying appropriate design and evaluation methodology and consolidating knowledge across efforts. To this end, this article presents a survey and framework intended to share knowledge and experiences of XAI design and evaluation methods across multiple disciplines. Aiming to support diverse design goals and evaluation methods in XAI research, after a thorough review of XAI related papers in the fields of machine learning, visualization, and human-computer interaction, we present a categorization of XAI design goals and evaluation methods. Our categorization presents the mapping between design goals for different XAI user groups and their evaluation methods. From our findings, we develop a framework with step-by-step design guidelines paired with evaluation methods to close the iterative design and evaluation cycles in multidisciplinary XAI teams. Further, we provide summarized ready-to-use tables of evaluation methods and recommendations for different goals in XAI research.
Collapse
|
16
|
Smith AG, Petersen J, Terrones-Campos C, Berthelsen AK, Forbes NJ, Darkner S, Specht L, Vogelius IR. RootPainter3D: Interactive-machine-learning enables rapid and accurate contouring for radiotherapy. Med Phys 2021; 49:461-473. [PMID: 34783028 DOI: 10.1002/mp.15353] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2021] [Revised: 09/22/2021] [Accepted: 10/28/2021] [Indexed: 11/07/2022] Open
Abstract
PURPOSE Organ-at-risk contouring is still a bottleneck in radiotherapy, with many deep learning methods falling short of promised results when evaluated on clinical data. We investigate the accuracy and time-savings resulting from the use of an interactive-machine-learning method for an organ-at-risk contouring task. METHODS We implement an open-source interactive-machine-learning software application that facilitates corrective-annotation for deep-learning generated contours on X-ray CT images. A trained-physician contoured 933 hearts using our software by delineating the first image, starting model training, and then correcting the model predictions for all subsequent images. These corrections were added into the training data, which was used for continuously training the assisting model. From the 933 hearts, the same physician also contoured the first 10 and last 10 in Eclipse (Varian) to enable comparison in terms of accuracy and duration. RESULTS We find strong agreement with manual delineations, with a dice score of 0.95. The annotations created using corrective-annotation also take less time to create as more images are annotated, resulting in substantial time savings compared to manual methods. After 923 images had been delineated, hearts took 2 min and 2 s to delineate on average, which includes time to evaluate the initial model prediction and assign the needed corrections, compared to 7 min and 1 s when delineating manually. CONCLUSIONS Our experiment demonstrates that interactive-machine-learning with corrective-annotation provides a fast and accessible way for non computer-scientists to train deep-learning models to segment their own structures of interest as part of routine clinical workflows.
Collapse
Affiliation(s)
- Abraham George Smith
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark.,Department of Oncology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark
| | - Jens Petersen
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark.,Department of Oncology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark
| | - Cynthia Terrones-Campos
- Department of Oncology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark.,Department of Infectious Diseases, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark
| | - Anne Kiil Berthelsen
- Department of Oncology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark.,Department of Clinical Physiology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark
| | - Nora Jarrett Forbes
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark.,Department of Oncology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark
| | - Sune Darkner
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark
| | - Lena Specht
- Department of Oncology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark
| | - Ivan Richter Vogelius
- Department of Oncology, Rigshospitalet, University of Copenhagen, Copenhagen, Denmark.,Department of Health and Medical Sciences, University of Copenhagen, Copenhagen, Denmark
| |
Collapse
|
17
|
Design of Generalized Search Interfaces for Health Informatics. INFORMATION 2021. [DOI: 10.3390/info12080317] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
In this paper, we investigate ontology-supported interfaces for health informatics search tasks involving large document sets. We begin by providing background on health informatics, machine learning, and ontologies. We review leading research on health informatics search tasks to help formulate high-level design criteria. We use these criteria to examine traditional design strategies for search interfaces. To demonstrate the utility of the criteria, we apply them to the design of ONTology-supported Search Interface (ONTSI), a demonstrative, prototype system. ONTSI allows users to plug-and-play document sets and expert-defined domain ontologies through a generalized search interface. ONTSI’s goal is to help align users’ common vocabulary with the domain-specific vocabulary of the plug-and-play document set. We describe the functioning and utility of ONTSI in health informatics search tasks through a workflow and a scenario. We conclude with a summary of ongoing evaluations, limitations, and future research.
Collapse
|
18
|
Gutzwiller RS, Reeder J. Dancing With Algorithms: Interaction Creates Greater Preference and Trust in Machine-Learned Behavior. HUMAN FACTORS 2021; 63:854-867. [PMID: 32048883 DOI: 10.1177/0018720820903893] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
OBJECTIVE We examined a method of machine learning (ML) to evaluate its potential to develop more trustworthy control of unmanned vehicle area search behaviors. BACKGROUND ML typically lacks interaction with the user. Novel interactive machine learning (IML) techniques incorporate user feedback, enabling observation of emerging ML behaviors, and human collaboration during ML of a task. This may enable trust and recognition of these algorithms. METHOD Participants judged and selected behaviors in a low and a high interaction condition (IML) over the course of behavior evolution using ML. User trust in the outputs, as well as preference, and ability to discriminate and recognize the behaviors were measured. RESULTS Compared to noninteractive techniques, IML behaviors were more trusted and preferred, as well as recognizable, separate from non-IML behaviors, and approached similar performance as pure ML models. CONCLUSION IML shows promise for creating behaviors by involving the user; this is the first extension of this technique for vehicle behavior model development targeting user satisfaction and is unique in its multifaceted evaluation of how users perceived, trusted, and implemented these learned controllers. APPLICATION There are many contexts where the brittleness of ML cannot be trusted, but the advantage of ML over traditional programmed behaviors may be large, as in some military operations where they could be scaled. IML in this early form appears to generate satisfactory behaviors without sacrificing performance, use, or trust in the behavior, but more work is necessary.
Collapse
Affiliation(s)
| | - John Reeder
- 41489 Naval Information Warfare Center, San Diego, CA, USA
| |
Collapse
|
19
|
Guckert M, Gumpfer N, Hannig J, Keller T, Urquhart N. A conceptual framework for establishing trust in real world intelligent systems. COGN SYST RES 2021. [DOI: 10.1016/j.cogsys.2021.04.001] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
20
|
Automatic Visual Attention Detection for Mobile Eye Tracking Using Pre-Trained Computer Vision Models and Human Gaze. SENSORS 2021; 21:s21124143. [PMID: 34208736 PMCID: PMC8235043 DOI: 10.3390/s21124143] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/12/2021] [Revised: 06/11/2021] [Accepted: 06/12/2021] [Indexed: 12/28/2022]
Abstract
Processing visual stimuli in a scene is essential for the human brain to make situation-aware decisions. These stimuli, which are prevalent subjects of diagnostic eye tracking studies, are commonly encoded as rectangular areas of interest (AOIs) per frame. Because it is a tedious manual annotation task, the automatic detection and annotation of visual attention to AOIs can accelerate and objectify eye tracking research, in particular for mobile eye tracking with egocentric video feeds. In this work, we implement two methods to automatically detect visual attention to AOIs using pre-trained deep learning models for image classification and object detection. Furthermore, we develop an evaluation framework based on the VISUS dataset and well-known performance metrics from the field of activity recognition. We systematically evaluate our methods within this framework, discuss potentials and limitations, and propose ways to improve the performance of future automatic visual attention detection methods.
Collapse
|
21
|
Gresse von Wangenheim C, Hauck JCR, Pacheco FS, Bertonceli Bueno MF. Visual tools for teaching machine learning in K-12: A ten-year systematic mapping. EDUCATION AND INFORMATION TECHNOLOGIES 2021; 26:5733-5778. [PMID: 33967587 PMCID: PMC8087535 DOI: 10.1007/s10639-021-10570-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/06/2020] [Accepted: 04/27/2021] [Indexed: 06/12/2023]
Abstract
Teaching Machine Learning in school helps students to be better prepared for a society rapidly changing due to the impact of Artificial Intelligence. This requires age-appropriate tools that allow students to develop a comprehensive understanding of Machine Learning in order to become creators of smart solutions. Following the trend of visual languages for introducing algorithms and programming in K-12, we present a ten-year systematic mapping of emerging visual tools that support the teaching of Machine Learning at this educational stage and analyze the tools concerning their educational characteristics, support for the development of ML models as well as their deployment and how the tools have been developed and evaluated. As a result, we encountered 16 tools targeting students mostly as part of short duration extracurricular activities. Tools mainly support the interactive development of ML models for image recognition tasks using supervised learning covering basic steps of the ML process. Being integrated into popular block-based programming languages (primarily Scratch and App Inventor), they also support the deployment of the created ML models as part of games or mobile applications. Findings indicate that the tools can effectively leverage students' understanding of Machine Learning, however, further studies regarding the design of the tools concerning educational aspects are required to better guide their effective adoption in schools and their enhancement to support the learning process more comprehensively.
Collapse
Affiliation(s)
| | - Jean C. R. Hauck
- Department of Informatics and Statistics, Federal University of Santa Catarina, Florianópolis, Brazil
| | - Fernando S. Pacheco
- Department of Electronics, Federal Institute of Santa Catarina, Florianópolis, Brazil
| | | |
Collapse
|
22
|
Maadi M, Akbarzadeh Khorshidi H, Aickelin U. A Review on Human-AI Interaction in Machine Learning and Insights for Medical Applications. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2021; 18:ijerph18042121. [PMID: 33671609 PMCID: PMC7926732 DOI: 10.3390/ijerph18042121] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/30/2020] [Revised: 02/08/2021] [Accepted: 02/12/2021] [Indexed: 11/19/2022]
Abstract
Objective: To provide a human–Artificial Intelligence (AI) interaction review for Machine Learning (ML) applications to inform how to best combine both human domain expertise and computational power of ML methods. The review focuses on the medical field, as the medical ML application literature highlights a special necessity of medical experts collaborating with ML approaches. Methods: A scoping literature review is performed on Scopus and Google Scholar using the terms “human in the loop”, “human in the loop machine learning”, and “interactive machine learning”. Peer-reviewed papers published from 2015 to 2020 are included in our review. Results: We design four questions to investigate and describe human–AI interaction in ML applications. These questions are “Why should humans be in the loop?”, “Where does human–AI interaction occur in the ML processes?”, “Who are the humans in the loop?”, and “How do humans interact with ML in Human-In-the-Loop ML (HILML)?”. To answer the first question, we describe three main reasons regarding the importance of human involvement in ML applications. To address the second question, human–AI interaction is investigated in three main algorithmic stages: 1. data producing and pre-processing; 2. ML modelling; and 3. ML evaluation and refinement. The importance of the expertise level of the humans in human–AI interaction is described to answer the third question. The number of human interactions in HILML is grouped into three categories to address the fourth question. We conclude the paper by offering a discussion on open opportunities for future research in HILML.
Collapse
|
23
|
Liu J, Dwyer T, Tack G, Gratzl S, Marriott K. Supporting the Problem-Solving Loop: Designing Highly Interactive Optimisation Systems. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2021; 27:1764-1774. [PMID: 33112748 DOI: 10.1109/tvcg.2020.3030364] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Efficient optimisation algorithms have become important tools for finding high-quality solutions to hard, real-world problems such as production scheduling, timetabling, or vehicle routing. These algorithms are typically "black boxes" that work on mathematical models of the problem to solve. However, many problems are difficult to fully specify, and require a "human in the loop" who collaborates with the algorithm by refining the model and guiding the search to produce acceptable solutions. Recently, the Problem-Solving Loop was introduced as a high-level model of such interactive optimisation. Here, we present and evaluate nine recommendations for the design of interactive visualisation tools supporting the Problem-Solving Loop. They range from the choice of visual representation for solutions and constraints to the use of a solution gallery to support exploration of alternate solutions. We first examined the applicability of the recommendations by investigating how well they had been supported in previous interactive optimisation tools. We then evaluated the recommendations in the context of the vehicle routing problem with time windows (VRPTW). To do so we built a sophisticated interactive visual system for solving VRPTW that was informed by the recommendations. Ten participants then used this system to solve a variety of routing problems. We report on participant comments and interaction patterns with the tool. These showed the tool was regarded as highly usable and the results generally supported the usefulness of the underlying recommendations.
Collapse
|
24
|
Abstract
This article attempts to bridge the gap between widely discussed ethical principles of Human-centered AI (HCAI) and practical steps for effective governance. Since HCAI systems are developed and implemented in multiple organizational structures, I propose 15 recommendations at three levels of governance: team, organization, and industry. The recommendations are intended to increase the reliability, safety, and trustworthiness of HCAI systems: (1) reliable systems based on sound software engineering practices, (2) safety culture through business management strategies, and (3) trustworthy certification by independent oversight. Software engineering practices within teams include audit trails to enable analysis of failures, software engineering workflows, verification and validation testing, bias testing to enhance fairness, and explainable user interfaces. The safety culture within organizations comes from management strategies that include leadership commitment to safety, hiring and training oriented to safety, extensive reporting of failures and near misses, internal review boards for problems and future plans, and alignment with industry standard practices. The trustworthiness certification comes from industry-wide efforts that include government interventions and regulation, accounting firms conducting external audits, insurance companies compensating for failures, non-governmental and civil society organizations advancing design principles, and professional organizations and research institutes developing standards, policies, and novel ideas. The larger goal of effective governance is to limit the dangers and increase the benefits of HCAI to individuals, organizations, and society.
Collapse
|
25
|
Eiband M, Völkel ST, Buschek D, Cook S, Hussmann H. A Method and Analysis to Elicit User-Reported Problems in Intelligent Everyday Applications. ACM T INTERACT INTEL 2020. [DOI: 10.1145/3370927] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
The complex nature of intelligent systems motivates work on supporting users during interaction, for example, through explanations. However, as of yet, there is little empirical evidence in regard to specific problems users face when applying such systems in everyday situations. This article contributes a novel method and analysis to investigate such problems
as reported by users:
We analysed 45,448 reviews of four apps on the Google Play Store (Facebook, Netflix, Google Maps, and Google Assistant) with sentiment analysis and topic modelling to reveal problems during interaction that can be attributed to the apps’ algorithmic decision-making. We enriched this data with users’ coping and support strategies through a follow-up online survey (N = 286). In particular, we found problems and strategies related to content, algorithm, user choice, and feedback. We discuss corresponding implications for designing user support, highlighting the importance of user control and explanations of output rather than processes.
Collapse
Affiliation(s)
| | | | - Daniel Buschek
- University of Bayreuth, Universitätsstraße, Bayreuth, Germany
| | - Sophia Cook
- LMU Munich, Frauenlobstraße, Munich, Germany
| | | |
Collapse
|
26
|
Interactive clustering: a scoping review. Artif Intell Rev 2020. [DOI: 10.1007/s10462-020-09913-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
27
|
Lindvall M, Sanner A, Petré F, Lindman K, Treanor D, Lundström C, Löwgren J. TissueWand, a Rapid Histopathology Annotation Tool. J Pathol Inform 2020; 11:27. [PMID: 33042606 PMCID: PMC7518350 DOI: 10.4103/jpi.jpi_5_20] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2020] [Revised: 03/23/2020] [Accepted: 05/19/2020] [Indexed: 12/28/2022] Open
Abstract
Background: Recent advancements in machine learning (ML) bring great possibilities for the development of tools to assist with diagnostic tasks within histopathology. However, these approaches typically require a large amount of ground truth training data in the form of image annotations made by human experts. As such annotation work is a very time-consuming task, there is a great need for tools that can assist in this process, saving time while not sacrificing annotation quality. Methods: In an iterative design process, we developed TissueWand – an interactive tool designed for efficient annotation of gigapixel-sized histopathological images, not being constrained to a predefined annotation task. Results: Several findings regarding appropriate interaction concepts were made, where a key design component was semi-automation based on rapid interaction feedback in a local region. In a user study, the resulting tool was shown to cause substantial speed-up compared to manual work while maintaining quality. Conclusions: The TissueWand tool shows promise to replace manual methods for early stages of dataset curation where no task-specific ML model yet exists to aid the effort.
Collapse
Affiliation(s)
- Martin Lindvall
- Sectra AB, Research Department, Linköping, Sweden.,Center for Medical Image Science and Visualization, Linköping University, Linköping, Sweden.,Department of Science and Technology (ITN), Linköping University, Norrköping, Sweden
| | | | | | - Karin Lindman
- Department of Clinical Pathology, Region Östergötland, Linköping, Sweden.,Department of Biomedical and Clinical Sciences, Linköping University, Linköping, Sweden
| | - Darren Treanor
- Center for Medical Image Science and Visualization, Linköping University, Linköping, Sweden.,Department of Cellular Pathology, Leeds Teaching Hospitals NHS Trust, Leeds, UK.,University of Leeds, Leeds, UK
| | - Claes Lundström
- Sectra AB, Research Department, Linköping, Sweden.,Center for Medical Image Science and Visualization, Linköping University, Linköping, Sweden.,Department of Science and Technology (ITN), Linköping University, Norrköping, Sweden
| | - Jonas Löwgren
- Department of Science and Technology (ITN), Linköping University, Norrköping, Sweden
| |
Collapse
|
28
|
Biswas R, Barz M, Sonntag D. Towards Explanatory Interactive Image Captioning Using Top-Down and Bottom-Up Features, Beam Search and Re-ranking. KUNSTLICHE INTELLIGENZ 2020. [DOI: 10.1007/s13218-020-00679-2] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
AbstractImage captioning is a challenging multimodal task. Significant improvements could be obtained by deep learning. Yet, captions generated by humans are still considered better, which makes it an interesting application for interactive machine learning and explainable artificial intelligence methods. In this work, we aim at improving the performance and explainability of the state-of-the-art method Show, Attend and Tell by augmenting their attention mechanism using additional bottom-up features. We compute visual attention on the joint embedding space formed by the union of high-level features and the low-level features obtained from the object specific salient regions of the input image. We embed the content of bounding boxes from a pre-trained Mask R-CNN model. This delivers state-of-the-art performance, while it provides explanatory features. Further, we discuss how interactive model improvement can be realized through re-ranking caption candidates using beam search decoders and explanatory features. We show that interactive re-ranking of beam search candidates has the potential to outperform the state-of-the-art in image captioning.
Collapse
|
29
|
Nadj M, Knaeble M, Li MX, Maedche A. Power to the Oracle? Design Principles for Interactive Labeling Systems in Machine Learning. KUNSTLICHE INTELLIGENZ 2020. [DOI: 10.1007/s13218-020-00634-1] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
30
|
Ming Y, Xu P, Cheng F, Qu H, Ren L. ProtoSteer: Steering Deep Sequence Model with Prototypes. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2020; 26:238-248. [PMID: 31514137 DOI: 10.1109/tvcg.2019.2934267] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Recently we have witnessed growing adoption of deep sequence models (e.g. LSTMs) in many application domains, including predictive health care, natural language processing, and log analysis. However, the intricate working mechanism of these models confines their accessibility to the domain experts. Their black-box nature also makes it a challenging task to incorporate domain-specific knowledge of the experts into the model. In ProtoSteer (Prototype Steering), we tackle the challenge of directly involving the domain experts to steer a deep sequence model without relying on model developers as intermediaries. Our approach originates in case-based reasoning, which imitates the common human problem-solving process of consulting past experiences to solve new problems. We utilize ProSeNet (Prototype Sequence Network), which learns a small set of exemplar cases (i.e., prototypes) from historical data. In ProtoSteer they serve both as an efficient visual summary of the original data and explanations of model decisions. With ProtoSteer the domain experts can inspect, critique, and revise the prototypes interactively. The system then incorporates user-specified prototypes and incrementally updates the model. We conduct extensive case studies and expert interviews in application domains including sentiment analysis on texts and predictive diagnostics based on vehicle fault logs. The results demonstrate that involvements of domain users can help obtain more interpretable models with concise prototypes while retaining similar accuracy.
Collapse
|
31
|
Clabaugh C, Mahajan K, Jain S, Pakkar R, Becerra D, Shi Z, Deng E, Lee R, Ragusa G, Matarić M. Long-Term Personalization of an In-Home Socially Assistive Robot for Children With Autism Spectrum Disorders. Front Robot AI 2019; 6:110. [PMID: 33501125 PMCID: PMC7805891 DOI: 10.3389/frobt.2019.00110] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2019] [Accepted: 10/15/2019] [Indexed: 11/25/2022] Open
Abstract
Socially assistive robots (SAR) have shown great potential to augment the social and educational development of children with autism spectrum disorders (ASD). As SAR continues to substantiate itself as an effective enhancement to human intervention, researchers have sought to study its longitudinal impacts in real-world environments, including the home. Computational personalization stands out as a central computational challenge as it is necessary to enable SAR systems to adapt to each child's unique and changing needs. Toward that end, we formalized personalization as a hierarchical human robot learning framework (hHRL) consisting of five controllers (disclosure, promise, instruction, feedback, and inquiry) mediated by a meta-controller that utilized reinforcement learning to personalize instruction challenge levels and robot feedback based on each user's unique learning patterns. We instantiated and evaluated the approach in a study with 17 children with ASD, aged 3–7 years old, over month-long interventions in their homes. Our findings demonstrate that the fully autonomous SAR system was able to personalize its instruction and feedback over time to each child's proficiency. As a result, every child participant showed improvements in targeted skills and long-term retention of intervention content. Moreover, all child users were engaged for a majority of the intervention, and their families reported the SAR system to be useful and adaptable. In summary, our results show that autonomous, personalized SAR interventions are both feasible and effective in providing long-term in-home developmental support for children with diverse learning needs.
Collapse
Affiliation(s)
- Caitlyn Clabaugh
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Kartik Mahajan
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Shomik Jain
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Roxanna Pakkar
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - David Becerra
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Zhonghao Shi
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Eric Deng
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Rhianna Lee
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Gisele Ragusa
- STEM Education Research Group, Division of Engineering Education, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| | - Maja Matarić
- Interaction Lab, Department of Computer Science, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
| |
Collapse
|
32
|
A Novel Hybrid Swarm Optimized Multilayer Neural Network for Spatial Prediction of Flash Floods in Tropical Areas Using Sentinel-1 SAR Imagery and Geospatial Data. SENSORS 2018; 18:s18113704. [PMID: 30384451 PMCID: PMC6263740 DOI: 10.3390/s18113704] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/26/2018] [Revised: 09/22/2018] [Accepted: 10/11/2018] [Indexed: 11/29/2022]
Abstract
Flash floods are widely recognized as one of the most devastating natural hazards in the world, therefore prediction of flash flood-prone areas is crucial for public safety and emergency management. This research proposes a new methodology for spatial prediction of flash floods based on Sentinel-1 SAR imagery and a new hybrid machine learning technique. The SAR imagery is used to detect flash flood inundation areas, whereas the new machine learning technique, which is a hybrid of the firefly algorithm (FA), Levenberg–Marquardt (LM) backpropagation, and an artificial neural network (named as FA-LM-ANN), was used to construct the prediction model. The Bac Ha Bao Yen (BHBY) area in the northwestern region of Vietnam was used as a case study. Accordingly, a Geographical Information System (GIS) database was constructed using 12 input variables (elevation, slope, aspect, curvature, topographic wetness index, stream power index, toposhade, stream density, rainfall, normalized difference vegetation index, soil type, and lithology) and subsequently the output of flood inundation areas was mapped. Using the database and FA-LM-ANN, the flash flood model was trained and verified. The model performance was validated via various performance metrics including the classification accuracy rate, the area under the curve, precision, and recall. Then, the flash flood model that produced the highest performance was compared with benchmarks, indicating that the combination of FA and LM backpropagation is proven to be very effective and the proposed FA-LM-ANN is a new and useful tool for predicting flash flood susceptibility.
Collapse
|