1
|
Nam KH, Kim DY, Kim DH, Lee JH, Lee JI, Kim MJ, Park JY, Hwang JH, Yun SS, Choi BK, Kim MG, Han IH. Conversational Artificial Intelligence for Spinal Pain Questionnaire: Validation and User Satisfaction. Neurospine 2022; 19:348-356. [PMID: 35577340 PMCID: PMC9260557 DOI: 10.14245/ns.2143080.540] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2021] [Accepted: 03/02/2022] [Indexed: 12/02/2022] Open
Abstract
Objective The purpose of our study is to develop a spoken dialogue system (SDS) for pain questionnaire in patients with spinal disease. We evaluate user satisfaction and validated the performance accuracy of the SDS in medical staff and patients.
Methods The SDS was developed to investigate pain and related psychological issues in patients with spinal diseases based on the pain questionnaire protocol. We recognized patients’ various answers, summarized important information, and documented them. User satisfaction and performance accuracy were evaluated in 30 potential users of SDS, including doctors, nurses, and patients and statistically analyzed.
Results The overall satisfaction score of 30 patients was 5.5 ± 1.4 out of 7 points. Satisfaction scores were 5.3 ± 0.8 for doctors, 6.0 ± 0.6 for nurses, and 5.3 ± 0.5 for patients. In terms of performance accuracy, the number of repetitions of the same question was 13, 16, and 33 (13.5%, 16.8%, and 34.7%) for doctors, nurses, and patients, respectively. The number of errors in the summarized comment by the SDS was 5, 0, and 11 (5.2%, 0.0%, and 11.6 %), respectively. The number of summarization omissions was 7, 5, and 7 (7.3%, 5.3%, and 7.4%), respectively.
Conclusion This is the first study in which voice-based conversational artificial intelligence (AI) was developed for a spinal pain questionnaire and validated by medical staff and patients. The conversational AI showed favorable results in terms of user satisfaction and performance accuracy. Conversational AI can be useful for the diagnosis and remote monitoring of various patients as well as for pain questionnaires in the future.
Collapse
Affiliation(s)
- Kyoung Hyup Nam
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Da Young Kim
- Human-Robot Interaction Center, Korea Institute of Robotics and Technology Convergence, Pohang, Korea
| | - Dong Hwan Kim
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Jung Hwan Lee
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Jae Il Lee
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Mi Jeong Kim
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Joo Young Park
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Jae Hyun Hwang
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Sang Seok Yun
- Division of Mechanical Convergence Engineering, Silla University, Busan, Korea
| | - Byung Kwan Choi
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
| | - Min Gyu Kim
- Human-Robot Interaction Center, Korea Institute of Robotics and Technology Convergence, Pohang, Korea
- Co-corresponding Author Min Gyu Kim Human-Robot Interaction Center, Korea Institute of Robotics and Technology Convergence, 39 Jigok-ro, Nam-gu, Pohang 37666, Korea
| | - In Ho Han
- Department of Neurosurgery, Medical Research Institute, Pusan National University Hospital, Busan, Korea
- Corresponding Author In Ho Han Department of Neurosurgery, Pusan National University Hospital, 179 Gudeok-ro, Seo-gu, Busan 49241, Korea
| |
Collapse
|
2
|
Von Gerich H, Moen H, Block LJ, Chu CH, DeForest H, Hobensack M, Michalowski M, Mitchell J, Nibber R, Olalia MA, Pruinelli L, Ronquillo CE, Topaz M, Peltonen LM. Artificial Intelligence -based technologies in nursing: A scoping literature review of the evidence. Int J Nurs Stud 2021; 127:104153. [DOI: 10.1016/j.ijnurstu.2021.104153] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2021] [Revised: 11/23/2021] [Accepted: 12/01/2021] [Indexed: 12/20/2022]
|
3
|
Mairittha T, Mairittha N, Inoue S. Automatic Labeled Dialogue Generation for Nursing Record Systems. J Pers Med 2020; 10:jpm10030062. [PMID: 32708593 PMCID: PMC7564988 DOI: 10.3390/jpm10030062] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2020] [Revised: 06/29/2020] [Accepted: 07/09/2020] [Indexed: 11/16/2022] Open
Abstract
The integration of digital voice assistants in nursing residences is becoming increasingly important to facilitate nursing productivity with documentation. A key idea behind this system is training natural language understanding (NLU) modules that enable the machine to classify the purpose of the user utterance (intent) and extract pieces of valuable information present in the utterance (entity). One of the main obstacles when creating robust NLU is the lack of sufficient labeled data, which generally relies on human labeling. This process is cost-intensive and time-consuming, particularly in the high-level nursing care domain, which requires abstract knowledge. In this paper, we propose an automatic dialogue labeling framework of NLU tasks, specifically for nursing record systems. First, we apply data augmentation techniques to create a collection of variant sample utterances. The individual evaluation result strongly shows a stratification rate, with regard to both fluency and accuracy in utterances. We also investigate the possibility of applying deep generative models for our augmented dataset. The preliminary character-based model based on long short-term memory (LSTM) obtains an accuracy of 90% and generates various reasonable texts with BLEU scores of 0.76. Secondly, we introduce an idea for intent and entity labeling by using feature embeddings and semantic similarity-based clustering. We also empirically evaluate different embedding methods for learning good representations that are most suitable to use with our data and clustering tasks. Experimental results show that fastText embeddings produce strong performances both for intent labeling and on entity labeling, which achieves an accuracy level of 0.79 and 0.78 f1-scores and 0.67 and 0.61 silhouette scores, respectively.
Collapse
|