1
|
Lovis C, Weber J, Liopyris K, Braun RP, Marghoob AA, Quigley EA, Nelson K, Prentice K, Duhaime E, Halpern AC, Rotemberg V. Agreement Between Experts and an Untrained Crowd for Identifying Dermoscopic Features Using a Gamified App: Reader Feasibility Study. JMIR Med Inform 2023; 11:e38412. [PMID: 36652282 PMCID: PMC9892985 DOI: 10.2196/38412] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2022] [Revised: 09/28/2022] [Accepted: 10/16/2022] [Indexed: 01/19/2023] Open
Abstract
BACKGROUND Dermoscopy is commonly used for the evaluation of pigmented lesions, but agreement between experts for identification of dermoscopic structures is known to be relatively poor. Expert labeling of medical data is a bottleneck in the development of machine learning (ML) tools, and crowdsourcing has been demonstrated as a cost- and time-efficient method for the annotation of medical images. OBJECTIVE The aim of this study is to demonstrate that crowdsourcing can be used to label basic dermoscopic structures from images of pigmented lesions with similar reliability to a group of experts. METHODS First, we obtained labels of 248 images of melanocytic lesions with 31 dermoscopic "subfeatures" labeled by 20 dermoscopy experts. These were then collapsed into 6 dermoscopic "superfeatures" based on structural similarity, due to low interrater reliability (IRR): dots, globules, lines, network structures, regression structures, and vessels. These images were then used as the gold standard for the crowd study. The commercial platform DiagnosUs was used to obtain annotations from a nonexpert crowd for the presence or absence of the 6 superfeatures in each of the 248 images. We replicated this methodology with a group of 7 dermatologists to allow direct comparison with the nonexpert crowd. The Cohen κ value was used to measure agreement across raters. RESULTS In total, we obtained 139,731 ratings of the 6 dermoscopic superfeatures from the crowd. There was relatively lower agreement for the identification of dots and globules (the median κ values were 0.526 and 0.395, respectively), whereas network structures and vessels showed the highest agreement (the median κ values were 0.581 and 0.798, respectively). This pattern was also seen among the expert raters, who had median κ values of 0.483 and 0.517 for dots and globules, respectively, and 0.758 and 0.790 for network structures and vessels. The median κ values between nonexperts and thresholded average-expert readers were 0.709 for dots, 0.719 for globules, 0.714 for lines, 0.838 for network structures, 0.818 for regression structures, and 0.728 for vessels. CONCLUSIONS This study confirmed that IRR for different dermoscopic features varied among a group of experts; a similar pattern was observed in a nonexpert crowd. There was good or excellent agreement for each of the 6 superfeatures between the crowd and the experts, highlighting the similar reliability of the crowd for labeling dermoscopic images. This confirms the feasibility and dependability of using crowdsourcing as a scalable solution to annotate large sets of dermoscopic images, with several potential clinical and educational applications, including the development of novel, explainable ML tools.
Collapse
Affiliation(s)
| | - Jochen Weber
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Konstantinos Liopyris
- Department of Dermatology, Andreas Syggros Hospital of Cutaneous and Venereal Diseases, University of Athens, Athens, Greece
| | - Ralph P Braun
- Department of Dermatology, University Hospital Zurich, Zurich, Switzerland
| | - Ashfaq A Marghoob
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Elizabeth A Quigley
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Kelly Nelson
- Department of Dermatology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | | | | | - Allan C Halpern
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Veronica Rotemberg
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| |
Collapse
|
2
|
Maleki F, Muthukrishnan N, Ovens K, Reinhold C, Forghani R. Machine Learning Algorithm Validation: From Essentials to Advanced Applications and Implications for Regulatory Certification and Deployment. Neuroimaging Clin N Am 2021; 30:433-445. [PMID: 33038994 DOI: 10.1016/j.nic.2020.08.004] [Citation(s) in RCA: 40] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
The deployment of machine learning (ML) models in the health care domain can increase the speed and accuracy of diagnosis and improve treatment planning and patient care. Translating academic research to applications that are deployable in clinical settings requires the ability to generalize and high reproducibility, which are contingent on a rigorous and sound methodology for the development and evaluation of ML models. This article describes the fundamental concepts and processes for ML model evaluation and highlights common workflows. It concludes with a discussion of the requirements for the deployment of ML models in clinical settings.
Collapse
Affiliation(s)
- Farhad Maleki
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada
| | - Nikesh Muthukrishnan
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada
| | - Katie Ovens
- Department of Computer Science, University of Saskatchewan, 176 Thorvaldson Bldg, 110 Science Place, Saskatoon S7N 5C9, Canada
| | - Caroline Reinhold
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada; Department of Radiology, McGill University, 1650 Cedar Avenue, Montreal, Quebec H3G 1A4, Canada
| | - Reza Forghani
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada; Department of Radiology, McGill University, 1650 Cedar Avenue, Montreal, Quebec H3G 1A4, Canada; Segal Cancer Centre, Lady Davis Institute for Medical Research, Jewish General Hospital, 3755 Cote Ste-Catherine Road, Montreal, Quebec H3T 1E2, Canada; Gerald Bronfman Department of Oncology, McGill University, Suite 720, 5100 Maisonneuve Boulevard West, Montreal, Quebec H4A3T2, Canada; Department of Otolaryngology - Head and Neck Surgery, Royal Victoria Hospital, McGill University Health Centre, 1001 boul. Decarie Boulevard, Montreal, Quebec H3A 3J1, Canada.
| |
Collapse
|
3
|
Park JH, Nadeem S, Boorboor S, Marino J, Kaufman A. CMed: Crowd Analytics for Medical Imaging Data. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2021; 27:2869-2880. [PMID: 31751242 PMCID: PMC7859862 DOI: 10.1109/tvcg.2019.2953026] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
We present a visual analytics framework, CMed, for exploring medical image data annotations acquired from crowdsourcing. CMed can be used to visualize, classify, and filter crowdsourced clinical data based on a number of different metrics such as detection rate, logged events, and clustering of the annotations. CMed provides several interactive linked visualization components to analyze the crowd annotation results for a particular video and the associated workers. Additionally, all results of an individual worker can be inspected using multiple linked views in our CMed framework. We allow a crowdsourcing application analyst to observe patterns and gather insights into the crowdsourced medical data, helping him/her design future crowdsourcing applications for optimal output from the workers. We demonstrate the efficacy of our framework with two medical crowdsourcing studies: polyp detection in virtual colonoscopy videos and lung nodule detection in CT thin-slab maximum intensity projection videos. We also provide experts' feedback to show the effectiveness of our framework. Lastly, we share the lessons we learned from our framework with suggestions for integrating our framework into a clinical workflow.
Collapse
|
4
|
Li J. Design and implementation of distributed asynchronous data aided computer information interaction system. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2020. [DOI: 10.3233/jifs-189299] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Under the influence of novel corona virus pneumonia epidemic prevention and control, it has put forward higher requirements for data storage and processing for personnel management system. The distributed asynchronous data aided computer information interaction system can solve the problem of multi node concurrent data processing. The traditional computer information interaction system has poor real-time performance, low precision and asynchronous data processing ability. The invocation features of message queuing asynchronous caching mode are combined with the standardization of Web services and cross language with cross platform access features in this paper. Through the combination of the two technologies, a flexible and universal asynchronous interaction architecture of distributed system is established. Based on Web service technology and system to system access, the call and response of tasks between modules are carried out in the system, which makes the interaction between the whole system have the characteristics of message driven. The test result shows that the system proposed in this paper has good real-time performance and strong data processing ability. It is suitable for the data interaction of distributed personal management system under the influence of novel corona virus pneumonia epidemic prevention and control.
Collapse
Affiliation(s)
- Jiwei Li
- Henan Logistics Vocational College, Zhengzhou, Henan, China
| |
Collapse
|
5
|
Patel BN, Rosenberg L, Willcox G, Baltaxe D, Lyons M, Irvin J, Rajpurkar P, Amrhein T, Gupta R, Halabi S, Langlotz C, Lo E, Mammarappallil J, Mariano AJ, Riley G, Seekins J, Shen L, Zucker E, Lungren M. Human-machine partnership with artificial intelligence for chest radiograph diagnosis. NPJ Digit Med 2019; 2:111. [PMID: 31754637 PMCID: PMC6861262 DOI: 10.1038/s41746-019-0189-7] [Citation(s) in RCA: 54] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2019] [Accepted: 10/23/2019] [Indexed: 12/17/2022] Open
Abstract
Human-in-the-loop (HITL) AI may enable an ideal symbiosis of human experts and AI models, harnessing the advantages of both while at the same time overcoming their respective limitations. The purpose of this study was to investigate a novel collective intelligence technology designed to amplify the diagnostic accuracy of networked human groups by forming real-time systems modeled on biological swarms. Using small groups of radiologists, the swarm-based technology was applied to the diagnosis of pneumonia on chest radiographs and compared against human experts alone, as well as two state-of-the-art deep learning AI models. Our work demonstrates that both the swarm-based technology and deep-learning technology achieved superior diagnostic accuracy than the human experts alone. Our work further demonstrates that when used in combination, the swarm-based technology and deep-learning technology outperformed either method alone. The superior diagnostic accuracy of the combined HITL AI solution compared to radiologists and AI alone has broad implications for the surging clinical AI deployment and implementation strategies in future practice.
Collapse
Affiliation(s)
- Bhavik N Patel
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Louis Rosenberg
- Unanimous AI, 2443 Fillmore Street #116, San Francisco, CA 94115-1814 USA
| | - Gregg Willcox
- Unanimous AI, 2443 Fillmore Street #116, San Francisco, CA 94115-1814 USA
| | - David Baltaxe
- Unanimous AI, 2443 Fillmore Street #116, San Francisco, CA 94115-1814 USA
| | - Mimi Lyons
- Unanimous AI, 2443 Fillmore Street #116, San Francisco, CA 94115-1814 USA
| | - Jeremy Irvin
- 3Department of Computer Science, Stanford University School of Medicine, 353 Serra Mall (Gates Building), Stanford, CA 94305 USA
| | - Pranav Rajpurkar
- 3Department of Computer Science, Stanford University School of Medicine, 353 Serra Mall (Gates Building), Stanford, CA 94305 USA
| | - Timothy Amrhein
- 4Department of Radiology, Duke University Medical Center, Box 3808 Erwin Rd, Durham, NC 27710 USA
| | - Rajan Gupta
- 4Department of Radiology, Duke University Medical Center, Box 3808 Erwin Rd, Durham, NC 27710 USA
| | - Safwan Halabi
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Curtis Langlotz
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Edward Lo
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Joseph Mammarappallil
- 4Department of Radiology, Duke University Medical Center, Box 3808 Erwin Rd, Durham, NC 27710 USA
| | - A J Mariano
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Geoffrey Riley
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Jayne Seekins
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Luyao Shen
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Evan Zucker
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| | - Matthew Lungren
- 1Department of Radiology, Stanford University School of Medicine, 300 Pasteur Dr., H1307, Stanford, CA 94305 USA
| |
Collapse
|
6
|
Fries JA, Varma P, Chen VS, Xiao K, Tejeda H, Saha P, Dunnmon J, Chubb H, Maskatia S, Fiterau M, Delp S, Ashley E, Ré C, Priest JR. Weakly supervised classification of aortic valve malformations using unlabeled cardiac MRI sequences. Nat Commun 2019; 10:3111. [PMID: 31308376 PMCID: PMC6629670 DOI: 10.1038/s41467-019-11012-3] [Citation(s) in RCA: 45] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2018] [Accepted: 06/13/2019] [Indexed: 11/23/2022] Open
Abstract
Biomedical repositories such as the UK Biobank provide increasing access to prospectively collected cardiac imaging, however these data are unlabeled, which creates barriers to their use in supervised machine learning. We develop a weakly supervised deep learning model for classification of aortic valve malformations using up to 4,000 unlabeled cardiac MRI sequences. Instead of requiring highly curated training data, weak supervision relies on noisy heuristics defined by domain experts to programmatically generate large-scale, imperfect training labels. For aortic valve classification, models trained with imperfect labels substantially outperform a supervised model trained on hand-labeled MRIs. In an orthogonal validation experiment using health outcomes data, our model identifies individuals with a 1.8-fold increase in risk of a major adverse cardiac event. This work formalizes a deep learning baseline for aortic valve classification and outlines a general strategy for using weak supervision to train machine learning models using unlabeled medical images at scale.
Collapse
Affiliation(s)
- Jason A Fries
- Department of Computer Science, Stanford University, Stanford, CA, 94305, USA.
- Center for Biomedical Informatics Research, Stanford University, Palo Alto, CA, 94305, USA.
| | - Paroma Varma
- Department of Electrical Engineering, Stanford University, Palo Alto, CA, 94305, USA
| | - Vincent S Chen
- Department of Computer Science, Stanford University, Stanford, CA, 94305, USA
| | - Ke Xiao
- Department of Pediatrics, Stanford University, Stanford, CA, 94304, USA
| | - Heliodoro Tejeda
- Department of Pediatrics, Stanford University, Stanford, CA, 94304, USA
| | - Priyanka Saha
- Department of Pediatrics, Stanford University, Stanford, CA, 94304, USA
| | - Jared Dunnmon
- Department of Computer Science, Stanford University, Stanford, CA, 94305, USA
| | - Henry Chubb
- Department of Pediatrics, Stanford University, Stanford, CA, 94304, USA
| | - Shiraz Maskatia
- Department of Pediatrics, Stanford University, Stanford, CA, 94304, USA
| | - Madalina Fiterau
- Department of Computer Science, Stanford University, Stanford, CA, 94305, USA
| | - Scott Delp
- Department of Bioengineering, Stanford University, Palo Alto, CA, 94305, USA
| | - Euan Ashley
- Department of Medicine, Stanford University, Stanford, CA, 94304, USA
- Chan Zuckerberg BioHub, San Francisco, CA, 94158, USA
| | - Christopher Ré
- Department of Computer Science, Stanford University, Stanford, CA, 94305, USA
- Chan Zuckerberg BioHub, San Francisco, CA, 94158, USA
| | - James R Priest
- Department of Pediatrics, Stanford University, Stanford, CA, 94304, USA
- Chan Zuckerberg BioHub, San Francisco, CA, 94158, USA
| |
Collapse
|
7
|
Kandel P, Wallace MB. Should We Resect and Discard Low Risk Diminutive Colon Polyps. Clin Endosc 2019; 52:239-246. [PMID: 30661337 PMCID: PMC6547333 DOI: 10.5946/ce.2018.136] [Citation(s) in RCA: 22] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/23/2018] [Accepted: 09/04/2018] [Indexed: 02/06/2023] Open
Abstract
Diminutive colorectal polyps <5 mm are very common and almost universally benign. The current strategy of resection with histological confirmation of all colorectal polyps is costly and may increase the risk of colonoscopy. Accurate, optical diagnosis without histology can be achieved with currently available endoscopic technologies. The American Society of Gastrointestinal Endoscopy Preservation and Incorporation of Valuable endoscopic Innovations supports strategies for optical diagnosis of small non neoplastic polyps as long as two criteria are met. For hyperplastic appearing polyps <5 mm in recto-sigmoid colon, the negative predictive value should be at least 90%. For diminutive low grade adenomatous appearing polyps, a resect and discard strategy should be sufficiently accurate such that post-polypectomy surveillance recommendations based on the optical diagnosis, agree with a histologically diagnosis at least 90% of the time. Although the resect and discard as well as diagnose and leave behind approach has major benefits with regard to both safety and cost, it has yet to be used widely in practice. To fully implement such as strategy, there is a need for better-quality training, quality assurance, and patient acceptance. In the article, we will review the current state of the science on optical diagnose of colorectal polyps and its implications for colonoscopy practice.
Collapse
Affiliation(s)
- Pujan Kandel
- Department of Gastroenterology and Hepatology, Mayo Clinic Florida, Jacksonville, FL, USA
| | - Michael B Wallace
- Department of Gastroenterology and Hepatology, Mayo Clinic Florida, Jacksonville, FL, USA
| |
Collapse
|
8
|
Heim E, Roß T, Seitel A, März K, Stieltjes B, Eisenmann M, Lebert J, Metzger J, Sommer G, Sauter AW, Schwartz FR, Termer A, Wagner F, Kenngott HG, Maier-Hein L. Large-scale medical image annotation with crowd-powered algorithms. J Med Imaging (Bellingham) 2018; 5:034002. [PMID: 30840724 DOI: 10.1117/1.jmi.5.3.034002] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2018] [Accepted: 07/26/2018] [Indexed: 01/07/2023] Open
Abstract
Accurate segmentations in medical images are the foundations for various clinical applications. Advances in machine learning-based techniques show great potential for automatic image segmentation, but these techniques usually require a huge amount of accurately annotated reference segmentations for training. The guiding hypothesis of this paper was that crowd-algorithm collaboration could evolve as a key technique in large-scale medical data annotation. As an initial step toward this goal, we evaluated the performance of untrained individuals to detect and correct errors made by three-dimensional (3-D) medical segmentation algorithms. To this end, we developed a multistage segmentation pipeline incorporating a hybrid crowd-algorithm 3-D segmentation algorithm integrated into a medical imaging platform. In a pilot study of liver segmentation using a publicly available dataset of computed tomography scans, we show that the crowd is able to detect and refine inaccurate organ contours with a quality similar to that of experts (engineers with domain knowledge, medical students, and radiologists). Although the crowds need significantly more time for the annotation of a slice, the annotation rate is extremely high. This could render crowdsourcing a key tool for cost-effective large-scale medical image annotation.
Collapse
Affiliation(s)
- Eric Heim
- German Cancer Research Center (DKFZ), Computer Assisted Medical Interventions, Heidelberg, Germany
| | - Tobias Roß
- German Cancer Research Center (DKFZ), Computer Assisted Medical Interventions, Heidelberg, Germany
| | - Alexander Seitel
- German Cancer Research Center (DKFZ), Computer Assisted Medical Interventions, Heidelberg, Germany
| | - Keno März
- German Cancer Research Center (DKFZ), Computer Assisted Medical Interventions, Heidelberg, Germany
| | - Bram Stieltjes
- University of Basel, University Hospital Basel, Radiology and Nuclear Medicine Clinic, Basel, Switzerland
| | - Matthias Eisenmann
- German Cancer Research Center (DKFZ), Computer Assisted Medical Interventions, Heidelberg, Germany
| | - Johannes Lebert
- University of Heidelberg, Department of General, Visceral and Transplant Surgery, Heidelberg, Germany
| | - Jasmin Metzger
- German Cancer Research Center (DKFZ), Medical Image Computing, Heidelberg, Germany
| | - Gregor Sommer
- University of Basel, University Hospital Basel, Radiology and Nuclear Medicine Clinic, Basel, Switzerland
| | - Alexander W Sauter
- University of Basel, University Hospital Basel, Radiology and Nuclear Medicine Clinic, Basel, Switzerland
| | - Fides Regina Schwartz
- University of Basel, University Hospital Basel, Radiology and Nuclear Medicine Clinic, Basel, Switzerland
| | - Andreas Termer
- University of Heidelberg, Department of General, Visceral and Transplant Surgery, Heidelberg, Germany
| | - Felix Wagner
- University of Heidelberg, Department of General, Visceral and Transplant Surgery, Heidelberg, Germany
| | - Hannes Götz Kenngott
- University of Heidelberg, Department of General, Visceral and Transplant Surgery, Heidelberg, Germany
| | - Lena Maier-Hein
- German Cancer Research Center (DKFZ), Computer Assisted Medical Interventions, Heidelberg, Germany
| |
Collapse
|
9
|
Créquit P, Mansouri G, Benchoufi M, Vivot A, Ravaud P. Mapping of Crowdsourcing in Health: Systematic Review. J Med Internet Res 2018; 20:e187. [PMID: 29764795 PMCID: PMC5974463 DOI: 10.2196/jmir.9330] [Citation(s) in RCA: 70] [Impact Index Per Article: 11.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2017] [Revised: 02/10/2018] [Accepted: 03/14/2018] [Indexed: 11/22/2022] Open
Abstract
Background Crowdsourcing involves obtaining ideas, needed services, or content by soliciting Web-based contributions from a crowd. The 4 types of crowdsourced tasks (problem solving, data processing, surveillance or monitoring, and surveying) can be applied in the 3 categories of health (promotion, research, and care). Objective This study aimed to map the different applications of crowdsourcing in health to assess the fields of health that are using crowdsourcing and the crowdsourced tasks used. We also describe the logistics of crowdsourcing and the characteristics of crowd workers. Methods MEDLINE, EMBASE, and ClinicalTrials.gov were searched for available reports from inception to March 30, 2016, with no restriction on language or publication status. Results We identified 202 relevant studies that used crowdsourcing, including 9 randomized controlled trials, of which only one had posted results at ClinicalTrials.gov. Crowdsourcing was used in health promotion (91/202, 45.0%), research (73/202, 36.1%), and care (38/202, 18.8%). The 4 most frequent areas of application were public health (67/202, 33.2%), psychiatry (32/202, 15.8%), surgery (22/202, 10.9%), and oncology (14/202, 6.9%). Half of the reports (99/202, 49.0%) referred to data processing, 34.6% (70/202) referred to surveying, 10.4% (21/202) referred to surveillance or monitoring, and 5.9% (12/202) referred to problem-solving. Labor market platforms (eg, Amazon Mechanical Turk) were used in most studies (190/202, 94%). The crowd workers’ characteristics were poorly reported, and crowdsourcing logistics were missing from two-thirds of the reports. When reported, the median size of the crowd was 424 (first and third quartiles: 167-802); crowd workers’ median age was 34 years (32-36). Crowd workers were mainly recruited nationally, particularly in the United States. For many studies (58.9%, 119/202), previous experience in crowdsourcing was required, and passing a qualification test or training was seldom needed (11.9% of studies; 24/202). For half of the studies, monetary incentives were mentioned, with mainly less than US $1 to perform the task. The time needed to perform the task was mostly less than 10 min (58.9% of studies; 119/202). Data quality validation was used in 54/202 studies (26.7%), mainly by attention check questions or by replicating the task with several crowd workers. Conclusions The use of crowdsourcing, which allows access to a large pool of participants as well as saving time in data collection, lowering costs, and speeding up innovations, is increasing in health promotion, research, and care. However, the description of crowdsourcing logistics and crowd workers’ characteristics is frequently missing in study reports and needs to be precisely reported to better interpret the study findings and replicate them.
Collapse
Affiliation(s)
- Perrine Créquit
- INSERM UMR1153, Methods Team, Epidemiology and Statistics Sorbonne Paris Cité Research Center, Paris Descartes University, Paris, France.,Centre d'Epidémiologie Clinique, Hôpital Hôtel Dieu, Assistance Publique des Hôpitaux de Paris, Paris, France.,Cochrane France, Paris, France
| | - Ghizlène Mansouri
- INSERM UMR1153, Methods Team, Epidemiology and Statistics Sorbonne Paris Cité Research Center, Paris Descartes University, Paris, France
| | - Mehdi Benchoufi
- Centre d'Epidémiologie Clinique, Hôpital Hôtel Dieu, Assistance Publique des Hôpitaux de Paris, Paris, France
| | - Alexandre Vivot
- INSERM UMR1153, Methods Team, Epidemiology and Statistics Sorbonne Paris Cité Research Center, Paris Descartes University, Paris, France.,Centre d'Epidémiologie Clinique, Hôpital Hôtel Dieu, Assistance Publique des Hôpitaux de Paris, Paris, France
| | - Philippe Ravaud
- INSERM UMR1153, Methods Team, Epidemiology and Statistics Sorbonne Paris Cité Research Center, Paris Descartes University, Paris, France.,Centre d'Epidémiologie Clinique, Hôpital Hôtel Dieu, Assistance Publique des Hôpitaux de Paris, Paris, France.,Cochrane France, Paris, France.,Department of Epidemiology, Columbia University, Mailman School of Public Health, New York, NY, United States
| |
Collapse
|
10
|
Patel R, Scuba B, Soetikno R, Kaltenbach T. Colorectal polyp characterization: Is my computer better than me? Endosc Int Open 2018; 6:E279-E280. [PMID: 29511730 PMCID: PMC5836845 DOI: 10.1055/s-0043-124967] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/04/2022] Open
Affiliation(s)
- Roshan Patel
- University of California, San Francisco, California, United States
| | - Bill Scuba
- University of Utah, Salt Lake City, Utah, United States
| | - Roy Soetikno
- San Francisco Veterans Affairs Healthcare System, San Francisco, California, United States
| | - Tonya Kaltenbach
- University of California, San Francisco, California, United States,San Francisco Veterans Affairs Healthcare System, San Francisco, California, United States,Corresponding author Tonya Kaltenbach University of California, San Francisco – VA San Francisco4150 Clement StreetSan Francisco, CA 94121+1-650-963-3535
| |
Collapse
|
11
|
Lee YJ, Arida JA, Donovan HS. The application of crowdsourcing approaches to cancer research: a systematic review. Cancer Med 2017; 6:2595-2605. [PMID: 28960834 PMCID: PMC5673951 DOI: 10.1002/cam4.1165] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2017] [Revised: 06/28/2017] [Accepted: 07/25/2017] [Indexed: 12/22/2022] Open
Abstract
Crowdsourcing is "the practice of obtaining participants, services, ideas, or content by soliciting contributions from a large group of people, especially via the Internet." (Ranard et al. J. Gen. Intern. Med. 29:187, 2014) Although crowdsourcing has been adopted in healthcare research and its potential for analyzing large datasets and obtaining rapid feedback has recently been recognized, no systematic reviews of crowdsourcing in cancer research have been conducted. Therefore, we sought to identify applications of and explore potential uses for crowdsourcing in cancer research. We conducted a systematic review of articles published between January 2005 and June 2016 on crowdsourcing in cancer research, using PubMed, CINAHL, Scopus, PsychINFO, and Embase. Data from the 12 identified articles were summarized but not combined statistically. The studies addressed a range of cancers (e.g., breast, skin, gynecologic, colorectal, prostate). Eleven studies collected data on the Internet using web-based platforms; one recruited participants in a shopping mall using paper-and-pen data collection. Four studies used Amazon Mechanical Turk for recruiting and/or data collection. Study objectives comprised categorizing biopsy images (n = 6), assessing cancer knowledge (n = 3), refining a decision support system (n = 1), standardizing survivorship care-planning (n = 1), and designing a clinical trial (n = 1). Although one study demonstrated that "the wisdom of the crowd" (NCI Budget Fact Book, 2017) could not replace trained experts, five studies suggest that distributed human intelligence could approximate or support the work of trained experts. Despite limitations, crowdsourcing has the potential to improve the quality and speed of research while reducing costs. Longitudinal studies should confirm and refine these findings.
Collapse
Affiliation(s)
- Young Ji Lee
- Department of Health and Community SystemsSchool of NursingUniversity of PittsburghPittsburghPennsylvania
- Department of Biomedical InformaticsSchool of MedicineUniversity of PittsburghPittsburghPennsylvania
| | - Janet A. Arida
- Department of Health and Community SystemsSchool of NursingUniversity of PittsburghPittsburghPennsylvania
| | - Heidi S. Donovan
- Department of Health and Community SystemsSchool of NursingUniversity of PittsburghPittsburghPennsylvania
- Department of Obstetrics, Gynecology, and Reproductive SciencesSchool of MedicineUniversity of PittsburghPittsburghPennsylvania
| |
Collapse
|
12
|
Abstract
OBJECTIVE Automated analysis of abdominal CT has advanced markedly over just the last few years. Fully automated assessment of organs, lymph nodes, adipose tissue, muscle, bowel, spine, and tumors are some examples where tremendous progress has been made. Computer-aided detection of lesions has also improved dramatically. CONCLUSION This article reviews the progress and provides insights into what is in store in the near future for automated analysis for abdominal CT, ultimately leading to fully automated interpretation.
Collapse
|
13
|
Meyer AND, Longhurst CA, Singh H. Crowdsourcing Diagnosis for Patients With Undiagnosed Illnesses: An Evaluation of CrowdMed. J Med Internet Res 2016; 18:e12. [PMID: 26769236 PMCID: PMC4731679 DOI: 10.2196/jmir.4887] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2015] [Revised: 10/15/2015] [Accepted: 11/30/2015] [Indexed: 11/30/2022] Open
Abstract
Background Despite visits to multiple physicians, many patients remain undiagnosed. A new online program, CrowdMed, aims to leverage the “wisdom of the crowd” by giving patients an opportunity to submit their cases and interact with case solvers to obtain diagnostic possibilities. Objective To describe CrowdMed and provide an independent assessment of its impact. Methods Patients submit their cases online to CrowdMed and case solvers sign up to help diagnose patients. Case solvers attempt to solve patients’ diagnostic dilemmas and often have an interactive online discussion with patients, including an exchange of additional diagnostic details. At the end, patients receive detailed reports containing diagnostic suggestions to discuss with their physicians and fill out surveys about their outcomes. We independently analyzed data collected from cases between May 2013 and April 2015 to determine patient and case solver characteristics and case outcomes. Results During the study period, 397 cases were completed. These patients previously visited a median of 5 physicians, incurred a median of US $10,000 in medical expenses, spent a median of 50 hours researching their illnesses online, and had symptoms for a median of 2.6 years. During this period, 357 active case solvers participated, of which 37.9% (132/348) were male and 58.3% (208/357) worked or studied in the medical industry. About half (50.9%, 202/397) of patients were likely to recommend CrowdMed to a friend, 59.6% (233/391) reported that the process gave insights that led them closer to the correct diagnoses, 57% (52/92) reported estimated decreases in medical expenses, and 38% (29/77) reported estimated improvement in school or work productivity. Conclusions Some patients with undiagnosed illnesses reported receiving helpful guidance from crowdsourcing their diagnoses during their difficult diagnostic journeys. However, further development and use of crowdsourcing methods to facilitate diagnosis requires long-term evaluation as well as validation to account for patients’ ultimate correct diagnoses.
Collapse
Affiliation(s)
- Ashley N D Meyer
- Houston Veterans Affairs Center for Innovations in Quality, Effectiveness and Safety, Health Services Research and Development, Michael E. DeBakey Veterans Affairs Medical Center, Houston, TX, United States
| | | | | |
Collapse
|
14
|
Manjunath KN, Gopalakrishna PK, Siddalingaswamy PC. Feasibility of computed tomography colonography as a diagnostic procedure in colon cancer screening in India. Asian Pac J Cancer Prev 2014; 15:5111-6. [PMID: 25040959 DOI: 10.7314/apjcp.2014.15.13.5111] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022] Open
Abstract
Computed Tomography Colonography (CTC) is a medical imaging technology used in identifying polyps and colon cancer masses in the large intestine. The technique has evolved a great deal since its invention and has become a routine diagnostic procedure in Western countries due to its non invasiveness and ease of use. The objective of our study was to explore the possibility of CTC application in Indian hospitals. This paper gives an overview of the procedure and its commercial viability. The explanation begins with the domain aspects from gastroenterologist perspective, the new way of thinking in polyp classification, the technical components of CTC procedure, and how engineering solutions have helped clinicians in solving the complexities involved in colon diagnosis. The colon cancer statistics in India and the results of single institution study we carried out with retrospective data is explained. By considering the increasing number of patients developing colon malignancies, the practicality of CTC in Indian hospitals is discussed. This paper does not reveal any technical aspects (algorithms) of engineering solutions implemented in CTC.
Collapse
Affiliation(s)
- Kanabagatte Nanjundappa Manjunath
- Department of Biomedical Engineering, Research Scholar, Manipal Institute of Technology, Manipal University, Manipal, India E-mail :
| | | | | |
Collapse
|
15
|
Ranard BL, Ha YP, Meisel ZF, Asch DA, Hill SS, Becker LB, Seymour AK, Merchant RM. Crowdsourcing--harnessing the masses to advance health and medicine, a systematic review. J Gen Intern Med 2014; 29:187-203. [PMID: 23843021 PMCID: PMC3889976 DOI: 10.1007/s11606-013-2536-8] [Citation(s) in RCA: 189] [Impact Index Per Article: 18.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/01/2013] [Revised: 05/06/2013] [Accepted: 05/20/2013] [Indexed: 11/30/2022]
Abstract
OBJECTIVE Crowdsourcing research allows investigators to engage thousands of people to provide either data or data analysis. However, prior work has not documented the use of crowdsourcing in health and medical research. We sought to systematically review the literature to describe the scope of crowdsourcing in health research and to create a taxonomy to characterize past uses of this methodology for health and medical research. DATA SOURCES PubMed, Embase, and CINAHL through March 2013. STUDY ELIGIBILITY CRITERIA Primary peer-reviewed literature that used crowdsourcing for health research. STUDY APPRAISAL AND SYNTHESIS METHODS Two authors independently screened studies and abstracted data, including demographics of the crowd engaged and approaches to crowdsourcing. RESULTS Twenty-one health-related studies utilizing crowdsourcing met eligibility criteria. Four distinct types of crowdsourcing tasks were identified: problem solving, data processing, surveillance/monitoring, and surveying. These studies collectively engaged a crowd of >136,395 people, yet few studies reported demographics of the crowd. Only one (5 %) reported age, sex, and race statistics, and seven (33 %) reported at least one of these descriptors. Most reports included data on crowdsourcing logistics such as the length of crowdsourcing (n = 18, 86 %) and time to complete crowdsourcing task (n = 15, 71 %). All articles (n = 21, 100 %) reported employing some method for validating or improving the quality of data reported from the crowd. LIMITATIONS Gray literature not searched and only a sample of online survey articles included. CONCLUSIONS AND IMPLICATIONS OF KEY FINDINGS Utilizing crowdsourcing can improve the quality, cost, and speed of a research project while engaging large segments of the public and creating novel science. Standardized guidelines are needed on crowdsourcing metrics that should be collected and reported to provide clarity and comparability in methods.
Collapse
Affiliation(s)
- Benjamin L Ranard
- Perelman School of Medicine, University of Pennsylvania, 423 Guardian Drive, Blockley Hall, Philadelphia, PA, 19104, USA
| | | | | | | | | | | | | | | |
Collapse
|
16
|
Hill S, Merchant R, Ungar L. LESSONS LEARNED ABOUT PUBLIC HEALTH FROM ONLINE CROWD SURVEILLANCE. BIG DATA 2013; 1:160-167. [PMID: 25045598 PMCID: PMC4102381 DOI: 10.1089/big.2013.0020] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
The Internet has forever changed the way people access information and make decisions about their healthcare needs. Patients now share information about their health at unprecedented rates on social networking sites such as Twitter and Facebook and on medical discussion boards. In addition to explicitly shared information about health conditions through posts, patients reveal data on their inner fears and desires about health when searching for health-related keywords on search engines. Data are also generated by the use of mobile phone applications that track users' health behaviors (e.g., eating and exercise habits) as well as give medical advice. The data generated through these applications are mined and repackaged by surveillance systems developed by academics, companies, and governments alike to provide insight to patients and healthcare providers for medical decisions. Until recently, most Internet research in public health has been surveillance focused or monitoring health behaviors. Only recently have researchers used and interacted with the crowd to ask questions and collect health-related data. In the future, we expect to move from this surveillance focus to the "ideal" of Internet-based patient-level interventions where healthcare providers help patients change their health behaviors. In this article, we highlight the results of our prior research on crowd surveillance and make suggestions for the future.
Collapse
Affiliation(s)
- Shawndra Hill
- Operations and Information Management Department, The Wharton School, University of Pennsylvania, Philadelphia, Pennsylvania
| | - Raina Merchant
- Department of Emergency Medicine, University of Pennsylvania, Philadelphia, Pennsylvania
| | - Lyle Ungar
- Department of Computer and Information Science, University of Pennsylvania, Philadelphia, Pennsylvania
| |
Collapse
|