1
|
Wilkinson J, Heal C, Antoniou GA, Flemyng E, Avenell A, Barbour V, Bordewijk EM, Brown NJL, Clarke M, Dumville J, Grohmann S, Gurrin LC, Hayden JA, Hunter KE, Lam E, Lasserson T, Li T, Lensen S, Liu J, Lundh A, Meyerowitz-Katz G, Mol BW, O'Connell NE, Parker L, Redman B, Seidler AL, Sheldrick K, Sydenham E, Dahly DL, van Wely M, Bero L, Kirkham JJ. A survey of experts to identify methods to detect problematic studies: Stage 1 of the INSPECT-SR Project. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2024:2024.03.18.24304479. [PMID: 38585914 PMCID: PMC10996715 DOI: 10.1101/2024.03.18.24304479] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/09/2024]
Abstract
Background Randomised controlled trials (RCTs) inform healthcare decisions. Unfortunately, some published RCTs contain false data, and some appear to have been entirely fabricated. Systematic reviews are performed to identify and synthesise all RCTs which have been conducted on a given topic. This means that any of these 'problematic studies' are likely to be included, but there are no agreed methods for identifying them. The INSPECT-SR project is developing a tool to identify problematic RCTs in systematic reviews of healthcare-related interventions. The tool will guide the user through a series of 'checks' to determine a study's authenticity. The first objective in the development process is to assemble a comprehensive list of checks to consider for inclusion. Methods We assembled an initial list of checks for assessing the authenticity of research studies, with no restriction to RCTs, and categorised these into five domains: Inspecting results in the paper; Inspecting the research team; Inspecting conduct, governance, and transparency; Inspecting text and publication details; Inspecting the individual participant data. We implemented this list as an online survey, and invited people with expertise and experience of assessing potentially problematic studies to participate through professional networks and online forums. Participants were invited to provide feedback on the checks on the list, and were asked to describe any additional checks they knew of, which were not featured in the list. Results Extensive feedback on an initial list of 102 checks was provided by 71 participants based in 16 countries across five continents. Fourteen new checks were proposed across the five domains, and suggestions were made to reword checks on the initial list. An updated list of checks was constructed, comprising 116 checks. Many participants expressed a lack of familiarity with statistical checks, and emphasized the importance of feasibility of the tool. Conclusions A comprehensive list of trustworthiness checks has been produced. The checks will be evaluated to determine which should be included in the INSPECT-SR tool.
Collapse
Affiliation(s)
- Jack Wilkinson
- Centre for Biostatistics, The University of Manchester, Manchester Academic Health Science Centre, Manchester, UK
| | - Calvin Heal
- Centre for Biostatistics, The University of Manchester, Manchester Academic Health Science Centre, Manchester, UK
| | - George A Antoniou
- Manchester Vascular Centre, Manchester University NHS Foundation Trust, Manchester, UK
- Division of Cardiovascular Sciences, School of Medical Sciences, Manchester Academic Health Science Centre, The University of Manchester, Manchester, UK
| | - Ella Flemyng
- Evidence Production and Methods Directorate, Cochrane Central Executive, London, UK
| | - Alison Avenell
- Health Services Research Unit, University of Aberdeen, Aberdeen, UK
| | | | - Esmee M Bordewijk
- Centre for Reproductive Medicine, Department of Obstetrics and Gynaecology, Amsterdam University Medical Center, Netherlands
| | | | - Mike Clarke
- Northern Ireland Methodology Hub, Queen's University Belfast, UK
| | - Jo Dumville
- Division of Nursing, Midwifery & Social Work, School of Health Sciences, The University of Manchester, Manchester, UK
- NIHR Manchester Biomedical Research Centre, Manchester University NHS Foundation Trust, Manchester Academic Health Science Centre, Manchester, UK
| | - Steph Grohmann
- Evidence Production and Methods Directorate, Cochrane Central Executive, London, UK
| | - Lyle C Gurrin
- School of Population and Global Health, The University of Melbourne, Australia
| | - Jill A Hayden
- Department of Community Health & Epidemiology, Dalhousie University, Canada
| | - Kylie E Hunter
- NHMRC Clinical Trials Centre, University of Sydney, Australia
| | - Emily Lam
- Independent lay member, unaffiliated, UK
| | - Toby Lasserson
- Evidence Production and Methods Directorate, Cochrane Central Executive, London, UK
| | - Tianjing Li
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado, USA
| | - Sarah Lensen
- Department of Obstetrics, Gynaecology and Newborth Health, Royal Women's Hospital, University of Melbourne, Melbourne, Australia
| | - Jianping Liu
- Director, Centre for Evidence-Based Chinese Medicine, Beijing University of Chinese Medicine, Beijing, China
| | - Andreas Lundh
- Cochrane Denmark & Centre for Evidence-Based Medicine Odense, Department of Clinical Research, University of Southern Denmark, Denmark
- Department of Respiratory Medicine and Infectious Diseases, Copenhagen University Hospital Bispebjerg and Frederiksberg, Denmark
| | | | - Ben W Mol
- Department of Obstetrics and Gynaecology, Monash University, Melbourne, Australia
| | - Neil E O'Connell
- Department of Health Sciences, Centre for Wellbeing Across the Lifecourse, Brunel University London, UK
| | - Lisa Parker
- Charles Perkins Centre, Faculty Medicine & Health, University of Sydney, Sydney, Australia
| | | | | | - Kyle Sheldrick
- Faculty of Medicine, University of New South Wales, Australia
| | | | - Darren L Dahly
- HRB Clinical Research Facility, University College Cork, Cork, Ireland
| | - Madelon van Wely
- Centre for Reproductive Medicine, Department of Obstetrics and Gynaecology, Amsterdam University Medical Center, Netherlands
| | - Lisa Bero
- University of Colorado Anschutz Medical Campus, Colorado, USA
| | - Jamie J Kirkham
- Centre for Biostatistics, The University of Manchester, Manchester Academic Health Science Centre, Manchester, UK
| |
Collapse
|
2
|
Barnett A. Automated detection of over- and under-dispersion in baseline tables in randomised controlled trials. F1000Res 2023; 11:783. [PMID: 37360941 PMCID: PMC10285343 DOI: 10.12688/f1000research.123002.2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 05/24/2023] [Indexed: 06/28/2023] Open
Abstract
Background: Papers describing the results of a randomised trial should include a baseline table that compares the characteristics of randomised groups. Researchers who fraudulently generate trials often unwittingly create baseline tables that are implausibly similar (under-dispersed) or have large differences between groups (over-dispersed). I aimed to create an automated algorithm to screen for under- and over-dispersion in the baseline tables of randomised trials. Methods: Using a cross-sectional study I examined 2,245 randomised controlled trials published in health and medical journals on PubMed Central. I estimated the probability that a trial's baseline summary statistics were under- or over-dispersed using a Bayesian model that examined the distribution of t-statistics for the between-group differences, and compared this with an expected distribution without dispersion. I used a simulation study to test the ability of the model to find under- or over-dispersion and compared its performance with an existing test of dispersion based on a uniform test of p-values. My model combined categorical and continuous summary statistics, whereas the uniform test used only continuous statistics. Results: The algorithm had a relatively good accuracy for extracting the data from baseline tables, matching well on the size of the tables and sample size. Using t-statistics in the Bayesian model out-performed the uniform test of p-values, which had many false positives for skewed, categorical and rounded data that were not under- or over-dispersed. For trials published on PubMed Central, some tables appeared under- or over-dispersed because they had an atypical presentation or had reporting errors. Some trials flagged as under-dispersed had groups with strikingly similar summary statistics. Conclusions: Automated screening for fraud of all submitted trials is challenging due to the widely varying presentation of baseline tables. The Bayesian model could be useful in targeted checks of suspected trials or authors.
Collapse
Affiliation(s)
- Adrian Barnett
- Australian Centre for Health Services Innovation & Centre for Healthcare Transformation, Queensland University of Technology, Kelvin Grove, Queensland, 4059, Australia
| |
Collapse
|
3
|
Barnett A. Automated detection of over- and under-dispersion in baseline tables in randomised controlled trials. F1000Res 2023; 11:783. [PMID: 37360941 PMCID: PMC10285343 DOI: 10.12688/f1000research.123002.1] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 05/24/2023] [Indexed: 10/12/2023] Open
Abstract
Background: Papers describing the results of a randomised trial should include a baseline table that compares the characteristics of randomised groups. Researchers who fraudulently generate trials often unwittingly create baseline tables that are implausibly similar (under-dispersed) or have large differences between groups (over-dispersed). I aimed to create an automated algorithm to screen for under- and over-dispersion in the baseline tables of randomised trials. Methods: Using a cross-sectional study I examined 2,245 randomised controlled trials published in health and medical journals on PubMed Central. I estimated the probability that a trial's baseline summary statistics were under- or over-dispersed using a Bayesian model that examined the distribution of t-statistics for the between-group differences, and compared this with an expected distribution without dispersion. I used a simulation study to test the ability of the model to find under- or over-dispersion and compared its performance with an existing test of dispersion based on a uniform test of p-values. My model combined categorical and continuous summary statistics, whereas the uniform test used only continuous statistics. Results: The algorithm had a relatively good accuracy for extracting the data from baseline tables, matching well on the size of the tables and sample size. Using t-statistics in the Bayesian model out-performed the uniform test of p-values, which had many false positives for skewed, categorical and rounded data that were not under- or over-dispersed. For trials published on PubMed Central, some tables appeared under- or over-dispersed because they had an atypical presentation or had reporting errors. Some trials flagged as under-dispersed had groups with strikingly similar summary statistics. Conclusions: Automated screening for fraud of all submitted trials is challenging due to the widely varying presentation of baseline tables. The Bayesian model could be useful in targeted checks of suspected trials or authors.
Collapse
Affiliation(s)
- Adrian Barnett
- Australian Centre for Health Services Innovation & Centre for Healthcare Transformation, Queensland University of Technology, Kelvin Grove, Queensland, 4059, Australia
| |
Collapse
|
4
|
Bordewijk EM, Li W, van Eekelen R, Wang R, Showell M, Mol BW, van Wely M. Methods to assess research misconduct in health-related research: A scoping review. J Clin Epidemiol 2021; 136:189-202. [PMID: 34033915 DOI: 10.1016/j.jclinepi.2021.05.012] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2021] [Revised: 05/11/2021] [Accepted: 05/12/2021] [Indexed: 01/17/2023]
Abstract
OBJECTIVE To give an overview of the available methods to investigate research misconduct in health-related research. STUDY DESIGN AND SETTING In this scoping review, we conducted a literature search in MEDLINE, Embase, The Cochrane CENTRAL Register of Studies Online (CRSO), and The Virtual Health Library portal up to July 2020. We included papers that mentioned and/or described methods for screening or assessing research misconduct in health-related research. We categorized identified methods into the following four groups according to their scopes: overall concern, textual concern, image concern, and data concern. RESULTS We included 57 papers reporting on 27 methods: two on overall concern, four on textual concern, three on image concern, and 18 on data concern. Apart from the methods to locate textual plagiarism and image manipulation, all other methods, be it theoretical or empirical, are based on examples, are not standardized, and lack formal validation. CONCLUSION Existing methods cover a wide range of issues regarding research misconduct. Although measures to counteract textual plagiarism are well implemented, tools to investigate other forms of research misconduct are rudimentary and labour-intensive. To cope with the rising challenge of research misconduct, further development of automatic tools and routine validation of these methods is needed. TRIAL REGISTRATION NUMBER Center for Open Science (OSF) (https://osf.io/mq89w).
Collapse
Affiliation(s)
- Esmee M Bordewijk
- Centre for Reproductive Medicine, Amsterdam UMC, Amsterdam, The Netherlands; Department of Obstetrics and Gynecology, Monash University, Clayton, Australia
| | - Wentao Li
- Department of Obstetrics and Gynecology, Monash University, Clayton, Australia.
| | - Rik van Eekelen
- Centre for Reproductive Medicine, Amsterdam UMC, Amsterdam, The Netherlands
| | - Rui Wang
- Department of Obstetrics and Gynecology, Monash University, Clayton, Australia
| | - Marian Showell
- Department of Obstetrics and Gynaecology, University of Auckland, Auckland, New Zealand
| | - Ben W Mol
- Department of Obstetrics and Gynecology, Monash University, Clayton, Australia
| | - Madelon van Wely
- Centre for Reproductive Medicine, Amsterdam UMC, Amsterdam, The Netherlands
| |
Collapse
|
5
|
Cragg WJ, Hurley C, Yorke-Edwards V, Stenning SP. Dynamic methods for ongoing assessment of site-level risk in risk-based monitoring of clinical trials: A scoping review. Clin Trials 2021; 18:245-259. [PMID: 33611927 PMCID: PMC8010889 DOI: 10.1177/1740774520976561] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/21/2023]
Abstract
Background/Aims It is increasingly recognised that reliance on frequent site visits for monitoring clinical trials is inefficient. Regulators and trialists have recently encouraged more risk-based monitoring. Risk assessment should take place before a trial begins to define the overarching monitoring strategy. It can also be done on an ongoing basis, to target sites for monitoring activity. Various methods have been proposed for such prioritisation, often using terms like ‘central statistical monitoring’, ‘triggered monitoring’ or, as in the International Conference on Harmonization Good Clinical Practice guidance, ‘targeted on-site monitoring’. We conducted a scoping review to identify such methods, to establish if any were supported by adequate evidence to allow wider implementation, and to guide future developments in this field of research. Methods We used seven publication databases, two sets of methodological conference abstracts and an Internet search engine to identify methods for using centrally held trial data to assess site conduct during a trial. We included only reports in English, and excluded reports published before 1996 or not directly relevant to our research question. We used reference and citation searches to find additional relevant reports. We extracted data using a predefined template. We contacted authors to request additional information about included reports. Results We included 30 reports in our final dataset, of which 21 were peer-reviewed publications. In all, 20 reports described central statistical monitoring methods (of which 7 focussed on detection of fraud or misconduct) and 9 described triggered monitoring methods; 21 reports included some assessment of their methods’ effectiveness, typically exploring the methods’ characteristics using real trial data without known integrity issues. Of the 21 with some effectiveness assessment, most contained limited information about whether or not concerns identified through central monitoring constituted meaningful problems. Several reports demonstrated good classification ability based on more than one classification statistic, but never without caveats of unclear reporting or other classification statistics being low or unavailable. Some reports commented on cost savings from reduced on-site monitoring, but none gave detailed costings for the development and maintenance of central monitoring methods themselves. Conclusion Our review identified various proposed methods, some of which could be combined within the same trial. The apparent emphasis on fraud detection may not be proportionate in all trial settings. Despite some promising evidence and some self-justifying benefits for data cleaning activity, many proposed methods have limitations that may currently prevent their routine use for targeting trial monitoring activity. The implementation costs, or uncertainty about these, may also be a barrier. We make recommendations for how the evidence-base supporting these methods could be improved.
Collapse
Affiliation(s)
- William J Cragg
- MRC Clinical Trials Unit at UCL, London, UK.,Clinical Trials Research Unit, Leeds Institute of Clinical Trials Research, University of Leeds, Leeds, UK
| | - Caroline Hurley
- Health Research Board-Trials Methodology Research Network (HRB-TMRN), National University of Ireland Galway, Galway, Ireland
| | | | | |
Collapse
|