1
|
Briney KA. Measuring data rot: An analysis of the continued availability of shared data from a Single University. PLoS One 2024; 19:e0304781. [PMID: 38838010 DOI: 10.1371/journal.pone.0304781] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Accepted: 05/17/2024] [Indexed: 06/07/2024] Open
Abstract
To determine where data is shared and what data is no longer available, this study analyzed data shared by researchers at a single university. 2166 supplemental data links were harvested from the university's institutional repository and web scraped using R. All links that failed to scrape or could not be tested algorithmically were tested for availability by hand. Trends in data availability by link type, age of publication, and data source were examined for patterns. Results show that researchers shared data in hundreds of places. About two-thirds of links to shared data were in the form of URLs and one-third were DOIs, with several FTP links and links directly to files. A surprising 13.4% of shared URL links pointed to a website homepage rather than a specific record on a website. After testing, 5.4% the 2166 supplemental data links were found to be no longer available. DOIs were the type of shared link that was least likely to disappear with a 1.7% loss, with URL loss at 5.9% averaged over time. Links from older publications were more likely to be unavailable, with a data disappearance rate estimated at 2.6% per year, as well as links to data hosted on journal websites. The results support best practice guidance to share data in a data repository using a permanent identifier.
Collapse
Affiliation(s)
- Kristin A Briney
- Caltech Library, California Institute of Technology, Pasadena, CA, United States of America
| |
Collapse
|
2
|
Neilson CJ, Premji Z. A study of search strategy availability statements and sharing practices for systematic reviews: Ask and you might receive. Res Synth Methods 2024; 15:441-449. [PMID: 38098285 DOI: 10.1002/jrsm.1696] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Revised: 11/27/2023] [Accepted: 12/06/2023] [Indexed: 04/26/2024]
Abstract
The literature search underpins data collection for all systematic reviews (SRs). The SR reporting guideline PRISMA, and its extensions, aim to facilitate research transparency and reproducibility, and ultimately improve the quality of research, by instructing authors to provide specific research materials and data upon publication of the manuscript. Search strategies are one item of data that are explicitly included in PRISMA and the critical appraisal tool AMSTAR2. Yet some authors use search availability statements implying that the search strategies are available upon request instead of providing strategies up front. We sought out reviews with search availability statements, characterized them, and requested the search strategies from authors via email. Over half of the included reviews cited PRISMA but less than a third included any search strategies. After requesting the strategies via email as instructed, we received replies from 46% of authors, and eventually received at least one search strategy from 36% of authors. Requesting search strategies via email has a low chance of success. Ask and you might receive-but you probably will not. SRs that do not make search strategies available are low quality at best according to AMSTAR2; Journal editors can and should enforce the requirement for authors to include their search strategies alongside their SR manuscripts.
Collapse
Affiliation(s)
| | - Zahra Premji
- University of Victoria, Libraries, Victoria, British Columbia, Canada
| |
Collapse
|
3
|
Law M, Couturier DL, Choodari-Oskooei B, Crout P, Gamble C, Jacko P, Pallmann P, Pilling M, Robertson DS, Robling M, Sydes MR, Villar SS, Wason J, Wheeler G, Williamson SF, Yap C, Jaki T. Medicines and Healthcare products Regulatory Agency's "Consultation on proposals for legislative changes for clinical trials": a response from the Trials Methodology Research Partnership Adaptive Designs Working Group, with a focus on data sharing. Trials 2023; 24:640. [PMID: 37798805 PMCID: PMC10552399 DOI: 10.1186/s13063-023-07576-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2022] [Accepted: 08/04/2023] [Indexed: 10/07/2023] Open
Abstract
In the UK, the Medicines and Healthcare products Regulatory Agency consulted on proposals "to improve and strengthen the UK clinical trials legislation to help us make the UK the best place to research and develop safe and innovative medicines". The purpose of the consultation was to help finalise the proposals and contribute to the drafting of secondary legislation. We discussed these proposals as members of the Trials Methodology Research Partnership Adaptive Designs Working Group, which is jointly funded by the Medical Research Council and the National Institute for Health and Care Research. Two topics arose frequently in the discussion: the emphasis on legislation, and the absence of questions on data sharing. It is our opinion that the proposals rely heavily on legislation to change practice. However, clinical trials are heterogeneous, and as a result some trials will struggle to comply with all of the proposed legislation. Furthermore, adaptive design clinical trials are even more heterogeneous than their non-adaptive counterparts, and face more challenges. Consequently, it is possible that increased legislation could have a greater negative impact on adaptive designs than non-adaptive designs. Overall, we are sceptical that the introduction of legislation will achieve the desired outcomes, with some exceptions. Meanwhile the topic of data sharing - making anonymised individual-level clinical trial data available to other investigators for further use - is entirely absent from the proposals and the consultation in general. However, as an aspect of the wider concept of open science and reproducible research, data sharing is an increasingly important aspect of clinical trials. The benefits of data sharing include faster innovation, improved surveillance of drug safety and effectiveness and decreasing participant exposure to unnecessary risk. There are already a number of UK-focused documents that discuss and encourage data sharing, for example, the Concordat on Open Research Data and the Medical Research Council's Data Sharing Policy. We strongly suggest that data sharing should be the norm rather than the exception, and hope that the forthcoming proposals on clinical trials invite discussion on this important topic.
Collapse
Affiliation(s)
- Martin Law
- Medical Research Council Biostatistics Unit, School of Clinical Medicine, University of Cambridge, Cambridge, UK.
- Royal Papworth Hospital NHS Foundation Trust, Cambridge, UK.
| | - Dominique-Laurent Couturier
- Medical Research Council Biostatistics Unit, School of Clinical Medicine, University of Cambridge, Cambridge, UK
- Cancer Research UK Cambridge Institute, University of Cambridge, Cambridge, UK
| | | | - Phillip Crout
- Medical Research Council Biostatistics Unit, School of Clinical Medicine, University of Cambridge, Cambridge, UK
| | - Carrol Gamble
- Liverpool Clinical Trials Centre, University of Liverpool, Liverpool, UK
| | - Peter Jacko
- Lancaster University Management School, Lancaster University, Lancaster, UK
- Berry Consultants, Abingdon, UK
| | | | - Mark Pilling
- Department of Public Health and Primary Care, School of Clinical Medicine, University of Cambridge, Cambridge, UK
| | - David S Robertson
- Medical Research Council Biostatistics Unit, School of Clinical Medicine, University of Cambridge, Cambridge, UK
| | | | - Matthew R Sydes
- University College London, London, UK
- British Heart Foundation Data Science Centre, Health Data Research UK, London, UK
| | - Sofía S Villar
- Medical Research Council Biostatistics Unit, School of Clinical Medicine, University of Cambridge, Cambridge, UK
| | - James Wason
- Biostatistics Research Group, Population Health Sciences Institute, Newcastle University, Newcastle upon Tyne, UK
| | - Graham Wheeler
- Imperial Clinical Trials Unit, Imperial College London, London, W12 7RH, UK
| | - S Faye Williamson
- Biostatistics Research Group, Population Health Sciences Institute, Newcastle University, Newcastle upon Tyne, UK
| | - Christina Yap
- Clinical Trials and Statistics Unit, The Institute of Cancer Research, London, UK
| | - Thomas Jaki
- Medical Research Council Biostatistics Unit, School of Clinical Medicine, University of Cambridge, Cambridge, UK
- Faculty for Informatics and Data Science, University of Regensburg, Regensburg, Germany
| |
Collapse
|
4
|
Blatch-Jones AJ, Recio Saucedo A, Giddins B. The use and acceptability of preprints in health and social care settings: A scoping review. PLoS One 2023; 18:e0291627. [PMID: 37713422 PMCID: PMC10503772 DOI: 10.1371/journal.pone.0291627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Accepted: 09/04/2023] [Indexed: 09/17/2023] Open
Abstract
BACKGROUND Preprints are open and accessible scientific manuscript or report that is shared publicly, through a preprint server, before being submitted to a journal. The value and importance of preprints has grown since its contribution during the public health emergency of the COVID-19 pandemic. Funders and publishers are establishing their position on the use of preprints, in grant applications and publishing models. However, the evidence supporting the use and acceptability of preprints varies across funders, publishers, and researchers. The scoping review explored the current evidence on the use and acceptability of preprints in health and social care settings by publishers, funders, and the research community throughout the research lifecycle. METHODS A scoping review was undertaken with no study or language limits. The search strategy was limited to the last five years (2017-2022) to capture changes influenced by COVID-19 (e.g., accelerated use and role of preprints in research). The review included international literature, including grey literature, and two databases were searched: Scopus and Web of Science (24 August 2022). RESULTS 379 titles and abstracts and 193 full text articles were assessed for eligibility. Ninety-eight articles met eligibility criteria and were included for full extraction. For barriers and challenges, 26 statements were grouped under four main themes (e.g., volume/growth of publications, quality assurance/trustworthiness, risks associated to credibility, and validation). For benefits and value, 34 statements were grouped under six themes (e.g., openness/transparency, increased visibility/credibility, open review process, open research, democratic process/systems, increased productivity/opportunities). CONCLUSIONS Preprints provide opportunities for rapid dissemination but there is a need for clear policies and guidance from journals, publishers, and funders. Cautionary measures are needed to maintain the quality and value of preprints, paying particular attention to how findings are translated to the public. More research is needed to address some of the uncertainties addressed in this review.
Collapse
Affiliation(s)
- Amanda Jane Blatch-Jones
- National Institute for Health and Care Research (NIHR) Coordinating Centre, School of Healthcare Enterprise and Innovation, University of Southampton, Southampton, Hampshire, United Kingdom
| | - Alejandra Recio Saucedo
- National Institute for Health and Care Research (NIHR) Coordinating Centre, School of Healthcare Enterprise and Innovation, University of Southampton, Southampton, Hampshire, United Kingdom
| | - Beth Giddins
- National Institute for Health and Care Research (NIHR) Coordinating Centre, School of Healthcare Enterprise and Innovation, University of Southampton, Southampton, Hampshire, United Kingdom
| |
Collapse
|
5
|
Schmidt L, Sinyor M, Webb RT, Marshall C, Knipe D, Eyles EC, John A, Gunnell D, Higgins JPT. A narrative review of recent tools and innovations toward automating living systematic reviews and evidence syntheses. ZEITSCHRIFT FUR EVIDENZ, FORTBILDUNG UND QUALITAT IM GESUNDHEITSWESEN 2023; 181:65-75. [PMID: 37596160 DOI: 10.1016/j.zefq.2023.06.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/14/2022] [Revised: 06/19/2023] [Accepted: 06/25/2023] [Indexed: 08/20/2023]
Abstract
Living reviews are an increasingly popular research paradigm. The purpose of a 'living' approach is to allow rapid collation, appraisal and synthesis of evolving evidence on an important research topic, enabling timely influence on patient care and public health policy. However, living reviews are time- and resource-intensive. The accumulation of new evidence and the possibility of developments within the review's research topic can introduce unique challenges into the living review workflow. To investigate the potential of software tools to support living systematic or rapid reviews, we present a narrative review informed by an examination of tools contained on the Systematic Review Toolbox website. We identified 11 tools with relevant functionalities and discuss the important features of these tools with respect to different steps of the living review workflow. Four tools (NestedKnowledge, SWIFT-ActiveScreener, DistillerSR, EPPI-Reviewer) covered multiple, successive steps of the review process, and the remaining tools addressed specific components of the workflow, including scoping and protocol formulation, reference retrieval, automated data extraction, write-up and dissemination of data. We identify several ways in which living reviews can be made more efficient and practical. Most of these focus on general workflow management, or automation through artificial intelligence and machine-learning, in the screening process. More sophisticated uses of automation mostly target living rapid reviews to increase the speed of production or evidence maps to broaden the scope of the map. We use a case study to highlight some of the barriers and challenges to incorporating tools into the living review workflow and processes. These include increased workload, the need for organisation, ensuring timely dissemination and challenges related to the development of bespoke automation tools to facilitate the review process. We describe how current end-user tools address these challenges, and which knowledge gaps remain that could be addressed by future tool development. Dedicated web presences for automatic dissemination of in-progress evidence updates, rather than solely relying on peer-reviewed journal publications, help to make the effort of a living evidence synthesis worthwhile. Despite offering basic living review functionalities, existing end-user tools could be further developed to be interoperable with other tools to support multiple workflow steps seamlessly, to address broader automatic evidence retrieval from a larger variety of sources, and to improve dissemination of evidence between review updates.
Collapse
Affiliation(s)
- Lena Schmidt
- National Institute for Health and Care Research Innovation Observatory, Population Health Sciences Institute, Newcastle University, Newcastle, UK; Sciome LLC, Research Triangle Park, North Carolina, USA.
| | - Mark Sinyor
- Department of Psychiatry, Sunnybrook Health Sciences Centre, Toronto, Canada; Department of Psychiatry, University of Toronto, Toronto, Canada
| | - Roger T Webb
- Division of Psychology and Mental Health, The University of Manchester, Manchester, UK; National Institute for Health and Care Research Greater Manchester Patient Safety Translational Research Centre (NIHR GM PSTRC), Manchester, UK
| | | | - Duleeka Knipe
- Population Health Sciences, Bristol Medical School, University of Bristol, Bristol, UK
| | - Emily C Eyles
- Population Health Sciences, Bristol Medical School, University of Bristol, Bristol, UK; The National Institute of Health and Care Research Applied Research Collaboration West (NIHR ARC West), University Hospitals Bristol NHS Foundation Trust, Bristol, UK
| | - Ann John
- Population Data Science, Swansea University, Swansea, UK; Public Health Wales NHS Trust, Wales, UK
| | - David Gunnell
- Population Health Sciences, Bristol Medical School, University of Bristol, Bristol, UK; The National Institute of Health and Care Research Biomedical Research Centre, University Hospitals Bristol NHS Foundation Trust and the University of Bristol, Bristol, UK
| | - Julian P T Higgins
- Population Health Sciences, Bristol Medical School, University of Bristol, Bristol, UK; The National Institute of Health and Care Research Applied Research Collaboration West (NIHR ARC West), University Hospitals Bristol NHS Foundation Trust, Bristol, UK; The National Institute of Health and Care Research Biomedical Research Centre, University Hospitals Bristol NHS Foundation Trust and the University of Bristol, Bristol, UK
| |
Collapse
|
6
|
Hamilton DG, Hong K, Fraser H, Rowhani-Farid A, Fidler F, Page MJ. Prevalence and predictors of data and code sharing in the medical and health sciences: systematic review with meta-analysis of individual participant data. BMJ 2023; 382:e075767. [PMID: 37433624 PMCID: PMC10334349 DOI: 10.1136/bmj-2023-075767] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 06/07/2023] [Indexed: 07/13/2023]
Abstract
OBJECTIVES To synthesise research investigating data and code sharing in medicine and health to establish an accurate representation of the prevalence of sharing, how this frequency has changed over time, and what factors influence availability. DESIGN Systematic review with meta-analysis of individual participant data. DATA SOURCES Ovid Medline, Ovid Embase, and the preprint servers medRxiv, bioRxiv, and MetaArXiv were searched from inception to 1 July 2021. Forward citation searches were also performed on 30 August 2022. REVIEW METHODS Meta-research studies that investigated data or code sharing across a sample of scientific articles presenting original medical and health research were identified. Two authors screened records, assessed the risk of bias, and extracted summary data from study reports when individual participant data could not be retrieved. Key outcomes of interest were the prevalence of statements that declared that data or code were publicly or privately available (declared availability) and the success rates of retrieving these products (actual availability). The associations between data and code availability and several factors (eg, journal policy, type of data, trial design, and human participants) were also examined. A two stage approach to meta-analysis of individual participant data was performed, with proportions and risk ratios pooled with the Hartung-Knapp-Sidik-Jonkman method for random effects meta-analysis. RESULTS The review included 105 meta-research studies examining 2 121 580 articles across 31 specialties. Eligible studies examined a median of 195 primary articles (interquartile range 113-475), with a median publication year of 2015 (interquartile range 2012-2018). Only eight studies (8%) were classified as having a low risk of bias. Meta-analyses showed a prevalence of declared and actual public data availability of 8% (95% confidence interval 5% to 11%) and 2% (1% to 3%), respectively, between 2016 and 2021. For public code sharing, both the prevalence of declared and actual availability were estimated to be <0.5% since 2016. Meta-regressions indicated that only declared public data sharing prevalence estimates have increased over time. Compliance with mandatory data sharing policies ranged from 0% to 100% across journals and varied by type of data. In contrast, success in privately obtaining data and code from authors historically ranged between 0% and 37% and 0% and 23%, respectively. CONCLUSIONS The review found that public code sharing was persistently low across medical research. Declarations of data sharing were also low, increasing over time, but did not always correspond to actual sharing of data. The effectiveness of mandatory data sharing policies varied substantially by journal and type of data, a finding that might be informative for policy makers when designing policies and allocating resources to audit compliance. SYSTEMATIC REVIEW REGISTRATION Open Science Framework doi:10.17605/OSF.IO/7SX8U.
Collapse
Affiliation(s)
- Daniel G Hamilton
- MetaMelb Research Group, School of BioSciences, University of Melbourne, Melbourne, VIC, Australia
- Melbourne Medical School, Faculty of Medicine, Dentistry, and Health Sciences, University of Melbourne, Melbourne, VIC, Australia
| | - Kyungwan Hong
- Department of Practice, Sciences, and Health Outcomes Research, University of Maryland School of Pharmacy, Baltimore, MD, USA
| | - Hannah Fraser
- MetaMelb Research Group, School of BioSciences, University of Melbourne, Melbourne, VIC, Australia
| | - Anisa Rowhani-Farid
- Department of Practice, Sciences, and Health Outcomes Research, University of Maryland School of Pharmacy, Baltimore, MD, USA
| | - Fiona Fidler
- MetaMelb Research Group, School of BioSciences, University of Melbourne, Melbourne, VIC, Australia
- School of Historical and Philosophical Studies, University of Melbourne, Melbourne, VIC, Australia
| | - Matthew J Page
- Methods in Evidence Synthesis Unit, School of Public Health and Preventive Medicine, Monash University, Melbourne, VIC, Australia
| |
Collapse
|
7
|
Héroux ME, Butler AA, Cashin AG, McCaughey EJ, Affleck AJ, Green MA, Cartwright A, Jones M, Kiely KM, van Schooten KS, Menant JC, Wewege M, Gandevia SC. Quality Output Checklist and Content Assessment (QuOCCA): a new tool for assessing research quality and reproducibility. BMJ Open 2022; 12:e060976. [PMID: 36167369 PMCID: PMC9516158 DOI: 10.1136/bmjopen-2022-060976] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/04/2022] Open
Abstract
Research must be well designed, properly conducted and clearly and transparently reported. Our independent medical research institute wanted a simple, generic tool to assess the quality of the research conducted by its researchers, with the goal of identifying areas that could be improved through targeted educational activities. Unfortunately, none was available, thus we devised our own. Here, we report development of the Quality Output Checklist and Content Assessment (QuOCCA), and its application to publications from our institute's scientists. Following consensus meetings and external review by statistical and methodological experts, 11 items were selected for the final version of the QuOCCA: research transparency (items 1-3), research design and analysis (items 4-6) and research reporting practices (items 7-11). Five pairs of raters assessed all 231 articles published in 2017 and 221 in 2018 by researchers at our institute. Overall, the results were similar between years and revealed limited engagement with several recommended practices highlighted in the QuOCCA. These results will be useful to guide educational initiatives and their effectiveness. The QuOCCA is brief and focuses on broadly applicable and relevant concepts to open, high-quality, reproducible and well-reported science. Thus, the QuOCCA could be used by other biomedical institutions and individual researchers to evaluate research publications, assess changes in research practice over time and guide the discussion about high-quality, open science. Given its generic nature, the QuOCCA may also be useful in other research disciplines.
Collapse
Affiliation(s)
- Martin E Héroux
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | - Annie A Butler
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | - Aidan G Cashin
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | - Euan J McCaughey
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- Queen Elizabeth National Spinal Injuries Unit, Queen Elizabeth University Hospital Campus, Glasgow, UK
| | - Andrew J Affleck
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- Department of Neuropathology, Royal Prince Alfred Hospital, Camperdown, New South Wales, Australia
| | - Michael A Green
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | | | - Matthew Jones
- University of New South Wales, Sydney, New South Wales, Australia
| | - Kim M Kiely
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | - Kimberley S van Schooten
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | - Jasmine C Menant
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | - Michael Wewege
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| | - Simon C Gandevia
- Neuroscience Research Australia, Sydney, New South Wales, Australia
- University of New South Wales, Sydney, New South Wales, Australia
| |
Collapse
|
8
|
DeVito NJ, Morton C, Cashin AG, Richards GC, Lee H. Sharing study materials in health and medical research. BMJ Evid Based Med 2022:bmjebm-2022-111987. [PMID: 36162960 DOI: 10.1136/bmjebm-2022-111987] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 09/03/2022] [Indexed: 11/04/2022]
Abstract
Making study materials available allows for a more comprehensive understanding of the scientific literature. Sharing can take many forms and include a wide variety of outputs including code and data. Biomedical research can benefit from increased transparency but faces unique challenges for sharing, for instance, confidentiality concerns around participants' medical data. Both general and specialised repositories exist to aid in sharing most study materials. Sharing may also require skills and resources to ensure that it is done safely and effectively. Educating researchers on how to best share their materials, and properly rewarding these practices, requires action from a variety of stakeholders including journals, funders and research institutions.
Collapse
Affiliation(s)
- Nicholas J DeVito
- Bennett Institute for Applied Data Science, Nuffield Department of Primary Care Health Sciences, University of Oxford, Oxford, Oxfordshire, UK
| | - Caroline Morton
- Bennett Institute for Applied Data Science, Nuffield Department of Primary Care Health Sciences, University of Oxford, Oxford, Oxfordshire, UK
| | - Aidan Gregory Cashin
- School of Health Sciences, University of New South Wales, Sydney, New South Wales, Australia
- Centre for Pain IMPACT, Neuroscience Research Australia, Randwick, New South Wales, Australia
| | - Georgia C Richards
- Centre for Evidence Based Medicine, Nuffield Department of Primary Care Health Sciences, University of Oxford, Oxford, Oxfordshire, UK
| | - Hopin Lee
- Centre for Statistics in Medicine & Rehabilitation Research in Oxford, Nuffield Department of Orthopaedics, Rheumatology and Musculoskeletal Sciences (NDORMS), University of Oxford, Oxford, Oxfordshire, UK
- School of Medicine and Public Health, The University of Newcastle, Callaghan, New South Wales, Australia
| |
Collapse
|
9
|
Anger M, Wendelborn C, Winkler EC, Schickhardt C. Neither carrots nor sticks? Challenges surrounding data sharing from the perspective of research funding agencies-A qualitative expert interview study. PLoS One 2022; 17:e0273259. [PMID: 36070283 PMCID: PMC9451069 DOI: 10.1371/journal.pone.0273259] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2022] [Accepted: 08/04/2022] [Indexed: 11/19/2022] Open
Abstract
BACKGROUND Data Sharing is widely recognised as crucial for accelerating scientific research and improving its quality. However, data sharing is still not a common practice. Funding agencies tend to facilitate the sharing of research data by both providing incentives and requiring data sharing as part of their policies and conditions for awarding grants. The goal of our article is to answer the following question: What challenges do international funding agencies see when it comes to their own efforts to foster and implement data sharing through their policies? METHODS We conducted a series of sixteen guideline-based expert interviews with representatives of leading international funding agencies. As contact persons for open science at their respective agencies, they offered their perspectives and experiences concerning their organisations' data sharing policies. We performed a qualitative content analysis of the interviews and categorised the challenges perceived by funding agencies. RESULTS We identify and illustrate six challenges surrounding data sharing policies as perceived by leading funding agencies: The design of clear policies, monitoring of compliance, sanctions for non-compliance, incentives, support, and limitations for funders' own capabilities. However, our interviews also show how funders approach potential solutions to overcome these challenges, for example by coordinating with other agencies or adjusting grant evaluation metrics to incentivise data sharing. DISCUSSION AND CONCLUSION Our interviews point to existing flaws in funders' data sharing policies, such as a lack of clarity, a lack of monitoring of funded researchers' data sharing behaviour, and a lack of incentives. A number of agencies could suggest potential solutions but often struggle with the overall complexity of data sharing and the implementation of these measures. Funders cannot solve each challenge by themselves, but they can play an active role and lead joint efforts towards a culture of data sharing.
Collapse
Affiliation(s)
- Michael Anger
- Section for Translational Medical Ethics, National Center for Tumor Diseases (NCT) Heidelberg, German Cancer Research Center (DKFZ), Heidelberg, Germany
| | - Christian Wendelborn
- Section for Translational Medical Ethics, National Center for Tumor Diseases (NCT) Heidelberg, German Cancer Research Center (DKFZ), Heidelberg, Germany
| | - Eva C. Winkler
- Section for Translational Medical Ethics/Department of Medical Oncology, National Center for Tumor Diseases (NCT) Heidelberg, University Hospital Heidelberg, Heidelberg, Germany
| | - Christoph Schickhardt
- Section for Translational Medical Ethics, National Center for Tumor Diseases (NCT) Heidelberg, German Cancer Research Center (DKFZ), Heidelberg, Germany
| |
Collapse
|
10
|
Gehanno JF, Grosjean J, Darmoni SJ, Rollin L. Reliability of citations of medRxiv preprints in articles published on COVID-19 in the world leading medical journals. PLoS One 2022; 17:e0264661. [PMID: 35947594 PMCID: PMC9365132 DOI: 10.1371/journal.pone.0264661] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2022] [Accepted: 07/25/2022] [Indexed: 11/18/2022] Open
Abstract
Introduction Preprints have been widely cited during the COVID-19 pandemics, even in the major medical journals. However, since subsequent publication of preprint is not always mentioned in preprint repositories, some may be inappropriately cited or quoted. Our objectives were to assess the reliability of preprint citations in articles on COVID-19, to the rate of publication of preprints cited in these articles and to compare, if relevant, the content of the preprints to their published version. Methods Articles published on COVID in 2020 in the BMJ, The Lancet, the JAMA and the NEJM were manually screened to identify all articles citing at least one preprint from medRxiv. We searched PubMed, Google and Google Scholar to assess if the preprint had been published in a peer-reviewed journal, and when. Published articles were screened to assess if the title, data or conclusions were identical to the preprint version. Results Among the 205 research articles on COVID published by the four major medical journals in 2020, 60 (29.3%) cited at least one medRxiv preprint. Among the 182 preprints cited, 124 were published in a peer-reviewed journal, with 51 (41.1%) before the citing article was published online and 73 (58.9%) later. There were differences in the title, the data or the conclusion between the preprint cited and the published version for nearly half of them. MedRxiv did not mentioned the publication for 53 (42.7%) of preprints. Conclusions More than a quarter of preprints citations were inappropriate since preprints were in fact already published at the time of publication of the citing article, often with a different content. Authors and editors should check the accuracy of the citations and of the quotations of preprints before publishing manuscripts that cite them.
Collapse
Affiliation(s)
- Jean-Francois Gehanno
- Department of Occupational Medicine, Rouen University Hospital, Rouen, France
- Inserm, Rouen University, Sorbonne University, University of Paris 13, Laboratory of Medical Informatics and Knowledge Engineering in e-Health, LIMICS, Paris, France
- * E-mail:
| | - Julien Grosjean
- Inserm, Rouen University, Sorbonne University, University of Paris 13, Laboratory of Medical Informatics and Knowledge Engineering in e-Health, LIMICS, Paris, France
- Department of Biomedical Informatics, Rouen University Hospital, Rouen France
| | - Stefan J. Darmoni
- Inserm, Rouen University, Sorbonne University, University of Paris 13, Laboratory of Medical Informatics and Knowledge Engineering in e-Health, LIMICS, Paris, France
- Department of Biomedical Informatics, Rouen University Hospital, Rouen France
| | - Laetitia Rollin
- Department of Occupational Medicine, Rouen University Hospital, Rouen, France
- Inserm, Rouen University, Sorbonne University, University of Paris 13, Laboratory of Medical Informatics and Knowledge Engineering in e-Health, LIMICS, Paris, France
| |
Collapse
|
11
|
Collins A, Alexander R. Reproducibility of COVID-19 pre-prints. Scientometrics 2022; 127:4655-4673. [PMID: 35813409 PMCID: PMC9252536 DOI: 10.1007/s11192-022-04418-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2021] [Accepted: 05/20/2022] [Indexed: 01/24/2023]
Abstract
To examine the reproducibility of COVID-19 research, we create a dataset of pre-prints posted to arXiv, bioRxiv, and medRxiv between 28 January 2020 and 30 June 2021 that are related to COVID-19. We extract the text from these pre-prints and parse them looking for keyword markers signaling the availability of the data and code underpinning the pre-print. For the pre-prints that are in our sample, we are unable to find markers of either open data or open code for 75% of those on arXiv, 67% of those on bioRxiv, and 79% of those on medRxiv.
Collapse
|
12
|
Jiao C, Li K, Fang Z. Data sharing practices across knowledge domains: A dynamic examination of data availability statements in PLOS ONE publications. J Inf Sci 2022. [DOI: 10.1177/01655515221101830] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
As the importance of research data gradually grows in sciences, data sharing has come to be encouraged and even mandated by journals and funders in recent years. Following this trend, the data availability statement has been increasingly embraced by academic communities as a means of sharing research data as part of research articles. This article presents a quantitative study of which mechanisms and repositories are used to share research data in PLOS ONE articles. We offer a dynamic examination of this topic from the disciplinary and temporal perspectives based on all statements in English-language research articles published between 2014 and 2020 in the journal. We find a slow yet steady growth in the use of data repositories to share data over time, as opposed to sharing data in the article and/or supplementary materials; this indicates improved compliance with the journal’s data sharing policies. We also find that multidisciplinary data repositories have been increasingly used over time, whereas some disciplinary repositories show a decreasing trend. Our findings can help academic publishers and funders to improve their data sharing policies and serve as an important baseline dataset for future studies on data sharing activities.
Collapse
Affiliation(s)
- Chenyue Jiao
- School of Information Sciences, University of Illinois Urbana-Champaign, USA
| | - Kai Li
- School of Information Resource Management, Renmin University of China, China
| | - Zhichao Fang
- Centre for Science and Technology Studies, Leiden University, The Netherlands
| |
Collapse
|
13
|
Data and code availability statements in systematic reviews of interventions were often missing or inaccurate: a content analysis. J Clin Epidemiol 2022; 147:1-10. [DOI: 10.1016/j.jclinepi.2022.03.003] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2021] [Revised: 02/22/2022] [Accepted: 03/03/2022] [Indexed: 11/30/2022]
|