Effects of double data extraction on errors in evidence synthesis: a crossover, multicenter, investigator-blinded, randomized controlled trial

Author:

Zhu Yi,Yang Xi,Doi Suhail A.R.,Furuya-Kanamori Luis,Lin Lifeng,Qin Zongshi,Tao Fangbiao,Xu Chang

Abstract

AbstractObjectivesThe objective was to investigate the role of duet extraction in reducing data errors in evidence synthesis for pharmaceutical and non-pharmaceutical interventions.DesignRandomized controlled trial.SettingUniversity teaching center and hospital evidence-based medicine center.ParticipantsEligible 100 participants were 2ndyear or above post-graduate students (e.g., masters, doctoral program), who were randomly (1:1) assigned for data extraction tasks of either 10 RCTs of pharmaceutical interventions or 10 of non-pharmaceutical interventions, followed by a cross-over pattern and a further double-checking process.InterventionThe intervention of this trial was double-checking process for data extraction.Primary and secondary outcome measuresThe primary outcome was the error rates for RCTs in the pharmaceutical versus non-pharmaceutical intervention group, in terms of both study level and cell level (2 by 2 table). The secondary outcome was the absolute difference in the error rates before and after the double-checking process for both the pharmaceutical and non-pharmaceutical intervention groups, again, in terms of both study level and cell level (2 by 2 table).ResultsThe error rates in RCTs of pharmaceutical and non-pharmaceutical groups were 64.65% and 59.90%, with an absolute difference of 4.75% and an odds ratio (OR) of 1.29 (95%CI: 1.06 to 1.57,P= 0.01) when measured at the study level. After double-checking, the error rates decreased to 44.88% and 39.54%, and the difference between the two groups remained at 5.34%, with the OR of 1.27 (95%CI: 1.1 to 1.46;P< 0.01). Similar results were observed when measured at the cell level.ConclusionDouble-checking reduced data extraction errors, but the error rate still remained high after the process. Further evidence synthesis research may consider to use triple data extraction to minimize potential errors.Trial registration numberChinese Clinical Trial Registry Center (Identifier: ChiCTR2200062206)Strengths and limitations of the studyThis is the third randomized trial focusing on data extraction strategies and the first one in the Asia-Pacific region.This is the first randomized trial that compares error rates of data extraction in trials of pharmaceutical interventions and non-pharmaceutical interventions.We validated the effectiveness of dual data extraction while also identifying its limitations, providing valuable evidence for future data extraction strategies.To ensure the feasibility of the trial, we restricted subjects to 2ndpost-graduate students or above, which may affects the representativeness of the sample.The readability of the chosen randomized controlled trials for pharmaceutical and non-pharmaceutical interventions may vary, potentially resulting in selection bias that can distort the outcomes.

Publisher

Cold Spring Harbor Laboratory

Reference22 articles.

1. Progress in evidence-based medicine: a quarter century on

2. Committee on Standards for Systematic Reviews of Comparative Effectiveness Research, Board on Health Care Services. In: Eden J , Levit L , Berg A , Morton S , editors. Finding what works in health care: standards for systematic reviews. Washington, DC: National Academies Press; 2011.

3. Frequency of data extraction errors and methods to increase data extraction quality: a methodological review

4. Automatic extraction of quantitative data from ClinicalTrials.gov to conduct meta-analyses;J Clin Epidemiol,2019

5. Data reproducibility issues and their potential impact on conclusions from evidence syntheses of randomized controlled trials in sleep medicine;Sleep Med Rev,2022

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3