Back to Search Start Over

Reproducible variability: assessing investigator discordance across 9 research teams attempting to reproduce the same observational study

Authors :
Anna Ostropolets
Yasser Albogami
Mitchell Conover
Juan M Banda
William A Baumgartner
Clair Blacketer
Priyamvada Desai
Scott L DuVall
Stephen Fortin
James P Gilbert
Asieh Golozar
Joshua Ide
Andrew S Kanter
David M Kern
Chungsoo Kim
Lana Y H Lai
Chenyu Li
Feifan Liu
Kristine E Lynch
Evan Minty
Maria Inês Neves
Ding Quan Ng
Tontel Obene
Victor Pera
Nicole Pratt
Gowtham Rao
Nadav Rappoport
Ines Reinecke
Paola Saroufim
Azza Shoaibi
Katherine Simon
Marc A Suchard
Joel N Swerdel
Erica A Voss
James Weaver
Linying Zhang
George Hripcsak
Patrick B Ryan
Medical Informatics
Ostropolets, Anna
Albogami, Yasser
Conover, Mitchell
Banda, Juan M
Pratt, Nicole
Ryan, Patrick B
Source :
Journal of the American Medical Informatics Association : JAMIA, 30(5), 859-868. Oxford University Press
Publication Year :
2023

Abstract

Objective Observational studies can impact patient care but must be robust and reproducible. Nonreproducibility is primarily caused by unclear reporting of design choices and analytic procedures. This study aimed to: (1) assess how the study logic described in an observational study could be interpreted by independent researchers and (2) quantify the impact of interpretations’ variability on patient characteristics. Materials and Methods Nine teams of highly qualified researchers reproduced a cohort from a study by Albogami et al. The teams were provided the clinical codes and access to the tools to create cohort definitions such that the only variable part was their logic choices. We executed teams’ cohort definitions against the database and compared the number of subjects, patient overlap, and patient characteristics. Results On average, the teams’ interpretations fully aligned with the master implementation in 4 out of 10 inclusion criteria with at least 4 deviations per team. Cohorts’ size varied from one-third of the master cohort size to 10 times the cohort size (2159–63 619 subjects compared to 6196 subjects). Median agreement was 9.4% (interquartile range 15.3–16.2%). The teams’ cohorts significantly differed from the master implementation by at least 2 baseline characteristics, and most of the teams differed by at least 5. Conclusions Independent research teams attempting to reproduce the study based on its free-text description alone produce different implementations that vary in the population size and composition. Sharing analytical code supported by a common data model and open-source tools allows reproducing a study unambiguously thereby preserving initial design choices.

Details

Language :
English
ISSN :
10675027
Database :
OpenAIRE
Journal :
Journal of the American Medical Informatics Association : JAMIA, 30(5), 859-868. Oxford University Press
Accession number :
edsair.doi.dedup.....c2d69de85b3cb0a18d90c106f56e7d55