Reproducible variability: assessing investigator discordance across 9 research teams attempting to reproduce the same observational study. (24th February 2023)
- Record Type:
- Journal Article
- Title:
- Reproducible variability: assessing investigator discordance across 9 research teams attempting to reproduce the same observational study. (24th February 2023)
- Main Title:
- Reproducible variability: assessing investigator discordance across 9 research teams attempting to reproduce the same observational study
- Authors:
- Ostropolets, Anna
Albogami, Yasser
Conover, Mitchell
Banda, Juan M
Baumgartner, William A
Blacketer, Clair
Desai, Priyamvada
DuVall, Scott L
Fortin, Stephen
Gilbert, James P
Golozar, Asieh
Ide, Joshua
Kanter, Andrew S
Kern, David M
Kim, Chungsoo
Lai, Lana Y H
Li, Chenyu
Liu, Feifan
Lynch, Kristine E
Minty, Evan
Neves, Maria Inês
Ng, Ding Quan
Obene, Tontel
Pera, Victor
Pratt, Nicole
Rao, Gowtham
Rappoport, Nadav
Reinecke, Ines
Saroufim, Paola
Shoaibi, Azza
Simon, Katherine
Suchard, Marc A
Swerdel, Joel N
Voss, Erica A
Weaver, James
Zhang, Linying
Hripcsak, George
Ryan, Patrick B
… (more) - Abstract:
- Abstract: Objective: Observational studies can impact patient care but must be robust and reproducible. Nonreproducibility is primarily caused by unclear reporting of design choices and analytic procedures. This study aimed to: (1) assess how the study logic described in an observational study could be interpreted by independent researchers and (2) quantify the impact of interpretations' variability on patient characteristics. Materials and Methods: Nine teams of highly qualified researchers reproduced a cohort from a study by Albogami et al. The teams were provided the clinical codes and access to the tools to create cohort definitions such that the only variable part was their logic choices. We executed teams' cohort definitions against the database and compared the number of subjects, patient overlap, and patient characteristics. Results: On average, the teams' interpretations fully aligned with the master implementation in 4 out of 10 inclusion criteria with at least 4 deviations per team. Cohorts' size varied from one-third of the master cohort size to 10 times the cohort size (2159–63 619 subjects compared to 6196 subjects). Median agreement was 9.4% (interquartile range 15.3–16.2%). The teams' cohorts significantly differed from the master implementation by at least 2 baseline characteristics, and most of the teams differed by at least 5. Conclusions: Independent research teams attempting to reproduce the study based on its free-text description alone produceAbstract: Objective: Observational studies can impact patient care but must be robust and reproducible. Nonreproducibility is primarily caused by unclear reporting of design choices and analytic procedures. This study aimed to: (1) assess how the study logic described in an observational study could be interpreted by independent researchers and (2) quantify the impact of interpretations' variability on patient characteristics. Materials and Methods: Nine teams of highly qualified researchers reproduced a cohort from a study by Albogami et al. The teams were provided the clinical codes and access to the tools to create cohort definitions such that the only variable part was their logic choices. We executed teams' cohort definitions against the database and compared the number of subjects, patient overlap, and patient characteristics. Results: On average, the teams' interpretations fully aligned with the master implementation in 4 out of 10 inclusion criteria with at least 4 deviations per team. Cohorts' size varied from one-third of the master cohort size to 10 times the cohort size (2159–63 619 subjects compared to 6196 subjects). Median agreement was 9.4% (interquartile range 15.3–16.2%). The teams' cohorts significantly differed from the master implementation by at least 2 baseline characteristics, and most of the teams differed by at least 5. Conclusions: Independent research teams attempting to reproduce the study based on its free-text description alone produce different implementations that vary in the population size and composition. Sharing analytical code supported by a common data model and open-source tools allows reproducing a study unambiguously thereby preserving initial design choices. … (more)
- Is Part Of:
- Journal of the American Medical Informatics Association. Volume 30:Number 5(2023)
- Journal:
- Journal of the American Medical Informatics Association
- Issue:
- Volume 30:Number 5(2023)
- Issue Display:
- Volume 30, Issue 5 (2023)
- Year:
- 2023
- Volume:
- 30
- Issue:
- 5
- Issue Sort Value:
- 2023-0030-0005-0000
- Page Start:
- 859
- Page End:
- 868
- Publication Date:
- 2023-02-24
- Subjects:
- reproducibility -- observational data -- credibility -- open science
Medical informatics -- Periodicals
Information Services -- Periodicals
Medical Informatics -- Periodicals
Médecine -- Informatique -- Périodiques
Informatica
Geneeskunde
Informatique médicale
Computer network resources
Electronic journals
610.285 - Journal URLs:
- http://jamia.bmj.com/ ↗
http://www.jamia.org ↗
http://www.pubmedcentral.nih.gov/tocrender.fcgi?journal=76 ↗
http://www.sciencedirect.com/science/journal/10675027 ↗
http://jamia.oxfordjournals.org/ ↗
http://www.oxfordjournals.org/en/ ↗ - DOI:
- 10.1093/jamia/ocad009 ↗
- Languages:
- English
- ISSNs:
- 1067-5027
- Deposit Type:
- Legaldeposit
- View Content:
- Available online (eLD content is only available in our Reading Rooms) ↗
- Physical Locations:
- British Library DSC - 4689.025000
British Library DSC - BLDSS-3PM
British Library STI - ELD Digital store - Ingest File:
- 27082.xml