Minimal reporting improvement after peer review in reports of COVID-19 prediction models: systematic review

J Clin Epidemiol. 2023 Feb:154:75-84. doi: 10.1016/j.jclinepi.2022.12.005. Epub 2022 Dec 14.

Abstract

Objectives: To assess improvement in the completeness of reporting coronavirus (COVID-19) prediction models after the peer review process.

Study design and setting: Studies included in a living systematic review of COVID-19 prediction models, with both preprint and peer-reviewed published versions available, were assessed. The primary outcome was the change in percentage adherence to the transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD) reporting guidelines between pre-print and published manuscripts.

Results: Nineteen studies were identified including seven (37%) model development studies, two external validations of existing models (11%), and 10 (53%) papers reporting on both development and external validation of the same model. Median percentage adherence among preprint versions was 33% (min-max: 10 to 68%). The percentage adherence of TRIPOD components increased from preprint to publication in 11/19 studies (58%), with adherence unchanged in the remaining eight studies. The median change in adherence was just 3 percentage points (pp, min-max: 0-14 pp) across all studies. No association was observed between the change in percentage adherence and preprint score, journal impact factor, or time between journal submission and acceptance.

Conclusions: The preprint reporting quality of COVID-19 prediction modeling studies is poor and did not improve much after peer review, suggesting peer review had a trivial effect on the completeness of reporting during the pandemic.

Keywords: Adherence; COVID-19; Peer review; Prediction modeling; Prognosis and diagnosis; Reporting guidelines; TRIPOD.

Publication types

  • Systematic Review
  • Research Support, Non-U.S. Gov't

MeSH terms

  • COVID-19* / epidemiology
  • Humans
  • Pandemics
  • Prognosis