The implications of unconfounding multisource performance ratings

J Appl Psychol. 2020 Mar;105(3):312-329. doi: 10.1037/apl0000434. Epub 2019 Jul 22.

Abstract

The multifaceted structure of multisource job performance ratings has been a subject of research and debate for over 30 years. However, progress in the field has been hampered by the confounding of effects relevant to the measurement design of multisource ratings and, as a consequence, the impact of ratee-, rater-, source-, and dimension-related effects on the reliability of multisource ratings remains unclear. In separate samples obtained from 2 different applications and measurement designs (N₁ [ratees] = 392, N₁ [raters] = 1,495; N₂ [ratees] = 342, N₂ [raters] = 2,636), we, for the first time, unconfounded all systematic effects commonly cited as being relevant to multisource ratings using a Bayesian generalizability theory approach. Our results suggest that the main contributors to the reliability of multisource ratings are source-related and general performance effects that are independent of dimension-related effects. In light of our findings, we discuss the interpretation and application of multisource ratings in organizational contexts. (PsycINFO Database Record (c) 2020 APA, all rights reserved).

MeSH terms

  • Adult
  • Bayes Theorem
  • Employee Performance Appraisal / standards*
  • Humans
  • Psychometrics / standards*
  • Reproducibility of Results
  • Work Performance*