Grade-Related Differential Item Functioning in General English Proficiency Test-Kids Listening

Front Psychol. 2021 Nov 25:12:767244. doi: 10.3389/fpsyg.2021.767244. eCollection 2021.

Abstract

Differential Item Functioning (DIF) analysis is always an indispensable methodology for detecting item and test bias in the arena of language testing. This study investigated grade-related DIF in the General English Proficiency Test-Kids (GEPT-Kids) listening section. Quantitative data were test scores collected from 791 test takers (Grade 5 = 398; Grade 6 = 393) from eight Chinese-speaking cities, and qualitative data were expert judgments collected from two primary school English teachers in Guangdong province. Two R packages "difR" and "difNLR" were used to perform five types of DIF analysis (two-parameter item response theory [2PL IRT] based Lord's chi-square and Raju's area tests, Mantel-Haenszel [MH], logistic regression [LR], and nonlinear regression [NLR] DIF methods) on the test scores, which altogether identified 16 DIF items. ShinyItemAnalysis package was employed to draw item characteristic curves (ICCs) for the 16 items in RStudio, which presented four different types of DIF effect. Besides, two experts identified reasons or sources for the DIF effect of four items. The study, therefore, may shed some light on the sustainable development of test fairness in the field of language testing: methodologically, a mixed-methods sequential explanatory design was adopted to guide further test fairness research using flexible methods to achieve research purposes; practically, the result indicates that DIF analysis does not necessarily imply bias. Instead, it only serves as an alarm that calls test developers' attention to further examine the appropriateness of test items.

Keywords: DIF; GEPT-Kids; grade; listening; mixed-methods approach.