Error curves for evaluating the quality of feature rankings

PeerJ Comput Sci. 2020 Dec 7:6:e310. doi: 10.7717/peerj-cs.310. eCollection 2020.

Abstract

In this article, we propose a method for evaluating feature ranking algorithms. A feature ranking algorithm estimates the importance of descriptive features when predicting the target variable, and the proposed method evaluates the correctness of these importance values by computing the error measures of two chains of predictive models. The models in the first chain are built on nested sets of top-ranked features, while the models in the other chain are built on nested sets of bottom ranked features. We investigate which predictive models are appropriate for building these chains, showing empirically that the proposed method gives meaningful results and can detect differences in feature ranking quality. This is first demonstrated on synthetic data, and then on several real-world classification benchmark problems.

Keywords: Error curves; Evaluation; Feature ranking.

Grants and funding

This work was supported by The Ad Futura Slovene Human Resources Development and Scholarship Fund, Slovenian Research Agency (through the grants J2-9230 and N2-0128 and a young researcher grant), the European Commission through the grants TAILOR (H2020-ICT-952215) and AI4EU (H2020-ICT-825619). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.