Automatic Tuning of the RBF Kernel Parameter for Batch-Mode Active Learning Algorithms: A Scalable Framework

IEEE Trans Cybern. 2019 Dec;49(12):4460-4472. doi: 10.1109/TCYB.2018.2869861. Epub 2018 Sep 28.

Abstract

Batch-mode active learning algorithms can select a batch of valuable unlabeled samples to manually annotate for reducing the total cost of labeling every unlabeled sample. To facilitate selection of valuable unlabeled samples, many batch-mode active learning algorithms map samples to the reproducing kernel Hilbert space induced by a radial-basis function (RBF) kernel. Setting a proper value to the parameter for the RBF kernel is crucial for such batch-mode active learning algorithms. In this paper, for automatic tuning of the kernel parameter, a hypothesis-margin-based criterion function is proposed. Three frameworks are also developed to incorporate the function of automatic tuning of the kernel parameter with existing batch-model active learning algorithms. In the proposed frameworks, the kernel parameter can be tuned in a single stage or in multiple stages. Tuning the kernel parameter in a single stage aims for the kernel parameter to be suitable for selecting the specified number of unlabeled samples. When the kernel parameter is tuned in multiple stages, the incorporated active learning algorithm can be enforced to make coarse-to-fine evaluations of the importance of unlabeled samples. The proposed framework can also improve the scalability of existing batch-mode active learning algorithms satisfying a decomposition property. Experimental results on data sets comprising hundreds to hundreds of thousands of samples have shown the feasibility of the proposed framework.