Entropy | |
Simple Stopping Criteria for Information Theoretic Feature Selection | |
José C. Príncipe1  Shujian Yu1  | |
[1] Computational NeuroEngineering Laboratory, University of Florida, Gainesville, FL 32611, USA; | |
关键词: feature selection; stopping criterion; conditional mutual information; multivariate matrix-based Rényi’s α-entropy functional; | |
DOI : 10.3390/e21010099 | |
来源: DOAJ |
【 摘 要 】
Feature selection aims to select the smallest feature subset that yields the minimum generalization error. In the rich literature in feature selection, information theory-based approaches seek a subset of features such that the mutual information between the selected features and the class labels is maximized. Despite the simplicity of this objective, there still remain several open problems in optimization. These include, for example, the automatic determination of the optimal subset size (i.e., the number of features) or a stopping criterion if the greedy searching strategy is adopted. In this paper, we suggest two stopping criteria by just monitoring the conditional mutual information (CMI) among groups of variables. Using the recently developed multivariate matrix-based Rényi’s
【 授权许可】
Unknown