Robust Variable Selection and Estimation Based on Kernel Modal Regression
AbstractModel-free variable selection has attracted increasing interest recently due to its flexibility in algorithmic design and outstanding performance in real-world applications. However, most of the existing statistical methods are formulated under the mean square error (MSE) criterion, and susceptible to non-Gaussian noise and outliers. As the MSE criterion requires the data to satisfy Gaussian noise condition, it potentially hampers the effectiveness of model-free methods in complex circumstances. To circumvent this issue, we present a new model-free variable selection algorithm by integrating kernel modal regression and gradient-based variable identification together. The derived modal regression estimator is related closely to information theoretic learning under the maximum correntropy criterion, and assures algorithmic robustness to complex noise by replacing learning of the conditional mean with the conditional mode. The gradient information of estimator offers a model-free metric to screen the key variables. In theory, we investigate the theoretical foundations of our new model on generalization-bound and variable selection consistency. In applications, the effectiveness of the proposed method is verified by data experiments. View Full-Text
Share & Cite This Article
Guo, C.; Song, B.; Wang, Y.; Chen, H.; Xiong, H. Robust Variable Selection and Estimation Based on Kernel Modal Regression. Entropy 2019, 21, 403.
Guo C, Song B, Wang Y, Chen H, Xiong H. Robust Variable Selection and Estimation Based on Kernel Modal Regression. Entropy. 2019; 21(4):403.Chicago/Turabian Style
Guo, Changying; Song, Biqin; Wang, Yingjie; Chen, Hong; Xiong, Huijuan. 2019. "Robust Variable Selection and Estimation Based on Kernel Modal Regression." Entropy 21, no. 4: 403.
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.