Large Margin Feature Selection for Support Vector Machine

Article Preview

Abstract:

Feature selection is an preprocessing step in pattern analysis and machine learning. In this paper, we design a algorithm for feature subset. We present L1-norm regularization technique for sparse feature weight. Margin loss are introduced to evaluate features, and we employs gradient descent to search the optimal solution to maximize margin. The proposed technique is tested on UCI data sets. Compared with four margin based loss functions for SVM, the proposed technique is effective and efficient.

You might also be interested in these eBooks

Info:

Periodical:

Pages:

161-164

Citation:

Online since:

January 2013

Export:

Price:

Permissions CCC:

Permissions PLS:

Сopyright:

© 2013 Trans Tech Publications Ltd. All Rights Reserved

Share:

Citation:

[1] Yvan Saeys, Inaki Inza, Pedro Larranaga, A review of feature selection techniques in bioinformatics, Bioinformatics, 19 (2007) 2507-2517.

DOI: 10.1093/bioinformatics/btm344

Google Scholar

[2] H. Liu, L. Yu, Toward integrating feature selection algorithms for classification and clustering, IEEE Transactions on Knowledge and Data Engineering, 17 (2005) 494-502.

DOI: 10.1109/tkde.2005.66

Google Scholar

[3] R. Kohavi, G. John, Wrapper for feature subset selection. Artificial Intelligence, 1997 234-273.

Google Scholar

[4] Gilad. Bachrach R, Navot A, and Tishby N, Margin based feature selection theory and algorithms, In proceedings of the 21'st International Conference on Machine Learning , (2004) 40.

DOI: 10.1145/1015330.1015352

Google Scholar

[5] Hanchuan Peng, Fuhui Long, and Chris Ding, Feature Selection Based on Mutual Information: Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence, 8 (2005) 1226-1236.

DOI: 10.1109/tpami.2005.159

Google Scholar

[6] Andrew Y. Ng, Feature selection, L1 vs. L2 regularization, and rotational invariance, Proceedings of the 21st International Conference on Machine Learning, Ban, Canada, (2004) 35-43.

Google Scholar

[7] Y. Freund, An Adaptive Version of the Boost by majority algorithm, International Conference on Machine Learning, 43(2001) 293-318.

Google Scholar

[8] Crammer,K., Gilad-Bachrach R., Navot A., Tishby N., Margin analysis of the lvq algorithm, Proc. 17'th Conference on Neural Information Processing Systems, (2002) 462-469.

Google Scholar

[9] URL: http : /www. ics . uci . edu/ mlearn/ MLRRepository. html, (1996).

Google Scholar