The Application of Boosting Algorithm in Data Mining

Article Preview

Abstract:

An algorithm which named Boosting algorithm occurred in the last ten years. It can raise the learning algorithm accuracy with multiple learning and obviously improve the efficiency of learning algorithm by adopting the principle of “comprehensive optimizing”. In addition, it can effectively develop the “weak learning algorithm” with low efficiency into a “strong learning algorithm” with high efficiency. The boosting, a new learning method to integration machine, is based in learning theory and show its good qualities in many fields. The paper elaborates and summarizes the basic ideas of Boosting algorithm and applies it into Data Mining (DM). In recent years, the DM has been extensively adopted out of labs such as in commerce, technological research and engineering technology. In this background, the paper tries improving traditional DM algorithm to solve these problems in industrial application.

You might also be interested in these eBooks

Info:

Periodical:

Pages:

258-261

Citation:

Online since:

July 2014

Export:

Price:

Permissions CCC:

Permissions PLS:

Сopyright:

© 2014 Trans Tech Publications Ltd. All Rights Reserved

Share:

Citation:

* - Corresponding Author

[1] H.A. Simon, Why Should Machines Learn?, in Machine Learning: An Artificial Intelligence Approach R.S. Miehalski, J.G. Carbonell, and T.M. Mitchell, Morgan Kaufmann Los AItos, (1983).

Google Scholar

[2] Dietterich.T. G Machine learning research: Four current directions[J]. AI Magazine, 18(4): 97—136. (1997).

Google Scholar

[3] Tom M. Mitchell: Machine Learning [J]. McGraw Hill, (1997).

Google Scholar

[4] Thomas G Dietterich. In The Handbook of Brain Theory and Neural Networks, Second Edition[J], (2002).

Google Scholar

[5] Michael Kearns, Leslie G. Valiant. Learning Boolean formulate or finite automata is as hard as factoring. Technical Report TR-14-88 Harvard University Aiken.

Google Scholar

[6] Friedman J, Hastie T, Tibshirani R. Additive logistic regression: a statistical view of boosting: [Technical Report], (1998).

Google Scholar

[7] Schapire R.E. Singer Y. Boost Texer: A system for multiclass multi-table text categorization. Machine Learning, (1998).

Google Scholar

[8] Schapire R.E. Singer Y. Using output codes to boost multiclass learning problems. In machine learning: Proc. Of the fourteenth international conference[J]. 313-321, (1997).

Google Scholar

[9] Dietterich T.G., An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting and randomization. AI Magazine[M], (18): 97-136, (1997).

Google Scholar