Loading [MathJax]/extensions/MathMenu.js
Evaluating the effect of rough set feature selection on the performance of decision trees | IEEE Conference Publication | IEEE Xplore

Evaluating the effect of rough set feature selection on the performance of decision trees


First Page of the Article

Date of Conference: 10-12 May 2006
Date Added to IEEE Xplore: 30 May 2006
Electronic ISBN:978-1-5090-9177-5
Conference Location: Atlanta, GA, USA

First Page of the Article

References is not available for this document.

I. Introduction

In pattern classification, a pattern consists of a set of features or attributes and the task is to classify unseen patterns into predefined categories using a classifier trained on a set of known patterns [17]. However, only some of these features are relevant to classification, whilst others are redundant, containing no information on the classification of patterns. In such circumstances, a classifier trained from the original patterns has a poor performance [17]. Another disadvantage is that a large number of features increases the training and prediction time of the classifier [17]. Feature selection removes the redundant features only from the original dataset while keeping all the important features [17]. Thus, classifiers with simpler structure and higher accuracy than the classifiers trained on the original dataset, can be obtained. 2 main feature selection approaches are the filter and the wrapper approaches [17]. The filter approach selects features based on the property of the data only as a pre-processing step of classifier training. The wrapper method selects features based on the performance of the classifier trained on the reduced datasets using the features.

Select All
1.
Q. Shen and A. Chouchoulas. A rough-fuzzy approach for generating classification rules, Pattern Recognition, volume 35, number 11, pages 341-354, 2002.
2.
Lech Polkowski, Rough Sets: Mathematical Foundations (Advances in soft Computing), Physica-Verlag, 2002.
3.
S. Vinterbo and A. Ohm. Minimal approximate hitting sets and rule templates, International Journal of Approximate Reasoning, volume 25, number 2, pages 123-143, 2000.
4.
J. R. Quinlan. C4.5: Programs for Machine Learning, San Mateo, CA: Morgan Kaufmann, 1993.
5.
J. Dougherty and R. Kohavi and M. Sahami. Supervised and Unsupervised Discretization of Continuous Features, Machine Learning: Proceedings of the Twelfth International Conference, Morgan Kaufmann Publishers, San Francisco, CA, pages 94-202, 1995.
6.
U. M. Fayyad and K. B. Irani. Multi-interval discretization of continuousvalued attributes for classification learning, Proceedings of 13th International Joint Conference on Artificial Intelligence, Morgan Kaufmann, pages 1022-1027, 1993.
7.
S. Hettich, C.L. Blake and C.J. Merz. UCI Repository of machine learning databases, http://www.ics.uci.edu/~mlearn/MLRepository.html, University of California, Irvine, Dept. of Information and Computer Sciences, 1998.
8.
I. E. Buchan and P. E. Bundred and D. J. Kitchiner and T. J. Cole. Tall stature is now a risk factor for obesity in three year olds: serial cross-sectional surveys 1988-2003, Evidence for Population Health Unit, Medical School, The University of Manchester, 2004.
9.
I. H. Witten and E. Frank. Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations, Morgan Kaufmann, San Francisco, 2nd edition, 2005.
10.
C. Shang and Q. Shen. Rough Feature Selection for Neural Network Based Image Classific ation, International Journal of Image and Graphics, volume 2, number 4, pages 541-555, 2002.
11.
M. Mitchell. An Introduction To Genetic Algorithms, MIT Press, 1996.
12.
T. Mitchell. Machine Learning, McGraw Hill, 1997.
13.
A. Skowron and C. Rauszer. The discernibility matrices and functions in information systems, Intelligent Decision Support: Handbook of Applications and Advances of the Rough Set Theory, R. Slowinski, Kluwer, 1992.
14.
A. Ohrn. Discernibility and Rough Sets in Medicine: Tools and Applications, Norwegian University of Science and Technology, Department of Computer and Information Science, Dec. 1999, NTNU report 1999:133, [http://www.idi.ntnu.no/aleks/thesis].
15.
N. Zhong and J. Dong and S. Ohsuga. Using Rough Sets with Heuristics for Feature Selection, Journal of Intelligent Information Systems, volume 16, number 3, pages 199-214, 2001.
16.
C. M. Bishop. Neural Networks for Pattern Recognition, Oxford University Press, 1995.
17.
I. Guyon and A. Elisseeff. An Introduction to Variable and Feature Selection, Journal of Machine Learning Research, volume 3, pages 1157 - 1182, 2003.
Contact IEEE to Subscribe

References

References is not available for this document.