The PDF file you selected should load here if your Web browser has a PDF reader plug-in installed (for example, a recent version of Adobe Acrobat Reader).

If you would like more information about how to print, save, and work with PDFs, Highwire Press provides a helpful Frequently Asked Questions about PDFs.

Alternatively, you can download the PDF file directly to your computer, from where it can be opened using a PDF reader. To download the PDF, click the Download link above.

Fullscreen Fullscreen Off


In data mining feature subset selection is a preprocessing step in classification that lessens dimensionality, eliminates unrelated data, increases accuracy and improves unambiguousness. The next step in classification is to produce enormous amount of rules from the reduced feature set from which high class rules are chosen to build effectual classifier. In this paper, Information Gain (IG) has been used to rank the features. Multi-Layer Perception (MLP) with back-propagation reduces features to achieve higher accuracy in classification. Artificial Neural Networks (ANN) classifier is used for classification. We handle the discretization of continuous valued features by dividing the series of values into a limited number of subsections. Wine Recognition data set taken from the UCI machine learning repository is used for testing. Original 13 features are drawn in classification. The thirteen features are reduced to five features. Experimental results show that the accuracy in training dataset is 98.62% and in the validation dataset is 96.06%. The accuracy difference between 13 features and 5 features in the training data is 5.54% and in validation data is 2.00%. We then build a Decision Tree and concentrate on discovering significant rules from the reduced data set that provide better classification.

Keywords

Back-Propagation, Classification, Decision Tree, Feature Subset Selection, Multi-Layer Perception
User