Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods
Kasturi Ghosh1 , Susmita Nandi2
Section:Research Paper, Product Type: Journal Paper
Volume-07 ,
Issue-01 , Page no. 1-11, Jan-2019
Online published on Jan 20, 2019
Copyright © Kasturi Ghosh, Susmita Nandi . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
View this paper at Google Scholar | DPI Digital Library
How to Cite this Paper
- IEEE Citation
- MLA Citation
- APA Citation
- BibTex Citation
- RIS Citation
IEEE Citation
IEEE Style Citation: Kasturi Ghosh, Susmita Nandi, “Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods,” International Journal of Computer Sciences and Engineering, Vol.07, Issue.01, pp.1-11, 2019.
MLA Citation
MLA Style Citation: Kasturi Ghosh, Susmita Nandi "Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods." International Journal of Computer Sciences and Engineering 07.01 (2019): 1-11.
APA Citation
APA Style Citation: Kasturi Ghosh, Susmita Nandi, (2019). Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods. International Journal of Computer Sciences and Engineering, 07(01), 1-11.
BibTex Citation
BibTex Style Citation:
@article{Ghosh_2019,
author = {Kasturi Ghosh, Susmita Nandi},
title = {Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {1 2019},
volume = {07},
Issue = {01},
month = {1},
year = {2019},
issn = {2347-2693},
pages = {1-11},
url = {https://www.ijcseonline.org/full_spl_paper_view.php?paper_id=583},
publisher = {IJCSE, Indore, INDIA},
}
RIS Citation
RIS Style Citation:
TY - JOUR
UR - https://www.ijcseonline.org/full_spl_paper_view.php?paper_id=583
TI - Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods
T2 - International Journal of Computer Sciences and Engineering
AU - Kasturi Ghosh, Susmita Nandi
PY - 2019
DA - 2019/01/20
PB - IJCSE, Indore, INDIA
SP - 1-11
IS - 01
VL - 07
SN - 2347-2693
ER -




Abstract
Feature selection (FS) is an important aspect of data mining. Now a days availability of information with hundreds of variables leads to high dimensional, irrelevant and redundant data. Thus FS techniques must be applied on the datasets before classification or rule generation. It basically aims at reducing the number of attributes by removing irrelevant or redundant ones, while trying to reduce computation time and improve performance of classifiers. In this paper three different FS methods are used, Correlation Based, Information Gain Based and Rough set Based FS method. A statistical analysis of three different classifier`s performance is also done in order to provide a detailed view.
Key-Words / Index Term
Data Mining (DM), Feature Selection (FS), Rough Set, Degree of Dependency, Decision Tree (J48 algorithm), Naive Bayes Algorithm (NB), K-Nearest Neighbor Algorithm (KNN), Classification, Statistical Analysis
References
[1] Imran Fareed Nizami, Muhammad Majid, Hammad Afzal and Khawar Khurshi, “Impact of Feature Selection Algorithms on Blind Image Quality Assessment”, Arabian Journal for Science and Engineering, pp 1–14, August 2017.
[2] Abdullah S. Ghareb, Abdul Razak Hamdan and Azuraliza Abu Bakar, “Integrating Noun-Based Feature Ranking and Selection Methods with Arabic Text Associative Classification Approach”, Arabian Journal for Science and Engineering, Vol.39, Issue.11, pp 7807–7822, November 2014.
[3] Z. Pawlak, Rough sets, International Journal of Computer and Information Sciences, 11, 341-356, 1982
[4] Javad Rahimipour Anaraki, Kerman, Iran, Mahdi Eftekhari, “Rough Set Based Feature Selection: A Review”, 5th Conference on Information and Knowledge Technology, IEEE, 2013.
[5] G. K. Gupta, “Introduction to Data Mining with Case Studies”, Prentice Hall of India New Delhi, 2006.
[6] P-N. Tan, M. Steinbach, V. Kumar, “Introduction to Data Mining”, Addison Wesley Publishing, 2006.
[7] O.Maimon and L.Rokach, “Data Mining and Knowledge Discovery”, Springer Science and Business Media, 2005.
[8] X. Niuniu and L. Yuxun, “Review of Decision Trees”, IEEE, 2010.
[9] Payam Emami Khoonsari and AhmadReza Motie, “A Comparison of Efficiency and Robustness of ID3 and C4.5 Algorithms Using Dynamic Test and Training Data Sets”, International Journal of Machine Learning and Computing, Vol.2, Issue.5, October 2012.
[10] V. Garcia, C. Debreuve, “Fast k Nearest Neighbor Search using GPU”, IEEE, 2008.
[11] A. Ashari I. Paryudi and A Min Tjoa, “Performance Comparison between Naïve Bayes Decision Tree and k-Nearest Neighbor in Searching Alternative Design in an Energy Simulation Tool”, International Journal of Advanced Computer Science and Applications, Vol.4, Issue. 11, 2013.
[12] Dougherty, J., R. Kohavi and M. Sahami, “Supervised and unsupervised discretization of continuous features”, Proceeding of the 12th International Conference on Machine Learning, 1995.
[13] https:// archive.ics.uci.edu/ ml/ datasets/ Diabetic + Retinopathy + Debrecen + Data + Set
[14] https:// archive.ics.uci.edu/ ml/ datasets/ EEG + Eye + State
[15] https:// archive.ics.uci.edu/ ml/ datasets/ cardiotocography
[16] https://archive.ics.uci.edu/ ml/ datasets/ Thoracic + Surgery + Data
[17] PIDD Dataset, https:// archive.ics.uci.edu/ ml/ datasets/ pima + indians + diabetes
[18] https:// archive.ics.uci.edu/ ml/ datasets/ ILPD + (Indian + Liver + Patient + Dataset)
[19] https:// archive.ics.uci.edu/ ml/ datasets/ breast + cancer + wisconsin + (original)