DOI:​10.4156/JCIT.VOL4.ISSUE3.14
Corpus ID: 16663667
Comparison of Classification Methods Based on the Type of Attributes and Sample Size
Reza Entezari-Maleki, A. Rezaei, B. Minaei-Bidgoli
Published 2009
Mathematics, Computer Science

J. Convergence Inf. Technol.
In this paper, the efficacy of seven data classification methods; Decision Tree (DT), k-Nearest Neighbor (k-NN), Logistic Regression (LogR), Naive Bayes (NB), C4.5, Support Vector Machine (SVM) and… Expand
View Via Publisher
Www4.Ncsu.Edu
Share This Paper




102 Citations
Highly Influential Citations
5
Background Citations
28
Methods Citations
24
Results Citations
2
Figures, Tables, and Topics from this paper
Table 1
Figure 1
Table 2
Figure 2
Table 3
Figure 3
Table 4
Figure 4
Table 5
Figure 5
Figure 6
Linear classifier
Support vector machine
Naive Bayes classifier
C4.5 algorithm
Decision tree
Logistic regression
K-nearest neighbors algorithm
102 Citations
Comparing Various Classification Algorithms by WEKA
M. Okhovvat, H. Naderi
2015
In knowledge discovery process classification is an important technique of data mining and widely used in various fields. The development of data-mining applications such as classification and… Expand
PDF
Comparing the Performance of Naive Bayes And Decision Tree Classification Using R
Kirtika Yadav, R. Thareja
Computer Science
2019
TLDR
This paper has taken student performance dataset that has 480 observations and classified these students into different groups and then calculated the accuracy of the classification by using the R language, using the Naive Bayes theorem. Expand
6 Citations
PDF
Classifying Patterns with a Neural Network and Performance Comparison between Naive Bayes, Decision Tree and k-Nearest Neighbor
K. Kalavathi, Nimitha Safar
2015
lassification is a data mining function that assigns items in a collection to target categories or classes. The goal of classification is to accurately predict the target class for each case in the… Expand
Feature Selection System for the Classification of Chronic Kidney Disease Based on Information Gain and Pearson Correlation
Yin Mon Aung
2020
Recently, large amounts of data are widely accessible in information structures, and students have attracted great attention of academics to turn such data into useful knowledge. Researchers also… Expand
A comparison of classification methods across different data complexity scenarios and datasets
Michael Scholz, Tristan Wimmer
Computer ScienceExpert Syst. Appl.
2021
TLDR
The investigation reveals that heterogeneous classifiers perform best on average, bagged CART is especially recommendable for datasets with low dimensionality and high sample size, kernel-based classification methods perform very well especially with a polynomial kernel, but require a rather long time for training and a nearest shrunken neighbor classifier is recommendable in case of unbalanced datasets. Expand
1 Citation
A Comparative Study of Classifiers in the Context of Papaya Disease Recognition
Md. Tarek Habib, Anup Majumder, R. N. Nandi, F. Ahmed, Mohammad Shorif Uddin
Computer Science
IJCCI

2018
TLDR
Nine prominent classification algorithms are compared in index of six performance metrics in a computer vision context to help build an online agro-medical expert system that recognizes the defects of fruit by diseases from an image that is taken using mobile or another handheld device. Expand
4 Citations
Evaluating the High Risk Groups for Suicide: A Comparison of Logistic Regression, Support Vector Machine, Decision Tree and Artificial Neural Network
P. Amini, H. Ahmadinia, J. Poorolajal, Mohammad Moqaddasi Amiri
Mathematics, Medicine
Iranian journal of public health
2016
TLDR
SVM had the best performance to classify fatal suicide attempts comparing to DT, LR and ANN, and showed the highest accuracy and the highest sensitivity for training sample, but the lowest sensitivity for the testing sample. Expand
24 Citations
On the Performance of Classification Techniques with Pixel Removal Applied to Digit Recognition
J. V. Roberts, I. Dialsingh
Computer Science
2016
TLDR
The results from the application of the k-Nearest Neighbour and Recursive Partitioning and Classification Trees methods exemplified the white noise effect in the trimming of the first few frames whilst the Naive Bayes and the Support Vector Machine did not.Expand
PDF
Application and comparison of advanced supervised classifiers in extraction of water bodies from remote sensing images
Arati Paul, Devarati Tripathi, D. Dutta
Computer Science
Sustainable Water Resources Management
2017
TLDR
The study has found that supervised classifier can extract water bodies with a good accuracy from remotely sensed images even with a fewer number of labeled samples, and linear classifiers also yield good accuracy in extracting water bodies across different sensor’s data. Expand
4 Citations
Naïve Bayes Classifier for Journal Quartile Classification
A. P. Wibawa, A. C. Kurniawan, +4 authorsYoungga Rega Nugraha
Computer Science
Int. J. Recent Contributions Eng. Sci. IT
2019
TLDR
The ability of Naive Bayes Classifier is shown to classify the quality of a journal commonly called Quartile with an accuracy of 71.60% and an error rate of 28.40%. Expand
...
1
2
3
4
...
References
SHOWING 1-10 OF 25 REFERENCES
Comparing naive Bayes, decision trees, and SVM with AUC and accuracy
Jin Huang, Jingjing Lu, C. Ling
Computer Science
Third IEEE International Conference on Data Mining
2003
TLDR
It is proved that AUC is, in general, a better measure (defined precisely) than accuracy for evaluating performance of learning algorithms.Expand
226 Citations
Comparison of the decision tree, artificial neural network, and linear regression methods based on the number and types of independent variables and sample size
Yong-Soo Kim
Mathematics, Computer Science
Expert Syst. Appl.
2008
TLDR
For continuous independent variables, a statistical technique was superior to data mining regardless of the number of variables and the sample size and the artificial neural network performance improved faster than that of the other methods as theNumber of classes of categorical variable increased. Expand
63 Citations
PDF
The use of the area under the ROC curve in the evaluation of machine learning algorithms
A. Bradley
Computer SciencePattern Recognit.
1997
TLDR
AUC exhibits a number of desirable properties when compared to overall accuracy: increased sensitivity in Analysis of Variance (ANOVA) tests; a standard error that decreased as both AUC and the number of test samples increased; decision threshold independent; and it is invariant to a priori class probabilities. Expand
A comparative study of K-Nearest Neighbour, Support Vector Machine and Multi-Layer Perceptron for Thalassemia screening
S. Amendolia, G. Cossu, M. Ganadu, B. Golosio, G. Masala, G. M. Mura
Computer Science
2003
TLDR
Both techniques enable thalassemia carriers to be discriminated from healthy subjects with 95% specificity, although the sensitivity of MLP is 92% while that of SVM is 83%.Expand
130 Citations
A comparative analysis of structural risk minimization by support vector machines and nearest neighbor rule
B. Karaçali, R. Ramanath, W. Snyder
Computer SciencePattern Recognit. Lett.
2004
TLDR
Experimental results indicate that the NNSRM formulation is not only computationally less expensive, but also much more robust to varying data representations than SVMs. Expand
30 Citations
PDF
A re-examination of text categorization methods
Yiming Yang, Xin Liu
Computer Science
SIGIR '99

1999
TLDR
The results show that SVM, kNN and LLSF signi cantly outperform NNet and NB when the number of positive training instances per category are small, and that all the methods perform comparably when the categories are over 300 instances. Expand
2,958 Citations
PDF
C4.5: Programs for Machine Learning
J. R. Quinlan
Computer Science
1992
TLDR
A complete guide to the C4.5 system as implemented in C for the UNIX environment, which starts from simple core learning methods and shows how they can be elaborated and extended to deal with typical problems such as missing data and over hitting.Expand
21,693 Citations
An experimental investigation of the impact of aggregation on the performance of data mining with logistic regression
A. Fadlalla
Engineering, Computer Science
Inf. Manag.
2005
TLDR
It is shown experimentally that, for this example, as long as aggregation does not shrink the sample size unduly, it does not significantly impair the performance of the logistic regression model for predicting the direction of the DJIA stock market index. Expand
18 Citations
Comparative analysis of logistic regression and artificial neural network for computer-aided diagnosis of breast masses.
Jae-Hong Song, S. Venkatesh, Emily A Conant, P. Arger, C. Sehgal
Medicine
Academic radiology

2005
TLDR
There was no difference in performance between logistic regression and the artificial neural network as measured by the area under the ROC curve, however, at a fixed 95% sensitivity, the artificial Neural network had higher (12%) specificity compared withlogistic regression value. Expand
78 Citations
PDF
Comparison of three methods for estimating the standard error of the area under the curve in ROC analysis of quantitative data.
K. Hajian-Tilaki, J. Hanley
Mathematics, Medicine
Academic radiology
2002
TLDR
Overall, the LABROC procedure and the method of DeLong et al yielded very close estimates of the SE of AUC, even with data generated from a nonbinormal model. Expand
35 Citations
...
1
2
3
...
SORT BY
Related Papers
Comparison of various classification algorithms on iris datasets using WEKA
Kanubhai G. Patel, J. Vala, J. Pandya, V. V. Nagar
Computer Science
2014
Classification is one of the most important task of data mining. Main task of data mining is data analysis. For study pu...
29 Citations
Naive Bayes vs. Decision Trees vs. Neural Networks in the Classification of Training Web Pages
Daniela Xhemali, C. Hinde, R. Stone
Computer Science
2009
Web classification has been attempted through many different technologies. In this study we concentrate on the compariso...
124 Citations
Show More
2/10
Abstract
Figures, Tables, and Topics
102 Citations
25 References
Related Papers
Stay Connected With Semantic Scholar
What Is Semantic Scholar?
Semantic Scholar is a free, AI-powered research tool for scientific literature, based at the Allen Institute for AI.
Learn More
About
About Us
Publishers
Beta Program
Contact
Research
Team
Datasets
Open Corpus
Supp.ai
Resources
Librarians
Tutorials
FAQ
API
Proudly built by AI2
Terms of ServicePrivacy Policy
By clicking accept or continuing to use the site, you agree to the terms outlined in our Privacy Policy, Terms of Service, and Dataset License
ACCEPT & CONTINUE