class boundary
Recently Published Documents


TOTAL DOCUMENTS

26
(FIVE YEARS 8)

H-INDEX

5
(FIVE YEARS 0)

Symmetry ◽  
2021 ◽  
Vol 13 (4) ◽  
pp. 615
Author(s):  
Liliya A. Demidova

The paper considers a solution to the problem of developing two-stage hybrid SVM-kNN classifiers with the aim to increase the data classification quality by refining the classification decisions near the class boundary defined by the SVM classifier. In the first stage, the SVM classifier with default parameters values is developed. Here, the training dataset is designed on the basis of the initial dataset. When developing the SVM classifier, a binary SVM algorithm or one-class SVM algorithm is used. Based on the results of the training of the SVM classifier, two variants of the training dataset are formed for the development of the kNN classifier: a variant that uses all objects from the original training dataset located inside the strip dividing the classes, and a variant that uses only those objects from the initial training dataset that are located inside the area containing all misclassified objects from the class dividing strip. In the second stage, the kNN classifier is developed using the new training dataset above-mentioned. The values of the parameters of the kNN classifier are determined during training to maximize the data classification quality. The data classification quality using the two-stage hybrid SVM-kNN classifier was assessed using various indicators on the test dataset. In the case of the improvement of the quality of classification near the class boundary defined by the SVM classifier using the kNN classifier, the two-stage hybrid SVM-kNN classifier is recommended for further use. The experimental results approve the feasibility of using two-stage hybrid SVM-kNN classifiers in the data classification problem. The experimental results obtained with the application of various datasets confirm the feasibility of using two-stage hybrid SVM-kNN classifiers in the data classification problem.


Symmetry ◽  
2021 ◽  
Vol 13 (2) ◽  
pp. 194
Author(s):  
Zhenhao Jiang ◽  
Tingting Pan ◽  
Chao Zhang ◽  
Jie Yang

Data imbalance is a thorny issue in machine learning. SMOTE is a famous oversampling method of imbalanced learning. However, it has some disadvantages such as sample overlapping, noise interference, and blindness of neighbor selection. In order to address these problems, we present a new oversampling method, OS-CCD, based on a new concept, the classification contribution degree. The classification contribution degree determines the number of synthetic samples generated by SMOTE for each positive sample. OS-CCD follows the spatial distribution characteristics of original samples on the class boundary, as well as avoids oversampling from noisy points. Experiments on twelve benchmark datasets demonstrate that OS-CCD outperforms six classical oversampling methods in terms of accuracy, F1-score, AUC, and ROC.


2020 ◽  
Vol 58 (8) ◽  
pp. 5782-5792
Author(s):  
Sihang Dang ◽  
Zongjie Cao ◽  
Zongyong Cui ◽  
Yiming Pi ◽  
Nengyuan Liu

2019 ◽  
Vol 6 (1-2) ◽  
pp. 215-221
Author(s):  
Anthony Ballas
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document