classification rule
Recently Published Documents


TOTAL DOCUMENTS

366
(FIVE YEARS 64)

H-INDEX

23
(FIVE YEARS 2)

2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Wang Zhouhuo

In order to solve the problem of large data classification of human resources, a new parallel classification algorithm of large data of human resources based on the Spark platform is proposed in this study. According to the spark platform, it can complete the update and distance calculation of the human resource big data clustering center and design the big data clustering process. Based on this, the K-means clustering method is introduced to mine frequent itemsets of large data and optimize the aggregation degree of similar large data. A fuzzy genetic algorithm is used to identify the balance of big data. This study adopts the selective integration method to study the unbalanced human resource database classifier in the process of transmission, introduces the decision contour matrix to construct the anomaly support model of the set of unbalanced human resource data classifier, identifies the features of the big data of human resource in parallel, repairs the relevance of the big data of human resource, introduces the improved ant colony algorithm, and finally realizes the design of the parallel classification algorithm of the big data of human resource. The experimental results show that the proposed algorithm has a low time cost, good classification effect, and ideal parallel classification rule complexity.


2021 ◽  
Vol 152 (A4) ◽  
Author(s):  
G A Gratsos ◽  
H N Psaraftis ◽  
P Zachariadis

The paper contributes a reasoned methodology and useful data to the debate that is taking place in the context of rational Goal Based Standards. It is hoped that the paper will generate further debate which should eventually lead to generally accepted conclusions on meaningful minimum design and classification rule standards.


2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Tristan Mary-Huard ◽  
Vittorio Perduca ◽  
Marie-Laure Martin-Magniette ◽  
Gilles Blanchard

Abstract In the context of finite mixture models one considers the problem of classifying as many observations as possible in the classes of interest while controlling the classification error rate in these same classes. Similar to what is done in the framework of statistical test theory, different type I and type II-like classification error rates can be defined, along with their associated optimal rules, where optimality is defined as minimizing type II error rate while controlling type I error rate at some nominal level. It is first shown that finding an optimal classification rule boils down to searching an optimal region in the observation space where to apply the classical Maximum A Posteriori (MAP) rule. Depending on the misclassification rate to be controlled, the shape of the optimal region is provided, along with a heuristic to compute the optimal classification rule in practice. In particular, a multiclass FDR-like optimal rule is defined and compared to the thresholded MAP rules that is used in most applications. It is shown on both simulated and real datasets that the FDR-like optimal rule may be significantly less conservative than the thresholded MAP rule.


2021 ◽  
Vol 2021 ◽  
pp. 1-8
Author(s):  
Vikas Maheshwari ◽  
Md Rashid Mahmood ◽  
Sumukham Sravanthi ◽  
N. Arivazhagan ◽  
A. ParimalaGandhi ◽  
...  

Increasing the growth of big data, particularly in healthcare-Internet of Things (IoT) and biomedical classes, tends to help patients by identifying the disease early through methods for the analysis of medical data. Hence, nanotechnology-based IOT biosensors play a significant role in the medical field. Problem. However, the consistency continues to decrease where missing data occurs in such medical data from nanotechnology-based IOT biosensors. Furthermore, each region has its own special features, which further lowers the accuracy of prediction. The proposed model initially reconstructs lost or partial data in order to address the challenge of handling the medical data structures with incomplete data. Methods. An adaptive architecture is proposed to enhance the computing capabilities to predict the disease automatically. The medical databases are managed by unpredictable environments. This optimized paradigm for diagnosis produces the fuzzy, genetically categorized decision tree algorithm. This work uses a normalized classifier namely fuzzy-based decision tree (FDT) algorithm for classifying the data collected via nanotechnology-based IOT biosensors, and this helps in the identification of nondeterministic instances from unstructured datasets relating to the medical diagnosis. The FDT algorithm is further enhanced by using genetic algorithms for effective classification of instances. Finally, the proposed system uses two larger datasets to verify the predictive precision. In order to describe a fuzzy decision tree algorithm based upon the fitness function value, a modified decision classification rule is used. The structure and unstructured databases are configured for processing. Results and Conclusions. This evaluation of test patterns helps to track the efficiency of FDT with optimized rules during the training and testing stages. The proposed method is validated against nanotechnology-based IOT biosensors data in terms of accuracy, sensitivity, specificity, and F -measure. The results of the simulation show that the proposed method achieves a higher rate of accuracy than the other methods. Other metrics relating to the model with and without feature selection show an improved sensitivity, specificity, and F -measure rate than the existing methods.


2021 ◽  
Vol 2021 ◽  
pp. 1-11
Author(s):  
Lin Wu

With the continuous improvement of living standards, people began to pay more and more attention to sports, and the impact of sports on human health and physique has been paid more and more attention. This study mainly analyzes the scientific impact of sports on human health and physique under the background of big data. Firstly, the big data analytic hierarchy process is used to construct the comprehensive evaluation structure system of sports on human health and physique. Then, an improved big data adaptive ant colony classification rule algorithm is proposed. Finally, the performance evaluation and physical impact analysis of the improved big data algorithm are carried out. The results show that compared with other algorithms, ACA ∗ (ant colony algorithm) based on big data has more obvious advantages in stability, optimization ability, running time, and convergence speed and is more suitable for practical application. In general, the improvement of the physical fitness level of the association members in 2019 mainly depends on the results of the improvement of the physical fitness level. In the future, we need to strengthen physical exercise, change living habits and traffic habits, and other methods to optimize the overall physical fitness.


2021 ◽  
Vol 4 (4) ◽  
Author(s):  
Giuseppe Caristi ◽  
Roberto Guarneri ◽  
Sabrin Lo Bosco

In this paper we show how it can be useful to the probability of intersections in the determination of a classification rule for raster conversions in Geographical Information System (GIS) and GRASS GIS for the Road Network Analysis (RNA). We use a geometric probabilities approach for irregular path considering these results for transportation planning operations. We study two particular problems with irregular tessellations, in order to have a situation more realistic respect to map GIS and considering also the maximum value of probability to narrow the range of possible probability values.


2021 ◽  
Vol 3 (1) ◽  
Author(s):  
Julika Ribbat-Idel ◽  
Florian Stellmacher ◽  
Florian Jann ◽  
Nicolas Kalms ◽  
Inke R. König ◽  
...  

Abstract Background Thrombus histology has become a potential diagnostic tool for the etiology assessment of patients with ischemic stroke caused by embolic proximal vessel occlusion. We validated a classification rule that differentiates between cardiac and arteriosclerotic emboli in individual stroke patients. We aim to describe in detail the development of this classification rule and disclose its reliability. Methods The classification rule is based on the hypothesis that cardiac emboli arise out of separation thrombi and arteriosclerotic emboli result from agglutinative thrombi. 125 emboli recovered by thrombectomy from stroke patients and 11 thrombi serving as references for cardiac (n = 5) and arteriosclerotic emboli (n = 6) were Hematoxylin and eosin, Elastica-van Gieson and CD61 stained and rated independently by two histopathologists blinded to the presumed etiology by several pre-defined criteria. Intra- and interobserver reliabilities of all criteria were determined. Out of the different criteria, three criteria with the most satisfactory reliability values were selected to compose the classification rule that was finally adjusted to the reference thrombi. Reliabilities of the classification rule were calculated by using the emboli of stroke patients. Results The classification rule reached intraobserver reliabilities for the two raters of 92.9% and 68.2%, respectively. Interobserver reliability was 69.9%. Conclusions A new classification rule for emboli obtained from thrombectomy was established. Within the limitations of histological investigations, it is reliable and able to distinguish between cardioembolic and arteriosclerotic emboli.


2021 ◽  
Vol 15 ◽  
Author(s):  
Anti Ingel ◽  
Raul Vicente

In this study, the information bottleneck method is proposed as an optimisation method for steady-state visual evoked potential (SSVEP)-based brain-computer interface (BCI). The information bottleneck is an information-theoretic optimisation method for solving problems with a trade-off between preserving meaningful information and compression. Its main practical application in machine learning is in representation learning or feature extraction. In this study, we use the information bottleneck to find optimal classification rule for a BCI. This is a novel application for the information bottleneck. This approach is particularly suitable for BCIs since the information bottleneck optimises the amount of information transferred by the BCI. Steady-state visual evoked potential-based BCIs often classify targets using very simple rules like choosing the class corresponding to the largest feature value. We call this classifier the arg max classifier. It is unlikely that this approach is optimal, and in this study, we propose a classification method specifically designed to optimise the performance measure of BCIs. This approach gives an advantage over standard machine learning methods, which aim to optimise different measures. The performance of the proposed algorithm is tested on two publicly available datasets in offline experiments. We use the standard power spectral density analysis (PSDA) and canonical correlation analysis (CCA) feature extraction methods on one dataset and show that the current approach outperforms most of the related studies on this dataset. On the second dataset, we use the task-related component analysis (TRCA) method and demonstrate that the proposed method outperforms the standard argmax classification rule in terms of information transfer rate when using a small number of classes. To our knowledge, this is the first time the information bottleneck is used in the context of SSVEP-based BCIs. The approach is unique in the sense that optimisation is done over the space of classification functions. It potentially improves the performance of BCIs and makes it easier to calibrate the system for different subjects.


Author(s):  
Tristan Mary-Huard ◽  
Sarmistha Das ◽  
Indranil Mukhopadhyay ◽  
Stephane Robin

Abstract Motivation Combining the results of different experiments to exhibit complex patterns or to improve statistical power is a typical aim of data integration. The starting point of the statistical analysis often comes as sets of p-values resulting from previous analyses, that need to be combined in a flexible way to explore complex hypotheses, while guaranteeing a low proportion of false discoveries. Results We introduce the generic concept of composed hypothesis, which corresponds to an arbitrary complex combination of simple hypotheses. We rephrase the problem of testing a composed hypothesis as a classification task, and show that finding items for which the composed null hypothesis is rejected boils down to fitting a mixture model and classify the items according to their posterior probabilities. We show that inference can be efficiently performed and provide a thorough classification rule to control for type I error. The performance and the usefulness of the approach are illustrated on simulations and on two different applications. The method is scalable, does not require any parameter tuning, and provided valuable biological insight on the considered application cases. Availability The QCH methodology is implemented in the qch R package hosted on CRAN.


Sign in / Sign up

Export Citation Format

Share Document