scholarly journals Decision Tree-based Machine Learning Algorithms to Classify Rice Plant Diseases

Rice is one of the most important foods on earth for human beings. India and China are two countries in the world mostly depend on rice. The output of this crop depends on the many parameters such as soil, water supply, pesticides used, time duration, and infected diseases. Rice Plant Disease (RPD) is one of the important factors that decrease the quantity and quality of rice. Identifying the type of rice plant disease and taking corrective action against the disease in time is always challenging for the farmers. Although the rice plant is affected by many diseases, Bacterial Leaf Blight (BLB), Brown Spot (BS), and Leaf Smut (LS) are major diseases. Identification of this disease is really challenging because the infected leaf has to be processed by the human eye. So in this paper, we focused on machine learning techniques to identify and classify the RPD. We have collected infected rice plant data from the UCI Machine Learning repository. The data set consists of 120 images of infected rice plants in which 40 images are BLB, 40 are BS, and 40 are LS. Experiments are conducted using Decision tree-based machine learning algorithms such as RandomForest, REPTree, and J48. In order to extract the numerical features from the infected images, we have used ColourLayoutFilter supported by WEKA. Experimental analysis is done using 65% data for training and 35% data for testing. The experiments unfold that the Random Forest algorithm is exceptional in predicting RPD.

2021 ◽  
Vol 11 (4) ◽  
pp. 251-264
Author(s):  
Radhika Bhagwat ◽  
Yogesh Dandawate

Plant diseases cause major yield and economic losses. To detect plant disease at early stages, selecting appropriate techniques is imperative as it affects the cost, diagnosis time, and accuracy. This research gives a comprehensive review of various plant disease detection methods based on the images used and processing algorithms applied. It systematically analyzes various traditional machine learning and deep learning algorithms used for processing visible and spectral range images, and comparatively evaluates the work done in literature in terms of datasets used, various image processing techniques employed, models utilized, and efficiency achieved. The study discusses the benefits and restrictions of each method along with the challenges to be addressed for rapid and accurate plant disease detection. Results show that for plant disease detection, deep learning outperforms traditional machine learning algorithms while visible range images are more widely used compared to spectral images.


2021 ◽  
Vol 75 (3) ◽  
pp. 83-93
Author(s):  
Zh. A. Buribayev ◽  
◽  
Zh. E. Amirgaliyeva ◽  
A.S. Ataniyazova ◽  
Z. M. Melis ◽  
...  

The article considers the relevance of the introduction of intelligent weed detection systems, in order to save herbicides and pesticides, as well as to obtain environmentally friendly products. A brief review of the researchers' scientific works is carried out, which describes the methods of identification, classification and discrimination of weeds developed by them based on machine learning algorithms, convolutional neural networks and deep learning algorithms. This research paper presents a program for detecting pests of agricultural land using the algorithms K-Nearest Neighbors, Random Forest and Decision Tree. The data set is collected from 4 types of weeds, such as amaranthus, ambrosia, bindweed and bromus. According to the results of the assessment, the accuracy of weed detection by the classifiers K-Nearest Neighbors, Random Forest and Decision Tree was 83.3%, 87.5%, and 80%. Quantitative results obtained on real data demonstrate that the proposed approach can provide good results in classifying low-resolution images of weeds.


2021 ◽  
Vol 11 (15) ◽  
pp. 6728
Author(s):  
Muhammad Asfand Hafeez ◽  
Muhammad Rashid ◽  
Hassan Tariq ◽  
Zain Ul Abideen ◽  
Saud S. Alotaibi ◽  
...  

Classification and regression are the major applications of machine learning algorithms which are widely used to solve problems in numerous domains of engineering and computer science. Different classifiers based on the optimization of the decision tree have been proposed, however, it is still evolving over time. This paper presents a novel and robust classifier based on a decision tree and tabu search algorithms, respectively. In the aim of improving performance, our proposed algorithm constructs multiple decision trees while employing a tabu search algorithm to consistently monitor the leaf and decision nodes in the corresponding decision trees. Additionally, the used tabu search algorithm is responsible to balance the entropy of the corresponding decision trees. For training the model, we used the clinical data of COVID-19 patients to predict whether a patient is suffering. The experimental results were obtained using our proposed classifier based on the built-in sci-kit learn library in Python. The extensive analysis for the performance comparison was presented using Big O and statistical analysis for conventional supervised machine learning algorithms. Moreover, the performance comparison to optimized state-of-the-art classifiers is also presented. The achieved accuracy of 98%, the required execution time of 55.6 ms and the area under receiver operating characteristic (AUROC) for proposed method of 0.95 reveals that the proposed classifier algorithm is convenient for large datasets.


2020 ◽  
Vol 98 (Supplement_4) ◽  
pp. 126-127
Author(s):  
Lucas S Lopes ◽  
Christine F Baes ◽  
Dan Tulpan ◽  
Luis Artur Loyola Chardulo ◽  
Otavio Machado Neto ◽  
...  

Abstract The aim of this project is to compare some of the state-of-the-art machine learning algorithms on the classification of steers finished in feedlots based on performance, carcass and meat quality traits. The precise classification of animals allows for fast, real-time decision making in animal food industry, such as culling or retention of herd animals. Beef production presents high variability in its numerous carcass and beef quality traits. Machine learning algorithms and software provide an opportunity to evaluate the interactions between traits to better classify animals. Four different treatment levels of wet distiller’s grain were applied to 97 Angus-Nellore animals and used as features for the classification problem. The C4.5 decision tree, Naïve Bayes (NB), Random Forest (RF) and Multilayer Perceptron (MLP) Artificial Neural Network algorithms were used to predict and classify the animals based on recorded traits measurements, which include initial and final weights, sheer force and meat color. The top performing classifier was the C4.5 decision tree algorithm with a classification accuracy of 96.90%, while the RF, the MLP and NB classifiers had accuracies of 55.67%, 39.17% and 29.89% respectively. We observed that the final decision tree model constructed with C4.5 selected only the dry matter intake (DMI) feature as a differentiator. When DMI was removed, no other feature or combination of features was sufficiently strong to provide good prediction accuracies for any of the classifiers. We plan to investigate in a follow-up study on a significantly larger sample size, the reasons behind DMI being a more relevant parameter than the other measurements.


Author(s):  
Jakub Gęca

The consequences of failures and unscheduled maintenance are the reasons why engineers have been trying to increase the reliability of industrial equipment for years. In modern solutions, predictive maintenance is a frequently used method. It allows to forecast failures and alert about their possibility. This paper presents a summary of the machine learning algorithms that can be used in predictive maintenance and comparison of their performance. The analysis was made on the basis of data set from Microsoft Azure AI Gallery. The paper presents a comprehensive approach to the issue including feature engineering, preprocessing, dimensionality reduction techniques, as well as tuning of model parameters in order to obtain the highest possible performance. The conducted research allowed to conclude that in the analysed case , the best algorithm achieved 99.92% accuracy out of over 122 thousand test data records. In conclusion, predictive maintenance based on machine learning represents the future of machine reliability in industry.


2020 ◽  
Vol 9 (3) ◽  
pp. 34
Author(s):  
Giovanna Sannino ◽  
Ivanoe De Falco ◽  
Giuseppe De Pietro

One of the most important physiological parameters of the cardiovascular circulatory system is Blood Pressure. Several diseases are related to long-term abnormal blood pressure, i.e., hypertension; therefore, the early detection and assessment of this condition are crucial. The identification of hypertension, and, even more the evaluation of its risk stratification, by using wearable monitoring devices are now more realistic thanks to the advancements in Internet of Things, the improvements of digital sensors that are becoming more and more miniaturized, and the development of new signal processing and machine learning algorithms. In this scenario, a suitable biomedical signal is represented by the PhotoPlethysmoGraphy (PPG) signal. It can be acquired by using a simple, cheap, and wearable device, and can be used to evaluate several aspects of the cardiovascular system, e.g., the detection of abnormal heart rate, respiration rate, blood pressure, oxygen saturation, and so on. In this paper, we take into account the Cuff-Less Blood Pressure Estimation Data Set that contains, among others, PPG signals coming from a set of subjects, as well as the Blood Pressure values of the latter that is the hypertension level. Our aim is to investigate whether or not machine learning methods applied to these PPG signals can provide better results for the non-invasive classification and evaluation of subjects’ hypertension levels. To this aim, we have availed ourselves of a wide set of machine learning algorithms, based on different learning mechanisms, and have compared their results in terms of the effectiveness of the classification obtained.


2019 ◽  
Vol 1 (1) ◽  
pp. 384-399 ◽  
Author(s):  
Thais de Toledo ◽  
Nunzio Torrisi

The Distributed Network Protocol (DNP3) is predominately used by the electric utility industry and, consequently, in smart grids. The Peekaboo attack was created to compromise DNP3 traffic, in which a man-in-the-middle on a communication link can capture and drop selected encrypted DNP3 messages by using support vector machine learning algorithms. The communication networks of smart grids are a important part of their infrastructure, so it is of critical importance to keep this communication secure and reliable. The main contribution of this paper is to compare the use of machine learning techniques to classify messages of the same protocol exchanged in encrypted tunnels. The study considers four simulated cases of encrypted DNP3 traffic scenarios and four different supervised machine learning algorithms: Decision tree, nearest-neighbor, support vector machine, and naive Bayes. The results obtained show that it is possible to extend a Peekaboo attack over multiple substations, using a decision tree learning algorithm, and to gather significant information from a system that communicates using encrypted DNP3 traffic.


Diagnostics ◽  
2019 ◽  
Vol 9 (3) ◽  
pp. 104 ◽  
Author(s):  
Ahmed ◽  
Yigit ◽  
Isik ◽  
Alpkocak

Leukemia is a fatal cancer and has two main types: Acute and chronic. Each type has two more subtypes: Lymphoid and myeloid. Hence, in total, there are four subtypes of leukemia. This study proposes a new approach for diagnosis of all subtypes of leukemia from microscopic blood cell images using convolutional neural networks (CNN), which requires a large training data set. Therefore, we also investigated the effects of data augmentation for an increasing number of training samples synthetically. We used two publicly available leukemia data sources: ALL-IDB and ASH Image Bank. Next, we applied seven different image transformation techniques as data augmentation. We designed a CNN architecture capable of recognizing all subtypes of leukemia. Besides, we also explored other well-known machine learning algorithms such as naive Bayes, support vector machine, k-nearest neighbor, and decision tree. To evaluate our approach, we set up a set of experiments and used 5-fold cross-validation. The results we obtained from experiments showed that our CNN model performance has 88.25% and 81.74% accuracy, in leukemia versus healthy and multiclass classification of all subtypes, respectively. Finally, we also showed that the CNN model has a better performance than other wellknown machine learning algorithms.


Sign in / Sign up

Export Citation Format

Share Document