scholarly journals Linear Regression Algorithm in Machine Learning through MATLAB

Author(s):  
Kalva Sindhu Priya

Abstract: In the present scenario, it is quite aware that almost every field is moving into machine based automation right from fundamentals to master level systems. Among them, Machine Learning (ML) is one of the important tool which is most similar to Artificial Intelligence (AI) by allowing some well known data or past experience in order to improve automatically or estimate the behavior or status of the given data through various algorithms. Modeling a system or data through Machine Learning is important and advantageous as it helps in the development of later and newer versions. Today most of the information technology giants such as Facebook, Uber, Google maps made Machine learning as a critical part of their ongoing operations for the better view of users. In this paper, various available algorithms in ML is given briefly and out of all the existing different algorithms, Linear Regression algorithm is used to predict a new set of values by taking older data as reference. However, a detailed predicted model is discussed clearly by building a code with the help of Machine Learning and Deep Learning tool in MATLAB/ SIMULINK. Keywords: Machine Learning (ML), Linear Regression algorithm, Curve fitting, Root Mean Squared Error

2020 ◽  
Vol 12 (5) ◽  
pp. 41-51
Author(s):  
Shaimaa Mahmoud ◽  
◽  
Mahmoud Hussein ◽  
Arabi Keshk

Opinion mining in social networks data is considered as one of most important research areas because a large number of users interact with different topics on it. This paper discusses the problem of predicting future products rate according to users’ comments. Researchers interacted with this problem by using machine learning algorithms (e.g. Logistic Regression, Random Forest Regression, Support Vector Regression, Simple Linear Regression, Multiple Linear Regression, Polynomial Regression and Decision Tree). However, the accuracy of these techniques still needs to be improved. In this study, we introduce an approach for predicting future products rate using LR, RFR, and SVR. Our data set consists of tweets and its rate from 1:5. The main goal of our approach is improving the prediction accuracy about existing techniques. SVR can predict future product rate with a Mean Squared Error (MSE) of 0.4122, Linear Regression model predict with a Mean Squared Error of 0.4986 and Random Forest Regression can predict with a Mean Squared Error of 0.4770. This is better than the existing approaches accuracy.


2020 ◽  
Author(s):  
Satish Kumar ◽  
Mohamed Rafiullah ◽  
Khalid Siddiqui

BACKGROUND Diabetic kidney disease (DKD) is a progressive disease that leads to loss of kidney function. As early intervention improves patient outcomes, it is essential to identify the patients who are at high risk of developing DKD. Artificial Intelligence methods apply different machine learning classification techniques to identify high-risk patients by building a predictive model from a given dataset. OBJECTIVE This study aims to find an accurate classification technique for predicting DKD by comparing different classification techniques applied to a DKD dataset using WEKA machine learning software. METHODS We analyzed the performance of nine different classification techniques on a DKD dataset with 410 instances and 18 attributes. 66% of the dataset was used to build a model, and 33% of the data was used for evaluating the model. The performance of classification techniques were assessed based on their execution time, accuracy, correctly and incorrectly classified instances, kappa statistics (K), mean absolute error, root mean squared error and true values of the confusion matrix. RESULTS Random Forest classifier was found to be the best performing technique with an accuracy of 76.5854% and a higher K value (0.5306) in comparison to other classifiers. Besides, it also showed the lowest root mean squared error rate (0.4007). From the confusion matrix, it was found that there were 46 false-positive instances and 50 false-negative instances from the Random Forest technique. CONCLUSIONS This study identified the Random Forest classification technique as the best performing classifier and accurate prediction method for DKD. CLINICALTRIAL NA


2021 ◽  
Vol 8 (3) ◽  
pp. 539
Author(s):  
Ayu Ahadi Ningrum ◽  
Iwan Syarif ◽  
Agus Indra Gunawan ◽  
Edi Satriyanto ◽  
Rosmaliati Muchtar

<p>Kualitas dan ketersediaan pasokan listrik menjadi hal yang sangat penting. Kegagalan pada transformator menyebabkan pemadaman listrik yang dapat menurunkan kualitas layanan kepada pelanggan. Oleh karena itu, pengetahuan tentang umur transformator sangat penting untuk menghindari terjadinya kerusakan transformator secara mendadak yang dapat mengurangi kualitas layanan pada pelanggan. Penelitian ini bertujuan untuk mengembangkan aplikasi yang dapat memprediksi umur transformator secara akurat menggunakan metode <em>Deep Learning-LSTM. LSTM </em>adalah metode yang dapat digunakan untuk mempelajari suatu pola pada data deret waktu. Data yang digunakan dalam penelitian ini bersumber dari 25 unit transformator yang meliputi data dari sensor arus, tegangan, dan suhu. Analisis performa yang digunakan untuk mengukur kinerja LSTM adalah <em>Root Mean Squared Error</em> (RMSE) dan <em>Squared Correlation (SC</em>). Selain LSTM, penelitian ini juga menerapkan <em>algoritma Multilayer Perceptron, Linear Regression,</em> dan <em>Gradient Boosting Regressor</em> sebagai algoritma pembanding.  Hasil eksperimen menunjukkan bahwa LSTM mempunyai kinerja yang sangat bagus setelah dilakukan pencarian komposisi data, seleksi fitur menggunakan algoritma KBest dan melakukan percobaan beberapa variasi parameter. Hasil penelitian menunjukkan bahwa metode <em>Deep Learning-LSTM</em> mempunyai kinerja yang lebih baik daripada 3 algoritma lain yaitu nilai RMSE= 0,0004 dan nilai <em>Squared Correlation</em>= 0,9690.</p><p> </p><p><em><strong>Abstract</strong></em></p><p><em></em><em>The quality and availability of the electricity supply is very important. Failures in the transformer cause power outages which can reduce the quality of service to customers. Therefore, knowledge of transformer life is very important to avoid sudden transformer damage which can reduce the quality of service to customers. This study aims to develop applications that can predict transformer life accurately using the Deep Learning-LSTM method. LSTM is a method that can be used to study a pattern in time series data. The data used in this research comes from 25 transformer units which include data from current, voltage, and temperature sensors. The performance analysis used to measure LSTM performance is Root Mean Squared Error (RMSE) and Squared Correlation (SC). Apart from LSTM, this research also applies the Multilayer Perceptron algorithm, Linear Regression, and Gradient Boosting Regressor as a comparison algorithm. The experimental results show that LSTM has a very good performance after searching for the composition of the data, selecting features using the KBest algorithm and experimenting with several parameter variations. The results showed that the Deep Learning-LSTM method had better performance than the other 3 algorithms, namely the value of RMSE = 0.0004 and the value of Squared Correlation = 0.9690.</em></p>


2021 ◽  
Vol 12 (11) ◽  
pp. 1940-1953
Author(s):  
Viratkumar K. Kothari, Et. al.

There is substantial archival data available in different forms, including manuscripts, printed papers, photographs, videos, audios, artefacts, sculptures, building, and others. Media content like photographs, audios, and videos are crucial content because such content conveys information well. The digital version of such media data is essential as it can be shared easily, available in the online or offline platform, easy to copy, easy to transport, easy to back up and easy to keep multiple copies at different places. The limitation of the digital version of media data is the lack of searchability as it hardly has any text that can be processed for OCR. These important data cannot be analysed and, therefore, cannot be used in a meaningful way. To make this data meaningful, one has to manually identify people in the images and tag them to create metadata. Most of the photographs were possible to search based on very basic metadata. This data, when hosted on the web platform, searching media data is becoming a challenge due to its data formats. Improvement in existing search functionality is required to improve the searchability of the photographs in terms of ease of usage, quick retrieval and efficiency. The recent revolution in machine learning, deep learning and artificial intelligence offers a variety of facilities to process media data and identify meaningful information out of it. This research paper explains the methods to process digital photographs to classify people in the given photographs, tag them and saves that information in the metadata. We will tune various hyperparameter to improve their accuracy. Machine learning, deep learning and artificial intelligence offers several benefits, including auto-identification of people, auto-tagging them, provide insights and finally, the most important part is it improves the searchability of photographs drastically. It was envisaged that about 85% of the manual tagging activity might be reduced and improves the searchability of photographs by 90%.


2020 ◽  
Vol 29 (2) ◽  
pp. e013
Author(s):  
İlker Ercanli

Aim of Study: As an innovative prediction technique, Artificial Intelligence technique based on a Deep Learning Algorithm (DLA) with various numbers of neurons and hidden layer alternatives were trained and evaluated to predict the relationships between total tree height (TTH) and diameter at breast height (DBH) with nonlinear least squared (NLS) regression models and nonlinear mixed effect (NLME) regression models.Area of Study: The data of this study were measured from even-aged, pure Turkish Pine (Pinus brutia Ten.) stands in the Kestel Forests located in the Bursa region of northwestern Turkey.Material and Methods: 1132 pairs of TTH-DBH measurements from 132 sample plots were used for modeling relationships between TTH, DBH, and stand attributes such as dominant height (Ho) and diameter (Do).Main Results: The combination of 100 # neurons and 8 # hidden layer in DLA resulted in the best predictive total height prediction values with Average Absolute Error (0.4188), max. Average Absolute Error (3.7598), Root Mean Squared Error (0.6942), Root Mean Squared error % (5.2164), Akaike Information Criteria (-345.4465), Bayesian Information Criterion (-330.836), the average Bias (0.0288) and the average Bias % (0.2166), and fitting abilities with r (0.9842) and Fit Index (0.9684). Also, the results of equivalence tests showed that the DLA technique successfully predicted the TTH in the validation dataset.Research highlights: These superior fitting scores coupled with the validation results in TTH predictions suggested that deep learning network models should be considered an alternative to the traditional nonlinear regression techniques and should be given importance as an innovative prediction technique.Keywords: Prediction; artificial intelligence; deep learning algorithms; number of neurons; hidden layer alternatives.Abbreviations: TTH (total tree height), DBH (diameter at breast height), OLS (ordinary least squares), NLME (nonlinear mixed effect), AIT (Artificial Intelligence Techniques), ANN (Artificial Neural Network), DLA (Deep Learning Algorithm), GPU (Graphical Processing Units), NLS (nonlinear least squared), RMSE (root mean squared error), AIC (Akaike information criteria), BIC (Bayesian information criterion), FI (fit index), AAE (average absolute error), BLUP (best linear unbiased predictor), TOST (two one-sided test method). 


2018 ◽  
Vol 15 (1) ◽  
pp. 6-28 ◽  
Author(s):  
Javier Pérez-Sianes ◽  
Horacio Pérez-Sánchez ◽  
Fernando Díaz

Background: Automated compound testing is currently the de facto standard method for drug screening, but it has not brought the great increase in the number of new drugs that was expected. Computer- aided compounds search, known as Virtual Screening, has shown the benefits to this field as a complement or even alternative to the robotic drug discovery. There are different methods and approaches to address this problem and most of them are often included in one of the main screening strategies. Machine learning, however, has established itself as a virtual screening methodology in its own right and it may grow in popularity with the new trends on artificial intelligence. Objective: This paper will attempt to provide a comprehensive and structured review that collects the most important proposals made so far in this area of research. Particular attention is given to some recent developments carried out in the machine learning field: the deep learning approach, which is pointed out as a future key player in the virtual screening landscape.


2020 ◽  
Vol 114 ◽  
pp. 242-245
Author(s):  
Jootaek Lee

The term, Artificial Intelligence (AI), has changed since it was first coined by John MacCarthy in 1956. AI, believed to have been created with Kurt Gödel's unprovable computational statements in 1931, is now called deep learning or machine learning. AI is defined as a computer machine with the ability to make predictions about the future and solve complex tasks, using algorithms. The AI algorithms are enhanced and become effective with big data capturing the present and the past while still necessarily reflecting human biases into models and equations. AI is also capable of making choices like humans, mirroring human reasoning. AI can help robots to efficiently repeat the same labor intensive procedures in factories and can analyze historic and present data efficiently through deep learning, natural language processing, and anomaly detection. Thus, AI covers a spectrum of augmented intelligence relating to prediction, autonomous intelligence relating to decision making, automated intelligence for labor robots, and assisted intelligence for data analysis.


Drones ◽  
2021 ◽  
Vol 5 (2) ◽  
pp. 37
Author(s):  
Bingsheng Wei ◽  
Martin Barczyk

We consider the problem of vision-based detection and ranging of a target UAV using the video feed from a monocular camera onboard a pursuer UAV. Our previously published work in this area employed a cascade classifier algorithm to locate the target UAV, which was found to perform poorly in complex background scenes. We thus study the replacement of the cascade classifier algorithm with newer machine learning-based object detection algorithms. Five candidate algorithms are implemented and quantitatively tested in terms of their efficiency (measured as frames per second processing rate), accuracy (measured as the root mean squared error between ground truth and detected location), and consistency (measured as mean average precision) in a variety of flight patterns, backgrounds, and test conditions. Assigning relative weights of 20%, 40% and 40% to these three criteria, we find that when flying over a white background, the top three performers are YOLO v2 (76.73 out of 100), Faster RCNN v2 (63.65 out of 100), and Tiny YOLO (59.50 out of 100), while over a realistic background, the top three performers are Faster RCNN v2 (54.35 out of 100, SSD MobileNet v1 (51.68 out of 100) and SSD Inception v2 (50.72 out of 100), leading us to recommend Faster RCNN v2 as the recommended solution. We then provide a roadmap for further work in integrating the object detector into our vision-based UAV tracking system.


2021 ◽  
Vol 54 (6) ◽  
pp. 1-35
Author(s):  
Ninareh Mehrabi ◽  
Fred Morstatter ◽  
Nripsuta Saxena ◽  
Kristina Lerman ◽  
Aram Galstyan

With the widespread use of artificial intelligence (AI) systems and applications in our everyday lives, accounting for fairness has gained significant importance in designing and engineering of such systems. AI systems can be used in many sensitive environments to make important and life-changing decisions; thus, it is crucial to ensure that these decisions do not reflect discriminatory behavior toward certain groups or populations. More recently some work has been developed in traditional machine learning and deep learning that address such challenges in different subdomains. With the commercialization of these systems, researchers are becoming more aware of the biases that these applications can contain and are attempting to address them. In this survey, we investigated different real-world applications that have shown biases in various ways, and we listed different sources of biases that can affect AI applications. We then created a taxonomy for fairness definitions that machine learning researchers have defined to avoid the existing bias in AI systems. In addition to that, we examined different domains and subdomains in AI showing what researchers have observed with regard to unfair outcomes in the state-of-the-art methods and ways they have tried to address them. There are still many future directions and solutions that can be taken to mitigate the problem of bias in AI systems. We are hoping that this survey will motivate researchers to tackle these issues in the near future by observing existing work in their respective fields.


2021 ◽  
Vol 10 (2) ◽  
pp. 205846012199029
Author(s):  
Rani Ahmad

Background The scope and productivity of artificial intelligence applications in health science and medicine, particularly in medical imaging, are rapidly progressing, with relatively recent developments in big data and deep learning and increasingly powerful computer algorithms. Accordingly, there are a number of opportunities and challenges for the radiological community. Purpose To provide review on the challenges and barriers experienced in diagnostic radiology on the basis of the key clinical applications of machine learning techniques. Material and Methods Studies published in 2010–2019 were selected that report on the efficacy of machine learning models. A single contingency table was selected for each study to report the highest accuracy of radiology professionals and machine learning algorithms, and a meta-analysis of studies was conducted based on contingency tables. Results The specificity for all the deep learning models ranged from 39% to 100%, whereas sensitivity ranged from 85% to 100%. The pooled sensitivity and specificity were 89% and 85% for the deep learning algorithms for detecting abnormalities compared to 75% and 91% for radiology experts, respectively. The pooled specificity and sensitivity for comparison between radiology professionals and deep learning algorithms were 91% and 81% for deep learning models and 85% and 73% for radiology professionals (p < 0.000), respectively. The pooled sensitivity detection was 82% for health-care professionals and 83% for deep learning algorithms (p < 0.005). Conclusion Radiomic information extracted through machine learning programs form images that may not be discernible through visual examination, thus may improve the prognostic and diagnostic value of data sets.


Sign in / Sign up

Export Citation Format

Share Document