scholarly journals The Application Of Fuzzy K-Nearest Neighbour Methods for A Student Graduation Rate

Author(s):  
Imam Ahmad ◽  
Heni Sulistiani ◽  
Hendrik Saputra

The absence of prediction system that can provide prediction analysis on the graduation rate of students becomes the reason for the research on the prediction of the level of graduation rate of students. Determining predictions of graduation rates of students in large numbers is not possible to do manually because it takes a long time. For that we need an algorithm that can categorize predictions of students' graduation rates in computing. The Fuzzy Method and KNN or K-Nearest Neighbor Methods are selected as the algorithm for the prediction process. In this study using 10 criteria as a material to predict students' graduation rate consisting of: NPM, Student Name, Semester 1 achievement index, Semester 2 achievement index, Semester 3 achievement index, Semester 4 achievement index, SPMB, origin SMA, Gender , and Study Period. Fuzzyfication process aims to change the value of the first semester achievement index until the fourth semester achievement index into three sets of fuzzy values are satisfactory, very satisfying, and cum laude. Make predictions to improve the quality of students and implement KNN method into prediction, where there are some attributes that have preprocess data so that obtained a value, and the value is compared with training data, so as to produce predictions of graduating students will be on time and graduating students will be late. This study produces a prediction of student pass rate and accuracy.

Author(s):  
Mohammad Imron ◽  
Satia Angga Kusumah

The student graduation rate is one of the indicators to improve the accreditation of a course. It is needed to monitor and evaluate student graduation tendencies, timely or not. One of them is to predict the graduation rate by utilizing the data mining technique. Data Mining Classification method used is the algorithm K-Nearest Neighbor (K-NN). The data used comes from student data, student value data, and student graduation data for the year 2010-2012 with a total of 2,189 records. The attributes used are gender, school of origin, IP study program Semester 1-6. The results showed that the K-NN method produced a high accuracy of 89.04%.


SinkrOn ◽  
2020 ◽  
Vol 4 (2) ◽  
pp. 42
Author(s):  
Rizki Muliono ◽  
Juanda Hakim Lubis ◽  
Nurul Khairina

Higher education plays a major role in improving the quality of education in Indonesia. The BAN-PT institution established by the government has a standard of higher education accreditation and study program accreditation. With the 4.0-based accreditation instrument, it encourages university leaders to improve the quality and quality of their education. One indicator that determines the accreditation of study programs is the timely graduation of students. This study uses the K-Nearest Neighbor algorithm to predict student graduation times. Students' GPA at the time of the seventh semester will be used as training data, and data of students who graduate are used as sample data. K-Nearest Neighbor works in accordance with the given sample data. The results of prediction testing on 60 data for students of 2015-2016, obtained the highest level of accuracy of 98.5% can be achieved when k = 3. Prediction results depend on the pattern of data entered, the more samples and training data used, the calculation of the K-Nearest Neighbor algorithm is also more accurate.


Respati ◽  
2018 ◽  
Vol 13 (2) ◽  
Author(s):  
Eri Sasmita Susanto ◽  
Kusrini Kusrini ◽  
Hanif Al Fatta

INTISARIPenelitian ini difokuskan untuk mengetahui uji kelayakan prediksi kelulusan mahasiswa Universitas AMIKOM Yogyakarta. Dalam hal ini penulis memilih algoritma K-Nearest Neighbors (K-NN) karena K-Nearest Neighbors (K-NN) merupakan algoritma  yang bisa digunakan untuk mengolah data yang bersifat numerik dan tidak membutuhkan skema estimasi parameter perulangan yang rumit, ini berarti bisa diaplikasikan untuk dataset berukuran besar.Input dari sistem ini adalah Data sampel berupa data mahasiswa tahun 2014-2015. pengujian pada penelitian ini menggunakn dua pengujian yaitu data testing dan data training. Kriteria yang digunakan dalam penelitian ini adalah , IP Semester 1-4, capaian SKS, Status Kelulusan. Output dari sistem ini berupa hasil prediksi kelulusan mahasiswa yang terbagi menjadi dua yaitu tepat waktu dan kelulusan tidak tepat waktu.Hasil pengujian menunjukkan bahwa Berdasarkan penerapan k=14 dan k-fold=5 menghasilkan performa yang terbaik dalam memprediksi kelulusan mahasiswa dengan metode K-Nearest Neighbor menggunakan indeks prestasi 4 semester dengan nilai akurasi= 98,46%, precision= 99.53% dan recall =97.64%.Kata kunci: Algoritma K-Nearest Neighbors, Prediksi Kelulusan, Data Testing, Data Training ABSTRACTThis research is focused on knowing the feasibility test of students' graduation prediction of AMIKOM University Yogyakarta. In this case the authors chose the K-Nearest Neighbors (K-NN) algorithm because K-Nearest Neighbors (K-NN) is an algorithm that can be used to process data that is numerical and does not require complicated repetitive parameter estimation scheme, this means it can be applied for large datasets.The input of this system is the sample data in the form of student data from 2014-2015. test in this research use two test that is data testing and training data. The criteria used in this study are, IP Semester 1-4, achievement of SKS, Graduation Status. The output of this system in the form of predicted results of student graduation which is divided into two that is timely and graduation is not timely.The result of the test shows that based on the application of k = 14 and k-fold = 5, the best performance in predicting the students' graduation using K-Nearest Neighbor method uses 4 semester achievement index with accuracy value = 98,46%, precision = 99.53% and recall = 97.64%.Keywords: K-Nearest Neighbors Algorithm, Graduation Prediction, Testing Data, Training Data


2020 ◽  
Vol 1 (1) ◽  
pp. 27-32
Author(s):  
Yuri Efenie ◽  
Miftahul Walid

In this research, trying to predict the salinity of sea water using the K-Nearest Neighbor method, this method serves to clarify the input data using the distance measurement method with training data, the variable used in this study is the value of the location of coordinates (latitude and longitude) and the output is in the form of salinity, the case study in this study is the southern waters of Sumenep, the system has been able to make an estimate but with an error rate of 1.00 so that there is a need for re-analysis because the data used is only small, the need for additional data so that the results will be more optimal, it is also necessary to experiment with changing methods or simplifying rules or by adding input variables in the system that have been created so that it produces better accuracy values, because the existing system still requires a long time in estimating.


Author(s):  
Sumarlin Sumarlin ◽  
Dewi Anggraini

Data on graduate students is an important part in determining the quality of a private and public university. Graduate data is included in important assessments in the accreditation process. Data from Uyelindo Kupang STIKOM graduates every year will continue to grow and accumulate like neglected data because it is rarely used. To maximize student data into information that can be used by universities, the data must be processed in this case used as training data in a study using data mining to obtain information in the form of predictions of graduation from Kupang Uyelindo STIKOM students. The method used in this study is K-Nearest Neighbor using rapidminer software to measure K-Nearest Neighbor's accuracy against student graduate data. The criteria used were in the form of student names, gender, cumulative achievement index (GPA) from semester 1 to 6. In applying the K-Nearest Neighbor algorithm can be used to produce predictions of student graduation. To measure the performance of the k-nearest neighbor algorithm, the Cross Validation, Confusion Matrix and ROC Curves methods are used, in this study using a 5-fold cross validation to predict student graduation. From 100 student dataset records Uyelindo Kupang STIKOM graduates obtained accuracy rate reached 82% and included a very good classification because it has an AUC value between 0.90-1.00, which is 0.971, so it can be concluded that the accuracy of testing of student graduation models using K-Nearest Neighbor (K-NN) algorithm is influenced by the number of data clusters. Accuracy and the highest AUC value of 5-fold validation is to cluster data k = 4 with the accuracy value of 90%.


Author(s):  
Tikaridha Hardiani

The students of Universitas ‘Aisyiyah Yogyakarta have been increasing including the number of students in the Faculty of Health Sciences. In 2016 the total number of UNISA students was 1851. The increasing number of students every year leads to great numbers of data stored in the university database. The data provide useful information for the university to predict student graduation or student study period whether they graduate on time with a study period of 4 years or late with a study period of more than 4 years. This can be processed by using a data mining technique that is the classification technique. Data needed in the classification technique are data of students who have graduated as training data and data of students who are still studying in the university as testing data. The training data were 501 records with 10 goals and the testing data were 428 records. Data mining process method used was the Cross-Industry Standard Prosses for Data Mining (CRISPDM). The algorithms used in this study were Naive Bayes, K-Nearest Neighbor (KNN) and Decision Tree. The three algorithms were compared to see the accuracy by using Rapidminer software. Based on the accuracy, it was found that the K-NN algorithm was the best in predicting student graduation with an accuracy of 91.82%. The K-NN algorithm showed that 100% of the students of Nursing study program of Universitas Aisyiyah Yogyakarta are predicted to graduate on time.


Mathematics ◽  
2021 ◽  
Vol 9 (8) ◽  
pp. 830
Author(s):  
Seokho Kang

k-nearest neighbor (kNN) is a widely used learning algorithm for supervised learning tasks. In practice, the main challenge when using kNN is its high sensitivity to its hyperparameter setting, including the number of nearest neighbors k, the distance function, and the weighting function. To improve the robustness to hyperparameters, this study presents a novel kNN learning method based on a graph neural network, named kNNGNN. Given training data, the method learns a task-specific kNN rule in an end-to-end fashion by means of a graph neural network that takes the kNN graph of an instance to predict the label of the instance. The distance and weighting functions are implicitly embedded within the graph neural network. For a query instance, the prediction is obtained by performing a kNN search from the training data to create a kNN graph and passing it through the graph neural network. The effectiveness of the proposed method is demonstrated using various benchmark datasets for classification and regression tasks.


Author(s):  
Aldi Nugroho ◽  
Osvaldo Richie Riady ◽  
Alexander Calvin ◽  
Derwin Suhartono

Students are an important asset in the world of education also an institution and therefore also need to pay attention to students' graduation rates on time. The ups and downs of the percentage of students' abilities in classroom learning is one important element for assessing university accreditation. Therefore, it is necessary to monitor and evaluate teaching and learning activities using the KNN Algorithm classification. By processing student complaints data and seeing the results of previous learning can obtain important things for higher education needs. In predicting graduation rates based on complaints, this study uses the K-Nearest Neighbor classification algorithm by grouping data k = 1, k = 2, k = 3 with the smallest value possible. In experiments using the KNN method the results were clearly visible and showed quite good accuracy. From the experiment it was concluded that if there were fewer complaints from one student it could minimize the level of student non-graduates at the university and ultimately produce good accreditation.


Machine Learning is empowering many aspects of day-to-day lives from filtering the content on social networks to suggestions of products that we may be looking for. This technology focuses on taking objects as image input to find new observations or show items based on user interest. The major discussion here is the Machine Learning techniques where we use supervised learning where the computer learns by the input data/training data and predict result based on experience. We also discuss the machine learning algorithms: Naïve Bayes Classifier, K-Nearest Neighbor, Random Forest, Decision Tress, Boosted Trees, Support Vector Machine, and use these classifiers on a dataset Malgenome and Drebin which are the Android Malware Dataset. Android is an operating system that is gaining popularity these days and with a rise in demand of these devices the rise in Android Malware. The traditional techniques methods which were used to detect malware was unable to detect unknown applications. We have run this dataset on different machine learning classifiers and have recorded the results. The experiment result provides a comparative analysis that is based on performance, accuracy, and cost.


2021 ◽  
Vol 87 (6) ◽  
pp. 445-455
Author(s):  
Yi Ma ◽  
Zezhong Zheng ◽  
Yutang Ma ◽  
Mingcang Zhu ◽  
Ran Huang ◽  
...  

Many manifold learning algorithms conduct an eigen vector analysis on a data-similarity matrix with a size of N×N, where N is the number of data points. Thus, the memory complexity of the analysis is no less than O(N2). We pres- ent in this article an incremental manifold learning approach to handle large hyperspectral data sets for land use identification. In our method, the number of dimensions for the high-dimensional hyperspectral-image data set is obtained with the training data set. A local curvature varia- tion algorithm is utilized to sample a subset of data points as landmarks. Then a manifold skeleton is identified based on the landmarks. Our method is validated on three AVIRIS hyperspectral data sets, outperforming the comparison algorithms with a k–nearest-neighbor classifier and achieving the second best performance with support vector machine.


Sign in / Sign up

Export Citation Format

Share Document