scholarly journals Rule Induction Algorithms Applied in Educational Data Set

Author(s):  
Walisson Carvalho ◽  
Cristiane Nobre ◽  
Luis Zarate
Keyword(s):  
Author(s):  
KAPIL SHARMA ◽  
SHEVETA VASHISHT

In this research work we use rule induction in data mining to obtain the accurate results with fast processing time. We using decision list induction algorithm to make order and unordered list of rules to coverage of maximum data from the data set. Using induction rule via association rule mining we can generate number of rules for training dataset to achieve accurate result with less error rate. We also use induction rule algorithms like confidence static and Shannon entropy to obtain the high rate of accurate results from the large dataset. This can also improves the traditional algorithms with good result.


2019 ◽  
Vol 6 (3) ◽  
pp. 341
Author(s):  
I Putu Pradnyana Iswara ◽  
Ahmad Afif Supianto

Pada penelitian ini kami mengimplementasikan algoritma klasifikasi untuk memberikan rekomendasi kepada mahasiswa keminatan apa yang lebih cocok diambil berdasarkan nilai-nilai mata kuliah prasyarat di semester-semester sebelumnya. Diharapkan dengan adanya rekomendasi ini semakin jelas pembatas antara disiplin ilmu yang ada pada Program Studi Sistem Informasi Universitas Brawijaya dimana terdapat 3 jenis jalur keminatan mata kuliah pilihan yaitu <em>Database</em>, Logika &amp; pemrograman dan Manajemen SI/TI. Data set yang terdiri dari data <em>training</em> dan data testing merupakan data akademik dari mahasiswa angkatan 2015 yang sudah mengambil mata kuliah pilihan, data target dari penelitian ini adalah data akademik mahasiswa angkatan 2016. Algoritma klasifikasi yang digunakan adalah <em>Rule Induction, CHAID, Random Forest, ID3, </em> dan <em>Naive Bayes</em>. Komposisi dari data <em>training</em> dan <em>testing</em> diubah-ubah untuk mengetahui pengaruh perubahan komposisi tersebut. Kelima algoritma tersebut diuji sebanyak 5 kali. Dari seluruh hasil pengujian didapatkan rata-rata akurasi dari kelima metode yang diusulkan berturut-turut adalah 66,48%, 67,49%, 80,62%, 86,90% dan 77,68%. Hasil tersebut menunjukkan bahwa algoritma dengan rata-rata akurasi tertinggi dimiliki oleh algoritma <em>ID3 </em>dikarenakan algoritmanya yang fleksibel dan dapat lebih akurat untuk menguji data yang digunakan<em>.</em>


2013 ◽  
Vol 4 (1) ◽  
pp. 135-145
Author(s):  
Kapil Sharma ◽  
Sheveta Vashisht ◽  
Heena Sharma ◽  
Jasreena kaur Bains ◽  
Richa Dhiman

Data Mining: extracting useful insights from large and detailed collections of data. With the increased possibilities in modern society for companies and institutions to gather data cheaply and efficiently, this subject has become of increasing importance. This interest has inspired a rapidly maturing research field with developments both on a theoretical, as well as on a practical level with the availability of a range of commercial tools. In this research work we use rule induction in data mining to obtain the accurate results with fast processing time. We using decision list induction algorithm to make order and unordered list of rules to coverage of maximum data from the data set. Using induction rule via association rule mining we can generate number of rules for training dataset to achieve accurate result with less error rate. We also use induction rule algorithms like confidence static and Shannon entropy to obtain the high rate of accurate results from the large dataset. This can also improves the traditional algorithms with good result.


Author(s):  
Marek Sikora ◽  
Beata Sikora

Improving prediction models applied in systems monitoring natural hazards and machineryA method of combining three analytic techniques including regression rule induction, thek-nearest neighbors method and time series forecasting by means of the ARIMA methodology is presented. A decrease in the forecasting error while solving problems that concern natural hazards and machinery monitoring in coal mines was the main objective of the combined application of these techniques. The M5 algorithm was applied as a basic method of developing prediction models. In spite of an intensive development of regression rule induction algorithms and fuzzy-neural systems, the M5 algorithm is still characterized by the generalization ability and unbeatable time of data model creation competitive with other systems. In the paper, two solutions designed to decrease the mean square error of the obtained rules are presented. One consists in introducing into a set of conditional variables the so-called meta-variable (an analogy to constructive induction) whose values are determined by an autoregressive or the ARIMA model. The other shows that limitation of a data set on which the M5 algorithm operates by thek-nearest neighbor method can also lead to error decreasing. Moreover, three application examples of the presented solutions for data collected by systems of natural hazards and machinery monitoring in coal mines are described. In Appendix, results of several benchmark data sets analyses are given as a supplement of the presented results.


1994 ◽  
Vol 144 ◽  
pp. 139-141 ◽  
Author(s):  
J. Rybák ◽  
V. Rušin ◽  
M. Rybanský

AbstractFe XIV 530.3 nm coronal emission line observations have been used for the estimation of the green solar corona rotation. A homogeneous data set, created from measurements of the world-wide coronagraphic network, has been examined with a help of correlation analysis to reveal the averaged synodic rotation period as a function of latitude and time over the epoch from 1947 to 1991.The values of the synodic rotation period obtained for this epoch for the whole range of latitudes and a latitude band ±30° are 27.52±0.12 days and 26.95±0.21 days, resp. A differential rotation of green solar corona, with local period maxima around ±60° and minimum of the rotation period at the equator, was confirmed. No clear cyclic variation of the rotation has been found for examinated epoch but some monotonic trends for some time intervals are presented.A detailed investigation of the original data and their correlation functions has shown that an existence of sufficiently reliable tracers is not evident for the whole set of examinated data. This should be taken into account in future more precise estimations of the green corona rotation period.


Author(s):  
Jules S. Jaffe ◽  
Robert M. Glaeser

Although difference Fourier techniques are standard in X-ray crystallography it has only been very recently that electron crystallographers have been able to take advantage of this method. We have combined a high resolution data set for frozen glucose embedded Purple Membrane (PM) with a data set collected from PM prepared in the frozen hydrated state in order to visualize any differences in structure due to the different methods of preparation. The increased contrast between protein-ice versus protein-glucose may prove to be an advantage of the frozen hydrated technique for visualizing those parts of bacteriorhodopsin that are embedded in glucose. In addition, surface groups of the protein may be disordered in glucose and ordered in the frozen state. The sensitivity of the difference Fourier technique to small changes in structure provides an ideal method for testing this hypothesis.


Author(s):  
D. E. Becker

An efficient, robust, and widely-applicable technique is presented for computational synthesis of high-resolution, wide-area images of a specimen from a series of overlapping partial views. This technique can also be used to combine the results of various forms of image analysis, such as segmentation, automated cell counting, deblurring, and neuron tracing, to generate representations that are equivalent to processing the large wide-area image, rather than the individual partial views. This can be a first step towards quantitation of the higher-level tissue architecture. The computational approach overcomes mechanical limitations, such as hysterisis and backlash, of microscope stages. It also automates a procedure that is currently done manually. One application is the high-resolution visualization and/or quantitation of large batches of specimens that are much wider than the field of view of the microscope.The automated montage synthesis begins by computing a concise set of landmark points for each partial view. The type of landmarks used can vary greatly depending on the images of interest. In many cases, image analysis performed on each data set can provide useful landmarks. Even when no such “natural” landmarks are available, image processing can often provide useful landmarks.


Author(s):  
Jaap Brink ◽  
Wah Chiu

Crotoxin complex is the principal neurotoxin of the South American rattlesnake, Crotalus durissus terrificus and has a molecular weight of 24 kDa. The protein is a heterodimer with subunit A assigneda chaperone function. Subunit B carries the lethal activity, which is exerted on both sides ofthe neuro-muscular junction, and which is thought to involve binding to the acetylcholine receptor. Insight in crotoxin complex’ mode of action can be gained from a 3 Å resolution structure obtained by electron crystallography. This abstract communicates our progress in merging the electron diffraction amplitudes into a 3-dimensional (3D) intensity data set close to completion. Since the thickness of crotoxin complex crystals varies from one crystal to the other, we chose to collect tilt series of electron diffraction patterns after determining their thickness. Furthermore, by making use of the symmetry present in these tilt data, intensities collected only from similar crystals will be merged.Suitable crystals of glucose-embedded crotoxin complex were searched for in the defocussed diffraction mode with the goniometer tilted to 55° of higher in a JEOL4000 electron cryo-microscopc operated at 400 kV with the crystals kept at -120°C in a Gatan 626 cryo-holder. The crystal thickness was measured using the local contrast of the crystal relative to the supporting film from search-mode images acquired using a 1024 x 1024 slow-scan CCD camera (model 679, Gatan Inc.).


Author(s):  
J. K. Samarabandu ◽  
R. Acharya ◽  
D. R. Pareddy ◽  
P. C. Cheng

In the study of cell organization in a maize meristem, direct viewing of confocal optical sections in 3D (by means of 3D projection of the volumetric data set, Figure 1) becomes very difficult and confusing because of the large number of nucleus involved. Numerical description of the cellular organization (e.g. position, size and orientation of each structure) and computer graphic presentation are some of the solutions to effectively study the structure of such a complex system. An attempt at data-reduction by means of manually contouring cell nucleus in 3D was reported (Summers et al., 1990). Apart from being labour intensive, this 3D digitization technique suffers from the inaccuracies of manual 3D tracing related to the depth perception of the operator. However, it does demonstrate that reducing stack of confocal images to a 3D graphic representation helps to visualize and analyze complex tissues (Figure 2). This procedure also significantly reduce computational burden in an interactive operation.


Author(s):  
M. Shlepr ◽  
C. M. Vicroy

The microelectronics industry is heavily tasked with minimizing contaminates at all steps of the manufacturing process. Particles are generated by physical and/or chemical fragmentation from a mothersource. The tools and macrovolumes of chemicals used for processing, the environment surrounding the process, and the circuits themselves are all potential particle sources. A first step in eliminating these contaminants is to identify their source. Elemental analysis of the particles often proves useful toward this goal, and energy dispersive spectroscopy (EDS) is a commonly used technique. However, the large variety of source materials and process induced changes in the particles often make it difficult to discern if the particles are from a common source.Ordination is commonly used in ecology to understand community relationships. This technique usespair-wise measures of similarity. Separation of the data set is based on discrimination functions. Theend product is a spatial representation of the data with the distance between points equaling the degree of dissimilarity.


Sign in / Sign up

Export Citation Format

Share Document