scholarly journals Data Mining to Select Relevant Variables Influencing External and Internal Workload of Elite Blind 5-a-Side Soccer

Author(s):  
José M. Gamonales ◽  
Kiko León ◽  
Daniel Rojas-Valverde ◽  
Braulio Sánchez-Ureña ◽  
Jesús Muñoz-Jiménez

(1) Background: Data mining has turned essential when exploring a large amount of information in performance analysis in sports. This study aimed to select the most relevant variables influencing the external and internal load in top-elite 5-a-side soccer (Sa5) using a data mining model considering some contextual indicators as match result, body mass index (BMI), scoring rate and age. (2) Methods: A total of 50 top-elite visually impaired soccer players (age 30.86 ± 11.2 years, weight 77.64 ± 9.78 kg, height 178.48 ± 7.9 cm) were monitored using magnetic, angular and rate gyroscope (MARG) sensors during an international Sa5 congested fixture tournament.; (3) Results: Fifteen external and internal load variables were extracted from a total of 49 time-related and peak variables derived from the MARG sensors using a principal component analysis as the most used data mining technique. The principal component analysis (PCA) model explained 80% of total variance using seven principal components. In contrast, the first principal component of the match was defined by jumps, take off by 24.8% of the total variance. Blind players usually performed a higher number of accelerations per min when losing a match. Scoring players execute higher DistanceExplosive and Distance21–24 km/h. And the younger players presented higher HRAVG and AccMax. (4) Conclusions: The influence of some contextual variables on external and internal load during top elite Sa5 official matches should be addressed by coaches, athletes, and medical staff. The PCA seems to be a useful statistical technique to select those relevant variables representing the team’s external and internal load. Besides, as a data reduction method, PCA allows administrating individualized training loads considering those relevant variables defining team load behavior.

1990 ◽  
Vol 55 (1) ◽  
pp. 55-62 ◽  
Author(s):  
Drahomír Hnyk

The principal component analysis has been applied to a data matrix formed by 7 usual substituent constants for 38 substituents. Three factors are able to explain 99.4% cumulative proportion of total variance. Several rotations have been carried out for the first two factors in order to obtain their physical meaning. The first factor is related to the resonance effect, whereas the second one expresses the inductive effect, and both together describe 97.5% cumulative proportion of total variance. Their mutual orthogonality does not directly follow from the rotations carried out. With the help of these factors the substituents are divided into four main classes, and some of them assume a special position.


2009 ◽  
Vol 147-149 ◽  
pp. 588-593 ◽  
Author(s):  
Marcin Derlatka ◽  
Jolanta Pauk

In the paper the procedure of processing biomechanical data has been proposed. It consists of selecting proper noiseless data, preprocessing data by means of model’s identification and Kernel Principal Component Analysis and next classification using decision tree. The obtained results of classification into groups (normal and two selected pathology of gait: Spina Bifida and Cerebral Palsy) were very good.


2021 ◽  
pp. 141-146
Author(s):  
Carlo Cusatelli ◽  
Massimiliano Giacalone ◽  
Eugenia Nissi

Well being is a multidimensional phenomenon, that cannot be measured by a single descriptive indicator and that, it should be represented by multiple dimensions. It requires, to be measured by combination of different dimensions that can be considered together as components of the phenomenon. This combination can be obtained by applying methodologies knows as Composite Indicators (CIs). CIs are largely used to have a comprehensive view on a phenomenon that cannot be captured by a single indicator. Principal Component Analysis (PCA) is one of the most popular multivariate statistical technique used for reducing data with many dimension, and often well being indicators are obtained using PCA. PCA is implicitly based on a reflective measurement model that it non suitable for all types of indicators. Mazziotta and Pareto (2013) in their paper discuss the use and misuse of PCA for measuring well-being. The classical PCA is not suitable for data collected on the territory because it does not take into account the spatial autocorrelation present in the data. The aim of this paper is to propose the use of Spatial Principal Component Analysis for measuring well being in the Italian Provinces.


Author(s):  
Yanwen Wang ◽  
Javad Garjami ◽  
Milena Tsvetkova ◽  
Nguyen Huu Hau ◽  
Kim-Hung Pho

Abstract Data mining, statistics, and data analysis are popular techniques to study datasets and extract knowledge from them. In this article, principal component analysis and factor analysis were applied to cluster thirteen different given arrangements about the Suras of the Holy Quran. The results showed that these thirteen arrangements can be categorized in two parts such that the first part includes Blachère, Davood, Grimm, Nöldeke, Bazargan, E’temad-al-Saltane and Muir, and the second part includes Ebn Nadim, Jaber, Ebn Abbas, Hazrat Ali, Khazan, and Al-Azhar.


Author(s):  
Zuhaira Muhammad Zain ◽  
Mona Alshenaifi ◽  
Abeer Aljaloud ◽  
Tamadhur Albednah ◽  
Reham Alghanim ◽  
...  

Breast cancer recurrence is among the most noteworthy fears faced by women. Nevertheless, with modern innovations in data mining technology, early recurrence prediction can help relieve these fears. Although medical information is typically complicated, and simplifying searches to the most relevant input is challenging, new sophisticated data mining techniques promise accurate predictions from high-dimensional data. In this study, the performances of three established data mining algorithms: Naïve Bayes (NB), k-nearest neighbor (KNN), and fast decision tree (REPTree), adopting the feature extraction algorithm, principal component analysis (PCA), for predicting breast cancer recurrence were contrasted. The comparison was conducted between models built in the absence and presence of PCA. The results showed that KNN produced better prediction without PCA (F-measure = 72.1%), whereas the other two techniques: NB and REPTree, improved when used with PCA (F-measure = 76.1% and 72.8%, respectively). This study can benefit the healthcare industry in assisting physicians in predicting breast cancer recurrence precisely.


2019 ◽  
Vol 8 (2) ◽  
pp. 32-39
Author(s):  
T. Mylsami ◽  
B. L. Shivakumar

In general the World Wide Web become the most useful information resource used for information retrievals and knowledge discoveries. But the Information on Web to be expand in size and density. The retrieval of the required information on the web is efficiently and effectively to be challenge one. For the tremendous growth of the web has created challenges for the search engine technology. Web mining is an area in which applies data mining techniques to deal the requirements. The following are the popular Web Mining algorithms, such as PageRanking (PR), Weighted PageRanking (WPR) and Hyperlink-Induced Topic Search (HITS), are quite commonly used algorithm to sort out and rank the search results. In among the page ranking algorithm uses web structure mining and web content mining to estimate the relevancy of a web site and not to deal the scalability problem and also visits of inlinks and outlinks of the pages. In recent days to access fast and efficient page ranking algorithm for webpage retrieval remains as a challenging. This paper proposed a new improved WPR algorithm which uses a Principal Component Analysis technique called (PWPR) based on mean value of page ranks. The proposed PWPR algorithm takes into account the importance of both the number of visits of inlinks and outlinks of the pages and distributes rank scores based on the popularity of the pages. The weight values of the pages is computed from the inlinks and outlinks with their mean values. But in PWPR method new data and updates are constantly arriving, the results of data mining applications become stale and obsolete over time. To solve this problem is a MapReduce (MR) framework is promising approach to refreshing mining results for mining big data .The proposed MR algorithm reduces the time complexity of the PWPR algorithm by reducing the number of iterations to reach a convergence point.


Sign in / Sign up

Export Citation Format

Share Document