data sequence
Recently Published Documents


TOTAL DOCUMENTS

134
(FIVE YEARS 32)

H-INDEX

10
(FIVE YEARS 1)

2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Zhou Xu ◽  
Guo Liwen ◽  
Zhang Jiuling ◽  
Qin Sijia ◽  
Zhu Yi

Accurate quantitative analysis and prediction of dust concentration in mines play a vital role in avoiding pneumoconiosis to a certain extent, improving industrial production efficiency, and protecting the ecological environment. The research has far-reaching significance for the prediction of dust concentration in mines in the future. Aiming at the shortcomings of the grey GM (1, 1) model in forecasting the data sequence with large random fluctuation, a grey Markov chain forecasting model is established. Firstly, considering the timeliness of monitoring data, the new dust concentration data is supplemented by using the method of cubic spline interpolation in the original data sequence. Therefore, the GM (1, 1) model is established by the method of metabolism. Then, the GM (1, 1) model is optimized by the theory of the Markov chain model. According to the relative error range generated during the prediction, the state interval is divided. Subsequently, the corresponding state probability transition matrix is constructed to obtain the grey Markov prediction model. The model was applied to the prediction of mine dust concentration and compared with the prediction results of the BP neural network model, grey prediction model, and ARIMA (1, 2, 1) model. The results showed that the prediction accuracy of the grey Markov model was significantly improved compared with other traditional prediction models. Therefore, the rationality and accuracy of this model in the prediction of mine dust concentration were verified.


2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Qiang Yi

Abstract The characteristics of non-Gaussian clutter in radar systems are different from standard waveforms. To fully filter to achieve the accuracy of radar detection, the paper developed a radar simulation system based on virtual reality technology. The article uses a non-Gaussian mathematical model to simulate and collect the clutter generated by the system and realise the generation of data sequence according to the power spectrum. The research results show that the radar cross-section modelling, target recognition, anti-recognition and data fusion technology of visible targets can all be well applied in this system.


2021 ◽  
Vol 13 (22) ◽  
pp. 4588
Author(s):  
Le Zhang ◽  
Anke Xue ◽  
Xiaodong Zhao ◽  
Shuwen Xu ◽  
Kecheng Mao

In this paper, an approach for radar clutter, especially sea and land clutter classification, is considered under the following conditions: the average amplitude levels of the clutter are close to each other, and the distributions of the clutter are unknown. The proposed approach divides the dataset into two parts. The first data sequence from sea and land is used to train the model to compute the parameters of the classifier, and the second data sequence from sea and land under the same conditions is used to test the performance of the algorithm. In order to find the essential structure of the data, a new data representation method based on the graph spectrum is utilized. The method reveals the nondominant correlation implied in the data, and it is quite different from the traditional method. Furthermore, this representation is combined with the support vector machine (SVM) artificial intelligence algorithm to solve the problem of sea and land clutter classification. We compare the proposed graph feature set with nine exciting valid features that have been used to classify sea clutter from the radar in other works, especially when the average amplitude levels of the two types of clutter are very close. The experimental results prove that the proposed extraction can represent the characteristics of the raw data efficiently in this application.


2021 ◽  
Vol 7 ◽  
pp. e752
Author(s):  
Omar Bin Samin ◽  
Maryam Omar ◽  
Musadaq Mansoor

Accurate disease classification in plants is important for a profound understanding of their growth and health. Recognizing diseases in plants from images is one of the critical and challenging problem in agriculture. In this research, a deep learning architecture model (CapPlant) is proposed that utilizes plant images to predict whether it is healthy or contain some disease. The prediction process does not require handcrafted features; rather, the representations are automatically extracted from input data sequence by architecture. Several convolutional layers are applied to extract and classify features accordingly. The last convolutional layer in CapPlant is replaced by state-of-the-art capsule layer to incorporate orientational and relative spatial relationship between different entities of a plant in an image to predict diseases more precisely. The proposed architecture is tested on the PlantVillage dataset, which contains more than 50,000 images of infected and healthy plants. Significant improvements in terms of prediction accuracy has been observed using the CapPlant model when compared with other plant disease classification models. The experimental results on the developed model have achieved an overall test accuracy of 93.01%, with F1 score of 93.07%.


2021 ◽  
pp. 265-306
Author(s):  
Magy Seif El-Nasr ◽  
Truong Huy Nguyen Dinh ◽  
Alessandro Canossa ◽  
Anders Drachen

This chapter is devoted to sequence game data analysis. It will first define what sequence data is and how it is represented, and then delve more deeply into how to develop models from such data. Sequence analysis has a lot of utility and is important as it conserves the sequence of player actions and can shed light on how players solved different problems within the different game levels. Further, sequence analysis can also be a great way to develop a more robust and accurate player model. The chapter will discuss such advantages in light of showcasing the use of sequential analysis for DOTA 2. Further, the chapter will also be a practical guide on how to develop models from sequence data using practical step-by-step labs. Please note that this chapter was written with Erica Kleinman (a PhD student at University of California at Santa Cruz).


Petir ◽  
2021 ◽  
Vol 14 (2) ◽  
pp. 159-169
Author(s):  
Endang Sunandar

There are various kinds of data sorting methods that we know of which are the Bubble Sort, Selection Sort, Insertion Sort, Quick Sort, Shell Sort, Heap Sort, and Radix Sort methods. All of these methods have advantages and disadvantages of each, whose use is determined based on needs. Each method has a different algorithm, where different algorithms affect the execution time. One interesting algorithm to be implemented on 2 variant models of data sorting is the Bubble Sort algorithm, the reason is that this algorithm has a fairly long and detailed process flow to produce an ordered data sequence from a previously unordered data sequence. Two (2) data sorting variant models that will be implemented using the Bubble Sort algorithm are: Ascending data sorting variants moving from left to right, and Descending data sorting variants moving from left to right. And the device used in implementing the Bubble Sort algorithm is the Java programming language.


Author(s):  
Yu-Ting Chen ◽  
Jeng-Min Chiou ◽  
Tzee-Ming Huang

Entropy ◽  
2021 ◽  
Vol 23 (8) ◽  
pp. 997
Author(s):  
Pham Thuc Hung ◽  
Kenji Yamanishi

In this paper, we propose a novel information criteria-based approach to select the dimensionality of the word2vec Skip-gram (SG). From the perspective of the probability theory, SG is considered as an implicit probability distribution estimation under the assumption that there exists a true contextual distribution among words. Therefore, we apply information criteria with the aim of selecting the best dimensionality so that the corresponding model can be as close as possible to the true distribution. We examine the following information criteria for the dimensionality selection problem: the Akaike’s Information Criterion (AIC), Bayesian Information Criterion (BIC), and Sequential Normalized Maximum Likelihood (SNML) criterion. SNML is the total codelength required for the sequential encoding of a data sequence on the basis of the minimum description length. The proposed approach is applied to both the original SG model and the SG Negative Sampling model to clarify the idea of using information criteria. Additionally, as the original SNML suffers from computational disadvantages, we introduce novel heuristics for its efficient computation. Moreover, we empirically demonstrate that SNML outperforms both BIC and AIC. In comparison with other evaluation methods for word embedding, the dimensionality selected by SNML is significantly closer to the optimal dimensionality obtained by word analogy or word similarity tasks.


2021 ◽  
Author(s):  
Xiaoli Li ◽  
Sridhar Krishnan ◽  
Ngok-Wah Ma

A newly developed grammar-based lossless source coding theory and its implementation was proposed in 1999 and 2000, respectively, by Yang and Kieffer. The code first transforms the original data sequence into an irreducible context-free grammar, which is then compressed using arithmetic coding. In the study of grammar-based coding for mammography applications, we encountered two issues: processing time and limited number of single-character grammar G variables. For the first issue, we discover a feature that can simplify the matching subsequence search in the irreducible grammar transform process. Using this discovery, an extended grammar code technique is proposed and the processing time of the grammar code can be significantly reduced. For the second issue, we propose to use double-character symbols to increase the number of grammar variables. Under the condition that all the G variables have the same probability of being used, our analysis shows that the double- and single-character approaches have the same compression rates. By using the methods proposed, we show that the grammar code can outperform three other schemes: Lempel-Ziv-Welch (LZW), arithmetic, and Huffman on compression ratio, and has similar error tolerance capabilities as LZW coding under similar circumstances.


2021 ◽  
Author(s):  
Xiaoli Li ◽  
Sridhar Krishnan ◽  
Ngok-Wah Ma

A newly developed grammar-based lossless source coding theory and its implementation was proposed in 1999 and 2000, respectively, by Yang and Kieffer. The code first transforms the original data sequence into an irreducible context-free grammar, which is then compressed using arithmetic coding. In the study of grammar-based coding for mammography applications, we encountered two issues: processing time and limited number of single-character grammar G variables. For the first issue, we discover a feature that can simplify the matching subsequence search in the irreducible grammar transform process. Using this discovery, an extended grammar code technique is proposed and the processing time of the grammar code can be significantly reduced. For the second issue, we propose to use double-character symbols to increase the number of grammar variables. Under the condition that all the G variables have the same probability of being used, our analysis shows that the double- and single-character approaches have the same compression rates. By using the methods proposed, we show that the grammar code can outperform three other schemes: Lempel-Ziv-Welch (LZW), arithmetic, and Huffman on compression ratio, and has similar error tolerance capabilities as LZW coding under similar circumstances.


Sign in / Sign up

Export Citation Format

Share Document