scholarly journals Comparing End-to-End Machine Learning Methods for Spectra Classification

2021 ◽  
Vol 11 (23) ◽  
pp. 11520
Author(s):  
Yue Sun ◽  
Sandor Brockhauser ◽  
Péter Hegedűs

In scientific research, spectroscopy and diffraction experimental techniques are widely used and produce huge amounts of spectral data. Learning patterns from spectra is critical during these experiments. This provides immediate feedback on the actual status of the experiment (e.g., time-resolved status of the sample), which helps guide the experiment. The two major spectral changes what we aim to capture are either the change in intensity distribution (e.g., drop or appearance) of peaks at certain locations, or the shift of those on the spectrum. This study aims to develop deep learning (DL) classification frameworks for one-dimensional (1D) spectral time series. In this work, we deal with the spectra classification problem from two different perspectives, one is a general two-dimensional (2D) space segmentation problem, and the other is a common 1D time series classification problem. We focused on the two proposed classification models under these two settings, the namely the end-to-end binned Fully Connected Neural Network (FCNN) with the automatically capturing weighting factors model and the convolutional SCT attention model. Under the setting of 1D time series classification, several other end-to-end structures based on FCNN, Convolutional Neural Network (CNN), ResNets, Long Short-Term Memory (LSTM), and Transformer were explored. Finally, we evaluated and compared the performance of these classification models based on the High Energy Density (HED) spectra dataset from multiple perspectives, and further performed the feature importance analysis to explore their interpretability. The results show that all the applied models can achieve 100% classification confidence, but the models applied under the 1D time series classification setting are superior. Among them, Transformer-based methods consume the least training time (0.449 s). Our proposed convolutional Spatial-Channel-Temporal (SCT) attention model uses 1.269 s, but its self-attention mechanism performed across spatial, channel, and temporal dimensions can suppress indistinguishable features better than others, and selectively focus on obvious features with high separability.

Symmetry ◽  
2021 ◽  
Vol 13 (7) ◽  
pp. 1186
Author(s):  
Yixiao Zhang ◽  
Ying Lei

Structural monitoring provides valuable information on the state of structural health, which is helpful for structural damage detection and structural state assessment. However, when the sensors are exposed to harsh environmental conditions, various anomalies caused by sensor failure or damage lead to abnormalities of the monitoring data. It is inefficient to remove abnormal data by manual elimination because of the massive number of data obtained by monitoring systems. In this paper, a data anomaly detection method based on structural vibration signals and a convolutional neural network (CNN) is proposed, which can automatically identify and eliminate abnormal data. First, the anomaly detection problem is modeled as a time series classification problem. Data preprocessing and data augmentation, including data expansion and down-sampling to construct new samples, are employed to process the original time series. For a small number of samples in the data set, randomly increase outliers, symmetrical flipping, and noise addition methods are used for data expansion, and samples with the same label are added without increasing the original samples. The down-sampling method of symmetrically extracting the maximum value and the minimum value at the same time can effectively reduce the dimensionality of the input sample, while retaining the characteristics of the data to the greatest extent. Using hyperparameter tuning of the classification weights, CNN is more effective in dealing with unbalanced training sets. Finally, the effectiveness of the proposed method is proved by the anomaly detection of acceleration data on a long-span bridge. For the anomaly detection problem modeled as a time series classification problem, the proposed method can effectively identify various abnormal patterns.


2021 ◽  
Vol 42 (18) ◽  
pp. 6921-6944
Author(s):  
Yi Chen ◽  
Yi He ◽  
Lifeng Zhang ◽  
Youdong Chen ◽  
Hongyu Pu ◽  
...  

Author(s):  
Yuheng Hu ◽  
Yili Hong

Residents often rely on newspapers and television to gather hyperlocal news for community awareness and engagement. More recently, social media have emerged as an increasingly important source of hyperlocal news. Thus far, the literature on using social media to create desirable societal benefits, such as civic awareness and engagement, is still in its infancy. One key challenge in this research stream is to timely and accurately distill information from noisy social media data streams to community members. In this work, we develop SHEDR (social media–based hyperlocal event detection and recommendation), an end-to-end neural event detection and recommendation framework with a particular use case for Twitter to facilitate residents’ information seeking of hyperlocal events. The key model innovation in SHEDR lies in the design of the hyperlocal event detector and the event recommender. First, we harness the power of two popular deep neural network models, the convolutional neural network (CNN) and long short-term memory (LSTM), in a novel joint CNN-LSTM model to characterize spatiotemporal dependencies for capturing unusualness in a region of interest, which is classified as a hyperlocal event. Next, we develop a neural pairwise ranking algorithm for recommending detected hyperlocal events to residents based on their interests. To alleviate the sparsity issue and improve personalization, our algorithm incorporates several types of contextual information covering topic, social, and geographical proximities. We perform comprehensive evaluations based on two large-scale data sets comprising geotagged tweets covering Seattle and Chicago. We demonstrate the effectiveness of our framework in comparison with several state-of-the-art approaches. We show that our hyperlocal event detection and recommendation models consistently and significantly outperform other approaches in terms of precision, recall, and F-1 scores. Summary of Contribution: In this paper, we focus on a novel and important, yet largely underexplored application of computing—how to improve civic engagement in local neighborhoods via local news sharing and consumption based on social media feeds. To address this question, we propose two new computational and data-driven methods: (1) a deep learning–based hyperlocal event detection algorithm that scans spatially and temporally to detect hyperlocal events from geotagged Twitter feeds; and (2) A personalized deep learning–based hyperlocal event recommender system that systematically integrates several contextual cues such as topical, geographical, and social proximity to recommend the detected hyperlocal events to potential users. We conduct a series of experiments to examine our proposed models. The outcomes demonstrate that our algorithms are significantly better than the state-of-the-art models and can provide users with more relevant information about the local neighborhoods that they live in, which in turn may boost their community engagement.


Open Physics ◽  
2021 ◽  
Vol 19 (1) ◽  
pp. 618-627
Author(s):  
Weixing Song ◽  
Jingjing Wu ◽  
Jianshe Kang ◽  
Jun Zhang

Abstract The aim of this study was to improve the low accuracy of equipment spare parts requirement predicting, which affects the quality and efficiency of maintenance support, based on the summary and analysis of the existing spare parts requirement predicting research. This article introduces the current latest popular long short-term memory (LSTM) algorithm which has the best effect on time series data processing to equipment spare parts requirement predicting, according to the time series characteristics of spare parts consumption data. A method for predicting the requirement for maintenance spare parts based on the LSTM recurrent neural network is proposed, and the network structure is designed in detail, the realization of network training and network prediction is given. The advantages of particle swarm algorithm are introduced to optimize the network parameters, and actual data of three types of equipment spare parts consumption are used for experiments. The performance comparison of predictive models such as BP neural network, generalized regression neural network, wavelet neural network, and squeeze-and-excitation network prove that the new method is effective and provides an effective method for scientifically predicting the requirement for maintenance spare parts and improving the quality of equipment maintenance.


2021 ◽  
Author(s):  
Fatemehalsadat Madaeni ◽  
Karem Chokmani ◽  
Rachid Lhissou ◽  
Saeid Homayuni ◽  
Yves Gauthier ◽  
...  

Abstract. In cold regions, ice-jam events result in severe flooding due to a rapid rise in water levels upstream of the jam. These floods threaten human safety and damage properties and infrastructures as the floods resulting from ice-jams are sudden. Hence, the ice-jam prediction tools can give an early warning to increase response time and minimize the possible corresponding damages. However, the ice-jam prediction has always been a challenging problem as there is no analytical method available for this purpose. Nonetheless, ice jams form when some hydro-meteorological conditions happen, a few hours to a few days before the event. The ice-jam prediction problem can be considered as a binary multivariate time-series classification. Deep learning techniques have been successfully applied for time-series classification in many fields such as finance, engineering, weather forecasting, and medicine. In this research, we successfully applied CNN, LSTM, and combined CN-LSTM networks for ice-jam prediction for all the rivers in Quebec. The results show that the CN-LSTM model yields the best results in the validation and generalization with F1 scores of 0.82 and 0.91, respectively. This demonstrates that CNN and LSTM models are complementary, and a combination of them further improves classification.


Sign in / Sign up

Export Citation Format

Share Document