scholarly journals Weighted-permutation entropy: A complexity measure for time series incorporating amplitude information

2013 ◽  
Vol 87 (2) ◽  
Author(s):  
Bilal Fadlallah ◽  
Badong Chen ◽  
Andreas Keil ◽  
José Príncipe
Sensors ◽  
2019 ◽  
Vol 19 (23) ◽  
pp. 5203 ◽  
Author(s):  
Li ◽  
Gao ◽  
Wang

Permutation entropy (PE), as one of the powerful complexity measures for analyzing time series, has advantages of easy implementation and high efficiency. In order to improve the performance of PE, some improved PE methods have been proposed through introducing amplitude information and distance information in recent years. Weighted-permutation entropy (W-PE) weight each arrangement pattern by using variance information, which has good robustness and stability in the case of high noise level and can extract complexity information from data with spike feature or abrupt amplitude change. Dispersion entropy (DE) introduces amplitude information by using the normal cumulative distribution function (NCDF); it not only can detect the change of simultaneous frequency and amplitude, but also is superior to the PE method in distinguishing different data sets. Reverse permutation entropy (RPE) is defined as the distance to white noise in the opposite trend with PE and W-PE, which has high stability for time series with varying lengths. To further improve the performance of PE, we propose a new complexity measure for analyzing time series, and term it as reverse dispersion entropy (RDE). RDE takes PE as its theoretical basis and combines the advantages of DE and RPE by introducing amplitude information and distance information. Simulation experiments were carried out on simulated and sensor signals, including mutation signal detection under different parameters, noise robustness testing, stability testing under different signal-to-noise ratios (SNRs), and distinguishing real data for different kinds of ships and faults. The experimental results show, compared with PE, W-PE, RPE, and DE, that RDE has better performance in detecting abrupt signal and noise robustness testing, and has better stability for simulated and sensor signal. Moreover, it also shows higher distinguishing ability than the other four kinds of PE for sensor signals.


2019 ◽  
Vol 513 ◽  
pp. 635-643 ◽  
Author(s):  
Francisco Traversaro ◽  
Nicolás Ciarrocchi ◽  
Florencia Pollo Cattaneo ◽  
Francisco Redelico

2018 ◽  
Vol 48 (10) ◽  
pp. 2877-2897
Author(s):  
Emad Ashtari Nezhad ◽  
Yadollah Waghei ◽  
G. R. Mohtashami Borzadaran ◽  
H. R. Nilli Sani ◽  
Hadi Alizadeh Noughabi

2021 ◽  
pp. 2150055
Author(s):  
Qin Zhou ◽  
Pengjian Shang

Cumulative residual entropy (CRE) has been suggested as a new measure to quantify uncertainty of nonlinear time series signals. Combined with permutation entropy and Rényi entropy, we introduce a generalized measure of CRE at multiple scales, namely generalized cumulative residual entropy (GCRE), and further propose a modification of GCRE procedure by the weighting scheme — weighted generalized cumulative residual entropy (WGCRE). The GCRE and WGCRE methods are performed on the synthetic series to study properties of parameters and verify the validity of measuring complexity of the series. After that, the GCRE and WGCRE methods are applied to the US, European and Chinese stock markets. Through data analysis and statistics comparison, the proposed methods can effectively distinguish stock markets with different characteristics.


Entropy ◽  
2019 ◽  
Vol 21 (4) ◽  
pp. 385 ◽  
Author(s):  
David Cuesta-Frau ◽  
Juan Pablo Murillo-Escobar ◽  
Diana Alexandra Orrego ◽  
Edilson Delgado-Trejos

Permutation Entropy (PE) is a time series complexity measure commonly used in a variety of contexts, with medicine being the prime example. In its general form, it requires three input parameters for its calculation: time series length N, embedded dimension m, and embedded delay τ . Inappropriate choices of these parameters may potentially lead to incorrect interpretations. However, there are no specific guidelines for an optimal selection of N, m, or τ , only general recommendations such as N > > m ! , τ = 1 , or m = 3 , … , 7 . This paper deals specifically with the study of the practical implications of N > > m ! , since long time series are often not available, or non-stationary, and other preliminary results suggest that low N values do not necessarily invalidate PE usefulness. Our study analyses the PE variation as a function of the series length N and embedded dimension m in the context of a diverse experimental set, both synthetic (random, spikes, or logistic model time series) and real–world (climatology, seismic, financial, or biomedical time series), and the classification performance achieved with varying N and m. The results seem to indicate that shorter lengths than those suggested by N > > m ! are sufficient for a stable PE calculation, and even very short time series can be robustly classified based on PE measurements before the stability point is reached. This may be due to the fact that there are forbidden patterns in chaotic time series, not all the patterns are equally informative, and differences among classes are already apparent at very short lengths.


Sign in / Sign up

Export Citation Format

Share Document