scholarly journals Assessment of Variability in Irregularly Sampled Time Series: Applications to Mental Healthcare

Mathematics ◽  
2020 ◽  
Vol 9 (1) ◽  
pp. 71
Author(s):  
Pablo Bonilla-Escribano ◽  
David Ramírez ◽  
Alejandro Porras-Segovia ◽  
Antonio Artés-Rodríguez

Variability is defined as the propensity at which a given signal is likely to change. There are many choices for measuring variability, and it is not generally known which ones offer better properties. This paper compares different variability metrics applied to irregularly (nonuniformly) sampled time series, which have important clinical applications, particularly in mental healthcare. Using both synthetic and real patient data, we identify the most robust and interpretable variability measures out of a set 21 candidates. Some of these candidates are also proposed in this work based on the absolute slopes of the time series. An additional synthetic data experiment shows that when the complete time series is unknown, as it happens with real data, a non-negligible bias that favors normalized and/or metrics based on the raw observations of the series appears. Therefore, only the results of the synthetic experiments, which have access to the full series, should be used to draw conclusions. Accordingly, the median absolute deviation of the absolute value of the successive slopes of the data is the best way of measuring variability for this kind of time series.

2018 ◽  
Vol 2018 ◽  
pp. 1-15
Author(s):  
Blanca Guillen ◽  
Jose L. Paredes ◽  
Rubén Medina

This paper proposes a simple yet effective approach for detecting activated voxels in fMRI data by exploiting the inherent sparsity property of the BOLD signal in temporal and spatial domains. In the time domain, the approach combines the General Linear Model (GLM) with a Least Absolute Deviation (LAD) based regression method regularized by the pseudonorm l0 to promote sparsity in the parameter vector of the model. In the spatial domain, detection of activated regions is based on thresholding the spatial map of estimated parameters associated with a particular stimulus. The threshold is calculated by exploiting the sparseness of the BOLD signal in the spatial domain assuming a Laplacian distribution model. The proposed approach is validated using synthetic and real fMRI data. For synthetic data, results show that the proposed approach is able to detect most activated voxels without any false activation. For real data, the method is evaluated through comparison with the SPM software. Results indicate that this approach can effectively find activated regions that are similar to those found by SPM, but using a much simpler approach. This study may lead to the development of robust spatial approaches to further simplifying the complexity of classical schemes.


2015 ◽  
Vol 41 (7) ◽  
pp. 692-713
Author(s):  
Donna M. Dudney ◽  
Benjamas Jirasakuldech ◽  
Thomas Zorn ◽  
Riza Emekter

Purpose – Variations in price/earnings (P/E) ratios are explained in a rational expectations framework by a number of fundamental factors, such as differences in growth expectations and risk. The purpose of this paper is to use a regression model and data from four sample periods (1996, 2000, 2001, and 2008) to separate the earnings/price (E/P) ratio into two parts – the portion of E/P that is related to fundamental determinants and a residual portion that cannot be explained by fundamentals. The authors use the residual portion as an indicator of over or undervaluation; a large negative residual is consistent with overvaluation while a large positive residual implies undervaluation. The authors find that stocks with larger negative residuals are associated with lower subsequent returns and reward-to-risk ratio, while stocks with larger positive residuals are associated with higher subsequent returns and reward-to-risk ratio. This pattern persists for both one and two-year holding periods. Design/methodology/approach – This study uses a regression methodology to decompose E/P into two parts – the portion of E/P than is related to fundamental determinants and a residual portion that cannot be explained by fundamentals. Focussing on the second portion allows us to isolate a potential indicator of stock over or undervaluation. Using a sample of stocks from four time periods (1996, 2000, 2001, and 2008, the authors calculate the residuals from a regression model of the fundamental determinants of cross-sectional variation in E/P. These residuals are then ranked and used to divide the stock sample into deciles, with the first decile containing the stocks with the highest negative residuals (indicating overvaluation) and the tenth decile containing stocks with the highest positive residuals (indicating undervaluation). Total returns for subsequent one and two-year holding periods are then calculated for each decile portfolio. Findings – The authors find that high positive residual stocks substantially outperform high negative residual stocks. This is true even after risk adjustments to the portfolio returns. The residual E/P appears to accurately predict relative stock performance with a relatively high degree of accuracy. Research limitations/implications – The findings of this paper provide some important implications for practitioners and investors, particularly for the stock selection, fund allocations, and portfolio strategies. Practitioners can still rely on a valuation measure such as E/P as a useful tool for making successful investment decisions and enhance portfolio performance. Investors can earn abnormal returns by allocating more weights on stocks with high E/P multiples. Portfolios of high E/P multiples or undervalued stocks are found to enjoy higher risk-adjusted returns after controlling for the fundamental factors. The most beneficial performance holding period return will be for a relatively short period of time ranging from one to two years. Relying on the E/P valuation ratios for a long-term investment may add little value. Practical implications – Practitioners and academics have long relied on the P/E ratio as an indicator of relative overvaluation. An increase in the absolute value of P/E, however, does not always indicate overvaluation. Instead, a high P/E ratio can simply reflect changes in the fundamental factors that affect P/E. The authors find that stocks with larger negative residuals are associated with lower subsequent returns and coefficients of variation, while stocks with larger positive residuals are associated with higher subsequent returns and coefficients of variation. This pattern persists for both one and two-year holding periods. Originality/value – The P/E ratio is widely used, particularly by practitioners, as a measure of relative stock valuation. The ratio has been used in both cross-sectional and time series comparisons as a metric for determining whether stocks are under or overvalued. An increase in the absolute value of P/E, however, does not always indicate overvaluation. Instead, a high P/E ratio can simply reflect changes in the fundamental factors that affect P/E. If interest rates are relatively low, for example, the time series P/E should be correspondingly higher. Similarly, if the risk of a stock is low, that stock’s P/E ratio should be higher than the P/E ratios of less risky stocks. The authors examine the cross-sectional behavior of the P/E (the authors actually use the E/P ratio for reasons explained below) after controlling for factors that are likely to fundamentally affect this ratio. These factors include the dividend payout ratio, risk measures, growth measures, and factors such as size and book to market that have been identified by Fama and French (1993) and others as important in explaining the cross-sectional variation in common stock returns. To control for changes in these primary determinants of E/P, the authors use a simple regression model. The residuals from this model represent the unexplained cross-sectional variation in E/P. The authors argue that this unexplained variation is a more reliable indicator than the raw E/P ratio of the relative under or overvaluation of stocks.


2021 ◽  
Vol 4 (1) ◽  
Author(s):  
August DuMont Schütte ◽  
Jürgen Hetzel ◽  
Sergios Gatidis ◽  
Tobias Hepp ◽  
Benedikt Dietz ◽  
...  

AbstractPrivacy concerns around sharing personally identifiable information are a major barrier to data sharing in medical research. In many cases, researchers have no interest in a particular individual’s information but rather aim to derive insights at the level of cohorts. Here, we utilise generative adversarial networks (GANs) to create medical imaging datasets consisting entirely of synthetic patient data. The synthetic images ideally have, in aggregate, similar statistical properties to those of a source dataset but do not contain sensitive personal information. We assess the quality of synthetic data generated by two GAN models for chest radiographs with 14 radiology findings and brain computed tomography (CT) scans with six types of intracranial haemorrhages. We measure the synthetic image quality by the performance difference of predictive models trained on either the synthetic or the real dataset. We find that synthetic data performance disproportionately benefits from a reduced number of classes. Our benchmark also indicates that at low numbers of samples per class, label overfitting effects start to dominate GAN training. We conducted a reader study in which trained radiologists discriminate between synthetic and real images. In accordance with our benchmark results, the classification accuracy of radiologists improves with an increasing resolution. Our study offers valuable guidelines and outlines practical conditions under which insights derived from synthetic images are similar to those that would have been derived from real data. Our results indicate that synthetic data sharing may be an attractive alternative to sharing real patient-level data in the right setting.


JAMIA Open ◽  
2020 ◽  
Author(s):  
Randi E Foraker ◽  
Sean C Yu ◽  
Aditi Gupta ◽  
Andrew P Michelson ◽  
Jose A Pineda Soto ◽  
...  

Abstract Background Synthetic data may provide a solution to researchers who wish to generate and share data in support of precision healthcare. Recent advances in data synthesis enable the creation and analysis of synthetic derivatives as if they were the original data; this process has significant advantages over data deidentification. Objectives To assess a big-data platform with data-synthesizing capabilities (MDClone Ltd., Beer Sheva, Israel) for its ability to produce data that can be used for research purposes while obviating privacy and confidentiality concerns. Methods We explored three use cases and tested the robustness of synthetic data by comparing the results of analyses using synthetic derivatives to analyses using the original data using traditional statistics, machine learning approaches, and spatial representations of the data. We designed these use cases with the purpose of conducting analyses at the observation level (Use Case 1), patient cohorts (Use Case 2), and population-level data (Use Case 3). Results For each use case, the results of the analyses were sufficiently statistically similar (P > 0.05) between the synthetic derivative and the real data to draw the same conclusions. Discussion and conclusion This article presents the results of each use case and outlines key considerations for the use of synthetic data, examining their role in clinical research for faster insights and improved data sharing in support of precision healthcare.


Author(s):  
Weibing Du ◽  
Weiqian Ji ◽  
Linjuan Xu ◽  
Shuangting Wang

Glacier melting is one of the important causes of glacier morphology change and can provide basic parameters for calculating glacier volume change and glacier mass balance, which, in turn, is important for evaluating water resources. However, it is difficult to obtain large-scale time series of glacier changes due to the cloudy and foggy conditions which are typical of mountain areas. Gravity-measuring satellites and laser altimetry satellites can monitor changes in glacier volume over a wide area, while synthetic-aperture radar satellites can monitoring glacier morphology with a high spatial and temporal resolution. In this article, an interferometric method using a short temporal baseline and a short spatial baseline, called the Small Baseline Subsets (SBAS) Interferometric Synthetic-Aperture Radar (InSAR) method, was used to study the average rate of glacier deformation on Karlik Mountain, in the Eastern Tienshan Mountains, China, by using 19 Sentinel-1A images from November 2017 to December 2018. Thus, a time series analysis of glacier deformation was conducted. It was found that the average glacier deformation in the study region was −11.77 ± 9.73 mm/year, with the observation sites generally moving away from the satellite along the Line of Sight (LOS). Taking the ridge line as the dividing line, it was found that the melting rate of southern slopes was higher than that of northern slopes. According to the perpendicular of the mountain direction, the mountain was divided into an area in the northwest with large glaciers (Area I) and an area in the southeast with small glaciers (Area II). It was found that the melting rate in the southeast area was larger than that in the northwest area. Additionally, through the analysis of temperature and precipitation data, it was found that precipitation played a leading role in glacier deformation in the study region. Through the statistical analysis of the deformation, it was concluded that the absolute value of deformation is large at elevations below 4200 m while the absolute value of the deformation is very small at elevations above 4500 m; the direction of deformation is always away from the satellite along the LOS and the absolute value of glacier deformation decreases with increasing elevation.


Sensors ◽  
2019 ◽  
Vol 19 (5) ◽  
pp. 1181 ◽  
Author(s):  
Jessamyn Dahmen ◽  
Diane Cook

Creation of realistic synthetic behavior-based sensor data is an important aspect of testing machine learning techniques for healthcare applications. Many of the existing approaches for generating synthetic data are often limited in terms of complexity and realism. We introduce SynSys, a machine learning-based synthetic data generation method, to improve upon these limitations. We use this method to generate synthetic time series data that is composed of nested sequences using hidden Markov models and regression models which are initially trained on real datasets. We test our synthetic data generation technique on a real annotated smart home dataset. We use time series distance measures as a baseline to determine how realistic the generated data is compared to real data and demonstrate that SynSys produces more realistic data in terms of distance compared to random data generation, data from another home, and data from another time period. Finally, we apply our synthetic data generation technique to the problem of generating data when only a small amount of ground truth data is available. Using semi-supervised learning we demonstrate that SynSys is able to improve activity recognition accuracy compared to using the small amount of real data alone.


Entropy ◽  
2021 ◽  
Vol 23 (12) ◽  
pp. 1586
Author(s):  
Dursun Aydın ◽  
Syed Ejaz Ahmed ◽  
Ersin Yılmaz

This paper focuses on the adaptive spline (A-spline) fitting of the semiparametric regression model to time series data with right-censored observations. Typically, there are two main problems that need to be solved in such a case: dealing with censored data and obtaining a proper A-spline estimator for the components of the semiparametric model. The first problem is traditionally solved by the synthetic data approach based on the Kaplan–Meier estimator. In practice, although the synthetic data technique is one of the most widely used solutions for right-censored observations, the transformed data’s structure is distorted, especially for heavily censored datasets, due to the nature of the approach. In this paper, we introduced a modified semiparametric estimator based on the A-spline approach to overcome data irregularity with minimum information loss and to resolve the second problem described above. In addition, the semiparametric B-spline estimator was used as a benchmark method to gauge the success of the A-spline estimator. To this end, a detailed Monte Carlo simulation study and a real data sample were carried out to evaluate the performance of the proposed estimator and to make a practical comparison.


Author(s):  
P.L. Nikolaev

This article deals with method of binary classification of images with small text on them Classification is based on the fact that the text can have 2 directions – it can be positioned horizontally and read from left to right or it can be turned 180 degrees so the image must be rotated to read the sign. This type of text can be found on the covers of a variety of books, so in case of recognizing the covers, it is necessary first to determine the direction of the text before we will directly recognize it. The article suggests the development of a deep neural network for determination of the text position in the context of book covers recognizing. The results of training and testing of a convolutional neural network on synthetic data as well as the examples of the network functioning on the real data are presented.


2020 ◽  
Vol 11 (1) ◽  
pp. 20
Author(s):  
Muhammad Ikbal Abdullah ◽  
Andi Chairil Furqan ◽  
Nina Yusnita Yamin ◽  
Fahri Eka Oktora

This study aims to analyze the sensitivity testing using measurements of realization of regional own-source revenues and operating expenditure and to analyze the extent of the effect of sample differences between Java and non-Java provinces by using samples outside of Java. By using sensitivity analysis, the results found the influence of audit opinion on the performance of the provincial government mediated by the realization of regional operating expenditure. More specifically, when using the measurement of the absolute value of the realization of regional operating expenditure it was found that there was a direct positive and significant influence of audit opinion on the performance of the Provincial Government. However, no significant effect of audit opinion was found on the realization value of regional operating expenditure and the effect of the realization value of regional operating expenditure on the performance of the Provincial Government. This result implies that an increase in audit opinion will be more likely to be used as an incentive for the Provincial Government to increase the realization of regional operating expenditure.


1977 ◽  
Vol 32 (11-12) ◽  
pp. 908-912 ◽  
Author(s):  
H. J. Schmidt ◽  
U. Schaum ◽  
J. P. Pichotka

Abstract The influence of five different methods of homogenisation (1. The method according to Potter and Elvehjem, 2. A modification of this method called Potter S, 3. The method of Dounce, 4. Homogenisation by hypersonic waves and 5. Coarce-grained homogenisation with the “Mikro-fleischwolf”) on the absolute value and stability of oxygen uptake of guinea pig liver homogenates has been investigated in simultaneous measurements. All homogenates showed a characteristic fall of oxygen uptake during measuring time (3 hours). The modified method according to Potter and Elvehjem called Potter S showed reproducible results without any influence by homogenisation intensity.


Sign in / Sign up

Export Citation Format

Share Document