scholarly journals Hybrid methodology for tuberculosis incidence time-series forecasting based on ARIMA and a NAR neural network

2017 ◽  
Vol 145 (6) ◽  
pp. 1118-1129 ◽  
Author(s):  
K. W. WANG ◽  
C. DENG ◽  
J. P. LI ◽  
Y. Y. ZHANG ◽  
X. Y. LI ◽  
...  

SUMMARYTuberculosis (TB) affects people globally and is being reconsidered as a serious public health problem in China. Reliable forecasting is useful for the prevention and control of TB. This study proposes a hybrid model combining autoregressive integrated moving average (ARIMA) with a nonlinear autoregressive (NAR) neural network for forecasting the incidence of TB from January 2007 to March 2016. Prediction performance was compared between the hybrid model and the ARIMA model. The best-fit hybrid model was combined with an ARIMA (3,1,0) × (0,1,1)12 and NAR neural network with four delays and 12 neurons in the hidden layer. The ARIMA-NAR hybrid model, which exhibited lower mean square error, mean absolute error, and mean absolute percentage error of 0·2209, 0·1373, and 0·0406, respectively, in the modelling performance, could produce more accurate forecasting of TB incidence compared to the ARIMA model. This study shows that developing and applying the ARIMA-NAR hybrid model is an effective method to fit the linear and nonlinear patterns of time-series data, and this model could be helpful in the prevention and control of TB.

BMJ Open ◽  
2021 ◽  
Vol 11 (1) ◽  
pp. e041040
Author(s):  
Yanling Zheng ◽  
Xueliang Zhang ◽  
Xijiang Wang ◽  
Kai Wang ◽  
Yan Cui

ObjectivesKashgar, located in Xinjiang, China has a high incidence of tuberculosis (TB) making prevention and control extremely difficult. In addition, there have been very few prediction studies on TB incidence here. We; therefore, considered it a high priority to do prediction analysis of TB incidence in Kashgar, and so provide a scientific reference for eventual prevention and control.DesignTime series study.Setting Kashgar, ChinaKashgar, China.MethodsWe used a single Box-Jenkins method and a Box-Jenkins and Elman neural network (ElmanNN) hybrid method to do prediction analysis of TB incidence in Kashgar. Root mean square error (RMSE), mean absolute error (MAE) and mean absolute percentage error (MAPE) were used to measure the prediction accuracy.ResultsAfter careful analysis, the single autoregression (AR) (1, 2, 8) model and the AR (1, 2, 8)-ElmanNN (AR-Elman) hybrid model were established, and the optimal neurons value of the AR-Elman hybrid model is 6. In the fitting dataset, the RMSE, MAE and MAPE were 6.15, 4.33 and 0.2858, respectively, for the AR (1, 2, 8) model, and 3.78, 3.38 and 0.1837, respectively, for the AR-Elman hybrid model. In the forecasting dataset, the RMSE, MAE and MAPE were 10.88, 8.75 and 0.2029, respectively, for the AR (1, 2, 8) model, and 8.86, 7.29 and 0.2006, respectively, for the AR-Elman hybrid model.ConclusionsBoth the single AR (1, 2, 8) model and the AR-Elman model could be used to predict the TB incidence in Kashgar, but the modelling and validation scale-dependent measures (RMSE, MAE and MAPE) in the AR (1, 2, 8) model were inferior to those in the AR-Elman hybrid model, which indicated that the AR-Elman hybrid model was better than the AR (1, 2, 8) model. The Box-Jenkins and ElmanNN hybrid method therefore can be highlighted in predicting the temporal trends of TB incidence in Kashgar, which may act as the potential for far-reaching implications for prevention and control of TB.


2020 ◽  
Vol 20 (1) ◽  
Author(s):  
Yihuai Huang ◽  
Chao Xu ◽  
Mengzhong Ji ◽  
Wei Xiang ◽  
Da He

Abstract Background Accurate forecasting of medical service demand is beneficial for the reasonable healthcare resource planning and allocation. The daily outpatient volume is characterized by randomness, periodicity and trend, and the time series methods, like ARIMA are often used for short-term outpatient visits forecasting. Therefore, to further enlarge the prediction horizon and improve the prediction accuracy, a hybrid prediction model integrating ARIMA and self-adaptive filtering method is proposed. Methods The ARIMA model is first used to identify the features like cyclicity and trend of the time series data and to estimate the model parameters. The parameters are then adjusted by the steepest descent algorithm in the adaptive filtering method to reduce the prediction error. The hybrid model is validated and compared with traditional ARIMA by several test sets from the Time Series Data Library (TSDL), a weekly emergency department (ED) visit case from literature study, and the real cases of prenatal examinations and B-ultrasounds in a maternal and child health care center (MCHCC) in Ningbo. Results For TSDL cases the prediction accuracy of the hybrid prediction is improved by 80–99% compared with the ARIMA model. For the weekly ED visit case, the forecasting results of the hybrid model are better than those of both traditional ARIMA and ANN model, and similar to the ANN combined data decomposition model mentioned in the literature. For the actual data of MCHCC in Ningbo, the MAPE predicted by the ARIMA model in the two departments was 18.53 and 27.69%, respectively, and the hybrid models were 2.79 and 1.25%, respectively. Conclusions The hybrid prediction model outperforms the traditional ARIMA model in both accurate predicting result with smaller average relative error and the applicability for short-term and medium-term prediction.


The challenging endeavor of a time series forecast model is to predict the future time series data accurately. Traditionally, the fundamental forecasting model in time series analysis is the autoregressive integrated moving average model or the ARIMA model requiring a model identification of a three-component vector which are the autoregressive order, the differencing order, and the moving average order before fitting coefficients of the model via the Box-Jenkins method. A model identification is analyzed via the sample autocorrelation function and the sample partial autocorrelation function which are effective tools for identifying the ARMA order but it is quite difficult for analysts. Even though a likelihood based-method is presented to automate this process by varying the ARIMA order and choosing the best one with the smallest criteria, such as Akaike information criterion. Nevertheless the obtained ARIMA model may not pass the residual diagnostic test. This paper presents the residual neural network model, called the self-identification ResNet-ARIMA order model to automatically learn the ARIMA order from known ARIMA time series data via sample autocorrelation function, the sample partial autocorrelation function and differencing time series images. In this work, the training time series data are randomly simulated and checked for stationary and invertibility properties before they are used. The result order from the model is used to generate and fit the ARIMA model by the Box-Jenkins method for predicting future values. The whole process of the forecasting time series algorithm is called the self-identification ResNet-ARIMA algorithm. The performance of the residual neural network model is evaluated by Precision, Recall and F1-score and is compared with the likelihood basedmethod and ResNET50. In addition, the performance of the forecasting time series algorithm is applied to the real world datasets to ensure the reliability by mean absolute percentage error, symmetric mean absolute percentage error, mean absolute error and root mean square error and this algorithm is confirmed with the residual diagnostic checks by the Ljung-Box test. From the experimental results, the new methodologies of this research outperforms other models in terms of identifying the order and predicting the future values.


2018 ◽  
Vol 73 ◽  
pp. 13008 ◽  
Author(s):  
Hasbi Yasin ◽  
Budi Warsito ◽  
Rukun Santoso ◽  
Suparti

Vector autoregressive model proposed for multivariate time series data. Neural Network, including Feed Forward Neural Network (FFNN), is the powerful tool for the nonlinear model. In autoregressive model, the input layer is the past values of the same series up to certain lag and the output layers is the current value. So, VAR-NN is proposed to predict the multivariate time series data using nonlinear approach. The optimal lag time in VAR are used as aid of selecting the input in VAR-NN. In this study we develop the soft computation tools of VAR-NN based on Graphical User Interface. In each number of neurons in hidden layer, the looping process is performed several times in order to get the best result. The best one is chosen by the least of Mean Absolute Percentage Error (MAPE) criteria. In this study, the model is applied in the two series of stock price data from Indonesia Stock Exchange. Evaluation of VAR-NN performance was based on train-validation and test-validation sample approach. Based on the empirical stock price data it can be concluded that VAR-NN yields perfect performance both in in-sample and in out-sample for non-linear function approximation. This is indicated by the MAPE value that is less than 1% .


PLoS ONE ◽  
2021 ◽  
Vol 16 (3) ◽  
pp. e0248597
Author(s):  
Guo-hua Ye ◽  
Mirxat Alim ◽  
Peng Guan ◽  
De-sheng Huang ◽  
Bao-sen Zhou ◽  
...  

Objective Hemorrhagic fever with renal syndrome (HFRS), one of the main public health concerns in mainland China, is a group of clinically similar diseases caused by hantaviruses. Statistical approaches have always been leveraged to forecast the future incidence rates of certain infectious diseases to effectively control their prevalence and outbreak potential. Compared to the use of one base model, model stacking can often produce better forecasting results. In this study, we fitted the monthly reported cases of HFRS in mainland China with a model stacking approach and compared its forecasting performance with those of five base models. Method We fitted the monthly reported cases of HFRS ranging from January 2004 to June 2019 in mainland China with an autoregressive integrated moving average (ARIMA) model; the Holt-Winter (HW) method, seasonal decomposition of the time series by LOESS (STL); a neural network autoregressive (NNAR) model; and an exponential smoothing state space model with a Box-Cox transformation; ARMA errors; and trend and seasonal components (TBATS), and we combined the forecasting results with the inverse rank approach. The forecasting performance was estimated based on several accuracy criteria for model prediction, including the mean absolute percentage error (MAPE), root-mean-squared error (RMSE) and mean absolute error (MAE). Result There was a slight downward trend and obvious seasonal periodicity inherent in the time series data for HFRS in mainland China. The model stacking method was selected as the best approach with the best performance in terms of both fitting (RMSE 128.19, MAE 85.63, MAPE 8.18) and prediction (RMSE 151.86, MAE 118.28, MAPE 13.16). Conclusion The results showed that model stacking by using the optimal mean forecasting weight of the five abovementioned models achieved the best performance in terms of predicting HFRS one year into the future. This study has corroborated the conclusion that model stacking is an easy way to enhance prediction accuracy when modeling HFRS.


Energies ◽  
2020 ◽  
Vol 13 (24) ◽  
pp. 6623
Author(s):  
Rial A. Rajagukguk ◽  
Raden A. A. Ramadhan ◽  
Hyun-Jin Lee

Presently, deep learning models are an alternative solution for predicting solar energy because of their accuracy. The present study reviews deep learning models for handling time-series data to predict solar irradiance and photovoltaic (PV) power. We selected three standalone models and one hybrid model for the discussion, namely, recurrent neural network (RNN), long short-term memory (LSTM), gated recurrent unit (GRU), and convolutional neural network-LSTM (CNN–LSTM). The selected models were compared based on the accuracy, input data, forecasting horizon, type of season and weather, and training time. The performance analysis shows that these models have their strengths and limitations in different conditions. Generally, for standalone models, LSTM shows the best performance regarding the root-mean-square error evaluation metric (RMSE). On the other hand, the hybrid model (CNN–LSTM) outperforms the three standalone models, although it requires longer training data time. The most significant finding is that the deep learning models of interest are more suitable for predicting solar irradiance and PV power than other conventional machine learning models. Additionally, we recommend using the relative RMSE as the representative evaluation metric to facilitate accuracy comparison between studies.


2020 ◽  
Author(s):  
Sina Ardabili ◽  
Amir Mosavi ◽  
Shahab S. Band ◽  
Annamaria R. Varkonyi-Koczy

Abstract Advancement of the novel models for time-series prediction of COVID-19 is of utmost importance. Machine learning (ML) methods have recently shown promising results. The present study aims to engage an artificial neural network-integrated by grey wolf optimizer for COVID-19 outbreak predictions by employing the Global dataset. Training and testing processes have been performed by time-series data related to January 22 to September 15, 2020 and validation has been performed by time-series data related to September 16 to October 15, 2020. Results have been evaluated by employing mean absolute percentage error (MAPE) and correlation coefficient (r) values. ANN-GWO provided a MAPE of 6.23, 13.15 and 11.4% for training, testing and validating phases, respectively. According to the results, the developed model could successfully cope with the prediction task.


Author(s):  
Muhammad Faheem Mushtaq ◽  
Urooj Akram ◽  
Muhammad Aamir ◽  
Haseeb Ali ◽  
Muhammad Zulqarnain

It is important to predict a time series because many problems that are related to prediction such as health prediction problem, climate change prediction problem and weather prediction problem include a time component. To solve the time series prediction problem various techniques have been developed over many years to enhance the accuracy of forecasting. This paper presents a review of the prediction of physical time series applications using the neural network models. Neural Networks (NN) have appeared as an effective tool for forecasting of time series.  Moreover, to resolve the problems related to time series data, there is a need of network with single layer trainable weights that is Higher Order Neural Network (HONN) which can perform nonlinearity mapping of input-output. So, the developers are focusing on HONN that has been recently considered to develop the input representation spaces broadly. The HONN model has the ability of functional mapping which determined through some time series problems and it shows the more benefits as compared to conventional Artificial Neural Networks (ANN). The goal of this research is to present the reader awareness about HONN for physical time series prediction, to highlight some benefits and challenges using HONN.


Mathematics ◽  
2021 ◽  
Vol 9 (10) ◽  
pp. 1122
Author(s):  
Oksana Mandrikova ◽  
Nadezhda Fetisova ◽  
Yuriy Polozov

A hybrid model for the time series of complex structure (HMTS) was proposed. It is based on the combination of function expansions in a wavelet series with ARIMA models. HMTS has regular and anomalous components. The time series components, obtained after expansion, have a simpler structure that makes it possible to identify the ARIMA model if the components are stationary. This allows us to obtain a more accurate ARIMA model for a time series of complicated structure and to extend the area for application. To identify the HMTS anomalous component, threshold functions are applied. This paper describes a technique to identify HMTS and proposes operations to detect anomalies. With the example of an ionospheric parameter time series, we show the HMTS efficiency, describe the results and their application in detecting ionospheric anomalies. The HMTS was compared with the nonlinear autoregression neural network NARX, which confirmed HMTS efficiency.


AI ◽  
2021 ◽  
Vol 2 (1) ◽  
pp. 48-70
Author(s):  
Wei Ming Tan ◽  
T. Hui Teo

Prognostic techniques attempt to predict the Remaining Useful Life (RUL) of a subsystem or a component. Such techniques often use sensor data which are periodically measured and recorded into a time series data set. Such multivariate data sets form complex and non-linear inter-dependencies through recorded time steps and between sensors. Many current existing algorithms for prognostic purposes starts to explore Deep Neural Network (DNN) and its effectiveness in the field. Although Deep Learning (DL) techniques outperform the traditional prognostic algorithms, the networks are generally complex to deploy or train. This paper proposes a Multi-variable Time Series (MTS) focused approach to prognostics that implements a lightweight Convolutional Neural Network (CNN) with attention mechanism. The convolution filters work to extract the abstract temporal patterns from the multiple time series, while the attention mechanisms review the information across the time axis and select the relevant information. The results suggest that the proposed method not only produces a superior accuracy of RUL estimation but it also trains many folds faster than the reported works. The superiority of deploying the network is also demonstrated on a lightweight hardware platform by not just being much compact, but also more efficient for the resource restricted environment.


Sign in / Sign up

Export Citation Format

Share Document