scholarly journals Modeling Provincial Covid-19 Epidemic Data Using an Adjusted Time-Dependent SIRD Model

Author(s):  
Luisa Ferrari ◽  
Giuseppe Gerardi ◽  
Giancarlo Manzi ◽  
Alessandra Micheletti ◽  
Federica Nicolussi ◽  
...  

In this paper, we develop a forecasting model for the spread of COVID-19 infection at a provincial (i.e., EU NUTS-3) level in Italy by using official data from the Italian Ministry of Health integrated with data extracted from daily official press conferences of regional authorities and local newspaper websites. This data integration is needed as COVID-19 death data are not available at the NUTS-3 level from official open data channels. An adjusted time-dependent SIRD model is used to predict the behavior of the epidemic; specifically, the number of susceptible, infected, deceased, recovered people and epidemiological parameters. Predictive model performance is evaluated using comparison with real data.

2020 ◽  
Author(s):  
Sudhansu Sekhar Singh ◽  
Dinakrushna Mohapatra

The role of mathematical modelling in predicting spread of an epidemic is of vital importance. The purpose of present study is to develop and apply a computational tool for predicting evolution of different epidemiological variables for COVID-19 in India. We propose a dynamic SIRD (Susceptible-Infected-Recovered-Dead) and SEIRD (Susceptible-Exposed-Infected-Recovered-Dead) model for this purpose. In the dynamic model, time dependent infection rate is assumed for estimating evolution of different variables of the model. Parameter estimation of the model is the first step of the analysis which is performed by least square optimization of priori data. In the second step of the analysis, simulation is carried out by using evaluated parameters for prediction of the outbreak. The computational model has been validated against real data for COVID-19 outbreak in Italy. Time to reach peak, peak infected cases and total reported cases were compared with actual data and found to be in very good agreement. Next the model is applied for the case of India and various Indian states to predict different epidemiological parameters. Priori data was taken from the beginning of nation-wide lockdown on 24 March to 6 July. It was found that peak of the outbreak may reach in the month of August-September with maximum 4-5 lakhs active cases at peak. Total number of reported cases all over India would be in between three to five millions. State wise, Maharashtra, Tamilnadu and Delhi would be worst affected.


2020 ◽  
Author(s):  
Eduardo Atem De Carvalho ◽  
Rogerio Atem De Carvalho

BACKGROUND Since the beginning of the COVID-19 pandemic, researchers and health authorities have sought to identify the different parameters that govern their infection and death cycles, in order to be able to make better decisions. In particular, a series of reproduction number estimation models have been presented, with different practical results. OBJECTIVE This article aims to present an effective and efficient model for estimating the Reproduction Number and to discuss the impacts of sub-notification on these calculations. METHODS The concept of Moving Average Method with Initial value (MAMI) is used, as well as a model for Rt, the Reproduction Number, is derived from experimental data. The models are applied to real data and their performance is presented. RESULTS Analyses on Rt and sub-notification effects for Germany, Italy, Sweden, United Kingdom, South Korea, and the State of New York are presented to show the performance of the methods here introduced. CONCLUSIONS We show that, with relatively simple mathematical tools, it is possible to obtain reliable values for time-dependent, incubation period-independent Reproduction Numbers (Rt). We also demonstrate that the impact of sub-notification is relatively low, after the initial phase of the epidemic cycle has passed.


2017 ◽  
Vol 29 (5) ◽  
pp. 529-542 ◽  
Author(s):  
Marko Intihar ◽  
Tomaž Kramberger ◽  
Dejan Dragan

The paper examines the impact of integration of macroeconomic indicators on the accuracy of container throughput time series forecasting model. For this purpose, a Dynamic factor analysis and AutoRegressive Integrated Moving-Average model with eXogenous inputs (ARIMAX) are used. Both methodologies are integrated into a novel four-stage heuristic procedure. Firstly, dynamic factors are extracted from external macroeconomic indicators influencing the observed throughput. Secondly, the family of ARIMAX models of different orders is generated based on the derived factors. In the third stage, the diagnostic and goodness-of-fit testing is applied, which includes statistical criteria such as fit performance, information criteria, and parsimony. Finally, the best model is heuristically selected and tested on the real data of the Port of Koper. The results show that by applying macroeconomic indicators into the forecasting model, more accurate future throughput forecasts can be achieved. The model is also used to produce future forecasts for the next four years indicating a more oscillatory behaviour in (2018-2020). Hence, care must be taken concerning any bigger investment decisions initiated from the management side. It is believed that the proposed model might be a useful reinforcement of the existing forecasting module in the observed port.


Author(s):  
Fiorella Pia Salvatore ◽  
Alessia Spada ◽  
Francesca Fortunato ◽  
Demetris Vrontis ◽  
Mariantonietta Fiore

The purpose of this paper is to investigate the determinants influencing the costs of cardiovascular disease in the regional health service in Italy’s Apulia region from 2014 to 2016. Data for patients with acute myocardial infarction (AMI), heart failure (HF), and atrial fibrillation (AF) were collected from the hospital discharge registry. Generalized linear models (GLM), and generalized linear mixed models (GLMM) were used to identify the role of random effects in improving the model performance. The study was based on socio-demographic variables and disease-specific variables (diagnosis-related group, hospitalization type, hospital stay, surgery, and economic burden of the hospital discharge form). Firstly, both models indicated an increase in health costs in 2016, and lower spending values for women (p < 0.001) were shown. GLMM indicates a significant increase in health expenditure with increasing age (p < 0.001). Day-hospital has the lowest cost, surgery increases the cost, and AMI is the most expensive pathology, contrary to AF (p < 0.001). Secondly, AIC and BIC assume the lowest values for the GLMM model, indicating the random effects’ relevance in improving the model performance. This study is the first that considers real data to estimate the economic burden of CVD from the regional health service’s perspective. It appears significant for its ability to provide a large set of estimates of the economic burden of CVD, providing information to managers for health management and planning.


2021 ◽  
Vol 2021 ◽  
pp. 1-8
Author(s):  
Shah Imran Alam ◽  
Ihtiram Raza Khan ◽  
Syed Imtiyaz Hassan ◽  
Farheen Siddiqui ◽  
M. Afshar Alam ◽  
...  

The benefits of open data were realised worldwide since the past decades, and the efforts to move more data under the license of open data intensified. There was a steep rise of open data in government repositories. In our study, we point out that privacy is one of the consistent and leading barriers among others. Strong privacy laws restrict data owners from opening the data freely. In this paper, we attempted to study the applied solutions and to the best of our knowledge, we found that anonymity-preserving algorithms did a substantial job to protect privacy in the release of the structured microdata. Such anonymity-preserving algorithms argue and compete in objectivethat not only could the released anonymized data preserve privacy but also the anonymized data preserve the required level of quality. K-anonymity algorithm was the foundation of many of its successor algorithms of all privacy-preserving algorithms. l-diversity claims to add another dimension of privacy protection. Both these algorithms used together are known to provide a good balance between privacy and quality control of the dataset as a whole entity. In this research, we have used the K-anonymity algorithm and compared the results with the addon of l-diversity. We discussed the gap and reported the benefits and loss with various combinations of K and l values, taken in combination with released data quality from an analyst’s perspective. We first used dummy fictitious data to explain the general expectations and then concluded the contrast in the findings with the real data from the food technology domain. The work contradicts the general assumptions with a specific set of evaluation parameters for data quality assessment. Additionally, it is intended to argue in favour of pushing for research contributions in the field of anonymity preservation and intensify the effort for major trends of research, considering its importance and potential to benefit people.


Author(s):  
A. Berveglieri ◽  
A. M. G. Tommaselli ◽  
E. Honkavaara

Hyperspectral camera operating in sequential acquisition mode produces spectral bands that are not recorded at the same instant, thus having different exterior orientation parameters (EOPs) for each band. The study presents experiments on bundle adjustment with time-dependent polynomial models for band orientation of hyperspectral cubes sequentially collected. The technique was applied to a Rikola camera model. The purpose was to investigate the behaviour of the estimated polynomial parameters and the feasibility of using a minimum of bands to estimate EOPs. Simulated and real data were produced for the analysis of parameters and accuracy in ground points. The tests considered conventional bundle adjustment and the polynomial models. The results showed that both techniques were comparable, indicating that the time-dependent polynomial model can be used to estimate the EOPs of all spectral bands, without requiring a bundle adjustment of each band. The accuracy of the block adjustment was analysed based on the discrepancy obtained from checkpoints. The root mean square error (RMSE) indicated an accuracy of 1&amp;thinsp;GSD in planimetry and 1.5&amp;thinsp;GSD in altimetry, when using a minimum of four bands per cube.


2020 ◽  
Vol 4 ◽  
pp. 128
Author(s):  
Per Liljenberg

Background: For diseases like Covid-19, where it has been difficult to identify the true number of infected people, or where the number of known cases is heavily influenced by the number of tests performed, hospitalizations and deaths play a significant role in understanding the epidemic and in determining the appropriate response. However, the Covid-19 deaths data reported by some countries display a significant weekly variability, which can make the interpretation and use of the death data in analysis and modeling difficult. Methods: We derive the mathematical relationship between the series of new daily deaths by reporting date and the series of deaths by death date. We then apply this formalism to the corresponding time-series reported by Sweden during the Covid-19 pandemic. Results: The practice of reporting new deaths daily, as is standard procedure during an outbreak in most countries and regions, should be viewed as a time-dependent filter, modulating the underlying true death curve. After having characterized the Swedish reporting process, we show how smoothing of the Swedish reported daily deaths series results in a curve distinctly different from the true death curve. We also comment on the use of nowcasting methods. Conclusions: Modelers and analysts using the series of new daily deaths by reporting date should take extra care when it is highly variable and when there is a significant reporting delay. It might be appropriate to instead use the series of deaths by death date combined with a nowcasting algorithm as basis for their analysis.


Author(s):  
Sachin Kumar ◽  
Karan Veer

Aims: The objective of this research is to predict the covid-19 cases in India based on the machine learning approaches. Background: Covid-19, a respiratory disease caused by one of the coronavirus family members, has led to a pandemic situation worldwide in 2020. This virus was detected firstly in Wuhan city of China in December 2019. This viral disease has taken less than three months to spread across the globe. Objective: In this paper, we proposed a regression model based on the Support vector machine (SVM) to forecast the number of deaths, the number of recovered cases, and total confirmed cases for the next 30 days. Method: For prediction, the data is collected from Github and the ministry of India's health and family welfare from March 14, 2020, to December 3, 2020. The model has been designed in Python 3.6 in Anaconda to forecast the forecasting value of corona trends until September 21, 2020. The proposed methodology is based on the prediction of values using SVM based regression model with polynomial, linear, rbf kernel. The dataset has been divided into train and test datasets with 40% and 60% test size and verified with real data. The model performance parameters are evaluated as a mean square error, mean absolute error, and percentage accuracy. Results and Conclusion: The results show that the polynomial model has obtained 95 % above accuracy score, linear scored above 90%, and rbf scored above 85% in predicting cumulative death, conformed cases, and recovered cases.


Author(s):  
Mariana Damova ◽  
Atanas Kiryakov ◽  
Maurice Grinberg ◽  
Michael K. Bergman ◽  
Frédérick Giasson ◽  
...  

The chapter introduces the process of design of two upper-level ontologies—PROTON and UMBEL—into reference ontologies and their integration in the so-called Reference Knowledge Stack (RKS). It is argued that RKS is an important step in the efforts of the Linked Open Data (LOD) project to transform the Web into a global data space with diverse real data, available for review and analysis. RKS is intended to make the interoperability between published datasets much more efficient than it is now. The approach discussed in the chapter consists of developing reference layers of upper-level ontologies by mapping them to certain LOD schemata and assigning instance data to them so they cover a reasonable portion of the LOD datasets. The chapter presents the methods (manual and semi-automatic) used in the creation of the RKS and gives examples that illustrate its advantages for managing highly heterogeneous data and its usefulness in real life knowledge intense applications.


Author(s):  
Khayra Bencherif ◽  
Mimoun Malki ◽  
Djamel Amar Bensaber

This article describes how the Linked Open Data Cloud project allows data providers to publish structured data on the web according to the Linked Data principles. In this context, several link discovery frameworks have been developed for connecting entities contained in knowledge bases. In order to achieve a high effectiveness for the link discovery task, a suitable link configuration is required to specify the similarity conditions. Unfortunately, such configurations are specified manually; which makes the link discovery task tedious and more difficult for the users. In this article, the authors address this drawback by proposing a novel approach for the automatic determination of link specifications. The proposed approach is based on a neural network model to combine a set of existing metrics into a compound one. The authors evaluate the effectiveness of the proposed approach in three experiments using real data sets from the LOD Cloud. In addition, the proposed approach is compared against link specifications approaches to show that it outperforms them in most experiments.


Sign in / Sign up

Export Citation Format

Share Document