scholarly journals Understanding Collective Human Mobility Spatiotemporal Patterns on Weekdays from Taxi Origin-Destination Point Data

Sensors ◽  
2019 ◽  
Vol 19 (12) ◽  
pp. 2812 ◽  
Author(s):  
Jing Yang ◽  
Yizhong Sun ◽  
Bowen Shang ◽  
Lei Wang ◽  
Jie Zhu

With the availability of large geospatial datasets, the study of collective human mobility spatiotemporal patterns provides a new way to explore urban spatial environments from the perspective of residents. In this paper, we constructed a classification model for mobility patterns that is suitable for taxi OD (Origin-Destination) point data, and it is comprised of three parts. First, a new aggregate unit, which uses a road intersection as the constraint condition, is designed for the analysis of the taxi OD point data. Second, the time series similarity measurement is improved by adding a normalization procedure and time windows to address the particular characteristics of the taxi time series data. Finally, the DBSCAN algorithm is used to classify the time series into different mobility patterns based on a proximity index that is calculated using the improved similarity measurement. In addition, we used the random forest algorithm to establish a correlation model between the mobility patterns and the regional functional characteristics. Based on the taxi OD point data from Nanjing, we delimited seven mobility patterns and illustrated that the regional functions have obvious driving effects on these mobility patterns. These findings are applicable to urban planning, traffic management and planning, and land use analyses in the future.

Sensors ◽  
2021 ◽  
Vol 21 (10) ◽  
pp. 3553
Author(s):  
Jeremy Watts ◽  
Anahita Khojandi ◽  
Rama Vasudevan ◽  
Fatta B. Nahab ◽  
Ritesh A. Ramdhani

Parkinson’s disease medication treatment planning is generally based on subjective data obtained through clinical, physician-patient interactions. The Personal KinetiGraph™ (PKG) and similar wearable sensors have shown promise in enabling objective, continuous remote health monitoring for Parkinson’s patients. In this proof-of-concept study, we propose to use objective sensor data from the PKG and apply machine learning to cluster patients based on levodopa regimens and response. The resulting clusters are then used to enhance treatment planning by providing improved initial treatment estimates to supplement a physician’s initial assessment. We apply k-means clustering to a dataset of within-subject Parkinson’s medication changes—clinically assessed by the MDS-Unified Parkinson’s Disease Rating Scale-III (MDS-UPDRS-III) and the PKG sensor for movement staging. A random forest classification model was then used to predict patients’ cluster allocation based on their respective demographic information, MDS-UPDRS-III scores, and PKG time-series data. Clinically relevant clusters were partitioned by levodopa dose, medication administration frequency, and total levodopa equivalent daily dose—with the PKG providing similar symptomatic assessments to physician MDS-UPDRS-III scores. A random forest classifier trained on demographic information, MDS-UPDRS-III scores, and PKG time-series data was able to accurately classify subjects of the two most demographically similar clusters with an accuracy of 86.9%, an F1 score of 90.7%, and an AUC of 0.871. A model that relied solely on demographic information and PKG time-series data provided the next best performance with an accuracy of 83.8%, an F1 score of 88.5%, and an AUC of 0.831, hence further enabling fully remote assessments. These computational methods demonstrate the feasibility of using sensor-based data to cluster patients based on their medication responses with further potential to assist with medication recommendations.


2021 ◽  
Author(s):  
Tetsuya Yamada ◽  
Shoi Shi

Comprehensive and evidence-based countermeasures against emerging infectious diseases have become increasingly important in recent years. COVID-19 and many other infectious diseases are spread by human movement and contact, but complex transportation networks in 21 century make it difficult to predict disease spread in rapidly changing situations. It is especially challenging to estimate the network of infection transmission in the countries that the traffic and human movement data infrastructure is not yet developed. In this study, we devised a method to estimate the network of transmission of COVID-19 from the time series data of its infection and applied it to determine its spread across areas in Japan. We incorporated the effects of soft lockdowns, such as the declaration of a state of emergency, and changes in the infection network due to government-sponsored travel promotion, and predicted the spread of infection using the Tokyo Olympics as a model. The models used in this study are available online, and our data-driven infection network models are scalable, whether it be at the level of a city, town, country, or continent, and applicable anywhere in the world, as long as the time-series data of infections per region is available. These estimations of effective distance and the depiction of infectious disease networks based on actual infection data are expected to be useful in devising data-driven countermeasures against emerging infectious diseases worldwide.


2018 ◽  
Vol 11 (4) ◽  
pp. 486-495
Author(s):  
Ke Yi Zhou ◽  
Shaolin Hu

Purpose The similarity measurement of time series is an important research in time series detection, which is a basic work of time series clustering, anomaly discovery, prediction and many other data mining problems. The purpose of this paper is to design a new similarity measurement algorithm to improve the performance of the original similarity measurement algorithm. The subsequence morphological information is taken into account by the proposed algorithm, and time series is represented by a pattern, so the similarity measurement algorithm is more accurate. Design/methodology/approach Following some previous researches on similarity measurement, an improved method is presented. This new method combines morphological representation and dynamic time warping (DTW) technique to measure the similarities of time series. After the segmentation of time series data into segments, three parameter values of median, point number and slope are introduced into the improved distance measurement formula. The effectiveness of the morphological weighted DTW algorithm (MW-DTW) is demonstrated by the example of momentum wheel data of an aircraft attitude control system. Findings The improved method is insensitive to the distortion and expansion of time axis and can be used to detect the morphological changes of time series data. Simulation results confirm that this method proposed in this paper has a high accuracy of similarity measurement. Practical implications This improved method has been used to solve the problem of similarity measurement in time series, which is widely emerged in different fields of science and engineering, such as the field of control, measurement, monitoring, process signal processing and economic analysis. Originality/value In the similarity measurement of time series, the distance between sequences is often used as the only detection index. The results of similarity measurement should not be affected by the longitudinal or transverse stretching and translation changes of the sequence, so it is necessary to incorporate the morphological changes of the sequence into similarity measurement. The MW-DTW is more suitable for the actual situation. At the same time, the MW-DTW algorithm reduces the computational complexity by transforming the computational object to subsequences.


2021 ◽  
Vol 5 (5) ◽  
pp. 619-635
Author(s):  
Harya Widiputra

The primary factor that contributes to the transmission of COVID-19 infection is human mobility. Positive instances added on a daily basis have a substantial positive association with the pace of human mobility, and the reverse is true. Thus, having the ability to predict human mobility trend during a pandemic is critical for policymakers to help in decreasing the rate of transmission in the future. In this regard, one approach that is commonly used for time-series data prediction is to build an ensemble with the aim of getting the best performance. However, building an ensemble often causes the performance of the model to decrease, due to the increasing number of parameters that are not being optimized properly. Consequently, the purpose of this study is to develop and evaluate a deep learning ensemble model, which is optimized using a genetic algorithm (GA) that incorporates a convolutional neural network (CNN) and a long short-term memory (LSTM). A CNN is used to conduct feature extraction from mobility time-series data, while an LSTM is used to do mobility prediction. The parameters of both layers are adjusted using GA. As a result of the experiments conducted with data from the Google Community Mobility Reports in Indonesia that ranges from the beginning of February 2020 to the end of December 2020, the GA-Optimized Multivariate CNN-LSTM ensemble outperforms stand-alone CNN and LSTM models, as well as the non-optimized CNN-LSTM model, in terms of predicting human movement in the future. This may be useful in assisting policymakers in anticipating future human mobility trends. Doi: 10.28991/esj-2021-01300 Full Text: PDF


2021 ◽  
Vol 14 (1) ◽  
pp. 140
Author(s):  
Johann Desloires ◽  
Dino Ienco ◽  
Antoine Botrel ◽  
Nicolas Ranc

Applications in which researchers aim to extract a single land type from remotely sensed data are quite common in practical scenarios: extract the urban footprint to make connections with socio-economic factors; map the forest extent to subsequently retrieve biophysical variables and detect a particular crop type to successively calibrate and deploy yield prediction models. In this scenario, the (positive) targeted class is well defined, while the negative class is difficult to describe. This one-class classification setting is also referred to as positive unlabelled learning (PUL) in the general field of machine learning. To deal with this challenging setting, when satellite image time series data are available, we propose a new framework named positive and unlabelled learning of satellite image time series (PUL-SITS). PUL-SITS involves two different stages: In the first one, a recurrent neural network autoencoder is trained to reconstruct only positive samples with the aim to higight reliable negative ones. In the second stage, both labelled and unlabelled samples are exploited in a semi-supervised manner to build the final binary classification model. To assess the quality of our approach, experiments were carried out on a real-world benchmark, namely Haute-Garonne, located in the southwest area of France. From this study site, we considered two different scenarios: a first one in which the process has the objective to map Cereals/Oilseeds cover versus the rest of the land cover classes and a second one in which the class of interest is the Forest land cover. The evaluation was carried out by comparing the proposed approach with recent competitors to deal with the considered positive and unlabelled learning scenarios.


2021 ◽  
Vol 2021 ◽  
pp. 1-11
Author(s):  
Zhongguo Yang ◽  
Irshad Ahmed Abbasi ◽  
Fahad Algarni ◽  
Sikandar Ali ◽  
Mingzhu Zhang

Nowadays, an Internet of Things (IoT) device consists of algorithms, datasets, and models. Due to good performance of deep learning methods, many devices integrated well-trained models in them. IoT empowers users to communicate and control physical devices to achieve vital information. However, these models are vulnerable to adversarial attacks, which largely bring potential risks to the normal application of deep learning methods. For instance, very little changes even one point in the IoT time-series data could lead to unreliable or wrong decisions. Moreover, these changes could be deliberately generated by following an adversarial attack strategy. We propose a robust IoT data classification model based on an encode-decode joint training model. Furthermore, thermometer encoding is taken as a nonlinear transformation to the original training examples that are used to reconstruct original time series examples through the encode-decode model. The trained ResNet model based on reconstruction examples is more robust to the adversarial attack. Experiments show that the trained model can successfully resist to fast gradient sign method attack to some extent and improve the security of the time series data classification model.


2020 ◽  
Vol 10 (24) ◽  
pp. 9050
Author(s):  
Ron Kremser ◽  
Niclas Grabowski ◽  
Roman Düssel ◽  
Albert Mulder ◽  
Dietmar Tutsch

In aluminium production, anode effects occur when the alumina content in the bath is so low that normal fused salt electrolysis cannot be maintained. This is followed by a rapid increase of pot voltage from about 4.3 V to values in the range from 10 to 80 V. As a result of a local depletion of oxide ions, the cryolite decomposes and forms climate-relevant perfluorocarbon (PFC) gases. The high pot voltage also causes a high energy input, which dissipates as heat. In order to ensure energy-efficient and climate-friendly operation, it is important to predict anode effects in advance so that they can be prevented by prophylactic actions like alumina feeding or beam downward movements. In this paper a classification model is trained with aggregated time series data from TRIMET Aluminium SE Essen (TAE) that is able to predict anode effects at least 1 min in advance. Due to a high imbalance in the class distribution of normal state and labeled anode effect state as well as possible model’s weaknesses the final F1 score of 32.4% is comparatively low. Nevertheless, the prediction provides an indication of possible anode effects and the process control system may react on it. Consequent practical implications will be discussed.


2008 ◽  
pp. 3272-3284
Author(s):  
Sagar Savla ◽  
Sharma Chakravarthy

Sensor-based applications, such as smart homes, require prediction of event occurrences for automating the environment using time-series data collected over a period of time. In these applications, it is important to predict events in tight and accurate intervals to effectively automate the application. This article deals with the discovery of significant intervals from time-series data. Although there is a considerable body of work on sequential mining of transactional data, most of them deal with time-point data and make several passes over the entire data set in order to discover frequently occurring patterns/events. We propose an approach in which significant intervals representing intrinsic nature of data are discovered in a single pass. In our approach, time-series data is folded over a periodicity (day, week, etc.) in which the intervals are formed. Significant intervals are discovered from this interval data that satisfy the criteria of minimum confidence and maximum interval length specified by the user. Both compression and working with intervals contribute towards improving the efficiency of the algorithm. In this article, we present a new single-pass algorithm for detecting significant intervals; discuss its characteristics, advantages, and disadvantages; and analyze it. Finally, we compare the performance of our algorithm with previously developed level-wise and SQL-based algorithms for significant interval discovery (SID).


Sign in / Sign up

Export Citation Format

Share Document