scholarly journals An IoT System for Social Distancing and Emergency Management in Smart Cities Using Multi-Sensor Data

Algorithms ◽  
2020 ◽  
Vol 13 (10) ◽  
pp. 254
Author(s):  
Rosario Fedele ◽  
Massimo Merenda

Smart cities need technologies that can be really applied to raise the quality of life and environment. Among all the possible solutions, Internet of Things (IoT)-based Wireless Sensor Networks (WSNs) have the potentialities to satisfy multiple needs, such as offering real-time plans for emergency management (due to accidental events or inadequate asset maintenance) and managing crowds and their spatiotemporal distribution in highly populated areas (e.g., cities or parks) to face biological risks (e.g., from a virus) by using strategies such as social distancing and movement restrictions. Consequently, the objective of this study is to present an IoT system, based on an IoT-WSN and on algorithms (Neural Network, NN, and Shortest Path Finding) that are able to recognize alarms, available exits, assembly points, safest and shortest paths, and overcrowding from real-time data gathered by sensors and cameras exploiting computer vision. Subsequently, this information is sent to mobile devices using a web platform and the Near Field Communication (NFC) technology. The results refer to two different case studies (i.e., emergency and monitoring) and show that the system is able to provide customized strategies and to face different situations, and that this is also applies in the case of a connectivity shutdown.

2020 ◽  
Vol 10 (17) ◽  
pp. 5882
Author(s):  
Federico Desimoni ◽  
Sergio Ilarri ◽  
Laura Po ◽  
Federica Rollo ◽  
Raquel Trillo-Lado

Modern cities face pressing problems with transportation systems including, but not limited to, traffic congestion, safety, health, and pollution. To tackle them, public administrations have implemented roadside infrastructures such as cameras and sensors to collect data about environmental and traffic conditions. In the case of traffic sensor data not only the real-time data are essential, but also historical values need to be preserved and published. When real-time and historical data of smart cities become available, everyone can join an evidence-based debate on the city’s future evolution. The TRAFAIR (Understanding Traffic Flows to Improve Air Quality) project seeks to understand how traffic affects urban air quality. The project develops a platform to provide real-time and predicted values on air quality in several cities in Europe, encompassing tasks such as the deployment of low-cost air quality sensors, data collection and integration, modeling and prediction, the publication of open data, and the development of applications for end-users and public administrations. This paper explicitly focuses on the modeling and semantic annotation of traffic data. We present the tools and techniques used in the project and validate our strategies for data modeling and its semantic enrichment over two cities: Modena (Italy) and Zaragoza (Spain). An experimental evaluation shows that our approach to publish Linked Data is effective.


Sensors ◽  
2018 ◽  
Vol 18 (9) ◽  
pp. 2994 ◽  
Author(s):  
Bhagya Silva ◽  
Murad Khan ◽  
Changsu Jung ◽  
Jihun Seo ◽  
Diyan Muhammad ◽  
...  

The Internet of Things (IoT), inspired by the tremendous growth of connected heterogeneous devices, has pioneered the notion of smart city. Various components, i.e., smart transportation, smart community, smart healthcare, smart grid, etc. which are integrated within smart city architecture aims to enrich the quality of life (QoL) of urban citizens. However, real-time processing requirements and exponential data growth withhold smart city realization. Therefore, herein we propose a Big Data analytics (BDA)-embedded experimental architecture for smart cities. Two major aspects are served by the BDA-embedded smart city. Firstly, it facilitates exploitation of urban Big Data (UBD) in planning, designing, and maintaining smart cities. Secondly, it occupies BDA to manage and process voluminous UBD to enhance the quality of urban services. Three tiers of the proposed architecture are liable for data aggregation, real-time data management, and service provisioning. Moreover, offline and online data processing tasks are further expedited by integrating data normalizing and data filtering techniques to the proposed work. By analyzing authenticated datasets, we obtained the threshold values required for urban planning and city operation management. Performance metrics in terms of online and offline data processing for the proposed dual-node Hadoop cluster is obtained using aforementioned authentic datasets. Throughput and processing time analysis performed with regard to existing works guarantee the performance superiority of the proposed work. Hence, we can claim the applicability and reliability of implementing proposed BDA-embedded smart city architecture in the real world.


2021 ◽  
Author(s):  
Goedele Verreydt ◽  
Niels Van Putte ◽  
Timothy De Kleyn ◽  
Joris Cool ◽  
Bino Maiheu

<p>Groundwater dynamics play a crucial role in the spreading of a soil and groundwater contamination. However, there is still a big gap in the understanding of the groundwater flow dynamics. Heterogeneities and dynamics are often underestimated and therefore not taken into account. They are of crucial input for successful management and remediation measures. The bulk of the mass of mass often is transported through only a small layer or section within the aquifer and is in cases of seepage into surface water very dependent to rainfall and occurring tidal effects.</p><p> </p><p>This study contains the use of novel real-time iFLUX sensors to map the groundwater flow dynamics over time. The sensors provide real-time data on groundwater flow rate and flow direction. The sensor probes consist of multiple bidirectional flow sensors that are superimposed. The probes can be installed directly in the subsoil, riverbed or monitoring well. The measurement setup is unique as it can perform measurements every second, ideal to map rapid changing flow conditions. The measurement range is between 0,5 and 500 cm per day.</p><p> </p><p>We will present the measurement principles and technical aspects of the sensor, together with two case studies.</p><p> </p><p>The first case study comprises the installation of iFLUX sensors in 4 different monitoring wells in a chlorinated solvent plume to map on the one hand the flow patterns in the plume, and on the other hand the flow dynamics that are influenced by the nearby popular trees. The foreseen remediation concept here is phytoremediation. The sensors were installed for a period of in total 4 weeks. Measurement frequency was 5 minutes. The flow profiles and time series will be presented together with the determined mass fluxes.</p><p> </p><p>A second case study was performed on behalf of the remediation of a canal riverbed. Due to industrial production of tar and carbon black in the past, the soil and groundwater next to the small canal ‘De Lieve’ in Ghent, Belgium, got contaminated with aliphatic and (poly)aromatic hydrocarbons. The groundwater contaminants migrate to the canal, impact the surface water quality and cause an ecological risk. The seepage flow and mass fluxes of contaminants into the surface water were measured with the novel iFLUX streambed sensors, installed directly in the river sediment. A site conceptual model was drawn and dimensioned based on the sensor data. The remediation concept to tackle the inflowing pollution: a hydraulic conductive reactive mat on the riverbed that makes use of the natural draining function of the waterbody, the adsorption capacity of a natural or secondary adsorbent and a future habitat for micro-organisms that biodegrade contaminants. The reactive mats were successfully installed and based on the mass flux calculations a lifespan of at least 10 years is expected for the adsorption material.  </p>


2015 ◽  
Vol 2015 ◽  
pp. 1-14 ◽  
Author(s):  
Woochul Kang ◽  
Jaeyong Chung

With ubiquitous deployment of sensors and network connectivity, amounts of real-time data for embedded systems are increasing rapidly and database capability is required for many embedded systems for systematic management of real-time data. In such embedded systems, supporting the timeliness of tasks accessing databases is an important problem. However, recent multicore-based embedded architectures pose a significant challenge for such data-intensive real-time tasks since the response time of accessing data can be significantly affected by potential intercore interferences. In this paper, we propose a novel feedback control scheme that supports the timeliness of data-intensive tasks against unpredictable intercore interferences. In particular, we use multiple inputs/multiple outputs (MIMO) control method that exploits multiple control knobs, for example, CPU frequency and the Quality-of-Data (QoD) to handle highly unpredictable workloads in multicore systems. Experimental results, using actual implementation, show that the proposed approach achieves the target Quality-of-Service (QoS) goals, such as task timeliness and Quality-of-Data (QoD) while consuming less energy compared to baseline approaches.


2009 ◽  
Vol 26 (3) ◽  
pp. 556-569 ◽  
Author(s):  
Ananda Pascual ◽  
Christine Boone ◽  
Gilles Larnicol ◽  
Pierre-Yves Le Traon

Abstract The timeliness of satellite altimeter measurements has a significant effect on their value for operational oceanography. In this paper, an Observing System Experiment (OSE) approach is used to assess the quality of real-time altimeter products, a key issue for robust monitoring and forecasting of the ocean state. In addition, the effect of two improved geophysical corrections and the number of missions that are combined in the altimeter products are also analyzed. The improved tidal and atmospheric corrections have a significant effect in coastal areas (0–100 km from the shore), and a comparison with tide gauge observations shows a slightly better agreement with the gridded delayed-time sea level anomalies (SLAs) with two altimeters [Jason-1 and European Remote Sensing Satellite-2 (ERS-2)/Envisat] using the new geophysical corrections (mean square differences in percent of tide gauge variance of 35.3%) than those with four missions [Jason-1, ERS/Envisat, Ocean Topography Experiment (TOPEX)/Poseidoninterlaced, and Geosat Follow-On] but using the old corrections (36.7%). In the deep ocean, however, the correction improvements have little influence. The performance of fast delivery products versus delayed-time data is compared using independent in situ data (tide gauge and drifter data). It clearly highlights the degradation of real-time SLA maps versus the delayed-time SLA maps: four altimeters are needed in real time to get the similar quality performance as two altimeters in delayed time (sea level error misfit around 36%, and zonal and meridional velocity estimation errors of 27% and 33%, respectively). This study proves that the continuous improvement of geophysical corrections is very important, and that it is essential to stay above a minimum threshold of four available altimetric missions to capture the main space and time oceanic scales in fast delivery products.


Author(s):  
Manjunath Ramachandra ◽  
Vikas Jain

The present day Internet traffic largely caters for the multimedia traffic throwing open new and unthinkable applications such as tele-surgery. The complexity of data transactions increases with a demand for in time and real time data transfers, demanding the limited resources of the network beyond their capabilities. It requires a prioritization of data transfers, controlled dumping of data over the network etc. To make the matter worse, the data from different origin combine together imparting long lasting detrimental features such as self similarity and long range dependency in to the traffic. The multimedia data fortunately is associated with redundancies that may be removed through efficient compression techniques. There exists a provision to control the compression or bitrates based on the availability of resources in the network. The traffic controller or shaper has to optimize the quality of the transferred multimedia data depending up on the state of the network. In this chapter, a novel traffic shaper is introduced considering the adverse properties of the network and counteract with the same.


Author(s):  
Suresh P. ◽  
Keerthika P. ◽  
Sathiyamoorthi V. ◽  
Logeswaran K. ◽  
Manjula Devi R. ◽  
...  

Cloud computing and big data analytics are the key parts of smart city development that can create reliable, secure, healthier, more informed communities while producing tremendous data to the public and private sectors. Since the various sectors of smart cities generate enormous amounts of streaming data from sensors and other devices, storing and analyzing this huge real-time data typically entail significant computing capacity. Most smart city solutions use a combination of core technologies such as computing, storage, databases, data warehouses, and advanced technologies such as analytics on big data, real-time streaming data, artificial intelligence, machine learning, and the internet of things (IoT). This chapter presents a theoretical and experimental perspective on the smart city services such as smart healthcare, water management, education, transportation and traffic management, and smart grid that are offered using big data management and cloud-based analytics services.


2020 ◽  
Vol 12 (23) ◽  
pp. 10175
Author(s):  
Fatima Abdullah ◽  
Limei Peng ◽  
Byungchul Tak

The volume of streaming sensor data from various environmental sensors continues to increase rapidly due to wider deployments of IoT devices at much greater scales than ever before. This, in turn, causes massive increase in the fog, cloud network traffic which leads to heavily delayed network operations. In streaming data analytics, the ability to obtain real time data insight is crucial for computational sustainability for many IoT enabled applications such as environmental monitors, pollution and climate surveillance, traffic control or even E-commerce applications. However, such network delays prevent us from achieving high quality real-time data analytics of environmental information. In order to address this challenge, we propose the Fog Sampling Node Selector (Fossel) technique that can significantly reduce the IoT network and processing delays by algorithmically selecting an optimal subset of fog nodes to perform the sensor data sampling. In addition, our technique performs a simple type of query executions within the fog nodes in order to further reduce the network delays by processing the data near the data producing devices. Our extensive evaluations show that Fossel technique outperforms the state-of-the-art in terms of latency reduction as well as in bandwidth consumption, network usage and energy consumption.


2020 ◽  
Vol 12 (13) ◽  
pp. 5324 ◽  
Author(s):  
Arabela Briciu ◽  
Victor-Alexandru Briciu ◽  
Androniki Kavoura

Global urbanization brings the urge to identify the most intelligent methods to cope with the challenges arising in the modern society. Sustainable and smart cities are the new target for urban development; their representatives are being forced to identify and develop new strategies to increase their city’s performance and ensure that it endures over time. Information and Communication Technology (ICT) contributes to this purpose. Brașov is one of the cities in Romania for which the process of modernization started years ago, and it is currently developing as a smart city. This paper focuses on the development of the city in terms of cultural tourism solutions by presenting a case study on the use of virtual reality with a mobile application and its evaluation on cultural heritage sites. The original contribution of the paper is to describe and analyze the quality of the mobile application by using a proposed analysis grid to identify the main elements of this app. The main findings suggest that the application may bring authenticity of experience through the lens of heritage preservation for further user engagement and participation in real-time, while suggestions are made for future enhancement. Implications are discussed for a) destination managers, b) for developers to improve the general quality of the mobile application in terms of design and features and to implement changes in the near future, and c) for visitors who engage in real-time and co-create experiences.


Sign in / Sign up

Export Citation Format

Share Document