scholarly journals A mobile crowd sensing framework for suspect investigation: An objectivity analysis and de-identification approach

2020 ◽  
Vol 17 (1) ◽  
pp. 253-269
Author(s):  
Alaoui El ◽  
Fazziki El ◽  
Fatima Ennaji ◽  
Mohamed Sadgal

The ubiquity of mobile devices and their advanced features have increased the use of crowdsourcing in many areas, such as the mobility in the smart cities. With the advent of high-quality sensors on smartphones, online communities can easily collect and share information. These information are of great importance for the institutions, which must analyze the facts by facilitating the data collecting on crimes and criminals, for example. This paper proposes an approach to develop a crowdsensing framework allowing a wider collaboration between the citizens and the authorities. In addition, this framework takes advantage of an objectivity analysis to ensure the participants? credibility and the information reliability, as law enforcement is often affected by unreliable and poor quality data. In addition, the proposed framework ensures the protection of users' private data through a de-identification process. Experimental results show that the proposed framework is an interesting tool to improve the quality of crowdsensing information in a government context.

10.28945/2584 ◽  
2002 ◽  
Author(s):  
Herna L. Viktor ◽  
Wayne Motha

Increasingly, large organizations are engaging in data warehousing projects in order to achieve a competitive advantage through the exploration of the information as contained therein. It is therefore paramount to ensure that the data warehouse includes high quality data. However, practitioners agree that the improvement of the quality of data in an organization is a daunting task. This is especially evident in data warehousing projects, which are often initiated “after the fact”. The slightest suspicion of poor quality data often hinders managers from reaching decisions, when they waste hours in discussions to determine what portion of the data should be trusted. Augmenting data warehousing with data mining methods offers a mechanism to explore these vast repositories, enabling decision makers to assess the quality of their data and to unlock a wealth of new knowledge. These methods can be effectively used with inconsistent, noisy and incomplete data that are commonplace in data warehouses.


2006 ◽  
Vol 21 (1) ◽  
pp. 67-70 ◽  
Author(s):  
Brian H. Toby

The definitions for important Rietveld error indices are defined and discussed. It is shown that while smaller error index values indicate a better fit of a model to the data, wrong models with poor quality data may exhibit smaller values error index values than some superb models with very high quality data.


Author(s):  
Vrushali Gajanan Kadam ◽  
Sharvari Chandrashekhar Tamane ◽  
Vijender Kumar Solanki

The world is growing and energy conservation is a very important challenge for the engineering domain. The emergence of smart cities is one possible solution for the same, as it claims that energy and resources are saved in the smart city infrastructure. This chapter is divided into five sections. Section 1 gives the past, present, and future of the living style. It gives the representation from rural, urban, to smart city. Section 2 gives the explanations of four pillars of big data, and through grid, a big data analysis is presented in the chapter. Section 3 started with the case study on smart grid. It comprises traffic congestion and their prospective solution through big data analytics. Section 4 starts from the mobile crowd sensing. It discusses a good elaboration on crowd sensing whereas Section 5 discusses the smart city approach. Important issues like lighting, parking, and traffic were taken into consideration.


2019 ◽  
pp. 23-34
Author(s):  
Harvey Goldstein ◽  
Ruth Gilbert

his chapter addresses data linkage which is key to using big administrative datasets to improve efficient and equitable services and policies. These benefits need to weigh against potential harms, which have mainly focussed on privacy. In this chapter we argue for the public and researchers to be alert also to other kinds of harms. These include misuses of big administrative data through poor quality data, misleading analyses, misinterpretation or misuse of findings, and restrictions limiting what questions can be asked and by whom, resulting in research not achieved and advances not made for the public benefit. Ensuring that big administrative data are validly used for public benefit requires increased transparency about who has access and whose access is denied, how data are processed, linked and analysed, and how analyses or algorithms are used in public and private services. Public benefits and especially trust require replicable analyses by many researchers not just a few data controllers. Wider use of big data will be helped by establishing a number of safe data repositories, fully accessible to researchers and their tools, and independent of the current monopolies on data processing, linkage, enhancement and uses of data.


2017 ◽  
Vol 49 (4) ◽  
pp. 415-424 ◽  
Author(s):  
Susan WILL-WOLF ◽  
Sarah JOVAN ◽  
Michael C. AMACHER

AbstractLichen element content is a reliable indicator for relative air pollution load in research and monitoring programmes requiring both efficiency and representation of many sites. We tested the value of costly rigorous field and handling protocols for sample element analysis using five lichen species. No relaxation of rigour was supported; four relaxed protocols generated data significantly different from rigorous protocols for many of the 20 validated elements. Minimally restrictive site selection criteria gave quality data from 86% of 81 permanent plots in northern Midwest USA; more restrictive criteria would likely reduce indicator reliability. Use of trained non-specialist field collectors was supported when target species choice considers the lichen community context. Evernia mesomorpha, Flavoparmelia caperata and Physcia aipolia/stellaris were successful target species. Non-specialists were less successful at distinguishing Parmelia sulcata and Punctelia rudecta from lookalikes, leading to few samples and some poor quality data.


2017 ◽  
Vol 2017 ◽  
pp. 1-16 ◽  
Author(s):  
Aamir Shakoor ◽  
Zahid Mahmood Khan ◽  
Muhammad Arshad ◽  
Hafiz Umar Farid ◽  
Muhammad Sultan ◽  
...  

The intensive abstraction of groundwater is causing a number of problems such as groundwater depletion and quality deterioration. To manage such problems, the data of 256 piezometers regarding groundwater levels and quality were acquired for the period of 2003 to 2012 in command area of Lower Chenab Canal (LCC), West Faisalabad, Pakistan. MODFLOW and MT3D models were calibrated for the period of 2003–2007 and validated for years 2008–2012 with respect to observed groundwater levels and quality data, respectively. After the successful calibration and validation, two pumping scenarios were developed up to year 2030: Scenario I (increase in pumping rate according to the historical trend) and Scenario II (adjusted canal water supplies and groundwater patterns). The predicted results of Scenario I revealed that, up to year 2030, the area under good quality groundwater reduced significantly from 50.35 to 28.95%, while marginal and hazardous groundwater quality area increased from 49.65 to 71.06%. Under Scenario II, the good quality groundwater area increased to 6.32% and 12.48% area possesses less hazardous quality of groundwater. It was concluded that the canal water supply should shift from good quality aquifer zone to poor quality aquifer zone for proficient management of groundwater at the study area.


Geophysics ◽  
2008 ◽  
Vol 73 (2) ◽  
pp. E51-E57 ◽  
Author(s):  
Jack P. Dvorkin

Laboratory data supported by granular-medium and inclusion theories indicate that Poisson’s ratio in gas-saturated sand lies within a range of 0–0.25, with typical values of approximately 0.15. However, some well log measurements, especially in slow gas formations, persistently produce a Poisson’s ratio as large as 0.3. If this measurement is not caused by poor-quality data, three in situ situations — patchy saturation, subresolution thin layering, and elastic anisotropy — provide a plausible explanation. In the patchy saturation situation, the well data must be corrected to produce realistic synthetic seismic traces. In the second and third cases, the effect observed in a well is likely to persist at the seismic scale.


Sign in / Sign up

Export Citation Format

Share Document