scholarly journals Data Reliability in a Citizen Science Protocol for Monitoring Stingless Bees Flight Activity

Insects ◽  
2021 ◽  
Vol 12 (9) ◽  
pp. 766
Author(s):  
Jailson N. Leocadio ◽  
Natalia P. Ghilardi-Lopes ◽  
Sheina Koffler ◽  
Celso Barbiéri ◽  
Tiago M. Francoy ◽  
...  

Although the quality of citizen science (CS) data is often a concern, evidence for high-quality CS data increases in the scientific literature. This study aimed to assess the data reliability of a structured CS protocol for monitoring stingless bees’ flight activity. We tested (1) data accuracy for replication among volunteers and for expert validation and (2) precision, comparing dispersion between citizen scientists and expert data. Two distinct activity dimensions were considered: (a) perception of flight activity and (b) flight activity counts (entrances, exits, and pollen load). No significant differences were found among groups regarding entrances and exits. However, replicator citizen scientists presented a higher chance of perceiving pollen than original data collectors and experts, likely a false positive. For those videos in which there was an agreement about pollen presence, the effective pollen counts were similar (with higher dispersion for citizen scientists), indicating the reliability of CS-collected data. The quality of the videos, a potential source of variance, did not influence the results. Increasing practical training could be an alternative to improve pollen data quality. Our study shows that CS provides reliable data for monitoring bee activity and highlights the relevance of a multi-dimensional approach for assessing CS data quality.

2018 ◽  
Vol 2 ◽  
pp. e26665
Author(s):  
Alan Stenhouse ◽  
Philip Roetman ◽  
Frank Grützner ◽  
Tahlia Perry ◽  
Lian Pin Koh

Field data collection by Citizen Scientists has been hugely assisted by the rapid development and spread of smart phones as well as apps that make use of the integrated technologies contained in these devices. We can improve the quality of the data by increasing utilisation of the device in-built sensors and improving the software user-interface. Improvements to data timeliness can be made by integrating directly with national and international biodiversity repositories, such as the Atlas of Living Australia (ALA). I will present two Citizen Science apps that we developed for the conservation of two of Australia’s iconic species – the koala and the echidna. First is the Koala Counter app used in the Great Koala Count 2 – a two-day Blitz-style population census. The aim was to improve both the recording of citizen science effort as well as to improve the recording of “absence” data which would improve population modelling. Our solution was to increase the transparent use of the phone sensors as well as providing an easy-to-use user interface. Second is the EchidnaCSI app – an observational tool for collecting sightings and samples of echidna. From a software developer’s perspective, I will provide details on multi-platform app development as well as collaboration and integration with the Australian national biodiversity repository – the Atlas of Living Australia. Preliminary analysis regarding data quality will be presented along with lessons learned and paths for future research. I also seek feedback and further ideas on possible enhancements or modifications that might usefully be made to improve these techniques.


2021 ◽  
Vol 3 ◽  
Author(s):  
Robert R. Downs ◽  
Hampapuram K. Ramapriyan ◽  
Ge Peng ◽  
Yaxing Wei

Information about data quality helps potential data users to determine whether and how data can be used and enables the analysis and interpretation of such data. Providing data quality information improves opportunities for data reuse by increasing the trustworthiness of the data. Recognizing the need for improving the quality of citizen science data, we describe quality assessment and quality control (QA/QC) issues for these data and offer perspectives on aspects of improving or ensuring citizen science data quality and for conducting research on related issues.


2015 ◽  
Vol 2015 ◽  
pp. 1-8 ◽  
Author(s):  
Li Jiang ◽  
Hao Chen ◽  
Yueqi Ouyang ◽  
Canbing Li

With the rapid development of information technology and the coming of the era of big data, various data are constantly emerging and present the characteristics of autonomy and heterogeneity. How to optimize data quality and evaluate the effect has become a challenging problem. Firstly, a heterogeneous data integration model based on retrospective audit is proposed to locate the original data source and match the data. Secondly, in order to improve the integrated data quality, a retrospective audit model and associative audit rules are proposed to fix incomplete and incorrect data from multiple heterogeneous data sources. The heterogeneous data integration model based on retrospective audit is divided into four modules including original heterogeneous data, data structure, data processing, and data retrospective audit. At last, some assessment criteria such as redundancy, sparsity, and accuracy are defined to evaluate the effect of the optimized data quality. Experimental results show that the quality of the integrated data is significantly higher than the quality of the original data.


2020 ◽  
Vol 27 (3) ◽  
Author(s):  
Jobel Santos Corrêa ◽  
Mauro Sampaio ◽  
Rodrigo de Castro Barros

Abstract The concept of Logistics 4.0 works closely to that of Industry 4.0. While Industry 4.0 proposes a disruptive change in manufacturing, Logistics 4.0 advocates a transformation in the way organizations buy, manufacture, sell, and deliver products. The objective of this paper is to identify, in Brazilian companies, the degree of interest in the investment in six emerging technologies applicable to logistics, according to scientific literature, as well as to identify the current perception of data quality of these companies. To achieve these objectives, an online survey was conducted. The research showed that the technologies that most interest Brazilian companies are Internet of Things (IoT) and cloud computing, both with 82% of investment intention. The two technologies that least interested companies are crowdsourcing and 3D printing, both with 68% investment disinterest among respondents.


Author(s):  
Kari Lahti ◽  
Mikko Heikkinen ◽  
Aino Juslén ◽  
Leif Schulman

The Finnish Biodiversity Information Facility (FinBIF) Research Infrastructure (Schulman et al. 2021) is a national service with a broad coverage of the components of biodiversity informatics (Bingham et al. 2017). Data flows are managed under a single information technology (IT) architecture. Services are available in a single, branded on-line portal. Data are collated from all relevant sources e.g., research institutes, scientific collections, public authorities and citizen science projects, whose data represent a major contribution. The challenge is to analyse, classify and share good quality data in a way that the user understands its utility. Need for quality data The philosophy of FinBIF is that all observation records are important, and that all data are assessed for quality and able to be annotated. The challenge is that, in practice, many users desire data with 100% reliability. In our experience, most user concerns about data quality are related to citizen science data. Researchers are usually able to manage raw data to serve their purposes. However, decision-making authorities often have less capacity to analyse the data and thus require data that can be used instantly. Therefore, we need tools to provide users the data that are the most relevant and reliable for their specific use. For all users, standardized metadata (information about datasets) are key, when the user has doubts about the fitness-for-use of a particular dataset. There is also a need to provide data in different formats to serve various users. Finally, the service has to be machine-actionable (using an application programming interface (API) and R-package) as well as human-accessible for viewing and downloading data. Quality assignment FinBIF data accuracy varies significantly within and between datasets, and observers. Two quality-based classifications suitable for filtering are therefore applied. The dataset origin filter is based on the quality of a whole dataset (e.g. citizen science project) and includes three broad classes assigned with an appropriate quality label: Datasets by Professionals, by Specialists and by Citizen Scientists. The observation reliability filter is based on a single observation and on annotations by FinBIF users. This classification includes Expert verified, Community verified, Unassessed (default for all records), Uncertain, and Erroneous. The dataset origin does not necessarily determine the quality of the individual records in it. Observations made by citizen scientists are often accurate, while there may be errors in the professionally collected data. Records are frequently subject to annotation, which raises their quality over time (e.g., iNaturalist). Naturally, evidence (e.g., media, detailed descriptions, specimens) is needed for reliable identification. Annotating data When observations are compiled at FinBIF’s portal (Laji.fi), they are initially “Unassessed” (unless they have otherwise been assessed at the original source). When annotating occurrences, volunteers can make various entries using the tools provided. The aim of the commentary is to improve the quality of the observation data. Annotators are divided into two categories with two different roles: As a basic user, anyone who has logged in at Laji.fi can make comments or tag observations for review by experts. Users defined as experts have wider rights than basic users and their comments carry more weight. The most desired actions of expert users are to classify observations into confidence levels or to give them new or refined identifications. As a basic user, anyone who has logged in at Laji.fi can make comments or tag observations for review by experts. Users defined as experts have wider rights than basic users and their comments carry more weight. The most desired actions of expert users are to classify observations into confidence levels or to give them new or refined identifications. Information about new comments passes to the observer if the observation is recorded by using the FinBIF Observation Management System “Notebook”. However, comments cannot yet be automatically forwarded e.g., to the primary data management systems at the original source. Annotations add extra indications of quality. They do not replace or delete the original information. Nevertheless, annotations can change a record’s taxonomic identification, and by default, a record will be handled based on its latest identification. R-package for researchers and Public Authority Portal (PAP) for decision makers FinBIF has produced an R programming language interface to its API, which makes the publicly available data in FinBIF accessible from within R. For authorities, the PAP offers direct access to all available species information to authorised users, including sensitive and restricted-use data.


2018 ◽  
pp. 1-9 ◽  
Author(s):  
Anupong Sirirungreung ◽  
Rangsiya Buasom ◽  
Chuleeporn Jiraphongsa ◽  
Suleeporn Sangrajrang

Purpose Data quality is a core value of cancer registries, which bring about greater understanding of cancer distribution and determinants. Thailand established its cancer registry in 1986; however, studies focusing on data reliability have been limited. This study aimed to assess the coding completeness and reliability of the National Cancer Institute (NCI) hospital-based cancer registry, Thailand. Methods This study was conducted using the reabstracting method. We focused on seven cancer sites—the colon, rectum, liver, lung, breast, cervix, and prostate—registered between 2012 and 2014 in the NCI hospital-based cancer registry. Missing data were identified for coding completeness calculation among important variables. The agreement rate and κ coefficient were computed to represent data reliability. Results For reabstracting, we retrieved 957 medical records from a total of 5,462. These were selected using the probability proportional to size method, stratified by topology, sex, and registered year. The overall coding completeness of the registered and reabstracted data was 89.9% and 93.6%, respectively. In addition, the overall agreement rate among variables ranged from 84.7% to 99.6%, and κ coefficient ranged from 0.619 to 0.995. The misclassification among unilateral organs caused lower coding completeness and agreement rate of laterality coding. The completeness of current residency could be improved using the reabstracting method. The lowest agreement rate was found among various categories of diagnosis basis. Sex misclassification for male breast cancer was identified. Conclusion The coding completeness and data reliability of the NCI hospital-based cancer registry met the standard in most critical variables. However, some challenges remain to improve the data quality. The reabstracting method could identify the critical points affecting the quality of cancer registry data.


Author(s):  
S. G. Grigoriev ◽  
M. V. Kurnosenko ◽  
A. M. Kostyuk

The article discusses possible forms of educational STEM projects in the field of electronics and device control using Arduino controllers. As you know, the implementation of such STEM projects can be carried out not only using various electronic constructors, but also using virtual modeling environments. The knowledge obtained during modeling in virtual environments makes it possible to increase the efficiency of face-to-face practical training with a real constructor, and to improve the quality of students’ knowledge. The use of virtual modeling environments in combination with the use of real constructors provides links between distance and full-time learning. A real constructors can be used simultaneously by both the teacher and the student, jointly practicing the features of solving practical problems. The article provides examples of using a virtual environment for preliminary prototyping of circuits available in the documentation for electronic constructors, to familiarize students with the basics of designing and assembling electronic circuits using the surface mounting method and on a breadboard, as well as programming controllers on the Arduino platform that control electronic devices. This approach allows students to accelerate the assimilation of various interdisciplinary knowledge in the field of natural sciences using STEM design.


2017 ◽  
Vol 4 (1) ◽  
pp. 25-31 ◽  
Author(s):  
Diana Effendi

Information Product Approach (IP Approach) is an information management approach. It can be used to manage product information and data quality analysis. IP-Map can be used by organizations to facilitate the management of knowledge in collecting, storing, maintaining, and using the data in an organized. The  process of data management of academic activities in X University has not yet used the IP approach. X University has not given attention to the management of information quality of its. During this time X University just concern to system applications used to support the automation of data management in the process of academic activities. IP-Map that made in this paper can be used as a basis for analyzing the quality of data and information. By the IP-MAP, X University is expected to know which parts of the process that need improvement in the quality of data and information management.   Index term: IP Approach, IP-Map, information quality, data quality. REFERENCES[1] H. Zhu, S. Madnick, Y. Lee, and R. Wang, “Data and Information Quality Research: Its Evolution and Future,” Working Paper, MIT, USA, 2012.[2] Lee, Yang W; at al, Journey To Data Quality, MIT Press: Cambridge, 2006.[3] L. Al-Hakim, Information Quality Management: Theory and Applications. Idea Group Inc (IGI), 2007.[4] “Access : A semiotic information quality framework: development and comparative analysis : Journal ofInformation Technology.” [Online]. Available: http://www.palgravejournals.com/jit/journal/v20/n2/full/2000038a.html. [Accessed: 18-Sep-2015].[5] Effendi, Diana, Pengukuran Dan Perbaikan Kualitas Data Dan Informasi Di Perguruan Tinggi MenggunakanCALDEA Dan EVAMECAL (Studi Kasus X University), Proceeding Seminar Nasional RESASTEK, 2012, pp.TIG.1-TI-G.6.


Sign in / Sign up

Export Citation Format

Share Document