Exploiting Predicted Answer in Label Aggregation to Make Better Use of the Crowd Wisdom

Author(s):  
Jiacheng Liu ◽  
Feilong Tang ◽  
Long Chen ◽  
Yanmin Zhu
2018 ◽  
Author(s):  
Ayelet Ben-Sasson ◽  
Eli Ben-Sasson ◽  
Kayla Jacobs ◽  
Rotem Malinovitch

BACKGROUND To lower barriers to developmental screening, we designed Baby CROINC (CROwd INtelligence Curation), a digital platform to help parents track and assess their children’s development through crowd wisdom. OBJECTIVE To understand users’ experiences using Baby CROINC in relation to users’ technological competence and attitudes, while considering the influence of their children’s presented developmental evaluations and parents’ actual use of the system. METHODS Mothers of 260 children (M age= 17.6 months, SD=13.7) used Baby CROINC for two weeks. They entered developmental milestones on their children’s developmental diary timeline and received statistical developmental percentile reports. Mothers then completed Usability and Technology Profile Questionnaires. RESULTS Mothers’ experiences of the Baby CROINC system usability were associated with their attitudes toward solving technological problems, mediated by frequency of engagement in Internet activities. Mothers with a proactive approach toward solving technology problems, engage in a wide range of Internet activities, and/or view the Internet as integral to their lives had a better experience with Baby CROINC than mothers who did not. The system’s perceived usability was not associated with the crowd-based child developmental percentiles or quantity of mothers’ usage of the system. CONCLUSIONS Parent’s user experiences correlate with their technology competence and problem solving attitude but is not correlated with their child’s developmental status. Developmental screening platforms need to solve the tension between requiring active engagement and encouraging proactive parenting.


Mathematics ◽  
2021 ◽  
Vol 9 (8) ◽  
pp. 875
Author(s):  
Jesus Cerquides ◽  
Mehmet Oğuz Mülâyim ◽  
Jerónimo Hernández-González ◽  
Amudha Ravi Shankar ◽  
Jose Luis Fernandez-Marquez

Over the last decade, hundreds of thousands of volunteers have contributed to science by collecting or analyzing data. This public participation in science, also known as citizen science, has contributed to significant discoveries and led to publications in major scientific journals. However, little attention has been paid to data quality issues. In this work we argue that being able to determine the accuracy of data obtained by crowdsourcing is a fundamental question and we point out that, for many real-life scenarios, mathematical tools and processes for the evaluation of data quality are missing. We propose a probabilistic methodology for the evaluation of the accuracy of labeling data obtained by crowdsourcing in citizen science. The methodology builds on an abstract probabilistic graphical model formalism, which is shown to generalize some already existing label aggregation models. We show how to make practical use of the methodology through a comparison of data obtained from different citizen science communities analyzing the earthquake that took place in Albania in 2019.


Author(s):  
Nikolay Sinyak ◽  
Singh Tajinder ◽  
Jaglan Madhu Kumari ◽  
Vitaliy Kozlovskiy

Ubiquitous growth in the text mining field is unprecedented, where social media mining is playing a significant role. Gigantic growth of text mining is becoming a potential source of crowd wisdom extraction and analysis especially in terms of text pre-processing and sentiment analysis. The analysis of a potential influence of sentiment on real estate markets controversially discussed by scholars of finance, valuation and market efficiency supporters. Therefore, it’s a significant task of current research purview which not only provide an appropriate platform for the contributors but also for active real estate market information seekers. Text mining has gained the widespread attention of real estate market information users which is almost on explosion level. Accessibility of data on such behemoth scale mandates regular and critical analysis of this information for various perspectives’ plausibility. Rich patterns of online social text can be exploited to extract the relevant real estate information effectively. As text mining plays a significant and crucial role in discovery of these insights therefore its challenges and contribution in social media analysis must be explored extensively. In this paper, we provide a brief about the current summary of the modern state of text mining in pre-processing and sentiment for the real estate market analysis. Empha-sis is placed on the resources and learning mechanism available to real estate researchers and practitioners, as well as the major text mining tasks of interest to the community. Thus, the main aim of this chapter is to expound and intellectualize the domains of social media which are accessible on an extraordinary range in the field of text mining real estate for predicting real estate market trends and value.


2020 ◽  
Vol 34 (04) ◽  
pp. 4667-4674 ◽  
Author(s):  
Shikun Li ◽  
Shiming Ge ◽  
Yingying Hua ◽  
Chunhui Zhang ◽  
Hao Wen ◽  
...  

Typically, learning a deep classifier from massive cleanly annotated instances is effective but impractical in many real-world scenarios. An alternative is collecting and aggregating multiple noisy annotations for each instance to train the classifier. Inspired by that, this paper proposes to learn deep classifier from multiple noisy annotators via a coupled-view learning approach, where the learning view from data is represented by deep neural networks for data classification and the learning view from labels is described by a Naive Bayes classifier for label aggregation. Such coupled-view learning is converted to a supervised learning problem under the mutual supervision of the aggregated and predicted labels, and can be solved via alternate optimization to update labels and refine the classifiers. To alleviate the propagation of incorrect labels, small-loss metric is proposed to select reliable instances in both views. A co-teaching strategy with class-weighted loss is further leveraged in the deep classifier learning, which uses two networks with different learning abilities to teach each other, and the diverse errors introduced by noisy labels can be filtered out by peer networks. By these strategies, our approach can finally learn a robust data classifier which less overfits to label noise. Experimental results on synthetic and real data demonstrate the effectiveness and robustness of the proposed approach.


2019 ◽  
Vol 19 (6) ◽  
pp. 552-552
Author(s):  
Thanuja Dharmadasa ◽  
Matthew C. Kiernan
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document