scholarly journals A Study on Instructional Humor: How Much Humor Is Used in Presentations?

2021 ◽  
Vol 12 (1) ◽  
pp. 7
Author(s):  
Vera Paola Shoda ◽  
Toshimasa Yamanaka

Humor is applied in pedagogy to create a positive learning environment. Recent research focuses on the theories, effects, individual differences, and qualitative aspects of humor for instruction. However, there is a lack of studies focusing on quantitative features. Therefore, this research explored the quantitative characteristics of instructional humor in a naturalistic setting and applied techniques from natural language processing (NLP). This paper describes the results of two studies. The first study focused on instructional humor frequency and the placement of humor, while the linguistic features of instructional humor and non-instructional humor were compared in the second study. Two corpora were used in this research: TED Talks and user-submitted jokes from “stupidstuff.org” The results found that educators used humor 12.92 times for popular talks, while less popular talks only had 3.92 times. Humor is also more commonly placed during the first parts of the talk and lessens toward the end. There were also significant differences between the linguistic features of instructional and non-instructional humor in terms of readability scores and sentiment. These results provide a substantial update on quantitative instructional humor research and help educators understand how to use humor in the classroom in terms of quantitative and linguistic features.

2020 ◽  
Vol 51 (2) ◽  
pp. 168-181 ◽  
Author(s):  
Joshua J. Underwood ◽  
Cornelia Kirchhoff ◽  
Haven Warwick ◽  
Maria A. Gartstein

During childhood, parents represent the most commonly used source of their child’s temperament information and, typically, do so by responding to questionnaires. Despite their wide-ranging applications, interviews present notorious data reduction challenges, as quantification of narratives has proven to be a labor-intensive process. However, for the purposes of this study, the labor-intensive nature may have conferred distinct advantages. The present study represents a demonstration project aimed at leveraging emerging technologies for this purpose. Specifically, we used Python natural language processing capabilities to analyze semistructured temperament interviews conducted with U.S. and German mothers of toddlers, expecting to identify differences between these two samples in the frequency of words used to describe individual differences, along with some similarities. Two different word lists were used: (a) a set of German personality words and (b) temperament-related words extracted from the Early Childhood Behavior Questionnaire (ECBQ). Analyses using the German trait word demonstrated that mothers from Germany described their toddlers as significantly more “cheerful” and “careful” compared with U.S. caregivers. According to U.S. mothers, their children were more “independent,” “emotional,” and “timid.” For the ECBQ analysis, German mothers described their children as “calm” and “careful” more often than U.S. mothers. U.S. mothers, however, referred to their children as “upset,” “happy,” and “frustrated” more frequently than German caregivers. The Python code developed herein illustrates this software as a viable research tool for cross-cultural investigations.


2019 ◽  
Author(s):  
Yizhao Ni ◽  
Drew Barzman ◽  
Alycia Bachtel ◽  
Marcus Griffey ◽  
Alexander Osborn ◽  
...  

BACKGROUND School violence has a far reaching effect, impacting the entire school population including staff, students and their families. Among youth attending the most violent schools, studies have reported higher dropout rates, poor school attendance, and poor scholastic achievement. It was noted that the largest crime-prevention results occurred when youth at elevated risk were given an individualized prevention program. However, much work is needed to establish an effective approach to identify at-risk subjects. OBJECTIVE In our earlier research, we developed a standardized risk assessment program to interview subjects, identify risk and protective factors, and evaluate risk for school violence. This study focused on developing natural language processing (NLP) and machine learning technologies to automate the risk assessment process. METHODS We prospectively recruited 131 students with behavioral concerns from 89 schools between 05/01/2015 and 04/30/2018. The subjects were interviewed with three innovative risk assessment scales and their risk of violence were determined by pediatric psychiatrists based on clinical judgment. Leveraging NLP technologies, different types of linguistic features were extracted from the interview content. Machine learning classifiers were then applied to predict risk of school violence for individual subjects. A two-stage feature selection was implemented to identify violence-related predictors. The performance was validated on the psychiatrist-generated reference standard of risk levels, where positive predictive value (PPV), sensitivity (SEN), negative predictive value (NPV), specificity (SPEC) and area under the ROC curve (AUC) were assessed. RESULTS Compared to subjects' demographics and socioeconomic information, use of linguistic features significantly improved classifiers' predictive performance (P<0.01). The best-performing classifier with n-gram features achieved 86.5%/86.5%/85.7%/85.7%/94.0% (PPV/SEN/NPV/SPEC/AUC) on the cross-validation set and 83.3%/93.8%/91.7%/78.6%/94.6% (PPV/SEN/NPV/SPEC/AUC) on the test data. The feature selection process identified a set of predictors covering the discussion of subjects' thoughts, perspectives, behaviors, individual characteristics, peers and family dynamics, and protective factors. CONCLUSIONS By analyzing the content from subject interviews, the NLP and machine learning algorithms showed good capacity for detecting risk of school violence. The feature selection uncovered multiple warning markers that could deliver useful clinical insights to assist personalizing intervention. Consequently, the developed approach offered the promise of an end-to-end computerized screening service for preventing school violence.


2022 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Krishnadas Nanath ◽  
Supriya Kaitheri ◽  
Sonia Malik ◽  
Shahid Mustafa

Purpose The purpose of this paper is to examine the factors that significantly affect the prediction of fake news from the virality theory perspective. The paper looks at a mix of emotion-driven content, sentimental resonance, topic modeling and linguistic features of news articles to predict the probability of fake news. Design/methodology/approach A data set of over 12,000 articles was chosen to develop a model for fake news detection. Machine learning algorithms and natural language processing techniques were used to handle big data with efficiency. Lexicon-based emotion analysis provided eight kinds of emotions used in the article text. The cluster of topics was extracted using topic modeling (five topics), while sentiment analysis provided the resonance between the title and the text. Linguistic features were added to the coding outcomes to develop a logistic regression predictive model for testing the significant variables. Other machine learning algorithms were also executed and compared. Findings The results revealed that positive emotions in a text lower the probability of news being fake. It was also found that sensational content like illegal activities and crime-related content were associated with fake news. The news title and the text exhibiting similar sentiments were found to be having lower chances of being fake. News titles with more words and content with fewer words were found to impact fake news detection significantly. Practical implications Several systems and social media platforms today are trying to implement fake news detection methods to filter the content. This research provides exciting parameters from a viral theory perspective that could help develop automated fake news detectors. Originality/value While several studies have explored fake news detection, this study uses a new perspective on viral theory. It also introduces new parameters like sentimental resonance that could help predict fake news. This study deals with an extensive data set and uses advanced natural language processing to automate the coding techniques in developing the prediction model.


2021 ◽  
Vol 13 ◽  
Author(s):  
Aparna Balagopalan ◽  
Benjamin Eyre ◽  
Jessica Robin ◽  
Frank Rudzicz ◽  
Jekaterina Novikova

Introduction: Research related to the automatic detection of Alzheimer's disease (AD) is important, given the high prevalence of AD and the high cost of traditional diagnostic methods. Since AD significantly affects the content and acoustics of spontaneous speech, natural language processing, and machine learning provide promising techniques for reliably detecting AD. There has been a recent proliferation of classification models for AD, but these vary in the datasets used, model types and training and testing paradigms. In this study, we compare and contrast the performance of two common approaches for automatic AD detection from speech on the same, well-matched dataset, to determine the advantages of using domain knowledge vs. pre-trained transfer models.Methods: Audio recordings and corresponding manually-transcribed speech transcripts of a picture description task administered to 156 demographically matched older adults, 78 with Alzheimer's Disease (AD) and 78 cognitively intact (healthy) were classified using machine learning and natural language processing as “AD” or “non-AD.” The audio was acoustically-enhanced, and post-processed to improve quality of the speech recording as well control for variation caused by recording conditions. Two approaches were used for classification of these speech samples: (1) using domain knowledge: extracting an extensive set of clinically relevant linguistic and acoustic features derived from speech and transcripts based on prior literature, and (2) using transfer-learning and leveraging large pre-trained machine learning models: using transcript-representations that are automatically derived from state-of-the-art pre-trained language models, by fine-tuning Bidirectional Encoder Representations from Transformer (BERT)-based sequence classification models.Results: We compared the utility of speech transcript representations obtained from recent natural language processing models (i.e., BERT) to more clinically-interpretable language feature-based methods. Both the feature-based approaches and fine-tuned BERT models significantly outperformed the baseline linguistic model using a small set of linguistic features, demonstrating the importance of extensive linguistic information for detecting cognitive impairments relating to AD. We observed that fine-tuned BERT models numerically outperformed feature-based approaches on the AD detection task, but the difference was not statistically significant. Our main contribution is the observation that when tested on the same, demographically balanced dataset and tested on independent, unseen data, both domain knowledge and pretrained linguistic models have good predictive performance for detecting AD based on speech. It is notable that linguistic information alone is capable of achieving comparable, and even numerically better, performance than models including both acoustic and linguistic features here. We also try to shed light on the inner workings of the more black-box natural language processing model by performing an interpretability analysis, and find that attention weights reveal interesting patterns such as higher attribution to more important information content units in the picture description task, as well as pauses and filler words.Conclusion: This approach supports the value of well-performing machine learning and linguistically-focussed processing techniques to detect AD from speech and highlights the need to compare model performance on carefully balanced datasets, using consistent same training parameters and independent test datasets in order to determine the best performing predictive model.


ReCALL ◽  
1999 ◽  
Vol 11 (S1) ◽  
pp. 12-19
Author(s):  
Arantza Díaz de llarraza ◽  
Aitor Maritxalar ◽  
Montse Maritxalar ◽  
Maite Oronoz

This paper presents IDAZKIDE, a prototype of an intelligent language learning environment (ICALL) for learners of Basque. The philosophy of the system is to make different Natural Language Processing tools simultaneously accessible to students to help them (mainly at the morphological level) to write in Basque, as well as to give advice, taking into account some characteristics of the student gathered in a student model.


Semiotica ◽  
2016 ◽  
Vol 2016 (209) ◽  
pp. 323-340 ◽  
Author(s):  
Jian Li ◽  
Le Cheng ◽  
Winnie Cheng

AbstractModality and negation, as two important linguistic features used to realise subjectivity, have been investigated within various disciplines, such as logic, linguistics and philosophy, and law. The interaction between modality and negation, as a relatively new and undeveloped domain, has however not been paid due attention in scholarship. This corpus-based study investigates three aspects of their interaction: the differentiation of the deontic value by negation, the categorization of deontic modality in Hong Kong legislation via negation, and distribution patterns of deontic modality, especially distribution patterns of the negation of modality, in Hong Kong legislation. This study shows that negation is a powerful linguistic mechanism not only for determining the nature and functions of modality, but also for determining the value of modality. This study also reveals that negation helps us to investigate the distribution of deontic modality in Hong Kong legislation and hence revisit the legal framework in Hong Kong. A study taking into account the discursive and professional aspects of the interaction between deontic modality and negation will provide a theoretical basis for the natural language processing of modality and negation in legislation and also offer important implications for the study of negation and modality in general contexts.


2010 ◽  
Vol 31 (3) ◽  
pp. 439-462 ◽  
Author(s):  
NICHOLAS D. DURAN ◽  
CHARLES HALL ◽  
PHILIP M. MCCARTHY ◽  
DANIELLE S. MCNAMARA

ABSTRACTThe words people use and the way they use them can reveal a great deal about their mental states when they attempt to deceive. The challenge for researchers is how to reliably distinguish the linguistic features that characterize these hidden states. In this study, we use a natural language processing tool called Coh-Metrix to evaluate deceptive and truthful conversations that occur within a context of computer-mediated communication. Coh-Metrix is unique in that it tracks linguistic features based on cognitive and social factors that are hypothesized to influence deception. The results from Coh-Metrix are compared to linguistic features reported in previous independent research, which used a natural language processing tool called Linguistic Inquiry and Word Count. The comparison reveals converging and contrasting alignment for several linguistic features and establishes new insights on deceptive language and its use in conversation.


Author(s):  
Longtu Zhang ◽  
Mamoru Komachi

Logographic and alphabetic languages (e.g., Chinese vs. English) have different writing systems linguistically. Languages belonging to the same writing system usually exhibit more sharing information, which can be used to facilitate natural language processing tasks such as neural machine translation (NMT). This article takes advantage of the logographic characters in Chinese and Japanese by decomposing them into smaller units, thus more optimally utilizing the information these characters share in the training of NMT systems in both encoding and decoding processes. Experiments show that the proposed method can robustly improve the NMT performance of both “logographic” language pairs (JA–ZH) and “logographic + alphabetic” (JA–EN and ZH–EN) language pairs in both supervised and unsupervised NMT scenarios. Moreover, as the decomposed sequences are usually very long, extra position features for the transformer encoder can help with the modeling of these long sequences. The results also indicate that, theoretically, linguistic features can be manipulated to obtain higher share token rates and further improve the performance of natural language processing systems.


Sign in / Sign up

Export Citation Format

Share Document