Linguistic Evidence in Security Law and Intelligence
Latest Publications


TOTAL DOCUMENTS

13
(FIVE YEARS 0)

H-INDEX

2
(FIVE YEARS 0)

Published By "University Library System, University Of Pittsburgh"

2327-5596

Author(s):  
Ángela Almela ◽  
Gema Alcaraz-Mármol ◽  
Arancha García-Pinar ◽  
Clara Pallejá

In this paper, the methods for developing a database of Spanish writing that can be used for forensic linguistic research are presented, including our data collection procedures. Specifically, the main instrument used for data collection has been translated into Spanish and adapted from Chaski (2001). It consists of ten tasks, by means of which the subjects are asked to write formal and informal texts about different topics. To date, 93 undergraduates from Spanish universities have already participated in the study and prisoners convicted of gender-based abuse have participated. A twofold analysis has been performed, since the data collected have been approached from a semantic and a morphosyntactic perspective. Regarding the semantic analysis, psycholinguistic categories have been used, many of them taken from the LIWC dictionary (Pennebaker et al., 2001). In order to obtain a more comprehensive depiction of the linguistic data, some other ad-hoc categories have been created, based on the corpus itself, using a double-check method for their validation so as to ensure inter-rater reliability. Furthermore, as regards morphosyntactic analysis, the natural language processing tool ALIAS TATTLER is being developed for Spanish.  Results shows that is it possible to differentiate non-abusers from abusers with strong accuracy based on linguistic features.


Author(s):  
Hans Van Halteren

This paper demonstrates how an author recognition system could be benchmarked, as a prerequisite for admission in court. The system used in the demonstration is the FEDERALES system, and the experimental data used were taken from the British National Corpus. The system was given several tasks, namely attributing a text sample to a specific text, verifying that a text sample was taken from a specific text, and verifying that a text sample was produced by a specific author. For the former two tasks, 1,099 texts with at least 10,000 words were used; for the latter 1,366 texts with known authors, which were verified against models for the 28 known authors for whom there were three or more texts. The experimental tasks were performed with different sampling methods (sequential samples or samples of concatenated random sentences), different sample sizes (1,000, 500, 250 or 125 words), varying amounts of training material (between 2 and 20 samples) and varying amounts of test material (1 or 3 samples). Under the best conditions, the system performed very well: with 7 training and 3 test samples of 1,000 words of randomly selected sentences, text attribution had an equal error rate of 0.06% and text verification an equal error rate of 1.3%; with 20 training and 3 test samples of 1,000 words of randomly selected sentences, author verification had an equal error rate of 7.5%. Under the worst conditions, with 2 training and 1 test sample of 125 words of sequential text, equal error rates for text attribution and text verification were 26.6% and 42.2%, and author verification did not perform better than chance. Furthermore, the quality degradation curves with slowly worsening conditions were not smooth, but contained steep drops. All in all, the results show the importance of having a benchmark which is as similar as possible to the actual court material for which the system is to be used, since the measured system quality differed greatly between evaluation scenarios and system degradation could not be predicted easily on the basis of the chosen scenario parameters.


Author(s):  
Michael J. Harris ◽  
Stefan Th. Gries ◽  
Viola G. Miglio

This article describes three studies in prosody and their potential application to the field of forensic linguistics. It begins with a brief introduction to prosody. It then proceeds to describe Miglio, Gries, & Harris (2014), a comparison of prosodic coding of new information by bilingual Spanish-English speakers and monolingual Spanish speakers. A description of Harris & Gries (2011) follows. This study compares the vowel duration variability of bilingual Spanish-English speakers and monolingual Spanish speakers, and touches upon corpus-based frequency effects and differences in linguistic aptitude between the two speaker groups. Finally, a portion of anongoing study is described (Harris in preparation). This section describes the use of prosodic variables and ensemble methods (or methods that use multiple learning algorithms) to classify languages, even in the case of impoverished data. All three experiments have implications and applications to the field of forensic linguistics, which are touched upon in each respective section and discussed in a more in-depth manner in the final section of this article. Furthermore, the applications of these methods to forensic linguistics are discussed in light of best practices for forensic linguistics, as outlined in Chaski (2013).


Author(s):  
Seung-Man Kang ◽  
Hyoungkeun Lee

This paper delves into the effect of SCAN and its cross-linguistic applicability by analyzing written statements in Korean. For this research, we conducted an experiment in which truth tellers were asked to write a true statement about a staged event and liars a fabricated one about the same event. We analyzed these two types of written statements using the criteria of SCAN. The results (accuracy rate, 81.6%) indicate that SCAN is effective in detecting deception despite the low internal consistency level among coders (Cronbach’s alpha level, 0.577). It was also shown that the SCAN criteria are not universally applicable across languages as the mode of using pronouns in Korean yields no significant difference between truthful and deceptive statements.


2013 ◽  
Vol 1 (1) ◽  
pp. 67-75 ◽  
Author(s):  
Tim Cole ◽  
JC Bruno Teboul ◽  
David E Zulawski ◽  
Douglas E Wicklander ◽  
Shane G Sturman

To date, few experimental studies have looked at the factors that influence people’s willingness to confess to something they did not do.  One widely cited experiment on the topic (i.e., Kassin & Kiechel, 1996) has suggested that false confessions are easy to obtain and that the use of false incriminating evidence increases the likelihood of obtaining one.  The present research attempted to replicate Kassin and Kiechel’s (1996) work using a different experimental task.  In the present experiment, unlike Kassin and Kiechel’s (1996) study, the participants were completely certain that they were not responsible for what had happened, thereby providing a different context for testing the idea that false incriminating evidence increases the likelihood of obtaining a false confession.  The results are discussed with respect to factors that may or may not increase individuals’ willingness to offer a false admission of guilt.


2013 ◽  
Vol 1 (1) ◽  
pp. 41-50 ◽  
Author(s):  
Isabel Picornell

Written witness statements are a unique source for the study of high-stakes textual deception. To date, however, there is no distinction in the way that they and other forms of verbal deception have been analysed, with written statements treated as extensions of transcribed versions of oral reports. Given the highly context-dependent nature of cues, it makes sense to take the characteristics of the medium into account when analysing for deceptive language. This study examines the characteristic features of witness narratives and proposes a new approach to search for deception cues. Narratives are treated as a progression of episodes over time, and deception as a progression of acts over time. This allows for the profiling of linguistic bundles in sequence, revealing the statements’ internal gradient, and deceivers’ choice of deceptive linguistic strategy. Study results suggest that, at least in the context of written witness statements, the weighting of individual features as deception cues is not static but depends on their interaction with other cues, and that detecting deceivers’ use of linguistic strategy is en effective vehicle for identifying deception.


2013 ◽  
Vol 1 (1) ◽  
pp. 99-103
Author(s):  
Carole E Chaski

ILER is a web-accessible platform for empirical research in forensic linguistics. It enables new researchers and practitioners to conduct mature research from literature review, experimental design, experiment review and approval, subject recruitment, human subjects protection and confidentiality agreement, subject demographic information collection, data collection, to data analysis.


2013 ◽  
Vol 1 (1) ◽  
pp. 76-98
Author(s):  
Harry Hollien

Linguistics and phonetics overlap in many areas. The essay to follow reviews some of the problems experienced by phoneticians in one of these regions. It may provide some insight for linguists when they are confronted by barriers in their own field. The present example involves individuals who are attempting to identify speakers from voice analysis. The fundamental challenge they face is, of course, caused by the thousands of variables associated with that task. Included here are differences among speakers’ gender, age, size, physiology, language, dialect, psychological/health states, background/education, reason for speaking, situation, environment, configuration of the acoustic channel -- plus many others. Many formal assessment procedures -- both aural-perceptual ones conducted by humans or machine/computer based systems -- have been proposed and/or used for the cited analyses. Unfortunately, however, few have enjoyed particularly high levels of success. Worse yet, reasonable progress has suffered from external impedances; the report to follow will outline some of them. Among the problems considered are: 1) competition (verification vs. identification, from voiceprints), 2) concept disputes 3) the continued undervaluation of relevant evidence and 4) markedly dissimilar philosophies of professionals from different disciplines. A response in the form of a short review of the data and concepts which clearly support the possibility of robust speaker identification is presented. Also included are suggestions as to how to enhance the effectiveness of disciplines such as ours.


2013 ◽  
Vol 1 (1) ◽  
pp. 51-66 ◽  
Author(s):  
Lauren B. Collister

This work explores the role of multimodal cues in detection of deception in a virtual world, an online community of World of Warcraft players. Case studies from a five-year ethnography are presented in three categories: small-scale deception in text, deception by avoidance, and large-scale deception in game-external modes. Each case study is analyzed in terms of how the affordances of the medium enabled or hampered deception as well as how the members of the community ultimately detected the deception. The ramifications of deception on the community are discussed, as well as the need for researchers to have a deep community knowledge when attempting to understand the role of deception in a complex society. Finally, recommendations are given for assessment of behavior in virtual worlds and the unique considerations that investigators must give to the rules and procedures of online communities.


Sign in / Sign up

Export Citation Format

Share Document