scholarly journals Visualization of Speech Perception Analysis via Phoneme Alignment: A Pilot Study

2022 ◽  
Vol 12 ◽  
Author(s):  
J. Tilak Ratnanather ◽  
Lydia C. Wang ◽  
Seung-Ho Bae ◽  
Erin R. O'Neill ◽  
Elad Sagi ◽  
...  

Objective: Speech tests assess the ability of people with hearing loss to comprehend speech with a hearing aid or cochlear implant. The tests are usually at the word or sentence level. However, few tests analyze errors at the phoneme level. So, there is a need for an automated program to visualize in real time the accuracy of phonemes in these tests.Method: The program reads in stimulus-response pairs and obtains their phonemic representations from an open-source digital pronouncing dictionary. The stimulus phonemes are aligned with the response phonemes via a modification of the Levenshtein Minimum Edit Distance algorithm. Alignment is achieved via dynamic programming with modified costs based on phonological features for insertion, deletions and substitutions. The accuracy for each phoneme is based on the F1-score. Accuracy is visualized with respect to place and manner (consonants) or height (vowels). Confusion matrices for the phonemes are used in an information transfer analysis of ten phonological features. A histogram of the information transfer for the features over a frequency-like range is presented as a phonemegram.Results: The program was applied to two datasets. One consisted of test data at the sentence and word levels. Stimulus-response sentence pairs from six volunteers with different degrees of hearing loss and modes of amplification were analyzed. Four volunteers listened to sentences from a mobile auditory training app while two listened to sentences from a clinical speech test. Stimulus-response word pairs from three lists were also analyzed. The other dataset consisted of published stimulus-response pairs from experiments of 31 participants with cochlear implants listening to 400 Basic English Lexicon sentences via different talkers at four different SNR levels. In all cases, visualization was obtained in real time. Analysis of 12,400 actual and random pairs showed that the program was robust to the nature of the pairs.Conclusion: It is possible to automate the alignment of phonemes extracted from stimulus-response pairs from speech tests in real time. The alignment then makes it possible to visualize the accuracy of responses via phonological features in two ways. Such visualization of phoneme alignment and accuracy could aid clinicians and scientists.

2000 ◽  
Vol 21 (1) ◽  
pp. 109-150 ◽  
Author(s):  
Nicola J. Woods

Colonial Englishes have been observed to contain both archaic and innovative linguistic features, and are thus seen to display evidence of both language preservation and change. This paper examines the use of certain phonological features of New Zealand English (NZE) and discusses their status as relics or innovations. Examination is made of the diphthong which occurs in the mouth lexical set and the front short vowels trap, dress and kit. Trends in usage are studied using real time analysis of speakers recorded in the 1940s and their present day descendants recorded in 1993/94. In this way, the development of NZE is charted by means of the analysis of different generations of the same New Zealand families.


1976 ◽  
Vol 19 (4) ◽  
pp. 718-748 ◽  
Author(s):  
Robert C. Bilger ◽  
Marilyn D. Wang

Consonant confusion matrices were obtained from 22 outpatient listeners with sensorineural hearing loss for four sets of CV and VC nonsense syllables, presented monaurally at SRT + 40 dB. Testing was typically conducted for six hours on each of two separate days. Overall performance and patterns of confusions were stable over time. Analysis of the matrices in terms of phonological features indicated that the patterns of consonant confusions varied both with degree and configuration of the subject’s loss. Scaling of intersubject similarity using a pairwise multidimensional scaling analysis resulted in consistent classification of subjects according to audiometric configuration into three groups-essentially normal hearing, flat or rising audiograms, and high-frequency hearing losses.


Author(s):  
R.P. Goehner ◽  
W.T. Hatfield ◽  
Prakash Rao

Computer programs are now available in various laboratories for the indexing and simulation of transmission electron diffraction patterns. Although these programs address themselves to the solution of various aspects of the indexing and simulation process, the ultimate goal is to perform real time diffraction pattern analysis directly off of the imaging screen of the transmission electron microscope. The program to be described in this paper represents one step prior to real time analysis. It involves the combination of two programs, described in an earlier paper(l), into a single program for use on an interactive basis with a minicomputer. In our case, the minicomputer is an INTERDATA 70 equipped with a Tektronix 4010-1 graphical display terminal and hard copy unit.A simplified flow diagram of the combined program, written in Fortran IV, is shown in Figure 1. It consists of two programs INDEX and TEDP which index and simulate electron diffraction patterns respectively. The user has the option of choosing either the indexing or simulating aspects of the combined program.


2013 ◽  
Vol 20 (3) ◽  
pp. 91-106 ◽  
Author(s):  
Rachel Pizarek ◽  
Valeriy Shafiro ◽  
Patricia McCarthy

Computerized auditory training (CAT) is a convenient, low-cost approach to improving communication of individuals with hearing loss or other communicative disorders. A number of CAT programs are being marketed to patients and audiologists. The present literature review is an examination of evidence for the effectiveness of CAT in improving speech perception in adults with hearing impairments. Six current CAT programs, used in 9 published studies, were reviewed. In all 9 studies, some benefit of CAT for speech perception was demonstrated. Although these results are encouraging, the overall quality of available evidence remains low, and many programs currently on the market have not yet been evaluated. Thus, caution is needed when selecting CAT programs for specific patients. It is hoped that future researchers will (a) examine a greater number of CAT programs using more rigorous experimental designs, (b) determine which program features and training regimens are most effective, and (c) indicate which patients may benefit from CAT the most.


2017 ◽  
Vol 2 (11) ◽  
pp. 79-90
Author(s):  
Courtney G. Scott ◽  
Trina M. Becker ◽  
Kenneth O. Simpson

The use of computer monitors to provide technology-based written feedback during clinical sessions, referred to as “bug-in-the-eye” (BITi) feedback, recently emerged in the literature with preliminary evidence to support its effectiveness (Carmel, Villatte, Rosenthal, Chalker & Comtois, 2015; Weck et al., 2016). This investigation employed a single-subject, sequential A-B design with two participants to observe the effects of implementing BITi feedback using a smartwatch on the clinical behavior of student clinicians (SCs). Baseline and treatment data on the stimulus-response-consequence (S-R-C) contingency completion rates of SCs were collected using 10 minute segments of recorded therapy sessions. All participants were students enrolled in a clinical practicum experience in a communication disorders and sciences (CDS) program. A celeration line, descriptive statistics, and stability band were used to analyze the data by slope, trend, and variability. Results demonstrated a significant correlative relationship between BITi feedback with a smartwatch and an increase in positive clinical behaviors. Based on qualitative interviews and exit rating scales, SCs reported BITi feedback was noninvasive and minimally distracting. Preliminary evidence suggests BITi feedback with a smartwatch may be an effective tool for providing real-time clinical feedback.


2020 ◽  
Vol 67 (4) ◽  
pp. 1197-1205 ◽  
Author(s):  
Yuki Totani ◽  
Susumu Kotani ◽  
Kei Odai ◽  
Etsuro Ito ◽  
Manabu Sakakibara

2021 ◽  
Vol 2021 (4) ◽  
pp. 7-16
Author(s):  
Sivaraman Eswaran ◽  
Aruna Srinivasan ◽  
Prasad Honnavalli

Sign in / Sign up

Export Citation Format

Share Document