Speech transformation solutions

2006 ◽  
Vol 14 (2) ◽  
pp. 411-442
Author(s):  
Dimitri Kanevsky ◽  
Sara Basson ◽  
Alexander Faisman ◽  
Leonid Rachevsky ◽  
Alex Zlatsin ◽  
...  

This paper outlines the background development of “intelligent” technologies such as speech recognition. Despite significant progress in the development of these technologies, they still fall short in many areas, and rapid advances in areas such as dictation are actually stalled. In this paper we have proposed semi-automatic solutions — smart integration of human and intelligent efforts. One such technique involves improvement to the speech recognition editing interface, thereby reducing the perception of errors to the viewer. Other techniques that are described in the paper are batch enrollment, which allows the user to reduce the amount of time required for enrollment, and content spotting, which can be used for applications that have repeated content flow, such as movies or museum tours. The paper also suggests a general concept of distributive training of speech recognition systems that is based on data collection across a network.

Author(s):  
Roger B. Garberg

Phoneme-based automatic speech recognition (ASR) technology enables designers to easily create custom command words or phrases that users can employ to request service operations. In this paper, I report results from two experiments concerning important dimensions of these ASR command vocabularies, including command naturalness/appropriateness and command recallability. Ease of recall is a critical dimension for assessing ASR commands used in multi-step applications since service subscribers may be engaged in several different cognitive activities that divide attention. Yet techniques for measuring command recallability can be difficult to implement owing to the time required for data collection and analysis. Results of these studies indicate the the dimension of command “naturalness” and memorability are closely related: under appropriate conditions, the simple procedures associated with measuring command naturalness or appropriateness can predict retrievability of command expressions.


2008 ◽  
Author(s):  
Kristie Nemeth ◽  
Nicole Arbuckle ◽  
Andrea Snead ◽  
Drew Bowers ◽  
Christopher Burneka ◽  
...  

Author(s):  
Conrad Bernath ◽  
Aitor Alvarez ◽  
Haritz Arzelus ◽  
Carlos David Martínez

Author(s):  
Sheng Li ◽  
Dabre Raj ◽  
Xugang Lu ◽  
Peng Shen ◽  
Tatsuya Kawahara ◽  
...  

Procedia CIRP ◽  
2021 ◽  
Vol 97 ◽  
pp. 130-135
Author(s):  
Christian Deuerlein ◽  
Moritz Langer ◽  
Julian Seßner ◽  
Peter Heß ◽  
Jörg Franke

Symmetry ◽  
2021 ◽  
Vol 13 (4) ◽  
pp. 634
Author(s):  
Alakbar Valizada ◽  
Natavan Akhundova ◽  
Samir Rustamov

In this paper, various methodologies of acoustic and language models, as well as labeling methods for automatic speech recognition for spoken dialogues in emergency call centers were investigated and comparatively analyzed. Because of the fact that dialogue speech in call centers has specific context and noisy, emotional environments, available speech recognition systems show poor performance. Therefore, in order to accurately recognize dialogue speeches, the main modules of speech recognition systems—language models and acoustic training methodologies—as well as symmetric data labeling approaches have been investigated and analyzed. To find an effective acoustic model for dialogue data, different types of Gaussian Mixture Model/Hidden Markov Model (GMM/HMM) and Deep Neural Network/Hidden Markov Model (DNN/HMM) methodologies were trained and compared. Additionally, effective language models for dialogue systems were defined based on extrinsic and intrinsic methods. Lastly, our suggested data labeling approaches with spelling correction are compared with common labeling methods resulting in outperforming the other methods with a notable percentage. Based on the results of the experiments, we determined that DNN/HMM for an acoustic model, trigram with Kneser–Ney discounting for a language model and using spelling correction before training data for a labeling method are effective configurations for dialogue speech recognition in emergency call centers. It should be noted that this research was conducted with two different types of datasets collected from emergency calls: the Dialogue dataset (27 h), which encapsulates call agents’ speech, and the Summary dataset (53 h), which contains voiced summaries of those dialogues describing emergency cases. Even though the speech taken from the emergency call center is in the Azerbaijani language, which belongs to the Turkic group of languages, our approaches are not tightly connected to specific language features. Hence, it is anticipated that suggested approaches can be applied to the other languages of the same group.


Sign in / Sign up

Export Citation Format

Share Document