Pilot points method for conditioning multiple-point statistical facies simulation on flow data

2018 ◽  
Vol 115 ◽  
pp. 219-233 ◽  
Author(s):  
Wei Ma ◽  
Behnam Jafarpour
Author(s):  
Meri L. Andreassen ◽  
Bonnie E. Smith ◽  
Thomas W. Guyette

Pressure-flow data are often used to provide information about the adequacy of velopharyngeal valving for speech. However, there is limited information available concerning simultaneous pressure-flow measurements for oral and nasal sound segments produced by normal speakers. This study provides normative pressure, flow, and velopharyngeal orifice area measurements for selected oral and nasal sound segments produced by 10 male and 10 female adult speakers. An aerodynamic categorization scheme of velopharyngeal function, including one typical category and three atypical categories (open, closed, and mixed) is proposed.


Author(s):  
Ramon Martins ◽  
Roney Thompson ◽  
Aristeu Silveira Neto ◽  
Gilmar MOMPEAN ◽  
João Rodrigo Andrade

Author(s):  
Kristopher D. Staller

Abstract Cold temperature failures are often difficult to resolve, especially those at extreme low levels (< -40°C). Momentary application of chill spray can confirm the failure mode, but is impractical during photoemission microscopy (PEM), laser scanning microscopy (LSM), and multiple point microprobing. This paper will examine relatively low-cost cold temperature systems that can hold samples at steady state extreme low temperatures and describe a case study where a cold temperature stage was combined with LSM soft defect localization (SDL) to rapidly identify the cause of a complex cold temperature failure mechanism.


2020 ◽  
Vol 17 (4) ◽  
pp. 497-506
Author(s):  
Sunil Patel ◽  
Ramji Makwana

Automatic classification of dynamic hand gesture is challenging due to the large diversity in a different class of gesture, Low resolution, and it is performed by finger. Due to a number of challenges many researchers focus on this area. Recently deep neural network can be used for implicit feature extraction and Soft Max layer is used for classification. In this paper, we propose a method based on a two-dimensional convolutional neural network that performs detection and classification of hand gesture simultaneously from multimodal Red, Green, Blue, Depth (RGBD) and Optical flow Data and passes this feature to Long-Short Term Memory (LSTM) recurrent network for frame-to-frame probability generation with Connectionist Temporal Classification (CTC) network for loss calculation. We have calculated an optical flow from Red, Green, Blue (RGB) data for getting proper motion information present in the video. CTC model is used to efficiently evaluate all possible alignment of hand gesture via dynamic programming and check consistency via frame-to-frame for the visual similarity of hand gesture in the unsegmented input stream. CTC network finds the most probable sequence of a frame for a class of gesture. The frame with the highest probability value is selected from the CTC network by max decoding. This entire CTC network is trained end-to-end with calculating CTC loss for recognition of the gesture. We have used challenging Vision for Intelligent Vehicles and Applications (VIVA) dataset for dynamic hand gesture recognition captured with RGB and Depth data. On this VIVA dataset, our proposed hand gesture recognition technique outperforms competing state-of-the-art algorithms and gets an accuracy of 86%


Sign in / Sign up

Export Citation Format

Share Document