Artificial Robot Ear Design for Sound Direction Estimation

Author(s):  
Youngjin Park ◽  
Sungmok Hwang
2011 ◽  
Vol 59 (3-4) ◽  
pp. 208-217 ◽  
Author(s):  
Sungmok Hwang ◽  
Youngjin Park ◽  
Youn-sik Park

2007 ◽  
Author(s):  
Sungmok Hwang ◽  
Youngjin Park ◽  
Younsik Park

Sensors ◽  
2021 ◽  
Vol 21 (15) ◽  
pp. 5228
Author(s):  
Jin-Cheol Kim ◽  
Hwi-Gu Jeong ◽  
Seongwook Lee

In this study, we propose a method to identify the type of target and simultaneously determine its moving direction in a millimeter-wave radar system. First, using a frequency-modulated continuous wave (FMCW) radar sensor with the center frequency of 62 GHz, radar sensor data for a pedestrian, a cyclist, and a car are obtained in the test field. Then, a You Only Look Once (YOLO)-based network is trained with the sensor data to perform simultaneous target classification and moving direction estimation. To generate input data suitable for the deep learning-based classifier, a method of converting the radar detection result into an image form is also proposed. With the proposed method, we can identify the type of each target and its direction of movement with an accuracy of over 95%. Moreover, the pre-trained classifier shows an identification accuracy of 85% even for newly acquired data that have not been used for training.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Georges Hattab ◽  
Adamantini Hatzipanayioti ◽  
Anna Klimova ◽  
Micha Pfeiffer ◽  
Peter Klausing ◽  
...  

AbstractRecent technological advances have made Virtual Reality (VR) attractive in both research and real world applications such as training, rehabilitation, and gaming. Although these other fields benefited from VR technology, it remains unclear whether VR contributes to better spatial understanding and training in the context of surgical planning. In this study, we evaluated the use of VR by comparing the recall of spatial information in two learning conditions: a head-mounted display (HMD) and a desktop screen (DT). Specifically, we explored (a) a scene understanding and then (b) a direction estimation task using two 3D models (i.e., a liver and a pyramid). In the scene understanding task, participants had to navigate the rendered the 3D models by means of rotation, zoom and transparency in order to substantially identify the spatial relationships among its internal objects. In the subsequent direction estimation task, participants had to point at a previously identified target object, i.e., internal sphere, on a materialized 3D-printed version of the model using a tracked pointing tool. Results showed that the learning condition (HMD or DT) did not influence participants’ memory and confidence ratings of the models. In contrast, the model type, that is, whether the model to be recalled was a liver or a pyramid significantly affected participants’ memory about the internal structure of the model. Furthermore, localizing the internal position of the target sphere was also unaffected by participants’ previous experience of the model via HMD or DT. Overall, results provide novel insights on the use of VR in a surgical planning scenario and have paramount implications in medical learning by shedding light on the mental model we make to recall spatial structures.


Sign in / Sign up

Export Citation Format

Share Document