Human emotion recognition based on facial expressions via deep learning on high-resolution images

Author(s):  
Yahia Said ◽  
Mohammad Barr
2020 ◽  
Vol 10 (12) ◽  
pp. 4282
Author(s):  
Ghada Zamzmi ◽  
Sivaramakrishnan Rajaraman ◽  
Sameer Antani

Medical images are acquired at different resolutions based on clinical goals or available technology. In general, however, high-resolution images with fine structural details are preferred for visual task analysis. Recognizing this significance, several deep learning networks have been proposed to enhance medical images for reliable automated interpretation. These deep networks are often computationally complex and require a massive number of parameters, which restrict them to highly capable computing platforms with large memory banks. In this paper, we propose an efficient deep learning approach, called Hydra, which simultaneously reduces computational complexity and improves performance. The Hydra consists of a trunk and several computing heads. The trunk is a super-resolution model that learns the mapping from low-resolution to high-resolution images. It has a simple architecture that is trained using multiple scales at once to minimize a proposed learning-loss function. We also propose to append multiple task-specific heads to the trained Hydra trunk for simultaneous learning of multiple visual tasks in medical images. The Hydra is evaluated on publicly available chest X-ray image collections to perform image enhancement, lung segmentation, and abnormality classification. Our experimental results support our claims and demonstrate that the proposed approach can improve the performance of super-resolution and visual task analysis in medical images at a remarkably reduced computational cost.


2022 ◽  
Vol 14 (2) ◽  
pp. 265
Author(s):  
Yanjun Wang ◽  
Shaochun Li ◽  
Fei Teng ◽  
Yunhao Lin ◽  
Mengjie Wang ◽  
...  

Accurate roof information of buildings can be obtained from UAV high-resolution images. The large-scale accurate recognition of roof types (such as gabled, flat, hipped, complex and mono-pitched roofs) of rural buildings is crucial for rural planning and construction. At present, most UAV high-resolution optical images only have red, green and blue (RGB) band information, which aggravates the problems of inter-class similarity and intra-class variability of image features. Furthermore, the different roof types of rural buildings are complex, spatially scattered, and easily covered by vegetation, which in turn leads to the low accuracy of roof type identification by existing methods. In response to the above problems, this paper proposes a method for identifying roof types of complex rural buildings based on visible high-resolution remote sensing images from UAVs. First, the fusion of deep learning networks with different visual features is investigated to analyze the effect of the different feature combinations of the visible difference vegetation index (VDVI) and Sobel edge detection features and UAV visible images on model recognition of rural building roof types. Secondly, an improved Mask R-CNN model is proposed to learn more complex features of different types of images of building roofs by using the ResNet152 feature extraction network with migration learning. After we obtained roof type recognition results in two test areas, we evaluated the accuracy of the results using the confusion matrix and obtained the following conclusions: (1) the model with RGB images incorporating Sobel edge detection features has the highest accuracy and enables the model to recognize more and more accurately the roof types of different morphological rural buildings, and the model recognition accuracy (Kappa coefficient (KC)) compared to that of RGB images is on average improved by 0.115; (2) compared with the original Mask R-CNN, U-Net, DeeplabV3 and PSPNet deep learning models, the improved Mask R-CNN model has the highest accuracy in recognizing the roof types of rural buildings, with F1-score, KC and OA averaging 0.777, 0.821 and 0.905, respectively. The method can obtain clear and accurate profiles and types of rural building roofs, and can be extended for green roof suitability evaluation, rooftop solar potential assessment, and other building roof surveys, management and planning.


Author(s):  
Wei Wang ◽  
Rongyuan Liu ◽  
Huiyun Yang ◽  
Ping Zhou ◽  
Xiangwen Zhang ◽  
...  

2022 ◽  
Author(s):  
Yaxing Li ◽  
Xiaofeng Jia ◽  
Xinming Wu ◽  
Zhicheng Geng

<p>Reverse time migration (RTM) is a technique used to obtain high-resolution images of underground reflectors; however, this method is computationally intensive when dealing with large amounts of seismic data. Multi-source RTM can significantly reduce the computational cost by processing multiple shots simultaneously. However, multi-source-based methods frequently result in crosstalk artifacts in the migrated images, causing serious interference in the imaging signals. Plane-wave migration, as a mainstream multi-source method, can yield migrated images with plane waves in different angles by implementing phase encoding of the source and receiver wavefields; however, this method frequently requires a trade-off between computational efficiency and imaging quality. We propose a method based on deep learning for removing crosstalk artifacts and enhancing the image quality of plane-wave migration images. We designed a convolutional neural network that accepts an input of seven plane-wave images at different angles and outputs a clear and enhanced image. We built 505 1024×256 velocity models, and employed each of them using plane-wave migration to produce raw images at 0°, ±20°, ±40°, and ±60° as input of the network. Labels are high-resolution images computed from the corresponding reflectivity models by convolving with a Ricker wavelet. Random sub-images with a size of 512×128 were used for training the network. Numerical examples demonstrated the effectiveness of the trained network in crosstalk removal and imaging enhancement. The proposed method is superior to both the conventional RTM and plane-wave RTM (PWRTM) in imaging resolution. Moreover, the proposed method requires only seven migrations, significantly improving the computational efficiency. In the numerical examples, the processing time required by our method was approximately 1.6% and 10% of that required by RTM and PWRTM, respectively.</p>


2019 ◽  
Vol 5 (4) ◽  
pp. 1
Author(s):  
MSUGH ORTIL ◽  
KOKO GRACE TWAKI ◽  
◽  

2019 ◽  
Author(s):  
Bagus Tris Atmaja

The demand for recognizing emotion in text has grown increasingly as human emotion can be expressed via text and manytechnologies, such as product reviews and speech transcription, can benefit from text emotion recognition. The study of text emotionrecognition was established some decades ago using unsupervised learning and a small amount of data. Advancements incomputation hardware and in the development of larger text corpus have enabled us to analyze emotion in the text by moresophisticated techniques. This paper presents a deep learning-based approach for the recognition of categorical and dimensionalemotion from both written and spoken texts. The result shows that the system performs better on both categorical and dimensional task( &gt; 60% accuracy and &lt; 20% error) with a larger dataset compared to a smaller dataset. We also found the recognition rate is affectedby both the size of the data and the number of emotion categories. On the dimensional task, a larger amount of data consistentlyprovided a better result. Recognition of categorical emotion on a spoken text is easier than on a written text, while on dimensional task,the written text yielded better performance.


2022 ◽  
Author(s):  
Yaxing Li ◽  
Xiaofeng Jia ◽  
Xinming Wu ◽  
Zhicheng Geng

<p>Reverse time migration (RTM) is a technique used to obtain high-resolution images of underground reflectors; however, this method is computationally intensive when dealing with large amounts of seismic data. Multi-source RTM can significantly reduce the computational cost by processing multiple shots simultaneously. However, multi-source-based methods frequently result in crosstalk artifacts in the migrated images, causing serious interference in the imaging signals. Plane-wave migration, as a mainstream multi-source method, can yield migrated images with plane waves in different angles by implementing phase encoding of the source and receiver wavefields; however, this method frequently requires a trade-off between computational efficiency and imaging quality. We propose a method based on deep learning for removing crosstalk artifacts and enhancing the image quality of plane-wave migration images. We designed a convolutional neural network that accepts an input of seven plane-wave images at different angles and outputs a clear and enhanced image. We built 505 1024×256 velocity models, and employed each of them using plane-wave migration to produce raw images at 0°, ±20°, ±40°, and ±60° as input of the network. Labels are high-resolution images computed from the corresponding reflectivity models by convolving with a Ricker wavelet. Random sub-images with a size of 512×128 were used for training the network. Numerical examples demonstrated the effectiveness of the trained network in crosstalk removal and imaging enhancement. The proposed method is superior to both the conventional RTM and plane-wave RTM (PWRTM) in imaging resolution. Moreover, the proposed method requires only seven migrations, significantly improving the computational efficiency. In the numerical examples, the processing time required by our method was approximately 1.6% and 10% of that required by RTM and PWRTM, respectively.</p>


Sign in / Sign up

Export Citation Format

Share Document