scholarly journals The effect of depth context in the segmentation of the colon in MRI volumes

Author(s):  
Ezenwoko Benson ◽  
Lukas Rier ◽  
Isawan Millican ◽  
Sue Pritchard ◽  
Carolyn Costigan ◽  
...  

ABSTRACTColonic volume content measurements can provide important information about the digestive tract physiology. Development of automated analyses will accelerate the translation of these measurements into clinical practice. In this paper, we test the effect of data dimension on the success of deep learning approaches to segment colons from MRI data. Deep learning network models were developed which used either 2D slices, complete 3D volumes and 2.5D partial volumes. These represent variations in the trade-off between the size and complexity of a network and its training regime, and the limitation of only being able to use a small section of the data at a time: full 3D networks, for example, have more image context available for decision making but require more powerful hardware to implement. For the datasets utilised here, 3D data was found to outperform 2.5D data, which in turn performed better than 2D datasets. The maximum Dice scores achieved by the networks were 0.898, 0.834 and 0.794 respectively. We also considered the effect of ablating varying amounts of data on the ability of the networks to label images correctly. We achieve dice scores of 0.829, 0.827 and 0.389 for 3D single slices ablation, 3D multi-slice ablation and 2.5D middle slice ablation.In addition, we examined another practical consideration of deep learning, that of how well a network performs on data from another acquisition device. Networks trained on images from a Philips Achieva MRI system yielded Dice scores of up to 0.77 in the 3D case when tested on images captured from a GE Medical Systems HDxt (both 1.5 Tesla) without any retraining. We also considered the effect of single versus multimodal MRI data showing that single modality dice scores can be boosted from 0.825 to 0.898 when adding an extra modality.

Sensors ◽  
2020 ◽  
Vol 20 (6) ◽  
pp. 1593 ◽  
Author(s):  
Yanlei Gu ◽  
Huiyang Zhang ◽  
Shunsuke Kamijo

Image based human behavior and activity understanding has been a hot topic in the field of computer vision and multimedia. As an important part, skeleton estimation, which is also called pose estimation, has attracted lots of interests. For pose estimation, most of the deep learning approaches mainly focus on the joint feature. However, the joint feature is not sufficient, especially when the image includes multi-person and the pose is occluded or not fully visible. This paper proposes a novel multi-task framework for the multi-person pose estimation. The proposed framework is developed based on Mask Region-based Convolutional Neural Networks (R-CNN) and extended to integrate the joint feature, body boundary, body orientation and occlusion condition together. In order to further improve the performance of the multi-person pose estimation, this paper proposes to organize the different information in serial multi-task models instead of the widely used parallel multi-task network. The proposed models are trained on the public dataset Common Objects in Context (COCO), which is further augmented by ground truths of body orientation and mutual-occlusion mask. Experiments demonstrate the performance of the proposed method for multi-person pose estimation and body orientation estimation. The proposed method can detect 84.6% of the Percentage of Correct Keypoints (PCK) and has an 83.7% Correct Detection Rate (CDR). Comparisons further illustrate the proposed model can reduce the over-detection compared with other methods.


One of the most serious global health threats is COVID-19 pandemic. The emphasis on increasing the diagnostic capability helps stopping its spread significantly. Therefore, to assist the radiologist or other medical professional to detect and identify the COVID-19 cases in the shortest possible time, we propose a computer-aided detection (CADe) system that uses the computed tomography (CT) scan images. This proposed boosted deep learning network (CLNet) is based on the implementation of Deep Learning (DL) networks as a complementary to the Compressive Learning (CL). We utilize our inception feature extraction technique in the measurement domain using CL to represent the data features into a new space with less dimensionality before accessing the Convolutional Neural Network. All original features have been contributed equally to the new space using a sensing matrix. Experiments performed on different compressed methods show promising results for COVID-19 detection.


IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 57566-57593 ◽  
Author(s):  
Abubakar Sulaiman Gezawa ◽  
Yan Zhang ◽  
Qicong Wang ◽  
Lei Yunqi

Author(s):  
Shengbo Liu ◽  
Pengyuan Fu ◽  
Lei Yan ◽  
Jian Wu ◽  
Yandong Zhao

Deep learning classification based on 3D point clouds has gained considerable research interest in recent years.The classification and quantitative analysis of wood defects are of great significance to the wood processing industry. In order to solve the problems of slow processing and low robustness of 3D data. This paper proposes an improvement based on littlepoint CNN lightweight deep learning network, adding BN layer. And based on the data set made by ourselves, the test is carried out. The new network bnlittlepoint CNN has been improved in speed and recognition rate. The correct rate of recognition for non defect log, non defect log and defect log as well as defect knot and dead knot can reach 95.6%.Finally, the "dead knot" and "loose knot" are quantitatively analyzed based on the "integral" idea, and the volume and surface area of the defect are obtained to a certain extent,the error is not more than 1.5% and the defect surface reconstruction is completed based on the triangulation idea.


Author(s):  
Ying Qu ◽  
Hairong Qi ◽  
Chiman Kwan

There are two mast cameras (Mastcam) onboard the Mars rover Curiosity. Both Mastcams are multispectral imagers with nine bands in each. The right Mastcam has three times higher resolution than the left. In this chapter, we apply some recently developed deep neural network models to enhance the left Mastcam images with help from the right Mastcam images. Actual Mastcam images were used to demonstrate the performance of the proposed algorithms.


Electronics ◽  
2020 ◽  
Vol 9 (4) ◽  
pp. 649
Author(s):  
Long Hoang ◽  
Suk-Hwan Lee ◽  
Ki-Ryong Kwon

3D shape recognition becomes necessary due to the popularity of 3D data resources. This paper aims to introduce the new method, hybrid deep learning network convolution neural network–support vector machine (CNN–SVM), for 3D recognition. The vertices of the 3D mesh are interpolated to be converted into Point Clouds; those Point Clouds are rotated for 3D data augmentation. We obtain and store the 2D projection of this 3D augmentation data in a 32 × 32 × 12 matrix, the input data of CNN–SVM. An eight-layer CNN is used as the algorithm for feature extraction, then SVM is applied for classifying feature extraction. Two big datasets, ModelNet40 and ModelNet10, of the 3D model are used for model validation. Based on our numerical experimental results, CNN–SVM is more accurate and efficient than other methods. The proposed method is 13.48% more accurate than the PointNet method in ModelNet10 and 8.5% more precise than 3D ShapeNets for ModelNet40. The proposed method works with both the 3D model in the augmented/virtual reality system and in the 3D Point Clouds, an output of the LIDAR sensor in autonomously driving cars.


2019 ◽  
Vol 10 ◽  
Author(s):  
Hanyu Zhang ◽  
Che-Lun Hung ◽  
Meiyuan Liu ◽  
Xiaoye Hu ◽  
Yi-Yang Lin

Automatic fake news detection is a challenging problem in deception detection. While evaluating the performance of deep learning-based models, if all the models are giving higher accuracy on a test dataset, it will make it harder to validate the performance of the deep learning models under consideration. So, we will need a complex problem to validate the performance of a deep learning model. LIAR is one such complex, much resent, labeled benchmark dataset which is publicly available for doing research on fake news detection to model statistical and machine learning approaches to combating fake news. In this work, a novel fake news detection system is implemented using Deep Neural Network models such as CNN, LSTM, BiLSTM, and the performance of their attention mechanism is evaluated by analyzing their performance in terms of Accuracy, Precision, Recall, and F1-score with training, validation and test datasets of LIAR.


Author(s):  
F. Matrone ◽  
A. Lingua ◽  
R. Pierdicca ◽  
E. S. Malinverni ◽  
M. Paolanti ◽  
...  

Abstract. The lack of benchmarking data for the semantic segmentation of digital heritage scenarios is hampering the development of automatic classification solutions in this field. Heritage 3D data feature complex structures and uncommon classes that prevent the simple deployment of available methods developed in other fields and for other types of data. The semantic classification of heritage 3D data would support the community in better understanding and analysing digital twins, facilitate restoration and conservation work, etc. In this paper, we present the first benchmark with millions of manually labelled 3D points belonging to heritage scenarios, realised to facilitate the development, training, testing and evaluation of machine and deep learning methods and algorithms in the heritage field. The proposed benchmark, available at http://archdataset.polito.it/, comprises datasets and classification results for better comparisons and insights into the strengths and weaknesses of different machine and deep learning approaches for heritage point cloud semantic segmentation, in addition to promoting a form of crowdsourcing to enrich the already annotated database.


Sign in / Sign up

Export Citation Format

Share Document