scholarly journals Classifying Upper Arm Gym-Workouts via Convolutional Neural Network by Imputing a Biopotential-Kinematic Relationship

2021 ◽  
Vol 11 (6) ◽  
pp. 2845
Author(s):  
Ji Hyeon Yoo ◽  
Ho Jin Jung ◽  
Yi Sue Jung ◽  
Yoonbee Kim ◽  
Changjae Lee ◽  
...  

This paper proposes a systemic approach to upper arm gym-workout classification according to spatio-temporal features depicted by biopotential as well as joint kinematics. The key idea of the proposed approach is to impute a biopotential-kinematic relationship by merging the joint kinematic data into a multichannel electromyography signal and visualizing the merged biopotential-kinematic data as an image. Under this approach, the biopotential-kinematic relationship can be imputed by counting on the functionality of a convolutional neural network: an automatic feature extractor followed by a classifier. First, while a professional trainer is demonstrating upper arm gym-workouts, electromyography and joint kinematic data are measured by an armband-type surface electromyography (sEMG) sensor and a RGB-d camera, respectively. Next, the measured data are augmented by adopting the amplitude adjusted Fourier Transform. Then, the augmented electromyography and joint kinematic data are visualized as one image by merging and calculating pixel components in three different ways. Lastly, for each visualized image type, upper arm gym-workout classification is performed via the convolutional neural network. To analyze classification accuracy, two-way rANOVA is performed with two factors: the level of data augmentation and visualized image type. The classification result substantiates that a biopotential-kinematic relationship can be successfully imputed by merging joint kinematic data in-between biceps- and triceps-electromyography channels and visualizing as a time-series heatmap image.

Friction ◽  
2021 ◽  
Author(s):  
Xiaobin Hu ◽  
Jian Song ◽  
Zhenhua Liao ◽  
Yuhong Liu ◽  
Jian Gao ◽  
...  

AbstractFinding the correct category of wear particles is important to understand the tribological behavior. However, manual identification is tedious and time-consuming. We here propose an automatic morphological residual convolutional neural network (M-RCNN), exploiting the residual knowledge and morphological priors between various particle types. We also employ data augmentation to prevent performance deterioration caused by the extremely imbalanced problem of class distribution. Experimental results indicate that our morphological priors are distinguishable and beneficial to largely boosting overall performance. M-RCNN demonstrates a much higher accuracy (0.940) than the deep residual network (0.845) and support vector machine (0.821). This work provides an effective solution for automatically identifying wear particles and can be a powerful tool to further analyze the failure mechanisms of artificial joints.


Author(s):  
Ramesh Adhikari ◽  
Suresh Pokharel

Data augmentation is widely used in image processing and pattern recognition problems in order to increase the richness in diversity of available data. It is commonly used to improve the classification accuracy of images when the available datasets are limited. Deep learning approaches have demonstrated an immense breakthrough in medical diagnostics over the last decade. A significant amount of datasets are needed for the effective training of deep neural networks. The appropriate use of data augmentation techniques prevents the model from over-fitting and thus increases the generalization capability of the network while testing afterward on unseen data. However, it remains a huge challenge to obtain such a large dataset from rare diseases in the medical field. This study presents the synthetic data augmentation technique using Generative Adversarial Networks to evaluate the generalization capability of neural networks using existing data more effectively. In this research, the convolutional neural network (CNN) model is used to classify the X-ray images of the human chest in both normal and pneumonia conditions; then, the synthetic images of the X-ray from the available dataset are generated by using the deep convolutional generative adversarial network (DCGAN) model. Finally, the CNN model is trained again with the original dataset and augmented data generated using the DCGAN model. The classification performance of the CNN model is improved by 3.2% when the augmented data were used along with the originally available dataset.


2018 ◽  
Vol 4 (9) ◽  
pp. 107 ◽  
Author(s):  
Mohib Ullah ◽  
Ahmed Mohammed ◽  
Faouzi Alaya Cheikh

Articulation modeling, feature extraction, and classification are the important components of pedestrian segmentation. Usually, these components are modeled independently from each other and then combined in a sequential way. However, this approach is prone to poor segmentation if any individual component is weakly designed. To cope with this problem, we proposed a spatio-temporal convolutional neural network named PedNet which exploits temporal information for spatial segmentation. The backbone of the PedNet consists of an encoder–decoder network for downsampling and upsampling the feature maps, respectively. The input to the network is a set of three frames and the output is a binary mask of the segmented regions in the middle frame. Irrespective of classical deep models where the convolution layers are followed by a fully connected layer for classification, PedNet is a Fully Convolutional Network (FCN). It is trained end-to-end and the segmentation is achieved without the need of any pre- or post-processing. The main characteristic of PedNet is its unique design where it performs segmentation on a frame-by-frame basis but it uses the temporal information from the previous and the future frame for segmenting the pedestrian in the current frame. Moreover, to combine the low-level features with the high-level semantic information learned by the deeper layers, we used long-skip connections from the encoder to decoder network and concatenate the output of low-level layers with the higher level layers. This approach helps to get segmentation map with sharp boundaries. To show the potential benefits of temporal information, we also visualized different layers of the network. The visualization showed that the network learned different information from the consecutive frames and then combined the information optimally to segment the middle frame. We evaluated our approach on eight challenging datasets where humans are involved in different activities with severe articulation (football, road crossing, surveillance). The most common CamVid dataset which is used for calculating the performance of the segmentation algorithm is evaluated against seven state-of-the-art methods. The performance is shown on precision/recall, F 1 , F 2 , and mIoU. The qualitative and quantitative results show that PedNet achieves promising results against state-of-the-art methods with substantial improvement in terms of all the performance metrics.


2020 ◽  
Author(s):  
Luis H. S. Vogado ◽  
Rodrigo M. S. Veras ◽  
Kelson R. T. Aires

Leukemia is a disorder that affects the bone marrow, causing uncontrolled production of leukocytes, impairing the transport of oxygen and causing blood coagulation problems. In this article, we propose a new computational tool, named LeukNet, a Convolutional Neural Network (CNN) architecture based on the VGG-16 convolutional blocks, to facilitate the leukemia diagnosis from blood smear images. We evaluated different architectures and fine-tuning methods using 18 datasets containing 3536 images with distinct characteristics of color, texture, contrast, and resolution. Additionally, data augmentation operations were applied to increase the training set by up to 20 times. The k-fold cross-validation (k = 5) results achieved 98.28% of accuracy. A cross-dataset validation technique, named LeaveOne-Dataset-Out Cross-Validation (LODOCV), is also proposed to evaluate the developed model’s generalization capability. The accuracy of using LODOCV on the ALL-IDB 1, ALL-IDB 2, and UFG datasets was 97.04%, 82.46%, and 70.24%, respectively, overcoming the current state-of-the-art results and offering new guidelines for image-based computer-aided diagnosis (CAD) systems in this area.


Sensors ◽  
2018 ◽  
Vol 18 (7) ◽  
pp. 2399 ◽  
Author(s):  
Cunwei Sun ◽  
Yuxin Yang ◽  
Chang Wen ◽  
Kai Xie ◽  
Fangqing Wen

The convolutional neural network (CNN) has made great strides in the area of voiceprint recognition; but it needs a huge number of data samples to train a deep neural network. In practice, it is too difficult to get a large number of training samples, and it cannot achieve a better convergence state due to the limited dataset. In order to solve this question, a new method using a deep migration hybrid model is put forward, which makes it easier to realize voiceprint recognition for small samples. Firstly, it uses Transfer Learning to transfer the trained network from the big sample voiceprint dataset to our limited voiceprint dataset for the further training. Fully-connected layers of a pre-training model are replaced by restricted Boltzmann machine layers. Secondly, the approach of Data Augmentation is adopted to increase the number of voiceprint datasets. Finally, we introduce fast batch normalization algorithms to improve the speed of the network convergence and shorten the training time. Our new voiceprint recognition approach uses the TLCNN-RBM (convolutional neural network mixed restricted Boltzmann machine based on transfer learning) model, which is the deep migration hybrid model that is used to achieve an average accuracy of over 97%, which is higher than that when using either CNN or the TL-CNN network (convolutional neural network based on transfer learning). Thus, an effective method for a small sample of voiceprint recognition has been provided.


Sign in / Sign up

Export Citation Format

Share Document