scholarly journals Reliability of Machine and Human Examiners for Detection of Laryngeal Penetration or Aspiration in Videofluoroscopic Swallowing Studies

2021 ◽  
Vol 10 (12) ◽  
pp. 2681
Author(s):  
Yuna Kim ◽  
Hyun-Il Kim ◽  
Geun-Seok Park ◽  
Seo-Young Kim ◽  
Sang-Il Choi ◽  
...  

Computer-assisted analysis is expected to improve the reliability of videofluoroscopic swallowing studies (VFSSs), but its usefulness is limited. Previously, we proposed a deep learning model that can detect laryngeal penetration or aspiration fully automatically in VFSS video images, but the evidence for its reliability was insufficient. This study aims to compare the intra- and inter-rater reliability of the computer model and human raters. The test dataset consisted of 173 video files from which the existence of laryngeal penetration or aspiration was judged by the computer and three physicians in two sessions separated by a one-month interval. Intra- and inter-rater reliability were calculated using Cohen’s kappa coefficient, the positive reliability ratio (PRR) and the negative reliability ratio (NRR). Intrarater reliability was almost perfect for the computer and two experienced physicians. Interrater reliability was moderate to substantial between the model and each human rater and between the human raters. The average PRR and NRR between the model and the human raters were similar to those between the human raters. The results demonstrate that the deep learning model can detect laryngeal penetration or aspiration from VFSS video as reliably as human examiners.

Diagnostics ◽  
2021 ◽  
Vol 11 (7) ◽  
pp. 1147
Author(s):  
Hyun-Il Kim ◽  
Yuna Kim ◽  
Bomin Kim ◽  
Dae Youp Shin ◽  
Seong Jae Lee ◽  
...  

Kinematic analysis of the hyoid bone in a videofluorosopic swallowing study (VFSS) is important for assessing dysphagia. However, calibrating the hyoid bone movement is time-consuming, and its reliability shows wide variation. Computer-assisted analysis has been studied to improve the efficiency and accuracy of hyoid bone identification and tracking, but its performance is limited. In this study, we aimed to design a robust network that can track hyoid bone movement automatically without human intervention. Using 69,389 frames from 197 VFSS files as the data set, a deep learning model for detection and trajectory prediction was constructed and trained by the BiFPN-U-Net(T) network. The present model showed improved performance when compared with the previous models: an area under the curve (AUC) of 0.998 for pixelwise accuracy, an accuracy of object detection of 99.5%, and a Dice similarity of 90.9%. The bounding box detection performance for the hyoid bone and reference objects was superior to that of other models, with a mean average precision of 95.9%. The estimation of the distance of hyoid bone movement also showed higher accuracy. The deep learning model proposed in this study could be used to detect and track the hyoid bone more efficiently and accurately in VFSS analysis.


2020 ◽  
Vol 22 (Supplement_2) ◽  
pp. ii148-ii148
Author(s):  
Yoshihiro Muragaki ◽  
Yutaka Matsui ◽  
Takashi Maruyama ◽  
Masayuki Nitta ◽  
Taiichi Saito ◽  
...  

Abstract INTRODUCTION It is useful to know the molecular subtype of lower-grade gliomas (LGG) when deciding on a treatment strategy. This study aims to diagnose this preoperatively. METHODS A deep learning model was developed to predict the 3-group molecular subtype using multimodal data including magnetic resonance imaging (MRI), positron emission tomography (PET), and computed tomography (CT). The performance was evaluated using leave-one-out cross validation with a dataset containing information from 217 LGG patients. RESULTS The model performed best when the dataset contained MRI, PET, and CT data. The model could predict the molecular subtype with an accuracy of 96.6% for the training dataset and 68.7% for the test dataset. The model achieved test accuracies of 58.5%, 60.4%, and 59.4% when the dataset contained only MRI, MRI and PET, and MRI and CT data, respectively. The conventional method used to predict mutations in the isocitrate dehydrogenase (IDH) gene and the codeletion of chromosome arms 1p and 19q (1p/19q) sequentially had an overall accuracy of 65.9%. This is 2.8 percent point lower than the proposed method, which predicts the 3-group molecular subtype directly. CONCLUSIONS AND FUTURE PERSPECTIVE A deep learning model was developed to diagnose the molecular subtype preoperatively based on multi-modality data in order to predict the 3-group classification directly. Cross-validation showed that the proposed model had an overall accuracy of 68.7% for the test dataset. This is the first model to double the expected value for a 3-group classification problem, when predicting the LGG molecular subtype. We plan to apply the techniques of heat map and/or segmentation for an increase in prediction accuracy.


2021 ◽  
Vol 11 (12) ◽  
pp. 3199-3208
Author(s):  
K. Ganapriya ◽  
N. Uma Maheswari ◽  
R. Venkatesh

Prediction of occurrence of a seizure would be of greater help to make necessary precaution for taking care of the patient. A Deep learning model, recurrent neural network (RNN), is designed for predicting the upcoming values in the EEG values. A deep data analysis is made to find the parameter that could best differentiate the normal values and seizure values. Next a recurrent neural network model is built for predicting the values earlier. Four different variants of recurrent neural networks are designed in terms of number of time stamps and the number of LSTM layers and the best model is identified. The best identified RNN model is used for predicting the values. The performance of the model is evaluated in terms of explained variance score and R2 score. The model founds to perform well number of elements in the test dataset is minimal and so this model can predict the seizure values only a few seconds earlier.


2021 ◽  
Author(s):  
Shinae Lee ◽  
Sang-il Oh ◽  
Junik Jo ◽  
Sumi Kang ◽  
Yooseok Shin ◽  
...  

Abstract The early detection of incipient dental caries enables preventive treatment, and bitewing radiography is a good diagnostic tool for posterior incipient caries. In the field of medical imaging, the utilization of deep learning with convolutional neural networks (CNNs) to process various types of images has been actively researched and has shown promising performance. In this study, we developed a CNN model using a U-shaped deep CNN (U-Net) for dental caries detection on bitewing radiographs and investigated whether this model can improve clinicians’ performance. In total, 304 bitewing radiographs were used to train the deep learning model and 50 radiographs were used for performance evaluation. The diagnostic performance of the CNN model on the total test dataset was as follows: precision, 63.29%; recall, 65.02%; and F1-score, 64.14%, showing quite accurate performance. When three dentists detected dental caries using the results of the CNN model as reference data, the overall diagnostic performance of all three clinicians significantly improved, as shown by an increased recall ratio (D1, 85.34%; D1', 92.15%; D2, 85.86%; D2', 93.72%; D3, 69.11%; D3', 79.06%). These increases were especially significant in the incipient and moderate caries subgroups. The deep learning model may help clinicians to diagnose dental caries more accurately.


2021 ◽  
Vol 1 (1) ◽  
pp. 44-46
Author(s):  
Ashar Mirza ◽  
Rishav Kumar Rajak

In this paper, we present a UNet architecture-based deep learning method that is used to segment polyp and instruments from the image data set provided in the MedAI Challenge2021. For the polyp segmentation task, we developed a UNet based algorithm for segmenting polyps in images taken from endoscopies. The main focus of this task is to achieve high segmentation metrics on the supplied test dataset. Similarly for the polyp segmentation task, in the instrument segmentation task, we have developed UNet based algorithms for segmenting instruments present in colonoscopy videos.


2020 ◽  
Author(s):  
Aaron E. Kornblith ◽  
Newton Addo ◽  
Ruolei Dong ◽  
Robert Rogers ◽  
Jacqueline Grupp-Phelan ◽  
...  

ABSTRACTThe pediatric Focused Assessment with Sonography for Trauma (FAST) is a sequence of ultrasound views rapidly performed by the clinician to diagnose hemorrhage. One limitation of FAST is inconsistent acquisition of required views. We sought to develop a deep learning model and classify FAST views using a heterogeneous dataset of pediatric FAST. This study of diagnostic test developed and tested a deep learning model for view classification of archived real-world pediatric FAST studies collected from two pediatric emergency departments. FAST frames were randomly distributed to training, validation, and test datasets in a 70:20:10 ratio; each patient was represented in only one dataset to maintain sample independence. The outcome was the prediction accuracy of the model in classifying FAST frames and video clips. FAST studies performed by 30 different clinicians from 699 injured children included 4,925 videos representing 1,062,612 frames from children who were a median of 9 years old. On test dataset, the overall view classification accuracy for the model was 93.4% (95% CI: 93.3-93.6) for frames and 97.8% (95% CI: 96.0-99.0) for video clips. Frames were correctly classified with an accuracy of 96.0% (95% CI: 95.9-96.1) for cardiac, 99.8% (95% CI: 99.8-99.8) for thoracic, 95.2% (95% CI: 95.0-95.3) for abdominal upper quadrants, and 95.9% (95% CI: 95.8-96.0) for suprapubic. A deep learning model can be developed to accurately classify pediatric FAST views. Accurate view classification is the important first step to support developing a consistent and accurate multi-stage deep learning model for pediatric FAST interpretation.


2021 ◽  
Vol 39 (15_suppl) ◽  
pp. e24085-e24085
Author(s):  
Yeong Hak Bang ◽  
Yoon Ho Choi ◽  
Mincheol Park ◽  
Geon Hee Lee ◽  
Soo-yong Shin ◽  
...  

e24085 Background: Breakthrough cancer pain (BTcP), a transitory flare of pain that occurs on a background of relatively well-controlled baseline pain, is a challenging clinical problem in managing cancer pain. We hypothesized that the BTcP could be predictable according to the patients’ previous observed patterns. In this study, we report on the development of a deep learning model that predicts hourly individual-level breakthrough pain for patients with cancer. Methods: We defined the BTcP as the pain with numerical rating scale (NRS) score 4 or above and developed models predicting the onset time of BTcP with the temporal resolution of 1 hour. The datasets which have more than 20 records of NRS score during hospitalization were included in our study. All the pain records were obtained from patients hospitalized on the wards of hematology-oncology in Samsung Medical Center between July 2016 to February 2020. The model used the time windows of 3 days to predict NRS scores over the next 24 hours. To capture irregular pain patterns, we created the sequence of average pain patterns over 24 hours from the previous 3 days and used it for normalization. We trained a Bi-directional long-short term memory (LSTM) based deep learning model. The model was validated using the holdout method with 20% of the datasets. Its performance was assessed with the area under the receiver operating characteristic curve (AUROC) and the area under the precision-recall curve (AURPC). Results: We included pain log data containing 2,905 admissions from 2,176 patients with solid cancer and 1,755 admissions from 1,082 patients with hematologic cancer in the analysis. The median age was 57 (interquartile range (IQR), 47-64), the most frequent type of cancer was lung cancer (18.0%), and most patients had stage 4 (60.7%). Among the 103,948 hours from patients in whole datasets, 1,091 (4.7%) hours were labeled as the period of BTcP. The patients have the records of NRS score with a median of 3 (IQR, 2.0-4.5) and BTcP with a median of 1.1 (IQR, 0.5-2.0) per day. We allocated approximately 20% of patients (653 patients with 932 admissions) to the holdout test dataset. Our model showed the AUROC 0.719 and AUPRC 0.680 for predicting the BTcP in the test dataset. Conclusions: Our study showed that cancer pain could be predictive by using a deep learning model. Though our exploratory study has a limitation of generalizability, future warranted subgroup analysis and verification research could make our model more applicable in a real-world setting.


2021 ◽  
Author(s):  
Di-Ge Ai ◽  
Yu He ◽  
Sheng-Hai Jin ◽  
Xue-Min Liu ◽  
Nian-Yi Sun ◽  
...  

BACKGROUND A standardized method for identifying medical laboratory observations, such as Logical Observation Identifiers Names and Codes (LOINC), is critical for creating accurate and effective public data models. However, such standards are not being used effectively. Standardized mapping facilitates consistency in medical terminologies and data sharing in multicenter treatment. OBJECTIVE To address the problem of standardizing laboratory test terminologies, a deep learning–based high-precision end-to-end terminology standardization matching system was developed to map laboratory test terms (LTTs) to LOINC. METHODS We manually constructed a laboratory test terminology mapping dataset containing 15,349 data items extracted from the information system of the Shengjing Hospital of China Medical University and matched 2,375 LOINC. We developed Attribute-wised Graph Attention Siamese Network (AGASN), a deep learning–based high-precision laboratory test terminology mapping model, to separately extract LTT features and LOINC term features and calculate the matching rate. We designed an attribute pooling mechanism to convert terminology strings to attribute sequences. Moreover, we developed a graph attention model based on attribute relations, which increased the interpretability of the proposed model. The problem of inconsistency in training and testing objectives was solved by improving the training objectives of the model. RESULTS The proposed a novel deep learning model achieved an accuracy of 82.33% ± 0.6% on the test dataset where the LOINC were visible, corresponding to a 10.9% improvement compared with that obtained using a random forest classifier. Furthermore, the proposed system achieved an accuracy of 63.14% ± 0.2% on the test dataset where the LOINC were invisible, constituting a 10.0% improvement compared with that obtained using SimCSE. Manual validation of the system performance showed accuracies of 82.33% and 70.66% on labeled and unlabeled datasets, respectively. Finally, we constructed a visual attribute relational strength network using an attribute graph attention model. CONCLUSIONS Herein, a Chinese laboratory test terminology mapping dataset was created and a deep learning system for the standardized mapping of LTTs was proposed. The results demonstrate that the proposed system can map LTTs to LOINC with high accuracy.


2020 ◽  
Vol 13 (4) ◽  
pp. 627-640 ◽  
Author(s):  
Avinash Chandra Pandey ◽  
Dharmveer Singh Rajpoot

Background: Sentiment analysis is a contextual mining of text which determines viewpoint of users with respect to some sentimental topics commonly present at social networking websites. Twitter is one of the social sites where people express their opinion about any topic in the form of tweets. These tweets can be examined using various sentiment classification methods to find the opinion of users. Traditional sentiment analysis methods use manually extracted features for opinion classification. The manual feature extraction process is a complicated task since it requires predefined sentiment lexicons. On the other hand, deep learning methods automatically extract relevant features from data hence; they provide better performance and richer representation competency than the traditional methods. Objective: The main aim of this paper is to enhance the sentiment classification accuracy and to reduce the computational cost. Method: To achieve the objective, a hybrid deep learning model, based on convolution neural network and bi-directional long-short term memory neural network has been introduced. Results: The proposed sentiment classification method achieves the highest accuracy for the most of the datasets. Further, from the statistical analysis efficacy of the proposed method has been validated. Conclusion: Sentiment classification accuracy can be improved by creating veracious hybrid models. Moreover, performance can also be enhanced by tuning the hyper parameters of deep leaning models.


Sign in / Sign up

Export Citation Format

Share Document