scholarly journals Deep Learning Models Comparison for Tissue Classification using Optical Coherence Tomography Images: Toward Smart Laser Osteotomy

OSA Continuum ◽  
2021 ◽  
Author(s):  
Yakub Bayhaqi ◽  
Arsham Hamidi ◽  
Ferda Canbaz ◽  
Alexander Navarini ◽  
Philippe Cattin ◽  
...  
2021 ◽  
Author(s):  
Viney Gupta ◽  
Shweta Birla ◽  
Toshit Varshney ◽  
Bindu I Somarajan ◽  
Shikha Gupta ◽  
...  

Abstract Objective: To predict the presence of Angle Dysgenesis on Anterior Segment Optical Coherence Tomography (ADoA) using deep learning and to correlate ADoA with mutations in known glaucoma genes. Design: A cross-sectional observational study. Participants: Eight hundred, high definition anterior segment optical coherence tomography (ASOCT) B-scans were included, out of which 340 images (One scan per eye) were used to build the machine learning (ML) model and the rest were used for validation of ADoA. Out of 340 images, 170 scans included PCG (n=27), JOAG (n=86) and POAG (n=57) eyes and the rest were controls. The genetic validation dataset consisted of another 393 images of patients with known mutations compared with 320 images of healthy controls Methods: ADoA was defined as the absence of Schlemm's canal(SC), the presence of extensive hyper-reflectivity over the region of trabecular meshwork or a hyper-reflective membrane (HM) over the region of the trabecular meshwork. Deep learning was used to classify a given ASOCT image as either having angle dysgenesis or not. ADoA was then specifically looked for, on ASOCT images of patients with mutations in the known genes for glaucoma (MYOC, CYP1B1, FOXC1 and LTBP2). Main Outcome measures: Using Deep learning to identify ADoA in patients with known gene mutations. Results: Our three optimized deep learning models showed an accuracy > 95%, specificity >97% and sensitivity >96% in detecting angle dysgenesis on ASOCT in the internal test dataset. The area under receiver operating characteristic (AUROC) curve, based on the external validation cohort were 0.91 (95% CI, 0.88 to 0.95), 0.80 (95% CI, 0.75 to 0.86) and 0.86 (95% CI, 0.80 to 0.91) for the three models. Amongst the patients with known gene mutations, ADoA was observed among all the patients with MYOC mutations, as it was also observed among those with CYP1B1, FOXC1 and with LTBP2 mutations compared to only 5% of those healthy controls (with no glaucoma mutations). Conclusions: Three deep learning models were developed for a consensus-based outcome to objectively identify ADoA among glaucoma patients. All patients with MYOC mutations had ADoA as predicted by the models.


2020 ◽  
Vol 9 (2) ◽  
pp. 51
Author(s):  
Michael G. Kawczynski ◽  
Thomas Bengtsson ◽  
Jian Dai ◽  
J. Jill Hopkins ◽  
Simon S. Gao ◽  
...  

2020 ◽  
Vol 11 (3) ◽  
pp. 1580 ◽  
Author(s):  
Zhe Jiang ◽  
Zhiyu Huang ◽  
Bin Qiu ◽  
Xiangxi Meng ◽  
Yunfei You ◽  
...  

2020 ◽  
pp. bjophthalmol-2020-317825
Author(s):  
Yonghao Li ◽  
Weibo Feng ◽  
Xiujuan Zhao ◽  
Bingqian Liu ◽  
Yan Zhang ◽  
...  

Background/aimsTo apply deep learning technology to develop an artificial intelligence (AI) system that can identify vision-threatening conditions in high myopia patients based on optical coherence tomography (OCT) macular images.MethodsIn this cross-sectional, prospective study, a total of 5505 qualified OCT macular images obtained from 1048 high myopia patients admitted to Zhongshan Ophthalmic Centre (ZOC) from 2012 to 2017 were selected for the development of the AI system. The independent test dataset included 412 images obtained from 91 high myopia patients recruited at ZOC from January 2019 to May 2019. We adopted the InceptionResnetV2 architecture to train four independent convolutional neural network (CNN) models to identify the following four vision-threatening conditions in high myopia: retinoschisis, macular hole, retinal detachment and pathological myopic choroidal neovascularisation. Focal Loss was used to address class imbalance, and optimal operating thresholds were determined according to the Youden Index.ResultsIn the independent test dataset, the areas under the receiver operating characteristic curves were high for all conditions (0.961 to 0.999). Our AI system achieved sensitivities equal to or even better than those of retina specialists as well as high specificities (greater than 90%). Moreover, our AI system provided a transparent and interpretable diagnosis with heatmaps.ConclusionsWe used OCT macular images for the development of CNN models to identify vision-threatening conditions in high myopia patients. Our models achieved reliable sensitivities and high specificities, comparable to those of retina specialists and may be applied for large-scale high myopia screening and patient follow-up.


2021 ◽  
Vol 11 (12) ◽  
pp. 5488
Author(s):  
Wei Ping Hsia ◽  
Siu Lun Tse ◽  
Chia Jen Chang ◽  
Yu Len Huang

The purpose of this article is to evaluate the accuracy of the optical coherence tomography (OCT) measurement of choroidal thickness in healthy eyes using a deep-learning method with the Mask R-CNN model. Thirty EDI-OCT of thirty patients were enrolled. A mask region-based convolutional neural network (Mask R-CNN) model composed of deep residual network (ResNet) and feature pyramid networks (FPNs) with standard convolution and fully connected heads for mask and box prediction, respectively, was used to automatically depict the choroid layer. The average choroidal thickness and subfoveal choroidal thickness were measured. The results of this study showed that ResNet 50 layers deep (R50) model and ResNet 101 layers deep (R101). R101 U R50 (OR model) demonstrated the best accuracy with an average error of 4.85 pixels and 4.86 pixels, respectively. The R101 ∩ R50 (AND model) took the least time with an average execution time of 4.6 s. Mask-RCNN models showed a good prediction rate of choroidal layer with accuracy rates of 90% and 89.9% for average choroidal thickness and average subfoveal choroidal thickness, respectively. In conclusion, the deep-learning method using the Mask-RCNN model provides a faster and accurate measurement of choroidal thickness. Comparing with manual delineation, it provides better effectiveness, which is feasible for clinical application and larger scale of research on choroid.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Reza Mirshahi ◽  
Pasha Anvari ◽  
Hamid Riazi-Esfahani ◽  
Mahsa Sardarinia ◽  
Masood Naseripour ◽  
...  

AbstractThe purpose of this study was to introduce a new deep learning (DL) model for segmentation of the fovea avascular zone (FAZ) in en face optical coherence tomography angiography (OCTA) and compare the results with those of the device’s built-in software and manual measurements in healthy subjects and diabetic patients. In this retrospective study, FAZ borders were delineated in the inner retinal slab of 3 × 3 enface OCTA images of 131 eyes of 88 diabetic patients and 32 eyes of 18 healthy subjects. To train a deep convolutional neural network (CNN) model, 126 enface OCTA images (104 eyes with diabetic retinopathy and 22 normal eyes) were used as training/validation dataset. Then, the accuracy of the model was evaluated using a dataset consisting of OCTA images of 10 normal eyes and 27 eyes with diabetic retinopathy. The CNN model was based on Detectron2, an open-source modular object detection library. In addition, automated FAZ measurements were conducted using the device’s built-in commercial software, and manual FAZ delineation was performed using ImageJ software. Bland–Altman analysis was used to show 95% limit of agreement (95% LoA) between different methods. The mean dice similarity coefficient of the DL model was 0.94 ± 0.04 in the testing dataset. There was excellent agreement between automated, DL model and manual measurements of FAZ in healthy subjects (95% LoA of − 0.005 to 0.026 mm2 between automated and manual measurement and 0.000 to 0.009 mm2 between DL and manual FAZ area). In diabetic eyes, the agreement between DL and manual measurements was excellent (95% LoA of − 0.063 to 0.095), however, there was a poor agreement between the automated and manual method (95% LoA of − 0.186 to 0.331). The presence of diabetic macular edema and intraretinal cysts at the fovea were associated with erroneous FAZ measurements by the device’s built-in software. In conclusion, the DL model showed an excellent accuracy in detection of FAZ border in enfaces OCTA images of both diabetic patients and healthy subjects. The DL and manual measurements outperformed the automated measurements of the built-in software.


2021 ◽  
Vol 11 (7) ◽  
pp. 3119
Author(s):  
Cristina L. Saratxaga ◽  
Jorge Bote ◽  
Juan F. Ortega-Morán ◽  
Artzai Picón ◽  
Elena Terradillos ◽  
...  

(1) Background: Clinicians demand new tools for early diagnosis and improved detection of colon lesions that are vital for patient prognosis. Optical coherence tomography (OCT) allows microscopical inspection of tissue and might serve as an optical biopsy method that could lead to in-situ diagnosis and treatment decisions; (2) Methods: A database of murine (rat) healthy, hyperplastic and neoplastic colonic samples with more than 94,000 images was acquired. A methodology that includes a data augmentation processing strategy and a deep learning model for automatic classification (benign vs. malignant) of OCT images is presented and validated over this dataset. Comparative evaluation is performed both over individual B-scan images and C-scan volumes; (3) Results: A model was trained and evaluated with the proposed methodology using six different data splits to present statistically significant results. Considering this, 0.9695 (±0.0141) sensitivity and 0.8094 (±0.1524) specificity were obtained when diagnosis was performed over B-scan images. On the other hand, 0.9821 (±0.0197) sensitivity and 0.7865 (±0.205) specificity were achieved when diagnosis was made considering all the images in the whole C-scan volume; (4) Conclusions: The proposed methodology based on deep learning showed great potential for the automatic characterization of colon polyps and future development of the optical biopsy paradigm.


Sign in / Sign up

Export Citation Format

Share Document