scholarly journals SD-Net: joint surgical gesture recognition and skill assessment

Author(s):  
Jinglu Zhang ◽  
Yinyu Nie ◽  
Yao Lyu ◽  
Xiaosong Yang ◽  
Jian Chang ◽  
...  

Abstract Purpose Surgical gesture recognition has been an essential task for providing intraoperative context-aware assistance and scheduling clinical resources. However, previous methods present limitations in catching long-range temporal information, and many of them require additional sensors. To address these challenges, we propose a symmetric dilated network, namely SD-Net, to jointly recognize surgical gestures and assess surgical skill levels only using RGB surgical video sequences. Methods We utilize symmetric 1D temporal dilated convolution layers to hierarchically capture gesture clues under different receptive fields such that features in different time span can be aggregated. In addition, a self-attention network is bridged in the middle to calculate the global frame-to-frame relativity. Results We evaluate our method on a robotic suturing task from the JIGSAWS dataset. The gesture recognition task largely outperforms the state of the arts on the frame-wise accuracy up to $$\sim $$ ∼ 6 points and the F1@50 score $$\sim $$ ∼ 8 points. We also keep the 100% predicted accuracy for the skill assessment task using LOSO validation scheme. Conclusion The results indicate that our architecture is able to obtain representative surgical video features by extensively considering the spatial, temporal and relational context from raw video input. Furthermore, the better performance in multi-task learning implies that surgical skill assessment has a complementary effects to gesture recognition task.

2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Joël L. Lavanchy ◽  
Joel Zindel ◽  
Kadir Kirtac ◽  
Isabell Twick ◽  
Enes Hosgor ◽  
...  

AbstractSurgical skills are associated with clinical outcomes. To improve surgical skills and thereby reduce adverse outcomes, continuous surgical training and feedback is required. Currently, assessment of surgical skills is a manual and time-consuming process which is prone to subjective interpretation. This study aims to automate surgical skill assessment in laparoscopic cholecystectomy videos using machine learning algorithms. To address this, a three-stage machine learning method is proposed: first, a Convolutional Neural Network was trained to identify and localize surgical instruments. Second, motion features were extracted from the detected instrument localizations throughout time. Third, a linear regression model was trained based on the extracted motion features to predict surgical skills. This three-stage modeling approach achieved an accuracy of 87 ± 0.2% in distinguishing good versus poor surgical skill. While the technique cannot reliably quantify the degree of surgical skill yet it represents an important advance towards automation of surgical skill assessment.


2007 ◽  
Vol 6 (2) ◽  
pp. 188-191 ◽  
Author(s):  
Sharif Al-Ruzzeh ◽  
Shishir Karthik ◽  
David O'Regan

Ophthalmology ◽  
2011 ◽  
Vol 118 (2) ◽  
pp. 427-427.e5 ◽  
Author(s):  
Karl C. Golnik ◽  
Hilary Beaver ◽  
Vinod Gauba ◽  
Andrew G. Lee ◽  
Eduardo Mayorga ◽  
...  

2015 ◽  
Vol 72 (5) ◽  
pp. 910-917 ◽  
Author(s):  
Giovanni Saggio ◽  
Alessandra Lazzaro ◽  
Laura Sbernini ◽  
Francesco Maria Carrano ◽  
Davide Passi ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document