Towards accurate surgical workflow recognition with convolutional networks and transformers

Author(s):  
Bokai Zhang ◽  
Julian Abbing ◽  
Amer Ghanem ◽  
Danyal Fer ◽  
Jocelyn Barker ◽  
...  
Author(s):  
Marco Cenzato ◽  
Roberto Stefini ◽  
Francesco Zenga ◽  
Maurizio Piparo ◽  
Alberto Debernardi ◽  
...  

Abstract Background Cerebellopontine angle (CPA) surgery carries the risk of lesioning the facial nerve. The goal of preserving the integrity of the facial nerve is usually pursued with intermittent electrical stimulation using a handheld probe that is alternated with the resection. We report our experience with continuous electrical stimulation delivered via the ultrasonic aspirator (UA) used for the resection of a series of vestibular schwannomas. Methods A total of 17 patients with vestibular schwannomas, operated on between 2010 and 2018, were included in this study. A constant-current stimulator was coupled to the UA used for the resection, delivering square-wave pulses throughout the resection. The muscle responses from upper and lower face muscles triggered by the electrical stimulation were displayed continuously on multichannel neurophysiologic equipment. The careful titration of the electrical stimulation delivered through the UA while tapering the current intensity with the progression of the resection was used as the main strategy. Results All operations were performed successfully, and facial nerve conduction was maintained in all patients except one, in whom a permanent lesion of the facial nerve followed a miscommunication to the neurosurgeon. Conclusion The coupling of the electrical stimulation to the UA provided the neurosurgeon with an efficient and cost-effective tool and allowed a safe resection. Positive responses were obtained from the facial muscles with low current intensity (lowest intensity: 0.1 mA). The availability of a resection tool paired with a stimulator allowed the surgeon to improve the surgical workflow because fewer interruptions were necessary to stimulate the facial nerve via a handheld probe.


Author(s):  
Hao Chen ◽  
Yue Xu ◽  
Feiran Huang ◽  
Zengde Deng ◽  
Wenbing Huang ◽  
...  

2021 ◽  
Vol 11 (15) ◽  
pp. 6975
Author(s):  
Tao Zhang ◽  
Lun He ◽  
Xudong Li ◽  
Guoqing Feng

Lipreading aims to recognize sentences being spoken by a talking face. In recent years, the lipreading method has achieved a high level of accuracy on large datasets and made breakthrough progress. However, lipreading is still far from being solved, and existing methods tend to have high error rates on the wild data and have the defects of disappearing training gradient and slow convergence. To overcome these problems, we proposed an efficient end-to-end sentence-level lipreading model, using an encoder based on a 3D convolutional network, ResNet50, Temporal Convolutional Network (TCN), and a CTC objective function as the decoder. More importantly, the proposed architecture incorporates TCN as a feature learner to decode feature. It can partly eliminate the defects of RNN (LSTM, GRU) gradient disappearance and insufficient performance, and this yields notable performance improvement as well as faster convergence. Experiments show that the training and convergence speed are 50% faster than the state-of-the-art method, and improved accuracy by 2.4% on the GRID dataset.


Sign in / Sign up

Export Citation Format

Share Document