Evaluating robotic-assisted surgery training videos with multi-task convolutional neural networks

J Robot Surg. 2022 Aug;16(4):917-925. doi: 10.1007/s11701-021-01316-2. Epub 2021 Oct 28.

Abstract

We seek to understand if an automated algorithm can replace human scoring of surgical trainees performing the urethrovesical anastomosis in radical prostatectomy with synthetic tissue. Specifically, we investigate neural networks for predicting the surgical proficiency score (GEARS score) from video clips. We evaluate videos of surgeons performing the urethral anastomosis using synthetic tissue. The algorithm tracks surgical instrument locations from video, saving the positions of key points on the instruments over time. These positional features are used to train a multi-task convolutional network to infer each sub-category of the GEARS score to determine the proficiency level of trainees. Experimental results demonstrate that the proposed method achieves good performance with scores matching manual inspection in 86.1% of all GEARS sub-categories. Furthermore, the model can detect the difference between proficiency (novice to expert) in 83.3% of videos. Evaluation of GEARS sub-categories with artificial neural networks is possible for novice and intermediate surgeons, but additional research is needed to understand if expert surgeons can be evaluated with a similar automated system.

Keywords: Deep learning; Keypoint detection; Robotic-assisted surgery; Skill evaluation; Surgical training.

MeSH terms

  • Clinical Competence
  • Humans
  • Male
  • Neural Networks, Computer
  • Prostatectomy / education
  • Robotic Surgical Procedures* / methods
  • Surgeons* / education