Skip to main content
Erschienen in: International Journal of Computer Assisted Radiology and Surgery 7/2023

30.05.2023 | Original Article

Using hand pose estimation to automate open surgery training feedback

verfasst von: Eddie Bkheet, Anne-Lise D’Angelo, Adam Goldbraikh, Shlomi Laufer

Erschienen in: International Journal of Computer Assisted Radiology and Surgery | Ausgabe 7/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Purpose

This research aims to facilitate the use of state-of-the-art computer vision algorithms for the automated training of surgeons and the analysis of surgical footage. By estimating 2D hand poses, we model the movement of the practitioner’s hands, and their interaction with surgical instruments, to study their potential benefit for surgical training.

Methods

We leverage pre-trained models on a publicly available hands dataset to create our own in-house dataset of 100 open surgery simulation videos with 2D hand poses. We also assess the ability of pose estimations to segment surgical videos into gestures and tool-usage segments and compare them to kinematic sensors and I3D features. Furthermore, we introduce 6 novel surgical dexterity proxies stemming from domain experts’ training advice, all of which our framework can automatically detect given raw video footage.

Results

State-of-the-art gesture segmentation accuracy of 88.35% on the open surgery simulation dataset is achieved with the fusion of 2D poses and I3D features from multiple angles. The introduced surgical skill proxies presented significant differences for novices compared to experts and produced actionable feedback for improvement.

Conclusion

This research demonstrates the benefit of pose estimations for open surgery by analyzing their effectiveness in gesture segmentation and skill assessment. Gesture segmentation using pose estimations achieved comparable results to physical sensors while being remote and markerless. Surgical dexterity proxies that rely on pose estimation proved they can be used to work toward automated training feedback. We hope our findings encourage additional collaboration on novel skill proxies to make surgical training more efficient.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
2.
Zurück zum Zitat Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller P-A (2018) Evaluating surgical skills from kinematic data using convolutional neural networks. In: MICCAI, pp 214–221 Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller P-A (2018) Evaluating surgical skills from kinematic data using convolutional neural networks. In: MICCAI, pp 214–221
11.
Zurück zum Zitat Gao Y, Vedula SS, Reiley CE, Ahmidi N, Varadarajan B, Lin HC, Tao L, Zappella L, B’ejar B, Yuh DD et al (2014) Jhu-isi gesture and skill assessment working set (jigsaws): a surgical activity dataset for human motion modeling. In: MICCAI Workshop: M2cai, vol 3, p 3 Gao Y, Vedula SS, Reiley CE, Ahmidi N, Varadarajan B, Lin HC, Tao L, Zappella L, B’ejar B, Yuh DD et al (2014) Jhu-isi gesture and skill assessment working set (jigsaws): a surgical activity dataset for human motion modeling. In: MICCAI Workshop: M2cai, vol 3, p 3
12.
Zurück zum Zitat van Amsterdam B, Funke I, Edwards E, Speidel S, Collins J, Sridhar A, Kelly J, Clarkson MJ, Stoyanov D (2022) Gesture recognition in robotic surgery with multimodal attention. IEEE Trans Med Imag 41(7):1677–1687CrossRef van Amsterdam B, Funke I, Edwards E, Speidel S, Collins J, Sridhar A, Kelly J, Clarkson MJ, Stoyanov D (2022) Gesture recognition in robotic surgery with multimodal attention. IEEE Trans Med Imag 41(7):1677–1687CrossRef
13.
Zurück zum Zitat Zhang M, Cheng X, Copeland D, Desai A, Guan MY, Brat GA, Yeung S (2020) Using computer vision to automate hand detection and tracking of surgeon movements in videos of open surgery. AMIA ... Annual Symposium proceedings. AMIA Symposium 2020, pp 1373–1382 Zhang M, Cheng X, Copeland D, Desai A, Guan MY, Brat GA, Yeung S (2020) Using computer vision to automate hand detection and tracking of surgeon movements in videos of open surgery. AMIA ... Annual Symposium proceedings. AMIA Symposium 2020, pp 1373–1382
14.
Zurück zum Zitat Louis N, Zhou L, Yule SJ, Dias RD, Manojlovich M, Pagani FD, Likosky DS, Corso JJ (2022) Temporally guided articulated hand pose tracking in surgical videos. IJCARS Louis N, Zhou L, Yule SJ, Dias RD, Manojlovich M, Pagani FD, Likosky DS, Corso JJ (2022) Temporally guided articulated hand pose tracking in surgical videos. IJCARS
15.
Zurück zum Zitat Goodman ED, Patel KK, Zhang Y, Locke W, Kennedy CJ, Mehrotra R, Ren S, Guan MY, Downing M, Chen HW, Clark JZ, Brat GA, Yeung S (2021) A real-time spatiotemporal AI model analyzes skill in open surgical videos. arXiv Goodman ED, Patel KK, Zhang Y, Locke W, Kennedy CJ, Mehrotra R, Ren S, Guan MY, Downing M, Chen HW, Clark JZ, Brat GA, Yeung S (2021) A real-time spatiotemporal AI model analyzes skill in open surgical videos. arXiv
18.
Zurück zum Zitat Schneider P, Memmesheimer R, Kramer I, Paulus D (2019) Gesture recognition in rgb videos using human body keypoints and dynamic time warping. In: Chalup S, Niemueller T, Suthakorn J, Williams M-A (eds.) RoboCup 2019: Robot world cup XXIII, pp 281–293. Springer Schneider P, Memmesheimer R, Kramer I, Paulus D (2019) Gesture recognition in rgb videos using human body keypoints and dynamic time warping. In: Chalup S, Niemueller T, Suthakorn J, Williams M-A (eds.) RoboCup 2019: Robot world cup XXIII, pp 281–293. Springer
19.
Zurück zum Zitat Ge Z, Liu S, Wang F, Li Z, Sun J (2021) YOLOX: Exceeding YOLO Series in 2021. arXiv Ge Z, Liu S, Wang F, Li Z, Sun J (2021) YOLOX: Exceeding YOLO Series in 2021. arXiv
20.
Zurück zum Zitat Chen K, Wang J, Pang J, Cao Y, Xiong Y, Li X, Sun S, Feng W, Liu Z, Xu J, Zhang Z, Cheng D, Zhu C, Cheng T, Zhao Q, Li B, Lu X, Zhu R, Wu Y, Dai J, Wang J, Shi J, Ouyang W, Loy CC, Lin D (2019) MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv Chen K, Wang J, Pang J, Cao Y, Xiong Y, Li X, Sun S, Feng W, Liu Z, Xu J, Zhang Z, Cheng D, Zhu C, Cheng T, Zhao Q, Li B, Lu X, Zhu R, Wu Y, Dai J, Wang J, Shi J, Ouyang W, Loy CC, Lin D (2019) MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv
21.
Zurück zum Zitat Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Dollár P, Zitnick CL (2014) Microsoft coco: common objects in context. In: Fleet D, Pajdla T, Schiele B, Tuytelaars T (eds.) ECCV, pp 740–755. Springer Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Dollár P, Zitnick CL (2014) Microsoft coco: common objects in context. In: Fleet D, Pajdla T, Schiele B, Tuytelaars T (eds.) ECCV, pp 740–755. Springer
23.
Zurück zum Zitat Sun K, Xiao B, Liu D, Wang J (2019) Deep high-resolution representation learning for human pose estimation. In: CVPR. IEEE Sun K, Xiao B, Liu D, Wang J (2019) Deep high-resolution representation learning for human pose estimation. In: CVPR. IEEE
24.
Zurück zum Zitat Xiao B, Wu H, Wei Y (2018) Simple baselines for human pose estimation and tracking. In: ECCV, pp 472–487 Xiao B, Wu H, Wei Y (2018) Simple baselines for human pose estimation and tracking. In: ECCV, pp 472–487
25.
Zurück zum Zitat Wang Y, Peng C, Liu Y (2019) Mask-pose cascaded CNN for 2D hand pose estimation from single color image. IEEE Trans Circuits Syst Video Technol 29(11):3258–3268CrossRef Wang Y, Peng C, Liu Y (2019) Mask-pose cascaded CNN for 2D hand pose estimation from single color image. IEEE Trans Circuits Syst Video Technol 29(11):3258–3268CrossRef
27.
Zurück zum Zitat Li S, Farha Y, Liu Y, Cheng M-M, Gall J (2020) MS-TCN++: Multi-stage temporal convolutional network for action segmentation. In: IEEE Transactions on pattern analysis and machine intelligence Li S, Farha Y, Liu Y, Cheng M-M, Gall J (2020) MS-TCN++: Multi-stage temporal convolutional network for action segmentation. In: IEEE Transactions on pattern analysis and machine intelligence
28.
Zurück zum Zitat Carreira J, Zisserman A (2017) Quo vadis, action recognition? a new model and the kinetics dataset. In: CVPR, pp 4724–4733. IEEE Carreira J, Zisserman A (2017) Quo vadis, action recognition? a new model and the kinetics dataset. In: CVPR, pp 4724–4733. IEEE
29.
Zurück zum Zitat Kay W, Carreira J, Simonyan K, Zhang B, Hillier C, Vijayanarasimhan S, Viola F, Green T, Back T, Natsev P, Suleyman M, Zisserman A (2017) The kinetics human action video dataset. arXiv Kay W, Carreira J, Simonyan K, Zhang B, Hillier C, Vijayanarasimhan S, Viola F, Green T, Back T, Natsev P, Suleyman M, Zisserman A (2017) The kinetics human action video dataset. arXiv
30.
Zurück zum Zitat Trott AT (2012) In: Wounds and Lacerations. Saunders, Fourth edition edn. W.B Trott AT (2012) In: Wounds and Lacerations. Saunders, Fourth edition edn. W.B
31.
Zurück zum Zitat Kantor J (2017) Atlas of suturing techniques: approaches to surgical wound, laceration, and cosmetic repair. McGraw-Hill Education Kantor J (2017) Atlas of suturing techniques: approaches to surgical wound, laceration, and cosmetic repair. McGraw-Hill Education
32.
Zurück zum Zitat Meloni S, Mastenbjork M (2019) In: Suture like a surgeon vol. 1. Medical Creations Meloni S, Mastenbjork M (2019) In: Suture like a surgeon vol. 1. Medical Creations
Metadaten
Titel
Using hand pose estimation to automate open surgery training feedback
verfasst von
Eddie Bkheet
Anne-Lise D’Angelo
Adam Goldbraikh
Shlomi Laufer
Publikationsdatum
30.05.2023
Verlag
Springer International Publishing
Erschienen in
International Journal of Computer Assisted Radiology and Surgery / Ausgabe 7/2023
Print ISSN: 1861-6410
Elektronische ISSN: 1861-6429
DOI
https://doi.org/10.1007/s11548-023-02947-6

Weitere Artikel der Ausgabe 7/2023

International Journal of Computer Assisted Radiology and Surgery 7/2023 Zur Ausgabe

Premium Partner