Skip to main content
Erschienen in: International Journal of Computer Assisted Radiology and Surgery 6/2013

01.11.2013 | Original Article

Speech and motion control for interventional radiology: requirements and feasibility

verfasst von: Andreas M. Hötker, Michael B. Pitton, Peter Mildenberger, Christoph Düber

Erschienen in: International Journal of Computer Assisted Radiology and Surgery | Ausgabe 6/2013

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Purpose

Interventional radiology is performed in a sterile environment where speech and motion control of image review is needed to simplify and expedite routine procedures. The requirements and limitations were defined by testing an interventional radiology test bed speech and motion control system.

Methods

Motion control software was implemented using the Microsoft\(^{\circledR }\) Kinect\(^{\circledR }\) (Microsoft Corp., USA) framework. The system was tested by 10 participants using a predefined set of six voice and six gesture commands under different lighting conditions to assess the influence of illumination on command recognition. The participants rated the convenience of the application and its possible use in everyday clinical routine. A basic set of voice or gesture commands required for interventional radiology were identified.

Results

The majority (93 %) of commands were recognized successfully. Speech commands were less prone to errors than gesture commands. Unwanted side effects occurred (e.g., accidentally issuing a gesture command) in about 30 % of cases. Dimmed lighting conditions did not have a measurable effect on the recognition rate. Six out of 10 participants would consider using the application in everyday routine. The necessary voice/gesture commands for interventional radiology were identified and integrated into the control system.

Conclusion

Speech and motion control of image review provides a new man–machine interface for radiological image handling that is especially useful in sterile environments due to no-touch navigation. Command recognition rates were high and remained stable under different lighting conditions. However, the rate of accidental triggering due to unintended commands should be reduced.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Dargar S, Nunno A, Sankaranarayanan G et al (2013) Microsoft kinect based head tracking for life size collaborative surgical simulation environments (LS-CollaSSLE). Stud Health Technol Inform 184:109–113PubMed Dargar S, Nunno A, Sankaranarayanan G et al (2013) Microsoft kinect based head tracking for life size collaborative surgical simulation environments (LS-CollaSSLE). Stud Health Technol Inform 184:109–113PubMed
2.
Zurück zum Zitat Guo X, Lopez LD, Yu Z et al (2013) A portable immersive surgery training system using RGB-D sensors. Stud Health Technol Inform 184:161–167 Guo X, Lopez LD, Yu Z et al (2013) A portable immersive surgery training system using RGB-D sensors. Stud Health Technol Inform 184:161–167
4.
Zurück zum Zitat Cook TS, Couch G, Couch TJ et al (2013) Using the microsoft kinect for patient size estimation and radiation dose normalization: proof of concept and initial validation. J Digit Imaging. doi:10.1007/s10278-012-9567-2 Cook TS, Couch G, Couch TJ et al (2013) Using the microsoft kinect for patient size estimation and radiation dose normalization: proof of concept and initial validation. J Digit Imaging. doi:10.​1007/​s10278-012-9567-2
5.
7.
Zurück zum Zitat Aggarwal JK, Ryoo MS (2011) Human activity analysis: a review. ACM Comput Surv (CSUR) 43(3):16CrossRef Aggarwal JK, Ryoo MS (2011) Human activity analysis: a review. ACM Comput Surv (CSUR) 43(3):16CrossRef
8.
Zurück zum Zitat Holte MB, Moeslund TB (2008) View invariant gesture recognition using 3D motion primitives. In: Acoustics, speech and signal processing, 2008. ICASSP 2008. IEEE International Conference on. IEEE, pp 797–800 Holte MB, Moeslund TB (2008) View invariant gesture recognition using 3D motion primitives. In: Acoustics, speech and signal processing, 2008. ICASSP 2008. IEEE International Conference on. IEEE, pp 797–800
9.
Zurück zum Zitat Ren Z, Meng J, Yuan J (2011) Depth camera based hand gesture recognition and its applications in human-computer-interaction. In: Communications and signal processing (ICICS) 2011 8th international conference on information: 1–5. doi:10.1109/ICICS.2011.6173545 Ren Z, Meng J, Yuan J (2011) Depth camera based hand gesture recognition and its applications in human-computer-interaction. In: Communications and signal processing (ICICS) 2011 8th international conference on information: 1–5. doi:10.​1109/​ICICS.​2011.​6173545
10.
Zurück zum Zitat Oikonomidis I, Kyriazis N, Argyros A (2011) Markerless and efficient 26-DOF hand pose recovery. In: Kimmel R, Klette R, Sugimoto A (eds) Computer vision—ACCV 2010, vol 6494. Springer, Berlin, pp 744–757 Oikonomidis I, Kyriazis N, Argyros A (2011) Markerless and efficient 26-DOF hand pose recovery. In: Kimmel R, Klette R, Sugimoto A (eds) Computer vision—ACCV 2010, vol 6494. Springer, Berlin, pp 744–757
11.
Zurück zum Zitat Oikonomidis I, Kyriazis N, Argyros A (2011) Efficient model-based 3d tracking of hand articulations using kinect. BMVC 2011 Oikonomidis I, Kyriazis N, Argyros A (2011) Efficient model-based 3d tracking of hand articulations using kinect. BMVC 2011
Metadaten
Titel
Speech and motion control for interventional radiology: requirements and feasibility
verfasst von
Andreas M. Hötker
Michael B. Pitton
Peter Mildenberger
Christoph Düber
Publikationsdatum
01.11.2013
Verlag
Springer Berlin Heidelberg
Erschienen in
International Journal of Computer Assisted Radiology and Surgery / Ausgabe 6/2013
Print ISSN: 1861-6410
Elektronische ISSN: 1861-6429
DOI
https://doi.org/10.1007/s11548-013-0841-7

Weitere Artikel der Ausgabe 6/2013

International Journal of Computer Assisted Radiology and Surgery 6/2013 Zur Ausgabe