Skip to main content

2016 | OriginalPaper | Buchkapitel

Medical Image Description Using Multi-task-loss CNN

verfasst von : Pavel Kisilev, Eli Sason, Ella Barkan, Sharbell Hashoul

Erschienen in: Deep Learning and Data Labeling for Medical Applications

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Automatic detection and classification of lesions in medical images remains one of the most important and challenging problems. In this paper, we present a new multi-task convolutional neural network (CNN) approach for detection and semantic description of lesions in diagnostic images. The proposed CNN-based architecture is trained to generate and rank rectangular regions of interests (ROI’s) surrounding suspicious areas. The highest score candidates are fed into the subsequent network layers. These layers are trained to generate semantic description of the remaining ROI’s.
During the training stage, our approach uses rectangular ground truth boxes; it does not require accurately delineated lesion contours. It has a clear advantage for supervised training on large datasets. Our system learns discriminative features which are shared in the Detection and the Description stages. This eliminates the need for hand-crafted features, and allows application of the method to new modalities and organs with minimal overhead. The proposed approach generates medical report by estimating standard radiological lexicon descriptors which are a basis for diagnosis. The proposed approach should help radiologists to understand a diagnostic decision of a computer aided diagnosis (CADx) system. We test the proposed method on proprietary and publicly available breast databases, and show that our method outperforms the competing approaches.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Guillaumin, M., Mensink, T., Verbeek, J.J., Schmid, C.: Tagprop: Discriminative metric learning in nearest neighbor models for image auto-annotation. In: ICCV (2009) Guillaumin, M., Mensink, T., Verbeek, J.J., Schmid, C.: Tagprop: Discriminative metric learning in nearest neighbor models for image auto-annotation. In: ICCV (2009)
2.
Zurück zum Zitat Farhadi, A., Hejrati, M., Sadeghi, M.A., Young, P., Rashtchian, C., Hockenmaier, J., Forsyth, D.: Every picture tells a story: generating sentences from images. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6316, pp. 15–29. Springer, Heidelberg (2010). doi:10.1007/978-3-642-15561-1_2 CrossRef Farhadi, A., Hejrati, M., Sadeghi, M.A., Young, P., Rashtchian, C., Hockenmaier, J., Forsyth, D.: Every picture tells a story: generating sentences from images. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6316, pp. 15–29. Springer, Heidelberg (2010). doi:10.​1007/​978-3-642-15561-1_​2 CrossRef
3.
Zurück zum Zitat V. Ordonez, G. Kulkarni and T. L. Berg. Im2Text: Describing Images Using 1 Million Captioned Photographs. In NIPS 2011, pages 1143–1151 V. Ordonez, G. Kulkarni and T. L. Berg. Im2Text: Describing Images Using 1 Million Captioned Photographs. In NIPS 2011, pages 1143–1151
4.
Zurück zum Zitat Elliott, D., Keller, F.: Image description using visual dependency representations. EMNLP 13, 1292–1302 (2013) Elliott, D., Keller, F.: Image description using visual dependency representations. EMNLP 13, 1292–1302 (2013)
5.
Zurück zum Zitat D’Orsi, C.J., Mendelson, E.B., Ikeda, D.M., et al.: Breast imaging reporting and data system: ACR BI-RADS - breast imaging atlas. American College of Radiology, Reston (2003) D’Orsi, C.J., Mendelson, E.B., Ikeda, D.M., et al.: Breast imaging reporting and data system: ACR BI-RADS - breast imaging atlas. American College of Radiology, Reston (2003)
6.
Zurück zum Zitat Weinreb, J., et al.: PI-RADS prostate imaging - reporting and data system: 2015, Version 2. Eur. Urol. 69(1), 16–40 (2016)CrossRef Weinreb, J., et al.: PI-RADS prostate imaging - reporting and data system: 2015, Version 2. Eur. Urol. 69(1), 16–40 (2016)CrossRef
7.
Zurück zum Zitat Mitchell, D., et al.: Li-RAD in liver lesions. Hepatology 61(3), 1056–1065 (2015)CrossRef Mitchell, D., et al.: Li-RAD in liver lesions. Hepatology 61(3), 1056–1065 (2015)CrossRef
8.
Zurück zum Zitat Oliver, A., Freixenet, J., Martí, J., Pérez, E., Pont, J., Denton, E.R., Zwiggelaar, R.: A review of automatic mass detection and segmentation in mammographic images. Med. Image Anal. 14(2), 87–110 (2010)CrossRef Oliver, A., Freixenet, J., Martí, J., Pérez, E., Pont, J., Denton, E.R., Zwiggelaar, R.: A review of automatic mass detection and segmentation in mammographic images. Med. Image Anal. 14(2), 87–110 (2010)CrossRef
9.
Zurück zum Zitat Wei, C.-H., Li, Y., Huang, P.J.: Mammogram retrieval through machine learning within BI-RADS standards. J. Biomed. Inform. 44(4), 607–614 (2011)CrossRef Wei, C.-H., Li, Y., Huang, P.J.: Mammogram retrieval through machine learning within BI-RADS standards. J. Biomed. Inform. 44(4), 607–614 (2011)CrossRef
10.
Zurück zum Zitat Rubin, D.L., Burnside, E.S., Shachter, R.: A bayesian network to assist mammography interpretation. In: Brandeau, M.L., Sainfort, F., Pierskalla, W.P. (eds.) Operations Research and Health Care. International Series in Operations Research & Management Science, vol. 70, pp. 695–720. Springer, New York (2004)CrossRef Rubin, D.L., Burnside, E.S., Shachter, R.: A bayesian network to assist mammography interpretation. In: Brandeau, M.L., Sainfort, F., Pierskalla, W.P. (eds.) Operations Research and Health Care. International Series in Operations Research & Management Science, vol. 70, pp. 695–720. Springer, New York (2004)CrossRef
11.
Zurück zum Zitat Narvaez, F., Diaz, G., Romero, E.: Automatic BI-RADS description of mammographic masses. In: Martí, J., Oliver, A., Freixenet, A., Martí, R. (eds.) Digital Mammography. Lecture Notes in Computer Science, vol. 6316, pp. 673–681. Springer, New York (2010)CrossRef Narvaez, F., Diaz, G., Romero, E.: Automatic BI-RADS description of mammographic masses. In: Martí, J., Oliver, A., Freixenet, A., Martí, R. (eds.) Digital Mammography. Lecture Notes in Computer Science, vol. 6316, pp. 673–681. Springer, New York (2010)CrossRef
12.
Zurück zum Zitat Kisilev, P., Walach, E., Hashoul, S., Barkan, E., Ophir, B., Alpert, S.: Semantic description of medical image findings: structured learning approach. In: BMVC (2015) Kisilev, P., Walach, E., Hashoul, S., Barkan, E., Ophir, B., Alpert, S.: Semantic description of medical image findings: structured learning approach. In: BMVC (2015)
13.
Zurück zum Zitat Heath, M., Bowyer, K., Kopans, D., Moore, R., Philip Kegelmeyer, W.: The digital database for screening mammography. In: Yaffe, M.J. (ed.) Proceedings of the Fifth International Workshop on Digital Mammography, pp. 212–218. Medical Physics Publishing, Madison (2001) Heath, M., Bowyer, K., Kopans, D., Moore, R., Philip Kegelmeyer, W.: The digital database for screening mammography. In: Yaffe, M.J. (ed.) Proceedings of the Fifth International Workshop on Digital Mammography, pp. 212–218. Medical Physics Publishing, Madison (2001)
14.
Zurück zum Zitat Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: Towards real-time object detection with region proposal networks. In: NIPS (2015) Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: Towards real-time object detection with region proposal networks. In: NIPS (2015)
15.
Zurück zum Zitat Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, R., Guadarrama, S., Darrell, T.: Caffe: Convolutional architecture for fast feature embedding (2014). arXiv:1408.5093 Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, R., Guadarrama, S., Darrell, T.: Caffe: Convolutional architecture for fast feature embedding (2014). arXiv:​1408.​5093
Metadaten
Titel
Medical Image Description Using Multi-task-loss CNN
verfasst von
Pavel Kisilev
Eli Sason
Ella Barkan
Sharbell Hashoul
Copyright-Jahr
2016
DOI
https://doi.org/10.1007/978-3-319-46976-8_13

Premium Partner