Skip to main content

2021 | OriginalPaper | Buchkapitel

This Looks Like That, Because ... Explaining Prototypes for Interpretable Image Recognition

verfasst von : Meike Nauta, Annemarie Jutte, Jesper Provoost, Christin Seifert

Erschienen in: Machine Learning and Principles and Practice of Knowledge Discovery in Databases

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Image recognition with prototypes is considered an interpretable alternative for black box deep learning models. Classification depends on the extent to which a test image “looks like” a prototype. However, perceptual similarity for humans can be different from the similarity learned by the classification model. Hence, only visualising prototypes can be insufficient for a user to understand what a prototype exactly represents, and why the model considers a prototype and an image to be similar. We address this ambiguity and argue that prototypes should be explained. We improve interpretability by automatically enhancing visual prototypes with quantitative information about visual characteristics deemed important by the classification model. Specifically, our method clarifies the meaning of a prototype by quantifying the influence of colour hue, shape, texture, contrast and saturation and can generate both global and local explanations. Because of the generality of our approach, it can improve the interpretability of any similarity-based method for prototypical image recognition. In our experiments, we apply our method to the existing Prototypical Part Network (ProtoPNet). Our analysis confirms that the global explanations are generalisable, and often correspond to the visually perceptible properties of a prototype. Our explanations are especially relevant for prototypes which might have been interpreted incorrectly otherwise. By explaining such ‘misleading’ prototypes, we improve the interpretability and simulatability of a prototype-based classification model. We also use our method to check whether visually similar prototypes have similar explanations, and are able to discover redundancy. Code is available at https://​github.​com/​M-Nauta/​Explaining_​Prototypes.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
We use the same methodology as ProtoPNet [7] in order to reproduce results, although it is known that there is some overlap between Caltech-UCSD and ImageNet.
 
Literatur
1.
Zurück zum Zitat Arik, S.Ö., Pfister, T.: Attention-based prototypical learning towards interpretable, confident and robust deep neural networks. CoRR abs/1902.06292 (2019) Arik, S.Ö., Pfister, T.: Attention-based prototypical learning towards interpretable, confident and robust deep neural networks. CoRR abs/1902.06292 (2019)
2.
Zurück zum Zitat Bau, D., Zhou, B., Khosla, A., Oliva, A., Torralba, A.: Network dissection: quantifying interpretability of deep visual representations. In: CVPR (2017) Bau, D., Zhou, B., Khosla, A., Oliva, A., Torralba, A.: Network dissection: quantifying interpretability of deep visual representations. In: CVPR (2017)
3.
Zurück zum Zitat Op de Beeck, H.P., Torfs, K., Wagemans, J.: Perceived shape similarity among unfamiliar objects and the organization of the human object vision pathway. J. Neurosci. 28(40), 10111–10123 (2008) Op de Beeck, H.P., Torfs, K., Wagemans, J.: Perceived shape similarity among unfamiliar objects and the organization of the human object vision pathway. J. Neurosci. 28(40), 10111–10123 (2008)
4.
Zurück zum Zitat Biehl, M., Hammer, B., Villmann, T.: Prototype-based models in machine learning. Wiley Interdisc. Rev. Cognitive Sci. 7(2), 92–111 (2016)CrossRef Biehl, M., Hammer, B., Villmann, T.: Prototype-based models in machine learning. Wiley Interdisc. Rev. Cognitive Sci. 7(2), 92–111 (2016)CrossRef
5.
Zurück zum Zitat Cavina-Pratesi, C., Kentridge, R., Heywood, C., Milner, A.: Separate channels for processing form, texture, and color: evidence from fMRI adaptation and visual object agnosia. Cereb. Cortex 20(10), 2319–2332 (2010) Cavina-Pratesi, C., Kentridge, R., Heywood, C., Milner, A.: Separate channels for processing form, texture, and color: evidence from fMRI adaptation and visual object agnosia. Cereb. Cortex 20(10), 2319–2332 (2010)
6.
Zurück zum Zitat Chaudhari, S., Polatkan, G., Ramanath, R., Mithal, V.: An attentive survey of attention models. CoRR abs/1904.02874 (2019) Chaudhari, S., Polatkan, G., Ramanath, R., Mithal, V.: An attentive survey of attention models. CoRR abs/1904.02874 (2019)
7.
Zurück zum Zitat Chen, C., Li, O., Tao, D., Barnett, A., Rudin, C., Su, J.K.: This looks like that: deep learning for interpretable image recognition. In: NeurIPS, pp. 8928–8939 (2019) Chen, C., Li, O., Tao, D., Barnett, A., Rudin, C., Su, J.K.: This looks like that: deep learning for interpretable image recognition. In: NeurIPS, pp. 8928–8939 (2019)
8.
Zurück zum Zitat Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F.A., Brendel, W.: Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In: ICLR (2019) Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F.A., Brendel, W.: Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In: ICLR (2019)
9.
Zurück zum Zitat Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge (2016) Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge (2016)
10.
Zurück zum Zitat Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., Pedreschi, D.: A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 51(5), 1–42 (2018)CrossRef Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., Pedreschi, D.: A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 51(5), 1–42 (2018)CrossRef
11.
Zurück zum Zitat Hosseini, H., Xiao, B., Jaiswal, M., Poovendran, R.: Assessing shape bias property of convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (June 2018) Hosseini, H., Xiao, B., Jaiswal, M., Poovendran, R.: Assessing shape bias property of convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (June 2018)
12.
Zurück zum Zitat Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: CVPR, pp. 4700–4708 (2017) Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: CVPR, pp. 4700–4708 (2017)
13.
Zurück zum Zitat King, M.L., Groen, I.I., Steel, A., Kravitz, D.J., Baker, C.I.: Similarity judgments and cortical visual responses reflect different properties of object and scene categories in naturalistic images. NeuroImage 197, 368–382 (2019)CrossRef King, M.L., Groen, I.I., Steel, A., Kravitz, D.J., Baker, C.I.: Similarity judgments and cortical visual responses reflect different properties of object and scene categories in naturalistic images. NeuroImage 197, 368–382 (2019)CrossRef
14.
Zurück zum Zitat Kourtzi, Z., Kanwisher, N.: Cortical regions involved in perceiving object shape. J. Neurosci. 20(9), 3310–3318 (2000)CrossRef Kourtzi, Z., Kanwisher, N.: Cortical regions involved in perceiving object shape. J. Neurosci. 20(9), 3310–3318 (2000)CrossRef
15.
Zurück zum Zitat Lipton, Z.C.: The mythos of model interpretability. Queue 16(3), 31–57 (2018)CrossRef Lipton, Z.C.: The mythos of model interpretability. Queue 16(3), 31–57 (2018)CrossRef
16.
Zurück zum Zitat Munzner, T., Maguire, E.: Visualization Analysis & Design. CRC Press, Boca Raton (2015) Munzner, T., Maguire, E.: Visualization Analysis & Design. CRC Press, Boca Raton (2015)
17.
Zurück zum Zitat Nauta, M., van Bree, R., Seifert, C.: Neural prototype trees for interpretable fine-grained image recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 14933–14943 (June 2021) Nauta, M., van Bree, R., Seifert, C.: Neural prototype trees for interpretable fine-grained image recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 14933–14943 (June 2021)
18.
Zurück zum Zitat Ritter, S., Barrett, D.G.T., Santoro, A., Botvinick, M.M.: Cognitive psychology for deep neural networks: a shape bias case study. In: ICML Proceedings of Machine Learning Research, vol. 70, pp. 2940–2949. PMLR (2017) Ritter, S., Barrett, D.G.T., Santoro, A., Botvinick, M.M.: Cognitive psychology for deep neural networks: a shape bias case study. In: ICML Proceedings of Machine Learning Research, vol. 70, pp. 2940–2949. PMLR (2017)
19.
Zurück zum Zitat Rosenfeld, A., Solbach, M.D., Tsotsos, J.K.: Totally looks like - how humans compare, compared to machines. In: CVPR Workshops (2018) Rosenfeld, A., Solbach, M.D., Tsotsos, J.K.: Totally looks like - how humans compare, compared to machines. In: CVPR Workshops (2018)
20.
Zurück zum Zitat Rossion, B., Pourtois, G.: Revisiting snodgrass and vanderwart’s object pictorial set: The role of surface detail in basic-level object recognition. Perception 33(2), 217–236 (2004). pMID: 15109163 Rossion, B., Pourtois, G.: Revisiting snodgrass and vanderwart’s object pictorial set: The role of surface detail in basic-level object recognition. Perception 33(2), 217–236 (2004). pMID: 15109163
21.
Zurück zum Zitat Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1(5), 206–215 (2019)CrossRef Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1(5), 206–215 (2019)CrossRef
22.
Zurück zum Zitat Saralajew, S., Holdijk, L., Rees, M., Asan, E., Villmann, T.: Classification-by-components: probabilistic modeling of reasoning over a set of components. In: NeurIPS, pp. 2792–2803 (2019) Saralajew, S., Holdijk, L., Rees, M., Asan, E., Villmann, T.: Classification-by-components: probabilistic modeling of reasoning over a set of components. In: NeurIPS, pp. 2792–2803 (2019)
23.
Zurück zum Zitat Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: Bengio, Y., LeCun, Y. (eds.) ICLR (2015) Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: Bengio, Y., LeCun, Y. (eds.) ICLR (2015)
24.
Zurück zum Zitat Striedter, G.F.: Neurobiology: a Functional Approach. Oxford University Press, Oxford (2016) Striedter, G.F.: Neurobiology: a Functional Approach. Oxford University Press, Oxford (2016)
25.
Zurück zum Zitat Welinder, P., et al.: Caltech-UCSD Birds 200. Technical Report. CNS-TR-2010-001 (2010) Welinder, P., et al.: Caltech-UCSD Birds 200. Technical Report. CNS-TR-2010-001 (2010)
27.
Zurück zum Zitat Zheng, H., Fu, J., Zha, Z., Luo, J., Mei, T.: Learning rich part hierarchies with progressive attention networks for fine-grained image recognition. IEEE Trans. Image Process. 29, 476–488 (2020)MathSciNetCrossRef Zheng, H., Fu, J., Zha, Z., Luo, J., Mei, T.: Learning rich part hierarchies with progressive attention networks for fine-grained image recognition. IEEE Trans. Image Process. 29, 476–488 (2020)MathSciNetCrossRef
Metadaten
Titel
This Looks Like That, Because ... Explaining Prototypes for Interpretable Image Recognition
verfasst von
Meike Nauta
Annemarie Jutte
Jesper Provoost
Christin Seifert
Copyright-Jahr
2021
DOI
https://doi.org/10.1007/978-3-030-93736-2_34

Premium Partner