Skip to main content
Top
Published in: Cognitive Computation 5/2015

01-10-2015

Affordance Estimation Enhances Artificial Visual Attention: Evidence from a Change-Blindness Study

Authors: Jan Tünnermann, Norbert Krüger, Bärbel Mertsching, Wail Mustafa

Published in: Cognitive Computation | Issue 5/2015

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Visual attention models are typically based on the concept of saliency, a conspicuity measure which considers features such as color, intensity or orientation. Much current research aims at modeling top-down interactions, which highly influence human attentional behavior. Typically, these are in the form of targets to be searched for or general characteristics (gist) of a scene. In humans, it has been shown that objects that afford actions, for example, graspable objects, strongly attract attention. Here, we integrate an artificial attention framework with a measure of affordances estimated from a sparse 3D scene representation. This work contributes further evidence for human attention being biased toward objects of high affordance, which for the first time is measured in an objective way. Furthermore, it demonstrates that artificial attention systems benefit from affordance estimation for predicting human attention. For technical systems, considering affordances provides mid-level influences that are not too specific or too general, but can guide attention toward potential action targets with respect to a system’s physical capabilities. Finally, the change-detection task we employ for model comparison constitutes a new method to evaluate artificial systems with respect to early human vision in natural scene perception.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Footnotes
2
The affordance model suggested edges of shelves twice, which could not be removed without also removing all their contents or leaving the contained objects suspiciously floating; the saliency model also suggested such an element once by pointing to empty space on a table between objects.
 
3
Note that some additional center bias is added due to an artifact: A vertical strip on the left in Fig. 6b shows no changes. This is because the left image of the stereo pair was used and no stereo correspondences (and thus no object representations and no estimated grasping possibilities) exist in the far left of the image.
 
4
All t tests and ANOVAs reported in this paper assume an alpha level of 0.05 and were performed on the arcsine-transformed relative frequencies. Whenever differences are not significant, we additionally report the mean and standard deviation of the differences and 95 % confidence intervals around the mean difference.
 
Literature
2.
go back to reference Simons DJ, Rensink RA. Change blindness: past, present, and future. Trends Cogn Sci. 2005;9(1):16–20.CrossRefPubMed Simons DJ, Rensink RA. Change blindness: past, present, and future. Trends Cogn Sci. 2005;9(1):16–20.CrossRefPubMed
3.
go back to reference Simons DJ, Levin DT. Failure to detect changes to people during a real-world interaction. Psychon Bull Rev. 1998;5:644–9.CrossRef Simons DJ, Levin DT. Failure to detect changes to people during a real-world interaction. Psychon Bull Rev. 1998;5:644–9.CrossRef
4.
go back to reference Rensink RA, O’Regan JK, Clark JJ. To see or not to see: the need for attention to perceive changes in scenes. Psychol Sci. 1997;8(5):368–73.CrossRef Rensink RA, O’Regan JK, Clark JJ. To see or not to see: the need for attention to perceive changes in scenes. Psychol Sci. 1997;8(5):368–73.CrossRef
5.
go back to reference Shore D, Klein RM. The effects of scene inversion on change blindness. J Gen Psychol. 2000;127:27–43.CrossRefPubMed Shore D, Klein RM. The effects of scene inversion on change blindness. J Gen Psychol. 2000;127:27–43.CrossRefPubMed
6.
go back to reference Kelley T, Chun M, Chua K. Effects of scene inversion on change detection of targets matched for visual salience. J Vis. 2003;3(1):1–5.CrossRefPubMed Kelley T, Chun M, Chua K. Effects of scene inversion on change detection of targets matched for visual salience. J Vis. 2003;3(1):1–5.CrossRefPubMed
7.
go back to reference Sampanes AC, Tseng P, Bridgeman B. The role of gist in scene recognition. Vis Res. 2008;48(21):2275–83.CrossRefPubMed Sampanes AC, Tseng P, Bridgeman B. The role of gist in scene recognition. Vis Res. 2008;48(21):2275–83.CrossRefPubMed
8.
go back to reference Tseng P, Tünnermann J, Roker-Knight N, Winter D, Scharlau I, Bridgeman B. Enhancing implicit change detection through action. Perception. 2010;39:1311–21.CrossRefPubMed Tseng P, Tünnermann J, Roker-Knight N, Winter D, Scharlau I, Bridgeman B. Enhancing implicit change detection through action. Perception. 2010;39:1311–21.CrossRefPubMed
9.
go back to reference Symes E, Tucker M, Ellis R, Vainio L, Ottoboni G. Grasp preparation improves change detection for congruent objects. J Exp Psychol Hum Percept Perform. 2008;34(4):854–71.CrossRefPubMed Symes E, Tucker M, Ellis R, Vainio L, Ottoboni G. Grasp preparation improves change detection for congruent objects. J Exp Psychol Hum Percept Perform. 2008;34(4):854–71.CrossRefPubMed
11.
go back to reference Tseng P, Bridgeman B, Juan CH. Take the matter into your own hands: a brief review of the effect of nearby-hands on visual processing. Vis Res. 2012;72:74–7.CrossRefPubMed Tseng P, Bridgeman B, Juan CH. Take the matter into your own hands: a brief review of the effect of nearby-hands on visual processing. Vis Res. 2012;72:74–7.CrossRefPubMed
12.
go back to reference Tünnermann J, Hilkenmeier F, Scharlau I. Change detection is enhanced for objects in near space; 2012. In: Poster presented at the 54. Tagung experimentell arbeitender Psychologen (TeaP)/54th Conference of Experimental Psychologists, Mannheim, Germany. Tünnermann J, Hilkenmeier F, Scharlau I. Change detection is enhanced for objects in near space; 2012. In: Poster presented at the 54. Tagung experimentell arbeitender Psychologen (TeaP)/54th Conference of Experimental Psychologists, Mannheim, Germany.
13.
go back to reference Stirk JA, Underwood G. Low-level visual saliency does not predict change detection in natural scenes. J Vis. 2007;7:1–10.CrossRefPubMed Stirk JA, Underwood G. Low-level visual saliency does not predict change detection in natural scenes. J Vis. 2007;7:1–10.CrossRefPubMed
14.
go back to reference Gibson JJ. The theory of affordances. In: Shaw R, Bransford J, editors. Perceiving, acting, and knowing: toward an ecological psychology. Hillsdale: Lawrence Erlbaum Associates; 1977. p. 67–82. Gibson JJ. The theory of affordances. In: Shaw R, Bransford J, editors. Perceiving, acting, and knowing: toward an ecological psychology. Hillsdale: Lawrence Erlbaum Associates; 1977. p. 67–82.
15.
go back to reference Craighero L, Fadiga L, Rizzolatti G, Umiltà C. Action for perception: a motor-visual attentional effect. J Exp Psychol Hum Percept Perform. 1999;25(6):1673–92.CrossRefPubMed Craighero L, Fadiga L, Rizzolatti G, Umiltà C. Action for perception: a motor-visual attentional effect. J Exp Psychol Hum Percept Perform. 1999;25(6):1673–92.CrossRefPubMed
16.
go back to reference Bekkering H, Neggers SFW. Visual search is modulated by action intentions. Psychol Sci. 2002;13(4):370–4.CrossRef Bekkering H, Neggers SFW. Visual search is modulated by action intentions. Psychol Sci. 2002;13(4):370–4.CrossRef
18.
go back to reference Handy TC, Grafton ST, Shroff NM, Ketay S, Gazzaniga MS. Graspable objects grab attention when the potential for action is recognized. Nat Neurosci. 2003;6(4):421–7.CrossRefPubMed Handy TC, Grafton ST, Shroff NM, Ketay S, Gazzaniga MS. Graspable objects grab attention when the potential for action is recognized. Nat Neurosci. 2003;6(4):421–7.CrossRefPubMed
19.
go back to reference Castellini C, Tommasi T, Noceti N, Odone F, Caputo B. Using object affordances to improve object recognition. IEEE Trans Auton Ment Dev. 2011;3(3):207–15.CrossRef Castellini C, Tommasi T, Noceti N, Odone F, Caputo B. Using object affordances to improve object recognition. IEEE Trans Auton Ment Dev. 2011;3(3):207–15.CrossRef
20.
go back to reference Detry R, Kraft D, Kroemer O, Bodenhagen L, Peters J, Krüger N, et al. Learning grasp affordance densities. Paladyn. 2011;2(1):1–17.CrossRef Detry R, Kraft D, Kroemer O, Bodenhagen L, Peters J, Krüger N, et al. Learning grasp affordance densities. Paladyn. 2011;2(1):1–17.CrossRef
21.
go back to reference Varadarajan KM, Vincze M. Affordance based part recognition for grasping and manipulation. In: Proceedings of the ICRA workshop on autonomous grasping; 2011. Varadarajan KM, Vincze M. Affordance based part recognition for grasping and manipulation. In: Proceedings of the ICRA workshop on autonomous grasping; 2011.
22.
go back to reference Stark L, Bowyer K. Generic recognition through qualitative reasoning about 3-D shape and object dunction. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 1991. p. 251–56. Stark L, Bowyer K. Generic recognition through qualitative reasoning about 3-D shape and object dunction. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 1991. p. 251–56.
23.
go back to reference Rivlin E, Dickinson SJ, Rosenfeld A. Recognition by functional parts [Function-based object recognition]. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 1994. p. 267–74. Rivlin E, Dickinson SJ, Rosenfeld A. Recognition by functional parts [Function-based object recognition]. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 1994. p. 267–74.
24.
go back to reference Itti L, Koch C, Niebur E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans Pattern Anal Mach Intell. 1998;20(11):1254–9.CrossRef Itti L, Koch C, Niebur E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans Pattern Anal Mach Intell. 1998;20(11):1254–9.CrossRef
25.
go back to reference Treisman AM, Gelade G. A feature-integration theory of attention. Cogn Psychol. 1980;12(1):97–136.CrossRefPubMed Treisman AM, Gelade G. A feature-integration theory of attention. Cogn Psychol. 1980;12(1):97–136.CrossRefPubMed
26.
go back to reference Koch C, Ullman S. Shifts in selective visual attention: towards the underlying neural circuitry. Hum Neurobiol. 1985;4(4):219–27.PubMed Koch C, Ullman S. Shifts in selective visual attention: towards the underlying neural circuitry. Hum Neurobiol. 1985;4(4):219–27.PubMed
27.
go back to reference Navalpakkam V, Itti L. An integrated model of top–down and bottom–up attention for optimal object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2006. p. 2049–56. Navalpakkam V, Itti L. An integrated model of top–down and bottom–up attention for optimal object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2006. p. 2049–56.
28.
go back to reference Aziz MZ, Mertsching B. Visual search in static and dynamic scenes using fine-grain top–down visual attention. In: Proceedings of the 6th international conference on computer vision systems. Springer, Berlin; 2008. p. 3–12. Aziz MZ, Mertsching B. Visual search in static and dynamic scenes using fine-grain top–down visual attention. In: Proceedings of the 6th international conference on computer vision systems. Springer, Berlin; 2008. p. 3–12.
29.
go back to reference Tünnermann J, Born C, Mertsching B. Top–down visual attention with complex templates. In: Proceedings of the international conference on computer vision theory and applications, vol. 1; 2013. p. 370–377. Tünnermann J, Born C, Mertsching B. Top–down visual attention with complex templates. In: Proceedings of the international conference on computer vision theory and applications, vol. 1; 2013. p. 370–377.
30.
go back to reference Borji A, Itti L. State-of-the-art in visual attention modeling. IEEE Trans Pattern Anal Mach Intell. 2013;35(1):185–207.CrossRefPubMed Borji A, Itti L. State-of-the-art in visual attention modeling. IEEE Trans Pattern Anal Mach Intell. 2013;35(1):185–207.CrossRefPubMed
31.
go back to reference Pugeault N, Wörgötter F, Krüger N. Visual primitives: local, condensed, and semantically rich visual descriptors and their applications in robotics. Int J Humanoid Robot (Special Issue on Cognitive Humanoid Vision). 2010;7(3):379–405. Pugeault N, Wörgötter F, Krüger N. Visual primitives: local, condensed, and semantically rich visual descriptors and their applications in robotics. Int J Humanoid Robot (Special Issue on Cognitive Humanoid Vision). 2010;7(3):379–405.
32.
go back to reference Aziz MZ, Mertsching B. Fast and robust generation of feature maps for region-based visual attention. IEEE Trans Image Process. 2008;17(5):633–44.CrossRefPubMed Aziz MZ, Mertsching B. Fast and robust generation of feature maps for region-based visual attention. IEEE Trans Image Process. 2008;17(5):633–44.CrossRefPubMed
33.
go back to reference Tünnermann J, Mertsching B. Region-based artificial visual attention in space and time. Cogn Comput. 2014;6(1):125–43.CrossRef Tünnermann J, Mertsching B. Region-based artificial visual attention in space and time. Cogn Comput. 2014;6(1):125–43.CrossRef
34.
go back to reference Aziz MZ, Shafik MS, Mertsching B, Munir A. Color segmentation for visual attention of mobile robots. In: Proceedings of the IEEE symposium on emerging technologies; 2005. p. 115–120. Aziz MZ, Shafik MS, Mertsching B, Munir A. Color segmentation for visual attention of mobile robots. In: Proceedings of the IEEE symposium on emerging technologies; 2005. p. 115–120.
35.
go back to reference Backer M, Tünnermann J, Mertsching B. Parallel k-means image segmentation using sort, scan and connected components on a GPU. In: Keller R, Kramer D, Weiss JP, editors. Facing the multicore-challenge III. vol. 7686 of lecture notes in computer science; 2013. p. 108–120. Backer M, Tünnermann J, Mertsching B. Parallel k-means image segmentation using sort, scan and connected components on a GPU. In: Keller R, Kramer D, Weiss JP, editors. Facing the multicore-challenge III. vol. 7686 of lecture notes in computer science; 2013. p. 108–120.
36.
go back to reference Aziz MZ, Mertsching B. Pop-out and IOR in static scenes with region based visual attention. In: Proceedings of the ICVS workshop on computational attention & applications. Bielefeld: Bielefeld University eCollections; 2007. doi:10.2390/biecoll-icvs2007-157. Aziz MZ, Mertsching B. Pop-out and IOR in static scenes with region based visual attention. In: Proceedings of the ICVS workshop on computational attention & applications. Bielefeld: Bielefeld University eCollections; 2007. doi:10.​2390/​biecoll-icvs2007-157.
37.
go back to reference Tünnermann J, Mertsching B. Continuous region-based processing of spatiotemporal saliency. In: Proceedings of the international conference on computer vision theory and applications, vol. 1; 2012. p. 230–239. Tünnermann J, Mertsching B. Continuous region-based processing of spatiotemporal saliency. In: Proceedings of the international conference on computer vision theory and applications, vol. 1; 2012. p. 230–239.
38.
go back to reference Krüger N, Janssen P, Kalkan S, Lappe M, Leonardis A, Piater J, et al. Deep hierarchies in the primate visual cortex: what can we learn for computer vision? IEEE Trans Pattern Anal Mach Intell. 2013;35(8):1847–71.CrossRefPubMed Krüger N, Janssen P, Kalkan S, Lappe M, Leonardis A, Piater J, et al. Deep hierarchies in the primate visual cortex: what can we learn for computer vision? IEEE Trans Pattern Anal Mach Intell. 2013;35(8):1847–71.CrossRefPubMed
39.
go back to reference Kootstra G, Popovic M, Jørgensen J, Kuklinski K, Miatliuk K, Kragic D, et al. Enabling grasping of unknown objects through a synergistic use of edge and surface information. Int J Robot Res. 2012;31(10):1190–213.CrossRef Kootstra G, Popovic M, Jørgensen J, Kuklinski K, Miatliuk K, Kragic D, et al. Enabling grasping of unknown objects through a synergistic use of edge and surface information. Int J Robot Res. 2012;31(10):1190–213.CrossRef
40.
go back to reference Olesen SM, Lyder S, Kraft D, Krüger N, Jessen JB. Real-time extraction of surface patches with associated uncertainties by means of Kinect cameras. J Real-Time Image Process. 2015;10(1):105–18.CrossRef Olesen SM, Lyder S, Kraft D, Krüger N, Jessen JB. Real-time extraction of surface patches with associated uncertainties by means of Kinect cameras. J Real-Time Image Process. 2015;10(1):105–18.CrossRef
41.
go back to reference Marat S, Rahman A, Pellerin D, Guyader N, Houzet D. Improving visual saliency by adding ‘face feature map’ and ‘center bias’. Cogn Comput. 2013;5(1):63–75.CrossRef Marat S, Rahman A, Pellerin D, Guyader N, Houzet D. Improving visual saliency by adding ‘face feature map’ and ‘center bias’. Cogn Comput. 2013;5(1):63–75.CrossRef
42.
go back to reference Schauerte B, Stiefelhagen R. How the distribution of salient objects in images influences salient object detection. In: Proceedings of the 20th international conference on image processing (ICIP); 2013. p. 1–5. Schauerte B, Stiefelhagen R. How the distribution of salient objects in images influences salient object detection. In: Proceedings of the 20th international conference on image processing (ICIP); 2013. p. 1–5.
43.
go back to reference Schauerte B, Stiefelhagen R. On the distribution of salient objects in web images and its influence on salient object detection. arXiv preprint arXiv:1501.03383; 2015. Schauerte B, Stiefelhagen R. On the distribution of salient objects in web images and its influence on salient object detection. arXiv preprint arXiv:​1501.​03383; 2015.
44.
go back to reference Born C, Tünnermann J, Mertsching B. Saliency from growing neural gas: learning pre-attentional structures for a flexible attention system. in review. Born C, Tünnermann J, Mertsching B. Saliency from growing neural gas: learning pre-attentional structures for a flexible attention system. in review.
45.
go back to reference Tünnermann J, Born C, Mertsching B. Integrating object affordances with artificial visual attention. In: Agapito L, Bronstein MM, Rother C, editors. Computer vision-ECCV 2014 Workshops. Switzerland: Springer; 2015. p. 427–437. Tünnermann J, Born C, Mertsching B. Integrating object affordances with artificial visual attention. In: Agapito L, Bronstein MM, Rother C, editors. Computer vision-ECCV 2014 Workshops. Switzerland: Springer; 2015. p. 427–437.
46.
go back to reference Ball F, Elzemann A, Busch NA. The scene and the unseen: Manipulating photographs for experiments on change blindness and scene memory. Behav Res Methods. 2014;46(3):689–701.CrossRefPubMed Ball F, Elzemann A, Busch NA. The scene and the unseen: Manipulating photographs for experiments on change blindness and scene memory. Behav Res Methods. 2014;46(3):689–701.CrossRefPubMed
47.
go back to reference Ma L, Xu K, Wong T, Jiang B, Hu S. Change blindness images. IEEE Trans Vis Comput Graph. 2013;19(11):1808–19.CrossRefPubMed Ma L, Xu K, Wong T, Jiang B, Hu S. Change blindness images. IEEE Trans Vis Comput Graph. 2013;19(11):1808–19.CrossRefPubMed
Metadata
Title
Affordance Estimation Enhances Artificial Visual Attention: Evidence from a Change-Blindness Study
Authors
Jan Tünnermann
Norbert Krüger
Bärbel Mertsching
Wail Mustafa
Publication date
01-10-2015
Publisher
Springer US
Published in
Cognitive Computation / Issue 5/2015
Print ISSN: 1866-9956
Electronic ISSN: 1866-9964
DOI
https://doi.org/10.1007/s12559-015-9329-9

Other articles of this Issue 5/2015

Cognitive Computation 5/2015 Go to the issue

Premium Partner