Skip to main content
Erschienen in: International Journal of Computer Vision 3/2014

01.05.2014

Visual Saliency with Statistical Priors

verfasst von: Jia Li, Yonghong Tian, Tiejun Huang

Erschienen in: International Journal of Computer Vision | Ausgabe 3/2014

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Visual saliency is a useful cue to locate the conspicuous image content. To estimate saliency, many approaches have been proposed to detect the unique or rare visual stimuli. However, such bottom-up solutions are often insufficient since the prior knowledge, which often indicates a biased selectivity on the input stimuli, is not taken into account. To solve this problem, this paper presents a novel approach to estimate image saliency by learning the prior knowledge. In our approach, the influences of the visual stimuli and the prior knowledge are jointly incorporated into a Bayesian framework. In this framework, the bottom-up saliency is calculated to pop-out the visual subsets that are probably salient, while the prior knowledge is used to recover the wrongly suppressed targets and inhibit the improperly popped-out distractors. Compared with existing approaches, the prior knowledge used in our approach, including the foreground prior and the correlation prior, is statistically learned from 9.6 million images in an unsupervised manner. Experimental results on two public benchmarks show that such statistical priors are effective to modulate the bottom-up saliency to achieve impressive improvements when compared with 10 state-of-the-art methods.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Fußnoten
1
The “winner-take-all” competition is not used in CS.
 
2
The face detection component is not activated and here we can treat CA as a bottom-up approach.
 
Literatur
Zurück zum Zitat Achanta, R., Hemami, S., Estrada, F., & Susstrunk, S. (2009). Frequency-tuned salient region detection. In: IEEE Conference on Computer Vision and Pattern Recognition. Achanta, R., Hemami, S., Estrada, F., & Susstrunk, S. (2009). Frequency-tuned salient region detection. In: IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Aziz, M. Z., & Mertsching, B. (2008). Fast and robust generation of feature maps for region-based visual attention. IEEE Transactions on Image Processing, 17(5), 633–644.CrossRefMathSciNet Aziz, M. Z., & Mertsching, B. (2008). Fast and robust generation of feature maps for region-based visual attention. IEEE Transactions on Image Processing, 17(5), 633–644.CrossRefMathSciNet
Zurück zum Zitat Borji, A., & Itti, L. (2012). Exploiting local and global patch rarities for saliency detection (pp. 478–485). In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. Borji, A., & Itti, L. (2012). Exploiting local and global patch rarities for saliency detection (pp. 478–485). In: Proc. IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Bruce, N. D., & Tsotsos, T. K. (2006). Saliency based on information maximization. Advances in Neural Information Processing Systems, 18, 155–162. Bruce, N. D., & Tsotsos, T. K. (2006). Saliency based on information maximization. Advances in Neural Information Processing Systems, 18, 155–162.
Zurück zum Zitat Cerf, M., Harel, J., Einhauser, W., & Koch, C. (2008). Predicting human gaze using low-level saliency combined with face detection (pp. 241–248). In: Advances in Neural Information Processing Systems. Cerf, M., Harel, J., Einhauser, W., & Koch, C. (2008). Predicting human gaze using low-level saliency combined with face detection (pp. 241–248). In: Advances in Neural Information Processing Systems.
Zurück zum Zitat Cheng, M. M., Zhang, G. X., Mitra, N. J., Huang, X., & Hu, S. M. (2011). Global contrast based salient region detection. In: IEEE Conference on Computer Vision and Pattern Recognition. Cheng, M. M., Zhang, G. X., Mitra, N. J., Huang, X., & Hu, S. M. (2011). Global contrast based salient region detection. In: IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Chikkerur, S., Serre, T., Tan, C., & Poggio, T. (2010). What and where: A bayesian inference theory of attention. Vision Research, 50(22), 2233–2247.CrossRef Chikkerur, S., Serre, T., Tan, C., & Poggio, T. (2010). What and where: A bayesian inference theory of attention. Vision Research, 50(22), 2233–2247.CrossRef
Zurück zum Zitat Elazary, L., & Itti, L. (2008). Interesting objects are visually salient. Journal of Vision, 8(3), 1–15.CrossRef Elazary, L., & Itti, L. (2008). Interesting objects are visually salient. Journal of Vision, 8(3), 1–15.CrossRef
Zurück zum Zitat Frith, C. (2005). The top in top-down attention. In L. Itti, G. Rees, & J. K. Tsotsos (Eds.), Neurobiology of attention (pp. 105–108). San Diego, CA: Elsevier.CrossRef Frith, C. (2005). The top in top-down attention. In L. Itti, G. Rees, & J. K. Tsotsos (Eds.), Neurobiology of attention (pp. 105–108). San Diego, CA: Elsevier.CrossRef
Zurück zum Zitat Goferman, S., Zelnik-Manor, L., & Tal, A. (2010). Context-aware saliency detection. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. Goferman, S., Zelnik-Manor, L., & Tal, A. (2010). Context-aware saliency detection. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Han, F., & Zhu, S. C. (2009). Bottom-up/top-down image parsing with attribute grammar. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(1), 59–73.CrossRefMathSciNet Han, F., & Zhu, S. C. (2009). Bottom-up/top-down image parsing with attribute grammar. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(1), 59–73.CrossRefMathSciNet
Zurück zum Zitat Harel, J., Koch, C., & Perona, P. (2006). Graph-based visual saliency. Advances in Neural Information Processing Systems, 19, 545–552. Harel, J., Koch, C., & Perona, P. (2006). Graph-based visual saliency. Advances in Neural Information Processing Systems, 19, 545–552.
Zurück zum Zitat Henderson, J. M. (2003). Human gaze control during real-world scene perception. Trends in Cognitive Sciences, 7(11), 498–504.CrossRef Henderson, J. M. (2003). Human gaze control during real-world scene perception. Trends in Cognitive Sciences, 7(11), 498–504.CrossRef
Zurück zum Zitat Hou, X., & Zhang, L. (2007). Saliency detection: A spectral residual approach. In: IEEE Conference on Computer Vision and Pattern Recognition. Hou, X., & Zhang, L. (2007). Saliency detection: A spectral residual approach. In: IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Hou, X., & Zhang, L. (2008). Dynamic visual attention: Searching for coding length increments. Advances in Neural Information Processing Systems, 21, 681–688. Hou, X., & Zhang, L. (2008). Dynamic visual attention: Searching for coding length increments. Advances in Neural Information Processing Systems, 21, 681–688.
Zurück zum Zitat Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(11), 1254–1259.CrossRef Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(11), 1254–1259.CrossRef
Zurück zum Zitat Judd, T., Ehinger, K., Durand, F., & Torralba, A. (2009). Learning to predict where humans look. In: IEEE International Conference on Computer Vision. Judd, T., Ehinger, K., Durand, F., & Torralba, A. (2009). Learning to predict where humans look. In: IEEE International Conference on Computer Vision.
Zurück zum Zitat Kienzle, W., Wichmann F.A., Scholkopf B., & Franz M.O. (2007). A nonparametric approach to bottom-up visual saliency (pp. 689–696). In: Advances in Neural Information Processing Systems. Kienzle, W., Wichmann F.A., Scholkopf B., & Franz M.O. (2007). A nonparametric approach to bottom-up visual saliency (pp. 689–696). In: Advances in Neural Information Processing Systems.
Zurück zum Zitat Li, J., Tian, Y., Huang, T., & Gao, W. (2010). Probabilistic multi-task learning for visual saliency estimation in video. International Journal of Computer Vision, 90(2), 150–165.CrossRef Li, J., Tian, Y., Huang, T., & Gao, W. (2010). Probabilistic multi-task learning for visual saliency estimation in video. International Journal of Computer Vision, 90(2), 150–165.CrossRef
Zurück zum Zitat Li, J., Levine, M. D., An, X., Xu, X., & He, H. (2013). Visual saliency based on scale-space analysis in the frequency domain. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(4), 996–1010.CrossRef Li, J., Levine, M. D., An, X., Xu, X., & He, H. (2013). Visual saliency based on scale-space analysis in the frequency domain. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(4), 996–1010.CrossRef
Zurück zum Zitat Liu, H., Jiang, S., Huang, Q., Xu, C., & Gao, W. (2007a). Region-based visual attention analysis with its application in image browsing on small displays (pp. 305–308). In: ACM International Conference on Multimedia. Liu, H., Jiang, S., Huang, Q., Xu, C., & Gao, W. (2007a). Region-based visual attention analysis with its application in image browsing on small displays (pp. 305–308). In: ACM International Conference on Multimedia.
Zurück zum Zitat Liu, T., Sun, J., Zheng, N. N., Tang, X., & Shum, H. Y. (2007b). Learning to detect a salient object. In: IEEE Conference on Computer Vision and Pattern Recognition. Liu, T., Sun, J., Zheng, N. N., Tang, X., & Shum, H. Y. (2007b). Learning to detect a salient object. In: IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Lu, Y., Zhang, W., Lu, H., & Xue, X. (2011). Salient object detection using concavity context (pp. 233–240). In: IEEE International Conference on Computer Vision. Lu, Y., Zhang, W., Lu, H., & Xue, X. (2011). Salient object detection using concavity context (pp. 233–240). In: IEEE International Conference on Computer Vision.
Zurück zum Zitat Meur, O. L., Callet, P. L., Barba, D., & Thoreau, D. (2006). A coherent computational approach to model bottom-up visual attention. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(5), 802–817.CrossRef Meur, O. L., Callet, P. L., Barba, D., & Thoreau, D. (2006). A coherent computational approach to model bottom-up visual attention. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(5), 802–817.CrossRef
Zurück zum Zitat Navalpakkam, V., & Itti, L. (2007). Search goal tunes visual features optimally. Neuron, 53, 605–617.CrossRef Navalpakkam, V., & Itti, L. (2007). Search goal tunes visual features optimally. Neuron, 53, 605–617.CrossRef
Zurück zum Zitat Parikh, D., Zitnick, C., & Chen, T. (2008). Determining patch saliency using low-level context. In: European Conference on Computer Vision. Parikh, D., Zitnick, C., & Chen, T. (2008). Determining patch saliency using low-level context. In: European Conference on Computer Vision.
Zurück zum Zitat Peters, R. J., & Itti, L. (2007). Beyond bottom-up: Incorporating task-dependent influences into a computational model of spatial attention. In: IEEE Conference on Computer Vision and Pattern Recognition. Peters, R. J., & Itti, L. (2007). Beyond bottom-up: Incorporating task-dependent influences into a computational model of spatial attention. In: IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Riche, N., Mancas, M., Gosselin, B., & Dutoit, T. (2012). Rare: a new bottom-up saliency model. In: IEEE International Conference on Image Processing. Riche, N., Mancas, M., Gosselin, B., & Dutoit, T. (2012). Rare: a new bottom-up saliency model. In: IEEE International Conference on Image Processing.
Zurück zum Zitat Sun, X., Yao, H., & Ji, R. (2012). What are we looking for: Towards statistical modeling of saccadic eye movements and visual saliency (pp. 1552–1559). In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. Sun, X., Yao, H., & Ji, R. (2012). What are we looking for: Towards statistical modeling of saccadic eye movements and visual saliency (pp. 1552–1559). In: Proc. IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Tatler, B. W., Baddeley, R. J., & Gilchrist, I. D. (2005). Visual correlates of fixation selection: Effects of scale and time. Vision Research, 45, 643–659.CrossRef Tatler, B. W., Baddeley, R. J., & Gilchrist, I. D. (2005). Visual correlates of fixation selection: Effects of scale and time. Vision Research, 45, 643–659.CrossRef
Zurück zum Zitat Torralba, A., Oliva, A., Castelhano, M., & Henderson, J. (2006). Contextual guidance of eye movements and attention in real-world scenes: The role of global features on object search. Psychological Review, 113(4), 766–786.CrossRef Torralba, A., Oliva, A., Castelhano, M., & Henderson, J. (2006). Contextual guidance of eye movements and attention in real-world scenes: The role of global features on object search. Psychological Review, 113(4), 766–786.CrossRef
Zurück zum Zitat Tseng, P. H., Carmi, R., Cameron, I. G. M., Munoz, D. P., & Itti, L. (2009). Quantifying center bias of observers in free viewing of dynamic natural scenes. Journal of Vision, 9(7), 1–16.CrossRef Tseng, P. H., Carmi, R., Cameron, I. G. M., Munoz, D. P., & Itti, L. (2009). Quantifying center bias of observers in free viewing of dynamic natural scenes. Journal of Vision, 9(7), 1–16.CrossRef
Zurück zum Zitat Vikram, T. N., Tscherepanow, M., & Wrede, B. (2012). A saliency map based on sampling an image into random rectangular regions of interest. Pattern Recognition, 45, 3114–3124.CrossRef Vikram, T. N., Tscherepanow, M., & Wrede, B. (2012). A saliency map based on sampling an image into random rectangular regions of interest. Pattern Recognition, 45, 3114–3124.CrossRef
Zurück zum Zitat Walther, D., & Koch, C. (2006). Modeling attention to salient proto-objects. Neural Networks, 19(9), 1395–1407.CrossRefMATH Walther, D., & Koch, C. (2006). Modeling attention to salient proto-objects. Neural Networks, 19(9), 1395–1407.CrossRefMATH
Zurück zum Zitat Wang, W., Wang, Y., Huang, Q., & Gao, W. (2010). Measuring visual saliency by site entropy rate. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. Wang, W., Wang, Y., Huang, Q., & Gao, W. (2010). Measuring visual saliency by site entropy rate. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Wolfe, J. M., Alvarez, G. A., & Horowitz, T. S. (2000). Attention is fast but volition is slow. Nature, 406, 691.CrossRef Wolfe, J. M., Alvarez, G. A., & Horowitz, T. S. (2000). Attention is fast but volition is slow. Nature, 406, 691.CrossRef
Zurück zum Zitat Wolfe, J. M. (2005). Guidance of visual search by preattentive information. In L. Itti, G. Rees, & J. Tsotsos (Eds.), Neurobiology of attention (pp. 101–104). San Diego, CA: Academic Press.CrossRef Wolfe, J. M. (2005). Guidance of visual search by preattentive information. In L. Itti, G. Rees, & J. Tsotsos (Eds.), Neurobiology of attention (pp. 101–104). San Diego, CA: Academic Press.CrossRef
Zurück zum Zitat Wu, T., & Zhu, S. C. (2011). A numerical study of the bottom-up and top-down inference processes in and-or graphs. International Journal of Computer Vision, 93(2), 226–252.CrossRefMATHMathSciNet Wu, T., & Zhu, S. C. (2011). A numerical study of the bottom-up and top-down inference processes in and-or graphs. International Journal of Computer Vision, 93(2), 226–252.CrossRefMATHMathSciNet
Zurück zum Zitat Yang, J., & Yang, M. H. (2012). Top-down visual saliency via joint crf and dictionary learning (pp. 2296–2303). In: Proc. IEEE Conference on Computer Vision and Pattern Recognition. Yang, J., & Yang, M. H. (2012). Top-down visual saliency via joint crf and dictionary learning (pp. 2296–2303). In: Proc. IEEE Conference on Computer Vision and Pattern Recognition.
Zurück zum Zitat Zhang, L., Tong, M. H., Marks, T. K., Shan, H., & Cottrell, G. W. (2008). Sun: A bayesian framework for saliency using natural statistics. Journal of Vision, 8(7), 1–20.CrossRef Zhang, L., Tong, M. H., Marks, T. K., Shan, H., & Cottrell, G. W. (2008). Sun: A bayesian framework for saliency using natural statistics. Journal of Vision, 8(7), 1–20.CrossRef
Zurück zum Zitat Zhao, Q., & Koch, C. (2011). Learning a saliency map using fixated locations in natural scenes. Journal of Vision, 11(3), 1–15.CrossRef Zhao, Q., & Koch, C. (2011). Learning a saliency map using fixated locations in natural scenes. Journal of Vision, 11(3), 1–15.CrossRef
Zurück zum Zitat Zhao, Q., & Koch, C. (2012). Learning visual saliency by combining feature maps in a nonlinear manner using adaboost. Journal of Vision, 12(6), 1–15.CrossRef Zhao, Q., & Koch, C. (2012). Learning visual saliency by combining feature maps in a nonlinear manner using adaboost. Journal of Vision, 12(6), 1–15.CrossRef
Metadaten
Titel
Visual Saliency with Statistical Priors
verfasst von
Jia Li
Yonghong Tian
Tiejun Huang
Publikationsdatum
01.05.2014
Verlag
Springer US
Erschienen in
International Journal of Computer Vision / Ausgabe 3/2014
Print ISSN: 0920-5691
Elektronische ISSN: 1573-1405
DOI
https://doi.org/10.1007/s11263-013-0678-0

Weitere Artikel der Ausgabe 3/2014

International Journal of Computer Vision 3/2014 Zur Ausgabe

Premium Partner