Elsevier

Neurocomputing

Volume 152, 25 March 2015, Pages 359-368
Neurocomputing

Saliency detection via background and foreground seed selection

https://doi.org/10.1016/j.neucom.2014.10.056Get rights and content

Abstract

In this paper, we propose a bottom-up visual saliency detection algorithm. Different from most previous methods that mainly concentrate on image object, we take both background and foreground into consideration. First, we collect background seeds from image border superpixels by boundary information and calculate a background-based saliency map. Second, we select foreground seeds by segmenting the first-stage saliency map via adaptive threshold and compute a foreground-based saliency map. Third, the two saliency maps are integrated by the proposed unified function. Finally, we refine the integrated result to obtain a more smooth and accurate saliency map. Moreover, the unified formula also proves to be effective in combining the proposed approach with other models. Experiments on publicly available data sets demonstrate that the proposed algorithm performs favorably against the state-of-the-art methods.

Introduction

Image saliency detection is an important preprocessing step in computer vision to reduce computational load by focusing on salient regions, which has been widely used in numerous applications including object detection [1] and recognition [2], image segmentation [3] and compression [4], video summarization [5], and so on. Saliency detection aims to estimate the probability of an image region appearing as part of foreground. Previous saliency detection approaches can be categorized as two main kinds: bottom-up and top-down. Top-down measures [6], [7], [8] learn task-driven models through training process, which requires specific prior knowledge. Bottom-up methods [9], [10], [11], [12] also known as stimuli-driven models, mainly detect saliency by global or local contrast without any prior knowledge.

Recent years have witnessed significant progress in visual saliency detection. Itti et al. [9] present a saliency map by combining three feature maps including color, intensity and orientation at different scales. Liu et al. [13] integrate three saliency detection results including multi-scale contrast, regional and global distribution by the weights derived from Conditional Random Field (CRF) learning. Achanta et al. [14] measure saliency by the dissimilarity of each pixel׳s color with the average color over the entire scene, which is simple and efficient but suffers a great difficulty in handling cluttered background. Their saliency maps have low resolution and tend to highlight object edge rather than interior. Taking context information into consideration, Goferman et al. [15] implement local contrast by computing the dissimilarity of a region only with its relevant context, defined as a set of appearance-based similar regions. The global region contrast based method [10] utilizes spatially weighted color contrast to measure saliency, which uniformly highlights the interior object but is sensitive to object size. Saliency estimation is defined as a Bayesian inference model in [11], [12]. Different from [11], Xie et al. [12] propose a coarse-to-fine strategy to roughly locate the object region by a convex hull and compute a prior map instead of the constant prior in [11]. Previous saliency detection models mainly focus on image object which is difficult to be dealt with due to its wide variation. However, prior knowledge based on background information could be more accurate and effective as reported by [16], [17], which measure saliency based on background priors.

In this paper, we propose a novel saliency detection algorithm based on both background and foreground priors. First, we obtain a border set by collecting the image border superpixels, which have been proved to be good visual cues for background priors in saliency detection [16], [17]. In addition, we remove the superpixels with strong image edges out of the border set to reduce the foreground noise (e.g., when objects appear at the image border) and thus obtain the background superpixel seeds. We calculate the color and spatial distances between each superpixel and the background seeds to obtain the background-based saliency map. Second, the foreground-based saliency detection is implemented based on the first-stage saliency map. We segment the map using an adaptive threshold and select the superpixels which are more likely to be the object as the foreground seeds. Then the foreground-based saliency map is computed by the color and spatial similarity with the foreground seeds, similar to the background-based saliency detection. Third, we propose a unified formula to integrate the two saliency maps: the background-based one which could highlight the objects and the foreground-based one which could suppress the background noises, therefore the unified map could benefit from both the two maps. Refined by saliency diffusion and Gaussian falloff, the final result could be more accurate. The main steps of the proposed method are shown in Fig. 1.

The main contributions of this work are three-fold:

  • 1.

    By the use of image edge information, we remove the foreground noise in the image borders to obtain more stable and reliable background prior knowledge.

  • 2.

    We devise two saliency detection models by computing the similarity of each image region with the selected background and foreground seeds, respectively.

  • 3.

    The proposed unification mechanism achieves more favorable results and is demonstrated to be effective for saliency detection by further experiments on other algorithms.

The rest of this paper is organized as follows. In Section 2, we describe the details of our algorithm, including the background and foreground based saliency maps and saliency unification and refinement. Experimental results of the proposed algorithm and comparisons with several previous methods are presented in Section 3. Section 4 concludes this paper.

Section snippets

Our approach

In this section, we present the details of the proposed saliency measure. To better capture the structural information of an input image, we use superpixels as the minimum processing units, which are generated by the simple linear iterative clustering (SLIC) algorithm [18].

Experiments

In this section, we evaluate our algorithm on three public available data sets: MSRA, ASD and SED. The MSRA data set [13] contains 5000 images labeled with bounding boxes by nine subjects. The ASD data set [14] contains 1000 images selected from the MSRA data set, where the salient objects are manually labeled with pixel-wise ground truth. The SED data set [22] consists of two subsets: SED1 (where each image contains a single object) and SED2 (where each image contains two objects). Each subset

Conclusion

In this paper, we propose a saliency detection algorithm via background and foreground seed selection. We firstly extract the image border superpixels and remove the foreground ones out of the border set by using boundary information, thus obtain the background seeds. Then we estimate the background-based saliency map via the feature dissimilarity with the selected background seeds. By thresholding the obtained saliency map, we then collect a set of foreground seeds, which are utilized to

Acknowledgments

This work was supported by the Fundamental Research Funds for the Central Universities under Grant No. DUT14YQ101, the Natural Science Foundation of China (NSFC) under Grant No. 61472060, the joint Foundation of China Education Ministry, and by the China Mobile Communication Corporation under Grant No. MCM20122071.

Jianpeng Wang received the B.E. degree in electronic information engineering from Dalian University of Technology, Dalian, China, in 2013. He is currently a master student majoring in signal and information processing in Dalian University of Technology, Dalian, China. His current research interests include saliency detection and running detection.

References (34)

  • A. Torralba, P. Sinha, Statistical context priming for object detection, in: Proceedings of the IEEE International...
  • U. Rutishauser, D. Walther, C. Koch, P. Perona, Is bottom-up attention useful for object recognition?, in: Computer...
  • V. Lempitsky, P. Kohli, C. Rother, T. Sharp, Image segmentation with a bounding box prior, in: IEEE 12th International...
  • C. Guo et al.

    A novel multiresolution spatiotemporal saliency detection model and its applications in image and video compression

    IEEE Trans. Image Process.

    (2010)
  • Y.-F. Ma, L. Lu, H.-J. Zhang, M. Li, A user attention model for video summarization, in: Proceedings of the Tenth ACM...
  • L. Marchesotti, C. Cifarelli, G. Csurka, A framework for visual saliency detection with applications to image...
  • J. Yang, M.-H. Yang, Top-down visual saliency via joint CRF and dictionary learning, in: Proceedings of IEEE Conference...
  • C. Kanan et al.

    Sun: top-down saliency using natural statistics

    Vis. Cognit.

    (2009)
  • L. Itti et al.

    A model of saliency-based visual attention for rapid scene analysis

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1998)
  • M.-M. Cheng, G.-X. Zhang, N.J. Mitra, X. Huang, S.-M. Hu, Global contrast based salient region detection, in:...
  • E. Rahtu, J. Kannala, M. Salo, J. Heikkilä, Segmenting salient objects from images and videos,...
  • Y. Xie, H. Lu, Visual saliency detection based on Bayesian model, in: Proceedings of IEEE International Conference on...
  • T. Liu et al.

    Learning to detect a salient object

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2011)
  • R. Achanta, S.S. Hemami, F.J. Estrada, S. Süsstrunk, Frequency-tuned salient region detection, in: Proceedings of IEEE...
  • S. Goferman et al.

    Context-aware saliency detection

    Pattern Anal. Mach. Intell.

    (2012)
  • Y. Wei, F. Wen, W. Zhu, J. Sun, Geodesic saliency using background priors, in: Proceedings of European Conference on...
  • X. Li, H. Lu, L. Zhang, X. Ruan, M.-H. Yang, Saliency detection via dense and sparse reconstruction, in: Proceedings of...
  • Cited by (0)

    Jianpeng Wang received the B.E. degree in electronic information engineering from Dalian University of Technology, Dalian, China, in 2013. He is currently a master student majoring in signal and information processing in Dalian University of Technology, Dalian, China. His current research interests include saliency detection and running detection.

    Huchuan Lu (M׳06) received the M.Sc. degree in signal and information processing and the Ph.D. degree in system engineering from the Dalian University of Technology (DUT), Dalian, China, in 1998 and 2008 respectively. He joined as a Faculty Member with the School of Information and Communication Engineering, DUT, in 1998, where he has been a Professor since 2012. His current research interests include computer vision, pattern recognition, visual tracking, and segmentation. Prof. Lu has been a member of the ACM since 2010. He is an Associate Editor of the IEEE Transaction on Systems, Man and Cybernetics: Part B.

    Xiaohui Li received the B.E. degree in electronic information engineering from Dalian University of Technology, Dalian, China, and the M.S. degree in signal and information processing in Dalian University of Technology, Dalian, China, in 2011 and 2014, respectively. Her research interest is in saliency detection.

    Na Tong received her B.E. degree in English Intensive, Electronic and Information Engineering, Dalian University of Technology (DUT), in 2012. She is currently a master student in the School of Information and Communication Engineering, Dalian University of Technology (DUT). Her research interest is in saliency detection and deep learning.

    Wei Liu received double bachelor degrees in electronic information engineering and English from Dalian University of Technology in 2013. She is now a first year master student in IIAU lab leaded by Prof. Huchuan Lu and her research field is image saliency detection.

    View full text