Research

SalientShape: Group Saliency in Image Collections

Ming-Ming Cheng1,4   Niloy J. Mitra2  Xiaolei Huang3  Shi-Min Hu1

1TNList, Tsinghua University, Beijing     2UCL     3Lehigh University     4Oxford Brookes University

Image

Figure. Our system explicitly extracts salient object regions from a group of related images with heteroge-neous quality offline (a-d, f) to enable efficient online (e) shape based query.

Abstract

Efficiently identifying salient objects in large image collections is essential for many applications including image retrieval, surveillance, image annotation, and object recognition. We propose a simple, fast, and effective algorithm for locating and segmenting salient objects by analysing image collections. As a key novelty, we introduce group saliency to achieve superior unsupervised salient object segmentation by extracting salient objects (in collections of pre-filtered images) that maximize between-image similarities and within-image distinctness. To evaluate our method, we construct a large benchmark dataset consisting of 15K images across multiple categories with 6000+ pixel-accurate ground truth annotations for salient object regions where applicable. In all our tests, group saliency consistently outperforms state-of-the-art single-image saliency algorithms, resulting in both higher precision and better recall. Our algorithm successfully handles image collections, of an order larger than any existing benchmark datasets, consisting of diverse and heterogeneous images from various inter-net sources.

Paper

SalientShape: Group Saliency in Image Collections. Ming-Ming Cheng, Niloy J. Mitra, Xiaolei Huang, Shi-Min Hu. The Visual Computer, 2013. [pdf] [supplementary] [bib]

Data

We introduce a labeled dataset of categorized images for evaluating sketch based image retrieval. Using Flickr, we downloaded about 3000 images for each of the 5 keywords: “butterfly”, “coffee mug”, “dog jump”, “giraffe”, and “plane”, together comprising of about 15000 images. For each image, if there is a non-ambiguous object with correct content matching with the query keyword and most part of the object is visible, we mark such an object region. Similar to MSRA10K, the salient regions are marked at a pixel level. We only label salient object region for objects with almost fully visible since partially occluded objects are is less useful for shape matching. Different from MSRA10K, the THUR15K dataset do not contain a salient region labeled for every image in the dataset, i.e., some images may not have any salient region. This dataset is used to evaluate shape based image retrieval performance.

Please read the notice first to see how to automatically get the password for unzip.

Comparisons with state of the art methods

Image

Figure. Evaluation results on our benchmark dataset. (a) Precision-recall curves for naive thresholding of saliency maps. S, G1, G2 represent single image saliency, group saliency after the 1st and 2nd iterations, respectively. Subscripts B, C, D, G, P represent groups of ‘butterfly’, ‘coffee mug’, ‘dog jump’, ‘giraffe’, ‘plane’, respectively. (b) Comparison of F-Measure for image groups using single image saliency segmentation methods (FT [1], SEG [37], RC [14]) vs. group saliency (GS) segmentation.

Image

Figure. SBIR comparison. In each group from left to right, first column shows images downloaded from Flickr using the corresponding keyword; second column shows our retrieval results obtained by comparing user-input sketch with group saliency segmentation results; third column shows corresponding sketch based retrieval results using SHoG [20]. Two input sketches with their retrieval results are shown in (e).

Learned appearance from Flickr images (without time consuming image annotation)

GroupSal

Links:

  1. Global Contrast based Salient Region Detection. Ming-Ming Cheng, Guo-Xin Zhang, Niloy J. Mitra, Xiaolei Huang, Shi-Min Hu. IEEE CVPR, 2011, p. 409-416.
  2. Unsupervised Joint Object Discovery and Segmentation in Internet Images, Michael Rubinstein, Armand Joulin, Johannes Kopf, Ce Liu, IEEE CVPR 2013.
  3. Unsupervised joint object discovery and segmentation in internet images, M. Rubinstein, A. Joulin, J. Kopf, and C. Liu, in IEEE CVPR, 2013, pp. 1939–1946. (Used the proposed saliency measure and showed that saliency-based segmentation produces state-of-the-art results on co-segmentation benchmarks, without using co-segmentation!)
  4. Unsupervised Object Discovery via Saliency-Guided Multiple Class Learning, Jun-Yan Zhu, Jiajun Wu, Yichen Wei, Eric Chang, and Zhuowen Tu, IEEE CVPR, 2012.
Locations of visitors to this page Locations of visitors to this page

9 thoughts on “SalientShape: Group Saliency in Image Collections

  • 请问THUR15K的全称是什么呢?

    Reply
  • 程老师您好,请问这篇文章有提供代码吗? 我在cmCoder里没有找到呢

    Reply
    • 都快十年了(2011年写的代码)。在电脑上找了半天,找到了这个名字的代码。打了个包放在了这里:http://mftp.mmcheng.net/Papers/GroupSalCode.zip 你看看吧,我也不确信是不是最终版。

      Reply
      • 谢谢程老师~ 我下载看看

        Reply
  • After refinement we left with few top quality images and how can we evaluate Prior valu
    Is id done by any leniar operation on GMM of each image or anything else ??
    Thanks in advance

    Reply
  • 老师好,最近看了您的这篇文章,我想问您一下Precision-recall curves是怎么评价算法优劣性的?

    Reply
  • Pingback: [14CVPR] BING: Binarized Normed Gradients for Objectness Estimation at 300fps | 增强视觉 | 计算机视觉 增强现实

Leave a Reply

Your email address will not be published. Required fields are marked *