Open Access
9 September 2017 Image segmentation via foreground and background semantic descriptors
Ding Yuan, Jingjing Qiang, Jihao Yin
Author Affiliations +
Abstract
In the field of image processing, it has been a challenging task to obtain a complete foreground that is not uniform in color or texture. Unlike other methods, which segment the image by only using low-level features, we present a segmentation framework, in which high-level visual features, such as semantic information, are used. First, the initial semantic labels were obtained by using the nonparametric method. Then, a subset of the training images, with a similar foreground to the input image, was selected. Consequently, the semantic labels could be further refined according to the subset. Finally, the input image was segmented by integrating the object affinity and refined semantic labels. State-of-the-art performance was achieved in experiments with the challenging MSRC 21 dataset.

1.

Introduction

Image segmentation is a fundamental problem in the field of computer vision. So far, abundant research has been published on this topic;15 however, segmenting the complete foreground objects, which are not uniform in color or texture, remains a challenging task. In addition to the local low-level image features, such as color, texture, and spatial position, an increasing amount of studies focus on segmenting images using high-level visual information.

Cosegmentation methods suggested by Refs. 67.8.9.10.11 employ foreground correspondence and jointly segmented objects, which have similar characteristics in a set of images. Rother et al.7 utilized histogram matching and a modified Markov random filed (MRF) framework formed by the difference of foreground region histograms. Sun et al.8 constructed an MRF framework, which reflected camera flash illumination changes in order to extract the foreground from the background. Kim et al.9 proposed a hierarchical framework for dividing the large image set into multiple subsets in order to perform segmentation by cosegmenting each subset separately with interimage connections. Inspired by the characteristic of linear anisotropic heat diffusion, Kim et al.10 suggested a cosegmentation model, in which the finite heat sources of temperature maximization corresponded to the maximized segmentation confidence. In Ref. 11, segmentation was modeled by an energy-minimization function, which combined local appearance and spatial consistency; however, in most existing studies on cosegmentation, only multiple images with common objects were handled, and different irregularly appearing objects were hardly dealt with.

In recent years, semantic segmentation aiming at assigning a semantic label to each pixel of a given image1218 has become a subject undergoing intense investigation in the field of computer vision. Especially, the techniques of deep neural networks have recently played an important role in the field of semantic segmentation. The segmentation accuracy has been greatly improved by applying the deep learning techniques,1922 on the condition that the huge dataset is collected to train the network.

Semantic information, such as high-level visual information, can provide an important cue for the segmentation of a complete foreground from the image. In this study, inspired by semantic segmentation methods, we propose a segmentation mechanism for achieving a complete and accurate foreground boundary. Inspired by nonparametric methods, the initial semantic labels were obtained by maximizing the normalized label likelihood score.23,24 Then, the foreground and background semantic descriptors were defined according to the initial semantic labels. With the aid of the two semantic descriptors, a subset of training images with similar foreground to the input image, was obtained. Subsequently, the semantic labels were further refined via object affinity and a semantic codebook. Finally, image segmentation was achieved by means of semantic labeling. For postprocessing, we adopted the Grab-Cut method25 and used it to merge separate regions.

The remainder of this paper is organized as follows: Sec. 2 describes initial semantic labeling using the nonparametric method; Sec. 3 describes our image segmentation scheme via foreground and background semantic descriptors; and the experimental results are presented in Sec. 4.

2.

Initial Semantic Labels Acquisition

Inspired by the nonparametric method,23,24,26 the initial semantic labels of the input image can be acquired as follows: in the beginning, an image subset D is obtained from the training set by applying the global GIST feature descriptor, such that the image subset D contains the most scenes similar to the input image.

The GIST descriptor can summarize the gradient information for local regions of an image, which provides a rough description of the scene. A GIST descriptor of the scene refers to the meaningful information that an observer can identify from a glimpse at the scene.27 The GIST can be represented at both perceptual and conceptual levels because it includes all levels of visual information. It can be constructed by two-dimensional Gabor wavelets.28 The Gabor wavelets of specific direction and scale can be considered as a local bandpass filter with respect to the corresponding direction and scale, whose response is exactly corresponding to the edges of specific directions in the image. At the beginning, the image is divided into patches. For each patch Pi of size r×c, the cascading of its convolution in each channel is defined as

Eq. (1)

GiP=cat[F(x,y)*Gm,n(x,y)],(x,y)Pi,
where cat(·) represents the cascade operation, and Gm,n(x,y) denotes the two-dimensional Gabor wavelet.

The average convolution of specific direction and scale for patch Pi is G¯iP=1r×cΣ(x,y)PiGiP(x,y), then the GIST descriptor can be expressed as

Eq. (2)

GG={G¯1P,G¯2P,,G¯NgP},
where Ng represents the number of patches in the image.

By detecting and combining the edge information among local patches, the GIST descriptor can describe the overall distribution of gradient information within the image.

In this study, the initial semantic labels were assigned to each superpixel, instead of individual pixels, due to the spatial supports among pixels. Specifically, the superpixels, within both the input image and the images in D, were obtained using the simple linear iterative clustering (SLIC) method.29 Then, we adopted three features, denoted as fk (k=1,2,3): the scale-invariant feature transform (SIFT) descriptor,30 color mean in Lab color space, and central location of the superpixel, in order to describe each superpixel. Is={s1,s2,,sN} denotes the set of superpixels of the input image, and Ds={n1,n2,,nM} denotes all the superpixels achieved from set D. For each superpixel si{s1,s2,,sN}, its neighborhood Nik was defined as a set of superpixels NikDs, which had the nearest Euclidean distance to si in terms of the k’th feature fik. In this work, Nik included its closest 15 superpixels.

Next, each superpixel si was assigned a semantic label lL, where L represented the set of semantic classes. The probability distribution of semantic labels was defined as the normalized label likelihood score. In this work, the normalized label likelihood score P(fik|l), of each superpixel si, was expressed by nonparametric density estimates

Eq. (3)

P(fik|l)=[n(l,Nik)+ε]/[n(l,D)][n(l¯,Nik)+ε]/[n(l¯,D)],
where l¯ was the complementary set of l. n(l,Nik) [or n(l,D)] and indicated the number of superpixels labeled l in the set Nik (or D), whereas n(l¯,Nik) [or n(l¯,D)] represented the number of superpixels that were not labeled l. The function of ε was to prevent zero likelihoods and smoothen the counts.

Then, the initial semantic label for each superpixel was achieved by maximizing the normalized label likelihood score

Eq. (4)

maxlk1ZP(fik|l),
where Z was a normalization factor under all sematic classes. However, the result of semantic labels was coarse and some superpixels were labeled incorrectly, as shown in Fig. 1.

Fig. 1

Initial semantic labels throughout the image. (a) Superpixels and (b) coarse initial labels.

JEI_26_5_053004_f001.png

3.

Image Segmentation via Semantic Descriptors

With regard to the input image, it was intuitively known that the segmentation would be guided effectively if there existed a subset of training images, which would have a foreground similar to the input image. Therefore, the subset of training data, named semantic retrieval set, had to be determined by utilizing the initial semantic labels prior to complete segmentation.

3.1.

Semantic Retrieval Set Determination via Foreground and Background Semantic Descriptor

In the beginning, the image could be divided into two segments according to its Lab color features by using the k-means method.31 Considering that peripheral regions often appear as background in images, we assigned the peripheral segment, mentioned above, a background label “0”; a foreground label “1” was assigned to the other segment. Then, the foreground semantic descriptor ffs and the background semantic descriptor fbs were defined in order to obtain the semantic retrieval set Ψ, such that the images in the set would have the most similar foreground objects to the input image.

The semantic descriptors were defined in a spatial pyramid structure. The segment labeled “foreground,” in the image, would be divided into equal grids with respect to different levels rs in the spatial pyramid. At each level, the semantic histogram was calculated within each grid. For instance, four semantic histograms h21, h22, h23, and h24 were obtained in the second level of the spatial pyramid, corresponding to the four equal grids, as shown in Fig. 2.

Fig. 2

Definition of semantic descriptors. At each level, the semantic histograms were calculated within each grid.

JEI_26_5_053004_f002.png

Then, the foreground semantic descriptor ffs was defined as the concatenation of all the semantic histograms within each grid at each pyramid level

Eq. (5)

ffs=[h1,h21,h22,h23,h24,,hrs,4rs1].

Similarly, we can also define the background semantic descriptor fbs. Experiments showed that the spatial pyramid layer rs=3 was a good compromise between capturing enough details and avoiding being sensitive to the noise.

In order to obtain the semantic retrieval set, we calculated the global GIST feature gf, foreground semantic descriptor ffs, and background semantic descriptor fbs, throughout the input image and all the images in the training set. Then, the similarity of the input image and the training set was defined as the Euclidean distance between the features

Eq. (6)

d=dg+dfs+λdbs,
where dg, dfs, and dbs were the Euclidean distances, with respect to gf, ffs, and fbs, between the input image and the training set. The purpose of λ was to control the influence of the background semantic labels in the case where some training images had been wrongly selected due to their large background area. The procedure of obtaining the semantic retrieval set Ψ was described in Algorithm 1.

Algorithm 1.

Semantic retrieval set Ψ selecting scheme.

1. Initialize λ(0)=0, and compute d(0)=dg+dfs+λ(0)dbs;
2. Search for a subset Ψ in an ascending order of distance d(0);
3. Set t=0;
4. while none of background label of subset Ψ is the same as the input image’s background && λ(t)0.1do
5. λ(t+1)=λ(t)+0.01
6. ifλ(t+1)0.1then
7.  d(t+1)=dg+dfs+λ(t+1)dbs;
8.  Search a new subset Ψ in an ascending order of distance d(t+1);
9.  t=t+1
10. end if
11. end while
12. returnλ(t), compute d(t)=dg+dfs+λ(t)dbs, and obtain final subset Ψ.

By applying Algorithm 1, the training images were arranged according to the ascending order of d, which corresponded exactly to the similarity of the input image. Finally, the semantic retrieval set Ψ was obtained by selecting the images corresponding to the smallest d. More images contained in set Ψ would provide more clues for labeling the input image, but they would also decrease the similarity to the input image; therefore, in this study, a maximum of four training images was selected for the formation of the semantic retrieval set Ψ.

Figure 3 shows the semantic retrieval set Ψ corresponding to the input “cow” image. Set Ψ also had “cows” appearing in the foreground, which meant that the coarse initial semantic labels were able to provide an effective cue on what semantic categories the foreground belonged to.

Fig. 3

(b) The semantic retrieval set of the (a) input image.

JEI_26_5_053004_f003.png

3.2.

Semantic Labels Assignment via Object Affinity

Once set Ψ was obtained, the semantic labels would be reassigned to each superpixel of the input image via object affinity.

Suppose si is a superpixel of the input image, and smj is a superpixel in the m’th image of Ψ; that is, mΨ. We denoted the distance between si and smj with respect to the k’th feature fik as Δdimjk. Then, the distance measure between si and smj was defined as

Eq. (7)

Δdimj=kωkΔdimjk,
where ω=[ω1,ω2,ω3] was the weight of different features; in this study, k=1, 2, 3 as mentioned in Sec. 2. Subsequently, within image m, the nearest neighborhood Nim of si in set Ψ was obtained via Δdimj.

Obviously, the semantic labels of Nim should have provided an important cue for assigning semantic labels to si, due to their high feature similarity. Moreover, the labels of the superpixels neighboring to si in the input image should have obeyed the smoothness constraint, which reflected the distribution of semantic labels in natural images. The above idea can be expressed as the concept of object affinity.

Thus, the semantic label likelihood of si, determined by the labels of Nim, was described as a Gaussian function

Eq. (8)

Gs(si)=1Kllexp(|Δdimj|2β12)δNim(mj),
where Kl was the number of superpixels sharing the same semantic class l within the neighborhood Nim; β1 was a damping parameter. The indicator function δNim(mj) was defined as
δNim(mj)={1,if  smjεNim0,otherwise.

Considering that the distribution of semantic labels tended to be smooth throughout natural images, we adopted the agglomerative clustering method10 in order to cluster the superpixels with respect to the Lab color feature. The semantic label propagation of the neighboring superpixels was achieved via object affinity

Eq. (9)

Sobject(si)=45·1|C(si)|·pεC(si)Gs(p)+15·Gs(si),
where C(si) was the cluster where si belonged.

3.3.

Initial Semantic Labels Refinement via Semantic Codebook

Although the initial semantic labels were coarse, they still offered a strong cue about the distribution of semantic labels. Now that the semantic retrieval set Ψ could also provide the probability of labels for each superpixel, we compared the similarity between initial semantic labels and semantic labels generated from the semantic retrieval set. Generally, the higher the similarity of the two semantic labels, the higher was the reliability of the semantic labels.

Hence, the initial semantic labels were refined according to a semantic codebook, which was constructed for measuring the similarity between initial semantic labels and the semantic labels in the semantic retrieval set Ψ. The semantic codebook of set Ψ was set as B={Blk|k=1,2,3.lLΨ}, where Blk was defined as the feature descriptor of all the superpixels labeled l in the k’th (k=1, 2, 3) feature channel (mentioned in Sec. 2) for a specific semantic class l; LΨ represented the set of semantic classes in set Ψ. Moreover, for any particular superpixel sj labeled l, its feature fljk (k=1,2,3) formed a codeword in the codebook.

For any superpixel si assigned a label lm initially, if its initial label lm was included in LΨ, such that lmLΨ, the similarity of label lm was calculated only with respect to Blm={Blmk|k=1,2,3} in the semantic codebook. However, if the initial label lmLΨ, the similarity was determined by examining all the codewords in B. Specifically, the similarity for superpixel si, which was initially labeled lm, was defined as

Eq. (10)

ΔHi={kωkΔHikif  lLΨminlkωkΔHilkif  lLΨ,
where ΔHik recorded the smallest distance between feature fik of the superpxiel si and the codeword in Blmk. ΔHilk was the smallest distance between feature fik and a particular codeword in the semantic codebook B.

Finally, for the initial semantic label of the superpxiel si, the semantic probability was refined as

Eq. (11)

Srefine(si)=exp(1|Ri|·|ΔHi|2β22),
where |Ri| was the size of the semantic region of superpixel si in the initial semantic labels; β2 was a damping parameter. In this work, we set the parameters as ω=[ω1,ω2,ω3]=[0.3,1,0.3], β1=1 and β2=0.035.

3.4.

From Semantic Labeling to Segmentation

In this section, we describe image segmentation by maximizing the linear combination of the object affinity and the refined probability of the initial semantic labels

Eq. (12)

Vseg=maxl[Sobject(si)+Srefine(si)].

The segmentation results with the semantic labels of sample images, shown in Fig. 4(b), also provided a bounding box indicating the possible foreground area. Finally, the Grab-Cut method25 was adopted as a postprocessing procedure to portray the boundary of the foreground precisely. The Grab-Cut method25 is a segmentation technique that uses graph cuts to perform segmentation. Before it is performed, a manually rectangular region of interest should be placed to indicate the location of the foreground in the image. The more precisely the rectangle could exactly encircle the object of interest, the more accurate the segmentation result is. If the rectangular region of interest is not perfectly placed, the good segmentation result cannot be obtained, as shown in Fig. 4.

Fig. 4

Segmentation results by using Grab-Cut method25 with respect to different artificial rectangular regions marked with green rectangles. (a) Segmentation results with respect to the precisely placed rectangle and (b) segmentation results with respect to the imperfect rectangle which do not encircle the foreground completely.

JEI_26_5_053004_f004.png

In this work, the semantic information can provide a bounding box indicating the possible foreground area. The Grab-Cut method is performed to further merge the neighboring regions assigned to the same semantic label in the bounding box, and at the same time, to precisely portray the boundary of the foreground via color features in the Lab space. Moreover, after performing the Grab-Cut method, for the areas where the new labels are not consistent with the segmentation result by using Eq. (12), the higher Vseg will enforce these areas to remain their previous labels. In the experiments, we set the threshold as 1.4. Figure 5 shows the experimental results by using our method with Grab-Cut as the postprocessing procedure, and the results by using the Grab-Cut method with respect to a precisely placed artificial rectangular region of interest. Compared to Fig. 5(c), it shows that the our result with Grab-Cut as postprocessing procedure is more crisp and the boundary of the foreground is more precisely portrayed as shown in Fig. 5(d). Also in Fig. 5(d), the green leaves among the red flowers in the middle of the image is correctly labeled as the background, whereas it is wrongly labeled as the foreground by only applying the Grab-Cut method, as shown in Fig. 5(e).

Fig. 5

Segmentation results by using our method and by directly using the Grab-Cut method. (a) Input image, (b) ground truth, (c) segmentation results with semantic labels, (d) our segmentation result with Grab-Cut as a postprocessing procedure, and (e) result by using Grab-Cut with respect to a manually rectangular region of interest marked in green.

JEI_26_5_053004_f005.png

More samples of postprocessing results are shown in Fig. 4(c). The foreground boundaries were eventually determined (Fig. 6).

Fig. 6

Segmentation results. (a) Input images, (b) segmentation results with semantic labels, and (c) postprocessing using Grab-Cut.

JEI_26_5_053004_f006.png

4.

Experimental Results

To verify the effectiveness of the proposed method, we conducted experiments on the MSRC 21 dataset, which contained 21 different classes with 276 training images and 256 testing images. In the experiments using the MSRC 21 dataset, the image subset D was allowed to include a maximum of 25 training images, such that enough scenes similar to the input image could be selected.

The segmentation performance was validated via the intersection-over-union score mentioned in Refs. 10 and 12

Eq. (13)

SIOU=maxl1|I|iIGTiRilGTiRil,
where GTi is the ground truth and Ril represents the region associated with l’th class in image i.

In the experiments, we tested all of the 14 image classes in the MSRC 21 dataset. The intersection-over-union score was adopted in order to evaluate the precision of our algorithm. Moreover, we compared our algorithm to other segmentation algorithms;2,6,1012,18,22 the results are listed in Table 1. The higher intersection-over-union score corresponded to higher segmentation precision. In addition, we also listed the precision of our semantic labeling named as “semantic label,” and the segmentation results by using the initial labels with the Grab-Cut as postprocessing is named as “initial+grab cut” in Table 1. “Ours” represents the results of our final segmentation with the Grab-Cut as postprocessing. The subscript represents the rank of segmentation accuracy by using different methods.

Table 1

Segmentation results evaluation on intersection-over-union score.

ClassOursInitial+Grab-CutSemantic labelRef. 12Ref. 11Ref. 10Ref. 6Ref. 2Ref. 18Ref. 22
Bike40.3542.1436.3739.9643.3229.9842.8313.71027.0959.51
Bird62.6250.9333.5748.3447.7529.98-34.3620.2971.81
Car68.2252.7448.0752.3659.7337.1852.5520.11035.3990.71
Cat57.6358.6241.5552.3431.9724.485.61033.5619.9985.11
Chair61.0153.1322.81054.3239.6528.7739.4624.1924.5849.04
Cow74.6263.9350.6543.2752.7433.5826.1944.8615.71078.51
Dog79.9260.6330.6850.8441.8633.07-43.6522.5984.61
Face55.2358.7242.2645.8570.0133.2840.8748.3428.19-
Flower66.1258.6343.7584.9151.9440.26-26.8839.67-
House63.0244.9538.9748.6451.0332.3866.4128.4944.46-
Plane46.0240.9444.1335.9521.6925.1733.4625.0816.71070.81
Sheep78.3270.8356.0766.3466.3460.8645.7838.01045.1985.31
Sign84.8177.1246.1659.5358.9443.27-42.4858.65
Tree74.1165.7466.8358.1767.0261.2655.9830.4964.65
Average65.1257.0342.9652.9450.2536.6840.9732.4931.61075.11

We also compared our results with the technique of deep neural network.22 In the experiment, we directly evaluated the released pretrained model trained with PASCAL-context dataset on the MSRC dataset and computed the segmentation accuracy of nine overlapping classes between the two datasets. The quantitative results have been listed in Table 1. Admittedly, the average accuracy of our results is lower than that of Ref. 22, which involved a large amount of training samples. However, our results have achieved the best average accuracy among the hand-designed feature-based methods. In addition, the segmentation accuracies of several classes, such as “cow,” “dog,” and “sheep” are comparable to that of Ref. 22. We also achieved a better result on class “chair” compared to Ref. 22. Details on the experimental results are listed in Table 1. Figure 7 shows segmentation results by using our proposed algorithm. We also compared our results with Ref. 22 visually, as shown in Figs. 7(c) and 7(f). For the results by using,22 only the overlapping classes between the MSRC and PASCAL-Context datasets are shown. For example, segmentation results by using22 are not shown from the 1st row to the 4th row in Fig. 7, as the PASCAL-Context dataset does not include the classes of "face", "flower", "sign" and "house", which is also illustrated in Table 1.

Fig. 7

Sample segmentation results. (a) and (d) Input images, (b) and (e) segmentation results by using our proposed method, and (c) and (f) segmentation results by using the technique of deep neural network.22

JEI_26_5_053004_f007.png

In the experiments, there are still some images which are very challenging. Our mechanism of semantic labeling and foreground segmentation depends on the color and SIFT features. Consequently, our method would probably fail, if dealing with the images in which the color or texture distribution of the foreground is similar to the background, as shown in Fig. 8.

Fig. 8

Incorrect segmentation results. (a) and (c) Input images and (b) and (d) segmentation results. The segmentation failed because the color or texture distribution of the foreground is similar to the background.

JEI_26_5_053004_f008.png

5.

Conclusion and Future Work

In this paper, an image segmentation framework based on semantic information was proposed. Unlike traditional methods based on low-level features, we adopted semantic information in order to distinguish the foreground from the background. In our study, the initial semantic labels were obtained using the nonparametric method. By searching for similar images in the training data, the input image was segmented via the combination of object affinity and semantic labels. Experimental testing using the MSRC 21 dataset demonstrated that our method performed well. In future work, segmentation of video data by means of semantic information will be investigated.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (No. 61005031).

References

1. 

D. Comaniciu and P. Meer, “Mean shift: a robust approach toward feature space analysis,” IEEE Trans. Pattern Anal. Mach. Intell., 24 603 –619 (2002). http://dx.doi.org/10.1109/34.1000236 Google Scholar

2. 

P. F. Felzenszwalb and D. P. Huttenlocher, “Efficient graph-based image segmentation,” Int. J. Comput. Vision, 59 (2), 167 –181 (2004). http://dx.doi.org/10.1023/B:VISI.0000022288.19776.77 IJCVEQ 0920-5691 Google Scholar

3. 

Z. Kato and T.-C. Pong, “A Markov random field image segmentation model for color textured images,” Image Vision Comput., 24 (10), 1103 –1114 (2006). http://dx.doi.org/10.1016/j.imavis.2006.03.005 IVCODK 0262-8856 Google Scholar

4. 

S. Ramalingam, P. Kohli and K. Alahari, “Exact inference in mufti-label CRFs with higher order cliques,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 1 –8 (2008). http://dx.doi.org/10.1109/CVPR.2008.4587401 Google Scholar

5. 

Y. Boykov and V. Kolmogorov, “An experimental comparison of min-cut/max-flow algorithms for energy minimization in vision,” IEEE Trans. Pattern Anal. Mach. Intell., 26 (9), 1124 –1137 (2004). http://dx.doi.org/10.1109/TPAMI.2004.60 ITPIDJ 0162-8828 Google Scholar

6. 

L. Mukherjee, V. Singh and J. Peng, “Scale invariant cosegmentation for image groups,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), (2011). http://dx.doi.org/10.1109/CVPR.2011.5995420 Google Scholar

7. 

C. Rother et al., “Cosegmentation of image pairs by histogram matching incorporating a global constraint into MRFs,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 993 –1000 (2006). http://dx.doi.org/10.1109/CVPR.2006.91 Google Scholar

8. 

J. Sun et al., “Flash cut: foreground extraction with flash and no-flash image pairs,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 1 –8 (2007). http://dx.doi.org/10.1109/CVPR.2007.383080 Google Scholar

9. 

E. Kim, H. Li and X. Huang, “A hierarchical image clustering cosegmentation framework,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), (2012). http://dx.doi.org/10.1109/CVPR.2012.6247737 Google Scholar

10. 

G. Kim et al., “Distributed cosegmentation via submodular optimization on anisotropic diffusion,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 169 –176 (2011). http://dx.doi.org/10.1109/ICCV.2011.6126239 Google Scholar

11. 

A. Joulin, F. Bach and J. Ponce, “Multi-class cosegmentation,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 542 –549 (2012). http://dx.doi.org/10.1109/CVPR.2012.6247719 Google Scholar

12. 

Y. Liu et al., “Weakly-supervised dual clustering for image semantic segmentation,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2075 –2082 (2013). http://dx.doi.org/10.1109/CVPR.2013.270 Google Scholar

13. 

L. Tao, F. Porikli and R. Vidal, “Sparse dictionaries for semantic segmentation,” in European Conf. on Computer Vision, 549 –564 (2014). Google Scholar

14. 

J. Yao, S. Fidler and R. Urtasun, “Describing the scene as a whole: joint object detection, scene classification and semantic segmentation,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 702 –709 (2012). http://dx.doi.org/10.1109/CVPR.2012.6247739 Google Scholar

15. 

C. Göring et al., “Semantic segmentation using GrabCut,” in Proc. of the Int. Conf. on Computer Vision Theory and Applications (VISAPP), (2012). Google Scholar

16. 

P. Arbelaez et al., “Semantic segmentation using regions and parts,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 3378 –3385 (2012). http://dx.doi.org/10.1109/CVPR.2012.6248077 Google Scholar

17. 

G. Csurka and F. Perronnin, “An efficient approach to semantic segmentation,” Int. J. Comput. Vision, 95 (2), 198 –212 (2011). http://dx.doi.org/10.1007/s11263-010-0344-8 IJCVEQ 0920-5691 Google Scholar

18. 

J. Yang, B. Price and S. Cohen, “Context driven scene parsing with attention to rare classes,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 3294 –3301 (2014). http://dx.doi.org/10.1109/CVPR.2014.415 Google Scholar

19. 

H. Zhao et al., “Pyramid scene parsing network,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2881 –2890 (2017). Google Scholar

20. 

Y. Li et al., “Fully convolutional instance-aware semantic segmentation,” in IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2359 –2367 (2017). Google Scholar

21. 

H. Noh, S. Hong and B. Han, “Learning deconvolution network for semantic segmentation,” in IEEE Int. Conf. on Computer Vision (ICCV), 1520 –1528 (2015). http://dx.doi.org/10.1109/ICCV.2015.178 Google Scholar

22. 

J. Long, E. Shelhamer and T. Darrel, “Fully convolutional networks for semantic segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., 39 (4), 640 –651 (2017). http://dx.doi.org/10.1109/TPAMI.2016.2572683 Google Scholar

23. 

G. Singh and J. Košecká, “Semantic context for nonparametric scene parsing and scene classification,” in Scene Understanding Workshop, CVPR, (2013). Google Scholar

24. 

G. Singh and J. Košecká, “Nonparametric scene parsing with adaptive feature relevance and semantic context,” in IEEE Conf. on Computer Vision and Pattern Recognition, 3151 –3157 (2013). http://dx.doi.org/10.1109/CVPR.2013.405 Google Scholar

25. 

C. Rother et al., “‘GrabCut’: interactive foreground extraction using iterated graph cuts,” ACM Trans. Graphics, 23 (3), 309 –314 (2004). http://dx.doi.org/10.1145/1186562.1015720 ATGRDF 0730-0301 Google Scholar

26. 

J. Tighe and S. Lazebnik, “SuperParsing: scalable nonparametric image parsing with superpixels,” Lect. Notes Comput. Sci., 6315 352 –365 (2010). http://dx.doi.org/10.1007/978-3-642-15555-0_26 Google Scholar

27. 

M. C. Potter, “Meaning in visual search,” Science, 187 965 –966 (1975). http://dx.doi.org/10.1126/science.1145183 SCIEAS 0036-8075 Google Scholar

28. 

A. Oliva and A. Torralba, “Modeling the shape of the scene: a holistic representation of the spatial envelope,” Int. J. Comput. Vision, 42 (3), 145 –175 (2001). http://dx.doi.org/10.1023/A:1011139631724 IJCVEQ 0920-5691 Google Scholar

29. 

R. Achanta and A. Shaji, “SLIC superpixels compared to state-of-the art superpixel methods,” IEEE Trans. Pattern Anal. Mach. Intell., 34 (11), 2274 –2282 (2012). http://dx.doi.org/10.1109/TPAMI.2012.120 ITPIDJ 0162-8828 Google Scholar

30. 

C. Liu et al., “SIFT flow: dense correspondence across scenes and its applications,” IEEE Trans Pattern Anal Mach Intell., 33 (5), 978 –994 (2011). http://dx.doi.org/10.1109/TPAMI.2010.147 IESEDJ 0098-5589 Google Scholar

31. 

J. A. Hartigan and M. A. Wong, “A K-means clustering algorithm,” Appl. Stat., 28 (1), 100 –108 (1979). http://dx.doi.org/10.2307/2346830 APSTAG Google Scholar

Biography

Ding Yuan received her PhD in mechanical and automation engineering from the Chinese University of Hong Kong and has worked in the field of computer vision for over 15 years. She is now an associate professor at the Image Processing Center, School of Astronautics, Beihang University.

Jingjing Qiang is a postgraduate student at the Image Processing Center, School of Astronautics, Beihang University.

Jihao Yin received his PhD in computer science from Beijing Institute of Technology. He is now an associate professor at the Image Processing Center, School of Astronautics, Beihang University. His research interests include image processing and remote sensing.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Ding Yuan, Jingjing Qiang, and Jihao Yin "Image segmentation via foreground and background semantic descriptors," Journal of Electronic Imaging 26(5), 053004 (9 September 2017). https://doi.org/10.1117/1.JEI.26.5.053004
Received: 18 April 2017; Accepted: 9 August 2017; Published: 9 September 2017
Lens.org Logo
CITATIONS
Cited by 2 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Image segmentation

Image retrieval

Information visualization

Visualization

Image processing algorithms and systems

Neural networks

Wavelets

Back to Top