nips nips2008 nips2008-95 knowledge-graph by maker-knowledge-mining

95 nips-2008-Grouping Contours Via a Related Image


Source: pdf

Author: Praveen Srinivasan, Liming Wang, Jianbo Shi

Abstract: Contours have been established in the biological and computer vision literature as a compact yet descriptive representation of object shape. While individual contours provide structure, they lack the large spatial support of region segments (which lack internal structure). We present a method for further grouping of contours in an image using their relationship to the contours of a second, related image. Stereo, motion, and similarity all provide cues that can aid this task; contours that have similar transformations relating them to their matching contours in the second image likely belong to a single group. To find matches for contours, we rely only on shape, which applies directly to all three modalities without modification, in contrast to the specialized approaches developed for each independently. Visually salient contours are extracted in each image, along with a set of candidate transformations for aligning subsets of them. For each transformation, groups of contours with matching shape across the two images are identified to provide a context for evaluating matches of individual contour points across the images. The resulting contexts of contours are used to perform a final grouping on contours in the original image while simultaneously finding matches in the related image, again by shape matching. We demonstrate grouping results on image pairs consisting of stereo, motion, and similar images. Our method also produces qualitatively better results against a baseline method that does not use the inferred contexts. 1

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 While individual contours provide structure, they lack the large spatial support of region segments (which lack internal structure). [sent-8, score-0.852]

2 We present a method for further grouping of contours in an image using their relationship to the contours of a second, related image. [sent-9, score-1.834]

3 Stereo, motion, and similarity all provide cues that can aid this task; contours that have similar transformations relating them to their matching contours in the second image likely belong to a single group. [sent-10, score-1.945]

4 Visually salient contours are extracted in each image, along with a set of candidate transformations for aligning subsets of them. [sent-12, score-0.849]

5 For each transformation, groups of contours with matching shape across the two images are identified to provide a context for evaluating matches of individual contour points across the images. [sent-13, score-1.715]

6 The resulting contexts of contours are used to perform a final grouping on contours in the original image while simultaneously finding matches in the related image, again by shape matching. [sent-14, score-2.214]

7 We demonstrate grouping results on image pairs consisting of stereo, motion, and similar images. [sent-15, score-0.322]

8 1 Introduction Researchers in biological vision have long hypothesized that image contours (ordered sets of edge pixels, or contour points) are a compact yet descriptive representation of object shape. [sent-17, score-1.224]

9 In computer vision, there has been substantial interest in extracting contours from images as well as using object models based on contours for object recognition ([15, 5]), and 3D image interpretation [11]. [sent-18, score-1.917]

10 We examine the problem of grouping contours in a single image aided by a related image, such as stereo pair, a frame from the same motion sequence, or a similar image. [sent-19, score-1.426]

11 Relative motion of contours in one image to their matching contours in the other provides a cue for grouping. [sent-20, score-1.959]

12 The contours themselves are detected bottom-up without a model, and are provided as input to our method. [sent-21, score-0.756]

13 While contours already represent groupings of edges, they typically lack large spatial support. [sent-22, score-0.847]

14 Region segments, on the other hand, have large spatial support, but lack the structure that contours provide. [sent-23, score-0.796]

15 Therefore, additional grouping of contours can give us both qualities. [sent-24, score-0.895]

16 This has important applications for object recognition and scene understanding, since these larger groups of contours are often large pieces of objects. [sent-25, score-0.877]

17 Figure 1 shows a single image in the 1st column, with contours; in the other columns, top row, are different images related by stereo, motion and similarity to the first, shown with their contours. [sent-26, score-0.438]

18 Below each of these images are idealized groupings of contours in the original image. [sent-27, score-0.933]

19 Note that internal contours on cars and buildings are grouped, providing rich, structured shape information over a larger image region. [sent-28, score-1.169]

20 1 Stereo Motion Similarity Image 2 Image 1 Grouping in image 1 using related images Figure 1: Contours (white) in the image on the left can be further grouped using the contours of a second, related image (top row). [sent-29, score-1.442]

21 2 Related Work Stereo, motion, and similar image matching have been studied largely in isolation, and often with different purposes in mind than perceptual grouping. [sent-31, score-0.324]

22 Much of the stereo literature focuses on perpixel depth recovery; however, as [7] noted, stereo can be used for perceptual grouping without requiring precise depth estimation. [sent-32, score-0.639]

23 These approaches to motion and stereo are largely region-based, and therefore do not provide the same internal structure that groups of contours provide. [sent-34, score-1.172]

24 Similar image matching has been used for object recognition [1], but is rarely applied to image segmentation. [sent-35, score-0.577]

25 In work on contours, [12] matched contour points in the context of aerial imagery, but use constraints such as ordering of matches along scanlines that are not appropriate for motion or similar images, and do not provide grouping information. [sent-36, score-0.72]

26 [9] grouped image pixels into contours according to similar motion using optical flow as a local cue. [sent-37, score-1.124]

27 [8] grouped and matched image regions across different images and unstable segmentations (as we do with contours), but the regions lack internal structure. [sent-39, score-0.394]

28 [2, 6] used stereo pairs of images to detect depth discontinuities as potential object boundaries. [sent-40, score-0.441]

29 However, these methods will not detect and group group contours in the interior of fronto-parallel surfaces. [sent-41, score-0.82]

30 A set of contours Ii , where the jth contour Ci Ii is an ordered subset of points in Ii : j C pkn ] Ii . [sent-48, score-1.027]

31 A transformation Ti that aligns a subset of contours (e. [sent-50, score-0.788]

32 A subset of contours Coni in each image, known as a context, such that the two subsets C C 0 1 I1 , Con2 0 1 I2 . [sent-55, score-0.756]

33 Coni = Con1 Con2 ; Con1 i i i i j Each Coni is vector that indicates which contours are in the context for image Ij . [sent-57, score-1.044]

34 j We further define the following variables on contours C1 = [pk1 pkn ]: j 1. [sent-59, score-0.789]

35 Good continuation - for contours that overlap significantly, we prefer that they are present in the same group, if they are grouped at all. [sent-69, score-0.794]

36 Common fate by shape: contours with similar transformations mapping them to their matches should be grouped. [sent-71, score-0.93]

37 We also require that each contour point in a grouped contour and its matching point in the second image have similar local shape. [sent-72, score-0.794]

38 Shape in each image is defined with respect to a subset of image contours known as a context; we will explain the importance of choosing the correct context for shape comparison, as first noted in [16]. [sent-73, score-1.418]

39 Maximality/simplicity: We would like to group as many of the contours as possible into as few groups as possible, while still maintaining the similarity of local shape described above. [sent-75, score-1.063]

40 This combination of the MRF standard graph matching technique with an LP for inferring context for accurate matching by shape is our main contribution. [sent-81, score-0.604]

41 4 Matching and Context Selection We can evaluate the hypothesis of a particular contour point match by comparing the local shape around the point and around its match. [sent-84, score-0.492]

42 Although local features such as curvature and simple proximity (in the case of roughly aligned contours) have been used for matching ([12]), inconsistencies in the input contours across two images make these them prone to error. [sent-85, score-1.049]

43 In a), the contours do not match, while in e), a “7” shape is common to both. [sent-89, score-0.947]

44 As an example, we use the shape context, an image feature that has been widely used for shape matching ([1]). [sent-98, score-0.706]

45 Briefly, a shape context provides a log-polar spatial histogram that records the number of points that fall into a particular bin. [sent-99, score-0.336]

46 In Figure 2 c,g), shape contexts (darker bins mean larger bin count) with large spatial support placed according to the rough alignment exhibit high dissimilarity in both cases, failing to find a match in a). [sent-100, score-0.519]

47 The large feature failed because contours in each image that had no match in the other image were used in computing the shape context. [sent-101, score-1.398]

48 Inferring which contours to include and which to omit would give better features, as in Figure 2 d),h). [sent-102, score-0.756]

49 This fixes the completeness problem, while retaining soundness: no combination of contours in the two images in a) can produce matching shapes. [sent-103, score-1.012]

50 Therefore, with rough alignment we can cast the first step of shape matching as context selection: which subset of contours, or context, to use for feature computation. [sent-104, score-0.478]

51 Given the correct context, matching individual contour points is much easier. [sent-105, score-0.379]

52 1 Selection Problem We can neatly summarize the effect of a particular context selection on a shape context as seen in Figure 3. [sent-107, score-0.45]

53 The bin counts for a shape context can be encoded as a vector, represented by the shape contexts and their vector representations alongside. [sent-110, score-0.676]

54 SC has dimensions nBins by nContours, where nBins is the number of bins in the shape context (and the length of the associated vector). [sent-112, score-0.321]

55 The entry SC(i j) is the bin count for bin i and contour j. [sent-113, score-0.352]

56 For each j 0 1 , which indicates contour Ci in an image Ii , we associate an selection indicator variable selj i whether or not the contour is selected; the vector of these indicator variables is seli . [sent-114, score-0.727]

57 Then the shape context bin counts realized by a particular selection of contours is SCseli , simply the multiplication of the matrix SC and the vector seli . [sent-115, score-1.254]

58 d) shows the effect on the shape context histogram of various context selections. [sent-116, score-0.401]

59 2 Shape context matching cost The effectiveness of selection depends significantly on the shape context matching cost. [sent-118, score-0.755]

60 Traditional matching costs (Chi-square, L1 , L2 ) only measure similarity, but selecting no contours in either image gives a perfect matching cost, since in both shape contexts, all bins will be 0. [sent-119, score-1.437]

61 While similarity is important, so is including more contours rather than fewer (maximality). [sent-120, score-0.789]

62 The selection process is run for each transformation to infer a context suitable for evaluating contour point matches via shape. [sent-122, score-0.5]

63 The intersection term encourages higher bin counts in each shape context and therefore the inclusion of more contours. [sent-124, score-0.394]

64 We begin with two C input images, where the contours are in rough alignment (by applying known Ti to I1 ). [sent-130, score-0.797]

65 Multiple shape contexts are placed in each image on a uniform grid (an approximation of Matchj l(j)=i , since we initially have no matches). [sent-131, score-0.465]

66 Our goal is to select contours in each image to minimize the sum of SCMatchCost for each pair of shape contexts. [sent-133, score-1.13]

67 For each shape context j in each image i, we compute the corresponding j j shape context matrix SCi . [sent-134, score-0.775]

68 SCi j for each image has been color coded to show the SCi matrix corresponding to each shape context. [sent-136, score-0.374]

69 SCi selci is then the realized bin counts for i all the shape contexts in image Ii under selection selci . [sent-139, score-0.692]

70 We seek to choose selc1 and selc2 such that SC1 selc1 SC2 selc2 in a shape sense; entries of SC1 selc1 and SC2 selc2 , or realized bin counts, are in correspondence, so we can score these pairs of bin counts using BinMatchCost. [sent-140, score-0.405]

71 Figure 5 depicts an idealized selection process for two images (only the contours are shown). [sent-156, score-0.931]

72 For groups of SIFT matches that describe similar transformations, a transformation Ti is extracted and warps the contours in image 1 to line up with those of image 2, in c). [sent-157, score-1.32]

73 The selection problem is formulated separately for each set of aligned contours d). [sent-158, score-0.84]

74 Two correct transforms align the car and person, and the selection result includes the respective contours (rows 1,2 of e). [sent-160, score-0.825]

75 We can view the context selection procedure for minimizing Fi as choosing the context of contours so as to best reduce the matching cost of the hypothesized inter-image matches for contours with label i, under the transformation Ti . [sent-162, score-2.081]

76 5 Graph Cuts for Group Assignment and Matching We previously computed context selections (as solutions to the SelectionCost LP), which found groups of contours in each image that have similar shape, Con = {Con1 , . [sent-165, score-1.115]

77 Note that a j contour C1 need not be selected as context in a particular group a in order to have lj = a. [sent-174, score-0.472]

78 j k The binary potentials φ(lj , lk ) encode the preference that overlapping contours C1 , C1 have the same label: 1 a=b φ(lj = a, lk = b) = (1) 1−τ a=b where 0 ≤ τ ≤ 1 controls the penalty of having different labels. [sent-177, score-0.808]

79 Two contours overlap if they contain at least one point in common. [sent-179, score-0.756]

80 , pkn ] can be matched in the second image with respect to the context {Con1 , Con2 }. [sent-183, score-0.338]

81 Instead of using all contours image 1 as nodes in the MN, we only allow contours were selected in at least one of the context Con1 ; likewise, we only permit i matches to points in image 2 that appear in a contour selected in at least one Con2 . [sent-186, score-2.319]

82 This better j allows us to deal with contours that appear only in one image and thus cannot be reliably grouped based on relative motion. [sent-187, score-0.977]

83 Columns 3,4: grouping results for our method and baseline; groups of contours are a single color. [sent-190, score-0.946]

84 If more than half the points in the final assignment ∗ lj for a contour were occluded, we marked the entire contour as occluded, and it was not displayed. [sent-200, score-0.573]

85 Since we omitted all selection information, all contours in the 1st image were included in the MN P as nodes, and their contour points were allowed to match to any contour point in I2 . [sent-201, score-1.527]

86 6 Experiments We tested our method and the baseline over stereo, motion and similar image pairs. [sent-204, score-0.35]

87 Input contours in each image were extracted automatically using the method of [15]. [sent-205, score-0.956]

88 SIFT matches were extracted from images, keeping only confident matches as described in [10]; matches proposing similar transformations were pruned to a small set, typically 10-20. [sent-206, score-0.403]

89 Because of the high quality of the inferred contexts, we used large shape contexts (radius 90 pixels, in images of size 400 by 500), which 7 made matching very robust. [sent-207, score-0.522]

90 Shape contexts were placed on a uniform grid atop the registered contours (via Ti ) with a spacing 50 pixels in the x and y dimensions. [sent-209, score-0.871]

91 Image pairs were taken from the Caltech 101 dataset [4] and from a stereo rig with 1m baseline mounted on a car from our lab (providing stereo and motion images). [sent-210, score-0.637]

92 Groups for stereo image pairs are colored according to disparity. [sent-214, score-0.408]

93 Due to the lack of large context, the baseline method is able to find a good match for a given contour point under almost any group hypothesis lj = a, since in cluttered regions, there are always nearby matches. [sent-215, score-0.518]

94 7 Conclusion We introduced the problem of grouping of contours in an image using a related image, such as stereo, motion or similar, as an important step for object recognition and scene understanding. [sent-219, score-1.271]

95 Grouping depends on the ability to match contours across images to determine their relative motion. [sent-220, score-0.958]

96 Selecting a good context for shape evaluation was key to robust simultaneous and grouping of contours across images. [sent-221, score-1.209]

97 Future work will include trying to learn 3D object models from stereo and motion images, and a probabilistic formulation of the matching framework. [sent-223, score-0.542]

98 Introducing learning to improve the grouping result is also an area of significant interest; some shape configurations are more reliable for matching than others. [sent-224, score-0.471]

99 Recognition of object contours from stereo images: an edge combination approach. [sent-258, score-1.034]

100 Contour context selection for object detection: A set-to-set contour matching approach. [sent-317, score-0.564]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('contours', 0.756), ('stereo', 0.225), ('contour', 0.216), ('shape', 0.191), ('image', 0.183), ('matching', 0.141), ('grouping', 0.139), ('sc', 0.132), ('motion', 0.123), ('lj', 0.119), ('coni', 0.114), ('context', 0.105), ('images', 0.099), ('matches', 0.098), ('contexts', 0.091), ('match', 0.085), ('selectioncost', 0.076), ('transformations', 0.076), ('bin', 0.068), ('matchj', 0.063), ('ti', 0.057), ('object', 0.053), ('groups', 0.051), ('groupings', 0.051), ('scmatchcost', 0.051), ('selection', 0.049), ('mn', 0.046), ('sift', 0.046), ('baseline', 0.044), ('pami', 0.044), ('sci', 0.043), ('grouped', 0.038), ('lp', 0.038), ('maximality', 0.038), ('pki', 0.038), ('selj', 0.038), ('ta', 0.038), ('aligned', 0.035), ('pkn', 0.033), ('similarity', 0.033), ('transformation', 0.032), ('group', 0.032), ('counts', 0.03), ('realized', 0.03), ('ii', 0.027), ('idealized', 0.027), ('graph', 0.026), ('lk', 0.026), ('occluded', 0.026), ('binmatchcost', 0.025), ('liming', 0.025), ('matchcostincontext', 0.025), ('nbins', 0.025), ('ngroups', 0.025), ('occlusionthresh', 0.025), ('qrn', 0.025), ('scmatchcos', 0.025), ('selci', 0.025), ('seli', 0.025), ('bins', 0.025), ('closest', 0.025), ('depth', 0.025), ('unary', 0.025), ('pixels', 0.024), ('min', 0.023), ('mismatch', 0.023), ('cost', 0.023), ('lack', 0.022), ('points', 0.022), ('criteria', 0.022), ('jianbo', 0.022), ('grasp', 0.022), ('buildings', 0.022), ('discontinuities', 0.022), ('rough', 0.022), ('mrf', 0.02), ('selections', 0.02), ('soundness', 0.02), ('ci', 0.02), ('car', 0.02), ('cuts', 0.02), ('dense', 0.02), ('alignment', 0.019), ('across', 0.018), ('fi', 0.018), ('seek', 0.018), ('swap', 0.018), ('spatial', 0.018), ('shapes', 0.017), ('recognition', 0.017), ('xes', 0.017), ('internal', 0.017), ('extracted', 0.017), ('segments', 0.017), ('matched', 0.017), ('potential', 0.017), ('completeness', 0.016), ('hypothesized', 0.016), ('edges', 0.016), ('proposing', 0.016)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000001 95 nips-2008-Grouping Contours Via a Related Image

Author: Praveen Srinivasan, Liming Wang, Jianbo Shi

Abstract: Contours have been established in the biological and computer vision literature as a compact yet descriptive representation of object shape. While individual contours provide structure, they lack the large spatial support of region segments (which lack internal structure). We present a method for further grouping of contours in an image using their relationship to the contours of a second, related image. Stereo, motion, and similarity all provide cues that can aid this task; contours that have similar transformations relating them to their matching contours in the second image likely belong to a single group. To find matches for contours, we rely only on shape, which applies directly to all three modalities without modification, in contrast to the specialized approaches developed for each independently. Visually salient contours are extracted in each image, along with a set of candidate transformations for aligning subsets of them. For each transformation, groups of contours with matching shape across the two images are identified to provide a context for evaluating matches of individual contour points across the images. The resulting contexts of contours are used to perform a final grouping on contours in the original image while simultaneously finding matches in the related image, again by shape matching. We demonstrate grouping results on image pairs consisting of stereo, motion, and similar images. Our method also produces qualitatively better results against a baseline method that does not use the inferred contexts. 1

2 0.39855045 6 nips-2008-A ``Shape Aware'' Model for semi-supervised Learning of Objects and its Context

Author: Abhinav Gupta, Jianbo Shi, Larry S. Davis

Abstract: We present an approach that combines bag-of-words and spatial models to perform semantic and syntactic analysis for recognition of an object based on its internal appearance and its context. We argue that while object recognition requires modeling relative spatial locations of image features within the object, a bag-of-word is sufficient for representing context. Learning such a model from weakly labeled data involves labeling of features into two classes: foreground(object) or “informative” background(context). We present a “shape-aware” model which utilizes contour information for efficient and accurate labeling of features in the image. Our approach iterates between an MCMC-based labeling and contour based labeling of features to integrate co-occurrence of features and shape similarity. 1

3 0.24526483 147 nips-2008-Multiscale Random Fields with Application to Contour Grouping

Author: Longin J. Latecki, Chengen Lu, Marc Sobel, Xiang Bai

Abstract: We introduce a new interpretation of multiscale random fields (MSRFs) that admits efficient optimization in the framework of regular (single level) random fields (RFs). It is based on a new operator, called append, that combines sets of random variables (RVs) to single RVs. We assume that a MSRF can be decomposed into disjoint trees that link RVs at different pyramid levels. The append operator is then applied to map RVs in each tree structure to a single RV. We demonstrate the usefulness of the proposed approach on a challenging task involving grouping contours of target shapes in images. It provides a natural representation of multiscale contour models, which is needed in order to cope with unstable contour decompositions. The append operator allows us to find optimal image segment labels using the classical framework of relaxation labeling. Alternative methods like Markov Chain Monte Carlo (MCMC) could also be used.

4 0.15586901 201 nips-2008-Robust Near-Isometric Matching via Structured Learning of Graphical Models

Author: Alex J. Smola, Julian J. Mcauley, Tibério S. Caetano

Abstract: Models for near-rigid shape matching are typically based on distance-related features, in order to infer matches that are consistent with the isometric assumption. However, real shapes from image datasets, even when expected to be related by “almost isometric” transformations, are actually subject not only to noise but also, to some limited degree, to variations in appearance and scale. In this paper, we introduce a graphical model that parameterises appearance, distance, and angle features and we learn all of the involved parameters via structured prediction. The outcome is a model for near-rigid shape matching which is robust in the sense that it is able to capture the possibly limited but still important scale and appearance variations. Our experimental results reveal substantial improvements upon recent successful models, while maintaining similar running times. 1

5 0.11666238 207 nips-2008-Shape-Based Object Localization for Descriptive Classification

Author: Geremy Heitz, Gal Elidan, Benjamin Packer, Daphne Koller

Abstract: Discriminative tasks, including object categorization and detection, are central components of high-level computer vision. Sometimes, however, we are interested in more refined aspects of the object in an image, such as pose or particular regions. In this paper we develop a method (LOOPS) for learning a shape and image feature model that can be trained on a particular object class, and used to outline instances of the class in novel images. Furthermore, while the training data consists of uncorresponded outlines, the resulting LOOPS model contains a set of landmark points that appear consistently across instances, and can be accurately localized in an image. Our model achieves state-of-the-art results in precisely outlining objects that exhibit large deformations and articulations in cluttered natural images. These localizations can then be used to address a range of tasks, including descriptive classification, search, and clustering. 1

6 0.10880589 116 nips-2008-Learning Hybrid Models for Image Annotation with Partially Labeled Data

7 0.10279786 157 nips-2008-Nonrigid Structure from Motion in Trajectory Space

8 0.094628647 118 nips-2008-Learning Transformational Invariants from Natural Movies

9 0.091668487 191 nips-2008-Recursive Segmentation and Recognition Templates for 2D Parsing

10 0.085031047 113 nips-2008-Kernelized Sorting

11 0.078332685 136 nips-2008-Model selection and velocity estimation using novel priors for motion patterns

12 0.07824865 148 nips-2008-Natural Image Denoising with Convolutional Networks

13 0.076961234 42 nips-2008-Cascaded Classification Models: Combining Models for Holistic Scene Understanding

14 0.073982388 208 nips-2008-Shared Segmentation of Natural Scenes Using Dependent Pitman-Yor Processes

15 0.073580228 119 nips-2008-Learning a discriminative hidden part model for human action recognition

16 0.071404755 246 nips-2008-Unsupervised Learning of Visual Sense Models for Polysemous Words

17 0.06367778 130 nips-2008-MCBoost: Multiple Classifier Boosting for Perceptual Co-clustering of Images and Visual Features

18 0.059039302 142 nips-2008-Multi-Level Active Prediction of Useful Image Annotations for Recognition

19 0.054956526 192 nips-2008-Reducing statistical dependencies in natural signals using radial Gaussianization

20 0.049637631 104 nips-2008-Improved Moves for Truncated Convex Models


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, -0.151), (1, -0.132), (2, 0.154), (3, -0.194), (4, 0.031), (5, -0.01), (6, -0.142), (7, -0.258), (8, 0.036), (9, -0.044), (10, -0.1), (11, -0.128), (12, 0.094), (13, -0.046), (14, 0.125), (15, -0.001), (16, 0.085), (17, 0.186), (18, -0.068), (19, 0.02), (20, -0.083), (21, -0.002), (22, -0.022), (23, 0.009), (24, 0.087), (25, 0.197), (26, 0.058), (27, 0.077), (28, 0.008), (29, 0.115), (30, -0.1), (31, 0.069), (32, 0.092), (33, -0.001), (34, -0.067), (35, 0.045), (36, -0.086), (37, -0.049), (38, -0.058), (39, 0.188), (40, -0.119), (41, 0.071), (42, 0.008), (43, -0.025), (44, 0.033), (45, -0.067), (46, 0.124), (47, -0.031), (48, -0.042), (49, 0.02)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.97285897 95 nips-2008-Grouping Contours Via a Related Image

Author: Praveen Srinivasan, Liming Wang, Jianbo Shi

Abstract: Contours have been established in the biological and computer vision literature as a compact yet descriptive representation of object shape. While individual contours provide structure, they lack the large spatial support of region segments (which lack internal structure). We present a method for further grouping of contours in an image using their relationship to the contours of a second, related image. Stereo, motion, and similarity all provide cues that can aid this task; contours that have similar transformations relating them to their matching contours in the second image likely belong to a single group. To find matches for contours, we rely only on shape, which applies directly to all three modalities without modification, in contrast to the specialized approaches developed for each independently. Visually salient contours are extracted in each image, along with a set of candidate transformations for aligning subsets of them. For each transformation, groups of contours with matching shape across the two images are identified to provide a context for evaluating matches of individual contour points across the images. The resulting contexts of contours are used to perform a final grouping on contours in the original image while simultaneously finding matches in the related image, again by shape matching. We demonstrate grouping results on image pairs consisting of stereo, motion, and similar images. Our method also produces qualitatively better results against a baseline method that does not use the inferred contexts. 1

2 0.85058832 147 nips-2008-Multiscale Random Fields with Application to Contour Grouping

Author: Longin J. Latecki, Chengen Lu, Marc Sobel, Xiang Bai

Abstract: We introduce a new interpretation of multiscale random fields (MSRFs) that admits efficient optimization in the framework of regular (single level) random fields (RFs). It is based on a new operator, called append, that combines sets of random variables (RVs) to single RVs. We assume that a MSRF can be decomposed into disjoint trees that link RVs at different pyramid levels. The append operator is then applied to map RVs in each tree structure to a single RV. We demonstrate the usefulness of the proposed approach on a challenging task involving grouping contours of target shapes in images. It provides a natural representation of multiscale contour models, which is needed in order to cope with unstable contour decompositions. The append operator allows us to find optimal image segment labels using the classical framework of relaxation labeling. Alternative methods like Markov Chain Monte Carlo (MCMC) could also be used.

3 0.79483604 6 nips-2008-A ``Shape Aware'' Model for semi-supervised Learning of Objects and its Context

Author: Abhinav Gupta, Jianbo Shi, Larry S. Davis

Abstract: We present an approach that combines bag-of-words and spatial models to perform semantic and syntactic analysis for recognition of an object based on its internal appearance and its context. We argue that while object recognition requires modeling relative spatial locations of image features within the object, a bag-of-word is sufficient for representing context. Learning such a model from weakly labeled data involves labeling of features into two classes: foreground(object) or “informative” background(context). We present a “shape-aware” model which utilizes contour information for efficient and accurate labeling of features in the image. Our approach iterates between an MCMC-based labeling and contour based labeling of features to integrate co-occurrence of features and shape similarity. 1

4 0.61560947 207 nips-2008-Shape-Based Object Localization for Descriptive Classification

Author: Geremy Heitz, Gal Elidan, Benjamin Packer, Daphne Koller

Abstract: Discriminative tasks, including object categorization and detection, are central components of high-level computer vision. Sometimes, however, we are interested in more refined aspects of the object in an image, such as pose or particular regions. In this paper we develop a method (LOOPS) for learning a shape and image feature model that can be trained on a particular object class, and used to outline instances of the class in novel images. Furthermore, while the training data consists of uncorresponded outlines, the resulting LOOPS model contains a set of landmark points that appear consistently across instances, and can be accurately localized in an image. Our model achieves state-of-the-art results in precisely outlining objects that exhibit large deformations and articulations in cluttered natural images. These localizations can then be used to address a range of tasks, including descriptive classification, search, and clustering. 1

5 0.5954594 201 nips-2008-Robust Near-Isometric Matching via Structured Learning of Graphical Models

Author: Alex J. Smola, Julian J. Mcauley, Tibério S. Caetano

Abstract: Models for near-rigid shape matching are typically based on distance-related features, in order to infer matches that are consistent with the isometric assumption. However, real shapes from image datasets, even when expected to be related by “almost isometric” transformations, are actually subject not only to noise but also, to some limited degree, to variations in appearance and scale. In this paper, we introduce a graphical model that parameterises appearance, distance, and angle features and we learn all of the involved parameters via structured prediction. The outcome is a model for near-rigid shape matching which is robust in the sense that it is able to capture the possibly limited but still important scale and appearance variations. Our experimental results reveal substantial improvements upon recent successful models, while maintaining similar running times. 1

6 0.46364892 157 nips-2008-Nonrigid Structure from Motion in Trajectory Space

7 0.37555972 119 nips-2008-Learning a discriminative hidden part model for human action recognition

8 0.37258118 116 nips-2008-Learning Hybrid Models for Image Annotation with Partially Labeled Data

9 0.36014003 148 nips-2008-Natural Image Denoising with Convolutional Networks

10 0.32860202 23 nips-2008-An ideal observer model of infant object perception

11 0.32182726 208 nips-2008-Shared Segmentation of Natural Scenes Using Dependent Pitman-Yor Processes

12 0.31023952 42 nips-2008-Cascaded Classification Models: Combining Models for Holistic Scene Understanding

13 0.30727342 246 nips-2008-Unsupervised Learning of Visual Sense Models for Polysemous Words

14 0.29308999 191 nips-2008-Recursive Segmentation and Recognition Templates for 2D Parsing

15 0.27823117 118 nips-2008-Learning Transformational Invariants from Natural Movies

16 0.27722415 113 nips-2008-Kernelized Sorting

17 0.25965813 41 nips-2008-Breaking Audio CAPTCHAs

18 0.24383064 232 nips-2008-The Conjoint Effect of Divisive Normalization and Orientation Selectivity on Redundancy Reduction

19 0.24106999 104 nips-2008-Improved Moves for Truncated Convex Models

20 0.22222747 142 nips-2008-Multi-Level Active Prediction of Useful Image Annotations for Recognition


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(6, 0.031), (7, 0.078), (12, 0.089), (15, 0.015), (28, 0.104), (57, 0.088), (59, 0.038), (63, 0.017), (66, 0.235), (71, 0.027), (77, 0.032), (81, 0.023), (83, 0.112)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.84422803 18 nips-2008-An Efficient Sequential Monte Carlo Algorithm for Coalescent Clustering

Author: Dilan Gorur, Yee W. Teh

Abstract: We propose an efficient sequential Monte Carlo inference scheme for the recently proposed coalescent clustering model [1]. Our algorithm has a quadratic runtime while those in [1] is cubic. In experiments, we were surprised to find that in addition to being more efficient, it is also a better sequential Monte Carlo sampler than the best in [1], when measured in terms of variance of estimated likelihood and effective sample size. 1

same-paper 2 0.7979843 95 nips-2008-Grouping Contours Via a Related Image

Author: Praveen Srinivasan, Liming Wang, Jianbo Shi

Abstract: Contours have been established in the biological and computer vision literature as a compact yet descriptive representation of object shape. While individual contours provide structure, they lack the large spatial support of region segments (which lack internal structure). We present a method for further grouping of contours in an image using their relationship to the contours of a second, related image. Stereo, motion, and similarity all provide cues that can aid this task; contours that have similar transformations relating them to their matching contours in the second image likely belong to a single group. To find matches for contours, we rely only on shape, which applies directly to all three modalities without modification, in contrast to the specialized approaches developed for each independently. Visually salient contours are extracted in each image, along with a set of candidate transformations for aligning subsets of them. For each transformation, groups of contours with matching shape across the two images are identified to provide a context for evaluating matches of individual contour points across the images. The resulting contexts of contours are used to perform a final grouping on contours in the original image while simultaneously finding matches in the related image, again by shape matching. We demonstrate grouping results on image pairs consisting of stereo, motion, and similar images. Our method also produces qualitatively better results against a baseline method that does not use the inferred contexts. 1

3 0.70910734 245 nips-2008-Unlabeled data: Now it helps, now it doesn't

Author: Aarti Singh, Robert Nowak, Xiaojin Zhu

Abstract: Empirical evidence shows that in favorable situations semi-supervised learning (SSL) algorithms can capitalize on the abundance of unlabeled training data to improve the performance of a learning task, in the sense that fewer labeled training data are needed to achieve a target error bound. However, in other situations unlabeled data do not seem to help. Recent attempts at theoretically characterizing SSL gains only provide a partial and sometimes apparently conflicting explanations of whether, and to what extent, unlabeled data can help. In this paper, we attempt to bridge the gap between the practice and theory of semi-supervised learning. We develop a finite sample analysis that characterizes the value of unlabeled data and quantifies the performance improvement of SSL compared to supervised learning. We show that there are large classes of problems for which SSL can significantly outperform supervised learning, in finite sample regimes and sometimes also in terms of error convergence rates.

4 0.61552614 207 nips-2008-Shape-Based Object Localization for Descriptive Classification

Author: Geremy Heitz, Gal Elidan, Benjamin Packer, Daphne Koller

Abstract: Discriminative tasks, including object categorization and detection, are central components of high-level computer vision. Sometimes, however, we are interested in more refined aspects of the object in an image, such as pose or particular regions. In this paper we develop a method (LOOPS) for learning a shape and image feature model that can be trained on a particular object class, and used to outline instances of the class in novel images. Furthermore, while the training data consists of uncorresponded outlines, the resulting LOOPS model contains a set of landmark points that appear consistently across instances, and can be accurately localized in an image. Our model achieves state-of-the-art results in precisely outlining objects that exhibit large deformations and articulations in cluttered natural images. These localizations can then be used to address a range of tasks, including descriptive classification, search, and clustering. 1

5 0.60892338 116 nips-2008-Learning Hybrid Models for Image Annotation with Partially Labeled Data

Author: Xuming He, Richard S. Zemel

Abstract: Extensive labeled data for image annotation systems, which learn to assign class labels to image regions, is difficult to obtain. We explore a hybrid model framework for utilizing partially labeled data that integrates a generative topic model for image appearance with discriminative label prediction. We propose three alternative formulations for imposing a spatial smoothness prior on the image labels. Tests of the new models and some baseline approaches on three real image datasets demonstrate the effectiveness of incorporating the latent structure. 1

6 0.59625506 42 nips-2008-Cascaded Classification Models: Combining Models for Holistic Scene Understanding

7 0.59599251 194 nips-2008-Regularized Learning with Networks of Features

8 0.595415 193 nips-2008-Regularized Co-Clustering with Dual Supervision

9 0.59304357 205 nips-2008-Semi-supervised Learning with Weakly-Related Unlabeled Data : Towards Better Text Categorization

10 0.58990169 32 nips-2008-Bayesian Kernel Shaping for Learning Control

11 0.5886662 26 nips-2008-Analyzing human feature learning as nonparametric Bayesian inference

12 0.58329272 63 nips-2008-Dimensionality Reduction for Data in Multiple Feature Representations

13 0.58190548 66 nips-2008-Dynamic visual attention: searching for coding length increments

14 0.58015561 120 nips-2008-Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text

15 0.57982802 246 nips-2008-Unsupervised Learning of Visual Sense Models for Polysemous Words

16 0.57952136 201 nips-2008-Robust Near-Isometric Matching via Structured Learning of Graphical Models

17 0.57836449 208 nips-2008-Shared Segmentation of Natural Scenes Using Dependent Pitman-Yor Processes

18 0.57588446 229 nips-2008-Syntactic Topic Models

19 0.57517022 79 nips-2008-Exploring Large Feature Spaces with Hierarchical Multiple Kernel Learning

20 0.57424563 176 nips-2008-Partially Observed Maximum Entropy Discrimination Markov Networks