nips nips2012 nips2012-360 knowledge-graph by maker-knowledge-mining

360 nips-2012-Visual Recognition using Embedded Feature Selection for Curvature Self-Similarity


Source: pdf

Author: Angela Eigenstetter, Bjorn Ommer

Abstract: Category-level object detection has a crucial need for informative object representations. This demand has led to feature descriptors of ever increasing dimensionality like co-occurrence statistics and self-similarity. In this paper we propose a new object representation based on curvature self-similarity that goes beyond the currently popular approximation of objects using straight lines. However, like all descriptors using second order statistics, ours also exhibits a high dimensionality. Although improving discriminability, the high dimensionality becomes a critical issue due to lack of generalization ability and curse of dimensionality. Given only a limited amount of training data, even sophisticated learning algorithms such as the popular kernel methods are not able to suppress noisy or superfluous dimensions of such high-dimensional data. Consequently, there is a natural need for feature selection when using present-day informative features and, particularly, curvature self-similarity. We therefore suggest an embedded feature selection method for SVMs that reduces complexity and improves generalization capability of object models. By successfully integrating the proposed curvature self-similarity representation together with the embedded feature selection in a widely used state-of-the-art object detection framework we show the general pertinence of the approach. 1

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 de Abstract Category-level object detection has a crucial need for informative object representations. [sent-4, score-0.464]

2 This demand has led to feature descriptors of ever increasing dimensionality like co-occurrence statistics and self-similarity. [sent-5, score-0.266]

3 In this paper we propose a new object representation based on curvature self-similarity that goes beyond the currently popular approximation of objects using straight lines. [sent-6, score-0.883]

4 Consequently, there is a natural need for feature selection when using present-day informative features and, particularly, curvature self-similarity. [sent-10, score-0.77]

5 We therefore suggest an embedded feature selection method for SVMs that reduces complexity and improves generalization capability of object models. [sent-11, score-0.581]

6 By successfully integrating the proposed curvature self-similarity representation together with the embedded feature selection in a widely used state-of-the-art object detection framework we show the general pertinence of the approach. [sent-12, score-1.224]

7 Starting with brightness values of image pixels and simple edge histograms [10] descriptors evolved and more sophisticated features like shape context [1] and wavelets [23] were suggested. [sent-14, score-0.263]

8 The probably most widely used and best performing image descriptors today are SIFT [18] and HOG [4] which model objects based on edge orientation histograms. [sent-15, score-0.314]

9 Furthermore it is noticeable that all descriptors that model the object boundary rely on image statistics that are primarily based on edge orientation. [sent-18, score-0.358]

10 However, it was shown in different studies within the perception community that besides orientation also curvature is an important cue when performing visual search tasks. [sent-20, score-0.622]

11 In our earlier work [21] we extended the modeling of object boundary contours beyond the widely used edge orientation histograms by utilizing curvature information to overcome the drawbacks of straight line approximations. [sent-21, score-1.224]

12 However, curvature can provide even more information about the object boundary. [sent-22, score-0.668]

13 By computing co-occurrences between discriminatively curved boundaries we build a curvature self-similarity descriptor that provides a more detailed and accurate object description. [sent-23, score-0.929]

14 While it was shown that self-similarity and co-occurrence lead to very robust and highly discriminative object representations, these second order image statistics are also pushing feature spaces to extremely 1 high dimensions. [sent-24, score-0.377]

15 Since the amount of training data stays more or less the same, the dimensionality of the object representation has to be reduced to prevent systems to suffer from curse of dimensionality and overfitting. [sent-25, score-0.377]

16 [5], for instance, suggested an approach that results in a 160000 dimensional descriptor which was evaluated on the ETHZ shape dataset which contains on average 30 positive object instances per category. [sent-28, score-0.433]

17 To exploit the full capabilities of high-dimensional representations applied in object detection we developed a new embedded feature selection method for SVM which reliable discards superfluous dimensions and therefore improves object detection performance. [sent-29, score-1.074]

18 The paper is organized as follows: First we will give a short overview on embedded feature selection methods for SVMs (Section 2. [sent-30, score-0.408]

19 After that we describe our new self-similarity descriptor based on curvature to go beyond the straight line approximation of objects to a more accurate description (Section 3). [sent-33, score-0.845]

20 In the experimental section at the end of the paper we evaluate the suggested curvature self-similarity descriptor along with our feature selection method. [sent-35, score-0.954]

21 [12] categorize feature selection methods into filters, wrappers and embedded methods. [sent-38, score-0.441]

22 Contrary to filters and wrappers embedded feature selection methods incorporate feature selection as a part of the learning process (for a review see [17]). [sent-39, score-0.716]

23 The focus of this paper is on embedded feature selection methods for SVMs, since most state-of-the-art detection systems use SVM as a classifier. [sent-40, score-0.526]

24 To directly integrate feature selection into the learning process of SVMs sparsity can be enforced on the model parameter w. [sent-41, score-0.275]

25 [30] suggested the doubly regularized SVM (DrSVM) which is not replacing the L2 regularization but adding an additional L1 regularization to automatically select dimensions during the learning process. [sent-45, score-0.175]

26 Contrary to linear SVM enforcing sparsity on the model parameter w does reduce dimensionality for non-linear kernel functions in the higher dimensional kernel space rather than in the number of input features. [sent-46, score-0.229]

27 To reduce the dimensionality for non-linear SVMs in the feature space one can introduce an additional selection vector θ ∈ [0, 1]n , where larger values of θi indicate more useful features. [sent-47, score-0.362]

28 The objective is then to find the best kernel of the form Kθ (x, z) = K(θ ∗ x, θ ∗ z), where x, z ∈ Rn are the feature vectors and ∗ is element-wise multiplication. [sent-48, score-0.186]

29 In this paper we integrate the scaling factors into the learning algorithm, but instead of using L2 norm constraint like in [11] on the scaling parameter θ we apply an L1 norm sparsity which is explicitly discarding dimensions of the input feature vector. [sent-51, score-0.212]

30 We are following the concept of embedded feature selection and therefore include the feature selection parameter θ directly in the SVM classifier. [sent-56, score-0.683]

31 We enforce sparsity of the feature selection parameter θ by the last constraint of Eq. [sent-60, score-0.275]

32 The optimization of the selection parameter θ starts at the canonical solution where all dimensions are set to one. [sent-75, score-0.191]

33 An overview of the suggested iterative dimensionality reduction algorithm is given in Algorithm 1. [sent-95, score-0.216]

34 3 Representing Curvature Self-Similarity Although several methods have been suggested for the robust estimation of curvature, it has been mainly represented indirectly in a contour based manner [1, 32] and to locate interest points at boundary points with high curvature value. [sent-96, score-0.642]

35 To design a more exact object representation that represents object curvedness in a natural way we revisit the idea of [21] and design a novel curvature self-similarity descriptor. [sent-97, score-0.908]

36 [29] showed that using color histograms directly is decreasing performance while using color self-similarity (CSS) as a feature is more appropriate. [sent-106, score-0.217]

37 To make use of the aforementioned advantages of global self-similarity we compute all pairwise curvature similarities across the whole image. [sent-109, score-0.542]

38 This results in a very high dimensional object representation. [sent-110, score-0.207]

39 As mentioned before such high dimensional representations have a natural need for dimensionality reduction which we fulfill by applying our embedded feature selection algorithm outlined in the previous section. [sent-111, score-0.609]

40 To describe complex objects it is not sufficient to build a self-similarity descriptor solely based on curvature information, since self-similarity of curvature leaves open many ambiguities. [sent-112, score-1.207]

41 To resolve these ambiguities we add 360 degree orientation information to get a more accurate descriptor. [sent-113, score-0.184]

42 We are using 360 degree orientation, since curved lines cannot be fully described by their 180 degree orientation. [sent-114, score-0.179]

43 This is different to straight lines, where 180 degree orientation gives us the full information about the line. [sent-115, score-0.289]

44 The tangent line has an orientation between 0 and 180 degrees. [sent-117, score-0.193]

45 Therefore, using a 180 degree orientation yields to high similarities between a left curved line segment and a right curved line segment. [sent-119, score-0.435]

46 As a first step we extract the curvature information and the corresponding 360 degree orientation of all edge pixels in the image. [sent-120, score-0.725]

47 To estimate the curvature we follow our approach presented in [21] and use the distance accumulation method of Han et al. [sent-121, score-0.522]

48 The perpendicular distance Dik 4 Figure 2: Our visualization shows the original images along with their curvature self-similarity matrices displaying the similarity between all pairs of curvature histogram cells. [sent-128, score-1.018]

49 While curvature self-similarity descriptor is similar for the same object category it looks quite different to other object categories is computed from Li to the point pk , using the euclidean distance. [sent-129, score-1.023]

50 To get the 360 degree orientation information we compute the gradient of the probabilistic boundary edge image [20] and extend the resulting 180 degree gradient orientation to a 360 degree orientation using the sign of the curvature. [sent-132, score-0.69]

51 Contrary to the original curvature feature proposed in [21] where histograms of curvature are computed using differently sized image regions we build our basic curvature feature using equally sized cells to make it more suitable for computing self-similarities. [sent-133, score-1.915]

52 We divide the image into nonoverlapping 8 × 8 pixel cells and build histograms over the curvature values in each cell. [sent-134, score-0.661]

53 Next we do the same for the 360 degree orientation and concatenate the two histograms. [sent-135, score-0.184]

54 This results in histograms of 28 bins, 10 bins representing the curvature and 18 bins representing the 360 degree orientation. [sent-136, score-0.689]

55 We follow the scheme that was applied to compute self similarities between color histograms [29] and use histogram intersection as a comparison measure to compute the similarities between different curvature histograms in the same bounding box. [sent-138, score-0.808]

56 Let D be the number of cells in an image, then computing all pairwise similarities results in a D2 large curvature self-similarity matrix. [sent-141, score-0.542]

57 To discard such superfluous dimensions we apply our embedded feature selection method to the proposed curvature self-similarity representation. [sent-147, score-0.951]

58 4 Experiments We evaluate our curvature self-similarity descriptor in combination with the suggested embedded dimensionality reduction algorithm for the object detection task on the PASCAL dataset [7]. [sent-148, score-1.234]

59 Our experiments show, that curvature self-similarity is providing complementary information to straight lines, while our feature selection algorithm is further improving performance by fulfilling its natural need for dimensionality reduction. [sent-151, score-0.962]

60 The common basic concept shared by many current detection systems are high-dimensional, holistic representations learned with a discriminative classifier, mostly an SVM [28]. [sent-152, score-0.217]

61 These systems rely on high-dimensional holistic image statistics primarily utilizing straight line approximations. [sent-154, score-0.253]

62 In this paper we explore a orthogonal direction to these extensions and focus on how one can improve on the basic system by extending the straight line representation of HOG to a more discriminative description using curvature self-similarity. [sent-155, score-0.696]

63 At the same time our aim is to reduce the dimensionality 5 Table 1: Average precision of our iterative feature reduction algorithm for linear and non-linear kernel function using our final feature vector consisting of HOG+Curv+CurvSS. [sent-156, score-0.548]

64 For linear kernel function we compare our feature selection (linSVM+FS) to L2 normalized linear SVM (linSVM) and to the doubly regularized SVM (DrSVM) [30]. [sent-157, score-0.4]

65 For non-linear kernel function we compare the fast intersection kernel SVM (FIKSVM) [19] with our feature selection (FIKSVM+FS) linSVM DrSVM linSVM + FS FIKSVM FIKSVM + FS aero 66. [sent-158, score-0.469]

66 In the first part of our experiments we adjust the selection parameter λ of our iterative dimensionality reduction technique via cross-validation. [sent-264, score-0.303]

67 Furthermore, we compare the performance of our feature selection algorithm to L2 regularized SVM [3, 19] and DrSVM [30]. [sent-265, score-0.307]

68 In the second part we evaluate the suggested curvature self-similarity feature after applying our feature selection method to it. [sent-266, score-0.958]

69 1 Evaluation of Feature Selection All experiments in this section are performed using our final feature vector consisting of HOG, curvature (Curv) and curvature self-similarity (CurvSS). [sent-268, score-1.148]

70 We apply our iterative dimensionality reduction algorithm in combination with linear L2 regularized SVM classifier (linSVM) [3] and nonlinear fast intersection kernel SVM (FIKSVM) by Maji et al. [sent-269, score-0.295]

71 All training and validation data from the PASCAL VOC 2007 dataset are used to train an SVM using our final object representation on all positive samples and randomly chosen negative samples. [sent-275, score-0.257]

72 Out of the collected set of samples every tenth sample is assigned to the hold out test set which is used to compare the performance of our feature selection method. [sent-278, score-0.275]

73 The remaining samples are randomly split into training and validation set of equal size which are used to perform the feature selection. [sent-279, score-0.186]

74 For each set we train an L2 norm SVM on all samples from the training and validation set using only the remaining dimensions of the feature vector. [sent-281, score-0.234]

75 Since DrSVM is solving a similar optimization problem as our suggested feature selection algorithm for a linear kernel this comparison is of particular interest. [sent-285, score-0.385]

76 It is evident that objects like the sheep are not defined by their boundary shape and are thus beyond the scope of approaches base on contour shape it is performing feature selection in the higher dimensional kernel space rather than in the original feature space. [sent-289, score-0.799]

77 Instead we compare our feature selection method to that of FIKSVM for the nonlinear case. [sent-290, score-0.275]

78 Our feature selection method reduces the dimensionality of the feature by up to 55% for the linear case and by up to 40% in the non-linear case, while the performance in average precision is constant or increases beyond the performance of linSVM and FIKSVM. [sent-291, score-0.567]

79 On average our feature selection increases performance about 1. [sent-292, score-0.275]

80 Our results confirm that our feature selection method reduces the amount of noisy dimensions of highdimensional representations and therefore increases the average precision compared to an linear and non-linear SVM classifier without applying any feature selection. [sent-299, score-0.535]

81 For the linear kernel we showed furthermore that the proposed feature selection algorithm achieves gain over the DrSVM. [sent-300, score-0.329]

82 2 Object Detection using Curvature Self-Similarity In this section we provide a structured evaluation of the parts of our final object detection system. [sent-302, score-0.291]

83 [8, 9] as baseline system, since it is the basis for many powerful object detection systems. [sent-304, score-0.291]

84 All detection results are measured in terms of average precision performing object detection on the PASCAL VOC 2007 dataset. [sent-305, score-0.453]

85 To the best of our knowledge neither curvature nor self-similarity was used to perform object detection on a dataset of similar complexity as the PASCAL dataset so far. [sent-306, score-0.786]

86 [5] evaluated their global self-similarity descriptor (GSS) on the simpler classification challenge on the PASCAL VOC 2007 dataset, while the object detection evaluation was performed on the ETHZ shape dataset. [sent-308, score-0.461]

87 However, we showed in [21], that including curvature already solves the detection task almost perfectly on the ETHZ dataset. [sent-309, score-0.613]

88 Since the proposed approach models the shape of curved object contours and reduces the dimensionality of the representation, we expect it to be of particular value for objects that are characterized by their shape and where their contours can be extracted using state-of-the-art methods. [sent-312, score-0.548]

89 Therefore state-of-the-art edge extraction fails to provide any basis for contour based approaches on these images and one can therefore only expect a significant gain on categories where proper edge information can be computed for a majority of the images. [sent-315, score-0.189]

90 Due to the large amount of data in the PASCAL database the usage of intersection kernel for object detection becomes comparable intractable. [sent-318, score-0.394]

91 Results of our final system consisting of HOG, curvature (Curv), curvature self-similarity (CurvSS) and our embedded feature selection method (FS) are reported in terms of average precision in Table 2. [sent-319, score-1.468]

92 7 and without feature selection to show the individual gain of the curvature self-similarity descriptor and our embedded feature selection algorithm. [sent-450, score-1.306]

93 The results show that the suggested self-similarity representation in combination with feature selection improves performance on most of the categories. [sent-451, score-0.361]

94 One can observe that curvature information in combination with our feature selection algorithm is already improving performance over the HOG baseline and that adding curvature self-similarity additionally increases performance by 1. [sent-454, score-1.265]

95 The gain obtained by applying our feature selection (FS) depends obviously on the dimensionality of the feature vector; the higher the dimensionality the more can be gained by removing noisy dimensions. [sent-456, score-0.581]

96 For HOG+Curv applying our feature selection is improving performance by 0. [sent-457, score-0.275]

97 The results underline that curvature information provides complementary information to straight lines and that feature selection is needed when dealing with high dimensional features like self-similarity. [sent-459, score-0.909]

98 An embedded feature selection method for SVMs has therefore been proposed in this paper, which has been demonstrated to successfully deal with high-dimensional descriptions and it increases the performance of linear and intersection kernel SVM. [sent-461, score-0.511]

99 Moreover, the proposed curvature self-similarity representation has been shown to add complementary information to widely used orientation histograms. [sent-462, score-0.652]

100 Discriminative structure learning of hierarchical representations for object detection. [sent-627, score-0.209]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('curvature', 0.495), ('hog', 0.282), ('curv', 0.279), ('drsvm', 0.224), ('fiksvm', 0.205), ('linsvm', 0.205), ('object', 0.173), ('fs', 0.169), ('svm', 0.154), ('curvss', 0.149), ('selection', 0.143), ('embedded', 0.133), ('feature', 0.132), ('descriptor', 0.128), ('orientation', 0.127), ('detection', 0.118), ('pascal', 0.115), ('straight', 0.105), ('voc', 0.102), ('dimensionality', 0.087), ('histograms', 0.085), ('cutting', 0.077), ('dik', 0.066), ('curved', 0.065), ('bundle', 0.065), ('pi', 0.06), ('degree', 0.057), ('deselaers', 0.057), ('suggested', 0.056), ('gss', 0.056), ('validation', 0.054), ('kernel', 0.054), ('objects', 0.051), ('ethz', 0.049), ('guyon', 0.049), ('sheep', 0.049), ('uous', 0.049), ('boundary', 0.049), ('intersection', 0.049), ('dimensions', 0.048), ('descriptors', 0.047), ('similarities', 0.047), ('edge', 0.046), ('contours', 0.044), ('reduction', 0.044), ('precision', 0.044), ('image', 0.043), ('cvpr', 0.043), ('contour', 0.042), ('shape', 0.042), ('svms', 0.041), ('eccv', 0.039), ('doubly', 0.039), ('super', 0.039), ('build', 0.038), ('aero', 0.037), ('cow', 0.037), ('curvedness', 0.037), ('eigenstetter', 0.037), ('iwr', 0.037), ('lss', 0.037), ('mbike', 0.037), ('pers', 0.037), ('sofa', 0.037), ('line', 0.037), ('plane', 0.036), ('representations', 0.036), ('dimensional', 0.034), ('holistic', 0.034), ('utilizing', 0.034), ('bottle', 0.033), ('monroy', 0.033), ('ommer', 0.033), ('wrappers', 0.033), ('bus', 0.033), ('hci', 0.033), ('felzenszwalb', 0.032), ('regularized', 0.032), ('discarding', 0.032), ('yi', 0.031), ('bike', 0.03), ('representation', 0.03), ('discriminatively', 0.03), ('xl', 0.03), ('iterative', 0.029), ('discriminative', 0.029), ('beyond', 0.029), ('tangent', 0.029), ('boat', 0.028), ('contrary', 0.028), ('images', 0.028), ('categories', 0.027), ('accumulation', 0.027), ('plant', 0.027), ('nal', 0.027), ('pk', 0.027), ('consisting', 0.026), ('bins', 0.026), ('chair', 0.026), ('artifacts', 0.026)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000006 360 nips-2012-Visual Recognition using Embedded Feature Selection for Curvature Self-Similarity

Author: Angela Eigenstetter, Bjorn Ommer

Abstract: Category-level object detection has a crucial need for informative object representations. This demand has led to feature descriptors of ever increasing dimensionality like co-occurrence statistics and self-similarity. In this paper we propose a new object representation based on curvature self-similarity that goes beyond the currently popular approximation of objects using straight lines. However, like all descriptors using second order statistics, ours also exhibits a high dimensionality. Although improving discriminability, the high dimensionality becomes a critical issue due to lack of generalization ability and curse of dimensionality. Given only a limited amount of training data, even sophisticated learning algorithms such as the popular kernel methods are not able to suppress noisy or superfluous dimensions of such high-dimensional data. Consequently, there is a natural need for feature selection when using present-day informative features and, particularly, curvature self-similarity. We therefore suggest an embedded feature selection method for SVMs that reduces complexity and improves generalization capability of object models. By successfully integrating the proposed curvature self-similarity representation together with the embedded feature selection in a widely used state-of-the-art object detection framework we show the general pertinence of the approach. 1

2 0.16262116 106 nips-2012-Dynamical And-Or Graph Learning for Object Shape Modeling and Detection

Author: Xiaolong Wang, Liang Lin

Abstract: This paper studies a novel discriminative part-based model to represent and recognize object shapes with an “And-Or graph”. We define this model consisting of three layers: the leaf-nodes with collaborative edges for localizing local parts, the or-nodes specifying the switch of leaf-nodes, and the root-node encoding the global verification. A discriminative learning algorithm, extended from the CCCP [23], is proposed to train the model in a dynamical manner: the model structure (e.g., the configuration of the leaf-nodes associated with the or-nodes) is automatically determined with optimizing the multi-layer parameters during the iteration. The advantages of our method are two-fold. (i) The And-Or graph model enables us to handle well large intra-class variance and background clutters for object shape detection from images. (ii) The proposed learning algorithm is able to obtain the And-Or graph representation without requiring elaborate supervision and initialization. We validate the proposed method on several challenging databases (e.g., INRIA-Horse, ETHZ-Shape, and UIUC-People), and it outperforms the state-of-the-arts approaches. 1

3 0.16218752 344 nips-2012-Timely Object Recognition

Author: Sergey Karayev, Tobias Baumgartner, Mario Fritz, Trevor Darrell

Abstract: In a large visual multi-class detection framework, the timeliness of results can be crucial. Our method for timely multi-class detection aims to give the best possible performance at any single point after a start time; it is terminated at a deadline time. Toward this goal, we formulate a dynamic, closed-loop policy that infers the contents of the image in order to decide which detector to deploy next. In contrast to previous work, our method significantly diverges from the predominant greedy strategies, and is able to learn to take actions with deferred values. We evaluate our method with a novel timeliness measure, computed as the area under an Average Precision vs. Time curve. Experiments are conducted on the PASCAL VOC object detection dataset. If execution is stopped when only half the detectors have been run, our method obtains 66% better AP than a random ordering, and 14% better performance than an intelligent baseline. On the timeliness measure, our method obtains at least 11% better performance. Our method is easily extensible, as it treats detectors and classifiers as black boxes and learns from execution traces using reinforcement learning. 1

4 0.15714325 197 nips-2012-Learning with Recursive Perceptual Representations

Author: Oriol Vinyals, Yangqing Jia, Li Deng, Trevor Darrell

Abstract: Linear Support Vector Machines (SVMs) have become very popular in vision as part of state-of-the-art object recognition and other classification tasks but require high dimensional feature spaces for good performance. Deep learning methods can find more compact representations but current methods employ multilayer perceptrons that require solving a difficult, non-convex optimization problem. We propose a deep non-linear classifier whose layers are SVMs and which incorporates random projection as its core stacking element. Our method learns layers of linear SVMs recursively transforming the original data manifold through a random projection of the weak prediction computed from each layer. Our method scales as linear SVMs, does not rely on any kernel computations or nonconvex optimization, and exhibits better generalization ability than kernel-based SVMs. This is especially true when the number of training samples is smaller than the dimensionality of data, a common scenario in many real-world applications. The use of random projections is key to our method, as we show in the experiments section, in which we observe a consistent improvement over previous –often more complicated– methods on several vision and speech benchmarks. 1

5 0.13133442 176 nips-2012-Learning Image Descriptors with the Boosting-Trick

Author: Tomasz Trzcinski, Mario Christoudias, Vincent Lepetit, Pascal Fua

Abstract: In this paper we apply boosting to learn complex non-linear local visual feature representations, drawing inspiration from its successful application to visual object detection. The main goal of local feature descriptors is to distinctively represent a salient image region while remaining invariant to viewpoint and illumination changes. This representation can be improved using machine learning, however, past approaches have been mostly limited to learning linear feature mappings in either the original input or a kernelized input feature space. While kernelized methods have proven somewhat effective for learning non-linear local feature descriptors, they rely heavily on the choice of an appropriate kernel function whose selection is often difficult and non-intuitive. We propose to use the boosting-trick to obtain a non-linear mapping of the input to a high-dimensional feature space. The non-linear feature mapping obtained with the boosting-trick is highly intuitive. We employ gradient-based weak learners resulting in a learned descriptor that closely resembles the well-known SIFT. As demonstrated in our experiments, the resulting descriptor can be learned directly from intensity patches achieving state-of-the-art performance. 1

6 0.12357195 168 nips-2012-Kernel Latent SVM for Visual Recognition

7 0.11983132 303 nips-2012-Searching for objects driven by context

8 0.11150084 201 nips-2012-Localizing 3D cuboids in single-view images

9 0.098223425 40 nips-2012-Analyzing 3D Objects in Cluttered Images

10 0.093152583 101 nips-2012-Discriminatively Trained Sparse Code Gradients for Contour Detection

11 0.091857731 209 nips-2012-Max-Margin Structured Output Regression for Spatio-Temporal Action Localization

12 0.089517586 311 nips-2012-Shifting Weights: Adapting Object Detectors from Image to Video

13 0.088905431 1 nips-2012-3D Object Detection and Viewpoint Estimation with a Deformable 3D Cuboid Model

14 0.084707782 188 nips-2012-Learning from Distributions via Support Measure Machines

15 0.0829935 357 nips-2012-Unsupervised Template Learning for Fine-Grained Object Recognition

16 0.076986834 62 nips-2012-Burn-in, bias, and the rationality of anchoring

17 0.076986834 116 nips-2012-Emergence of Object-Selective Features in Unsupervised Feature Learning

18 0.075766653 179 nips-2012-Learning Manifolds with K-Means and K-Flats

19 0.075547054 228 nips-2012-Multilabel Classification using Bayesian Compressed Sensing

20 0.07097023 185 nips-2012-Learning about Canonical Views from Internet Image Collections


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.183), (1, 0.022), (2, -0.16), (3, -0.052), (4, 0.158), (5, -0.081), (6, 0.001), (7, -0.011), (8, 0.012), (9, -0.05), (10, -0.026), (11, 0.016), (12, 0.131), (13, -0.083), (14, 0.067), (15, 0.081), (16, -0.008), (17, -0.005), (18, -0.033), (19, 0.012), (20, 0.018), (21, -0.036), (22, 0.018), (23, -0.032), (24, -0.071), (25, 0.005), (26, -0.037), (27, 0.027), (28, -0.043), (29, -0.034), (30, -0.096), (31, 0.017), (32, -0.012), (33, 0.062), (34, 0.041), (35, 0.01), (36, 0.015), (37, 0.047), (38, 0.063), (39, -0.016), (40, 0.145), (41, 0.045), (42, 0.026), (43, 0.015), (44, 0.057), (45, -0.03), (46, -0.032), (47, -0.17), (48, -0.008), (49, 0.066)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.94517654 360 nips-2012-Visual Recognition using Embedded Feature Selection for Curvature Self-Similarity

Author: Angela Eigenstetter, Bjorn Ommer

Abstract: Category-level object detection has a crucial need for informative object representations. This demand has led to feature descriptors of ever increasing dimensionality like co-occurrence statistics and self-similarity. In this paper we propose a new object representation based on curvature self-similarity that goes beyond the currently popular approximation of objects using straight lines. However, like all descriptors using second order statistics, ours also exhibits a high dimensionality. Although improving discriminability, the high dimensionality becomes a critical issue due to lack of generalization ability and curse of dimensionality. Given only a limited amount of training data, even sophisticated learning algorithms such as the popular kernel methods are not able to suppress noisy or superfluous dimensions of such high-dimensional data. Consequently, there is a natural need for feature selection when using present-day informative features and, particularly, curvature self-similarity. We therefore suggest an embedded feature selection method for SVMs that reduces complexity and improves generalization capability of object models. By successfully integrating the proposed curvature self-similarity representation together with the embedded feature selection in a widely used state-of-the-art object detection framework we show the general pertinence of the approach. 1

2 0.70814574 106 nips-2012-Dynamical And-Or Graph Learning for Object Shape Modeling and Detection

Author: Xiaolong Wang, Liang Lin

Abstract: This paper studies a novel discriminative part-based model to represent and recognize object shapes with an “And-Or graph”. We define this model consisting of three layers: the leaf-nodes with collaborative edges for localizing local parts, the or-nodes specifying the switch of leaf-nodes, and the root-node encoding the global verification. A discriminative learning algorithm, extended from the CCCP [23], is proposed to train the model in a dynamical manner: the model structure (e.g., the configuration of the leaf-nodes associated with the or-nodes) is automatically determined with optimizing the multi-layer parameters during the iteration. The advantages of our method are two-fold. (i) The And-Or graph model enables us to handle well large intra-class variance and background clutters for object shape detection from images. (ii) The proposed learning algorithm is able to obtain the And-Or graph representation without requiring elaborate supervision and initialization. We validate the proposed method on several challenging databases (e.g., INRIA-Horse, ETHZ-Shape, and UIUC-People), and it outperforms the state-of-the-arts approaches. 1

3 0.701563 101 nips-2012-Discriminatively Trained Sparse Code Gradients for Contour Detection

Author: Ren Xiaofeng, Liefeng Bo

Abstract: Finding contours in natural images is a fundamental problem that serves as the basis of many tasks such as image segmentation and object recognition. At the core of contour detection technologies are a set of hand-designed gradient features, used by most approaches including the state-of-the-art Global Pb (gPb) operator. In this work, we show that contour detection accuracy can be significantly improved by computing Sparse Code Gradients (SCG), which measure contrast using patch representations automatically learned through sparse coding. We use K-SVD for dictionary learning and Orthogonal Matching Pursuit for computing sparse codes on oriented local neighborhoods, and apply multi-scale pooling and power transforms before classifying them with linear SVMs. By extracting rich representations from pixels and avoiding collapsing them prematurely, Sparse Code Gradients effectively learn how to measure local contrasts and find contours. We improve the F-measure metric on the BSDS500 benchmark to 0.74 (up from 0.71 of gPb contours). Moreover, our learning approach can easily adapt to novel sensor data such as Kinect-style RGB-D cameras: Sparse Code Gradients on depth maps and surface normals lead to promising contour detection using depth and depth+color, as verified on the NYU Depth Dataset. 1

4 0.69732261 357 nips-2012-Unsupervised Template Learning for Fine-Grained Object Recognition

Author: Shulin Yang, Liefeng Bo, Jue Wang, Linda G. Shapiro

Abstract: Fine-grained recognition refers to a subordinate level of recognition, such as recognizing different species of animals and plants. It differs from recognition of basic categories, such as humans, tables, and computers, in that there are global similarities in shape and structure shared cross different categories, and the differences are in the details of object parts. We suggest that the key to identifying the fine-grained differences lies in finding the right alignment of image regions that contain the same object parts. We propose a template model for the purpose, which captures common shape patterns of object parts, as well as the cooccurrence relation of the shape patterns. Once the image regions are aligned, extracted features are used for classification. Learning of the template model is efficient, and the recognition results we achieve significantly outperform the stateof-the-art algorithms. 1

5 0.66462445 168 nips-2012-Kernel Latent SVM for Visual Recognition

Author: Weilong Yang, Yang Wang, Arash Vahdat, Greg Mori

Abstract: Latent SVMs (LSVMs) are a class of powerful tools that have been successfully applied to many applications in computer vision. However, a limitation of LSVMs is that they rely on linear models. For many computer vision tasks, linear models are suboptimal and nonlinear models learned with kernels typically perform much better. Therefore it is desirable to develop the kernel version of LSVM. In this paper, we propose kernel latent SVM (KLSVM) – a new learning framework that combines latent SVMs and kernel methods. We develop an iterative training algorithm to learn the model parameters. We demonstrate the effectiveness of KLSVM using three different applications in visual recognition. Our KLSVM formulation is very general and can be applied to solve a wide range of applications in computer vision and machine learning. 1

6 0.65187889 176 nips-2012-Learning Image Descriptors with the Boosting-Trick

7 0.65047634 201 nips-2012-Localizing 3D cuboids in single-view images

8 0.63172477 1 nips-2012-3D Object Detection and Viewpoint Estimation with a Deformable 3D Cuboid Model

9 0.62496269 344 nips-2012-Timely Object Recognition

10 0.61993474 303 nips-2012-Searching for objects driven by context

11 0.59632033 146 nips-2012-Graphical Gaussian Vector for Image Categorization

12 0.57618231 197 nips-2012-Learning with Recursive Perceptual Representations

13 0.5671109 209 nips-2012-Max-Margin Structured Output Regression for Spatio-Temporal Action Localization

14 0.55220377 210 nips-2012-Memorability of Image Regions

15 0.53287804 87 nips-2012-Convolutional-Recursive Deep Learning for 3D Object Classification

16 0.52878559 40 nips-2012-Analyzing 3D Objects in Cluttered Images

17 0.52028948 188 nips-2012-Learning from Distributions via Support Measure Machines

18 0.52027345 289 nips-2012-Recognizing Activities by Attribute Dynamics

19 0.51931518 306 nips-2012-Semantic Kernel Forests from Multiple Taxonomies

20 0.51481128 48 nips-2012-Augmented-SVM: Automatic space partitioning for combining multiple non-linear dynamics


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(0, 0.024), (21, 0.014), (38, 0.077), (39, 0.027), (42, 0.019), (54, 0.022), (55, 0.019), (74, 0.532), (76, 0.099), (80, 0.054), (92, 0.032)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.90898186 202 nips-2012-Locally Uniform Comparison Image Descriptor

Author: Andrew Ziegler, Eric Christiansen, David Kriegman, Serge J. Belongie

Abstract: Keypoint matching between pairs of images using popular descriptors like SIFT or a faster variant called SURF is at the heart of many computer vision algorithms including recognition, mosaicing, and structure from motion. However, SIFT and SURF do not perform well for real-time or mobile applications. As an alternative very fast binary descriptors like BRIEF and related methods use pairwise comparisons of pixel intensities in an image patch. We present an analysis of BRIEF and related approaches revealing that they are hashing schemes on the ordinal correlation metric Kendall’s tau. Here, we introduce Locally Uniform Comparison Image Descriptor (LUCID), a simple description method based on linear time permutation distances between the ordering of RGB values of two image patches. LUCID is computable in linear time with respect to the number of pixels and does not require floating point computation. 1

same-paper 2 0.90278685 360 nips-2012-Visual Recognition using Embedded Feature Selection for Curvature Self-Similarity

Author: Angela Eigenstetter, Bjorn Ommer

Abstract: Category-level object detection has a crucial need for informative object representations. This demand has led to feature descriptors of ever increasing dimensionality like co-occurrence statistics and self-similarity. In this paper we propose a new object representation based on curvature self-similarity that goes beyond the currently popular approximation of objects using straight lines. However, like all descriptors using second order statistics, ours also exhibits a high dimensionality. Although improving discriminability, the high dimensionality becomes a critical issue due to lack of generalization ability and curse of dimensionality. Given only a limited amount of training data, even sophisticated learning algorithms such as the popular kernel methods are not able to suppress noisy or superfluous dimensions of such high-dimensional data. Consequently, there is a natural need for feature selection when using present-day informative features and, particularly, curvature self-similarity. We therefore suggest an embedded feature selection method for SVMs that reduces complexity and improves generalization capability of object models. By successfully integrating the proposed curvature self-similarity representation together with the embedded feature selection in a widely used state-of-the-art object detection framework we show the general pertinence of the approach. 1

3 0.89857531 40 nips-2012-Analyzing 3D Objects in Cluttered Images

Author: Mohsen Hejrati, Deva Ramanan

Abstract: We present an approach to detecting and analyzing the 3D configuration of objects in real-world images with heavy occlusion and clutter. We focus on the application of finding and analyzing cars. We do so with a two-stage model; the first stage reasons about 2D shape and appearance variation due to within-class variation (station wagons look different than sedans) and changes in viewpoint. Rather than using a view-based model, we describe a compositional representation that models a large number of effective views and shapes using a small number of local view-based templates. We use this model to propose candidate detections and 2D estimates of shape. These estimates are then refined by our second stage, using an explicit 3D model of shape and viewpoint. We use a morphable model to capture 3D within-class variation, and use a weak-perspective camera model to capture viewpoint. We learn all model parameters from 2D annotations. We demonstrate state-of-the-art accuracy for detection, viewpoint estimation, and 3D shape reconstruction on challenging images from the PASCAL VOC 2011 dataset. 1

4 0.84021997 337 nips-2012-The Lovász ϑ function, SVMs and finding large dense subgraphs

Author: Vinay Jethava, Anders Martinsson, Chiranjib Bhattacharyya, Devdatt Dubhashi

Abstract: The Lov´ sz ϑ function of a graph, a fundamental tool in combinatorial optimizaa tion and approximation algorithms, is computed by solving a SDP. In this paper we establish that the Lov´ sz ϑ function is equivalent to a kernel learning problem a related to one class SVM. This interesting connection opens up many opportunities bridging graph theoretic algorithms and machine learning. We show that there exist graphs, which we call SVM − ϑ graphs, on which the Lov´ sz ϑ function a can be approximated well by a one-class SVM. This leads to novel use of SVM techniques for solving algorithmic problems in large graphs e.g. identifying a √ 1 planted clique of size Θ( n) in a random graph G(n, 2 ). A classic approach for this problem involves computing the ϑ function, however it is not scalable due to SDP computation. We show that the random graph with a planted clique is an example of SVM − ϑ graph. As a consequence a SVM based approach easily identifies the clique in large graphs and is competitive with the state-of-the-art. We introduce the notion of common orthogonal labelling and show that it can be computed by solving a Multiple Kernel learning problem. It is further shown that such a labelling is extremely useful in identifying a large common dense subgraph in multiple graphs, which is known to be a computationally difficult problem. The proposed algorithm achieves an order of magnitude scalability compared to state of the art methods. 1

5 0.80957139 339 nips-2012-The Time-Marginalized Coalescent Prior for Hierarchical Clustering

Author: Levi Boyles, Max Welling

Abstract: We introduce a new prior for use in Nonparametric Bayesian Hierarchical Clustering. The prior is constructed by marginalizing out the time information of Kingman’s coalescent, providing a prior over tree structures which we call the Time-Marginalized Coalescent (TMC). This allows for models which factorize the tree structure and times, providing two benefits: more flexible priors may be constructed and more efficient Gibbs type inference can be used. We demonstrate this on an example model for density estimation and show the TMC achieves competitive experimental results. 1

6 0.76053518 3 nips-2012-A Bayesian Approach for Policy Learning from Trajectory Preference Queries

7 0.6898917 357 nips-2012-Unsupervised Template Learning for Fine-Grained Object Recognition

8 0.68886524 185 nips-2012-Learning about Canonical Views from Internet Image Collections

9 0.6843105 274 nips-2012-Priors for Diversity in Generative Latent Variable Models

10 0.67061681 201 nips-2012-Localizing 3D cuboids in single-view images

11 0.64788312 176 nips-2012-Learning Image Descriptors with the Boosting-Trick

12 0.63625228 8 nips-2012-A Generative Model for Parts-based Object Segmentation

13 0.62550831 210 nips-2012-Memorability of Image Regions

14 0.61600626 101 nips-2012-Discriminatively Trained Sparse Code Gradients for Contour Detection

15 0.606713 106 nips-2012-Dynamical And-Or Graph Learning for Object Shape Modeling and Detection

16 0.59927386 235 nips-2012-Natural Images, Gaussian Mixtures and Dead Leaves

17 0.59121275 303 nips-2012-Searching for objects driven by context

18 0.5748145 146 nips-2012-Graphical Gaussian Vector for Image Categorization

19 0.56754929 168 nips-2012-Kernel Latent SVM for Visual Recognition

20 0.55183691 81 nips-2012-Context-Sensitive Decision Forests for Object Detection