nips nips2001 nips2001-191 knowledge-graph by maker-knowledge-mining

191 nips-2001-Transform-invariant Image Decomposition with Similarity Templates


Source: pdf

Author: Chris Stauffer, Erik Miller, Kinh Tieu

Abstract: Recent work has shown impressive transform-invariant modeling and clustering for sets of images of objects with similar appearance. We seek to expand these capabilities to sets of images of an object class that show considerable variation across individual instances (e.g. pedestrian images) using a representation based on pixel-wise similarities, similarity templates. Because of its invariance to the colors of particular components of an object, this representation enables detection of instances of an object class and enables alignment of those instances. Further, this model implicitly represents the regions of color regularity in the class-specific image set enabling a decomposition of that object class into component regions. 1

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 Transform-invariant image decomposition with similarity templates Chris Stauffer, Erik Miller, and Kinh Tieu MIT Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 {stauffer,emiller,tieu}@ai. [sent-1, score-0.633]

2 edu Abstract Recent work has shown impressive transform-invariant modeling and clustering for sets of images of objects with similar appearance. [sent-3, score-0.322]

3 We seek to expand these capabilities to sets of images of an object class that show considerable variation across individual instances (e. [sent-4, score-0.358]

4 pedestrian images) using a representation based on pixel-wise similarities, similarity templates. [sent-6, score-0.562]

5 Because of its invariance to the colors of particular components of an object, this representation enables detection of instances of an object class and enables alignment of those instances. [sent-7, score-0.507]

6 Further, this model implicitly represents the regions of color regularity in the class-specific image set enabling a decomposition of that object class into component regions. [sent-8, score-0.833]

7 In particular, we are interested in modeling classes of objects that are characterized by similarities and differences between image pixels rather than by the values of those pixels. [sent-10, score-0.486]

8 For instance, images of pedestrians (at a certain scale and pose) can be characterized by a few regions of regularity (RORs) such as shirt, pants, background, and head, that have fixed properties such as constant color or constant texture within the region, but tend to be different from each other. [sent-11, score-1.113]

9 The particular color (or texture) of those regions is largely irrelevant. [sent-12, score-0.528]

10 We shall refer to sets of images that fit this general description as images characterized by regions of regularity, or ICRORs. [sent-13, score-0.557]

11 Jojic and Frey [1] and others [2] have investigated transform-invariant modeling and clustering for images of a particular object (e. [sent-14, score-0.348]

12 Their work with transformed component analysis (TCA) shows promise for handling considerable variation within the images resulting from lighting or slight misalignments. [sent-19, score-0.307]

13 However, because these models rely on an image set with a fixed mean or mixture of means, they are not directly applicable to ICRORs. [sent-20, score-0.164]

14 We would also like to address transform-invariant modeling, but use a model which is invariant to the particular color of component regions. [sent-21, score-0.416]

15 One simple way to achieve this is to use edge templates to model local differences in image color. [sent-22, score-0.298]

16 In contrast, we have chosen to model global similarities in color using a similarity template (ST). [sent-23, score-0.918]

17 While representations of pixel similarity have previously been exploited for segmentation of single images [3, 4], we have chosen to use them for aggregate modeling of image sets. [sent-24, score-1.121]

18 Similarity templates enable alignment of image sets and decomposition of images into class-specific pixel regions. [sent-25, score-0.897]

19 We note also that registration of two ICRORs can be accomplished by minimizing the mutual information between corresponding pixels [5]. [sent-26, score-0.18]

20 But, there is no obvious way of extending this method to large sets of images without a combinatorial explosion. [sent-27, score-0.205]

21 Section 4 covers their application to decomposing a class-specific set of images into component regions. [sent-31, score-0.244]

22 2 Similarity templates This section begins with a brief explanation of the similarity template followed by the mechanics of computing and comparing similarity templates. [sent-33, score-0.905]

23 A similarity template S for an N -pixel image is an N xN matrix. [sent-34, score-0.619]

24 The element Si,j represents the probability that pixel locations pi and pj would result from choosing a region and drawing (iid) two samples (pixel locations) from it. [sent-35, score-0.81]

25 More formally, Si,j = p(r)p(pi |r)p(pj |r), (1) r where p(r) is the probability of choosing region r and p(pi |r) is the probability of choosing pixel location pi from region r. [sent-36, score-0.89]

26 1 The “ideal” similarity template Consider sampling pixel pairs as described above from an N -pixel image of a particular object (e. [sent-38, score-0.931]

27 , a pedestrian) segmented by an oracle into disjoint regions (e. [sent-40, score-0.143]

28 Assuming each region is equally likely to be sampled and that the pixels in the region are selected with uniform probability, then 1 1 ( R )( Sr )2 if ri = rj Si,j = (2) 0 otherwise, where R is the number of regions, Sr is the number of pixels in region r, and ri is the region label of pi . [sent-43, score-1.425]

29 If two pixels are from the same region, the corresponding 1 value is the product of the probability R of choosing a particular region and the 1 2 probability ( Sr ) of drawing that pixel pair. [sent-44, score-0.709]

30 This can be interpreted as a block diagonal co-occurrence matrix of sampled pixel pairs. [sent-45, score-0.237]

31 In this ideal case, two images of different pedestrians with the same body size and shape would result in the same similarity template regardless of the colors of their clothes, since the ST is a function only of the segmentation. [sent-46, score-1.122]

32 An ST of an image without a pedestrian would exhibit different statistics. [sent-47, score-0.402]

33 Note that even the ST of an image of a blank wall (segmented as a single region) would be different because pixels that are in different regions under the ideal pedestrian ST would be in the same region. [sent-48, score-0.746]

34 Unfortunately, images do not typically come with labeled regions, and so computation of a similarity template is impossible. [sent-49, score-0.681]

35 Using this observation, we can approximate true similarity templates from unsegmented images. [sent-51, score-0.429]

36 2 Computing similarity templates For the purposes of this paper, our model for similarity is based solely on color. [sent-53, score-0.703]

37 If each latent region had a constant but unique color and the regions were of equal 2 size, then as σi approaches zero this process reconstructs the “ideal” similarity template defined in Equation 1. [sent-57, score-1.281]

38 Although region colors are neither constant nor unique, this approximation has proven to work well in practice. [sent-58, score-0.412]

39 It is possible to add a spatial prior based on the relative pixel location to model the fact that similarities tend to local, but we will rely on the statistics of the images in our data set to determine whether (and to what extent) this is the case. [sent-59, score-0.529]

40 Also, it may be possible to achieve better results using a more complex color model (e. [sent-60, score-0.395]

41 , hsv with full covariance) or broadening the measure of similarity to include other modalities (e. [sent-62, score-0.308]

42 Figure 1 shows two views of the same similarity template. [sent-66, score-0.274]

43 The first view represents each pixel’s similarity to every other pixel. [sent-67, score-0.319]

44 The second view contains a sub-image for each pixel which highlights the pixels that are most likely produced by the same region. [sent-68, score-0.475]

45 Pixels in the shirt tend to highlight the entire shirt and the pants (to a lesser amount). [sent-69, score-0.572]

46 Pixels in the background tend to be very dissimilar to all pixels in the foreground. [sent-70, score-0.275]

47 3 Aggregate similarity templates (AST) We assume each estimated ST is a noisy measurement of the true underlying joint distribution. [sent-72, score-0.455]

48 Hence we compute an aggregate similarity template (AST) as the ¯ mean S of the ST estimates over an entire class-specific set of K images: 1 ¯ Si,j = K K ˜k Si,j . [sent-73, score-0.625]

49 Note that this is a less restrictive assumption than assuming edges of regions are in correspondence across an image set, since regions have greater support. [sent-75, score-0.394]

50 (a) (b) Figure 1: (a) The N xN aggregate similarity template for pedestrian data set. [sent-77, score-0.884]

51 Each sub-image highlights the pixels that are most similar to the pixel it represents. [sent-81, score-0.454]

52 4 Comparing similarity templates ˜ To compare an estimated similarity template S to an aggregate similarity template ¯ S we evaluate their dot product1 : ¯ ˜ s(S, S) = ¯ ˜ Si,j Si,j . [sent-83, score-1.53]

53 By thresholding the ratio of the dot product of a particular image patch under and AST trained on pedestrian image patches versus an AST trained on random image patches, we can determine whether a person is present in the image. [sent-85, score-0.737]

54 3 Data set alignment In this paper, we investigate a more difficult problem: alignment of a set of images. [sent-87, score-0.192]

55 To explore this problem, we created a set of 128x64 images of simulated pedestrians. [sent-88, score-0.205]

56 These pedestrians were generated by creating four independently-colored regions corresponding to shirts, pants, head, and background. [sent-89, score-0.316]

57 Then, independent Gaussian noise was added to each pixel (σ = . [sent-92, score-0.237]

58 Finally the images were translated uniformly up to 25% of the size of the object. [sent-94, score-0.205]

59 Figure 2: A set of randomly generated “pedestrian” images used in alignment experimetns. [sent-97, score-0.301]

60 [2], we iteratively estimated the latent variables (translations) that maximized the probability of the image STs to the AST and re-estimated the AST. [sent-99, score-0.214]

61 4 Decomposing the similarity template This section explains how to derive a factorized representation from the AST that will be useful for recognition of particular instances of a class and for further refinement of detection. [sent-102, score-0.554]

62 This representation is also useful in approximating the template to avoid the O(N 2 ) storage requirements. [sent-103, score-0.231]

63 An AST represents the similarity of pixels within an image across an entire classspecific data set. [sent-104, score-0.67]

64 Also, rather than treating pixel brightness (darkness, redness, blueness, or hue) as a value to be reconstructed in the decomposition, we chose to represent pixel similarity. [sent-112, score-0.494]

65 In contrast to simply treating images as additive mixtures of basis functions [9], our decomposition will get the same results on a database of images of digits written in black on white paper or in white on a black board and color images introduce no difficulties for our methods. [sent-113, score-1.225]

66 Given the number of regions R, it is possible to estimate the priors for each region p(r) and the probability of each region producing each pixel p(pi |r). [sent-116, score-0.803]

67 Because our model S is symmetric, this case can be updated with only two rules: ˆ S(pi , pj ) pnew (pi |r) ∝ p(pi |r) p(r)p(pj |r) ¯ , and (7) S(pi , pj ) p j pnew (r) ∝ p(r) pi pj ˆ S(pi , pj ) p(pj |r)p(pi |r) ¯ . [sent-118, score-0.715]

68 S(pi , pj ) (8) 50 100 150 200 250 300 350 400 450 500 50 100 150 200 250 300 350 400 450 500 Figure 3: The similarity template and the corresponding automatically generated binary decomposition of the images in the pedestrian data set. [sent-119, score-1.122]

69 The root node represents every pixel in the image. [sent-120, score-0.261]

70 The more underlying regions we allow our model, the closer our estimate will approximate the true joint distribution. [sent-124, score-0.138]

71 These region models tend to represent parts of the object class. [sent-125, score-0.321]

72 p(pi |r) will tend to have high probabilities for a set of pixels belonging to the same region. [sent-126, score-0.22]

73 We take advantage of the fact that aligned pedestrian images are symmetric about the vertical axis by adding a “reflected” aggregate similarity template to the aggregate similarity template. [sent-127, score-1.56]

74 Rather than performing a straight R-way decomposition of the AST to obtain R pixel region models, we extracted a hierarchical segmentation in the form of a binary tree. [sent-129, score-0.63]

75 Given the initial region-conditioned marginals p(pi |r0 ) and p(pi |r1 ), each pixel was assigned to the region with higher likelihood. [sent-130, score-0.501]

76 For instance, the mean pixel value can be calculated as a weighted mean where the pixels are weighted by p(pi |r). [sent-135, score-0.417]

77 Figure 3 shows the resulting hierarchical segmentation for the pedestrian AST. [sent-138, score-0.364]

78 In our experience, a color histogram of all the pixels within a pedestrian is not useful for recognition and was almost useless for data mining applications. [sent-140, score-0.856]

79 It determines a color model over each region that our algorithm has determined contain similar color information within this class of objects. [sent-142, score-1.039]

80 This allows us to obtain robust estimates of color in the regions Figure 4: Results of automatic clustering on three components: shirt, pants, and the background. [sent-143, score-0.543]

81 Each shows the feature, the most unusual examples of that region, followed by the 12 most likely examples for the eight prototypical colors of that region. [sent-144, score-0.255]

82 Further, as a result of our probabilistic segmentation, the values of p(pi |r) indicate which pixels are most regular in a region which enables us to weight the contribution of each pixel to the color model. [sent-146, score-1.07]

83 For the case of pedestrian-conditional color models, the regions roughly correspond to shirt color, pant color, feet color, head color, and some background color regions. [sent-147, score-1.327]

84 The colors in a region of a single image can be modeled by color histograms, Gaussians, or mixtures of Gaussians. [sent-148, score-0.95]

85 These region models can be clustered across images to determine a density of shirt colors, pant colors, and other region colors within a particular environment. [sent-149, score-1.191]

86 This enables not only an efficient factored color component codebook, but anomaly detection based on particular regions and higher order models of co-occurrences between particular types of regions. [sent-150, score-0.612]

87 To illustrate the effectiveness of our representation we chose the simplest model for the colors in each region–a single Gaussian in RGB space. [sent-151, score-0.214]

88 The mean and variance of each Gaussian was computed by weighting the pixels represented by the corresponding node by p(pi |r). [sent-152, score-0.18]

89 This biases the estimate towards the “most similar” pixels in the region (e. [sent-153, score-0.407]

90 , the center of the shirt or the center of the legs). [sent-155, score-0.213]

91 This allows us to represent the colors of each pedestrian image with 31 means and variances corresponding to the (2treeheight − 1) nodes. [sent-156, score-0.587]

92 We investigated unsupervised clustering on components of the conditional color model. [sent-157, score-0.431]

93 We fit a mixture of eight Gaussians to the 924 color means for each region. [sent-158, score-0.449]

94 Figure 4 shows the 12 pedestrians with the highest probability under each of the eight models and the 12 most unusual pedestrians with respect to that region for three of the nodes of the tree: shirt color, pant color, and color of the background. [sent-159, score-1.377]

95 In the future, similarity templates could be applied to different modalities including texture similarity, depth similarity, or motion similarity. [sent-166, score-0.563]

96 While computationally intensive, we believe that similarity templates can provide a unified approach to the extraction of possible class-specific targets from an image database, alignment of the candidate images, and precomputation of meaningful features of that class. [sent-167, score-0.668]

97 For the case of pedestrians, it could detect potential pedestrians in a database, align them, derive a model of pedestrians, and extract the parameters for each pedestrian. [sent-168, score-0.248]

98 We have introduced a new image representation based on pixel-wise similarity. [sent-170, score-0.172]

99 We have shown its application in both alignment and decomposition of pedestrian images. [sent-171, score-0.416]

100 “Similarity templates for detection and recognition,” submitted to CVPR (2001). [sent-210, score-0.187]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('color', 0.395), ('similarity', 0.274), ('pedestrian', 0.259), ('pixel', 0.237), ('region', 0.227), ('shirt', 0.213), ('images', 0.205), ('pedestrians', 0.204), ('ast', 0.202), ('template', 0.202), ('colors', 0.185), ('pixels', 0.18), ('pi', 0.157), ('templates', 0.155), ('aggregate', 0.149), ('image', 0.143), ('pj', 0.121), ('regions', 0.112), ('pants', 0.106), ('alignment', 0.096), ('stau', 0.085), ('di', 0.082), ('segmentation', 0.081), ('st', 0.069), ('erent', 0.066), ('pant', 0.064), ('decomposition', 0.061), ('head', 0.056), ('texture', 0.056), ('background', 0.055), ('object', 0.054), ('ideal', 0.052), ('latent', 0.051), ('objects', 0.049), ('similarities', 0.047), ('regularity', 0.044), ('cvpr', 0.044), ('align', 0.044), ('rgb', 0.043), ('rors', 0.043), ('tend', 0.04), ('sr', 0.039), ('decomposing', 0.039), ('marginals', 0.037), ('highlights', 0.037), ('legs', 0.037), ('shirts', 0.037), ('feet', 0.037), ('pnew', 0.037), ('unusual', 0.037), ('miller', 0.036), ('clustering', 0.036), ('lab', 0.035), ('characterized', 0.035), ('modalities', 0.034), ('ectively', 0.034), ('viola', 0.034), ('jojic', 0.034), ('cbcl', 0.034), ('erences', 0.034), ('eight', 0.033), ('june', 0.033), ('detection', 0.032), ('database', 0.032), ('modeling', 0.032), ('segmented', 0.031), ('lighting', 0.031), ('enables', 0.031), ('representation', 0.029), ('instances', 0.028), ('promise', 0.028), ('patches', 0.028), ('cuts', 0.027), ('aligned', 0.027), ('across', 0.027), ('black', 0.026), ('zi', 0.026), ('joint', 0.026), ('white', 0.025), ('blue', 0.025), ('factorization', 0.025), ('hierarchical', 0.024), ('depth', 0.024), ('represents', 0.024), ('drawing', 0.023), ('seek', 0.023), ('within', 0.022), ('collected', 0.022), ('xn', 0.022), ('view', 0.021), ('choosing', 0.021), ('particular', 0.021), ('considerable', 0.021), ('mixture', 0.021), ('er', 0.021), ('symmetric', 0.021), ('treating', 0.02), ('motion', 0.02), ('unique', 0.02), ('iteratively', 0.02)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.99999875 191 nips-2001-Transform-invariant Image Decomposition with Similarity Templates

Author: Chris Stauffer, Erik Miller, Kinh Tieu

Abstract: Recent work has shown impressive transform-invariant modeling and clustering for sets of images of objects with similar appearance. We seek to expand these capabilities to sets of images of an object class that show considerable variation across individual instances (e.g. pedestrian images) using a representation based on pixel-wise similarities, similarity templates. Because of its invariance to the colors of particular components of an object, this representation enables detection of instances of an object class and enables alignment of those instances. Further, this model implicitly represents the regions of color regularity in the class-specific image set enabling a decomposition of that object class into component regions. 1

2 0.14860475 89 nips-2001-Grouping with Bias

Author: Stella X. Yu, Jianbo Shi

Abstract: With the optimization of pattern discrimination as a goal, graph partitioning approaches often lack the capability to integrate prior knowledge to guide grouping. In this paper, we consider priors from unitary generative models, partially labeled data and spatial attention. These priors are modelled as constraints in the solution space. By imposing uniformity condition on the constraints, we restrict the feasible space to one of smooth solutions. A subspace projection method is developed to solve this constrained eigenproblema We demonstrate that simple priors can greatly improve image segmentation results. 1

3 0.13088034 46 nips-2001-Categorization by Learning and Combining Object Parts

Author: Bernd Heisele, Thomas Serre, Massimiliano Pontil, Thomas Vetter, Tomaso Poggio

Abstract: We describe an algorithm for automatically learning discriminative components of objects with SVM classifiers. It is based on growing image parts by minimizing theoretical bounds on the error probability of an SVM. Component-based face classifiers are then combined in a second stage to yield a hierarchical SVM classifier. Experimental results in face classification show considerable robustness against rotations in depth and suggest performance at significantly better level than other face detection systems. Novel aspects of our approach are: a) an algorithm to learn component-based classification experts and their combination, b) the use of 3-D morphable models for training, and c) a maximum operation on the output of each component classifier which may be relevant for biological models of visual recognition.

4 0.10623703 34 nips-2001-Analog Soft-Pattern-Matching Classifier using Floating-Gate MOS Technology

Author: Toshihiko Yamasaki, Tadashi Shibata

Abstract: A flexible pattern-matching analog classifier is presented in conjunction with a robust image representation algorithm called Principal Axes Projection (PAP). In the circuit, the functional form of matching is configurable in terms of the peak position, the peak height and the sharpness of the similarity evaluation. The test chip was fabricated in a 0.6-µm CMOS technology and successfully applied to hand-written pattern recognition and medical radiograph analysis using PAP as a feature extraction pre-processing step for robust image coding. The separation and classification of overlapping patterns is also experimentally demonstrated. 1 I ntr o du c ti o n Pattern classification using template matching techniques is a powerful tool in implementing human-like intelligent systems. However, the processing is computationally very expensive, consuming a lot of CPU time when implemented as software running on general-purpose computers. Therefore, software approaches are not practical for real-time applications. For systems working in mobile environment, in particular, they are not realistic because the memory and computational resources are severely limited. The development of analog VLSI chips having a fully parallel template matching architecture [1,2] would be a promising solution in such applications because they offer an opportunity of low-power operation as well as very compact implementation. In order to build a real human-like intelligent system, however, not only the pattern representation algorithm but also the matching hardware itself needs to be made flexible and robust in carrying out the pattern matching task. First of all, two-dimensional patterns need to be represented by feature vectors having substantially reduced dimensions, while at the same time preserving the human perception of similarity among patterns in the vector space mapping. For this purpose, an image representation algorithm called Principal Axes Projection (PAP) has been de- veloped [3] and its robust nature in pattern recognition has been demonstrated in the applications to medical radiograph analysis [3] and hand-written digits recognition [4]. However, the demonstration so far was only carried out by computer simulation. Regarding the matching hardware, high-flexibility analog template matching circuits have been developed for PAP vector representation. The circuits are flexible in a sense that the matching criteria (the weight to elements, the strictness in matching) are configurable. In Ref. [5], the fundamental characteristics of the building block circuits were presented, and their application to simple hand-written digits was presented in Ref. [6]. The purpose of this paper is to demonstrate the robust nature of the hardware matching system by experiments. The classification of simple hand-written patterns and the cephalometric landmark identification in gray-scale medical radiographs have been carried out and successful results are presented. In addition, multiple overlapping patterns can be separated without utilizing a priori knowledge, which is one of the most difficult problems at present in artificial intelligence. 2 I ma g e re pr es e n tati on by P AP PAP is a feature extraction technique using the edge information. The input image (64x64 pixels) is first subjected to pixel-by-pixel spatial filtering operations to detect edges in four directions: horizontal (HR); vertical (VR); +45 degrees (+45); and –45 degrees (-45). Each detected edge is represented by a binary flag and four edge maps are generated. The two-dimensional bit array in an edge map is reduced to a one-dimensional array of numerals by projection. The horizontal edge flags are accumulated in the horizontal direction and projected onto vertical axis. The vertical, +45-degree and –45-degree edge flags are similarly projected onto horizontal, -45-degree and +45-degree axes, respectively. Therefore the method is called “Principal Axes Projection (PAP)” [3,4]. Then each projection data set is series connected in the order of HR, +45, VR, -45 to form a feature vector. Neighboring four elements are averaged and merged to one element and a 64-dimensional vector is finally obtained. This vector representation very well preserves the human perception of similarity in the vector space. In the experiments below, we have further reduced the feature vector to 16 dimensions by merging each set of four neighboring elements into one, without any significant degradation in performance. C i r cui t c o nf i g ura ti ons A B C VGG A B C VGG IOUT IOUT 1 1 2 2 4 4 1 VIN 13 VIN RST RST £ ¡ ¤¢  £ ¥ §¦  3 Figure 1: Schematic of vector element matching circuit: (a) pyramid (gain reduction) type; (b) plateau (feedback) type. The capacitor area ratio is indicated in the figure. The basic functional form of the similarity evaluation is generated by the shortcut current flowing in a CMOS inverter as in Refs. [7,8,9]. However, their circuits were utilized to form radial basis functions and only the peak position was programmable. In our circuits, not only the peak position but also the peak height and the sharpness of the peak response shape are made configurable to realize flexible matching operations [5]. Two types of the element matching circuit are shown in Fig. 1. They evaluate the similarity between two vector elements. The result of the evaluation is given as an output current (IOUT ) from the pMOS current mirror. The peak position is temporarily memorized by auto-zeroing of the CMOS inverter. The common-gate transistor with VGG stabilizes the voltage supply to the inverter. By controlling the gate bias VGG, the peak height can be changed. This corresponds to multiplying a weight factor to the element. The sharpness of the functional form is taken as the strictness of the similarity evaluation. In the pyramid type circuit (Fig. 1(a)), the sharpness is controlled by the gain reduction in the input. In the plateau type (Fig. 1(b)), the output voltage of the inverter is fed back to input nodes and the sharpness changes in accordance with the amount of the feedback.    ¥£¡ ¦¤¢   £¨ 9&% ¦©§ (!! #$ 5 !' #$ &% 9 9 4 92 !¦ A1@9  ¨¥  5 4 52 (!  5 8765  9) 0 1 ¥ 1 ¨

5 0.10222959 134 nips-2001-On Kernel-Target Alignment

Author: Nello Cristianini, John Shawe-Taylor, André Elisseeff, Jaz S. Kandola

Abstract: We introduce the notion of kernel-alignment, a measure of similarity between two kernel functions or between a kernel and a target function. This quantity captures the degree of agreement between a kernel and a given learning task, and has very natural interpretations in machine learning, leading also to simple algorithms for model selection and learning. We analyse its theoretical properties, proving that it is sharply concentrated around its expected value, and we discuss its relation with other standard measures of performance. Finally we describe some of the algorithms that can be obtained within this framework, giving experimental results showing that adapting the kernel to improve alignment on the labelled data significantly increases the alignment on the test set, giving improved classification accuracy. Hence, the approach provides a principled method of performing transduction. Keywords: Kernels, alignment, eigenvectors, eigenvalues, transduction 1

6 0.099298976 54 nips-2001-Contextual Modulation of Target Saliency

7 0.09333314 84 nips-2001-Global Coordination of Local Linear Models

8 0.088057593 182 nips-2001-The Fidelity of Local Ordinal Encoding

9 0.087808549 111 nips-2001-Learning Lateral Interactions for Feature Binding and Sensory Segmentation

10 0.085928358 189 nips-2001-The g Factor: Relating Distributions on Features to Distributions on Images

11 0.085068293 75 nips-2001-Fast, Large-Scale Transformation-Invariant Clustering

12 0.084899433 170 nips-2001-Spectral Kernel Methods for Clustering

13 0.082881011 153 nips-2001-Product Analysis: Learning to Model Observations as Products of Hidden Variables

14 0.07155209 77 nips-2001-Fast and Robust Classification using Asymmetric AdaBoost and a Detector Cascade

15 0.068730511 80 nips-2001-Generalizable Relational Binding from Coarse-coded Distributed Representations

16 0.067998081 122 nips-2001-Model Based Population Tracking and Automatic Detection of Distribution Changes

17 0.067121409 74 nips-2001-Face Recognition Using Kernel Methods

18 0.065418094 127 nips-2001-Multi Dimensional ICA to Separate Correlated Sources

19 0.063515469 65 nips-2001-Effective Size of Receptive Fields of Inferior Temporal Visual Cortex Neurons in Natural Scenes

20 0.059340622 129 nips-2001-Multiplicative Updates for Classification by Mixture Models


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, -0.166), (1, -0.009), (2, -0.085), (3, -0.064), (4, -0.056), (5, 0.027), (6, -0.23), (7, -0.025), (8, 0.049), (9, 0.004), (10, 0.058), (11, -0.014), (12, 0.186), (13, 0.016), (14, -0.154), (15, -0.036), (16, 0.026), (17, -0.014), (18, 0.031), (19, 0.059), (20, -0.023), (21, 0.11), (22, -0.073), (23, -0.075), (24, -0.013), (25, -0.133), (26, 0.007), (27, -0.076), (28, 0.052), (29, 0.025), (30, 0.08), (31, -0.059), (32, -0.142), (33, 0.12), (34, 0.082), (35, -0.021), (36, 0.198), (37, 0.086), (38, 0.057), (39, -0.007), (40, 0.086), (41, 0.138), (42, 0.049), (43, -0.088), (44, 0.063), (45, 0.063), (46, 0.002), (47, 0.194), (48, -0.038), (49, -0.041)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.97778606 191 nips-2001-Transform-invariant Image Decomposition with Similarity Templates

Author: Chris Stauffer, Erik Miller, Kinh Tieu

Abstract: Recent work has shown impressive transform-invariant modeling and clustering for sets of images of objects with similar appearance. We seek to expand these capabilities to sets of images of an object class that show considerable variation across individual instances (e.g. pedestrian images) using a representation based on pixel-wise similarities, similarity templates. Because of its invariance to the colors of particular components of an object, this representation enables detection of instances of an object class and enables alignment of those instances. Further, this model implicitly represents the regions of color regularity in the class-specific image set enabling a decomposition of that object class into component regions. 1

2 0.67170382 182 nips-2001-The Fidelity of Local Ordinal Encoding

Author: Javid Sadr, Sayan Mukherjee, Keith Thoresz, Pawan Sinha

Abstract: A key question in neuroscience is how to encode sensory stimuli such as images and sounds. Motivated by studies of response properties of neurons in the early cortical areas, we propose an encoding scheme that dispenses with absolute measures of signal intensity or contrast and uses, instead, only local ordinal measures. In this scheme, the structure of a signal is represented by a set of equalities and inequalities across adjacent regions. In this paper, we focus on characterizing the fidelity of this representation strategy. We develop a regularization approach for image reconstruction from ordinal measures and thereby demonstrate that the ordinal representation scheme can faithfully encode signal structure. We also present a neurally plausible implementation of this computation that uses only local update rules. The results highlight the robustness and generalization ability of local ordinal encodings for the task of pattern classification. 1

3 0.54438555 89 nips-2001-Grouping with Bias

Author: Stella X. Yu, Jianbo Shi

Abstract: With the optimization of pattern discrimination as a goal, graph partitioning approaches often lack the capability to integrate prior knowledge to guide grouping. In this paper, we consider priors from unitary generative models, partially labeled data and spatial attention. These priors are modelled as constraints in the solution space. By imposing uniformity condition on the constraints, we restrict the feasible space to one of smooth solutions. A subspace projection method is developed to solve this constrained eigenproblema We demonstrate that simple priors can greatly improve image segmentation results. 1

4 0.49166569 34 nips-2001-Analog Soft-Pattern-Matching Classifier using Floating-Gate MOS Technology

Author: Toshihiko Yamasaki, Tadashi Shibata

Abstract: A flexible pattern-matching analog classifier is presented in conjunction with a robust image representation algorithm called Principal Axes Projection (PAP). In the circuit, the functional form of matching is configurable in terms of the peak position, the peak height and the sharpness of the similarity evaluation. The test chip was fabricated in a 0.6-µm CMOS technology and successfully applied to hand-written pattern recognition and medical radiograph analysis using PAP as a feature extraction pre-processing step for robust image coding. The separation and classification of overlapping patterns is also experimentally demonstrated. 1 I ntr o du c ti o n Pattern classification using template matching techniques is a powerful tool in implementing human-like intelligent systems. However, the processing is computationally very expensive, consuming a lot of CPU time when implemented as software running on general-purpose computers. Therefore, software approaches are not practical for real-time applications. For systems working in mobile environment, in particular, they are not realistic because the memory and computational resources are severely limited. The development of analog VLSI chips having a fully parallel template matching architecture [1,2] would be a promising solution in such applications because they offer an opportunity of low-power operation as well as very compact implementation. In order to build a real human-like intelligent system, however, not only the pattern representation algorithm but also the matching hardware itself needs to be made flexible and robust in carrying out the pattern matching task. First of all, two-dimensional patterns need to be represented by feature vectors having substantially reduced dimensions, while at the same time preserving the human perception of similarity among patterns in the vector space mapping. For this purpose, an image representation algorithm called Principal Axes Projection (PAP) has been de- veloped [3] and its robust nature in pattern recognition has been demonstrated in the applications to medical radiograph analysis [3] and hand-written digits recognition [4]. However, the demonstration so far was only carried out by computer simulation. Regarding the matching hardware, high-flexibility analog template matching circuits have been developed for PAP vector representation. The circuits are flexible in a sense that the matching criteria (the weight to elements, the strictness in matching) are configurable. In Ref. [5], the fundamental characteristics of the building block circuits were presented, and their application to simple hand-written digits was presented in Ref. [6]. The purpose of this paper is to demonstrate the robust nature of the hardware matching system by experiments. The classification of simple hand-written patterns and the cephalometric landmark identification in gray-scale medical radiographs have been carried out and successful results are presented. In addition, multiple overlapping patterns can be separated without utilizing a priori knowledge, which is one of the most difficult problems at present in artificial intelligence. 2 I ma g e re pr es e n tati on by P AP PAP is a feature extraction technique using the edge information. The input image (64x64 pixels) is first subjected to pixel-by-pixel spatial filtering operations to detect edges in four directions: horizontal (HR); vertical (VR); +45 degrees (+45); and –45 degrees (-45). Each detected edge is represented by a binary flag and four edge maps are generated. The two-dimensional bit array in an edge map is reduced to a one-dimensional array of numerals by projection. The horizontal edge flags are accumulated in the horizontal direction and projected onto vertical axis. The vertical, +45-degree and –45-degree edge flags are similarly projected onto horizontal, -45-degree and +45-degree axes, respectively. Therefore the method is called “Principal Axes Projection (PAP)” [3,4]. Then each projection data set is series connected in the order of HR, +45, VR, -45 to form a feature vector. Neighboring four elements are averaged and merged to one element and a 64-dimensional vector is finally obtained. This vector representation very well preserves the human perception of similarity in the vector space. In the experiments below, we have further reduced the feature vector to 16 dimensions by merging each set of four neighboring elements into one, without any significant degradation in performance. C i r cui t c o nf i g ura ti ons A B C VGG A B C VGG IOUT IOUT 1 1 2 2 4 4 1 VIN 13 VIN RST RST £ ¡ ¤¢  £ ¥ §¦  3 Figure 1: Schematic of vector element matching circuit: (a) pyramid (gain reduction) type; (b) plateau (feedback) type. The capacitor area ratio is indicated in the figure. The basic functional form of the similarity evaluation is generated by the shortcut current flowing in a CMOS inverter as in Refs. [7,8,9]. However, their circuits were utilized to form radial basis functions and only the peak position was programmable. In our circuits, not only the peak position but also the peak height and the sharpness of the peak response shape are made configurable to realize flexible matching operations [5]. Two types of the element matching circuit are shown in Fig. 1. They evaluate the similarity between two vector elements. The result of the evaluation is given as an output current (IOUT ) from the pMOS current mirror. The peak position is temporarily memorized by auto-zeroing of the CMOS inverter. The common-gate transistor with VGG stabilizes the voltage supply to the inverter. By controlling the gate bias VGG, the peak height can be changed. This corresponds to multiplying a weight factor to the element. The sharpness of the functional form is taken as the strictness of the similarity evaluation. In the pyramid type circuit (Fig. 1(a)), the sharpness is controlled by the gain reduction in the input. In the plateau type (Fig. 1(b)), the output voltage of the inverter is fed back to input nodes and the sharpness changes in accordance with the amount of the feedback.    ¥£¡ ¦¤¢   £¨ 9&% ¦©§ (!! #$ 5 !' #$ &% 9 9 4 92 !¦ A1@9  ¨¥  5 4 52 (!  5 8765  9) 0 1 ¥ 1 ¨

5 0.46880576 111 nips-2001-Learning Lateral Interactions for Feature Binding and Sensory Segmentation

Author: Heiko Wersing

Abstract: We present a new approach to the supervised learning of lateral interactions for the competitive layer model (CLM) dynamic feature binding architecture. The method is based on consistency conditions, which were recently shown to characterize the attractor states of this linear threshold recurrent network. For a given set of training examples the learning problem is formulated as a convex quadratic optimization problem in the lateral interaction weights. An efficient dimension reduction of the learning problem can be achieved by using a linear superposition of basis interactions. We show the successful application of the method to a medical image segmentation problem of fluorescence microscope cell images.

6 0.44974291 75 nips-2001-Fast, Large-Scale Transformation-Invariant Clustering

7 0.38944852 46 nips-2001-Categorization by Learning and Combining Object Parts

8 0.37890372 54 nips-2001-Contextual Modulation of Target Saliency

9 0.37627167 189 nips-2001-The g Factor: Relating Distributions on Features to Distributions on Images

10 0.36927727 153 nips-2001-Product Analysis: Learning to Model Observations as Products of Hidden Variables

11 0.2733238 84 nips-2001-Global Coordination of Local Linear Models

12 0.27163088 53 nips-2001-Constructing Distributed Representations Using Additive Clustering

13 0.25488251 170 nips-2001-Spectral Kernel Methods for Clustering

14 0.25442341 134 nips-2001-On Kernel-Target Alignment

15 0.25099885 158 nips-2001-Receptive field structure of flow detectors for heading perception

16 0.2503114 193 nips-2001-Unsupervised Learning of Human Motion Models

17 0.25011578 77 nips-2001-Fast and Robust Classification using Asymmetric AdaBoost and a Detector Cascade

18 0.24634704 127 nips-2001-Multi Dimensional ICA to Separate Correlated Sources

19 0.24623883 110 nips-2001-Learning Hierarchical Structures with Linear Relational Embedding

20 0.2438467 176 nips-2001-Stochastic Mixed-Signal VLSI Architecture for High-Dimensional Kernel Machines


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(14, 0.036), (17, 0.014), (19, 0.024), (20, 0.011), (27, 0.137), (30, 0.075), (36, 0.013), (38, 0.024), (49, 0.013), (59, 0.035), (72, 0.09), (79, 0.033), (83, 0.018), (85, 0.237), (91, 0.157)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.88230813 191 nips-2001-Transform-invariant Image Decomposition with Similarity Templates

Author: Chris Stauffer, Erik Miller, Kinh Tieu

Abstract: Recent work has shown impressive transform-invariant modeling and clustering for sets of images of objects with similar appearance. We seek to expand these capabilities to sets of images of an object class that show considerable variation across individual instances (e.g. pedestrian images) using a representation based on pixel-wise similarities, similarity templates. Because of its invariance to the colors of particular components of an object, this representation enables detection of instances of an object class and enables alignment of those instances. Further, this model implicitly represents the regions of color regularity in the class-specific image set enabling a decomposition of that object class into component regions. 1

2 0.80728173 27 nips-2001-Activity Driven Adaptive Stochastic Resonance

Author: Gregor Wenning, Klaus Obermayer

Abstract: Cortical neurons might be considered as threshold elements integrating in parallel many excitatory and inhibitory inputs. Due to the apparent variability of cortical spike trains this yields a strongly fluctuating membrane potential, such that threshold crossings are highly irregular. Here we study how a neuron could maximize its sensitivity w.r.t. a relatively small subset of excitatory input. Weak signals embedded in fluctuations is the natural realm of stochastic resonance. The neuron's response is described in a hazard-function approximation applied to an Ornstein-Uhlenbeck process. We analytically derive an optimality criterium and give a learning rule for the adjustment of the membrane fluctuations, such that the sensitivity is maximal exploiting stochastic resonance. We show that adaptation depends only on quantities that could easily be estimated locally (in space and time) by the neuron. The main results are compared with simulations of a biophysically more realistic neuron model. 1

3 0.70798481 157 nips-2001-Rates of Convergence of Performance Gradient Estimates Using Function Approximation and Bias in Reinforcement Learning

Author: Gregory Z. Grudic, Lyle H. Ungar

Abstract: We address two open theoretical questions in Policy Gradient Reinforcement Learning. The first concerns the efficacy of using function approximation to represent the state action value function, . Theory is presented showing that linear function approximation representations of can degrade the rate of convergence of performance gradient estimates by a factor of relative to when no function approximation of is used, where is the number of possible actions and is the number of basis functions in the function approximation representation. The second concerns the use of a bias term in estimating the state action value function. Theory is presented showing that a non-zero bias term can improve the rate of convergence of performance gradient estimates by , where is the number of possible actions. Experimental evidence is presented showing that these theoretical results lead to significant improvement in the convergence properties of Policy Gradient Reinforcement Learning algorithms.       ¤ ¨ ¦ ¢ ©§¥¤£¡ ¦ ¤ ¨ £¡ ¨ ¤¢  ¢

4 0.70628786 150 nips-2001-Probabilistic Inference of Hand Motion from Neural Activity in Motor Cortex

Author: Yun Gao, Michael J. Black, Elie Bienenstock, Shy Shoham, John P. Donoghue

Abstract: Statistical learning and probabilistic inference techniques are used to infer the hand position of a subject from multi-electrode recordings of neural activity in motor cortex. First, an array of electrodes provides training data of neural firing conditioned on hand kinematics. We learn a nonparametric representation of this firing activity using a Bayesian model and rigorously compare it with previous models using cross-validation. Second, we infer a posterior probability distribution over hand motion conditioned on a sequence of neural test data using Bayesian inference. The learned firing models of multiple cells are used to define a nonGaussian likelihood term which is combined with a prior probability for the kinematics. A particle filtering method is used to represent, update, and propagate the posterior distribution over time. The approach is compared with traditional linear filtering methods; the results suggest that it may be appropriate for neural prosthetic applications.

5 0.70518333 8 nips-2001-A General Greedy Approximation Algorithm with Applications

Author: T. Zhang

Abstract: Greedy approximation algorithms have been frequently used to obtain sparse solutions to learning problems. In this paper, we present a general greedy algorithm for solving a class of convex optimization problems. We derive a bound on the rate of approximation for this algorithm, and show that our algorithm includes a number of earlier studies as special cases.

6 0.70472223 95 nips-2001-Infinite Mixtures of Gaussian Process Experts

7 0.70449078 13 nips-2001-A Natural Policy Gradient

8 0.70334125 29 nips-2001-Adaptive Sparseness Using Jeffreys Prior

9 0.70285928 131 nips-2001-Neural Implementation of Bayesian Inference in Population Codes

10 0.70257866 84 nips-2001-Global Coordination of Local Linear Models

11 0.70121187 100 nips-2001-Iterative Double Clustering for Unsupervised and Semi-Supervised Learning

12 0.70039093 185 nips-2001-The Method of Quantum Clustering

13 0.7003535 132 nips-2001-Novel iteration schemes for the Cluster Variation Method

14 0.69852084 57 nips-2001-Correlation Codes in Neuronal Populations

15 0.69670033 143 nips-2001-PAC Generalization Bounds for Co-training

16 0.69662201 56 nips-2001-Convolution Kernels for Natural Language

17 0.69656098 88 nips-2001-Grouping and dimensionality reduction by locally linear embedding

18 0.69651598 190 nips-2001-Thin Junction Trees

19 0.69609761 135 nips-2001-On Spectral Clustering: Analysis and an algorithm

20 0.69596219 16 nips-2001-A Parallel Mixture of SVMs for Very Large Scale Problems