iccv iccv2013 iccv2013-45 knowledge-graph by maker-knowledge-mining

45 iccv-2013-Affine-Constrained Group Sparse Coding and Its Application to Image-Based Classifications


Source: pdf

Author: Yu-Tseh Chi, Mohsen Ali, Muhammad Rushdi, Jeffrey Ho

Abstract: This paper proposes a novel approach for sparse coding that further improves upon the sparse representation-based classification (SRC) framework. The proposed framework, Affine-Constrained Group Sparse Coding (ACGSC), extends the current SRC framework to classification problems with multiple input samples. Geometrically, the affineconstrained group sparse coding essentially searches for the vector in the convex hull spanned by the input vectors that can best be sparse coded using the given dictionary. The resulting objectivefunction is still convex and can be efficiently optimized using iterative block-coordinate descent scheme that is guaranteed to converge. Furthermore, we provide a form of sparse recovery result that guarantees, at least theoretically, that the classification performance of the constrained group sparse coding should be at least as good as the group sparse coding. We have evaluated the proposed approach using three different recognition experiments that involve illumination variation of faces and textures, and face recognition under occlusions. Prelimi- nary experiments have demonstrated the effectiveness of the proposed approach, and in particular, the results from the recognition/occlusion experiment are surprisingly accurate and robust.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 edu @ Abstract This paper proposes a novel approach for sparse coding that further improves upon the sparse representation-based classification (SRC) framework. [sent-5, score-0.703]

2 The proposed framework, Affine-Constrained Group Sparse Coding (ACGSC), extends the current SRC framework to classification problems with multiple input samples. [sent-6, score-0.172]

3 Geometrically, the affineconstrained group sparse coding essentially searches for the vector in the convex hull spanned by the input vectors that can best be sparse coded using the given dictionary. [sent-7, score-1.353]

4 Furthermore, we provide a form of sparse recovery result that guarantees, at least theoretically, that the classification performance of the constrained group sparse coding should be at least as good as the group sparse coding. [sent-9, score-1.409]

5 The dictionary is obtained virtually without cost by simply stacking the training samples into a matrix D. [sent-17, score-0.216]

6 edu s ing testing, a test image x is sparse coded with respect to the dictionary by optimizing a (convex) objective function E(c; D) of the sparse coefficient c that is usually a sum of an c? [sent-24, score-0.701]

7 The proposed affineconstrained group sparse coding (ACGSC) model aims to further improve the effectiveness of SRC by addressing two of these shortcomings: its generalizability and its extension for multiple input images. [sent-35, score-0.712]

8 Therefore, there is a need to properly generalize SRC for classification problems that require decision on a group of test data x1, . [sent-38, score-0.42]

9 Right: Illustration of the selected atoms from D and the sparse coefficients c of x∗ shown on the right. [sent-54, score-0.398]

10 (Figure best viewed in color) ent setting for SRC, and a straightforward application of SRC would be to apply SRC individually to each xi and generate the final classification decision by pooling or voting among the individual classification results. [sent-56, score-0.391]

11 Group sparse coding [1] (GSC) offers a solution that takes into account the potential relation among [x1, . [sent-58, score-0.392]

12 , xk] by sparse coding all data simultaneously using the objective function mCin E(C;X,D) = ? [sent-61, score-0.392]

13 The matrix C of sparse coefficients can be used as in SRC to generate classification decision by applying voting or pooling across its rows. [sent-66, score-0.482]

14 For classification problems in computer vision, this pa- per argues that a generalization of the group sparse coding, affine-constrained group sparse coding, using the following objective function, offers a more principled and flexible approach: E(a, c; X, D) = ? [sent-68, score-0.916]

15 , is a k-dimensional vector with nonnegative components satisfying the affine constraint a1 + . [sent-74, score-0.225]

16 (2), the ACGSC enlarges its feasible domain by including a kdimensional vector a. [sent-79, score-0.171]

17 However, the feature vector c used in classification is in fact a vector not a matrix, and comparing with GSC, the classification decision based on Eq. [sent-80, score-0.252]

18 Geometrically, ACGSC is easy to explain as it simply searches for the vector in the convex hull S generated by x1, . [sent-82, score-0.321]

19 xk that can best be sparse coded using the dictionary D. [sent-85, score-0.431]

20 From classification viewpoint, ACGSC benefits from multiple inputs by using a linear generative model to “align” the test images with the dictionary elements. [sent-86, score-0.319]

21 For example, if the test (face) images were taken under illumination conditions that are very different from the ones for the dictionary elements, sparse coding test images individually or in group can be expected to provide poor classification results. [sent-87, score-1.086]

22 Heuristically, this can be explained as a kind of generalized misalignment (in terms of illumination effects) between the training and test images. [sent-88, score-0.218]

23 For ACGSC, the coefficient a is used to “align” the test images with the dictionary el- ements, and it is precisely this online “alignment” of the input images xi that provides ACGSC with an edge over other SRC methods. [sent-89, score-0.339]

24 Second, in most applications, the linear generative model given by Xa is valid only for restricted a, and the nonnegative affine constraint proposed here is sufficiently general to provide a bounded and tractable domain for efficient optimization. [sent-93, score-0.275]

25 We remark that for image-based applications, the test images xi have nonnegative intensity values, and therefore, their convex hull would never contain the zero vector (i. [sent-94, score-0.576]

26 Likewise, for a typical group of data xi, their convex hull will not contain the zero vector. [sent-97, score-0.486]

27 The argument in favor of affine-constrained group sparse coding relies mainly on a form of sparse recovery guarantee presented in Theorem 1below. [sent-100, score-0.895]

28 We propose a novel sparse representation-based classification framework based on affine-constrained group sparse coding, and it provides a principled extension of the current SRC framework to classification problems with multiple input samples. [sent-104, score-0.946]

29 We provide theoretical analysis of the proposed frame668822 work in the form of a sparse recovery result. [sent-107, score-0.271]

30 Theory and Method Let x1, · · · , xk denote a group of input test data, and D is the given dictionary. [sent-113, score-0.359]

31 Our proposed affine-constrained group sparse coding sxeeks to minimize the objective function: ECGSC(a, c; X, D) = ? [sent-115, score-0.594]

32 2 + λ Ψ(c) (4) subject to the nonnegative affine constraint on the group coefficient a ai = 1, and a1, a2 , . [sent-117, score-0.566]

33 arse coding [1] (GSC), there are no group eco theaftficients a an? [sent-122, score-0.393]

34 d the sparse coefficients c are given as a matrix. [sent-123, score-0.291]

35 A schematic illustration of the difference between the group sparse coding and our constrained version is shown in Fig. [sent-124, score-0.69]

36 The GSC-based classification scheme sparse codes the input feature vectors xi simultaneously. [sent-126, score-0.399]

37 While some group sparsity can be claimed for this approach based on the appropriate regularizer Ψ, it is generally difficult to provide any guarantee on the behavior of the sparse coefficient matrix C. [sent-127, score-0.523]

38 On the other hand, for our constrained version, the discrete set of the input vectors has been completed to form a convex set S, and our approach is designed to capture any vector in this convex set that can best be sparse coded by the dictionary D. [sent-128, score-0.74]

39 The situation here shares some similarity with the LP-relaxation of integer programming [24] or the convexification of a non- convex program [18], in which one enlarges the feasible domain in order to achieve convexity and thereby, efficiently compute approximate solution. [sent-129, score-0.302]

40 It is clear that the optimization problem is indeed convex and it is completely tractable as the feasible domain and objective function are both convex. [sent-132, score-0.256]

41 The only complication is the projection onto the simplex defined by the group co- (? [sent-134, score-0.202]

42 Figure 2: Illustration of the difference between group sparse coding and constrained group sparse coding, and its effect on classification results. [sent-139, score-1.138]

43 The cone represents the subspace spanned by a block of the dictionary D. [sent-140, score-0.327]

44 Shaded areas represent the convex hull spanned columns in X. [sent-141, score-0.451]

45 None of the xi lie within the subspace; however some of the points on the convex hull do and the proposed algorithm is designed to capture these points. [sent-142, score-0.34]

46 Theoretical Guarantee Given a dictionary D, a vector x has sparsity s if it can be written exactly as a linear combination of s columns of D. [sent-145, score-0.247]

47 An important result that underlies all SRC frameworks is the guarantee provided by the sparse recovery result that for a feature vector x with sparsity bounded from above by a constant depending on D [4, 6], x can be recovered by minimizing the ? [sent-146, score-0.388]

48 We remark that the two programs, while related, are in fact different, with most sparse recovery results given by minimizing (P1). [sent-156, score-0.315]

49 i gA typical SRC method will determine its classification based on its sparse coefficients obtained by minimizing the program (Pλ1). [sent-158, score-0.401]

50 Compared to them, our proposed framework enlarges (thPe optimization domain by introducing the group coefficients a, and it is possible that with larger domain, spurious and incorrect solutions could arise. [sent-159, score-0.447]

51 The following theorem rules out this possibility, at least when the sparse vector x can be exactly recovered by a typical SRC method and classified correctly: (Pλ1) Theorem 1. [sent-160, score-0.269]

52 gF Purthermore, we assume that the global minimum of ECGSC, given a group of input adta ttha eX g, lios unique. [sent-163, score-0.266]

53 The proofis straightforward and it consists ofchecking the sparse vector x also corresponds to the global minimum of ECGSC (a, c). [sent-168, score-0.233]

54 Since x is a sparse vector that can be recovered exactly by minimizing (Pλ1) in Eq. [sent-170, score-0.24]

55 (h6a)t, we bleet c cboev eirtse sparse yco ebfyfi mcieinntism, aznindg we have x = Dc. [sent-171, score-0.201]

56 This shows that (a, c) is the global minimum of the convex function ECGSC (a, c), regardless wmhineitmheurm x oisf on eth ceo boundary toiof nth Ee convex hull S. [sent-178, score-0.447]

57 However, the behavior of GSC in this case is difficult to predict because other (noisy) input feature vectors will affect the sparse coding of the (noiseless) input vector, and the result of the subsequence pooling based on the matrix C can be uncertain. [sent-185, score-0.505]

58 Second, if there is a sparse vector x lying inside the convex hull S spanned by x1, . [sent-186, score-0.586]

59 In the standard ACGSC, the online reconstructed image X a is simply the convex combination of the input images as the columns of X. [sent-194, score-0.281]

60 This model can be augmented using a known image domain partition and define the reconstructed image as a composite image that uses a different convex combination in each region of the partition. [sent-195, score-0.233]

61 (a): ai are the group coefficients corresponding to the sample xi. [sent-207, score-0.359]

62 The nonnegative affine constraint here is a1 + a2 = 1. [sent-208, score-0.225]

63 The vector ap can then be optimized individually under the nonnegative affine constraints given in Eq. [sent-271, score-0.329]

64 Sparse representations have been successfully applied to many vision problems such as face recognition [12], [17], [22], image classification [23], [5], [7], denoising and inpainting [14], and other areas [25]. [sent-279, score-0.228]

65 Interestingly, one of the original motivations for using sparse representations in solving vision problems is the realization that sparse representations of visual signals are produced in the first stage of visual processing by human brain [15]. [sent-280, score-0.46]

66 In many cases, simply replacing the original features with their sparse representations leads to surprisingly better recognition/classification results [25]. [sent-281, score-0.261]

67 Group sparse coding was first proposed in [1] and its extension to block-structured dictionaries has been studied in [2, 19]. [sent-282, score-0.422]

68 In these methods, group sparsity was promoted using a matrix norm that encourages features in the same group to share the same atoms in the dictionary. [sent-283, score-0.559]

69 From classification viewpoint, [1] has two undesirable features: First, the classification is based on matrices (multiple vectors) and this increase in dimension complicates the process. [sent-284, score-0.22]

70 Second, while promoting group sparsity, [1] does not go beyond the test data xi themselves to search for potentially more useful features that are better-aligned with the dictionary. [sent-285, score-0.334]

71 Our proposed framework addresses both shortcomings by introducing the group coefficients a in Eq. [sent-286, score-0.36]

72 However, none of the cited work has investigated face recognition under more realistic scenario when there are large differences between the illumination conditions of the training and test images. [sent-292, score-0.261]

73 The training images were used to simulate the well-lit images such as passport photos, and the test images were used to simulate images (e. [sent-297, score-0.177]

74 We used the training images directly as atoms of the dictionary D [5, 26]. [sent-305, score-0.273]

75 We have also compared with two group-regularized sparse coding algorithms proposed by [1] [2]. [sent-329, score-0.392]

76 Their algorithm promotes the data in a group (columns of X) to share same dictionary atoms. [sent-340, score-0.369]

77 In [2], block structure was added to D together with the group structure on the data X. [sent-341, score-0.295]

78 We applied these two methods to every test sample since they do not utilize group structure on data and the results are also shown in Fig. [sent-357, score-0.278]

79 2, is that our method goes beyond the input test images and searches for a more “aligned” image on a larger domain (the convex hull spanned by X) while the group-regularized methods can only rely on the input images that are poorly poorly “aligned” with the dictionary. [sent-361, score-0.612]

80 This is because, in their experiments, they randomly chose half of the dataset as atoms of D and the other half as test samples. [sent-374, score-0.213]

81 Therefore their dictionaries are 50% larger than ours and the variability between training and test samples are much more limited due to the random selection of training images. [sent-375, score-0.192]

82 4) than the mean images if the convex hull spanned by columns of X lies within the subspaces spanned by the blocks in D. [sent-377, score-0.581]

83 Randomly selected ng test samples (X) from the occluded images of person p. [sent-433, score-0.201]

84 Lastly, we compared with the results from directly applying regular sparse coding (Wright et. [sent-464, score-0.392]

85 8(d) shows the part-based group coefficients (ap) × ap of the test samples after the optimization. [sent-475, score-0.488]

86 8(e) shows the part-based affine combination of the test images. [sent-483, score-0.185]

87 Due to the occlusion of the scarf, the test samples were incorrectly matched to training images from a male subject with beard and mustache. [sent-489, score-0.189]

88 2Fo =r e5 6ac1h2 t iemxtaugrees category, we randomly chose 20 images as the training samples and the rest as test samples. [sent-498, score-0.189]

89 (d) Weights of the part-wise group coefficients overlaid on the corresponding test samples. [sent-505, score-0.368]

90 (e) The part-wise affine combination of the test samples after optimization. [sent-507, score-0.235]

91 Use the training samples as columns of the dictionary D. [sent-521, score-0.282]

92 Since there is no group structure defined in their framework, we computed the sparse coefficients for all the test images individually. [sent-534, score-0.569]

93 We have also presented a form of sparse recovery result and based on this result, we have argued that, at least in theory, the classification performance using the proposed method should be as good as if not better than the one using existing SRC-based methods. [sent-561, score-0.381]

94 We have evaluated the proposed approach using three experiments that involves face recognition, texture classification and face recognition under occlusions. [sent-562, score-0.331]

95 We believe that it is possible to obtain a stronger form of the sparse recovery result under noisy assumption, providing a better un- derstanding of the power and limitation of the proposed algorithm. [sent-565, score-0.271]

96 Furthermore, we will also investigate useful and effective prior for the group coefficients a and the resulting (usually non-convex) optimization problem. [sent-566, score-0.292]

97 Block and group regularized sparse modeling for dictionary learning. [sent-581, score-0.536]

98 Multi-layer group sparse coding for concurrent image classification and annotation. [sent-608, score-0.704]

99 Learning multiscale sparse representations for image and video restoration (preprint). [sent-656, score-0.23]

100 Efficient learning of sparse representations with an energy-based model. [sent-662, score-0.23]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('acgsc', 0.421), ('src', 0.312), ('ecgsc', 0.252), ('group', 0.202), ('sparse', 0.201), ('gsc', 0.199), ('coding', 0.191), ('hull', 0.153), ('dictionary', 0.133), ('convex', 0.131), ('dc', 0.128), ('wright', 0.121), ('nonnegative', 0.116), ('classification', 0.11), ('affine', 0.109), ('atoms', 0.107), ('mcin', 0.103), ('xa', 0.102), ('spanned', 0.101), ('ip', 0.097), ('block', 0.093), ('coefficients', 0.09), ('face', 0.089), ('aixi', 0.084), ('test', 0.076), ('enlarges', 0.075), ('ap', 0.07), ('recovery', 0.07), ('ai', 0.067), ('columns', 0.066), ('illumination', 0.063), ('np', 0.061), ('chi', 0.057), ('affineconstrained', 0.056), ('cecgsc', 0.056), ('embodied', 0.056), ('xi', 0.056), ('reconstructed', 0.052), ('ak', 0.051), ('samples', 0.05), ('domain', 0.05), ('curet', 0.05), ('xk', 0.049), ('pooling', 0.049), ('elhamifar', 0.049), ('coded', 0.048), ('sparsity', 0.048), ('misalignment', 0.046), ('feasible', 0.046), ('ieee', 0.044), ('remark', 0.044), ('rows', 0.043), ('hayman', 0.043), ('aip', 0.043), ('texture', 0.043), ('coefficient', 0.042), ('ng', 0.042), ('ci', 0.042), ('noiseless', 0.041), ('azimuth', 0.04), ('elevation', 0.04), ('barcelona', 0.04), ('recovered', 0.039), ('scarf', 0.038), ('shortcomings', 0.038), ('searches', 0.037), ('cis', 0.037), ('princeton', 0.037), ('conference', 0.037), ('schematic', 0.036), ('yale', 0.034), ('promotes', 0.034), ('individually', 0.034), ('simulate', 0.034), ('ho', 0.034), ('training', 0.033), ('occluded', 0.033), ('pages', 0.033), ('sapiro', 0.033), ('completed', 0.033), ('input', 0.032), ('minimum', 0.032), ('decision', 0.032), ('constrained', 0.031), ('surprisingly', 0.031), ('wiley', 0.031), ('guarantee', 0.03), ('chose', 0.03), ('cropped', 0.03), ('framework', 0.03), ('aij', 0.03), ('extension', 0.03), ('subject', 0.03), ('varma', 0.03), ('theory', 0.029), ('representations', 0.029), ('blocks', 0.029), ('theorem', 0.029), ('illustration', 0.029), ('indeed', 0.029)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.9999997 45 iccv-2013-Affine-Constrained Group Sparse Coding and Its Application to Image-Based Classifications

Author: Yu-Tseh Chi, Mohsen Ali, Muhammad Rushdi, Jeffrey Ho

Abstract: This paper proposes a novel approach for sparse coding that further improves upon the sparse representation-based classification (SRC) framework. The proposed framework, Affine-Constrained Group Sparse Coding (ACGSC), extends the current SRC framework to classification problems with multiple input samples. Geometrically, the affineconstrained group sparse coding essentially searches for the vector in the convex hull spanned by the input vectors that can best be sparse coded using the given dictionary. The resulting objectivefunction is still convex and can be efficiently optimized using iterative block-coordinate descent scheme that is guaranteed to converge. Furthermore, we provide a form of sparse recovery result that guarantees, at least theoretically, that the classification performance of the constrained group sparse coding should be at least as good as the group sparse coding. We have evaluated the proposed approach using three different recognition experiments that involve illumination variation of faces and textures, and face recognition under occlusions. Prelimi- nary experiments have demonstrated the effectiveness of the proposed approach, and in particular, the results from the recognition/occlusion experiment are surprisingly accurate and robust.

2 0.35474074 20 iccv-2013-A Max-Margin Perspective on Sparse Representation-Based Classification

Author: Zhaowen Wang, Jianchao Yang, Nasser Nasrabadi, Thomas Huang

Abstract: Sparse Representation-based Classification (SRC) is a powerful tool in distinguishing signal categories which lie on different subspaces. Despite its wide application to visual recognition tasks, current understanding of SRC is solely based on a reconstructive perspective, which neither offers any guarantee on its classification performance nor provides any insight on how to design a discriminative dictionary for SRC. In this paper, we present a novel perspective towards SRC and interpret it as a margin classifier. The decision boundary and margin of SRC are analyzed in local regions where the support of sparse code is stable. Based on the derived margin, we propose a hinge loss function as the gauge for the classification performance of SRC. A stochastic gradient descent algorithm is implemented to maximize the margin of SRC and obtain more discriminative dictionaries. Experiments validate the effectiveness of the proposed approach in predicting classification performance and improving dictionary quality over reconstructive ones. Classification results competitive with other state-ofthe-art sparse coding methods are reported on several data sets.

3 0.23330744 188 iccv-2013-Group Sparsity and Geometry Constrained Dictionary Learning for Action Recognition from Depth Maps

Author: Jiajia Luo, Wei Wang, Hairong Qi

Abstract: Human action recognition based on the depth information provided by commodity depth sensors is an important yet challenging task. The noisy depth maps, different lengths of action sequences, and free styles in performing actions, may cause large intra-class variations. In this paper, a new framework based on sparse coding and temporal pyramid matching (TPM) is proposed for depthbased human action recognition. Especially, a discriminative class-specific dictionary learning algorithm isproposed for sparse coding. By adding the group sparsity and geometry constraints, features can be well reconstructed by the sub-dictionary belonging to the same class, and the geometry relationships among features are also kept in the calculated coefficients. The proposed approach is evaluated on two benchmark datasets captured by depth cameras. Experimental results show that the proposed algorithm repeatedly hqi } @ ut k . edu GB ImagesR epth ImagesD setkonlSy0 896.5170d4ept.3h021 .x02y 19.876504.dep3th02.1 x02. achieves superior performance to the state of the art algorithms. Moreover, the proposed dictionary learning method also outperforms classic dictionary learning approaches.

4 0.2077198 384 iccv-2013-Semi-supervised Robust Dictionary Learning via Efficient l-Norms Minimization

Author: Hua Wang, Feiping Nie, Weidong Cai, Heng Huang

Abstract: Representing the raw input of a data set by a set of relevant codes is crucial to many computer vision applications. Due to the intrinsic sparse property of real-world data, dictionary learning, in which the linear decomposition of a data point uses a set of learned dictionary bases, i.e., codes, has demonstrated state-of-the-art performance. However, traditional dictionary learning methods suffer from three weaknesses: sensitivity to noisy and outlier samples, difficulty to determine the optimal dictionary size, and incapability to incorporate supervision information. In this paper, we address these weaknesses by learning a Semi-Supervised Robust Dictionary (SSR-D). Specifically, we use the ℓ2,0+ norm as the loss function to improve the robustness against outliers, and develop a new structured sparse regularization com, , tom. . cai@sydney . edu . au , heng@uta .edu make the learning tasks easier to deal with and reduce the computational cost. For example, in image tagging, instead of using the raw pixel-wise features, semi-local or patch- based features, such as SIFT and geometric blur, are usually more desirable to achieve better performance. In practice, finding a set of compact features bases, also referred to as dictionary, with enhanced representative and discriminative power, plays a significant role in building a successful computer vision system. In this paper, we explore this important problem by proposing a novel formulation and its solution for learning Semi-Supervised Robust Dictionary (SSRD), where we examine the challenges in dictionary learning, and seek opportunities to overcome them and improve the dictionary qualities. 1.1. Challenges in Dictionary Learning to incorporate the supervision information in dictionary learning, without incurring additional parameters. Moreover, the optimal dictionary size is automatically learned from the input data. Minimizing the derived objective function is challenging because it involves many non-smooth ℓ2,0+ -norm terms. We present an efficient algorithm to solve the problem with a rigorous proof of the convergence of the algorithm. Extensive experiments are presented to show the superior performance of the proposed method.

5 0.20234209 161 iccv-2013-Fast Sparsity-Based Orthogonal Dictionary Learning for Image Restoration

Author: Chenglong Bao, Jian-Feng Cai, Hui Ji

Abstract: In recent years, how to learn a dictionary from input images for sparse modelling has been one very active topic in image processing and recognition. Most existing dictionary learning methods consider an over-complete dictionary, e.g. the K-SVD method. Often they require solving some minimization problem that is very challenging in terms of computational feasibility and efficiency. However, if the correlations among dictionary atoms are not well constrained, the redundancy of the dictionary does not necessarily improve the performance of sparse coding. This paper proposed a fast orthogonal dictionary learning method for sparse image representation. With comparable performance on several image restoration tasks, the proposed method is much more computationally efficient than the over-complete dictionary based learning methods.

6 0.17811944 276 iccv-2013-Multi-attributed Dictionary Learning for Sparse Coding

7 0.17644432 354 iccv-2013-Robust Dictionary Learning by Error Source Decomposition

8 0.15241142 197 iccv-2013-Hierarchical Joint Max-Margin Learning of Mid and Top Level Representations for Visual Recognition

9 0.13524927 258 iccv-2013-Low-Rank Sparse Coding for Image Classification

10 0.13247843 398 iccv-2013-Sparse Variation Dictionary Learning for Face Recognition with a Single Training Sample per Person

11 0.1321547 14 iccv-2013-A Generalized Iterated Shrinkage Algorithm for Non-convex Sparse Coding

12 0.12540516 232 iccv-2013-Latent Space Sparse Subspace Clustering

13 0.11985205 97 iccv-2013-Coupling Alignments with Recognition for Still-to-Video Face Recognition

14 0.11642219 244 iccv-2013-Learning View-Invariant Sparse Representations for Cross-View Action Recognition

15 0.11545324 114 iccv-2013-Dictionary Learning and Sparse Coding on Grassmann Manifolds: An Extrinsic Solution

16 0.10813922 6 iccv-2013-A Convex Optimization Framework for Active Learning

17 0.10672189 422 iccv-2013-Toward Guaranteed Illumination Models for Non-convex Objects

18 0.10555744 51 iccv-2013-Anchored Neighborhood Regression for Fast Example-Based Super-Resolution

19 0.10258877 359 iccv-2013-Robust Object Tracking with Online Multi-lifespan Dictionary Learning

20 0.10064779 177 iccv-2013-From Point to Set: Extend the Learning of Distance Metrics


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.224), (1, 0.099), (2, -0.12), (3, -0.048), (4, -0.272), (5, -0.094), (6, -0.049), (7, 0.015), (8, -0.011), (9, -0.059), (10, 0.016), (11, 0.038), (12, -0.009), (13, 0.02), (14, -0.068), (15, 0.017), (16, -0.025), (17, 0.011), (18, 0.016), (19, 0.051), (20, -0.007), (21, -0.041), (22, -0.009), (23, -0.058), (24, 0.029), (25, -0.014), (26, 0.025), (27, 0.041), (28, 0.109), (29, -0.009), (30, 0.03), (31, -0.049), (32, -0.062), (33, -0.031), (34, 0.013), (35, 0.04), (36, 0.03), (37, 0.01), (38, 0.02), (39, 0.031), (40, 0.062), (41, -0.001), (42, 0.023), (43, 0.09), (44, -0.053), (45, -0.007), (46, 0.018), (47, -0.016), (48, -0.054), (49, 0.084)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.95798308 45 iccv-2013-Affine-Constrained Group Sparse Coding and Its Application to Image-Based Classifications

Author: Yu-Tseh Chi, Mohsen Ali, Muhammad Rushdi, Jeffrey Ho

Abstract: This paper proposes a novel approach for sparse coding that further improves upon the sparse representation-based classification (SRC) framework. The proposed framework, Affine-Constrained Group Sparse Coding (ACGSC), extends the current SRC framework to classification problems with multiple input samples. Geometrically, the affineconstrained group sparse coding essentially searches for the vector in the convex hull spanned by the input vectors that can best be sparse coded using the given dictionary. The resulting objectivefunction is still convex and can be efficiently optimized using iterative block-coordinate descent scheme that is guaranteed to converge. Furthermore, we provide a form of sparse recovery result that guarantees, at least theoretically, that the classification performance of the constrained group sparse coding should be at least as good as the group sparse coding. We have evaluated the proposed approach using three different recognition experiments that involve illumination variation of faces and textures, and face recognition under occlusions. Prelimi- nary experiments have demonstrated the effectiveness of the proposed approach, and in particular, the results from the recognition/occlusion experiment are surprisingly accurate and robust.

2 0.91832274 20 iccv-2013-A Max-Margin Perspective on Sparse Representation-Based Classification

Author: Zhaowen Wang, Jianchao Yang, Nasser Nasrabadi, Thomas Huang

Abstract: Sparse Representation-based Classification (SRC) is a powerful tool in distinguishing signal categories which lie on different subspaces. Despite its wide application to visual recognition tasks, current understanding of SRC is solely based on a reconstructive perspective, which neither offers any guarantee on its classification performance nor provides any insight on how to design a discriminative dictionary for SRC. In this paper, we present a novel perspective towards SRC and interpret it as a margin classifier. The decision boundary and margin of SRC are analyzed in local regions where the support of sparse code is stable. Based on the derived margin, we propose a hinge loss function as the gauge for the classification performance of SRC. A stochastic gradient descent algorithm is implemented to maximize the margin of SRC and obtain more discriminative dictionaries. Experiments validate the effectiveness of the proposed approach in predicting classification performance and improving dictionary quality over reconstructive ones. Classification results competitive with other state-ofthe-art sparse coding methods are reported on several data sets.

3 0.84776795 354 iccv-2013-Robust Dictionary Learning by Error Source Decomposition

Author: Zhuoyuan Chen, Ying Wu

Abstract: Sparsity models have recently shown great promise in many vision tasks. Using a learned dictionary in sparsity models can in general outperform predefined bases in clean data. In practice, both training and testing data may be corrupted and contain noises and outliers. Although recent studies attempted to cope with corrupted data and achieved encouraging results in testing phase, how to handle corruption in training phase still remains a very difficult problem. In contrast to most existing methods that learn the dictionaryfrom clean data, this paper is targeted at handling corruptions and outliers in training data for dictionary learning. We propose a general method to decompose the reconstructive residual into two components: a non-sparse component for small universal noises and a sparse component for large outliers, respectively. In addition, , further analysis reveals the connection between our approach and the “partial” dictionary learning approach, updating only part of the prototypes (or informative codewords) with remaining (or noisy codewords) fixed. Experiments on synthetic data as well as real applications have shown satisfactory per- formance of this new robust dictionary learning approach.

4 0.83494353 384 iccv-2013-Semi-supervised Robust Dictionary Learning via Efficient l-Norms Minimization

Author: Hua Wang, Feiping Nie, Weidong Cai, Heng Huang

Abstract: Representing the raw input of a data set by a set of relevant codes is crucial to many computer vision applications. Due to the intrinsic sparse property of real-world data, dictionary learning, in which the linear decomposition of a data point uses a set of learned dictionary bases, i.e., codes, has demonstrated state-of-the-art performance. However, traditional dictionary learning methods suffer from three weaknesses: sensitivity to noisy and outlier samples, difficulty to determine the optimal dictionary size, and incapability to incorporate supervision information. In this paper, we address these weaknesses by learning a Semi-Supervised Robust Dictionary (SSR-D). Specifically, we use the ℓ2,0+ norm as the loss function to improve the robustness against outliers, and develop a new structured sparse regularization com, , tom. . cai@sydney . edu . au , heng@uta .edu make the learning tasks easier to deal with and reduce the computational cost. For example, in image tagging, instead of using the raw pixel-wise features, semi-local or patch- based features, such as SIFT and geometric blur, are usually more desirable to achieve better performance. In practice, finding a set of compact features bases, also referred to as dictionary, with enhanced representative and discriminative power, plays a significant role in building a successful computer vision system. In this paper, we explore this important problem by proposing a novel formulation and its solution for learning Semi-Supervised Robust Dictionary (SSRD), where we examine the challenges in dictionary learning, and seek opportunities to overcome them and improve the dictionary qualities. 1.1. Challenges in Dictionary Learning to incorporate the supervision information in dictionary learning, without incurring additional parameters. Moreover, the optimal dictionary size is automatically learned from the input data. Minimizing the derived objective function is challenging because it involves many non-smooth ℓ2,0+ -norm terms. We present an efficient algorithm to solve the problem with a rigorous proof of the convergence of the algorithm. Extensive experiments are presented to show the superior performance of the proposed method.

5 0.830791 14 iccv-2013-A Generalized Iterated Shrinkage Algorithm for Non-convex Sparse Coding

Author: Wangmeng Zuo, Deyu Meng, Lei Zhang, Xiangchu Feng, David Zhang

Abstract: In many sparse coding based image restoration and image classification problems, using non-convex ?p-norm minimization (0 ≤ p < 1) can often obtain better results than timhei convex 0?1 -norm m 1)ini camniza otfiteonn. Ab naiunm bbeetrt of algorithms, e.g., iteratively reweighted least squares (IRLS), iteratively thresholding method (ITM-?p), and look-up table (LUT), have been proposed for non-convex ?p-norm sparse coding, while some analytic solutions have been suggested for some specific values of p. In this paper, by extending the popular soft-thresholding operator, we propose a generalized iterated shrinkage algorithm (GISA) for ?p-norm non-convex sparse coding. Unlike the analytic solutions, the proposed GISA algorithm is easy to implement, and can be adopted for solving non-convex sparse coding problems with arbitrary p values. Compared with LUT, GISA is more general and does not need to compute and store the look-up tables. Compared with IRLS and ITM-?p, GISA is theoretically more solid and can achieve more accurate solutions. Experiments on image restoration and sparse coding based face recognition are conducted to validate the performance of GISA. ××

6 0.78945225 398 iccv-2013-Sparse Variation Dictionary Learning for Face Recognition with a Single Training Sample per Person

7 0.77360433 161 iccv-2013-Fast Sparsity-Based Orthogonal Dictionary Learning for Image Restoration

8 0.75886577 197 iccv-2013-Hierarchical Joint Max-Margin Learning of Mid and Top Level Representations for Visual Recognition

9 0.74884754 188 iccv-2013-Group Sparsity and Geometry Constrained Dictionary Learning for Action Recognition from Depth Maps

10 0.74148703 276 iccv-2013-Multi-attributed Dictionary Learning for Sparse Coding

11 0.73654246 258 iccv-2013-Low-Rank Sparse Coding for Image Classification

12 0.697559 34 iccv-2013-Abnormal Event Detection at 150 FPS in MATLAB

13 0.68122792 114 iccv-2013-Dictionary Learning and Sparse Coding on Grassmann Manifolds: An Extrinsic Solution

14 0.63377118 401 iccv-2013-Stacked Predictive Sparse Coding for Classification of Distinct Regions in Tumor Histopathology

15 0.63354337 96 iccv-2013-Coupled Dictionary and Feature Space Learning with Applications to Cross-Domain Image Synthesis and Recognition

16 0.63294721 51 iccv-2013-Anchored Neighborhood Regression for Fast Example-Based Super-Resolution

17 0.62637788 422 iccv-2013-Toward Guaranteed Illumination Models for Non-convex Objects

18 0.59985524 292 iccv-2013-Non-convex P-Norm Projection for Robust Sparsity

19 0.54004747 167 iccv-2013-Finding Causal Interactions in Video Sequences

20 0.53814107 244 iccv-2013-Learning View-Invariant Sparse Representations for Cross-View Action Recognition


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(2, 0.055), (7, 0.021), (26, 0.092), (31, 0.058), (40, 0.011), (42, 0.186), (48, 0.015), (64, 0.047), (67, 0.151), (73, 0.037), (78, 0.02), (89, 0.156), (95, 0.016), (97, 0.031), (98, 0.018)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.88614094 45 iccv-2013-Affine-Constrained Group Sparse Coding and Its Application to Image-Based Classifications

Author: Yu-Tseh Chi, Mohsen Ali, Muhammad Rushdi, Jeffrey Ho

Abstract: This paper proposes a novel approach for sparse coding that further improves upon the sparse representation-based classification (SRC) framework. The proposed framework, Affine-Constrained Group Sparse Coding (ACGSC), extends the current SRC framework to classification problems with multiple input samples. Geometrically, the affineconstrained group sparse coding essentially searches for the vector in the convex hull spanned by the input vectors that can best be sparse coded using the given dictionary. The resulting objectivefunction is still convex and can be efficiently optimized using iterative block-coordinate descent scheme that is guaranteed to converge. Furthermore, we provide a form of sparse recovery result that guarantees, at least theoretically, that the classification performance of the constrained group sparse coding should be at least as good as the group sparse coding. We have evaluated the proposed approach using three different recognition experiments that involve illumination variation of faces and textures, and face recognition under occlusions. Prelimi- nary experiments have demonstrated the effectiveness of the proposed approach, and in particular, the results from the recognition/occlusion experiment are surprisingly accurate and robust.

2 0.87587368 279 iccv-2013-Multi-stage Contextual Deep Learning for Pedestrian Detection

Author: Xingyu Zeng, Wanli Ouyang, Xiaogang Wang

Abstract: Cascaded classifiers1 have been widely used in pedestrian detection and achieved great success. These classifiers are trained sequentially without joint optimization. In this paper, we propose a new deep model that can jointly train multi-stage classifiers through several stages of backpropagation. It keeps the score map output by a classifier within a local region and uses it as contextual information to support the decision at the next stage. Through a specific design of the training strategy, this deep architecture is able to simulate the cascaded classifiers by mining hard samples to train the network stage-by-stage. Each classifier handles samples at a different difficulty level. Unsupervised pre-training and specifically designed stage-wise supervised training are used to regularize the optimization problem. Both theoretical analysis and experimental results show that the training strategy helps to avoid overfitting. Experimental results on three datasets (Caltech, ETH and TUD-Brussels) show that our approach outperforms the state-of-the-art approaches.

3 0.85920066 259 iccv-2013-Manifold Based Face Synthesis from Sparse Samples

Author: Hongteng Xu, Hongyuan Zha

Abstract: Data sparsity has been a thorny issuefor manifold-based image synthesis, and in this paper we address this critical problem by leveraging ideas from transfer learning. Specifically, we propose methods based on generating auxiliary data in the form of synthetic samples using transformations of the original sparse samples. To incorporate the auxiliary data, we propose a weighted data synthesis method, which adaptively selects from the generated samples for inclusion during the manifold learning process via a weighted iterative algorithm. To demonstrate the feasibility of the proposed method, we apply it to the problem of face image synthesis from sparse samples. Compared with existing methods, the proposed method shows encouraging results with good performance improvements.

4 0.85809743 161 iccv-2013-Fast Sparsity-Based Orthogonal Dictionary Learning for Image Restoration

Author: Chenglong Bao, Jian-Feng Cai, Hui Ji

Abstract: In recent years, how to learn a dictionary from input images for sparse modelling has been one very active topic in image processing and recognition. Most existing dictionary learning methods consider an over-complete dictionary, e.g. the K-SVD method. Often they require solving some minimization problem that is very challenging in terms of computational feasibility and efficiency. However, if the correlations among dictionary atoms are not well constrained, the redundancy of the dictionary does not necessarily improve the performance of sparse coding. This paper proposed a fast orthogonal dictionary learning method for sparse image representation. With comparable performance on several image restoration tasks, the proposed method is much more computationally efficient than the over-complete dictionary based learning methods.

5 0.84743327 44 iccv-2013-Adapting Classification Cascades to New Domains

Author: Vidit Jain, Sachin Sudhakar Farfade

Abstract: Classification cascades have been very effective for object detection. Such a cascade fails to perform well in data domains with variations in appearances that may not be captured in the training examples. This limited generalization severely restricts the domains for which they can be used effectively. A common approach to address this limitation is to train a new cascade of classifiers from scratch for each of the new domains. Building separate detectors for each of the different domains requires huge annotation and computational effort, making it not scalable to a large number of data domains. Here we present an algorithm for quickly adapting a pre-trained cascade of classifiers using a small number oflabeledpositive instancesfrom a different yet similar data domain. In our experiments with images of human babies and human-like characters from movies, we demonstrate that the adapted cascade significantly outperforms both of the original cascade and the one trained from scratch using the given training examples. –

6 0.84703124 427 iccv-2013-Transfer Feature Learning with Joint Distribution Adaptation

7 0.84696144 330 iccv-2013-Proportion Priors for Image Sequence Segmentation

8 0.84657687 280 iccv-2013-Multi-view 3D Reconstruction from Uncalibrated Radially-Symmetric Cameras

9 0.84579796 398 iccv-2013-Sparse Variation Dictionary Learning for Face Recognition with a Single Training Sample per Person

10 0.84461927 277 iccv-2013-Multi-channel Correlation Filters

11 0.84416837 14 iccv-2013-A Generalized Iterated Shrinkage Algorithm for Non-convex Sparse Coding

12 0.84290642 328 iccv-2013-Probabilistic Elastic Part Model for Unsupervised Face Detector Adaptation

13 0.84264708 93 iccv-2013-Correlation Adaptive Subspace Segmentation by Trace Lasso

14 0.84167594 54 iccv-2013-Attribute Pivots for Guiding Relevance Feedback in Image Search

15 0.84141976 26 iccv-2013-A Practical Transfer Learning Algorithm for Face Verification

16 0.84108543 392 iccv-2013-Similarity Metric Learning for Face Recognition

17 0.83841777 80 iccv-2013-Collaborative Active Learning of a Kernel Machine Ensemble for Recognition

18 0.83828145 106 iccv-2013-Deep Learning Identity-Preserving Face Space

19 0.83698195 79 iccv-2013-Coherent Object Detection with 3D Geometric Context from a Single Image

20 0.83624923 52 iccv-2013-Attribute Adaptation for Personalized Image Search