cvpr cvpr2013 cvpr2013-66 knowledge-graph by maker-knowledge-mining

66 cvpr-2013-Block and Group Regularized Sparse Modeling for Dictionary Learning


Source: pdf

Author: Yu-Tseh Chi, Mohsen Ali, Ajit Rajwade, Jeffrey Ho

Abstract: This paper proposes a dictionary learning framework that combines the proposed block/group (BGSC) or reconstructed block/group (R-BGSC) sparse coding schemes with the novel Intra-block Coherence Suppression Dictionary Learning (ICS-DL) algorithm. An important and distinguishing feature of the proposed framework is that all dictionary blocks are trained simultaneously with respect to each data group while the intra-block coherence being explicitly minimized as an important objective. We provide both empirical evidence and heuristic support for this feature that can be considered as a direct consequence of incorporating both the group structure for the input data and the block structure for the dictionary in the learning process. The optimization problems for both the dictionary learning and sparse coding can be solved efficiently using block-gradient descent, and the details of the optimization algorithms are presented. We evaluate the proposed methods using well-known datasets, and favorable comparisons with state-of-the-art dictionary learning methods demonstrate the viability and validity of the proposed framework.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 in Abstract This paper proposes a dictionary learning framework that combines the proposed block/group (BGSC) or reconstructed block/group (R-BGSC) sparse coding schemes with the novel Intra-block Coherence Suppression Dictionary Learning (ICS-DL) algorithm. [sent-8, score-0.749]

2 An important and distinguishing feature of the proposed framework is that all dictionary blocks are trained simultaneously with respect to each data group while the intra-block coherence being explicitly minimized as an important objective. [sent-9, score-1.123]

3 We provide both empirical evidence and heuristic support for this feature that can be considered as a direct consequence of incorporating both the group structure for the input data and the block structure for the dictionary in the learning process. [sent-10, score-0.982]

4 The optimization problems for both the dictionary learning and sparse coding can be solved efficiently using block-gradient descent, and the details of the optimization algorithms are presented. [sent-11, score-0.749]

5 We evaluate the proposed methods using well-known datasets, and favorable comparisons with state-of-the-art dictionary learning methods demonstrate the viability and validity of the proposed framework. [sent-12, score-0.631]

6 Introduction Sparse modeling and dictionary learning have emerged recently as an effective and popular paradigm for solving many important learning problems in computer vision. [sent-14, score-0.611]

7 λDgΨ(D) + λCΩ(C(g)), (1) where the X(g) are vectors/matrices generated from the data X, and Ψ, are regularizers on the learned dictionary D and sparse coefficients C(g) , respectively. [sent-19, score-0.804]

8 In dictionary learning, Ω(C) is usually based on various sparsity- Ω Figure1:Ilustraion fthepro sedBlock/GroupSarseCoding algorithm. [sent-20, score-0.57]

9 A group of data X(g) on the left is sparsely coded with respect to the dictionary D with block structure D[1] · · · D[b] . [sent-21, score-0.925]

10 promoting norms that depend on the extra structures placed on D, and it is the regularizer Ψ(D) that largely determines the nature of the dictionary D. [sent-22, score-0.584]

11 (1) provides the elegant theme, its variations are often composed of extra structures placed on the dictionary D ( [15, 10, 7, 13]), and less frequently, different ways of generating sparsely-coded data X(g) for training the dictionary. [sent-25, score-0.572]

12 For classification, a block structure is often imposed on D and hierarchical structures could be further specified using these blocks ([16, 10]), with the aim of endowing the learned dictionary certain predictive power. [sent-27, score-1.044]

13 To promote sparsity, the block structure on D is often accompanied by an appropriate block-based ? [sent-28, score-0.303]

14 On the other hand, for X(g) , a common approach is to generate a collection of groups of data vectors {xg1 , ··· , xgk} from X and to simultaneously sparse ctoordse { xthe ,d·a·ta· xvect}ors fr oinm e Xac han dda ttao group nXe(og)u [l1y] . [sent-34, score-0.282]

15 In a noiseless setting, our proposed problem of encoding sparse representations for a group of data samples X using 333777557 the minimum number of blocks from D can be cast as the following optimization program: P? [sent-38, score-0.473]

16 cator function ,p = 1, 2 and C[i] is the wi-thheer e bl Io(c·k) (ss aunb i-nmdaictriatxo) ro ffu nCc ithonat , corresponds t Co the i-th block of D as shown in Fig. [sent-45, score-0.267]

17 (3) We will call this program Block/Group Sparse Coding (BGSC) as it incorporates both the group structure in data and block structure in the dictionary. [sent-55, score-0.482]

18 In some applications of which the main concern is identifying contributing blocks rather than finding the sparse representation [7], the following optimization program is considered: P? [sent-56, score-0.374]

19 m Sharing of dictionary atoms for data in the same group had been shown to increase the discriminative power of the dictionary ([1]). [sent-85, score-1.386]

20 With the block structure added to dictionary D, our proposed BGSC and R-BGSC algorithms promote a group of data to share only few blocks of D for encoding. [sent-86, score-1.17]

21 Therefore, incorporating these SC algorithms in a dictionary learning framework, which iteratively updates coefficients of data and updates atoms of D, will result in training each block of D using only few groups of data. [sent-87, score-1.224]

22 This means that, for example, a badly written digit ’9’, which looks like a ’7’, when grouped together with other normally written ’9’s, will be encoded using atoms these ’9’s used. [sent-88, score-0.418]

23 The badly written ’9’ will, in turns, be used to train the atoms in D that represent ’9’s rather than those that represent ’7’s. [sent-89, score-0.31]

24 Another novelty of our framework is that we do not assign a class of signals to specific blocks of a dictionary, unlike other Sparse Representation based Classification (SRC) [6, 17] and [13]. [sent-90, score-0.266]

25 This would allow some blocks to store shared features between some different classes. [sent-91, score-0.232]

26 [13] trained a single dictionary block for each group of data. [sent-94, score-0.901]

27 This method increases the redundancy of the information encoded in the learned dictionary as the informationcommonto two ormore groups (acommon scenario in many classification problems) will need to be stored separately within each block. [sent-95, score-0.68]

28 Since one dictionary block is assigned to each class, the redundancy induced in the dictionary needs to be reduced for greater efficiency. [sent-96, score-1.36]

29 This is done by removing dictionary elements whose mutual dot product has an absolute value greater than an arbitrary-chosen threshold (e. [sent-97, score-0.584]

30 In particular, our proposal to encode data from a single class using multiple blocks obviates the need to even incorporate an explicit inter-block coherence minimization term unlike [13]. [sent-102, score-0.524]

31 3 in) 1 [ 5w],he thne na (2k − 1)μB < 1− (na − 1)μS, (6) where na and k are the size and the rank of a block, respectively, and μB and μS are inter- and intra-block coherence defined in Section 2. [sent-107, score-0.506]

32 A way to achieve minimum μS is to make atoms orthonormal within each block [11, 3]. [sent-110, score-0.478]

33 However, such dictionaries (over-complete dictionary with union of orthonormal basis) do not perform as well as those with more flexible structure [14]. [sent-111, score-0.688]

34 For example, in SRC-based face recognition, each block contains atoms representing faces of the same person. [sent-112, score-0.424]

35 It does not make sense to impose strict orthogonality on each block. [sent-113, score-0.173]

36 Therefore, rather than imposing strong orthogonality constraint on each block, we propose a dictionary learning algorithm that minimizes only intrablock coherence. [sent-114, score-0.856]

37 The proposed dictionary learning framework learns the dictionary D by minimizing the objective function given in Eq. [sent-115, score-1.153]

38 (16) measures the mutual coherence within each block of D. [sent-117, score-0.537]

39 The corresponding sparse coding can be either BGSC or R-BGSC. [sent-118, score-0.171]

40 Besides the novel sparse coding algorithms, BGSC and R-BGSC, there are three specific features that distinguish our dictionary framework from existing methods: 1. [sent-119, score-0.716]

41 4 minimizes the intrablock coherence as one of its main objectives. [sent-122, score-0.395]

42 Our framework does not require to assign a class or a group of data to block(s) in the dictionary as in [13]. [sent-124, score-0.717]

43 This allows some blocks of the dictionary to be shared by different classes. [sent-125, score-0.777]

44 The dictionary is trained simultaneously with respect to each group of training samples X(g) using our proposed block/group regularized SC algorithm. [sent-127, score-0.757]

45 We will start with sparse coding algorithms first and work our way towards the full dictionary learning algorithm. [sent-133, score-0.749]

46 We denote scalars with lower-case letters, vectors with bold lower-case letters, matrices with upper-case letters, and the i-th block and group of a matrix (or vector) with Z[i] , and Z(i) , respectively. [sent-134, score-0.356]

47 imization steps only for one specific group of data X and its corresponding sparse coefficients C. [sent-165, score-0.34]

48 (13), the block sparsity of C depends on the value of λ. [sent-265, score-0.242]

49 When there is no block structure on D, BGSC is equivalent to the group sparse coding (GSC) in [1]. [sent-273, score-0.59]

50 Intra-Block Coherence Suppression Dictionary Learning The intra-block coherence is defined as μS(D) = miax? [sent-302, score-0.256]

51 =j (15) set of the atoms in block μB tis o fde tfhineed a as μB n(D b)l c=k ? [sent-318, score-0.424]

52 As mentioned in the Introduction, it is necessary to have a dictionary updating algorithm that minimizes the intrablock coherence. [sent-324, score-0.684]

53 =r where cr is the r-th row of C and dr is in block b. [sent-353, score-0.363]

54 It is clear from the first three terms of the above equation why group-regularized SC algorithms tend to generate high intra-block coherence blocks. [sent-356, score-0.256]

55 As we can see, the value of dr depends not only on how much it is being used to encode X (1st and 3rd term) but also on how much other dk’s are being used to encode X. [sent-357, score-0.187]

56 Since, BGSC and R-BGSC minimize the number of blocks to be used for encoding X, the atoms dr are likely in the same block as dk. [sent-358, score-0.753]

57 For example, if the coefficient C of X has only one non-zero block, then the atoms dr and dk, which correspond to the non-zero rows of coefficients c’s in the above equation, are all in the same block. [sent-359, score-0.441]

58 This justifies putting the intra-block coherence suppressing regularizer term in Eq. [sent-361, score-0.363]

59 badly written ’9’s that look like a ’7’, into one group and hence allow them to act as one different class and to be used to train the dictionary blocks corresponding to the wrong classes. [sent-367, score-1.022]

60 Note that it is not uncommon to add a post-processing step to make atoms in D unit vectors or requiring ? [sent-371, score-0.182]

61 We collected 15 groups of data for each digit where each group contained 50 randomly chosen images from the same class. [sent-382, score-0.26]

62 Generate a random dictionary D with nb blocks and each block contains na columns (atoms) (a total of nb na columns). [sent-384, score-1.493]

63 Iteratively compute coefficients using BGSC and update the dictionary using ICS-DL algorithm. [sent-386, score-0.683]

64 Use the coefficients of the training data to train 10 onevs-all linear SVMs[2]. [sent-388, score-0.204]

65 Compute the sparse coefficients of the test samples using either BGSC or R-BGSC. [sent-390, score-0.258]

66 are 333787880 Table 1demonstrates the impact of the dictionary’s block structure on the error rates. [sent-393, score-0.315]

67 For the experiment in the last column of Table 1, we assign two blocks to each digit. [sent-397, score-0.239]

68 The results show that the error rates are similar when the number of blocks (nb) is greater than 10 even though the number of classes of this dataset is 10. [sent-398, score-0.305]

69 2(a), the sparse coefficients of most of the training data have 3 to 6 active blocks when nb = 20. [sent-401, score-0.585]

70 The last column of Table 1 shows that the hard assignment of blocks to classes results in higher error rate even though the size of the dictionary is twice as large as those of the first three experiments in Table 1. [sent-402, score-0.83]

71 As mentioned in the Introduction, we did not assign blocks to classes and prefer using more blocks for encoding data with larger variability. [sent-403, score-0.447]

72 We can see that ’7’ and ’9’ share two blocks of dictionary due to their similarity. [sent-407, score-0.753]

73 However, they each have an exclusive block with large coefficients (darker in color) to allow them to encode the difference. [sent-408, score-0.413]

74 52 †: Assign each digit to two blocks of the dictionary. [sent-417, score-0.297]

75 When β = 0, our ICS-DL algorithm does not suppress intra-block coherence and is hence equivalent to the dictionary learning algorithm in [1]. [sent-421, score-0.873]

76 We stopped the training roughly after 200 iterations when the dictionary update did not change much. [sent-427, score-0.597]

77 The results in Table 2 suggest that suppressing the intra-block coherence can indeed improve the performance. [sent-428, score-0.324]

78 In the extreme case when imposing a strict orthogonality on the blocks using the UOB-DL, the error rate increases to 4. [sent-430, score-0.493]

79 These results provide an empirical support for not using strict orthogonality constraint. [sent-432, score-0.173]

80 However, our ICS-DL algorithm does not impose any inter-block orthogonality constraint on the dictionary as SISF-DL does. [sent-434, score-0.649]

81 12 To further demonstrate the intra-block coherence suppressing property of our ICS-DL algorithm, we plot the intra-block coherence values of the dictionaries trained with β = 0 and β = 200, respectively, in Fig. [sent-444, score-0.645]

82 Solid and dotted lines indicate the coherence and error, respectively. [sent-447, score-0.297]

83 The red solid line demonstrates that our ICS-DL method can keep the intra-block coherence at a low value. [sent-448, score-0.292]

84 On the contrary, without the intra-block coherence suppression term, the blue solid line shows that the coherence value becomes comparably large with increasing number of iterations. [sent-449, score-0.583]

85 Once we have a trained dictionary, we used the coefficients of training samples to train ten linear SVMs. [sent-451, score-0.236]

86 Another way to obtain coefficients of training samples is to recompute them individually. [sent-453, score-0.197]

87 The dictionary was trained with β = 400, λtrain = 0. [sent-456, score-0.545]

88 Finally, we compared our results with other state-of-theart results using dictionary learning algorithms ([13, 12]) shown in Table 3. [sent-466, score-0.578]

89 We also compare with the UOBDL ([11]) which imposes strict orthogonality constraint on blocks. [sent-467, score-0.173]

90 The results show that our algorithms outperform other dictionary learning methods, even the one specially tailored for hand-written digits recognition [8]. [sent-468, score-0.578]

91 Although Table 2 suggests that suppressing intra-block coherence of D improves the classification performance, imposing a strict orthogonality on the blocks, however, does not result in any improvement. [sent-469, score-0.558]

92 (b) Intra-block coherence (solid) and error rates (dotted) of two dictionaries (red for β = 200 and blue for β = 0). [sent-482, score-0.418]

93 The scenarios differ in terms of how the training samples are organized for computing the coefficients and which proposed SC algorithms were used. [sent-485, score-0.197]

94 First column in the legend (separated by ’ | ’) indicates how the coefficients of the training samples are computed, in groups (G) or individually (I) . [sent-486, score-0.254]

95 Tumhen s ienco thned l ecgoelunmdn (s ienpdaicraatteeds wbyh i’c|h’) SinCd algorithm i tsh eus cedoe tffoi compute hthee coefficients of the training samples. [sent-487, score-0.19]

96 However, due to the amount and complexity of this dataset, we were not able to fully exploit different dictionary structures and parameters to obtain a reasonable result. [sent-490, score-0.545]

97 Conclusion We have proposed a novel dictionary learning framework that includes two novel block/group regularized sparse coding algorithms and one novel dictionary learning algorithm. [sent-496, score-1.366]

98 Experimental comparisons with several state-of-the-art dictionary learning methods are favorable, and in particular, for hand-written digit recognition experiment, the proposed framework outperformed these state-of-the-art dictionary learning algorithms. [sent-497, score-1.245]

99 5Our dictionary size is 5 times smaller than what was used in SISF-DL. [sent-540, score-0.545]

100 Classification and clustering via dictionary learning with structured incoherence and shared features. [sent-590, score-0.602]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('dictionary', 0.545), ('bgsc', 0.4), ('coherence', 0.256), ('block', 0.242), ('blocks', 0.208), ('atoms', 0.182), ('coefficients', 0.138), ('na', 0.125), ('nb', 0.124), ('dr', 0.121), ('group', 0.114), ('orthogonality', 0.104), ('programs', 0.093), ('digit', 0.089), ('mcin', 0.089), ('sparse', 0.088), ('intrablock', 0.086), ('coding', 0.083), ('program', 0.078), ('ramirez', 0.076), ('strict', 0.069), ('dk', 0.069), ('suppressing', 0.068), ('dictionaries', 0.065), ('usps', 0.061), ('ajit', 0.057), ('rajwade', 0.057), ('groups', 0.057), ('badly', 0.055), ('dc', 0.055), ('orthonormal', 0.054), ('minimizes', 0.053), ('qc', 0.051), ('error', 0.049), ('sc', 0.049), ('rates', 0.048), ('sprechmann', 0.044), ('dotted', 0.041), ('regularizer', 0.039), ('regularized', 0.039), ('equivalent', 0.039), ('mutual', 0.039), ('train', 0.039), ('combinatorial', 0.037), ('supplementary', 0.037), ('promote', 0.037), ('clarity', 0.037), ('solid', 0.036), ('mnist', 0.036), ('elhamifar', 0.036), ('imposing', 0.035), ('suppression', 0.035), ('written', 0.034), ('derivation', 0.034), ('letters', 0.034), ('encode', 0.033), ('learning', 0.033), ('regularizers', 0.033), ('icassp', 0.032), ('samples', 0.032), ('cast', 0.031), ('assign', 0.031), ('presentation', 0.03), ('objective', 0.03), ('iterate', 0.029), ('norm', 0.029), ('sapiro', 0.028), ('favorable', 0.028), ('cc', 0.028), ('redundancy', 0.028), ('frobenius', 0.028), ('rate', 0.028), ('training', 0.027), ('proved', 0.027), ('class', 0.027), ('ieee', 0.027), ('mairal', 0.027), ('formula', 0.026), ('classification', 0.026), ('viability', 0.025), ('hec', 0.025), ('miax', 0.025), ('depen', 0.025), ('ffu', 0.025), ('endowing', 0.025), ('fthepro', 0.025), ('gsc', 0.025), ('haend', 0.025), ('india', 0.025), ('poczos', 0.025), ('szabo', 0.025), ('thned', 0.025), ('iterations', 0.025), ('shared', 0.024), ('encoded', 0.024), ('structure', 0.024), ('nxe', 0.023), ('oinm', 0.023), ('bruckstein', 0.023), ('jeffrey', 0.023)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000002 66 cvpr-2013-Block and Group Regularized Sparse Modeling for Dictionary Learning

Author: Yu-Tseh Chi, Mohsen Ali, Ajit Rajwade, Jeffrey Ho

Abstract: This paper proposes a dictionary learning framework that combines the proposed block/group (BGSC) or reconstructed block/group (R-BGSC) sparse coding schemes with the novel Intra-block Coherence Suppression Dictionary Learning (ICS-DL) algorithm. An important and distinguishing feature of the proposed framework is that all dictionary blocks are trained simultaneously with respect to each data group while the intra-block coherence being explicitly minimized as an important objective. We provide both empirical evidence and heuristic support for this feature that can be considered as a direct consequence of incorporating both the group structure for the input data and the block structure for the dictionary in the learning process. The optimization problems for both the dictionary learning and sparse coding can be solved efficiently using block-gradient descent, and the details of the optimization algorithms are presented. We evaluate the proposed methods using well-known datasets, and favorable comparisons with state-of-the-art dictionary learning methods demonstrate the viability and validity of the proposed framework.

2 0.43988907 392 cvpr-2013-Separable Dictionary Learning

Author: Simon Hawe, Matthias Seibert, Martin Kleinsteuber

Abstract: Many techniques in computer vision, machine learning, and statistics rely on the fact that a signal of interest admits a sparse representation over some dictionary. Dictionaries are either available analytically, or can be learned from a suitable training set. While analytic dictionaries permit to capture the global structure of a signal and allow a fast implementation, learned dictionaries often perform better in applications as they are more adapted to the considered class of signals. In imagery, unfortunately, the numerical burden for (i) learning a dictionary and for (ii) employing the dictionary for reconstruction tasks only allows to deal with relatively small image patches that only capture local image information. The approach presented in this paper aims at overcoming these drawbacks by allowing a separable structure on the dictionary throughout the learning process. On the one hand, this permits larger patch-sizes for the learning phase, on the other hand, the dictionary is applied efficiently in reconstruction tasks. The learning procedure is based on optimizing over a product of spheres which updates the dictionary as a whole, thus enforces basic dictionary proper- , ties such as mutual coherence explicitly during the learning procedure. In the special case where no separable structure is enforced, our method competes with state-of-the-art dictionary learning methods like K-SVD.

3 0.42742762 296 cvpr-2013-Multi-level Discriminative Dictionary Learning towards Hierarchical Visual Categorization

Author: Li Shen, Shuhui Wang, Gang Sun, Shuqiang Jiang, Qingming Huang

Abstract: For the task of visual categorization, the learning model is expected to be endowed with discriminative visual feature representation and flexibilities in processing many categories. Many existing approaches are designed based on a flat category structure, or rely on a set of pre-computed visual features, hence may not be appreciated for dealing with large numbers of categories. In this paper, we propose a novel dictionary learning method by taking advantage of hierarchical category correlation. For each internode of the hierarchical category structure, a discriminative dictionary and a set of classification models are learnt for visual categorization, and the dictionaries in different layers are learnt to exploit the discriminative visual properties of different granularity. Moreover, the dictionaries in lower levels also inherit the dictionary of ancestor nodes, so that categories in lower levels are described with multi-scale visual information using our dictionary learning approach. Experiments on ImageNet object data subset and SUN397 scene dataset demonstrate that our approach achieves promising performance on data with large numbers of classes compared with some state-of-the-art methods, and is more efficient in processing large numbers of categories.

4 0.39724949 257 cvpr-2013-Learning Structured Low-Rank Representations for Image Classification

Author: Yangmuzi Zhang, Zhuolin Jiang, Larry S. Davis

Abstract: An approach to learn a structured low-rank representation for image classification is presented. We use a supervised learning method to construct a discriminative and reconstructive dictionary. By introducing an ideal regularization term, we perform low-rank matrix recovery for contaminated training data from all categories simultaneously without losing structural information. A discriminative low-rank representation for images with respect to the constructed dictionary is obtained. With semantic structure information and strong identification capability, this representation is good for classification tasks even using a simple linear multi-classifier. Experimental results demonstrate the effectiveness of our approach.

5 0.39154169 58 cvpr-2013-Beta Process Joint Dictionary Learning for Coupled Feature Spaces with Application to Single Image Super-Resolution

Author: Li He, Hairong Qi, Russell Zaretzki

Abstract: This paper addresses the problem of learning overcomplete dictionaries for the coupled feature spaces, where the learned dictionaries also reflect the relationship between the two spaces. A Bayesian method using a beta process prior is applied to learn the over-complete dictionaries. Compared to previous couple feature spaces dictionary learning algorithms, our algorithm not only provides dictionaries that customized to each feature space, but also adds more consistent and accurate mapping between the two feature spaces. This is due to the unique property of the beta process model that the sparse representation can be decomposed to values and dictionary atom indicators. The proposed algorithm is able to learn sparse representations that correspond to the same dictionary atoms with the same sparsity but different values in coupled feature spaces, thus bringing consistent and accurate mapping between coupled feature spaces. Another advantage of the proposed method is that the number of dictionary atoms and their relative importance may be inferred non-parametrically. We compare the proposed approach to several state-of-the-art dictionary learning methods super-resolution. tionaries learned resolution results ods. by applying this method to single image The experimental results show that dicby our method produces the best supercompared to other state-of-the-art meth-

6 0.38401625 315 cvpr-2013-Online Robust Dictionary Learning

7 0.37985057 185 cvpr-2013-Generalized Domain-Adaptive Dictionaries

8 0.26396668 422 cvpr-2013-Tag Taxonomy Aware Dictionary Learning for Region Tagging

9 0.25376928 125 cvpr-2013-Dictionary Learning from Ambiguously Labeled Data

10 0.23566265 5 cvpr-2013-A Bayesian Approach to Multimodal Visual Dictionary Learning

11 0.20804578 419 cvpr-2013-Subspace Interpolation via Dictionary Learning for Unsupervised Domain Adaptation

12 0.20711973 302 cvpr-2013-Multi-task Sparse Learning with Beta Process Prior for Action Recognition

13 0.16092104 220 cvpr-2013-In Defense of Sparsity Based Face Recognition

14 0.15973334 204 cvpr-2013-Histograms of Sparse Codes for Object Detection

15 0.14446653 399 cvpr-2013-Single-Sample Face Recognition with Image Corruption and Misalignment via Sparse Illumination Transfer

16 0.14218049 67 cvpr-2013-Blocks That Shout: Distinctive Parts for Scene Classification

17 0.12549315 178 cvpr-2013-From Local Similarity to Global Coding: An Application to Image Classification

18 0.11924907 442 cvpr-2013-Transfer Sparse Coding for Robust Image Representation

19 0.11503229 88 cvpr-2013-Compressible Motion Fields

20 0.093567133 164 cvpr-2013-Fast Convolutional Sparse Coding


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.184), (1, -0.183), (2, -0.32), (3, 0.373), (4, -0.134), (5, -0.147), (6, 0.133), (7, 0.142), (8, -0.044), (9, 0.11), (10, 0.017), (11, 0.075), (12, 0.017), (13, 0.023), (14, 0.029), (15, 0.007), (16, 0.035), (17, 0.053), (18, -0.03), (19, 0.011), (20, -0.013), (21, 0.034), (22, 0.024), (23, 0.036), (24, 0.029), (25, -0.009), (26, -0.044), (27, -0.014), (28, -0.061), (29, 0.016), (30, 0.035), (31, 0.031), (32, -0.005), (33, -0.012), (34, 0.005), (35, -0.012), (36, 0.03), (37, -0.016), (38, 0.0), (39, -0.004), (40, 0.009), (41, -0.013), (42, -0.016), (43, 0.036), (44, 0.025), (45, 0.001), (46, -0.016), (47, -0.042), (48, -0.002), (49, -0.001)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.9786616 392 cvpr-2013-Separable Dictionary Learning

Author: Simon Hawe, Matthias Seibert, Martin Kleinsteuber

Abstract: Many techniques in computer vision, machine learning, and statistics rely on the fact that a signal of interest admits a sparse representation over some dictionary. Dictionaries are either available analytically, or can be learned from a suitable training set. While analytic dictionaries permit to capture the global structure of a signal and allow a fast implementation, learned dictionaries often perform better in applications as they are more adapted to the considered class of signals. In imagery, unfortunately, the numerical burden for (i) learning a dictionary and for (ii) employing the dictionary for reconstruction tasks only allows to deal with relatively small image patches that only capture local image information. The approach presented in this paper aims at overcoming these drawbacks by allowing a separable structure on the dictionary throughout the learning process. On the one hand, this permits larger patch-sizes for the learning phase, on the other hand, the dictionary is applied efficiently in reconstruction tasks. The learning procedure is based on optimizing over a product of spheres which updates the dictionary as a whole, thus enforces basic dictionary proper- , ties such as mutual coherence explicitly during the learning procedure. In the special case where no separable structure is enforced, our method competes with state-of-the-art dictionary learning methods like K-SVD.

same-paper 2 0.97106701 66 cvpr-2013-Block and Group Regularized Sparse Modeling for Dictionary Learning

Author: Yu-Tseh Chi, Mohsen Ali, Ajit Rajwade, Jeffrey Ho

Abstract: This paper proposes a dictionary learning framework that combines the proposed block/group (BGSC) or reconstructed block/group (R-BGSC) sparse coding schemes with the novel Intra-block Coherence Suppression Dictionary Learning (ICS-DL) algorithm. An important and distinguishing feature of the proposed framework is that all dictionary blocks are trained simultaneously with respect to each data group while the intra-block coherence being explicitly minimized as an important objective. We provide both empirical evidence and heuristic support for this feature that can be considered as a direct consequence of incorporating both the group structure for the input data and the block structure for the dictionary in the learning process. The optimization problems for both the dictionary learning and sparse coding can be solved efficiently using block-gradient descent, and the details of the optimization algorithms are presented. We evaluate the proposed methods using well-known datasets, and favorable comparisons with state-of-the-art dictionary learning methods demonstrate the viability and validity of the proposed framework.

3 0.95501423 58 cvpr-2013-Beta Process Joint Dictionary Learning for Coupled Feature Spaces with Application to Single Image Super-Resolution

Author: Li He, Hairong Qi, Russell Zaretzki

Abstract: This paper addresses the problem of learning overcomplete dictionaries for the coupled feature spaces, where the learned dictionaries also reflect the relationship between the two spaces. A Bayesian method using a beta process prior is applied to learn the over-complete dictionaries. Compared to previous couple feature spaces dictionary learning algorithms, our algorithm not only provides dictionaries that customized to each feature space, but also adds more consistent and accurate mapping between the two feature spaces. This is due to the unique property of the beta process model that the sparse representation can be decomposed to values and dictionary atom indicators. The proposed algorithm is able to learn sparse representations that correspond to the same dictionary atoms with the same sparsity but different values in coupled feature spaces, thus bringing consistent and accurate mapping between coupled feature spaces. Another advantage of the proposed method is that the number of dictionary atoms and their relative importance may be inferred non-parametrically. We compare the proposed approach to several state-of-the-art dictionary learning methods super-resolution. tionaries learned resolution results ods. by applying this method to single image The experimental results show that dicby our method produces the best supercompared to other state-of-the-art meth-

4 0.93534255 315 cvpr-2013-Online Robust Dictionary Learning

Author: Cewu Lu, Jiaping Shi, Jiaya Jia

Abstract: Online dictionary learning is particularly useful for processing large-scale and dynamic data in computer vision. It, however, faces the major difficulty to incorporate robust functions, rather than the square data fitting term, to handle outliers in training data. In thispaper, wepropose a new online framework enabling the use of ?1 sparse data fitting term in robust dictionary learning, notably enhancing the usability and practicality of this important technique. Extensive experiments have been carried out to validate our new framework.

5 0.91267234 257 cvpr-2013-Learning Structured Low-Rank Representations for Image Classification

Author: Yangmuzi Zhang, Zhuolin Jiang, Larry S. Davis

Abstract: An approach to learn a structured low-rank representation for image classification is presented. We use a supervised learning method to construct a discriminative and reconstructive dictionary. By introducing an ideal regularization term, we perform low-rank matrix recovery for contaminated training data from all categories simultaneously without losing structural information. A discriminative low-rank representation for images with respect to the constructed dictionary is obtained. With semantic structure information and strong identification capability, this representation is good for classification tasks even using a simple linear multi-classifier. Experimental results demonstrate the effectiveness of our approach.

6 0.88759232 296 cvpr-2013-Multi-level Discriminative Dictionary Learning towards Hierarchical Visual Categorization

7 0.83294654 125 cvpr-2013-Dictionary Learning from Ambiguously Labeled Data

8 0.80004168 185 cvpr-2013-Generalized Domain-Adaptive Dictionaries

9 0.73955595 5 cvpr-2013-A Bayesian Approach to Multimodal Visual Dictionary Learning

10 0.66871858 422 cvpr-2013-Tag Taxonomy Aware Dictionary Learning for Region Tagging

11 0.6451382 220 cvpr-2013-In Defense of Sparsity Based Face Recognition

12 0.60605127 302 cvpr-2013-Multi-task Sparse Learning with Beta Process Prior for Action Recognition

13 0.55881327 204 cvpr-2013-Histograms of Sparse Codes for Object Detection

14 0.53651947 83 cvpr-2013-Classification of Tumor Histology via Morphometric Context

15 0.52802324 442 cvpr-2013-Transfer Sparse Coding for Robust Image Representation

16 0.43043804 399 cvpr-2013-Single-Sample Face Recognition with Image Corruption and Misalignment via Sparse Illumination Transfer

17 0.41539788 419 cvpr-2013-Subspace Interpolation via Dictionary Learning for Unsupervised Domain Adaptation

18 0.41077337 164 cvpr-2013-Fast Convolutional Sparse Coding

19 0.40515473 178 cvpr-2013-From Local Similarity to Global Coding: An Application to Image Classification

20 0.37019795 421 cvpr-2013-Supervised Kernel Descriptors for Visual Recognition


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(10, 0.109), (16, 0.023), (19, 0.182), (26, 0.079), (28, 0.014), (33, 0.285), (67, 0.052), (69, 0.099), (87, 0.07)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.94603097 463 cvpr-2013-What's in a Name? First Names as Facial Attributes

Author: Huizhong Chen, Andrew C. Gallagher, Bernd Girod

Abstract: This paper introduces a new idea in describing people using their first names, i.e., the name assigned at birth. We show that describing people in terms of similarity to a vector of possible first names is a powerful description of facial appearance that can be used for face naming and building facial attribute classifiers. We build models for 100 common first names used in the United States and for each pair, construct a pairwise firstname classifier. These classifiers are built using training images downloaded from the internet, with no additional user interaction. This gives our approach important advantages in building practical systems that do not require additional human intervention for labeling. We use the scores from each pairwise name classifier as a set of facial attributes. We show several surprising results. Our name attributes predict the correct first names of test faces at rates far greater than chance. The name attributes are applied to gender recognition and to age classification, outperforming state-of-the-art methods with all training images automatically gathered from the internet.

2 0.92734593 356 cvpr-2013-Representing and Discovering Adversarial Team Behaviors Using Player Roles

Author: Patrick Lucey, Alina Bialkowski, Peter Carr, Stuart Morgan, Iain Matthews, Yaser Sheikh

Abstract: In this paper, we describe a method to represent and discover adversarial group behavior in a continuous domain. In comparison to other types of behavior, adversarial behavior is heavily structured as the location of a player (or agent) is dependent both on their teammates and adversaries, in addition to the tactics or strategies of the team. We present a method which can exploit this relationship through the use of a spatiotemporal basis model. As players constantly change roles during a match, we show that employing a “role-based” representation instead of one based on player “identity” can best exploit the playing structure. As vision-based systems currently do not provide perfect detection/tracking (e.g. missed or false detections), we show that our compact representation can effectively “denoise ” erroneous detections as well as enabling temporal analysis, which was previously prohibitive due to the dimensionality of the signal. To evaluate our approach, we used a fully instrumented field-hockey pitch with 8 fixed highdefinition (HD) cameras and evaluated our approach on approximately 200,000 frames of data from a state-of-the- art real-time player detector and compare it to manually labelled data.

3 0.91692013 197 cvpr-2013-Hallucinated Humans as the Hidden Context for Labeling 3D Scenes

Author: Yun Jiang, Hema Koppula, Ashutosh Saxena

Abstract: For scene understanding, one popular approach has been to model the object-object relationships. In this paper, we hypothesize that such relationships are only an artifact of certain hidden factors, such as humans. For example, the objects, monitor and keyboard, are strongly spatially correlated only because a human types on the keyboard while watching the monitor. Our goal is to learn this hidden human context (i.e., the human-object relationships), and also use it as a cue for labeling the scenes. We present Infinite Factored Topic Model (IFTM), where we consider a scene as being generated from two types of topics: human configurations and human-object relationships. This enables our algorithm to hallucinate the possible configurations of the humans in the scene parsimoniously. Given only a dataset of scenes containing objects but not humans, we show that our algorithm can recover the human object relationships. We then test our algorithm on the task ofattribute and object labeling in 3D scenes and show consistent improvements over the state-of-the-art.

same-paper 4 0.88981593 66 cvpr-2013-Block and Group Regularized Sparse Modeling for Dictionary Learning

Author: Yu-Tseh Chi, Mohsen Ali, Ajit Rajwade, Jeffrey Ho

Abstract: This paper proposes a dictionary learning framework that combines the proposed block/group (BGSC) or reconstructed block/group (R-BGSC) sparse coding schemes with the novel Intra-block Coherence Suppression Dictionary Learning (ICS-DL) algorithm. An important and distinguishing feature of the proposed framework is that all dictionary blocks are trained simultaneously with respect to each data group while the intra-block coherence being explicitly minimized as an important objective. We provide both empirical evidence and heuristic support for this feature that can be considered as a direct consequence of incorporating both the group structure for the input data and the block structure for the dictionary in the learning process. The optimization problems for both the dictionary learning and sparse coding can be solved efficiently using block-gradient descent, and the details of the optimization algorithms are presented. We evaluate the proposed methods using well-known datasets, and favorable comparisons with state-of-the-art dictionary learning methods demonstrate the viability and validity of the proposed framework.

5 0.87379742 377 cvpr-2013-Sample-Specific Late Fusion for Visual Category Recognition

Author: Dong Liu, Kuan-Ting Lai, Guangnan Ye, Ming-Syan Chen, Shih-Fu Chang

Abstract: Late fusion addresses the problem of combining the prediction scores of multiple classifiers, in which each score is predicted by a classifier trained with a specific feature. However, the existing methods generally use a fixed fusion weight for all the scores of a classifier, and thus fail to optimally determine the fusion weight for the individual samples. In this paper, we propose a sample-specific late fusion method to address this issue. Specifically, we cast the problem into an information propagation process which propagates the fusion weights learned on the labeled samples to individual unlabeled samples, while enforcing that positive samples have higher fusion scores than negative samples. In this process, we identify the optimal fusion weights for each sample and push positive samples to top positions in the fusion score rank list. We formulate our problem as a L∞ norm constrained optimization problem and apply the Alternating Direction Method of Multipliers for the optimization. Extensive experiment results on various visual categorization tasks show that the proposed method consis- tently and significantly beats the state-of-the-art late fusion methods. To the best knowledge, this is the first method supporting sample-specific fusion weight learning.

6 0.87084872 64 cvpr-2013-Blessing of Dimensionality: High-Dimensional Feature and Its Efficient Compression for Face Verification

7 0.86498559 61 cvpr-2013-Beyond Point Clouds: Scene Understanding by Reasoning Geometry and Physics

8 0.86455315 119 cvpr-2013-Detecting and Aligning Faces by Image Retrieval

9 0.8617655 231 cvpr-2013-Joint Detection, Tracking and Mapping by Semantic Bundle Adjustment

10 0.86145777 292 cvpr-2013-Multi-agent Event Detection: Localization and Role Assignment

11 0.86070836 445 cvpr-2013-Understanding Bayesian Rooms Using Composite 3D Object Models

12 0.86028564 70 cvpr-2013-Bottom-Up Segmentation for Top-Down Detection

13 0.85987514 311 cvpr-2013-Occlusion Patterns for Object Class Detection

14 0.85934389 446 cvpr-2013-Understanding Indoor Scenes Using 3D Geometric Phrases

15 0.85830009 248 cvpr-2013-Learning Collections of Part Models for Object Recognition

16 0.85756963 86 cvpr-2013-Composite Statistical Inference for Semantic Segmentation

17 0.85700965 372 cvpr-2013-SLAM++: Simultaneous Localisation and Mapping at the Level of Objects

18 0.85670674 172 cvpr-2013-Finding Group Interactions in Social Clutter

19 0.8563385 365 cvpr-2013-Robust Real-Time Tracking of Multiple Objects by Volumetric Mass Densities

20 0.85503495 441 cvpr-2013-Tracking Sports Players with Context-Conditioned Motion Models