iccv iccv2013 iccv2013-408 knowledge-graph by maker-knowledge-mining

408 iccv-2013-Super-resolution via Transform-Invariant Group-Sparse Regularization


Source: pdf

Author: Carlos Fernandez-Granda, Emmanuel J. Candès

Abstract: We present a framework to super-resolve planar regions found in urban scenes and other man-made environments by taking into account their 3D geometry. Such regions have highly structured straight edges, but this prior is challenging to exploit due to deformations induced by the projection onto the imaging plane. Our method factors out such deformations by using recently developed tools based on convex optimization to learn a transform that maps the image to a domain where its gradient has a simple group-sparse structure. This allows to obtain a novel convex regularizer that enforces global consistency constraints between the edges of the image. Computational experiments with real images show that this data-driven approach to the design of regularizers promoting transform-invariant group sparsity is very effective at high super-resolution factors. We view our approach as complementary to most recent superresolution methods, which tend to focus on hallucinating high-frequency textures.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 edu Abstract We present a framework to super-resolve planar regions found in urban scenes and other man-made environments by taking into account their 3D geometry. [sent-2, score-0.19]

2 Such regions have highly structured straight edges, but this prior is challenging to exploit due to deformations induced by the projection onto the imaging plane. [sent-3, score-0.314]

3 Our method factors out such deformations by using recently developed tools based on convex optimization to learn a transform that maps the image to a domain where its gradient has a simple group-sparse structure. [sent-4, score-0.398]

4 This allows to obtain a novel convex regularizer that enforces global consistency constraints between the edges of the image. [sent-5, score-0.436]

5 Computational experiments with real images show that this data-driven approach to the design of regularizers promoting transform-invariant group sparsity is very effective at high super-resolution factors. [sent-6, score-0.143]

6 The aim is to obtain a higher-resolution image by upsampling a single image. [sent-12, score-0.289]

7 We can only hope to reconstruct certain very specific structures (see [4] for theoretical results on the super-resolution of pointwise objects) or to hallucinate high-frequency textures that are visually pleasing [1]. [sent-14, score-0.214]

8 In this work we consider planar regions taken from 3D scenes that have straight edges aligned in a few main directions, such as the one in Figure 1. [sent-15, score-0.518]

9 They are ubiquitous in urban environments and recent large-scale urban 3D mapping efforts (such as the Apple 3D map) make such data readily availEmmanuel J. [sent-17, score-0.096]

10 Existing super-resolution techniques can be applied to this class of textures to obtain reasonably good upsampling results up to factors of three or four. [sent-22, score-0.496]

11 In this work, we explore the possibility of attaining higher upsampling factors by harnessing such prior knowledge for images with structured edges. [sent-24, score-0.414]

12 Unfortunately, the non-uniform blur and deformations induced by the projection of 3D surfaces onto the imaging plane make it very challenging to exploit prior knowledge about the structure of the data directly. [sent-25, score-0.163]

13 In fact, in our quest to super-resolve regions with structured edges we face two fundamental questions that are at the core ofmany problems in computer vision: 1. [sent-26, score-0.301]

14 A solution in the case of highly structured 3D scenes is to use recent advances in the recovery of low-rank textures [19], defined as low-rank structures deformed by affine or projective transformations. [sent-36, score-0.335]

15 In the case of images with highly structured edges, the sparsity pattern of the gradient tends to follow a lowrank pattern, as illustrated by Figure 1. [sent-38, score-0.184]

16 Learning the domain transform that reveals the lowrank structure of the data. [sent-40, score-0.238]

17 Under the assumptions that the edge structure of the 3D scene is approximately low rank, this data-driven procedure produces a convex regularizer that allows to super-resolve the image very effectively. [sent-43, score-0.37]

18 These example-based methods enforce local consistency to produce sharp-looking edges and are able to hallucinate high-frequency textures very effectively at moderate upsampling factors, especially if prior knowledge about these textures is available [9, 7, 14]. [sent-47, score-0.849]

19 However, they are not well adapted to deal with global features, such as the straight edges in Figure 1. [sent-48, score-0.39]

20 To recapitulate, our main contribution is a principled methodology for the super-resolution of planar regions with regular structures, which achieves high-quality results at upsampling factors that are problematic for other methods. [sent-51, score-0.538]

21 Directional total variation Consider the problem of designing a regularizer adapted to the problem of super-resolving images with sharp edges oriented in a few main directions. [sent-59, score-0.582]

22 1 norm of the gradient [11], also known as the total variation (TV) of the image, or related non-convex penalties [10] in order to obtain an estimate with a sparse gradient (see also [15] for a recent approach that takes discretization into account). [sent-61, score-0.19]

23 Unfortunately, minimizing the total variation often fails to superresolve two-dimensional edges, even in the case of very simple piecewise-constant images such as the checkerboard shown in Figure 3. [sent-62, score-0.13]

24 This failure is largely due to the fact that the regularizer is agnostic to the orientation of the edges in the image, and in particular to the correlation between the orientation of nearby edges. [sent-63, score-0.377]

25 This suggests resorting to a regularizer that is better adapted to the high-level structure of the image gradient. [sent-64, score-0.257]

26 Let us assume that, as is the case for the checkerboard in Figure 3, we happen to know the directions of most edges in the image. [sent-65, score-0.333]

27 In this case, the gradient in the image is not only sparse, but group sparse [18], since its nonzero elements are grouped along horizontal and vertical lines. [sent-66, score-0.246]

28 As a result, a more suitable regularizer is the directional total variation (DTV) of the image, defined as DTV(I) =x? [sent-67, score-0.296]

29 It is designed to favor edges that are aligned horizontally and vertically. [sent-83, score-0.337]

30 A similar regularizer has been proposed for multiple change-point detection in time-series analysis [3]. [sent-84, score-0.181]

31 The top of Figure 3 compares the results of minimizing the TV and DTV cost functions to perform nonblind deblurring of a checkerboard image. [sent-85, score-0.251]

32 Transform-invariant regularization 21 Gaussian The DTV regularizer proposed in Section 2. [sent-93, score-0.276]

33 We seldom encounter images where the edges are perfectly aligned horizontally and vertically. [sent-95, score-0.337]

34 We consider images such that there exists an affine or projective transform τ for which most of the edges of I◦τ are aligned vertically and horizontally. [sent-97, score-0.536]

35 eInd general yiitn igs not straightforward to design a regularizer adapted to such a model. [sent-100, score-0.257]

36 The reason is that the gradient is no longer group sparse along a few main directions. [sent-101, score-0.142]

37 However, it is group sparse modulo the transform τ. [sent-102, score-0.208]

38 2) × a cost function that promotes straight edges in the transformed image, where TI-DTV stands for transforminvariant directional total variation. [sent-104, score-0.47]

39 2) is invariant to affine or projective transforms of I, as long as we are able to estimate them a priori. [sent-106, score-0.117]

40 As we will see, this allows to factor out significant deformations induced by the camera projection. [sent-107, score-0.148]

41 At the bottom of Figure 4 we can see the results of minimizing the TI-DTV cost function to perform nonblind deblurring of a tilted checkerboard. [sent-108, score-0.158]

42 Transform-invariant low-rank textures 21 In order to use the cost function proposed in (2. [sent-116, score-0.195]

43 2), it is necessary to learn a transform “τ” mapping the image to a domain where its edges are mostly aligned vertically and horizontally. [sent-117, score-0.464]

44 We propose doing this by exploiting the fact that images with vertical and horizontal edges tend to be approximately low rank when viewed as a matrix. [sent-118, score-0.35]

45 This is obviously the case for the checkerboard in Figure 4, but holds much more broadly. [sent-119, score-0.093]

46 The main edges are indeed aligned horizontally and vertically by the transformation associated to the low-rank texture. [sent-121, score-0.416]

47 An image with low-rank structured edges might lose its sharpness at low resolutions, but it remains an approximately low-rank texture. [sent-122, score-0.314]

48 This is crucial for our interests, since we can consequently use TILT to learn the transform associated to the edge structure and then apply the regularizer proposed in Section 2. [sent-123, score-0.339]

49 The authors of [19] develop robust computational tools that allow to extract low-rank textures distorted by affine or projective transformations. [sent-126, score-0.353]

50 In essence, TILT al33333381 lows to compute a transform τ from an image I such that I τ = L + E, where L is a low-rank matrix and E ac◦ cIo ◦un τts = =for L sparse d wevheiaretio Lns i sfro am lo twh-er alnowk- mraantkri xm oanddel. [sent-128, score-0.203]

51 We refer to [19] for more details on transform-invariant low-rank textures and on how to solve Problem (2. [sent-135, score-0.15]

52 3), but we would like to mention that the presence of the sparse term E is of vital importance if we apply TILT to retrieve low-rank textures from low-resolution inputs. [sent-136, score-0.24]

53 The reason is that it accounts for artifacts caused by blur and pixelation. [sent-137, score-0.105]

54 This is illustrated by Figure 6, which shows the low-rank and sparse components obtained from a blurry image. [sent-138, score-0.147]

55 Super-resolution via TI-DTV regularization We finally have all the tools to tackle the problem of super-resolving an image obtained from a 3D scene with structured edges. [sent-144, score-0.216]

56 We propose to leverage a data-driven convex regularizer adapted to the 3D geometry revealed by TILT. [sent-145, score-0.316]

57 4) for a downsampling operator D : RN1×N2 → Rn1×n2 and a blurring ksaemrnpelli nKg ∈p rRatNo1r × DN2 :. [sent-147, score-0.146]

58 We apply TILT to the low-resolution image ILR in order to obtain a transform τ which reveals the low-rank edge structure of the image. [sent-150, score-0.201]

59 In practice, we upsample ILR using bicubic interpolation before learning the transformation. [sent-151, score-0.39]

60 5) where λ and β are regularization parameters,TV represents the usual total variation operator for discrete images (i. [sent-181, score-0.186]

61 the sum of the horizontal and vertical finite differences) and Aτ is a linear operator that maps the image to the domain where we seek to penalize the directional total variation. [sent-183, score-0.281]

62 For color images we apply this procedure to the illuminance channel and upsample the chrominance components Cb and Cr using bicubic interpolation. [sent-184, score-0.362]

63 In practice, we use a Gaussian kernel with a standard deviation σ slightly greater than the upsampling factor divided by two. [sent-191, score-0.346]

64 This framework consists in casting the problem as a conic program, determining the dual problem and applying a firstorder method, such as accelerated gradient descent, to solve a smoothed version of the dual. [sent-199, score-0.133]

65 5) by implementing functions to apply the operator Aτ and to compute the dual of the mixed ? [sent-204, score-0.163]

66 This can be done by implementing Aτ with a sparse matrix that samples the transformed image on a new grid using bilinear interpolation. [sent-211, score-0.096]

67 5), another option to super-resolve our class of images of interest is to work directly in the rectified domain penalizing the DTV norm. [sent-226, score-0.1]

68 More precisely, one can compute the rectified lowresolution image ILτR = ILR ◦ τ and then solve the optimiza- tion problem mI˜HτiRn? [sent-227, score-0.135]

69 Adequate modeling of the downsampling operator is crucial to super-resolve effectively [12], so it not surprising that the results for this alternative method are not as sharp as those obtained with TI-DTV regularization, as shown in Figure 7. [sent-249, score-0.144]

70 6) to super-resolve at an upsampling factor of 8 using geometric information obtained from the low-resolution image. [sent-253, score-0.346]

71 We focus on qualitative comparisons, since there is no clear metric capable of quantifying the quality of super-resolved images (for instance, at high upsampling factors the mean-square error can be better for blurry images that do not enhance any features of interest). [sent-263, score-0.482]

72 In our first example, we take large planar regions from five images in the SUN database [16], shown in Figure 8, and compare our method with other representative super-resolution methods developed in the literature. [sent-265, score-0.142]

73 Although we apply the algorithms to the whole planar region, zoomed-in areas are shown due to space limitations. [sent-266, score-0.138]

74 Interpolation algorithms are represented by bicubic interpolation, which we compute using the Matlab function imresize. [sent-267, score-0.249]

75 eter, since TV regularization is usually the method of choice to promote sharp edges in image processing. [sent-289, score-0.343]

76 It is actually noted in [17] that exemplar-based methods have difficulties dealing with highly structured textures such as building facades, because it is difficult to build a dictionary that can exhaust edges in all directions and scales. [sent-292, score-0.458]

77 Planar regions with structured edges extracted from images in the SUN database [16]. [sent-295, score-0.301]

78 33333403 online for this algorithm, which allows to apply an upsampling factor of4. [sent-297, score-0.379]

79 For the rest ofthe methods, including ours, we apply an upsampling factor of 8. [sent-298, score-0.379]

80 In all cases bicubic interpolation produces images that are very blurry. [sent-300, score-0.379]

81 The results for TV regularization are sharper, but they contain significant artifacts which make edges appear wobbly instead of straight. [sent-301, score-0.359]

82 Despite its reduced upsampling factor, the sparse-coding algorithm is also not capable of super-resolving edges effectively and its results are only slightly better than those of bicubic interpolation. [sent-302, score-0.78]

83 In contrast, TI-DTV regularization produces clear straight edges that correspond to the global geometry of the planar surface, yielding upsampled images that are significantly sharper than the rest. [sent-303, score-0.669]

84 As expected the edges align mostly horizontally and vertically following the low-rank structure. [sent-305, score-0.354]

85 For the top example in Figure 12, where the lowresolution image has size 120x136, the running time required by TI-DTV regularization is of 123. [sent-306, score-0.175]

86 Super-resolution of text Text follows the model that we consider to some extent, since most letters contain horizontal or vertical edges. [sent-313, score-0.104]

87 As a result, TI-DTV regularization is capable of effectively super-resolving letter or characters printed on distorted surfaces. [sent-314, score-0.174]

88 To demonstrate this, Figure 11 shows four such examples and compares the results obtained from bicubic interpolation and our method. [sent-315, score-0.342]

89 TI-DTV regularization is clearly superior in all cases, despite the fact that some letters have edges that are not aligned with the low-rank structure (see the following section). [sent-317, score-0.353]

90 Limitations As we have made clear throughout this paper, our method is geared to the super-resolution of planar surfaces that are approximately low-rank and have straight edges that are oriented following the low-rank structure. [sent-320, score-0.504]

91 If these conditions are not met, the algorithm might produce artifacts in regions that resemble horizontal or vertical edges. [sent-321, score-0.209]

92 In any case, the method often degrades gracefully in regions that do not have straight edges with the right orientation. [sent-324, score-0.351]

93 In Figure 10, for example, TI-DTV does not produce any artifacts around the arc and at the same time super-resolves sharply the edges in the rest of the image. [sent-325, score-0.298]

94 In comparison, TV makes the arc look sharper but generates obvious artifacts. [sent-326, score-0.099]

95 Bicubic TV TI-DTV that does not completely conform to the transform-invariant lowrank model using bicubic interpolation, TV regularization and TIDTV regularization. [sent-329, score-0.412]

96 Conclusion and extensions We believe that developing tools capable of constraining non-local image structure is a crucial step towards achieving high-quality super-resolution at large upsampling factors. [sent-332, score-0.388]

97 Our contributions are the introduction of a principled methodology in which such constraints are imposed through data-driven non-parametric regularizers and a robust implementation of this methodology for a particular class of images, which yields state-of-the-art results. [sent-333, score-0.161]

98 Templates for convex cone problems with applications to sparse signal recovery. [sent-362, score-0.116]

99 Results from super-resolving the images in Figure 8 using bicubic interpolation, total-variation regularization, sparse coding [17] and our proposed algorithm. [sent-495, score-0.306]

100 The upsampling factor was 4 for sparse coding and 8 for the rest of the methods. [sent-496, score-0.403]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('dtv', 0.327), ('upsampling', 0.289), ('ilr', 0.254), ('bicubic', 0.249), ('tv', 0.205), ('tilt', 0.198), ('edges', 0.196), ('regularizer', 0.181), ('tfocs', 0.181), ('textures', 0.15), ('straight', 0.118), ('srf', 0.109), ('planar', 0.105), ('hr', 0.105), ('regularization', 0.095), ('interpolation', 0.093), ('checkerboard', 0.093), ('blurry', 0.09), ('cand', 0.086), ('transform', 0.082), ('lowresolution', 0.08), ('horizontally', 0.079), ('vertically', 0.079), ('directional', 0.078), ('adapted', 0.076), ('projective', 0.074), ('fgranda', 0.073), ('hirn', 0.073), ('nonblind', 0.073), ('artifacts', 0.068), ('structured', 0.068), ('lowrank', 0.068), ('sharper', 0.065), ('hallucinate', 0.064), ('aligned', 0.062), ('regularizers', 0.061), ('convex', 0.059), ('sparse', 0.057), ('factors', 0.057), ('factor', 0.057), ('rectified', 0.055), ('operator', 0.054), ('blurring', 0.054), ('deformations', 0.054), ('vertical', 0.054), ('upsampled', 0.053), ('tools', 0.053), ('sharp', 0.052), ('approximately', 0.05), ('methodology', 0.05), ('horizontal', 0.05), ('urban', 0.048), ('upsample', 0.048), ('conic', 0.048), ('ford', 0.048), ('gradient', 0.048), ('capable', 0.046), ('cost', 0.045), ('domain', 0.045), ('promoting', 0.045), ('directions', 0.044), ('superresolution', 0.044), ('affine', 0.043), ('reveals', 0.043), ('edge', 0.043), ('sun', 0.042), ('deblurring', 0.04), ('designing', 0.04), ('implementing', 0.039), ('downsampling', 0.038), ('mi', 0.037), ('regions', 0.037), ('variation', 0.037), ('produces', 0.037), ('blur', 0.037), ('dual', 0.037), ('induced', 0.037), ('group', 0.037), ('super', 0.036), ('stanford', 0.035), ('surfaces', 0.035), ('mathematical', 0.035), ('arc', 0.034), ('apply', 0.033), ('distorted', 0.033), ('stands', 0.033), ('solvers', 0.033), ('tog', 0.032), ('blurred', 0.032), ('hallucinating', 0.032), ('kme', 0.032), ('sfro', 0.032), ('chrominance', 0.032), ('mraantkri', 0.032), ('otfo', 0.032), ('modulo', 0.032), ('anford', 0.032), ('truck', 0.032), ('lns', 0.032), ('tracted', 0.032)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.99999964 408 iccv-2013-Super-resolution via Transform-Invariant Group-Sparse Regularization

Author: Carlos Fernandez-Granda, Emmanuel J. Candès

Abstract: We present a framework to super-resolve planar regions found in urban scenes and other man-made environments by taking into account their 3D geometry. Such regions have highly structured straight edges, but this prior is challenging to exploit due to deformations induced by the projection onto the imaging plane. Our method factors out such deformations by using recently developed tools based on convex optimization to learn a transform that maps the image to a domain where its gradient has a simple group-sparse structure. This allows to obtain a novel convex regularizer that enforces global consistency constraints between the edges of the image. Computational experiments with real images show that this data-driven approach to the design of regularizers promoting transform-invariant group sparsity is very effective at high super-resolution factors. We view our approach as complementary to most recent superresolution methods, which tend to focus on hallucinating high-frequency textures.

2 0.20754518 209 iccv-2013-Image Guided Depth Upsampling Using Anisotropic Total Generalized Variation

Author: David Ferstl, Christian Reinbacher, Rene Ranftl, Matthias Ruether, Horst Bischof

Abstract: In this work we present a novel method for the challenging problem of depth image upsampling. Modern depth cameras such as Kinect or Time of Flight cameras deliver dense, high quality depth measurements but are limited in their lateral resolution. To overcome this limitation we formulate a convex optimization problem using higher order regularization for depth image upsampling. In this optimization an anisotropic diffusion tensor, calculated from a high resolution intensity image, is used to guide the upsampling. We derive a numerical algorithm based on a primaldual formulation that is efficiently parallelized and runs at multiple frames per second. We show that this novel upsampling clearly outperforms state of the art approaches in terms of speed and accuracy on the widely used Middlebury 2007 datasets. Furthermore, we introduce novel datasets with highly accurate groundtruth, which, for the first time, enable to benchmark depth upsampling methods using real sensor data.

3 0.20205586 156 iccv-2013-Fast Direct Super-Resolution by Simple Functions

Author: Chih-Yuan Yang, Ming-Hsuan Yang

Abstract: The goal of single-image super-resolution is to generate a high-quality high-resolution image based on a given low-resolution input. It is an ill-posed problem which requires exemplars or priors to better reconstruct the missing high-resolution image details. In this paper, we propose to split the feature space into numerous subspaces and collect exemplars to learn priors for each subspace, thereby creating effective mapping functions. The use of split input space facilitates both feasibility of using simple functionsfor super-resolution, and efficiency ofgenerating highresolution results. High-quality high-resolution images are reconstructed based on the effective learned priors. Experimental results demonstrate that theproposed algorithmperforms efficiently and effectively over state-of-the-art methods.

4 0.14714007 35 iccv-2013-Accurate Blur Models vs. Image Priors in Single Image Super-resolution

Author: Netalee Efrat, Daniel Glasner, Alexander Apartsin, Boaz Nadler, Anat Levin

Abstract: Over the past decade, single image Super-Resolution (SR) research has focused on developing sophisticated image priors, leading to significant advances. Estimating and incorporating the blur model, that relates the high-res and low-res images, has received much less attention, however. In particular, the reconstruction constraint, namely that the blurred and downsampled high-res output should approximately equal the low-res input image, has been either ignored or applied with default fixed blur models. In this work, we examine the relative importance ofthe imageprior and the reconstruction constraint. First, we show that an accurate reconstruction constraint combined with a simple gradient regularization achieves SR results almost as good as those of state-of-the-art algorithms with sophisticated image priors. Second, we study both empirically and theoretically the sensitivity of SR algorithms to the blur model assumed in the reconstruction constraint. We find that an accurate blur model is more important than a sophisticated image prior. Finally, using real camera data, we demonstrate that the default blur models of various SR algorithms may differ from the camera blur, typically leading to over- smoothed results. Our findings highlight the importance of accurately estimating camera blur in reconstructing raw low- res images acquired by an actual camera.

5 0.11577915 103 iccv-2013-Deblurring by Example Using Dense Correspondence

Author: Yoav Hacohen, Eli Shechtman, Dani Lischinski

Abstract: This paper presents a new method for deblurring photos using a sharp reference example that contains some shared content with the blurry photo. Most previous deblurring methods that exploit information from other photos require an accurately registered photo of the same static scene. In contrast, our method aims to exploit reference images where the shared content may have undergone substantial photometric and non-rigid geometric transformations, as these are the kind of reference images most likely to be found in personal photo albums. Our approach builds upon a recent method for examplebased deblurring using non-rigid dense correspondence (NRDC) [11] and extends it in two ways. First, we suggest exploiting information from the reference image not only for blur kernel estimation, but also as a powerful local prior for the non-blind deconvolution step. Second, we introduce a simple yet robust technique for spatially varying blur estimation, rather than assuming spatially uniform blur. Unlike the aboveprevious method, which hasproven successful only with simple deblurring scenarios, we demonstrate that our method succeeds on a variety of real-world examples. We provide quantitative and qualitative evaluation of our method and show that it outperforms the state-of-the-art.

6 0.10495016 129 iccv-2013-Dynamic Scene Deblurring

7 0.10410433 18 iccv-2013-A Joint Intensity and Depth Co-sparse Analysis Model for Depth Map Super-resolution

8 0.093795024 174 iccv-2013-Forward Motion Deblurring

9 0.08799386 12 iccv-2013-A General Dense Image Matching Framework Combining Direct and Feature-Based Costs

10 0.08781863 51 iccv-2013-Anchored Neighborhood Regression for Fast Example-Based Super-Resolution

11 0.086327121 45 iccv-2013-Affine-Constrained Group Sparse Coding and Its Application to Image-Based Classifications

12 0.082793124 404 iccv-2013-Structured Forests for Fast Edge Detection

13 0.080072112 421 iccv-2013-Total Variation Regularization for Functions with Values in a Manifold

14 0.077681489 82 iccv-2013-Compensating for Motion during Direct-Global Separation

15 0.076532193 26 iccv-2013-A Practical Transfer Learning Algorithm for Face Verification

16 0.074199691 256 iccv-2013-Locally Affine Sparse-to-Dense Matching for Motion and Occlusion Estimation

17 0.072882876 343 iccv-2013-Real-World Normal Map Capture for Nearly Flat Reflective Surfaces

18 0.071810134 422 iccv-2013-Toward Guaranteed Illumination Models for Non-convex Objects

19 0.070489012 293 iccv-2013-Nonparametric Blind Super-resolution

20 0.070442796 271 iccv-2013-Modeling the Calibration Pipeline of the Lytro Camera for High Quality Light-Field Image Reconstruction


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.181), (1, -0.075), (2, -0.061), (3, -0.017), (4, -0.097), (5, 0.021), (6, -0.021), (7, -0.086), (8, 0.028), (9, -0.101), (10, -0.001), (11, -0.091), (12, 0.04), (13, -0.069), (14, -0.017), (15, 0.046), (16, -0.017), (17, -0.034), (18, -0.028), (19, -0.009), (20, 0.003), (21, 0.065), (22, -0.016), (23, -0.063), (24, 0.015), (25, 0.063), (26, 0.048), (27, -0.062), (28, -0.055), (29, -0.069), (30, -0.014), (31, -0.018), (32, 0.004), (33, 0.064), (34, -0.036), (35, -0.023), (36, -0.075), (37, -0.021), (38, 0.042), (39, -0.018), (40, 0.124), (41, -0.083), (42, 0.042), (43, 0.058), (44, -0.049), (45, 0.034), (46, -0.094), (47, 0.101), (48, -0.026), (49, 0.015)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.92026305 408 iccv-2013-Super-resolution via Transform-Invariant Group-Sparse Regularization

Author: Carlos Fernandez-Granda, Emmanuel J. Candès

Abstract: We present a framework to super-resolve planar regions found in urban scenes and other man-made environments by taking into account their 3D geometry. Such regions have highly structured straight edges, but this prior is challenging to exploit due to deformations induced by the projection onto the imaging plane. Our method factors out such deformations by using recently developed tools based on convex optimization to learn a transform that maps the image to a domain where its gradient has a simple group-sparse structure. This allows to obtain a novel convex regularizer that enforces global consistency constraints between the edges of the image. Computational experiments with real images show that this data-driven approach to the design of regularizers promoting transform-invariant group sparsity is very effective at high super-resolution factors. We view our approach as complementary to most recent superresolution methods, which tend to focus on hallucinating high-frequency textures.

2 0.73483878 156 iccv-2013-Fast Direct Super-Resolution by Simple Functions

Author: Chih-Yuan Yang, Ming-Hsuan Yang

Abstract: The goal of single-image super-resolution is to generate a high-quality high-resolution image based on a given low-resolution input. It is an ill-posed problem which requires exemplars or priors to better reconstruct the missing high-resolution image details. In this paper, we propose to split the feature space into numerous subspaces and collect exemplars to learn priors for each subspace, thereby creating effective mapping functions. The use of split input space facilitates both feasibility of using simple functionsfor super-resolution, and efficiency ofgenerating highresolution results. High-quality high-resolution images are reconstructed based on the effective learned priors. Experimental results demonstrate that theproposed algorithmperforms efficiently and effectively over state-of-the-art methods.

3 0.69324982 35 iccv-2013-Accurate Blur Models vs. Image Priors in Single Image Super-resolution

Author: Netalee Efrat, Daniel Glasner, Alexander Apartsin, Boaz Nadler, Anat Levin

Abstract: Over the past decade, single image Super-Resolution (SR) research has focused on developing sophisticated image priors, leading to significant advances. Estimating and incorporating the blur model, that relates the high-res and low-res images, has received much less attention, however. In particular, the reconstruction constraint, namely that the blurred and downsampled high-res output should approximately equal the low-res input image, has been either ignored or applied with default fixed blur models. In this work, we examine the relative importance ofthe imageprior and the reconstruction constraint. First, we show that an accurate reconstruction constraint combined with a simple gradient regularization achieves SR results almost as good as those of state-of-the-art algorithms with sophisticated image priors. Second, we study both empirically and theoretically the sensitivity of SR algorithms to the blur model assumed in the reconstruction constraint. We find that an accurate blur model is more important than a sophisticated image prior. Finally, using real camera data, we demonstrate that the default blur models of various SR algorithms may differ from the camera blur, typically leading to over- smoothed results. Our findings highlight the importance of accurately estimating camera blur in reconstructing raw low- res images acquired by an actual camera.

4 0.63215178 364 iccv-2013-SGTD: Structure Gradient and Texture Decorrelating Regularization for Image Decomposition

Author: Qiegen Liu, Jianbo Liu, Pei Dong, Dong Liang

Abstract: This paper presents a novel structure gradient and texture decorrelating regularization (SGTD) for image decomposition. The motivation of the idea is under the assumption that the structure gradient and texture components should be properly decorrelated for a successful decomposition. The proposed model consists of the data fidelity term, total variation regularization and the SGTD regularization. An augmented Lagrangian method is proposed to address this optimization issue, by first transforming the unconstrained problem to an equivalent constrained problem and then applying an alternating direction method to iteratively solve the subproblems. Experimental results demonstrate that the proposed method presents better or comparable performance as state-of-the-art methods do.

5 0.62824994 135 iccv-2013-Efficient Image Dehazing with Boundary Constraint and Contextual Regularization

Author: Gaofeng Meng, Ying Wang, Jiangyong Duan, Shiming Xiang, Chunhong Pan

Abstract: unkown-abstract

6 0.62338281 98 iccv-2013-Cross-Field Joint Image Restoration via Scale Map

7 0.57995898 112 iccv-2013-Detecting Irregular Curvilinear Structures in Gray Scale and Color Imagery Using Multi-directional Oriented Flux

8 0.57904702 293 iccv-2013-Nonparametric Blind Super-resolution

9 0.56557637 271 iccv-2013-Modeling the Calibration Pipeline of the Lytro Camera for High Quality Light-Field Image Reconstruction

10 0.56000209 292 iccv-2013-Non-convex P-Norm Projection for Robust Sparsity

11 0.55955738 18 iccv-2013-A Joint Intensity and Depth Co-sparse Analysis Model for Depth Map Super-resolution

12 0.54862857 55 iccv-2013-Automatic Kronecker Product Model Based Detection of Repeated Patterns in 2D Urban Images

13 0.5483734 14 iccv-2013-A Generalized Iterated Shrinkage Algorithm for Non-convex Sparse Coding

14 0.54496843 413 iccv-2013-Target-Driven Moire Pattern Synthesis by Phase Modulation

15 0.51456338 148 iccv-2013-Example-Based Facade Texture Synthesis

16 0.51217127 51 iccv-2013-Anchored Neighborhood Regression for Fast Example-Based Super-Resolution

17 0.51074284 96 iccv-2013-Coupled Dictionary and Feature Space Learning with Applications to Cross-Domain Image Synthesis and Recognition

18 0.50521398 312 iccv-2013-Perceptual Fidelity Aware Mean Squared Error

19 0.49678025 407 iccv-2013-Subpixel Scanning Invariant to Indirect Lighting Using Quadratic Code Length

20 0.49432355 422 iccv-2013-Toward Guaranteed Illumination Models for Non-convex Objects


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(2, 0.037), (7, 0.013), (26, 0.065), (31, 0.506), (42, 0.096), (64, 0.024), (73, 0.029), (89, 0.155)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.90600336 345 iccv-2013-Recognizing Text with Perspective Distortion in Natural Scenes

Author: Trung Quy Phan, Palaiahnakote Shivakumara, Shangxuan Tian, Chew Lim Tan

Abstract: This paper presents an approach to text recognition in natural scene images. Unlike most existing works which assume that texts are horizontal and frontal parallel to the image plane, our method is able to recognize perspective texts of arbitrary orientations. For individual character recognition, we adopt a bag-of-keypoints approach, in which Scale Invariant Feature Transform (SIFT) descriptors are extracted densely and quantized using a pre-trained vocabulary. Following [1, 2], the context information is utilized through lexicons. We formulate word recognition as finding the optimal alignment between the set of characters and the list of lexicon words. Furthermore, we introduce a new dataset called StreetViewText-Perspective, which contains texts in street images with a great variety of viewpoints. Experimental results on public datasets and the proposed dataset show that our method significantly outperforms the state-of-the-art on perspective texts of arbitrary orientations.

same-paper 2 0.87411302 408 iccv-2013-Super-resolution via Transform-Invariant Group-Sparse Regularization

Author: Carlos Fernandez-Granda, Emmanuel J. Candès

Abstract: We present a framework to super-resolve planar regions found in urban scenes and other man-made environments by taking into account their 3D geometry. Such regions have highly structured straight edges, but this prior is challenging to exploit due to deformations induced by the projection onto the imaging plane. Our method factors out such deformations by using recently developed tools based on convex optimization to learn a transform that maps the image to a domain where its gradient has a simple group-sparse structure. This allows to obtain a novel convex regularizer that enforces global consistency constraints between the edges of the image. Computational experiments with real images show that this data-driven approach to the design of regularizers promoting transform-invariant group sparsity is very effective at high super-resolution factors. We view our approach as complementary to most recent superresolution methods, which tend to focus on hallucinating high-frequency textures.

3 0.84167439 72 iccv-2013-Characterizing Layouts of Outdoor Scenes Using Spatial Topic Processes

Author: Dahua Lin, Jianxiong Xiao

Abstract: In this paper, we develop a generative model to describe the layouts of outdoor scenes the spatial configuration of regions. Specifically, the layout of an image is represented as a composite of regions, each associated with a semantic topic. At the heart of this model is a novel stochastic process called Spatial Topic Process, which generates a spatial map of topics from a set of coupled Gaussian processes, thus allowing the distributions of topics to vary continuously across the image plane. A key aspect that distinguishes this model from previous ones consists in its capability of capturing dependencies across both locations and topics while allowing substantial variations in the layouts. We demonstrate the practical utility of the proposed model by testing it on scene classification, semantic segmentation, and layout hallucination. –

4 0.82430422 357 iccv-2013-Robust Matrix Factorization with Unknown Noise

Author: Deyu Meng, Fernando De_La_Torre

Abstract: Many problems in computer vision can be posed as recovering a low-dimensional subspace from highdimensional visual data. Factorization approaches to lowrank subspace estimation minimize a loss function between an observed measurement matrix and a bilinear factorization. Most popular loss functions include the L2 and L1 losses. L2 is optimal for Gaussian noise, while L1 is for Laplacian distributed noise. However, real data is often corrupted by an unknown noise distribution, which is unlikely to be purely Gaussian or Laplacian. To address this problem, this paper proposes a low-rank matrix factorization problem with a Mixture of Gaussians (MoG) noise model. The MoG model is a universal approximator for any continuous distribution, and hence is able to model a wider range of noise distributions. The parameters of the MoG model can be estimated with a maximum likelihood method, while the subspace is computed with standard approaches. We illustrate the benefits of our approach in extensive syn- thetic and real-world experiments including structure from motion, face modeling and background subtraction.

5 0.81599569 38 iccv-2013-Action Recognition with Actons

Author: Jun Zhu, Baoyuan Wang, Xiaokang Yang, Wenjun Zhang, Zhuowen Tu

Abstract: With the improved accessibility to an exploding amount of video data and growing demands in a wide range of video analysis applications, video-based action recognition/classification becomes an increasingly important task in computer vision. In this paper, we propose a two-layer structure for action recognition to automatically exploit a mid-level “acton ” representation. The weakly-supervised actons are learned via a new max-margin multi-channel multiple instance learning framework, which can capture multiple mid-level action concepts simultaneously. The learned actons (with no requirement for detailed manual annotations) observe theproperties ofbeing compact, informative, discriminative, and easy to scale. The experimental results demonstrate the effectiveness ofapplying the learned actons in our two-layer structure, and show the state-ofthe-art recognition performance on two challenging action datasets, i.e., Youtube and HMDB51.

6 0.7234112 275 iccv-2013-Motion-Aware KNN Laplacian for Video Matting

7 0.68209589 269 iccv-2013-Modeling Occlusion by Discriminative AND-OR Structures

8 0.62741601 73 iccv-2013-Class-Specific Simplex-Latent Dirichlet Allocation for Image Classification

9 0.6215151 376 iccv-2013-Scene Text Localization and Recognition with Oriented Stroke Detection

10 0.59562576 137 iccv-2013-Efficient Salient Region Detection with Soft Image Abstraction

11 0.58926082 210 iccv-2013-Image Retrieval Using Textual Cues

12 0.58922148 315 iccv-2013-PhotoOCR: Reading Text in Uncontrolled Conditions

13 0.56395936 180 iccv-2013-From Where and How to What We See

14 0.55817902 415 iccv-2013-Text Localization in Natural Images Using Stroke Feature Transform and Text Covariance Descriptors

15 0.55433846 173 iccv-2013-Fluttering Pattern Generation Using Modified Legendre Sequence for Coded Exposure Imaging

16 0.55021191 19 iccv-2013-A Learning-Based Approach to Reduce JPEG Artifacts in Image Matting

17 0.54486406 192 iccv-2013-Handwritten Word Spotting with Corrected Attributes

18 0.54248315 287 iccv-2013-Neighbor-to-Neighbor Search for Fast Coding of Feature Vectors

19 0.54148388 156 iccv-2013-Fast Direct Super-Resolution by Simple Functions

20 0.53720129 328 iccv-2013-Probabilistic Elastic Part Model for Unsupervised Face Detector Adaptation