nips nips2011 nips2011-275 knowledge-graph by maker-knowledge-mining

275 nips-2011-Structured Learning for Cell Tracking


Source: pdf

Author: Xinghua Lou, Fred A. Hamprecht

Abstract: We study the problem of learning to track a large quantity of homogeneous objects such as cell tracking in cell culture study and developmental biology. Reliable cell tracking in time-lapse microscopic image sequences is important for modern biomedical research. Existing cell tracking methods are usually kept simple and use only a small number of features to allow for manual parameter tweaking or grid search. We propose a structured learning approach that allows to learn optimum parameters automatically from a training set. This allows for the use of a richer set of features which in turn affords improved tracking compared to recently reported methods on two public benchmark sequences. 1

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 de Abstract We study the problem of learning to track a large quantity of homogeneous objects such as cell tracking in cell culture study and developmental biology. [sent-6, score-0.915]

2 Reliable cell tracking in time-lapse microscopic image sequences is important for modern biomedical research. [sent-7, score-0.651]

3 Existing cell tracking methods are usually kept simple and use only a small number of features to allow for manual parameter tweaking or grid search. [sent-8, score-0.834]

4 We propose a structured learning approach that allows to learn optimum parameters automatically from a training set. [sent-9, score-0.209]

5 This allows for the use of a richer set of features which in turn affords improved tracking compared to recently reported methods on two public benchmark sequences. [sent-10, score-0.4]

6 1 Introduction One distinguishing property of life is its temporal dynamics, and it is hence only natural that time lapse experiments play a crucial role in current research on signaling pathways, drug discovery and developmental biology [17]. [sent-11, score-0.102]

7 Such experiments yield a very large number of images, and reliable automated cell tracking emerges naturally as a prerequisite for further quantitative analysis. [sent-12, score-0.602]

8 Even today, cell tracking remains a challenging problem in dense populations, in the presence of complex behavior or when image quality is poor. [sent-13, score-0.651]

9 Existing cell tracking methods can broadly be categorized as deformable models, stochastic filtering and object association. [sent-14, score-0.706]

10 Deformable models combine detection, segmentation and tracking by initializing a set of models (e. [sent-15, score-0.386]

11 active contours) in the first frame and updating them in subsequent frames (e. [sent-17, score-0.179]

12 Object association methods approximate and simplify the problem by separating the detection and association steps: once object candidates have been detected and characterized, a second step suggests associations between object candidates at different frames. [sent-27, score-1.058]

13 This class of methods scales well [21, 16, 13] and allows the tracking of thousands of cells in 3D [19]. [sent-28, score-0.342]

14 This was first accomplished by casting tracking as a local affinity prediction problem such as binary classification with either offline [1] or online learning [11, 5, 15], weakly supervised learning with imperfect oracles [27], manifold appearance model learning [25], or ranking [10, 18]. [sent-31, score-0.382]

15 However, these local methods fail to capture the very important dependency among associations, hence the resulting local affinities do not necessarily guarantee a better global association [26]. [sent-32, score-0.129]

16 To address this limitation, [26] extended the RankBoost method from [18] to rank global associations represented as a Conditional Random Field (CRF). [sent-33, score-0.31]

17 Firstly, it depends on a set of artificially generated false association samples that can make the training data particularly imbalanced and the training procedure too expensive 1 for large-scale tracking problems. [sent-35, score-0.621]

18 We first present an extended formulation of the object association models proposed in the literature. [sent-42, score-0.233]

19 We hence, secondly, propose to use structured learning to automatically learn optimum parameters from a training set, and hence profit fully from this richer description. [sent-44, score-0.209]

20 In section 2, we present the extended object association models and a structured learning approach for global affinity learning. [sent-47, score-0.329]

21 In section 3, an evaluation shows that our framework inherits the runtime advantage of object association while addressing many of its limitations. [sent-48, score-0.233]

22 1 Association Hypotheses and Scoring We assume that a previous detection and segmentation step has identified object candidates in all frames, see Fig. [sent-51, score-0.313]

23 We set out to find that set of object associations that best explains these observations. [sent-53, score-0.414]

24 To this end, we admit the following set E of standard events [21, 13]: a cell can move or divide and it can appear or disappear. [sent-54, score-0.514]

25 In addition, we allow two cells to (seemingly) merge, to account for occlusion or undersegmentation; and a cell can (seemingly) split, to allow for the lifting of occlusion or oversegmentation. [sent-55, score-0.26]

26 These additional hypotheses are useful to account for the errors that typically occur in the detection and segmentation step in crowded or noisy data. [sent-56, score-0.207]

27 The distinction between division and split is reasonable given that typical fluorescence stains endow the anaphase with a distinctive appearance. [sent-57, score-0.153]

28 Given a pair of object candidate lists x = {C, C } in two neighboring frames, there is a multitude of possible association hypotheses, see Fig. [sent-60, score-0.311]

29 We have two tasks: firstly, to allow only consistent associations (e. [sent-62, score-0.31]

30 making sure that each cell in the second frame is accounted for only once); and secondly to identify, among the multitude of consistent hypotheses, the one that is most compatible with the observations, and with what we have learned from the training data. [sent-64, score-0.543]

31 We express this compatibility of the association between c ∈ P(C) and c ∈ P(C ) by event e ∈ E e e as an inner product fc,c we . [sent-65, score-0.169]

32 Here, fc,c is a feature vector that characterizes the discrepancy (if any) between object candidates c and c ; and we is a parameter vector that encodes everything we 2 have learned from the training data. [sent-66, score-0.296]

33 Summing over all object candidates in either of the frames and over all types of events gives the following compatibility function: e e fc,c , we zc,c L(x, z; w) = (1) e∈E c∈P(C) c ∈P(C ) e e zc,c = 1 with zc,c ∈ {0, 1} e zc,c = 1 and s. [sent-67, score-0.441]

34 (2) e∈E c ∈P(C ) e∈E c∈P(C) The constraints in the last line involve binary indicator variables z that reflect the consistency requirements: each candidate in the first frame must have a single fate, and each candidate from the second frame a unique history. [sent-69, score-0.296]

35 As an important technical detail, note that P(C) := C ∪ (C ⊗ C) is a set comprising each object candidate, as well as all ordered pairs of object candidates from a frame1 . [sent-70, score-0.325]

36 This allows us to conveniently subsume cell divisions, splits and mergers in the above equation. [sent-71, score-0.303]

37 the global affinity measure, states how well a set of associations z matches the observations f (x) computed from the raw data x, given the knowledge w from the training set. [sent-74, score-0.385]

38 The remaining tasks, discussed next, are how to learn the parameters w from the training data (section 2. [sent-75, score-0.113]

39 2); given these, how to find the best possible associations z (section 2. [sent-76, score-0.31]

40 2 Structured Max-Margin Parameter Learning In learning the parameters automatically from a training set, we pursue two goals: first, to go beyond manual parameter tweaking in obtaining the best possible performance; and second, to make the process as facile as possible for the user. [sent-80, score-0.249]

41 This is under the assumption that most experimentalists find it easier to specify what a correct tracking should look like, rather than what value a more-or-less obscure parameter should have. [sent-81, score-0.385]

42 ∗ Given N training frame pairs X = {xn } and their correct associations Z ∗ = {zn }, n = 1, . [sent-82, score-0.495]

43 , N , the best set of parameters is the optimizer of arg min R(w; X, Z ∗ ) + λΩ(w) (3) w Here, R(w; X, Z ∗ ) measures the empirical loss of the current parametrization w given the training data X, Z ∗ . [sent-85, score-0.181]

44 The empirical loss is given by N 1 ∗ ˆ ˆ R(w; X, Z ∗ ) = N i=1 ∆(zn , zn (w; xn )). [sent-93, score-0.25]

45 Here ∆(z ∗ , z) is a loss function that measures the discrepancy between a true association z ∗ and a prediction by specifying the fraction of missed events w. [sent-94, score-0.278]

46 Importantly, both the input (objects from a frame pair) and output (associations between objects) in this learning problem are structured. [sent-100, score-0.11]

47 We hence resort to max-margin structured learning [2] to exploit the structure and dependency within the association hypotheses. [sent-101, score-0.225]

48 In comparison to other aforementioned learning methods, structured learning allows us to directly learn the global affinity measure, avoid generating many artificial false association samples, and drop any assumptions on the signs of the features. [sent-102, score-0.263]

49 In particular, we attempt to find the decision boundary that maximizes the margin between the ∗ correct association zn and the closest runner-up solution. [sent-104, score-0.341]

50 3 ∗ that the score of zn be greater than that of any other solution. [sent-107, score-0.212]

51 1 N N n=1 ξn + λΩ(w) ∗ ∗ ˆ ˆ ˆ ∀n, ∀zn ∈ Zn : L(xn , zn ; w) − L(xn , zn ; w) ≥ ∆(zn , zn ) − ξn , (5) ∗ ˆ where Zn is the set of possible consistent associations and ∆(zn , zn ) − ξn is known as “marginrescaling” [24]. [sent-111, score-1.158]

52 Iteratively find, first, the optimum associations for the current ∗ ˆ ˆ w by solving, for all n, zn = arg maxz {L(xn , z; w) + ∆(zn , z)}. [sent-118, score-0.522]

53 Use all these zn to identify the most violated constraint, and add it to Eq. [sent-119, score-0.212]

54 For a given parametrization, the optimum associations can be found by integer linear programming (ILP) [16, 21, 13]. [sent-124, score-0.31]

55 Our framework has been implemented in Matlab and C++, including a labeling GUI for the generation of training set associations, feature extraction, model inference and the bundle method. [sent-125, score-0.138]

56 To reduce the search space and eliminate hypotheses with no prospect of being realized, we constrain the hypotheses to a k-nearest neighborhood with distance thresholding. [sent-126, score-0.144]

57 division and split) and resolve ambiguity in model inference, we need rich features to characterize different events. [sent-132, score-0.156]

58 In additional to basic features such as size/position [21] and intensity histogram [16], we also designed new features such as “shape compactness” for oversegmentation and “angle pattern” for division. [sent-133, score-0.221]

59 Shape compactness relates the summed areas of two object candidates to the area of their union’s convex hull. [sent-134, score-0.33]

60 Features can be defined on a pair of object candidates or on an individual object candidate only. [sent-136, score-0.363]

61 The two datasets show a certain degree of variations such as illumination, cell density and image compression artifacts (Fig. [sent-142, score-0.36]

62 action=show movie;query=243867 3 4 GFP stained cell nuclei were segmented using the method in [19], yielding an F-measure over 99. [sent-152, score-0.303]

63 Full ground truth associations for training and evaluation were generated with a Matlab GUI tool at a rate of approximately 20 frames/hour. [sent-154, score-0.385]

64 The Mitocheck sequence exhibits higher cell density, larger intensity variability and “blockness” artifacts due to image compression. [sent-163, score-0.422]

65 Task 1: Efficient Tracking for a Given Sequence We first evaluate our method on a task that is frequently encountered in practice: the user simply wishes to obtain a good tracking for a given sequence with the smallest possible effort. [sent-164, score-0.342]

66 For a fair comparison, we extended Padfield’s method [21] to account for the six events described in section 2. [sent-165, score-0.111]

67 A detailed analysis of the error counts for specific events shows that the method accounts well for moves, but has difficulty with disappearance and split events. [sent-172, score-0.231]

68 To study the difference between manual tweaking and learning of the parameters, we used the learning framework presented here to optimize the model and obtained a reduction of the total loss from 1. [sent-174, score-0.212]

69 Note that the learned parametrization actually deteriorates the detection of divisions because the learning aims at minimizing the overall loss across all events. [sent-178, score-0.253]

70 With 37 features included and their weights optimized using structured learning, our model fully profits from this richer description and achieves a total loss of only 0. [sent-180, score-0.192]

71 30% (4th row) which is a significant improvement over [21, 16] (2nd/7th row) and manual tweaking (6th row). [sent-181, score-0.174]

72 Though a certain amount of efforts is needed for creating the training set, our method allows experimentalists to contribute their expertise in an intuitive fashion. [sent-182, score-0.118]

73 They afford the following observations: Firstly, features on cell size and shape are generally of high importance, which is in line with the assumption in [21]. [sent-187, score-0.417]

74 Secondly, the correlations of the features with the final association score are 5 Table 3: Performance comparison on the DCellIQ dataset. [sent-188, score-0.187]

75 The header row shows the number of events occurring for moves, divisions, appearance, disappearance, splits and mergers. [sent-189, score-0.246]

76 mov div app dis spl mer total loss 10156 104 78 76 54 55 Padfield et al. [sent-191, score-0.414]

77 Figure 3: Some diverging associations by [21] (top) and our method (bottom). [sent-204, score-0.31]

78 For example, shape compactness is positively correlated with split but negatively with division. [sent-207, score-0.182]

79 This is in line with the intuition that an oversegmentation conserves compact shape, while a true division seemingly pushes the daughters far away from each other (in the present kind of data, where only DNA is labeled). [sent-208, score-0.178]

80 Task 2: Tracking for High-Throughput Experiments The experiment described in the foregoing draws both training and test samples from the same time lapse experiment. [sent-210, score-0.167]

81 To emulate this situation, we have used the parameters w trained in the foregoing on the DCellIQ sequence [16] and used these to estimate the tracking of the Mitocheck dataset. [sent-212, score-0.385]

82 The main focus of the Mitocheck project is on accurate detection of mitosis (cell division). [sent-213, score-0.122]

83 Despite the difference in illumination and cell density from the training data, and despite the segmentation artifacts caused by the compression of the image sequence, our method shows a high generalization capability and obtains a total loss of 0. [sent-214, score-0.517]

84 2% of 384 mitosis events which is a significant improvement over the mitosis detection rate reported in [12] (81. [sent-217, score-0.307]

85 We sample positive associations from the ground truth and randomly generate false associations. [sent-221, score-0.31]

86 The predicted probabilities by the RF classifiers are used to compute the overall association score as in Eq. [sent-223, score-0.129]

87 Since we have multiple competing events (one cell can only have a single 6 mov div app dis spl mer Feature Importance (L2) Importance 0. [sent-226, score-0.747]

88 su fa fa an in d me m th th gl te ev an er er e n. [sent-244, score-0.209]

89 p e i a s sh s cc nt tt um ap iz en en er e e tr si n ov er m co ev ic ty di la as mp en it st p s ac ne y an wi ev tn ss ce th en es n s d t b e diiff d o or ss di f . [sent-246, score-1.374]

90 1 di di ff Importance Feature Importance (L1) 0. [sent-273, score-0.438]

91 Parameters weighing the features for different events are colored differently. [sent-275, score-0.169]

92 The header row shows the number of events occurring for moves, divisions, appearance, disappearance, splits and mergers. [sent-281, score-0.246]

93 mov div app dis spl mer total loss 22520 384 310 304 127 132 Padfield et al. [sent-283, score-0.414]

94 To test the sensitivity of the results to the training data used, we drew different numbers of training image pairs randomly from the entire sequence and used the remaining pairs for testing. [sent-296, score-0.199]

95 According to the one-standard-error-rule, associations between at least 15 or 20 image pairs are desirable, which can be accomplished in well below an hour of annotation work. [sent-300, score-0.359]

96 L2) 25 20 L1 Regularization L2 Regularization 15 10 5 0 10 20 30 40 50 60 70 Number of constraints Number of frame pairs for training Figure 5: Learning curve of structured learning Figure 6: Convergence rates of structured learn(with L2 regularization). [sent-315, score-0.377]

97 4 Conclusion & Future Work We present a new cell tracking scheme that uses more expressive features and comes with a structured learning framework to train the larger number of parameters involved. [sent-318, score-0.756]

98 We currently work on further improvement of the tracking by considering more than two frames at a time, and on an active learning scheme that should reduce the amount of required training inputs. [sent-320, score-0.486]

99 CellCognition: time-resolved phenotype annotation in highthroughput live cell imaging. [sent-404, score-0.26]

100 Cell population tracking and lineage construction with spatiotemporal context. [sent-440, score-0.391]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('tracking', 0.342), ('associations', 0.31), ('cell', 0.26), ('ff', 0.24), ('zn', 0.212), ('mitocheck', 0.198), ('dcelliq', 0.148), ('en', 0.136), ('association', 0.129), ('mach', 0.12), ('candidates', 0.117), ('events', 0.111), ('frame', 0.11), ('tweaking', 0.109), ('pad', 0.109), ('object', 0.104), ('di', 0.099), ('divisions', 0.099), ('division', 0.098), ('nt', 0.098), ('structured', 0.096), ('move', 0.087), ('med', 0.087), ('moves', 0.08), ('te', 0.079), ('af', 0.076), ('training', 0.075), ('mitosis', 0.074), ('spl', 0.074), ('hypotheses', 0.072), ('frames', 0.069), ('parametrization', 0.068), ('compactness', 0.065), ('app', 0.065), ('mov', 0.065), ('disappearance', 0.065), ('mer', 0.065), ('manual', 0.065), ('nity', 0.065), ('bundle', 0.063), ('res', 0.063), ('ss', 0.063), ('intensity', 0.062), ('shape', 0.062), ('anal', 0.06), ('div', 0.06), ('features', 0.058), ('secondly', 0.058), ('ev', 0.056), ('divide', 0.056), ('split', 0.055), ('developmental', 0.053), ('artifacts', 0.051), ('cvpr', 0.05), ('didif', 0.049), ('diiff', 0.049), ('dzyubachyk', 0.049), ('evenness', 0.049), ('executables', 0.049), ('fate', 0.049), ('father', 0.049), ('header', 0.049), ('imag', 0.049), ('lapse', 0.049), ('lineage', 0.049), ('lou', 0.049), ('rankboost', 0.049), ('image', 0.049), ('ap', 0.049), ('detection', 0.048), ('dis', 0.047), ('regularization', 0.046), ('segmentation', 0.044), ('summed', 0.044), ('row', 0.043), ('experimentalists', 0.043), ('nuclei', 0.043), ('io', 0.043), ('gui', 0.043), ('crowded', 0.043), ('oversegmentation', 0.043), ('foregoing', 0.043), ('ilp', 0.043), ('si', 0.043), ('rf', 0.043), ('splits', 0.043), ('heidelberg', 0.042), ('firstly', 0.042), ('sh', 0.041), ('appearance', 0.04), ('compatibility', 0.04), ('multitude', 0.04), ('iz', 0.04), ('loss', 0.038), ('candidate', 0.038), ('learn', 0.038), ('afford', 0.037), ('pushes', 0.037), ('ze', 0.037), ('er', 0.037)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000001 275 nips-2011-Structured Learning for Cell Tracking

Author: Xinghua Lou, Fred A. Hamprecht

Abstract: We study the problem of learning to track a large quantity of homogeneous objects such as cell tracking in cell culture study and developmental biology. Reliable cell tracking in time-lapse microscopic image sequences is important for modern biomedical research. Existing cell tracking methods are usually kept simple and use only a small number of features to allow for manual parameter tweaking or grid search. We propose a structured learning approach that allows to learn optimum parameters automatically from a training set. This allows for the use of a richer set of features which in turn affords improved tracking compared to recently reported methods on two public benchmark sequences. 1

2 0.17845595 180 nips-2011-Multiple Instance Filtering

Author: Kamil A. Wnuk, Stefano Soatto

Abstract: We propose a robust filtering approach based on semi-supervised and multiple instance learning (MIL). We assume that the posterior density would be unimodal if not for the effect of outliers that we do not wish to explicitly model. Therefore, we seek for a point estimate at the outset, rather than a generic approximation of the entire posterior. Our approach can be thought of as a combination of standard finite-dimensional filtering (Extended Kalman Filter, or Unscented Filter) with multiple instance learning, whereby the initial condition comes with a putative set of inlier measurements. We show how both the state (regression) and the inlier set (classification) can be estimated iteratively and causally by processing only the current measurement. We illustrate our approach on visual tracking problems whereby the object of interest (target) moves and evolves as a result of occlusions and deformations, and partial knowledge of the target is given in the form of a bounding box (training set). 1

3 0.13648948 148 nips-2011-Learning Probabilistic Non-Linear Latent Variable Models for Tracking Complex Activities

Author: Angela Yao, Juergen Gall, Luc V. Gool, Raquel Urtasun

Abstract: A common approach for handling the complexity and inherent ambiguities of 3D human pose estimation is to use pose priors learned from training data. Existing approaches however, are either too simplistic (linear), too complex to learn, or can only learn latent spaces from “simple data”, i.e., single activities such as walking or running. In this paper, we present an efficient stochastic gradient descent algorithm that is able to learn probabilistic non-linear latent spaces composed of multiple activities. Furthermore, we derive an incremental algorithm for the online setting which can update the latent space without extensive relearning. We demonstrate the effectiveness of our approach on the task of monocular and multi-view tracking and show that our approach outperforms the state-of-the-art. 1

4 0.12875326 303 nips-2011-Video Annotation and Tracking with Active Learning

Author: Carl Vondrick, Deva Ramanan

Abstract: We introduce a novel active learning framework for video annotation. By judiciously choosing which frames a user should annotate, we can obtain highly accurate tracks with minimal user effort. We cast this problem as one of active learning, and show that we can obtain excellent performance by querying frames that, if annotated, would produce a large expected change in the estimated object track. We implement a constrained tracker and compute the expected change for putative annotations with efficient dynamic programming algorithms. We demonstrate our framework on four datasets, including two benchmark datasets constructed with key frame annotations obtained by Amazon Mechanical Turk. Our results indicate that we could obtain equivalent labels for a small fraction of the original cost. 1

5 0.10966277 1 nips-2011-$\theta$-MRF: Capturing Spatial and Semantic Structure in the Parameters for Scene Understanding

Author: Congcong Li, Ashutosh Saxena, Tsuhan Chen

Abstract: For most scene understanding tasks (such as object detection or depth estimation), the classifiers need to consider contextual information in addition to the local features. We can capture such contextual information by taking as input the features/attributes from all the regions in the image. However, this contextual dependence also varies with the spatial location of the region of interest, and we therefore need a different set of parameters for each spatial location. This results in a very large number of parameters. In this work, we model the independence properties between the parameters for each location and for each task, by defining a Markov Random Field (MRF) over the parameters. In particular, two sets of parameters are encouraged to have similar values if they are spatially close or semantically close. Our method is, in principle, complementary to other ways of capturing context such as the ones that use a graphical model over the labels instead. In extensive evaluation over two different settings, of multi-class object detection and of multiple scene understanding tasks (scene categorization, depth estimation, geometric labeling), our method beats the state-of-the-art methods in all the four tasks. 1

6 0.092906147 35 nips-2011-An ideal observer model for identifying the reference frame of objects

7 0.089360818 255 nips-2011-Simultaneous Sampling and Multi-Structure Fitting with Adaptive Reversible Jump MCMC

8 0.087630346 166 nips-2011-Maximal Cliques that Satisfy Hard Constraints with Application to Deformable Object Model Learning

9 0.085529819 154 nips-2011-Learning person-object interactions for action recognition in still images

10 0.08089488 151 nips-2011-Learning a Tree of Metrics with Disjoint Visual Features

11 0.080577999 197 nips-2011-On Tracking The Partition Function

12 0.079276606 169 nips-2011-Maximum Margin Multi-Label Structured Prediction

13 0.078940012 193 nips-2011-Object Detection with Grammar Models

14 0.078788273 138 nips-2011-Joint 3D Estimation of Objects and Scene Layout

15 0.076885037 247 nips-2011-Semantic Labeling of 3D Point Clouds for Indoor Scenes

16 0.076821618 68 nips-2011-Demixed Principal Component Analysis

17 0.075896718 114 nips-2011-Hierarchical Multitask Structured Output Learning for Large-scale Sequence Segmentation

18 0.075158127 62 nips-2011-Continuous-Time Regression Models for Longitudinal Networks

19 0.071830787 304 nips-2011-Why The Brain Separates Face Recognition From Object Recognition

20 0.068203144 227 nips-2011-Pylon Model for Semantic Segmentation


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.213), (1, 0.09), (2, -0.078), (3, 0.111), (4, 0.052), (5, 0.009), (6, 0.007), (7, -0.066), (8, -0.017), (9, 0.09), (10, 0.018), (11, -0.073), (12, 0.027), (13, -0.026), (14, 0.014), (15, -0.045), (16, 0.021), (17, 0.03), (18, -0.051), (19, 0.043), (20, 0.027), (21, 0.048), (22, -0.008), (23, 0.044), (24, -0.037), (25, -0.053), (26, -0.204), (27, 0.111), (28, -0.073), (29, -0.036), (30, 0.062), (31, -0.048), (32, 0.113), (33, 0.001), (34, -0.055), (35, 0.019), (36, -0.005), (37, -0.082), (38, 0.078), (39, -0.044), (40, -0.019), (41, -0.005), (42, -0.051), (43, 0.053), (44, 0.07), (45, -0.029), (46, -0.047), (47, -0.055), (48, -0.075), (49, -0.035)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.92931587 275 nips-2011-Structured Learning for Cell Tracking

Author: Xinghua Lou, Fred A. Hamprecht

Abstract: We study the problem of learning to track a large quantity of homogeneous objects such as cell tracking in cell culture study and developmental biology. Reliable cell tracking in time-lapse microscopic image sequences is important for modern biomedical research. Existing cell tracking methods are usually kept simple and use only a small number of features to allow for manual parameter tweaking or grid search. We propose a structured learning approach that allows to learn optimum parameters automatically from a training set. This allows for the use of a richer set of features which in turn affords improved tracking compared to recently reported methods on two public benchmark sequences. 1

2 0.8262862 180 nips-2011-Multiple Instance Filtering

Author: Kamil A. Wnuk, Stefano Soatto

Abstract: We propose a robust filtering approach based on semi-supervised and multiple instance learning (MIL). We assume that the posterior density would be unimodal if not for the effect of outliers that we do not wish to explicitly model. Therefore, we seek for a point estimate at the outset, rather than a generic approximation of the entire posterior. Our approach can be thought of as a combination of standard finite-dimensional filtering (Extended Kalman Filter, or Unscented Filter) with multiple instance learning, whereby the initial condition comes with a putative set of inlier measurements. We show how both the state (regression) and the inlier set (classification) can be estimated iteratively and causally by processing only the current measurement. We illustrate our approach on visual tracking problems whereby the object of interest (target) moves and evolves as a result of occlusions and deformations, and partial knowledge of the target is given in the form of a bounding box (training set). 1

3 0.75364667 303 nips-2011-Video Annotation and Tracking with Active Learning

Author: Carl Vondrick, Deva Ramanan

Abstract: We introduce a novel active learning framework for video annotation. By judiciously choosing which frames a user should annotate, we can obtain highly accurate tracks with minimal user effort. We cast this problem as one of active learning, and show that we can obtain excellent performance by querying frames that, if annotated, would produce a large expected change in the estimated object track. We implement a constrained tracker and compute the expected change for putative annotations with efficient dynamic programming algorithms. We demonstrate our framework on four datasets, including two benchmark datasets constructed with key frame annotations obtained by Amazon Mechanical Turk. Our results indicate that we could obtain equivalent labels for a small fraction of the original cost. 1

4 0.66808361 148 nips-2011-Learning Probabilistic Non-Linear Latent Variable Models for Tracking Complex Activities

Author: Angela Yao, Juergen Gall, Luc V. Gool, Raquel Urtasun

Abstract: A common approach for handling the complexity and inherent ambiguities of 3D human pose estimation is to use pose priors learned from training data. Existing approaches however, are either too simplistic (linear), too complex to learn, or can only learn latent spaces from “simple data”, i.e., single activities such as walking or running. In this paper, we present an efficient stochastic gradient descent algorithm that is able to learn probabilistic non-linear latent spaces composed of multiple activities. Furthermore, we derive an incremental algorithm for the online setting which can update the latent space without extensive relearning. We demonstrate the effectiveness of our approach on the task of monocular and multi-view tracking and show that our approach outperforms the state-of-the-art. 1

5 0.62689471 193 nips-2011-Object Detection with Grammar Models

Author: Ross B. Girshick, Pedro F. Felzenszwalb, David A. McAllester

Abstract: Compositional models provide an elegant formalism for representing the visual appearance of highly variable objects. While such models are appealing from a theoretical point of view, it has been difficult to demonstrate that they lead to performance advantages on challenging datasets. Here we develop a grammar model for person detection and show that it outperforms previous high-performance systems on the PASCAL benchmark. Our model represents people using a hierarchy of deformable parts, variable structure and an explicit model of occlusion for partially visible objects. To train the model, we introduce a new discriminative framework for learning structured prediction models from weakly-labeled data. 1

6 0.56124353 255 nips-2011-Simultaneous Sampling and Multi-Structure Fitting with Adaptive Reversible Jump MCMC

7 0.54437208 290 nips-2011-Transfer Learning by Borrowing Examples for Multiclass Object Detection

8 0.53549558 166 nips-2011-Maximal Cliques that Satisfy Hard Constraints with Application to Deformable Object Model Learning

9 0.52891034 169 nips-2011-Maximum Margin Multi-Label Structured Prediction

10 0.52776814 35 nips-2011-An ideal observer model for identifying the reference frame of objects

11 0.52669054 154 nips-2011-Learning person-object interactions for action recognition in still images

12 0.5230096 138 nips-2011-Joint 3D Estimation of Objects and Scene Layout

13 0.52173185 197 nips-2011-On Tracking The Partition Function

14 0.46694085 247 nips-2011-Semantic Labeling of 3D Point Clouds for Indoor Scenes

15 0.45039773 1 nips-2011-$\theta$-MRF: Capturing Spatial and Semantic Structure in the Parameters for Scene Understanding

16 0.44627589 192 nips-2011-Nonstandard Interpretations of Probabilistic Programs for Efficient Inference

17 0.43083557 277 nips-2011-Submodular Multi-Label Learning

18 0.4254961 7 nips-2011-A Machine Learning Approach to Predict Chemical Reactions

19 0.41231298 252 nips-2011-ShareBoost: Efficient multiclass learning with feature sharing

20 0.41092712 279 nips-2011-Target Neighbor Consistent Feature Weighting for Nearest Neighbor Classification


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(0, 0.015), (4, 0.036), (20, 0.499), (26, 0.021), (31, 0.04), (33, 0.024), (43, 0.046), (45, 0.089), (57, 0.026), (74, 0.041), (83, 0.033), (84, 0.013), (99, 0.051)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.89873451 290 nips-2011-Transfer Learning by Borrowing Examples for Multiclass Object Detection

Author: Joseph J. Lim, Antonio Torralba, Ruslan Salakhutdinov

Abstract: Despite the recent trend of increasingly large datasets for object detection, there still exist many classes with few training examples. To overcome this lack of training data for certain classes, we propose a novel way of augmenting the training data for each class by borrowing and transforming examples from other classes. Our model learns which training instances from other classes to borrow and how to transform the borrowed examples so that they become more similar to instances from the target class. Our experimental results demonstrate that our new object detector, with borrowed and transformed examples, improves upon the current state-of-the-art detector on the challenging SUN09 object detection dataset. 1

same-paper 2 0.8929345 275 nips-2011-Structured Learning for Cell Tracking

Author: Xinghua Lou, Fred A. Hamprecht

Abstract: We study the problem of learning to track a large quantity of homogeneous objects such as cell tracking in cell culture study and developmental biology. Reliable cell tracking in time-lapse microscopic image sequences is important for modern biomedical research. Existing cell tracking methods are usually kept simple and use only a small number of features to allow for manual parameter tweaking or grid search. We propose a structured learning approach that allows to learn optimum parameters automatically from a training set. This allows for the use of a richer set of features which in turn affords improved tracking compared to recently reported methods on two public benchmark sequences. 1

3 0.88532639 119 nips-2011-Higher-Order Correlation Clustering for Image Segmentation

Author: Sungwoong Kim, Sebastian Nowozin, Pushmeet Kohli, Chang D. Yoo

Abstract: For many of the state-of-the-art computer vision algorithms, image segmentation is an important preprocessing step. As such, several image segmentation algorithms have been proposed, however, with certain reservation due to high computational load and many hand-tuning parameters. Correlation clustering, a graphpartitioning algorithm often used in natural language processing and document clustering, has the potential to perform better than previously proposed image segmentation algorithms. We improve the basic correlation clustering formulation by taking into account higher-order cluster relationships. This improves clustering in the presence of local boundary ambiguities. We first apply the pairwise correlation clustering to image segmentation over a pairwise superpixel graph and then develop higher-order correlation clustering over a hypergraph that considers higher-order relations among superpixels. Fast inference is possible by linear programming relaxation, and also effective parameter learning framework by structured support vector machine is possible. Experimental results on various datasets show that the proposed higher-order correlation clustering outperforms other state-of-the-art image segmentation algorithms.

4 0.87550735 305 nips-2011-k-NN Regression Adapts to Local Intrinsic Dimension

Author: Samory Kpotufe

Abstract: Many nonparametric regressors were recently shown to converge at rates that depend only on the intrinsic dimension of data. These regressors thus escape the curse of dimension when high-dimensional data has low intrinsic dimension (e.g. a manifold). We show that k-NN regression is also adaptive to intrinsic dimension. In particular our rates are local to a query x and depend only on the way masses of balls centered at x vary with radius. Furthermore, we show a simple way to choose k = k(x) locally at any x so as to nearly achieve the minimax rate at x in terms of the unknown intrinsic dimension in the vicinity of x. We also establish that the minimax rate does not depend on a particular choice of metric space or distribution, but rather that this minimax rate holds for any metric space and doubling measure. 1

5 0.78831798 260 nips-2011-Sparse Features for PCA-Like Linear Regression

Author: Christos Boutsidis, Petros Drineas, Malik Magdon-Ismail

Abstract: Principal Components Analysis (PCA) is often used as a feature extraction procedure. Given a matrix X ∈ Rn×d , whose rows represent n data points with respect to d features, the top k right singular vectors of X (the so-called eigenfeatures), are arbitrary linear combinations of all available features. The eigenfeatures are very useful in data analysis, including the regularization of linear regression. Enforcing sparsity on the eigenfeatures, i.e., forcing them to be linear combinations of only a small number of actual features (as opposed to all available features), can promote better generalization error and improve the interpretability of the eigenfeatures. We present deterministic and randomized algorithms that construct such sparse eigenfeatures while provably achieving in-sample performance comparable to regularized linear regression. Our algorithms are relatively simple and practically efficient, and we demonstrate their performance on several data sets.

6 0.53792888 223 nips-2011-Probabilistic Joint Image Segmentation and Labeling

7 0.5157252 154 nips-2011-Learning person-object interactions for action recognition in still images

8 0.50445241 227 nips-2011-Pylon Model for Semantic Segmentation

9 0.48739725 303 nips-2011-Video Annotation and Tracking with Active Learning

10 0.47189099 1 nips-2011-$\theta$-MRF: Capturing Spatial and Semantic Structure in the Parameters for Scene Understanding

11 0.46734571 59 nips-2011-Composite Multiclass Losses

12 0.46325305 166 nips-2011-Maximal Cliques that Satisfy Hard Constraints with Application to Deformable Object Model Learning

13 0.45895788 263 nips-2011-Sparse Manifold Clustering and Embedding

14 0.45844731 208 nips-2011-Optimistic Optimization of a Deterministic Function without the Knowledge of its Smoothness

15 0.45220694 304 nips-2011-Why The Brain Separates Face Recognition From Object Recognition

16 0.44888827 247 nips-2011-Semantic Labeling of 3D Point Clouds for Indoor Scenes

17 0.44826919 266 nips-2011-Spatial distance dependent Chinese restaurant processes for image segmentation

18 0.44325721 180 nips-2011-Multiple Instance Filtering

19 0.44267726 76 nips-2011-Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials

20 0.44134146 25 nips-2011-Adaptive Hedge