nips nips2006 nips2006-111 knowledge-graph by maker-knowledge-mining

111 nips-2006-Learning Motion Style Synthesis from Perceptual Observations


Source: pdf

Author: Lorenzo Torresani, Peggy Hackney, Christoph Bregler

Abstract: This paper presents an algorithm for synthesis of human motion in specified styles. We use a theory of movement observation (Laban Movement Analysis) to describe movement styles as points in a multi-dimensional perceptual space. We cast the task of learning to synthesize desired movement styles as a regression problem: sequences generated via space-time interpolation of motion capture data are used to learn a nonlinear mapping between animation parameters and movement styles in perceptual space. We demonstrate that the learned model can apply a variety of motion styles to pre-recorded motion sequences and it can extrapolate styles not originally included in the training data. 1

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 edu Abstract This paper presents an algorithm for synthesis of human motion in specified styles. [sent-6, score-0.633]

2 We use a theory of movement observation (Laban Movement Analysis) to describe movement styles as points in a multi-dimensional perceptual space. [sent-7, score-0.613]

3 We cast the task of learning to synthesize desired movement styles as a regression problem: sequences generated via space-time interpolation of motion capture data are used to learn a nonlinear mapping between animation parameters and movement styles in perceptual space. [sent-8, score-2.21]

4 We demonstrate that the learned model can apply a variety of motion styles to pre-recorded motion sequences and it can extrapolate styles not originally included in the training data. [sent-9, score-1.702]

5 1 Introduction Human motion perception can be generally thought of as the result of interaction of two factors, traditionally termed content and style. [sent-10, score-0.653]

6 ), while style denotes the particular way that action is performed. [sent-14, score-0.391]

7 In computer animation, the separation of the underlying content of a movement from its stylistic characteristics is particularly important. [sent-15, score-0.394]

8 For example, a system that can synthesize stylistic variations of a given action would be a useful tool for animators. [sent-16, score-0.319]

9 In this work we address such a problem by proposing a system that applies user-specified styles to motion sequences. [sent-17, score-0.812]

10 Specifically, given as input a target motion style and an arbitrary animation or pre-recorded motion, we want to synthesize a novel sequence that preserves the content of the original input motion but exhibits style similar to the user-specified target. [sent-18, score-2.406]

11 Concatenative synthesis techniques [15, 1, 11] are based on the simple idea of generating novel movements by concatenation of motion capture snippets. [sent-20, score-0.792]

12 Since motion is produced by cutting and pasting pre-recorded examples, the resulting animations achieve realism similar to that of pure motion-capture play back. [sent-21, score-0.65]

13 Snippet concatenation can produce novel content by generating arbitrarily complex new movements. [sent-22, score-0.32]

14 However, this approach is restricted to synthesize only the subset of styles originally contained in the input database. [sent-23, score-0.447]

15 Sample-based concatenation techniques are unable to produce novel stylistic variations and cannot generalize style differences from the existing examples. [sent-24, score-0.68]

16 Unfortunately, most of these methods learn simple parametric motion models that are unable to fully capture the subtleties and complexities of human movement. [sent-26, score-0.616]

17 The aim is to maintain the animated precision of motion capture data, while introducing the flexibility of style changes achievable by learned parametric models. [sent-29, score-0.983]

18 Our system builds on the observation that stylistically novel, yet highly realistic animations can be generated via space-time interpolation of pairs of motion sequences. [sent-30, score-0.992]

19 We propose to learn not a parametric function of the motion, but rather a parametric function of how the interpolation or extrapolation weights applied to data snippets relate to the styles of the output sequences. [sent-31, score-0.775]

20 This allows us to create motions with arbitrary styles without compromising animation quality. [sent-32, score-0.654]

21 Several researchers have previously proposed the use of motion interpolation for synthesis of novel movement [18, 6, 10]. [sent-33, score-1.061]

22 These approaches are based on the na¨ve assumption that motion interpolation ı produces styles corresponding precisely to the interpolation of the styles of the original sequences. [sent-34, score-1.613]

23 In this paper we experimentally demonstrate that styles generated through motion interpolation are a rather complex function of styles and contents of the original snippets. [sent-35, score-1.365]

24 We propose to explicitly learn the mapping between motion blending parameters and resulting animation styles. [sent-36, score-0.724]

25 This enables our animation system not only to generate arbitrary stylistic variations of a given action, but, more importantly, to synthesize sequences matching user-specified stylistic characteristics. [sent-37, score-0.722]

26 [16], in which interpolation models parameterized by style attributes are learned for several actions, such as walking or reaching. [sent-39, score-0.727]

27 This feature allows our algorithm to be used for style editing of sequences without content specification by the user. [sent-41, score-0.604]

28 2 The LMA Framework In computer animation literature motion style is a vaguely defined concept. [sent-43, score-1.068]

29 In our work, we describe motion styles according to a movement notation system, called Laban Movement Analysis or LMA [7]. [sent-44, score-0.917]

30 Instead, it targets the subtle differences in motion style, e. [sent-49, score-0.523]

31 Generally, additional features not present in motion capture data, such as eye gaze, are necessary to detect this factor. [sent-64, score-0.532]

32 We model styles as points in a three-dimensional perceptual space derived by translating the LMA-Effort notations for each of these factors into numerical values ranging in the interval [−3, 3]. [sent-66, score-0.333]

33 3 Overview of the system The key-idea of our work is to learn motion style synthesis from a training set of computer-generated animations. [sent-67, score-1.05]

34 This set of supervised data is used to learn a mapping between the space of motion styles and the animation system parameters. [sent-69, score-1.043]

35 1 Training: Learning the Style of Motion Interpolation In order to train our system to synthesize motion styles, we employ a corpus of human motion sequences recorded with a motion capture system. [sent-72, score-1.798]

36 We represent the motion as a time-varying vector of joint angles. [sent-73, score-0.524]

37 In the training stage each motion sequence is manually segmented by an LMA human expert into fragments corresponding to fundamental actions or units of motions. [sent-74, score-0.835]

38 We apply a motion matching algorithm to identify fragment pairs (Xi , Xj ) containing similar actions. [sent-77, score-0.578]

39 Our motion matching algorithm is based on dynamic-time warping. [sent-78, score-0.541]

40 Both the synthesized animations Xα as well as the ”seed” i,j motion capture data Xi are labeled with LMA-Effort values by an LMA expert. [sent-83, score-0.73]

41 i,j A non-linear regression model [5] is fitted to the LMA labels and the parameters α of the space-time interpolation algorithm. [sent-85, score-0.336]

42 2 Testing: Style Transfer ¯ At testing stage we are given a motion sequence Y, and a user-specified motion style e. [sent-88, score-1.428]

43 The goal ¯ is to apply style e to the input sequence Y, without modifying the content of the motion. [sent-89, score-0.602]

44 First, we use dynamic-time warping to segment the input sequence into snippets Yi , such that each snippet matches the content of a set of analogous motions {Xi1 , . [sent-90, score-0.649]

45 , XiK }, we determine the one that ik ¯ provides the best approximation to the target style e. [sent-97, score-0.452]

46 4 Matching motion content The objective of the matching algorithm is to identify pairs of sequences having similar motion content or consisting of analogous activities. [sent-100, score-1.431]

47 The method should ignore variations in the style with which movements are performed. [sent-101, score-0.421]

48 Previous work [2, 12] has shown that the differences in movement styles can be found by examining the parameters of timing and movement acceleration. [sent-102, score-0.625]

49 Thus we compare the content of two motions by identifying similar spatial body poses while allowing for potentially large differences in timing. [sent-104, score-0.385]

50 Specifically, we define the content similarity between motion snippets Xi and Xj , as the minimum sum of their squared joint angle differences SSD(Xi , Xi ) under a dynamic time warping path. [sent-105, score-0.911]

51 5 Space-time interpolation A time warping strategy is also employed to synthesize novel animations from the pairs of contentmatching examples found by the algorithm outlined in the previous section. [sent-109, score-0.739]

52 Given matching snippets Xi and Xj , the objective is to generate a stylistically novel sequence that maintains the content of the two original motions. [sent-110, score-0.44]

53 The idea is to induce changes in style by acting primarily on the timings of the motions. [sent-111, score-0.367]

54 we estimate joint angles Xj at time steps q∗ (n0 )) then the resampled motion will k correspond to playing sequence j with the timing of sequence i. [sent-127, score-0.708]

55 , n1 j can be cho1 T sen, such that q∗ (n1 ) = k, and these parameter values can be used to synthesize motion i with the k timing of motion j. [sent-132, score-1.154]

56 It is also possible to smoothly interpolate between these two scenarios according to an interpolation parameter α ∈ [0, 1] to produce intermediate time warps. [sent-133, score-0.338]

57 New stylistic versions of motions i and j can be produced by estimating the joint angles Xi and Xj at p∗ (nα ) and q∗ (nα ), respectively. [sent-140, score-0.342]

58 From these two synchronized sequences, a novel motion Xα can be generated by averaging the joint angles according i,j to mixing coefficients (1 − α) and α: Xα (k) = (1 − α)Xi (p∗ (nα )) + αXj (q∗ (nα )). [sent-142, score-0.647]

59 The synthei,j k k sized motion Xα will display content similar to that of Xi and Xj , but it will have distinct style. [sent-143, score-0.653]

60 6 Learning style interpolation Given a pair of content-matching snippets Xi and Xj , our goal is to determine the parameter α that needs to be applied to space-time interpolation in order to produce a motion Xα exhibiting target i,j ¯ style e. [sent-145, score-1.984]

61 The training data for this supervised learning task consists of our seed motion sequences {Xi } in the database, a set of interpolated motions {Xα }, and the i,j corresponding LMA-Effort qualities {ei }, {eα } observed by an LMA human expert. [sent-147, score-0.897]

62 In order to avoid overfitting, we compress further the motion data by projecting it onto a low-dimensional linear subspace computed using Principal Component Analysis (PCA). [sent-149, score-0.493]

63 In many of the test cases, we found it was sufficient to retain only the first two or three principal components in order to obtain a discriminative representation of the motion contents. [sent-150, score-0.493]

64 While kernel ridge regression requires us to store all training examples in order to evaluate function f at a given input, support vector regression overcomes this limitation by using an ǫ-insensitive loss function [17]. [sent-163, score-0.301]

65 7 Testing: Style Transfer We can restate our initial objective as follows: given an input motion sequence Y in unknown style, ¯ and a target motion style e specified by LMA-Effort values, we want to synthesize a sequence having ¯ style e and content analogous to that of motion Y. [sent-165, score-2.671]

66 A na¨ve approach to this problem is to seek in ı the motion database a pair of sequences having content similar to Y and whose interpolation can ¯ approximate style e. [sent-166, score-1.418]

67 The learned function f can be used to determine the pair of motions and the ¯ interpolation parameter α that produce the best approximation to e. [sent-167, score-0.495]

68 The idea is to determine the concatenation of database motion examples [X1 , . [sent-172, score-0.651]

69 , XN ] determined by the ¯ method outlined in the previous section to synthesize a version of motion Y in style e. [sent-210, score-1.028]

70 The final goal then is to replace each snippet Yi with a pairwise blend of examples ¯ in its cluster so as to produce a motion exhibiting style e. [sent-226, score-1.027]

71 We then select the pair (ik∗ , il∗ ) providing the minimum deviation ¯ from the target style e. [sent-228, score-0.404]

72 25] rather than [0,1], we give the algorithm the ability to extrapolate from existing motion styles. [sent-236, score-0.518]

73 Given optimal parameters (α∗ , k ∗ , l∗ ), space-time interpolation of fragments Xik∗ and Xil∗ with parameter value α∗ produces an animation with content similar to that of Yi and style approximating ¯ the desired target e. [sent-237, score-1.179]

74 The final animation is obtained by concatenating all of the fragments generated via interpolation with optimal parameters. [sent-239, score-0.615]

75 8 Experiments The system was tested using a motion database consisting of 12 sequences performed by different professional dancers. [sent-240, score-0.65]

76 All fragments were then automatically clustered into 5 content groups using the SSD criterion outlined in section 4. [sent-243, score-0.322]

77 The motions were recorded using a marker-based motion capture system. [sent-244, score-0.694]

78 The joint angles were represented with exponential maps [13], which have the property of being locally linear and thus particularly suitable for motion interpolation. [sent-246, score-0.579]

79 From these 60 motion fragments, 105 novel motions were synthesized with space-time interpolation using random values of α in the range [−0. [sent-247, score-1.015]

80 5 α Figure 1: Sample LMA-Effort attributes estimated by kernel ridge regression on three different pairs of motions (Xi , Xj ) and for α varying in [-0. [sent-259, score-0.385]

81 From this set of motions, 85 training examples were randomly selected to train the style regression models. [sent-265, score-0.48]

82 We include in our analysis the linear style interpolation model, commonly used in previous work. [sent-269, score-0.643]

83 This model assumes that the style of a sequence generated via motion interpolation is equal to the interpolation of the styles of the two seed motions: eα = αei + (1 − α)ej . [sent-270, score-1.782]

84 Overall, non-linear regression models proved to be much superior to the linear interpolation function, indicating that the style of sequences generates via space-time interpolation is a complex function of the original styles and motions. [sent-277, score-1.34]

85 Figure 1 shows the LMA-Effort qualities predicted by kernel ridge regression while varying α for three different sample values of the inputs (Xi , Xj , ei , ej ). [sent-278, score-0.33]

86 Several additional motion examples performed by dancers not included in the training data were used to evaluate the complete pipeline of the motion synthesis algorithm. [sent-282, score-1.148]

87 The input sequences were always correctly segmented by the dynamic programming algorithm into the five fragments associated with the actions in the phrase. [sent-283, score-0.315]

88 In order to test the generalization ability of our system, the target styles in this experiment were chosen to be considerably different from those in the training set. [sent-289, score-0.344]

89 All of the synthesized sequences were visually inspected by LMA experts and, for the great majority, they were found to be consistent with the style target labels. [sent-290, score-0.522]

90 9 Discussions and Future Work We have presented a novel technique that learns motion style synthesis from artificially-generated examples. [sent-291, score-1.012]

91 Animations produced by our system have quality similar to pure motion capture playback. [sent-292, score-0.567]

92 Furthermore, we have shown that, even with a small database, it is possible to use pair-wise interpolation or extrapolation to generate new styles. [sent-293, score-0.299]

93 In previous LMA-based animation systems [3], heuristic and hand-designed rules have been adopted to implement the style changes associated to LMA-Effort variations. [sent-294, score-0.575]

94 Although our algorithm has shown to produce good results with small training data, we expect that larger databases with a wider variety of motion contents and styles are needed in order to build an effective animation system. [sent-296, score-1.07]

95 Multi-way, as opposed to pair-wise, interpolation might lead to synthesis of more varied motion styles. [sent-297, score-0.878]

96 Our future work will focus on the recognition of LMA categories in motion capture data. [sent-299, score-0.532]

97 Research in this area might point to methods for learning person-specific styles and to techniques for transferring individual movement signatures to arbitrary motion sequences. [sent-300, score-0.917]

98 Morphable models for the analysis and synthesis of complex motion patterns. [sent-344, score-0.602]

99 Interactive control of avatars animated with human motion data. [sent-377, score-0.555]

100 Motion texture: A two-level statistical model for character motion synthesis. [sent-384, score-0.493]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('motion', 0.493), ('style', 0.367), ('styles', 0.284), ('interpolation', 0.276), ('animation', 0.208), ('lma', 0.203), ('motions', 0.162), ('content', 0.16), ('animations', 0.157), ('movement', 0.14), ('synthesize', 0.137), ('fragments', 0.131), ('synthesis', 0.109), ('snippets', 0.109), ('ridge', 0.098), ('stylistic', 0.094), ('flow', 0.092), ('graphics', 0.092), ('xik', 0.087), ('concatenation', 0.083), ('snippet', 0.078), ('sequences', 0.077), ('qualities', 0.074), ('xi', 0.074), ('ssd', 0.068), ('warping', 0.065), ('laban', 0.063), ('regression', 0.06), ('xj', 0.057), ('angles', 0.055), ('sequence', 0.049), ('perceptual', 0.049), ('matching', 0.048), ('ik', 0.048), ('july', 0.047), ('concatenative', 0.047), ('poles', 0.047), ('xil', 0.047), ('database', 0.045), ('novel', 0.043), ('opposing', 0.041), ('synthesized', 0.041), ('acm', 0.041), ('path', 0.04), ('capture', 0.039), ('target', 0.037), ('fragment', 0.037), ('seed', 0.037), ('tj', 0.037), ('ei', 0.037), ('frames', 0.036), ('system', 0.035), ('attributes', 0.035), ('frame', 0.034), ('produce', 0.034), ('body', 0.033), ('human', 0.031), ('joint', 0.031), ('adverbs', 0.031), ('animated', 0.031), ('blends', 0.031), ('stylistically', 0.031), ('zj', 0.031), ('ej', 0.031), ('outlined', 0.031), ('il', 0.031), ('timing', 0.031), ('examples', 0.03), ('parametric', 0.03), ('differences', 0.03), ('mse', 0.03), ('kernel', 0.03), ('variations', 0.029), ('segmented', 0.029), ('actions', 0.029), ('smoothly', 0.028), ('contents', 0.028), ('eik', 0.027), ('lorenzo', 0.027), ('rose', 0.027), ('input', 0.026), ('walking', 0.026), ('stage', 0.026), ('movements', 0.025), ('extrapolate', 0.025), ('touch', 0.025), ('synchronized', 0.025), ('exhibiting', 0.025), ('endpoint', 0.025), ('action', 0.024), ('transactions', 0.024), ('ti', 0.024), ('expert', 0.024), ('learn', 0.023), ('learned', 0.023), ('dynamic', 0.023), ('tij', 0.023), ('extrapolation', 0.023), ('yi', 0.023), ('training', 0.023)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000005 111 nips-2006-Learning Motion Style Synthesis from Perceptual Observations

Author: Lorenzo Torresani, Peggy Hackney, Christoph Bregler

Abstract: This paper presents an algorithm for synthesis of human motion in specified styles. We use a theory of movement observation (Laban Movement Analysis) to describe movement styles as points in a multi-dimensional perceptual space. We cast the task of learning to synthesize desired movement styles as a regression problem: sequences generated via space-time interpolation of motion capture data are used to learn a nonlinear mapping between animation parameters and movement styles in perceptual space. We demonstrate that the learned model can apply a variety of motion styles to pre-recorded motion sequences and it can extrapolate styles not originally included in the training data. 1

2 0.32339579 134 nips-2006-Modeling Human Motion Using Binary Latent Variables

Author: Graham W. Taylor, Geoffrey E. Hinton, Sam T. Roweis

Abstract: We propose a non-linear generative model for human motion data that uses an undirected model with binary latent variables and real-valued “visible” variables that represent joint angles. The latent and visible variables at each time step receive directed connections from the visible variables at the last few time-steps. Such an architecture makes on-line inference efficient and allows us to use a simple approximate learning procedure. After training, the model finds a single set of parameters that simultaneously capture several different kinds of motion. We demonstrate the power of our approach by synthesizing various motion sequences and by performing on-line filling in of data lost during motion capture. Website: http://www.cs.toronto.edu/∼gwtaylor/publications/nips2006mhmublv/

3 0.26599023 31 nips-2006-Analysis of Contour Motions

Author: Ce Liu, William T. Freeman, Edward H. Adelson

Abstract: A reliable motion estimation algorithm must function under a wide range of conditions. One regime, which we consider here, is the case of moving objects with contours but no visible texture. Tracking distinctive features such as corners can disambiguate the motion of contours, but spurious features such as T-junctions can be badly misleading. It is difficult to determine the reliability of motion from local measurements, since a full rank covariance matrix can result from both real and spurious features. We propose a novel approach that avoids these points altogether, and derives global motion estimates by utilizing information from three levels of contour analysis: edgelets, boundary fragments and contours. Boundary fragment are chains of orientated edgelets, for which we derive motion estimates from local evidence. The uncertainties of the local estimates are disambiguated after the boundary fragments are properly grouped into contours. The grouping is done by constructing a graphical model and marginalizing it using importance sampling. We propose two equivalent representations in this graphical model, reversible switch variables attached to the ends of fragments and fragment chains, to capture both local and global statistics of boundaries. Our system is successfully applied to both synthetic and real video sequences containing high-contrast boundaries and textureless regions. The system produces good motion estimates along with properly grouped and completed contours.

4 0.10096779 147 nips-2006-Non-rigid point set registration: Coherent Point Drift

Author: Andriy Myronenko, Xubo Song, Miguel Á. Carreira-Perpiñán

Abstract: We introduce Coherent Point Drift (CPD), a novel probabilistic method for nonrigid registration of point sets. The registration is treated as a Maximum Likelihood (ML) estimation problem with motion coherence constraint over the velocity field such that one point set moves coherently to align with the second set. We formulate the motion coherence constraint and derive a solution of regularized ML estimation through the variational approach, which leads to an elegant kernel form. We also derive the EM algorithm for the penalized ML optimization with deterministic annealing. The CPD method simultaneously finds both the non-rigid transformation and the correspondence between two point sets without making any prior assumption of the transformation model except that of motion coherence. This method can estimate complex non-linear non-rigid transformations, and is shown to be accurate on 2D and 3D examples and robust in the presence of outliers and missing points.

5 0.094559774 46 nips-2006-Blind source separation for over-determined delayed mixtures

Author: Lars Omlor, Martin Giese

Abstract: Blind source separation, i.e. the extraction of unknown sources from a set of given signals, is relevant for many applications. A special case of this problem is dimension reduction, where the goal is to approximate a given set of signals by superpositions of a minimal number of sources. Since in this case the signals outnumber the sources the problem is over-determined. Most popular approaches for addressing this problem are based on purely linear mixing models. However, many applications like the modeling of acoustic signals, EMG signals, or movement trajectories, require temporal shift-invariance of the extracted components. This case has only rarely been treated in the computational literature, and specifically for the case of dimension reduction almost no algorithms have been proposed. We present a new algorithm for the solution of this problem, which is based on a timefrequency transformation (Wigner-Ville distribution) of the generative model. We show that this algorithm outperforms classical source separation algorithms for linear mixtures, and also a related method for mixtures with delays. In addition, applying the new algorithm to trajectories of human gaits, we demonstrate that it is suitable for the extraction of spatio-temporal components that are easier to interpret than components extracted with other classical algorithms. 1

6 0.090265326 120 nips-2006-Learning to Traverse Image Manifolds

7 0.08607927 112 nips-2006-Learning Nonparametric Models for Probabilistic Imitation

8 0.083289146 108 nips-2006-Large Scale Hidden Semi-Markov SVMs

9 0.081810154 153 nips-2006-Online Clustering of Moving Hyperplanes

10 0.077681057 186 nips-2006-Support Vector Machines on a Budget

11 0.07741116 66 nips-2006-Detecting Humans via Their Pose

12 0.077341594 148 nips-2006-Nonlinear physically-based models for decoding motor-cortical population activity

13 0.072508223 84 nips-2006-Generalized Regularized Least-Squares Learning with Predefined Features in a Hilbert Space

14 0.063220538 45 nips-2006-Blind Motion Deblurring Using Image Statistics

15 0.054596022 79 nips-2006-Fast Iterative Kernel PCA

16 0.049436152 8 nips-2006-A Nonparametric Approach to Bottom-Up Visual Saliency

17 0.048527591 110 nips-2006-Learning Dense 3D Correspondence

18 0.048404448 95 nips-2006-Implicit Surfaces with Globally Regularised and Compactly Supported Basis Functions

19 0.046697803 65 nips-2006-Denoising and Dimension Reduction in Feature Space

20 0.044591386 42 nips-2006-Bayesian Image Super-resolution, Continued


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, -0.175), (1, 0.015), (2, 0.093), (3, -0.048), (4, 0.012), (5, -0.092), (6, -0.049), (7, -0.098), (8, -0.013), (9, 0.221), (10, -0.097), (11, -0.157), (12, -0.218), (13, 0.115), (14, -0.061), (15, -0.061), (16, -0.385), (17, -0.095), (18, -0.215), (19, -0.047), (20, -0.067), (21, 0.09), (22, -0.038), (23, 0.183), (24, 0.027), (25, -0.051), (26, -0.062), (27, 0.01), (28, -0.001), (29, 0.008), (30, 0.031), (31, 0.019), (32, 0.059), (33, -0.007), (34, 0.095), (35, 0.133), (36, -0.106), (37, 0.055), (38, -0.088), (39, -0.071), (40, 0.092), (41, 0.1), (42, 0.043), (43, -0.069), (44, 0.063), (45, -0.024), (46, 0.016), (47, 0.052), (48, 0.051), (49, -0.026)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.97278684 111 nips-2006-Learning Motion Style Synthesis from Perceptual Observations

Author: Lorenzo Torresani, Peggy Hackney, Christoph Bregler

Abstract: This paper presents an algorithm for synthesis of human motion in specified styles. We use a theory of movement observation (Laban Movement Analysis) to describe movement styles as points in a multi-dimensional perceptual space. We cast the task of learning to synthesize desired movement styles as a regression problem: sequences generated via space-time interpolation of motion capture data are used to learn a nonlinear mapping between animation parameters and movement styles in perceptual space. We demonstrate that the learned model can apply a variety of motion styles to pre-recorded motion sequences and it can extrapolate styles not originally included in the training data. 1

2 0.85815269 31 nips-2006-Analysis of Contour Motions

Author: Ce Liu, William T. Freeman, Edward H. Adelson

Abstract: A reliable motion estimation algorithm must function under a wide range of conditions. One regime, which we consider here, is the case of moving objects with contours but no visible texture. Tracking distinctive features such as corners can disambiguate the motion of contours, but spurious features such as T-junctions can be badly misleading. It is difficult to determine the reliability of motion from local measurements, since a full rank covariance matrix can result from both real and spurious features. We propose a novel approach that avoids these points altogether, and derives global motion estimates by utilizing information from three levels of contour analysis: edgelets, boundary fragments and contours. Boundary fragment are chains of orientated edgelets, for which we derive motion estimates from local evidence. The uncertainties of the local estimates are disambiguated after the boundary fragments are properly grouped into contours. The grouping is done by constructing a graphical model and marginalizing it using importance sampling. We propose two equivalent representations in this graphical model, reversible switch variables attached to the ends of fragments and fragment chains, to capture both local and global statistics of boundaries. Our system is successfully applied to both synthetic and real video sequences containing high-contrast boundaries and textureless regions. The system produces good motion estimates along with properly grouped and completed contours.

3 0.72710508 134 nips-2006-Modeling Human Motion Using Binary Latent Variables

Author: Graham W. Taylor, Geoffrey E. Hinton, Sam T. Roweis

Abstract: We propose a non-linear generative model for human motion data that uses an undirected model with binary latent variables and real-valued “visible” variables that represent joint angles. The latent and visible variables at each time step receive directed connections from the visible variables at the last few time-steps. Such an architecture makes on-line inference efficient and allows us to use a simple approximate learning procedure. After training, the model finds a single set of parameters that simultaneously capture several different kinds of motion. We demonstrate the power of our approach by synthesizing various motion sequences and by performing on-line filling in of data lost during motion capture. Website: http://www.cs.toronto.edu/∼gwtaylor/publications/nips2006mhmublv/

4 0.4594734 153 nips-2006-Online Clustering of Moving Hyperplanes

Author: René Vidal

Abstract: We propose a recursive algorithm for clustering trajectories lying in multiple moving hyperplanes. Starting from a given or random initial condition, we use normalized gradient descent to update the coefficients of a time varying polynomial whose degree is the number of hyperplanes and whose derivatives at a trajectory give an estimate of the vector normal to the hyperplane containing that trajectory. As time proceeds, the estimates of the hyperplane normals are shown to track their true values in a stable fashion. The segmentation of the trajectories is then obtained by clustering their associated normal vectors. The final result is a simple recursive algorithm for segmenting a variable number of moving hyperplanes. We test our algorithm on the segmentation of dynamic scenes containing rigid motions and dynamic textures, e.g., a bird floating on water. Our method not only segments the bird motion from the surrounding water motion, but also determines patterns of motion in the scene (e.g., periodic motion) directly from the temporal evolution of the estimated polynomial coefficients. Our experiments also show that our method can deal with appearing and disappearing motions in the scene.

5 0.36111325 112 nips-2006-Learning Nonparametric Models for Probabilistic Imitation

Author: David B. Grimes, Daniel R. Rashid, Rajesh P. Rao

Abstract: Learning by imitation represents an important mechanism for rapid acquisition of new behaviors in humans and robots. A critical requirement for learning by imitation is the ability to handle uncertainty arising from the observation process as well as the imitator’s own dynamics and interactions with the environment. In this paper, we present a new probabilistic method for inferring imitative actions that takes into account both the observations of the teacher as well as the imitator’s dynamics. Our key contribution is a nonparametric learning method which generalizes to systems with very different dynamics. Rather than relying on a known forward model of the dynamics, our approach learns a nonparametric forward model via exploration. Leveraging advances in approximate inference in graphical models, we show how the learned forward model can be directly used to plan an imitating sequence. We provide experimental results for two systems: a biomechanical model of the human arm and a 25-degrees-of-freedom humanoid robot. We demonstrate that the proposed method can be used to learn appropriate motor inputs to the model arm which imitates the desired movements. A second set of results demonstrates dynamically stable full-body imitation of a human teacher by the humanoid robot. 1

6 0.34629777 147 nips-2006-Non-rigid point set registration: Coherent Point Drift

7 0.32453209 108 nips-2006-Large Scale Hidden Semi-Markov SVMs

8 0.28059146 148 nips-2006-Nonlinear physically-based models for decoding motor-cortical population activity

9 0.2804141 120 nips-2006-Learning to Traverse Image Manifolds

10 0.27107713 66 nips-2006-Detecting Humans via Their Pose

11 0.2379079 46 nips-2006-Blind source separation for over-determined delayed mixtures

12 0.2374672 40 nips-2006-Bayesian Detection of Infrequent Differences in Sets of Time Series with Shared Structure

13 0.23283099 186 nips-2006-Support Vector Machines on a Budget

14 0.21618368 140 nips-2006-Multiple Instance Learning for Computer Aided Diagnosis

15 0.21470536 95 nips-2006-Implicit Surfaces with Globally Regularised and Compactly Supported Basis Functions

16 0.20476308 84 nips-2006-Generalized Regularized Least-Squares Learning with Predefined Features in a Hilbert Space

17 0.18538214 101 nips-2006-Isotonic Conditional Random Fields and Local Sentiment Flow

18 0.18354955 128 nips-2006-Manifold Denoising

19 0.18342167 189 nips-2006-Temporal dynamics of information content carried by neurons in the primary visual cortex

20 0.17876643 25 nips-2006-An Application of Reinforcement Learning to Aerobatic Helicopter Flight


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(1, 0.146), (3, 0.048), (7, 0.041), (8, 0.043), (9, 0.036), (12, 0.014), (13, 0.304), (20, 0.011), (22, 0.051), (44, 0.05), (57, 0.069), (65, 0.035), (69, 0.034), (71, 0.012)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.80259657 111 nips-2006-Learning Motion Style Synthesis from Perceptual Observations

Author: Lorenzo Torresani, Peggy Hackney, Christoph Bregler

Abstract: This paper presents an algorithm for synthesis of human motion in specified styles. We use a theory of movement observation (Laban Movement Analysis) to describe movement styles as points in a multi-dimensional perceptual space. We cast the task of learning to synthesize desired movement styles as a regression problem: sequences generated via space-time interpolation of motion capture data are used to learn a nonlinear mapping between animation parameters and movement styles in perceptual space. We demonstrate that the learned model can apply a variety of motion styles to pre-recorded motion sequences and it can extrapolate styles not originally included in the training data. 1

2 0.7310794 126 nips-2006-Logistic Regression for Single Trial EEG Classification

Author: Ryota Tomioka, Kazuyuki Aihara, Klaus-Robert Müller

Abstract: We propose a novel framework for the classification of single trial ElectroEncephaloGraphy (EEG), based on regularized logistic regression. Framed in this robust statistical framework no prior feature extraction or outlier removal is required. We present two variations of parameterizing the regression function: (a) with a full rank symmetric matrix coefficient and (b) as a difference of two rank=1 matrices. In the first case, the problem is convex and the logistic regression is optimal under a generative model. The latter case is shown to be related to the Common Spatial Pattern (CSP) algorithm, which is a popular technique in Brain Computer Interfacing. The regression coefficients can also be topographically mapped onto the scalp similarly to CSP projections, which allows neuro-physiological interpretation. Simulations on 162 BCI datasets demonstrate that classification accuracy and robustness compares favorably against conventional CSP based classifiers. 1

3 0.71066499 79 nips-2006-Fast Iterative Kernel PCA

Author: Nicol N. Schraudolph, Simon Günter, S.v.n. Vishwanathan

Abstract: We introduce two methods to improve convergence of the Kernel Hebbian Algorithm (KHA) for iterative kernel PCA. KHA has a scalar gain parameter which is either held constant or decreased as 1/t, leading to slow convergence. Our KHA/et algorithm accelerates KHA by incorporating the reciprocal of the current estimated eigenvalues as a gain vector. We then derive and apply Stochastic MetaDescent (SMD) to KHA/et; this further speeds convergence by performing gain adaptation in RKHS. Experimental results for kernel PCA and spectral clustering of USPS digits as well as motion capture and image de-noising problems confirm that our methods converge substantially faster than conventional KHA. 1

4 0.68945479 168 nips-2006-Reducing Calibration Time For Brain-Computer Interfaces: A Clustering Approach

Author: Matthias Krauledat, Michael Schröder, Benjamin Blankertz, Klaus-Robert Müller

Abstract: Up to now even subjects that are experts in the use of machine learning based BCI systems still have to undergo a calibration session of about 20-30 min. From this data their (movement) intentions are so far infered. We now propose a new paradigm that allows to completely omit such calibration and instead transfer knowledge from prior sessions. To achieve this goal we first define normalized CSP features and distances in-between. Second, we derive prototypical features across sessions: (a) by clustering or (b) by feature concatenation methods. Finally, we construct a classifier based on these individualized prototypes and show that, indeed, classifiers can be successfully transferred to a new session for a number of subjects.

5 0.51947826 51 nips-2006-Clustering Under Prior Knowledge with Application to Image Segmentation

Author: Dong S. Cheng, Vittorio Murino, Mário Figueiredo

Abstract: This paper proposes a new approach to model-based clustering under prior knowledge. The proposed formulation can be interpreted from two different angles: as penalized logistic regression, where the class labels are only indirectly observed (via the probability density of each class); as finite mixture learning under a grouping prior. To estimate the parameters of the proposed model, we derive a (generalized) EM algorithm with a closed-form E-step, in contrast with other recent approaches to semi-supervised probabilistic clustering which require Gibbs sampling or suboptimal shortcuts. We show that our approach is ideally suited for image segmentation: it avoids the combinatorial nature Markov random field priors, and opens the door to more sophisticated spatial priors (e.g., wavelet-based) in a simple and computationally efficient way. Finally, we extend our formulation to work in unsupervised, semi-supervised, or discriminative modes. 1

6 0.51644415 65 nips-2006-Denoising and Dimension Reduction in Feature Space

7 0.51489657 152 nips-2006-Online Classification for Complex Problems Using Simultaneous Projections

8 0.51433998 153 nips-2006-Online Clustering of Moving Hyperplanes

9 0.51342201 179 nips-2006-Sparse Representation for Signal Classification

10 0.51197898 138 nips-2006-Multi-Task Feature Learning

11 0.51173824 32 nips-2006-Analysis of Empirical Bayesian Methods for Neuroelectromagnetic Source Localization

12 0.51101375 106 nips-2006-Large Margin Hidden Markov Models for Automatic Speech Recognition

13 0.51069248 84 nips-2006-Generalized Regularized Least-Squares Learning with Predefined Features in a Hilbert Space

14 0.50849879 3 nips-2006-A Complexity-Distortion Approach to Joint Pattern Alignment

15 0.50820935 108 nips-2006-Large Scale Hidden Semi-Markov SVMs

16 0.50667417 83 nips-2006-Generalized Maximum Margin Clustering and Unsupervised Kernel Learning

17 0.50644767 20 nips-2006-Active learning for misspecified generalized linear models

18 0.50568503 127 nips-2006-MLLE: Modified Locally Linear Embedding Using Multiple Weights

19 0.50467497 87 nips-2006-Graph Laplacian Regularization for Large-Scale Semidefinite Programming

20 0.50437135 178 nips-2006-Sparse Multinomial Logistic Regression via Bayesian L1 Regularisation