nips nips2008 nips2008-60 knowledge-graph by maker-knowledge-mining
Source: pdf
Author: Jeremy Lewi, Robert Butera, David M. Schneider, Sarah Woolley, Liam Paninski
Abstract: Sequential optimal design methods hold great promise for improving the efficiency of neurophysiology experiments. However, previous methods for optimal experimental design have incorporated only weak prior information about the underlying neural system (e.g., the sparseness or smoothness of the receptive field). Here we describe how to use stronger prior information, in the form of parametric models of the receptive field, in order to construct optimal stimuli and further improve the efficiency of our experiments. For example, if we believe that the receptive field is well-approximated by a Gabor function, then our method constructs stimuli that optimally constrain the Gabor parameters (orientation, spatial frequency, etc.) using as few experimental trials as possible. More generally, we may believe a priori that the receptive field lies near a known sub-manifold of the full parameter space; in this case, our method chooses stimuli in order to reduce the uncertainty along the tangent space of this sub-manifold as rapidly as possible. Applications to simulated and real data indicate that these methods may in many cases improve the experimental efficiency. 1
Reference: text
sentIndex sentText sentNum sentScore
1 Designing neurophysiology experiments to optimally constrain receptive field models along parametric submanifolds. [sent-1, score-0.455]
2 edu Abstract Sequential optimal design methods hold great promise for improving the efficiency of neurophysiology experiments. [sent-12, score-0.275]
3 However, previous methods for optimal experimental design have incorporated only weak prior information about the underlying neural system (e. [sent-13, score-0.171]
4 Here we describe how to use stronger prior information, in the form of parametric models of the receptive field, in order to construct optimal stimuli and further improve the efficiency of our experiments. [sent-16, score-0.324]
5 For example, if we believe that the receptive field is well-approximated by a Gabor function, then our method constructs stimuli that optimally constrain the Gabor parameters (orientation, spatial frequency, etc. [sent-17, score-0.284]
6 More generally, we may believe a priori that the receptive field lies near a known sub-manifold of the full parameter space; in this case, our method chooses stimuli in order to reduce the uncertainty along the tangent space of this sub-manifold as rapidly as possible. [sent-19, score-0.817]
7 1 Introduction A long standing problem in neuroscience has been collecting enough data to robustly estimate the response function of a neuron. [sent-21, score-0.206]
8 To make optimizing the design tractable, we typically need to assume our knowledge has some nice mathematical representation. [sent-23, score-0.212]
9 This restriction often makes it difficult to include the types of prior beliefs held by neurophysiologists; for example that the receptive field has some parametric form such as a Gabor function [7]. [sent-24, score-0.264]
10 edu/∼liam/ 1 p(θ|µt , Ct ) p(θ|µb , Cb ) TµM,t M θ2 µM,t M µt θ2 θ2 θ1 θ1 θ1 Figure 1: A schematic illustrating how we use the manifold to improve stimulus design. [sent-30, score-0.449]
11 Our method begins with a Gaussian approximation of the posterior on the full model space after t trials, p(θ|µt , C t ). [sent-31, score-0.249]
12 The next step involves constructing the tangent space approximation of the manifold M on which θ is believed to lie, as illustrated in the middle plot; M is indicated in blue. [sent-33, score-0.858]
13 The MAP estimate (blue dot) is projected onto the manifold to obtain µM,t (green dot). [sent-34, score-0.396]
14 We then compute the tangent space (dashed red line) by taking the derivative of the manifold at µM,t . [sent-35, score-0.846]
15 The tangent space is the space spanned by vectors in the direction parallel to M at µM,t . [sent-36, score-0.541]
16 By definition, in the neighborhood of µM,t , moving along the manifold is roughly equivalent to moving along the tangent space. [sent-37, score-0.813]
17 Thus, the tangent space provides a good local approximation of M. [sent-38, score-0.514]
18 In the right panel we compute p(θ|µb,t , Cb,t ) by evaluating p(θ|µt , C t ) on the tangent space. [sent-39, score-0.407]
19 the problem of incorporating this strong prior knowledge into an existing algorithm for optimizing neurophysiology experiments [8]. [sent-41, score-0.327]
20 We use the manifold to design an experiment which will provide the largest reduction in our uncertainty about the unknown parameters. [sent-45, score-0.545]
21 To make the computations tractable we approximate the manifold using the tangent space evaluated at the maximum a posteriori (MAP) estimate of the parameters projected onto the manifold. [sent-46, score-0.918]
22 Despite this rather crude approximation of the geometry of the manifold, our simulations show that this method can significantly improve the informativeness of our experiments. [sent-47, score-0.16]
23 2 Methods We begin by summarizing the three key elements of an existing algorithm for optimizing neurophysiology experiments. [sent-49, score-0.204]
24 We model the neuron’s response function as a mapping between the neuron’s input at time t, st , and its response, rt . [sent-51, score-0.33]
25 In this context, optimizing the experimental design means picking the input for which observing the response will provide the most information about the parameters θ defining the conditional response function. [sent-60, score-0.391]
26 The likelihood of the response depends on the firing rate, λt , which is a function of the input, λt = E(rt ) = f θT st , (1) where f () is some nonlinear function which is assumed known1 . [sent-62, score-0.265]
27 This approximation is justified by the log-concavity of the likelihood function and asymptotic normality of the posterior distribution given sufficient data [12]. [sent-66, score-0.137]
28 Since the purpose of an experiment is to identify the best model, we optimize the design by maximizing the conditional mutual information between rt+1 and θ given st+1 , I(θ; rt+1 |st+1 ). [sent-70, score-0.276]
29 The mutual information measures how much we expect observing the response to st+1 will reduce our uncertainty about θ. [sent-71, score-0.263]
30 We pick the optimal input by maximizing the mutual information with respect to st+1 ; as discussed in [8], this step, along with the updating of the posterior mean and covariance (µt , C t ), may be computed efficiently enough for real-time implementation in many cases. [sent-72, score-0.245]
31 For the computation of the mutual information to be tractable, the space of candidate models, Θ, must have some convenient form so that we can derive a suitable expression for the mutual information. [sent-75, score-0.239]
32 The problem with incorporating prior knowledge is that if we restrict the model to some complicated subset of model space we will no longer be able to efficiently integrate over the set of candidate models. [sent-78, score-0.23]
33 We address this problem by showing how local geometric approximations to the parameter sub-manifold can be used to guide optimal sampling while still maintaining a flexible, tractable representation of the posterior distribution on the full model space. [sent-79, score-0.19]
34 In many experiments, neurophysiologists expect a-priori that the receptive field of a neuron will have some low-dimensional parametric structure; e. [sent-80, score-0.402]
35 g the receptive field might be well-approximated by a Gabor function [13], or by a difference of Gaussians [14], or by a low rank spatiotemporal matrix [15, 13]. [sent-81, score-0.253]
36 (2) The vector, φ, essentially enumerates the points on the manifold and Ψ() is a function which maps these points into Θ space. [sent-83, score-0.344]
37 The basic idea is that we want to run experiments which can identify exactly where on the manifold the optimal model lies. [sent-87, score-0.344]
38 Since M can have some arbitrary nonlinear shape, computing the informativeness of a stimulus using just the models on the manifold is not easy. [sent-88, score-0.613]
39 Rather, we maintain a Gaussian approximation of the posterior on the full space, Θ. [sent-91, score-0.182]
40 However, when optimizing our stimuli we combine our posterior with our knowledge of M in order to do a better job of maximizing the informativeness of each experiment. [sent-92, score-0.363]
41 3 Computing the mutual information I(rt+1 ; θ|st+1 , s1:t , r1:t ) entails an integral over model space weighted by the posterior probability on each model. [sent-94, score-0.283]
42 We integrate over model space because the informativeness of an experiment clearly depends on what we already know (i. [sent-95, score-0.223]
43 Furthermore, the informativeness of an experiment will depend on the outcome. [sent-98, score-0.156]
44 Unfortunately, since the manifold in general has some arbitrary nonlinear shape we cannot easily compute integrals over the manifold. [sent-100, score-0.388]
45 Furthermore, we do not want to continue to restrict ourselves to models on the manifold if the data indicates our prior knowledge is wrong. [sent-101, score-0.46]
46 We can solve both problems by making use of the tangent space of the manifold, as illustrated in Figure 1 [16]. [sent-102, score-0.474]
47 The tangent space is a linear space which provides a local approximation of the manifold. [sent-103, score-0.581]
48 Since the tangent space is a linear subspace of Θ, integrating over θ in the tangent space is much easier than integrating over all θ on the manifold; in fact, the methods introduced in [8] may be applied directly to this case. [sent-104, score-1.049]
49 The tangent space is a local linear approximation evaluated at a particular point, µM,t , on the manifold. [sent-105, score-0.514]
50 For µM,t we use the projection of µt onto the manifold (i. [sent-106, score-0.427]
51 To find models on the manifold close to µM,t we want to perturb the parameters φ about the values corresponding to µM,t . [sent-111, score-0.41]
52 Thus, the subspace formed by linear combinations of the partial derivatives approximates the set of models on the manifold close to µM,t . [sent-115, score-0.472]
53 This subspace is the tangent space, TµM,t M = {θ : θ = µM,t + B b, ∀b ∈ Rdim(M) } B = orth ∂Ψ ∂Ψ . [sent-116, score-0.501]
54 ∂φ1 ∂φd , (3) where orth is an orthonormal basis for the column space of its argument. [sent-119, score-0.151]
55 Here Tx M denotes the tangent space at the point x. [sent-120, score-0.474]
56 ) We now use our Gaussian posterior on the full parameter space to compute the posterior likelihood of the models in the tangent space. [sent-123, score-0.713]
57 Since the tangent space is a subspace of Θ, restricting our Gaussian approximation, p(θ|µt , C t ), to the tangent space means we are taking a slice through our Gaussian approximation of the posterior. [sent-124, score-1.027]
58 The result is a Gaussian distribution on the tangent space whose parameters may be obtained using the standard Gaussian conditioning formula: ptan (θ|µb,t , Cb,t ) = µb,t = −Cb,t B T C −1 (µM,t − µt ) t N (b; µb,t , Cb,t ) if 0 if Cb,t = (B T C −1 B)−1 t ∃ b s. [sent-126, score-0.59]
59 Now, rather than optimizing the stimulus by trying to squeeze the uncertainty p(θ|r1:t , s1:t , M) on the nonlinear manifold M down as much as possible (a very difficult task in general), we pick the stimulus which best reduces the uncertainty ptan (θ|µb,t , Cb,t ) on the vector space TµM,t . [sent-128, score-0.911]
60 Finally, to handle the possibility that θ ∈ M, every so / often we optimize the stimulus using the full posterior p(θ|µt , C t ). [sent-130, score-0.275]
61 tangent space design described in the text; an i. [sent-143, score-0.6]
62 design which did not use the assumption that θ corresponds to a low rank STRF. [sent-148, score-0.211]
63 We see that using the tangent space to optimize the design leads to much faster convergence to the true parameters; in addition, either infomax design significantly outperforms the iid design here. [sent-152, score-0.946]
64 In this case the true STRF did not in fact lie on the manifold M (chosen to be the set of rank-2 matrices here); thus, these results also show that our knowledge of M does not need to be exact in order to improve the experimental design. [sent-153, score-0.45]
65 Initial conditions: start with a log-concave (approximately Gaussian) posterior given t previous trials, summarized by the posterior mean, µt and covariance, C t . [sent-157, score-0.194]
66 Compute the tangent space of M at µM,t using Eqn. [sent-162, score-0.474]
67 Compute the posterior restricted to the tangent space, ptan (θ|µb,t , Cb,t ), using the standard Gaussian conditioning formula (Eqn. [sent-165, score-0.62]
68 Update the posterior by recursively updating the posterior mean and covariance: µt → µt+1 and C t → C t+1 (again, as in [8]), and return to step 1. [sent-170, score-0.194]
69 1 Results Low rank models To test our methods in a realistic, high-dimensional setting, we simulated a typical auditory neurophysiology [17, 15, 18] experiment. [sent-172, score-0.282]
70 Here, the objective is to to identify the spectro-temporal receptive field (STRF) of the neuron. [sent-173, score-0.168]
71 The input and receptive field of the neuron are usually represented in the frequency domain because the cochlea is known to perform a frequency decomposition of sound. [sent-174, score-0.396]
72 Several researchers, however, have shown that low-rank assumptions can be used to produce accurate approximations of the receptive field while significantly reducing the number of unknown parameters [19, 13, 15, 20]. [sent-178, score-0.168]
73 A low rank assumption is a more general version of the space-time separable assumption that is often used when studying visual receptive fields [21]. [sent-179, score-0.253]
74 Mathematically, a low-rank assumption means that the matrix corresponding to the STRF can be written as a sum of rank one matrices, Θ = M at θ = U V T (6) where M at indicates the matrix formed by reshaping the vector θ to form the STRF. [sent-180, score-0.161]
75 The columns of U and V are the principal components of the column and row spaces of Θ respectively, and encode the spectral and temporal properties of the STRF, respectively. [sent-182, score-0.126]
76 We simulated an auditory experiment using an STRF fitted to the actual response of a neuron in the Mesencephalicus lateralis pars dorsalis (MLd) of an adult male zebra finch [18]. [sent-183, score-0.363]
77 For the manifold we used the set of θ corresponding to rank-2 matrices. [sent-186, score-0.344]
78 The manifold of rank r matrices is convenient because we can easily project any θ onto M by reshaping θ as a matrix and then computing its singular-value-decomposition (SVD). [sent-189, score-0.57]
79 The results clearly show that using the tangent space to design the experiments leads to much faster convergence to the true parameters. [sent-196, score-0.636]
80 2 Real birdsong data We also tested our method by using it to reshuffle the data collected during an actual experiment to find an ordering which provided a faster decrease in the error of the fitted model. [sent-199, score-0.127]
81 We compared a design which randomly shuffled the trials to a design which used our info. [sent-201, score-0.309]
82 To constrain the models we assume the STRF is low-rank and that its principal components are smooth. [sent-206, score-0.135]
83 The smoothing prior means that if we take the Fourier transform of the principal components, the Fourier coefficients of high frequencies should be zero with high probability. [sent-207, score-0.179]
84 In other words, each principal component (the columns of U and V ) should be a linear combination of sinusoidal functions with low frequencies. [sent-208, score-0.17]
85 (7) Each column of F and T is a sine or cosine function representing one of the basis functions of the principal spectral (columns of F ) or temporal (columns of T ) components of the STRF. [sent-210, score-0.151]
86 Each column of ν and η determines how we form one of the principal components by combining sine and cosine functions. [sent-211, score-0.151]
87 ω is a diagonal matrix which specifies the projection of Θ onto each principal 6 Eθlog p(r|st,θt) 0 −0. [sent-212, score-0.162]
88 design which uses the tangent space, and a shuffled design. [sent-221, score-0.533]
89 The plot shows the expected log-likelihood (prediction accuracy) of the spike trains in response to a birdsong in the test set. [sent-223, score-0.16]
90 Using a rank 2 manifold to constrain the model produces slightly better fits of the data. [sent-224, score-0.485]
91 The sinusoidal functions corresponding to the columns of F and T should have frequencies {0, . [sent-227, score-0.146]
92 mf and mt are the largest integers such that fo,f mf and fo,t mt are less than the Nyquist frequency. [sent-235, score-0.238]
93 Now to enforce a smoothing prior we can simply restrict the columns of F and T to sinusoids with low frequencies. [sent-236, score-0.132]
94 To project Θ onto the manifold we simply need to compute ν, ω and η by evaluating the SVD of F T ΘT . [sent-237, score-0.396]
95 Furthermore, incorporating the low-rank assumption using the tangent space improves the info. [sent-241, score-0.521]
96 Thus, the different designs could choose radically different stimulus sets; in contrast, when re-analyzing the data offline, all we can do is reshuffle the trials, but the stimulus sets remain the same in the info. [sent-247, score-0.317]
97 4 Conclusion We have provided a method for incorporating detailed prior information in existing algorithms for the information-theoretic optimal design of neurophysiology experiments. [sent-250, score-0.367]
98 These methods use realistic assumptions about the neuron’s response function and choose significantly more informative stimuli, leading to faster convergence to the true response function using fewer experimental trials. [sent-251, score-0.246]
99 We expect that the inclusion of this strong prior information will help experimentalists contend with the high dimensionality of neural response functions. [sent-252, score-0.183]
100 We plot µt for trials in the interval over which the expected log-likelihood of the different designs differed the most in Fig. [sent-261, score-0.164]
wordName wordTfidf (topN-words)
[('strf', 0.467), ('tangent', 0.407), ('manifold', 0.344), ('receptive', 0.168), ('neurophysiology', 0.149), ('design', 0.126), ('glm', 0.12), ('informativeness', 0.12), ('st', 0.116), ('rt', 0.109), ('designs', 0.107), ('stimulus', 0.105), ('response', 0.105), ('neuron', 0.102), ('posterior', 0.097), ('paninski', 0.096), ('trial', 0.092), ('mutual', 0.086), ('gabor', 0.086), ('rank', 0.085), ('nch', 0.082), ('ptan', 0.082), ('shuf', 0.082), ('principal', 0.079), ('eld', 0.078), ('zebra', 0.072), ('mf', 0.072), ('space', 0.067), ('perturb', 0.066), ('frequency', 0.063), ('et', 0.062), ('georgia', 0.062), ('stimuli', 0.06), ('neuroscience', 0.06), ('trials', 0.057), ('constrain', 0.056), ('birdsong', 0.055), ('khz', 0.055), ('lewi', 0.055), ('liam', 0.055), ('mld', 0.055), ('orth', 0.055), ('shuffled', 0.055), ('woolley', 0.055), ('frequencies', 0.055), ('optimizing', 0.055), ('onto', 0.052), ('neurophysiol', 0.051), ('parametric', 0.051), ('tractable', 0.048), ('neurophysiologists', 0.048), ('reshaping', 0.048), ('reshuf', 0.048), ('auditory', 0.048), ('incorporating', 0.047), ('columns', 0.047), ('mt', 0.047), ('columbia', 0.045), ('full', 0.045), ('prior', 0.045), ('sinusoidal', 0.044), ('bird', 0.044), ('jeremy', 0.044), ('nonlinear', 0.044), ('tan', 0.041), ('robustly', 0.041), ('matrices', 0.041), ('approximation', 0.04), ('restrict', 0.04), ('subspace', 0.039), ('sine', 0.039), ('svd', 0.039), ('uncertainty', 0.039), ('tted', 0.037), ('experiment', 0.036), ('faster', 0.036), ('ring', 0.035), ('lie', 0.034), ('conditioning', 0.034), ('expect', 0.033), ('entails', 0.033), ('cosine', 0.033), ('manifolds', 0.033), ('derivatives', 0.032), ('integrating', 0.031), ('pick', 0.031), ('projection', 0.031), ('furthermore', 0.031), ('knowledge', 0.031), ('along', 0.031), ('dot', 0.03), ('iid', 0.03), ('gaussian', 0.03), ('map', 0.03), ('partial', 0.029), ('orthonormal', 0.029), ('mathematically', 0.029), ('formed', 0.028), ('derivative', 0.028), ('optimize', 0.028)]
simIndex simValue paperId paperTitle
same-paper 1 0.99999952 60 nips-2008-Designing neurophysiology experiments to optimally constrain receptive field models along parametric submanifolds
Author: Jeremy Lewi, Robert Butera, David M. Schneider, Sarah Woolley, Liam Paninski
Abstract: Sequential optimal design methods hold great promise for improving the efficiency of neurophysiology experiments. However, previous methods for optimal experimental design have incorporated only weak prior information about the underlying neural system (e.g., the sparseness or smoothness of the receptive field). Here we describe how to use stronger prior information, in the form of parametric models of the receptive field, in order to construct optimal stimuli and further improve the efficiency of our experiments. For example, if we believe that the receptive field is well-approximated by a Gabor function, then our method constructs stimuli that optimally constrain the Gabor parameters (orientation, spatial frequency, etc.) using as few experimental trials as possible. More generally, we may believe a priori that the receptive field lies near a known sub-manifold of the full parameter space; in this case, our method chooses stimuli in order to reduce the uncertainty along the tangent space of this sub-manifold as rapidly as possible. Applications to simulated and real data indicate that these methods may in many cases improve the experimental efficiency. 1
2 0.21074036 151 nips-2008-Non-parametric Regression Between Manifolds
Author: Florian Steinke, Matthias Hein
Abstract: This paper discusses non-parametric regression between Riemannian manifolds. This learning problem arises frequently in many application areas ranging from signal processing, computer vision, over robotics to computer graphics. We present a new algorithmic scheme for the solution of this general learning problem based on regularized empirical risk minimization. The regularization functional takes into account the geometry of input and output manifold, and we show that it implements a prior which is particularly natural. Moreover, we demonstrate that our algorithm performs well in a difficult surface registration problem. 1
3 0.14672862 51 nips-2008-Convergence and Rate of Convergence of a Manifold-Based Dimension Reduction Algorithm
Author: Andrew Smith, Hongyuan Zha, Xiao-ming Wu
Abstract: We study the convergence and the rate of convergence of a local manifold learning algorithm: LTSA [13]. The main technical tool is the perturbation analysis on the linear invariant subspace that corresponds to the solution of LTSA. We derive a worst-case upper bound of errors for LTSA which naturally leads to a convergence result. We then derive the rate of convergence for LTSA in a special case. 1
4 0.11034069 231 nips-2008-Temporal Dynamics of Cognitive Control
Author: Jeremy Reynolds, Michael C. Mozer
Abstract: Cognitive control refers to the flexible deployment of memory and attention in response to task demands and current goals. Control is often studied experimentally by presenting sequences of stimuli, some demanding a response, and others modulating the stimulus-response mapping. In these tasks, participants must maintain information about the current stimulus-response mapping in working memory. Prominent theories of cognitive control use recurrent neural nets to implement working memory, and optimize memory utilization via reinforcement learning. We present a novel perspective on cognitive control in which working memory representations are intrinsically probabilistic, and control operations that maintain and update working memory are dynamically determined via probabilistic inference. We show that our model provides a parsimonious account of behavioral and neuroimaging data, and suggest that it offers an elegant conceptualization of control in which behavior can be cast as optimal, subject to limitations on learning and the rate of information processing. Moreover, our model provides insight into how task instructions can be directly translated into appropriate behavior and then efficiently refined with subsequent task experience. 1
5 0.092538595 206 nips-2008-Sequential effects: Superstition or rational behavior?
Author: Angela J. Yu, Jonathan D. Cohen
Abstract: In a variety of behavioral tasks, subjects exhibit an automatic and apparently suboptimal sequential effect: they respond more rapidly and accurately to a stimulus if it reinforces a local pattern in stimulus history, such as a string of repetitions or alternations, compared to when it violates such a pattern. This is often the case even if the local trends arise by chance in the context of a randomized design, such that stimulus history has no real predictive power. In this work, we use a normative Bayesian framework to examine the hypothesis that such idiosyncrasies may reflect the inadvertent engagement of mechanisms critical for adapting to a changing environment. We show that prior belief in non-stationarity can induce experimentally observed sequential effects in an otherwise Bayes-optimal algorithm. The Bayesian algorithm is shown to be well approximated by linear-exponential filtering of past observations, a feature also apparent in the behavioral data. We derive an explicit relationship between the parameters and computations of the exact Bayesian algorithm and those of the approximate linear-exponential filter. Since the latter is equivalent to a leaky-integration process, a commonly used model of neuronal dynamics underlying perceptual decision-making and trial-to-trial dependencies, our model provides a principled account of why such dynamics are useful. We also show that parameter-tuning of the leaky-integration process is possible, using stochastic gradient descent based only on the noisy binary inputs. This is a proof of concept that not only can neurons implement near-optimal prediction based on standard neuronal dynamics, but that they can also learn to tune the processing parameters without explicitly representing probabilities. 1
6 0.082877003 109 nips-2008-Interpreting the neural code with Formal Concept Analysis
7 0.081202999 188 nips-2008-QUIC-SVD: Fast SVD Using Cosine Trees
8 0.075023562 172 nips-2008-Optimal Response Initiation: Why Recent Experience Matters
9 0.073098503 24 nips-2008-An improved estimator of Variance Explained in the presence of noise
10 0.072929285 227 nips-2008-Supervised Exponential Family Principal Component Analysis via Convex Optimization
11 0.070880689 38 nips-2008-Bio-inspired Real Time Sensory Map Realignment in a Robotic Barn Owl
12 0.069823027 194 nips-2008-Regularized Learning with Networks of Features
13 0.067809701 30 nips-2008-Bayesian Experimental Design of Magnetic Resonance Imaging Sequences
14 0.063243881 240 nips-2008-Tracking Changing Stimuli in Continuous Attractor Neural Networks
15 0.062945768 81 nips-2008-Extracting State Transition Dynamics from Multiple Spike Trains with Correlated Poisson HMM
16 0.062073871 67 nips-2008-Effects of Stimulus Type and of Error-Correcting Code Design on BCI Speller Performance
17 0.060732294 179 nips-2008-Phase transitions for high-dimensional joint support recovery
18 0.059442054 59 nips-2008-Dependent Dirichlet Process Spike Sorting
19 0.058178499 241 nips-2008-Transfer Learning by Distribution Matching for Targeted Advertising
20 0.05809366 244 nips-2008-Unifying the Sensory and Motor Components of Sensorimotor Adaptation
topicId topicWeight
[(0, -0.191), (1, 0.039), (2, 0.098), (3, 0.099), (4, 0.023), (5, 0.015), (6, 0.007), (7, 0.054), (8, 0.059), (9, 0.045), (10, 0.057), (11, -0.014), (12, -0.125), (13, 0.037), (14, -0.056), (15, 0.067), (16, 0.072), (17, 0.138), (18, 0.028), (19, -0.063), (20, -0.061), (21, -0.049), (22, 0.008), (23, -0.123), (24, -0.091), (25, 0.069), (26, 0.032), (27, 0.065), (28, 0.043), (29, 0.153), (30, -0.04), (31, -0.077), (32, 0.15), (33, 0.016), (34, 0.012), (35, -0.285), (36, -0.165), (37, 0.094), (38, 0.069), (39, -0.04), (40, 0.051), (41, -0.089), (42, -0.037), (43, 0.092), (44, -0.014), (45, 0.084), (46, -0.011), (47, 0.005), (48, -0.007), (49, -0.109)]
simIndex simValue paperId paperTitle
same-paper 1 0.94619495 60 nips-2008-Designing neurophysiology experiments to optimally constrain receptive field models along parametric submanifolds
Author: Jeremy Lewi, Robert Butera, David M. Schneider, Sarah Woolley, Liam Paninski
Abstract: Sequential optimal design methods hold great promise for improving the efficiency of neurophysiology experiments. However, previous methods for optimal experimental design have incorporated only weak prior information about the underlying neural system (e.g., the sparseness or smoothness of the receptive field). Here we describe how to use stronger prior information, in the form of parametric models of the receptive field, in order to construct optimal stimuli and further improve the efficiency of our experiments. For example, if we believe that the receptive field is well-approximated by a Gabor function, then our method constructs stimuli that optimally constrain the Gabor parameters (orientation, spatial frequency, etc.) using as few experimental trials as possible. More generally, we may believe a priori that the receptive field lies near a known sub-manifold of the full parameter space; in this case, our method chooses stimuli in order to reduce the uncertainty along the tangent space of this sub-manifold as rapidly as possible. Applications to simulated and real data indicate that these methods may in many cases improve the experimental efficiency. 1
2 0.79869735 151 nips-2008-Non-parametric Regression Between Manifolds
Author: Florian Steinke, Matthias Hein
Abstract: This paper discusses non-parametric regression between Riemannian manifolds. This learning problem arises frequently in many application areas ranging from signal processing, computer vision, over robotics to computer graphics. We present a new algorithmic scheme for the solution of this general learning problem based on regularized empirical risk minimization. The regularization functional takes into account the geometry of input and output manifold, and we show that it implements a prior which is particularly natural. Moreover, we demonstrate that our algorithm performs well in a difficult surface registration problem. 1
3 0.61911339 51 nips-2008-Convergence and Rate of Convergence of a Manifold-Based Dimension Reduction Algorithm
Author: Andrew Smith, Hongyuan Zha, Xiao-ming Wu
Abstract: We study the convergence and the rate of convergence of a local manifold learning algorithm: LTSA [13]. The main technical tool is the perturbation analysis on the linear invariant subspace that corresponds to the solution of LTSA. We derive a worst-case upper bound of errors for LTSA which naturally leads to a convergence result. We then derive the rate of convergence for LTSA in a special case. 1
4 0.52205414 172 nips-2008-Optimal Response Initiation: Why Recent Experience Matters
Author: Matt Jones, Sachiko Kinoshita, Michael C. Mozer
Abstract: In most cognitive and motor tasks, speed-accuracy tradeoffs are observed: Individuals can respond slowly and accurately, or quickly yet be prone to errors. Control mechanisms governing the initiation of behavioral responses are sensitive not only to task instructions and the stimulus being processed, but also to the recent stimulus history. When stimuli can be characterized on an easy-hard dimension (e.g., word frequency in a naming task), items preceded by easy trials are responded to more quickly, and with more errors, than items preceded by hard trials. We propose a rationally motivated mathematical model of this sequential adaptation of control, based on a diffusion model of the decision process in which difficulty corresponds to the drift rate for the correct response. The model assumes that responding is based on the posterior distribution over which response is correct, conditioned on the accumulated evidence. We derive this posterior as a function of the drift rate, and show that higher estimates of the drift rate lead to (normatively) faster responding. Trial-by-trial tracking of difficulty thus leads to sequential effects in speed and accuracy. Simulations show the model explains a variety of phenomena in human speeded decision making. We argue this passive statistical mechanism provides a more elegant and parsimonious account than extant theories based on elaborate control structures. 1
5 0.52130479 188 nips-2008-QUIC-SVD: Fast SVD Using Cosine Trees
Author: Michael P. Holmes, Jr. Isbell, Charles Lee, Alexander G. Gray
Abstract: The Singular Value Decomposition is a key operation in many machine learning methods. Its computational cost, however, makes it unscalable and impractical for applications involving large datasets or real-time responsiveness, which are becoming increasingly common. We present a new method, QUIC-SVD, for fast approximation of the whole-matrix SVD based on a new sampling mechanism called the cosine tree. Our empirical tests show speedups of several orders of magnitude over exact SVD. Such scalability should enable QUIC-SVD to accelerate and enable a wide array of SVD-based methods and applications. 1
6 0.51553035 126 nips-2008-Localized Sliced Inverse Regression
7 0.46246696 124 nips-2008-Load and Attentional Bayes
8 0.45397335 231 nips-2008-Temporal Dynamics of Cognitive Control
9 0.39908183 67 nips-2008-Effects of Stimulus Type and of Error-Correcting Code Design on BCI Speller Performance
10 0.39879793 109 nips-2008-Interpreting the neural code with Formal Concept Analysis
11 0.39186102 8 nips-2008-A general framework for investigating how far the decoding process in the brain can be simplified
12 0.37136066 90 nips-2008-Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population activity
13 0.36214927 206 nips-2008-Sequential effects: Superstition or rational behavior?
14 0.3596696 38 nips-2008-Bio-inspired Real Time Sensory Map Realignment in a Robotic Barn Owl
15 0.341934 240 nips-2008-Tracking Changing Stimuli in Continuous Attractor Neural Networks
16 0.33537969 122 nips-2008-Learning with Consistency between Inductive Functions and Kernels
17 0.32768229 23 nips-2008-An ideal observer model of infant object perception
18 0.3245399 46 nips-2008-Characterizing response behavior in multisensory perception with conflicting cues
19 0.3150512 66 nips-2008-Dynamic visual attention: searching for coding length increments
20 0.31390849 61 nips-2008-Diffeomorphic Dimensionality Reduction
topicId topicWeight
[(6, 0.061), (7, 0.135), (12, 0.039), (15, 0.021), (25, 0.011), (28, 0.181), (57, 0.063), (59, 0.019), (63, 0.027), (71, 0.019), (77, 0.046), (83, 0.035), (94, 0.277)]
simIndex simValue paperId paperTitle
1 0.84455538 136 nips-2008-Model selection and velocity estimation using novel priors for motion patterns
Author: Shuang Wu, Hongjing Lu, Alan L. Yuille
Abstract: Psychophysical experiments show that humans are better at perceiving rotation and expansion than translation. These findings are inconsistent with standard models of motion integration which predict best performance for translation [6]. To explain this discrepancy, our theory formulates motion perception at two levels of inference: we first perform model selection between the competing models (e.g. translation, rotation, and expansion) and then estimate the velocity using the selected model. We define novel prior models for smooth rotation and expansion using techniques similar to those in the slow-and-smooth model [17] (e.g. Green functions of differential operators). The theory gives good agreement with the trends observed in human experiments. 1
same-paper 2 0.80802393 60 nips-2008-Designing neurophysiology experiments to optimally constrain receptive field models along parametric submanifolds
Author: Jeremy Lewi, Robert Butera, David M. Schneider, Sarah Woolley, Liam Paninski
Abstract: Sequential optimal design methods hold great promise for improving the efficiency of neurophysiology experiments. However, previous methods for optimal experimental design have incorporated only weak prior information about the underlying neural system (e.g., the sparseness or smoothness of the receptive field). Here we describe how to use stronger prior information, in the form of parametric models of the receptive field, in order to construct optimal stimuli and further improve the efficiency of our experiments. For example, if we believe that the receptive field is well-approximated by a Gabor function, then our method constructs stimuli that optimally constrain the Gabor parameters (orientation, spatial frequency, etc.) using as few experimental trials as possible. More generally, we may believe a priori that the receptive field lies near a known sub-manifold of the full parameter space; in this case, our method chooses stimuli in order to reduce the uncertainty along the tangent space of this sub-manifold as rapidly as possible. Applications to simulated and real data indicate that these methods may in many cases improve the experimental efficiency. 1
3 0.80673313 215 nips-2008-Sparse Signal Recovery Using Markov Random Fields
Author: Volkan Cevher, Marco F. Duarte, Chinmay Hegde, Richard Baraniuk
Abstract: Compressive Sensing (CS) combines sampling and compression into a single subNyquist linear measurement process for sparse and compressible signals. In this paper, we extend the theory of CS to include signals that are concisely represented in terms of a graphical model. In particular, we use Markov Random Fields (MRFs) to represent sparse signals whose nonzero coefficients are clustered. Our new model-based recovery algorithm, dubbed Lattice Matching Pursuit (LaMP), stably recovers MRF-modeled signals using many fewer measurements and computations than the current state-of-the-art algorithms.
4 0.76102078 17 nips-2008-Algorithms for Infinitely Many-Armed Bandits
Author: Yizao Wang, Jean-yves Audibert, Rémi Munos
Abstract: We consider multi-armed bandit problems where the number of arms is larger than the possible number of experiments. We make a stochastic assumption on the mean-reward of a new selected arm which characterizes its probability of being a near-optimal arm. Our assumption is weaker than in previous works. We describe algorithms based on upper-confidence-bounds applied to a restricted set of randomly selected arms and provide upper-bounds on the resulting expected regret. We also derive a lower-bound which matches (up to a logarithmic factor) the upper-bound in some cases. 1
5 0.65758157 71 nips-2008-Efficient Sampling for Gaussian Process Inference using Control Variables
Author: Neil D. Lawrence, Magnus Rattray, Michalis K. Titsias
Abstract: Sampling functions in Gaussian process (GP) models is challenging because of the highly correlated posterior distribution. We describe an efficient Markov chain Monte Carlo algorithm for sampling from the posterior process of the GP model. This algorithm uses control variables which are auxiliary function values that provide a low dimensional representation of the function. At each iteration, the algorithm proposes new values for the control variables and generates the function from the conditional GP prior. The control variable input locations are found by minimizing an objective function. We demonstrate the algorithm on regression and classification problems and we use it to estimate the parameters of a differential equation model of gene regulation. 1
6 0.65161014 45 nips-2008-Characterizing neural dependencies with copula models
7 0.64820254 62 nips-2008-Differentiable Sparse Coding
8 0.64801073 137 nips-2008-Modeling Short-term Noise Dependence of Spike Counts in Macaque Prefrontal Cortex
9 0.64555532 63 nips-2008-Dimensionality Reduction for Data in Multiple Feature Representations
10 0.64490032 221 nips-2008-Stochastic Relational Models for Large-scale Dyadic Data using MCMC
11 0.64475393 129 nips-2008-MAS: a multiplicative approximation scheme for probabilistic inference
12 0.64390701 79 nips-2008-Exploring Large Feature Spaces with Hierarchical Multiple Kernel Learning
13 0.64283758 4 nips-2008-A Scalable Hierarchical Distributed Language Model
14 0.64280182 138 nips-2008-Modeling human function learning with Gaussian processes
15 0.64191729 135 nips-2008-Model Selection in Gaussian Graphical Models: High-Dimensional Consistency of \boldmath$\ell 1$-regularized MLE
16 0.64180171 66 nips-2008-Dynamic visual attention: searching for coding length increments
17 0.64135003 99 nips-2008-High-dimensional support union recovery in multivariate regression
18 0.64089167 118 nips-2008-Learning Transformational Invariants from Natural Movies
19 0.64053619 218 nips-2008-Spectral Clustering with Perturbed Data
20 0.64028609 54 nips-2008-Covariance Estimation for High Dimensional Data Vectors Using the Sparse Matrix Transform