nips nips2007 nips2007-104 knowledge-graph by maker-knowledge-mining
Source: pdf
Author: Maneesh Sahani, Byron M. Yu, John P. Cunningham, Krishna V. Shenoy
Abstract: Neural spike trains present challenges to analytical efforts due to their noisy, spiking nature. Many studies of neuroscientific and neural prosthetic importance rely on a smoothed, denoised estimate of the spike train’s underlying firing rate. Current techniques to find time-varying firing rates require ad hoc choices of parameters, offer no confidence intervals on their estimates, and can obscure potentially important single trial variability. We present a new method, based on a Gaussian Process prior, for inferring probabilistically optimal estimates of firing rate functions underlying single or multiple neural spike trains. We test the performance of the method on simulated data and experimentally gathered neural spike trains, and we demonstrate improvements over conventional estimators. 1
Reference: text
sentIndex sentText sentNum sentScore
1 uk 3 Abstract Neural spike trains present challenges to analytical efforts due to their noisy, spiking nature. [sent-8, score-0.821]
2 Many studies of neuroscientific and neural prosthetic importance rely on a smoothed, denoised estimate of the spike train’s underlying firing rate. [sent-9, score-0.622]
3 Current techniques to find time-varying firing rates require ad hoc choices of parameters, offer no confidence intervals on their estimates, and can obscure potentially important single trial variability. [sent-10, score-0.207]
4 We present a new method, based on a Gaussian Process prior, for inferring probabilistically optimal estimates of firing rate functions underlying single or multiple neural spike trains. [sent-11, score-0.713]
5 We test the performance of the method on simulated data and experimentally gathered neural spike trains, and we demonstrate improvements over conventional estimators. [sent-12, score-0.677]
6 Even when experimental conditions are repeated closely, the same neuron may produce quite different spike trains from trial to trial. [sent-14, score-0.812]
7 This variability may be due to both randomness in the spiking process and to differences in cognitive processing on different experimental trials. [sent-15, score-0.089]
8 One common view is that a spike train is generated from a smooth underlying function of time (the firing rate) and that this function carries a significant portion of the neural information. [sent-16, score-0.66]
9 If this is the case, questions of neuroscientific and neural prosthetic importance may require an accurate estimate of the firing rate. [sent-17, score-0.074]
10 Unfortunately, these estimates are complicated by the fact that spike data gives only a sparse observation of its underlying rate. [sent-18, score-0.571]
11 Typically, researchers average across many trials to find a smooth estimate (averaging out spiking noise). [sent-19, score-0.118]
12 However, averaging across many roughly similar trials can obscure important temporal features [1]. [sent-20, score-0.094]
13 Thus, estimating the underlying rate from only one spike train (or a small number of spike trains believed to be generated from the same underlying rate) is an important but challenging problem. [sent-21, score-1.509]
14 The most common approach to the problem has been to collect spikes from multiple trials in a peristimulus-time histogram (PSTH), which is then sometimes smoothed by convolution or splines [2], [3]. [sent-22, score-0.164]
15 Bin sizes and smoothness parameters are typically chosen ad hoc (but see [4], [5]) and the result is fundamentally a multi-trial analysis. [sent-23, score-0.077]
16 An alternative is to convolve a single spike train with a kernel. [sent-24, score-0.584]
17 Again, the kernel shape and time scale are frequently ad hoc. [sent-25, score-0.099]
18 1 More recently, point process likelihood methods have been adapted to spike data [6]–[8]. [sent-28, score-0.542]
19 These methods optimize (implicitly or explicitly) the conditional intensity function λ(t|x(t), H(t)) — which gives the probability of a spike in [t, t + dt), given an underlying rate function x(t) and the history of previous spikes H(t) — with respect to x(t). [sent-29, score-0.713]
20 In a regression setting, this rate x(t) may be learned as a function of an observed covariate, such as a sensory stimulus or limb movement. [sent-30, score-0.082]
21 In other fields, a theory of generalized Cox processes has been developed, where the point process is conditionally Poisson, and x(t) is obtained by applying a link function to a draw from a random process, often a Gaussian process (GP) (e. [sent-39, score-0.105]
22 However, the link function, which ensures a nonnegative intensity, introduces possibly undesirable artifacts. [sent-43, score-0.09]
23 For instance, an exponential link leads to a process that grows less smooth as the intensity increases. [sent-44, score-0.12]
24 First, we adapt the theory of GP-driven point processes to incorporate a history-dependent conditional likelihood, suitable for spike trains. [sent-46, score-0.509]
25 We show that GP methods employing evidence optimization outperform both kernel smoothing and maximum-likelihood point process models. [sent-49, score-0.104]
26 2 Gaussian Process Model For Spike Trains Spike trains can often be well modelled by gamma-interval point processes [6], [10]. [sent-50, score-0.256]
27 We assume the underlying nonnegative firing rate x(t) : t ∈ [0, T ] is a draw from a GP, and then we assume that our spike train is a conditionally inhomogeneous gamma-interval process (IGIP), given x(t). [sent-51, score-0.894]
28 The spike train is represented by a list of spike times y = {y0 , . [sent-52, score-1.093]
29 [6]) can be written as: p(yi | yi−1 , x(t)) = γx(yi ) γ Γ(γ) γ−1 yi x(u)du yi−1 yi exp −γ x(u)du . [sent-58, score-0.15]
30 (2) yi−1 The true p0 (·) and pT (·) under this gamma-interval spiking model are not closed form, so we simplify these distributions as intervals of an inhomogeneous Poisson process (IP). [sent-59, score-0.242]
31 The discretized IGIP output process is now (ignoring terms that scale with ∆): 1 The IGIP is one of a class of renewal models that works well for spike data (much better than inhomogeneous Poisson; see [6], [10]). [sent-67, score-0.738]
32 Other log-concave renewal models such as the inhomogeneous inverse-Gaussian interval can be chosen, and the implementation details remain unchanged. [sent-68, score-0.196]
33 2 N p(y | x) = i=1 γxyi γ Γ(γ) yi −1 yi −1 γ−1 exp −γ xk ∆ k=yi−1 y0 −1 · xy0 exp − xk ∆ k=yi−1 n−1 xk ∆ · exp − k=0 (3) xk ∆ , k=yN where the final two terms are p0 (·) and pT (·), respectively [11]. [sent-69, score-0.354]
34 Our goal is to estimate a smoothly varying firing rate function from spike times. [sent-70, score-0.591]
35 Since our intensity function is nonnegative, however, it is sensible to treat µ instead as a hyperparameter and let it be optimized to a positive value. [sent-79, score-0.105]
36 As written, the model assumes only one observed spike train; it may be that we have m trials believed to be generated from the same firing rate profile. [sent-82, score-0.653]
37 Our method naturally incorporates this case: define m p(y(i) | x), where y(i) denotes the ith spike train observed. [sent-83, score-0.584]
38 1 Finding an Optimal Firing Rate Estimate Algorithmic Approach Ideally, we would calculate the posterior on firing rate p(x | y) = θ p(x | y, θ)p(θ)dθ (integrating over the hyperparameters θ), but this problem is intractable. [sent-86, score-0.146]
39 We consider two approximations: replacing the integral by evaluation at the modal θ, and replacing the integral with a sum over a discrete grid of θ values. [sent-87, score-0.099]
40 We first consider choosing a modal hyperparameter set (ML-II model selection, see [12]), i. [sent-88, score-0.13]
41 p(x | y) ≈ q(x | y, θ ∗ ) where q(·) is some approximate posterior, and θ∗ = argmax p(θ | y) = argmax p(θ)p(y | θ) = argmax p(θ) θ θ θ p(y | x, θ)p(x | θ)dx. [sent-90, score-0.135]
42 Below we show this integrand is log concave in x. [sent-98, score-0.071]
43 Further, since we are modelling a non-zero mean GP, most of the Laplace approximated probability mass lies in the nonnegative orthant (as is the case with the true posterior). [sent-100, score-0.117]
44 Accordingly, we write: 2 Another reasonable approach would consider each trial as having a different rate function x that is a draw from a GP with a nonstationary mean function µ(t). [sent-101, score-0.156]
45 Instead of inferring a mean rate function x ∗ , we would learn a distribution of means. [sent-102, score-0.105]
46 The Laplace approximation also naturally provides confidence intervals from the approximated posterior covariance (Σ −1 + Λ∗ )−1 . [sent-109, score-0.072]
47 The approximate integrated posterior can be written as p(x | y) = Eθ|y [p(x | y, θ)] ≈ j q(x | y, θj )q(θj | y) for some choice of samples θj (which again gives confidence intervals on the estimates). [sent-113, score-0.072]
48 This approximate integration consistently yields better results than a modal hyperparameter set, so we will only consider approximate integration for the remainder of this report. [sent-115, score-0.18]
49 For the Laplace approximation at any value of θ, we require the modal estimate of firing rate x ∗ , which is simply the MAP estimator: x∗ = argmax p(x | y) = argmax p(y | x)p(x). [sent-116, score-0.245]
50 Typically a link or squashing function would be included to enforce nonnegativity in x, but this can distort the intensity space in unintended ways. [sent-118, score-0.116]
51 Accordingly, we can use a log barrier Newton method to efficiently solve for the global MAP estimator of firing rate x∗ [15]. [sent-145, score-0.112]
52 In the case of multiple spike train observations, we need only add extra terms of negative log likelihood from the observation model. [sent-146, score-0.614]
53 With 1 ms time resolution (or similar), this method would be restricted to spike trains lasting less than a second, and even this problem would be burdensome. [sent-158, score-0.837]
54 Since we have evenly spaced resolution of our data x in time indices ti , Σ is Toeplitz; thus multiplication by Σ can be done using Fast Fourier Transform (FFT) methods [16]. [sent-165, score-0.084]
55 For the modal hyperparameter scheme (as opposed to approximately integrating over the hyperparameters), gradients of Eq. [sent-168, score-0.153]
56 These data sizes translate to seconds of spike data at 1 ms resolution, long enough for most electrophysiological trials. [sent-172, score-0.551]
57 4 Results We tested the methods developed here using both simulated neural data, where the true firing rate was known by construction, and in real neural spike trains, where the true firing rate was estimated by a PSTH that averaged many similar trials. [sent-174, score-0.828]
58 We compared the IGIP-likelihood GP method (hereafter, GP IGIP) to other rate estimators (kernel smoothers, Bayesian Adaptive Regressions Splines or BARS [3], and variants of the GP method) using root mean squared difference (RMS) to the true firing rate. [sent-177, score-0.082]
59 PSTH and kernel methods approximate the mean conditional intensity λ(t) = EH(t) [λ(t|x(t), H(t))]. [sent-178, score-0.119]
60 For a renewal process, we know (by the time rescaling theorem [7], [11]) that λ(t) = x(t), and thus we can compare the GP IGIP (which finds x(t)) directly to the kernel methods. [sent-179, score-0.162]
61 To confirm that hyperparameter optimization improves performance, we also compared GP IGIP results to maximum likelihood (ML) estimates of x(t) using fixed hyperparameters θ. [sent-180, score-0.12]
62 To evaluate the importance of an observation model with spike history dependence (the IGIP of Eq. [sent-182, score-0.509]
63 3), we also compared GP IGIP to an inhomogeneous Poisson (GP IP) observation model (again with a GP prior on x(t); simply γ = 1 in Eq. [sent-183, score-0.105]
64 For the approximate integration, we chose a grid consisting of the empirical mean rate for µ (that is, total spike count N divided by total time T ) 2 and (γ, log(σf ), log(κ)) ∈ [1, 2, 4] × [4, . [sent-193, score-0.617]
65 1; 4 spike trains 50 16 14 40 Firing Rate (spikes/sec) Firing Rate (spikes/sec) 45 35 30 25 20 15 10 10 8 6 4 2 5 0 12 0 0. [sent-219, score-0.765]
66 3; 1 spike train Figure 1: Sample GP firing rate estimate. [sent-237, score-0.666]
67 1 represent experimentally gathered firing rate profiles (according to the methods in [17]). [sent-240, score-0.129]
68 In each of the plots, the empirical average firing rate of the spike trains is shown in bold red. [sent-241, score-0.887]
69 For simulated spike trains, the spike trains were generated from each of these empirical average firing rates using an IGIP (γ = 4, comparable to fits to real neural data). [sent-242, score-1.42]
70 For real neural data, the spike train(s) were selected as a subset of the roughly 200 experimentally recorded spike trains that were used to construct the firing rate profile. [sent-243, score-1.447]
71 These spike trains are shown as a train of black dots, each dot indicating a spike event time (the y-axis position is not meaningful). [sent-244, score-1.349]
72 This spike train or group of spike trains is the only input given to each of the fitting models. [sent-245, score-1.349]
73 In thin green and magenta, we have two kernel smoothed estimates of firing rates; each represents the spike trains convolved with a normal distribution of a specified standard deviation (50 and 100 ms). [sent-246, score-0.894]
74 We also smoothed these spike trains with adaptive kernel [18], fixed ML (as described above), BARS [3], and 150 ms kernel smoothers. [sent-247, score-1.015]
75 The light blue envelopes around the bold blue GP firing rate estimate represent the 95% confidence intervals. [sent-252, score-0.172]
76 We then ran all methods 100 times on each firing rate profile, using (separately) simulated and real neural spike trains. [sent-256, score-0.709]
77 The four panels correspond to the same rate profiles shown in Fig. [sent-262, score-0.108]
78 In each panel, the top, middle, and bottom bar graphs correspond to the method on 1, 4, and 8 spike trains, respectively. [sent-264, score-0.509]
79 6 50 GP Methods GP IP Fixed ML short Kernel Smoothers medium long adaptive BARS 50 % 0 50 % % 0 BARS % 0 50 Kernel Smoothers medium long adaptive 0 50 % short % 0 50 GP Methods GP IP Fixed ML 0 (a) L20061107. [sent-271, score-0.208]
80 1; 1,4,8 spike trains 50 GP Methods GP IP Fixed ML short Kernel Smoothers medium long adaptive BARS (b) L20061107. [sent-273, score-0.869]
81 1; 1,4,8 spike trains 50 % 0 50 % % 0 BARS % 0 50 Kernel Smoothers medium long adaptive 0 50 % short % 0 50 GP Methods GP IP Fixed ML 0 (c) L20061107. [sent-275, score-0.869]
82 3; 1,4,8 spike trains Figure 2: Average percent RMS improvement of GP IGIP method (with model selection) vs. [sent-279, score-0.765]
83 Blue improvement bars are for simulated spike trains; red improvement bars are for real neural spike trains. [sent-282, score-1.252]
84 the long bandwidth kernel in panel (b), real spike trains, 4 and 8 spike trains), outperforming the GP IGIP method requires an optimal kernel choice, which can not be judged from the data alone. [sent-286, score-1.218]
85 In particular, the adaptive kernel method generally performed more poorly than GP IGIP. [sent-287, score-0.102]
86 5 Discussion We have demonstrated a new method that accurately estimates underlying neural firing rate functions and provides confidence intervals, given one or a few spike trains as input. [sent-291, score-0.946]
87 Estimating underlying firing rates is especially challenging due to the inherent noise in spike trains. [sent-293, score-0.576]
88 Having only a few spike trains deprives the method of many trials to reduce spiking noise. [sent-294, score-0.883]
89 It is important here to remember why we care about single trial or small number of trial estimates, since we believe that in general the neural processing on repeated trials is not identical. [sent-295, score-0.193]
90 In this study we show both simulated and real neural spike trains. [sent-297, score-0.627]
91 Simulated data provides a good test environment for this method, since the underlying firing rate is known, but it lacks the experimental proof of real neural spike trains (where spiking does not exactly follow a gamma-interval process). [sent-298, score-1.009]
92 Taken together, however, we believe the real and simulated data give good evidence of the general improvements offered by this method. [sent-300, score-0.114]
93 In simulation, GP IGIP generally outperforms the other smoothers (though, by considerably less than in other panels). [sent-304, score-0.087]
94 This may indicate that, in the low firing rate regime, the IGIP is a poor model for real neural spiking. [sent-306, score-0.149]
95 Furthermore, some neural spike trains may be inherently ill-suited to analysis. [sent-309, score-0.802]
96 With spike trains of only a few spikes/sec, it will be impossible for any method to find interesting structure in the firing rate. [sent-311, score-0.765]
97 Several studies have investigated the inhomogeneous gamma and other more general models (e. [sent-313, score-0.105]
98 [6], [19]), including the inhomogeneous inverse gaussian (IIG) interval and inhomogeneous Markov interval (IMI) processes. [sent-315, score-0.21]
99 The methods of this paper apply immediately to any log-concave inhomogeneous renewal process in which inhomogeneity is generated by time-rescaling (this includes the IIG and several others). [sent-316, score-0.229]
100 Another direction for this work is to consider significant nonstationarity in the spike data. [sent-318, score-0.509]
wordName wordTfidf (topN-words)
[('spike', 0.509), ('gp', 0.455), ('igip', 0.402), ('trains', 0.256), ('ring', 0.219), ('inhomogeneous', 0.105), ('laplace', 0.101), ('firing', 0.095), ('renewal', 0.091), ('smoothers', 0.087), ('rate', 0.082), ('yi', 0.075), ('train', 0.075), ('ip', 0.074), ('modal', 0.073), ('kernel', 0.071), ('trials', 0.062), ('bars', 0.058), ('hyperparameter', 0.057), ('spiking', 0.056), ('ml', 0.054), ('simulated', 0.051), ('xk', 0.051), ('nonnegative', 0.051), ('yn', 0.049), ('intervals', 0.048), ('intensity', 0.048), ('kass', 0.048), ('inversions', 0.048), ('ryu', 0.048), ('trial', 0.047), ('medium', 0.047), ('fixed', 0.047), ('argmax', 0.045), ('psth', 0.044), ('se', 0.043), ('hessian', 0.042), ('ms', 0.042), ('integrand', 0.041), ('bold', 0.04), ('hyperparameters', 0.04), ('sec', 0.039), ('underlying', 0.039), ('link', 0.039), ('neural', 0.037), ('pro', 0.037), ('barbieri', 0.037), ('gilja', 0.037), ('iig', 0.037), ('imi', 0.037), ('matern', 0.037), ('orthant', 0.037), ('prosthetic', 0.037), ('xyi', 0.037), ('pt', 0.036), ('smoothed', 0.035), ('spikes', 0.035), ('poisson', 0.034), ('improvements', 0.033), ('process', 0.033), ('dence', 0.032), ('determinants', 0.032), ('neuroscienti', 0.032), ('afshar', 0.032), ('obscure', 0.032), ('santhanam', 0.032), ('splines', 0.032), ('ventura', 0.032), ('adaptive', 0.031), ('ti', 0.031), ('rms', 0.03), ('log', 0.03), ('resolution', 0.03), ('real', 0.03), ('nonnegativity', 0.029), ('modelling', 0.029), ('tj', 0.029), ('panel', 0.028), ('rates', 0.028), ('ad', 0.028), ('nonstationary', 0.027), ('maneesh', 0.027), ('panels', 0.026), ('grid', 0.026), ('short', 0.026), ('meaningful', 0.026), ('comp', 0.026), ('spline', 0.026), ('integration', 0.025), ('smoothness', 0.025), ('blue', 0.025), ('hoc', 0.024), ('experimentally', 0.024), ('posterior', 0.024), ('multiplication', 0.023), ('gathered', 0.023), ('inferring', 0.023), ('gradients', 0.023), ('estimates', 0.023), ('sacri', 0.022)]
simIndex simValue paperId paperTitle
same-paper 1 0.99999994 104 nips-2007-Inferring Neural Firing Rates from Spike Trains Using Gaussian Processes
Author: Maneesh Sahani, Byron M. Yu, John P. Cunningham, Krishna V. Shenoy
Abstract: Neural spike trains present challenges to analytical efforts due to their noisy, spiking nature. Many studies of neuroscientific and neural prosthetic importance rely on a smoothed, denoised estimate of the spike train’s underlying firing rate. Current techniques to find time-varying firing rates require ad hoc choices of parameters, offer no confidence intervals on their estimates, and can obscure potentially important single trial variability. We present a new method, based on a Gaussian Process prior, for inferring probabilistically optimal estimates of firing rate functions underlying single or multiple neural spike trains. We test the performance of the method on simulated data and experimentally gathered neural spike trains, and we demonstrate improvements over conventional estimators. 1
2 0.31813487 36 nips-2007-Better than least squares: comparison of objective functions for estimating linear-nonlinear models
Author: Tatyana Sharpee
Abstract: This paper compares a family of methods for characterizing neural feature selectivity with natural stimuli in the framework of the linear-nonlinear model. In this model, the neural firing rate is a nonlinear function of a small number of relevant stimulus components. The relevant stimulus dimensions can be found by maximizing one of the family of objective functions, R´ nyi divergences of different e orders [1, 2]. We show that maximizing one of them, R´ nyi divergence of ore der 2, is equivalent to least-square fitting of the linear-nonlinear model to neural data. Next, we derive reconstruction errors in relevant dimensions found by maximizing R´ nyi divergences of arbitrary order in the asymptotic limit of large spike e numbers. We find that the smallest errors are obtained with R´ nyi divergence of e order 1, also known as Kullback-Leibler divergence. This corresponds to finding relevant dimensions by maximizing mutual information [2]. We numerically test how these optimization schemes perform in the regime of low signal-to-noise ratio (small number of spikes and increasing neural noise) for model visual neurons. We find that optimization schemes based on either least square fitting or information maximization perform well even when number of spikes is small. Information maximization provides slightly, but significantly, better reconstructions than least square fitting. This makes the problem of finding relevant dimensions, together with the problem of lossy compression [3], one of examples where informationtheoretic measures are no more data limited than those derived from least squares. 1
3 0.29059827 94 nips-2007-Gaussian Process Models for Link Analysis and Transfer Learning
Author: Kai Yu, Wei Chu
Abstract: This paper aims to model relational data on edges of networks. We describe appropriate Gaussian Processes (GPs) for directed, undirected, and bipartite networks. The inter-dependencies of edges can be effectively modeled by adapting the GP hyper-parameters. The framework suggests an intimate connection between link prediction and transfer learning, which were traditionally two separate research topics. We develop an efficient learning algorithm that can handle a large number of observations. The experimental results on several real-world data sets verify superior learning capacity. 1
4 0.2158999 170 nips-2007-Robust Regression with Twinned Gaussian Processes
Author: Andrew Naish-guzman, Sean Holden
Abstract: We propose a Gaussian process (GP) framework for robust inference in which a GP prior on the mixing weights of a two-component noise model augments the standard process over latent function values. This approach is a generalization of the mixture likelihood used in traditional robust GP regression, and a specialization of the GP mixture models suggested by Tresp [1] and Rasmussen and Ghahramani [2]. The value of this restriction is in its tractable expectation propagation updates, which allow for faster inference and model selection, and better convergence than the standard mixture. An additional benefit over the latter method lies in our ability to incorporate knowledge of the noise domain to influence predictions, and to recover with the predictive distribution information about the outlier distribution via the gating process. The model has asymptotic complexity equal to that of conventional robust methods, but yields more confident predictions on benchmark problems than classical heavy-tailed models and exhibits improved stability for data with clustered corruptions, for which they fail altogether. We show further how our approach can be used without adjustment for more smoothly heteroscedastic data, and suggest how it could be extended to more general noise models. We also address similarities with the work of Goldberg et al. [3].
5 0.21198839 140 nips-2007-Neural characterization in partially observed populations of spiking neurons
Author: Jonathan W. Pillow, Peter E. Latham
Abstract: Point process encoding models provide powerful statistical methods for understanding the responses of neurons to sensory stimuli. Although these models have been successfully applied to neurons in the early sensory pathway, they have fared less well capturing the response properties of neurons in deeper brain areas, owing in part to the fact that they do not take into account multiple stages of processing. Here we introduce a new twist on the point-process modeling approach: we include unobserved as well as observed spiking neurons in a joint encoding model. The resulting model exhibits richer dynamics and more highly nonlinear response properties, making it more powerful and more flexible for fitting neural data. More importantly, it allows us to estimate connectivity patterns among neurons (both observed and unobserved), and may provide insight into how networks process sensory input. We formulate the estimation procedure using variational EM and the wake-sleep algorithm, and illustrate the model’s performance using a simulated example network consisting of two coupled neurons.
6 0.1924119 17 nips-2007-A neural network implementing optimal state estimation based on dynamic spike train decoding
7 0.18963268 33 nips-2007-Bayesian Inference for Spiking Neuron Models with a Sparsity Prior
9 0.18464299 32 nips-2007-Bayesian Co-Training
10 0.17035876 35 nips-2007-Bayesian binning beats approximate alternatives: estimating peri-stimulus time histograms
11 0.16931045 205 nips-2007-Theoretical Analysis of Learning with Reward-Modulated Spike-Timing-Dependent Plasticity
12 0.16636334 212 nips-2007-Using Deep Belief Nets to Learn Covariance Kernels for Gaussian Processes
13 0.14357729 135 nips-2007-Multi-task Gaussian Process Prediction
14 0.13008668 117 nips-2007-Learning to classify complex patterns using a VLSI network of spiking neurons
15 0.11491485 103 nips-2007-Inferring Elapsed Time from Stochastic Neural Processes
16 0.11057498 161 nips-2007-Random Projections for Manifold Learning
17 0.10860728 195 nips-2007-The Generalized FITC Approximation
18 0.10720682 164 nips-2007-Receptive Fields without Spike-Triggering
19 0.076859549 14 nips-2007-A configurable analog VLSI neural network with spiking neurons and self-regulating plastic synapses
20 0.076489076 79 nips-2007-Efficient multiple hyperparameter learning for log-linear models
topicId topicWeight
[(0, -0.246), (1, 0.188), (2, 0.286), (3, 0.183), (4, -0.045), (5, -0.072), (6, -0.178), (7, -0.258), (8, -0.008), (9, 0.055), (10, -0.251), (11, -0.047), (12, -0.074), (13, 0.096), (14, -0.071), (15, -0.036), (16, 0.05), (17, 0.06), (18, 0.083), (19, 0.094), (20, -0.084), (21, -0.1), (22, 0.014), (23, 0.149), (24, 0.006), (25, -0.106), (26, 0.032), (27, -0.068), (28, -0.008), (29, -0.002), (30, 0.077), (31, -0.083), (32, 0.013), (33, 0.162), (34, 0.011), (35, 0.037), (36, -0.053), (37, 0.009), (38, -0.039), (39, -0.002), (40, -0.052), (41, 0.037), (42, 0.025), (43, 0.033), (44, 0.014), (45, 0.027), (46, -0.028), (47, -0.026), (48, -0.014), (49, 0.062)]
simIndex simValue paperId paperTitle
same-paper 1 0.96858817 104 nips-2007-Inferring Neural Firing Rates from Spike Trains Using Gaussian Processes
Author: Maneesh Sahani, Byron M. Yu, John P. Cunningham, Krishna V. Shenoy
Abstract: Neural spike trains present challenges to analytical efforts due to their noisy, spiking nature. Many studies of neuroscientific and neural prosthetic importance rely on a smoothed, denoised estimate of the spike train’s underlying firing rate. Current techniques to find time-varying firing rates require ad hoc choices of parameters, offer no confidence intervals on their estimates, and can obscure potentially important single trial variability. We present a new method, based on a Gaussian Process prior, for inferring probabilistically optimal estimates of firing rate functions underlying single or multiple neural spike trains. We test the performance of the method on simulated data and experimentally gathered neural spike trains, and we demonstrate improvements over conventional estimators. 1
2 0.74147731 36 nips-2007-Better than least squares: comparison of objective functions for estimating linear-nonlinear models
Author: Tatyana Sharpee
Abstract: This paper compares a family of methods for characterizing neural feature selectivity with natural stimuli in the framework of the linear-nonlinear model. In this model, the neural firing rate is a nonlinear function of a small number of relevant stimulus components. The relevant stimulus dimensions can be found by maximizing one of the family of objective functions, R´ nyi divergences of different e orders [1, 2]. We show that maximizing one of them, R´ nyi divergence of ore der 2, is equivalent to least-square fitting of the linear-nonlinear model to neural data. Next, we derive reconstruction errors in relevant dimensions found by maximizing R´ nyi divergences of arbitrary order in the asymptotic limit of large spike e numbers. We find that the smallest errors are obtained with R´ nyi divergence of e order 1, also known as Kullback-Leibler divergence. This corresponds to finding relevant dimensions by maximizing mutual information [2]. We numerically test how these optimization schemes perform in the regime of low signal-to-noise ratio (small number of spikes and increasing neural noise) for model visual neurons. We find that optimization schemes based on either least square fitting or information maximization perform well even when number of spikes is small. Information maximization provides slightly, but significantly, better reconstructions than least square fitting. This makes the problem of finding relevant dimensions, together with the problem of lossy compression [3], one of examples where informationtheoretic measures are no more data limited than those derived from least squares. 1
3 0.68802035 35 nips-2007-Bayesian binning beats approximate alternatives: estimating peri-stimulus time histograms
Author: Dominik Endres, Mike Oram, Johannes Schindelin, Peter Foldiak
Abstract: The peristimulus time histogram (PSTH) and its more continuous cousin, the spike density function (SDF) are staples in the analytic toolkit of neurophysiologists. The former is usually obtained by binning spike trains, whereas the standard method for the latter is smoothing with a Gaussian kernel. Selection of a bin width or a kernel size is often done in an relatively arbitrary fashion, even though there have been recent attempts to remedy this situation [1, 2]. We develop an exact Bayesian, generative model approach to estimating PSTHs and demonstate its superiority to competing methods. Further advantages of our scheme include automatic complexity control and error bars on its predictions. 1
4 0.56386352 33 nips-2007-Bayesian Inference for Spiking Neuron Models with a Sparsity Prior
Author: Sebastian Gerwinn, Matthias Bethge, Jakob H. Macke, Matthias Seeger
Abstract: Generalized linear models are the most commonly used tools to describe the stimulus selectivity of sensory neurons. Here we present a Bayesian treatment of such models. Using the expectation propagation algorithm, we are able to approximate the full posterior distribution over all weights. In addition, we use a Laplacian prior to favor sparse solutions. Therefore, stimulus features that do not critically influence neural activity will be assigned zero weights and thus be effectively excluded by the model. This feature selection mechanism facilitates both the interpretation of the neuron model as well as its predictive abilities. The posterior distribution can be used to obtain confidence intervals which makes it possible to assess the statistical significance of the solution. In neural data analysis, the available amount of experimental measurements is often limited whereas the parameter space is large. In such a situation, both regularization by a sparsity prior and uncertainty estimates for the model parameters are essential. We apply our method to multi-electrode recordings of retinal ganglion cells and use our uncertainty estimate to test the statistical significance of functional couplings between neurons. Furthermore we used the sparsity of the Laplace prior to select those filters from a spike-triggered covariance analysis that are most informative about the neural response. 1
5 0.5629217 94 nips-2007-Gaussian Process Models for Link Analysis and Transfer Learning
Author: Kai Yu, Wei Chu
Abstract: This paper aims to model relational data on edges of networks. We describe appropriate Gaussian Processes (GPs) for directed, undirected, and bipartite networks. The inter-dependencies of edges can be effectively modeled by adapting the GP hyper-parameters. The framework suggests an intimate connection between link prediction and transfer learning, which were traditionally two separate research topics. We develop an efficient learning algorithm that can handle a large number of observations. The experimental results on several real-world data sets verify superior learning capacity. 1
6 0.53193909 140 nips-2007-Neural characterization in partially observed populations of spiking neurons
7 0.52304244 170 nips-2007-Robust Regression with Twinned Gaussian Processes
8 0.50680202 205 nips-2007-Theoretical Analysis of Learning with Reward-Modulated Spike-Timing-Dependent Plasticity
9 0.50112879 32 nips-2007-Bayesian Co-Training
10 0.47115475 177 nips-2007-Simplified Rules and Theoretical Analysis for Information Bottleneck Optimization and PCA with Spiking Neurons
11 0.43047786 17 nips-2007-A neural network implementing optimal state estimation based on dynamic spike train decoding
12 0.40397337 164 nips-2007-Receptive Fields without Spike-Triggering
13 0.39244717 195 nips-2007-The Generalized FITC Approximation
14 0.38240743 212 nips-2007-Using Deep Belief Nets to Learn Covariance Kernels for Gaussian Processes
15 0.37895557 103 nips-2007-Inferring Elapsed Time from Stochastic Neural Processes
16 0.37578917 133 nips-2007-Modelling motion primitives and their timing in biologically executed movements
17 0.35190928 135 nips-2007-Multi-task Gaussian Process Prediction
18 0.30994812 97 nips-2007-Hidden Common Cause Relations in Relational Learning
19 0.29698098 117 nips-2007-Learning to classify complex patterns using a VLSI network of spiking neurons
20 0.2849856 25 nips-2007-An in-silico Neural Model of Dynamic Routing through Neuronal Coherence
topicId topicWeight
[(5, 0.036), (13, 0.028), (16, 0.176), (18, 0.027), (19, 0.017), (21, 0.106), (25, 0.133), (31, 0.023), (34, 0.029), (35, 0.041), (46, 0.013), (47, 0.076), (49, 0.019), (83, 0.089), (85, 0.031), (87, 0.018), (90, 0.058)]
simIndex simValue paperId paperTitle
1 0.84584492 170 nips-2007-Robust Regression with Twinned Gaussian Processes
Author: Andrew Naish-guzman, Sean Holden
Abstract: We propose a Gaussian process (GP) framework for robust inference in which a GP prior on the mixing weights of a two-component noise model augments the standard process over latent function values. This approach is a generalization of the mixture likelihood used in traditional robust GP regression, and a specialization of the GP mixture models suggested by Tresp [1] and Rasmussen and Ghahramani [2]. The value of this restriction is in its tractable expectation propagation updates, which allow for faster inference and model selection, and better convergence than the standard mixture. An additional benefit over the latter method lies in our ability to incorporate knowledge of the noise domain to influence predictions, and to recover with the predictive distribution information about the outlier distribution via the gating process. The model has asymptotic complexity equal to that of conventional robust methods, but yields more confident predictions on benchmark problems than classical heavy-tailed models and exhibits improved stability for data with clustered corruptions, for which they fail altogether. We show further how our approach can be used without adjustment for more smoothly heteroscedastic data, and suggest how it could be extended to more general noise models. We also address similarities with the work of Goldberg et al. [3].
same-paper 2 0.84342498 104 nips-2007-Inferring Neural Firing Rates from Spike Trains Using Gaussian Processes
Author: Maneesh Sahani, Byron M. Yu, John P. Cunningham, Krishna V. Shenoy
Abstract: Neural spike trains present challenges to analytical efforts due to their noisy, spiking nature. Many studies of neuroscientific and neural prosthetic importance rely on a smoothed, denoised estimate of the spike train’s underlying firing rate. Current techniques to find time-varying firing rates require ad hoc choices of parameters, offer no confidence intervals on their estimates, and can obscure potentially important single trial variability. We present a new method, based on a Gaussian Process prior, for inferring probabilistically optimal estimates of firing rate functions underlying single or multiple neural spike trains. We test the performance of the method on simulated data and experimentally gathered neural spike trains, and we demonstrate improvements over conventional estimators. 1
3 0.83556062 33 nips-2007-Bayesian Inference for Spiking Neuron Models with a Sparsity Prior
Author: Sebastian Gerwinn, Matthias Bethge, Jakob H. Macke, Matthias Seeger
Abstract: Generalized linear models are the most commonly used tools to describe the stimulus selectivity of sensory neurons. Here we present a Bayesian treatment of such models. Using the expectation propagation algorithm, we are able to approximate the full posterior distribution over all weights. In addition, we use a Laplacian prior to favor sparse solutions. Therefore, stimulus features that do not critically influence neural activity will be assigned zero weights and thus be effectively excluded by the model. This feature selection mechanism facilitates both the interpretation of the neuron model as well as its predictive abilities. The posterior distribution can be used to obtain confidence intervals which makes it possible to assess the statistical significance of the solution. In neural data analysis, the available amount of experimental measurements is often limited whereas the parameter space is large. In such a situation, both regularization by a sparsity prior and uncertainty estimates for the model parameters are essential. We apply our method to multi-electrode recordings of retinal ganglion cells and use our uncertainty estimate to test the statistical significance of functional couplings between neurons. Furthermore we used the sparsity of the Laplace prior to select those filters from a spike-triggered covariance analysis that are most informative about the neural response. 1
4 0.8290894 60 nips-2007-Contraction Properties of VLSI Cooperative Competitive Neural Networks of Spiking Neurons
Author: Emre Neftci, Elisabetta Chicca, Giacomo Indiveri, Jean-jeacques Slotine, Rodney J. Douglas
Abstract: A non–linear dynamic system is called contracting if initial conditions are forgotten exponentially fast, so that all trajectories converge to a single trajectory. We use contraction theory to derive an upper bound for the strength of recurrent connections that guarantees contraction for complex neural networks. Specifically, we apply this theory to a special class of recurrent networks, often called Cooperative Competitive Networks (CCNs), which are an abstract representation of the cooperative-competitive connectivity observed in cortex. This specific type of network is believed to play a major role in shaping cortical responses and selecting the relevant signal among distractors and noise. In this paper, we analyze contraction of combined CCNs of linear threshold units and verify the results of our analysis in a hybrid analog/digital VLSI CCN comprising spiking neurons and dynamic synapses. 1
5 0.82706857 206 nips-2007-Topmoumoute Online Natural Gradient Algorithm
Author: Nicolas L. Roux, Pierre-antoine Manzagol, Yoshua Bengio
Abstract: Guided by the goal of obtaining an optimization algorithm that is both fast and yields good generalization, we study the descent direction maximizing the decrease in generalization error or the probability of not increasing generalization error. The surprising result is that from both the Bayesian and frequentist perspectives this can yield the natural gradient direction. Although that direction can be very expensive to compute we develop an efficient, general, online approximation to the natural gradient descent which is suited to large scale problems. We report experimental results showing much faster convergence in computation time and in number of iterations with TONGA (Topmoumoute Online natural Gradient Algorithm) than with stochastic gradient descent, even on very large datasets.
6 0.79786009 140 nips-2007-Neural characterization in partially observed populations of spiking neurons
7 0.76615256 36 nips-2007-Better than least squares: comparison of objective functions for estimating linear-nonlinear models
8 0.75302935 195 nips-2007-The Generalized FITC Approximation
9 0.74001187 164 nips-2007-Receptive Fields without Spike-Triggering
10 0.6995827 205 nips-2007-Theoretical Analysis of Learning with Reward-Modulated Spike-Timing-Dependent Plasticity
11 0.6921683 79 nips-2007-Efficient multiple hyperparameter learning for log-linear models
12 0.68410277 87 nips-2007-Fast Variational Inference for Large-scale Internet Diagnosis
13 0.67660952 213 nips-2007-Variational Inference for Diffusion Processes
14 0.66854197 35 nips-2007-Bayesian binning beats approximate alternatives: estimating peri-stimulus time histograms
15 0.6678499 28 nips-2007-Augmented Functional Time Series Representation and Forecasting with Gaussian Processes
16 0.66781098 177 nips-2007-Simplified Rules and Theoretical Analysis for Information Bottleneck Optimization and PCA with Spiking Neurons
17 0.66729975 138 nips-2007-Near-Maximum Entropy Models for Binary Neural Representations of Natural Images
18 0.66647416 174 nips-2007-Selecting Observations against Adversarial Objectives
19 0.66604447 117 nips-2007-Learning to classify complex patterns using a VLSI network of spiking neurons
20 0.66584796 125 nips-2007-Markov Chain Monte Carlo with People