nips nips2013 nips2013-266 knowledge-graph by maker-knowledge-mining
Source: pdf
Author: Marius Pachitariu, Biljana Petreska, Maneesh Sahani
Abstract: Population neural recordings with long-range temporal structure are often best understood in terms of a common underlying low-dimensional dynamical process. Advances in recording technology provide access to an ever-larger fraction of the population, but the standard computational approaches available to identify the collective dynamics scale poorly with the size of the dataset. We describe a new, scalable approach to discovering low-dimensional dynamics that underlie simultaneously recorded spike trains from a neural population. We formulate the Recurrent Linear Model (RLM) by generalising the Kalman-filter-based likelihood calculation for latent linear dynamical systems to incorporate a generalised-linear observation process. We show that RLMs describe motor-cortical population data better than either directly-coupled generalised-linear models or latent linear dynamical system models with generalised-linear observations. We also introduce the cascaded generalised-linear model (CGLM) to capture low-dimensional instantaneous correlations in neural populations. The CGLM describes the cortical recordings better than either Ising or Gaussian models and, like the RLM, can be fit exactly and quickly. The CGLM can also be seen as a generalisation of a lowrank Gaussian model, in this case factor analysis. The computational tractability of the RLM and CGLM allow both to scale to very high-dimensional neural data. 1
Reference: text
sentIndex sentText sentNum sentScore
1 Recurrent linear models of simultaneously-recorded neural populations Marius Pachitariu, Biljana Petreska, Maneesh Sahani Gatsby Computational Neuroscience Unit University College London, UK {marius,biljana,maneesh}@gatsby. [sent-1, score-0.062]
2 uk Abstract Population neural recordings with long-range temporal structure are often best understood in terms of a common underlying low-dimensional dynamical process. [sent-4, score-0.172]
3 Advances in recording technology provide access to an ever-larger fraction of the population, but the standard computational approaches available to identify the collective dynamics scale poorly with the size of the dataset. [sent-5, score-0.079]
4 We describe a new, scalable approach to discovering low-dimensional dynamics that underlie simultaneously recorded spike trains from a neural population. [sent-6, score-0.152]
5 We formulate the Recurrent Linear Model (RLM) by generalising the Kalman-filter-based likelihood calculation for latent linear dynamical systems to incorporate a generalised-linear observation process. [sent-7, score-0.26]
6 We show that RLMs describe motor-cortical population data better than either directly-coupled generalised-linear models or latent linear dynamical system models with generalised-linear observations. [sent-8, score-0.263]
7 We also introduce the cascaded generalised-linear model (CGLM) to capture low-dimensional instantaneous correlations in neural populations. [sent-9, score-0.205]
8 The CGLM describes the cortical recordings better than either Ising or Gaussian models and, like the RLM, can be fit exactly and quickly. [sent-10, score-0.038]
9 The CGLM can also be seen as a generalisation of a lowrank Gaussian model, in this case factor analysis. [sent-11, score-0.061]
10 The computational tractability of the RLM and CGLM allow both to scale to very high-dimensional neural data. [sent-12, score-0.033]
11 1 Introduction Many essential neural computations are implemented by large populations of neurons working in concert, and recent studies have sought both to monitor increasingly large groups of neurons [1, 2] and to characterise their collective behaviour [3, 4]. [sent-13, score-0.278]
12 In this paper we introduce a new computational tool to model coordinated behaviour in very large neural data sets. [sent-14, score-0.11]
13 While we explicitly discuss only multi-electrode extracellular recordings, the same model can be readily used to characterise 2-photon calcium-marker image data, EEG, fMRI or even large-scale biologically-faithful simulations. [sent-15, score-0.081]
14 Populational neural data may be represented at each time point by a vector yt with as many dimensions as neurons, and as many indices t as time points in the experiment. [sent-16, score-0.634]
15 For spiking neurons, yt will have positive integer elements corresponding to the number of spikes fired by each neuron in the time interval corresponding to the t-th bin. [sent-17, score-0.627]
16 As others have before [5, 6], we assume that the coordinated activity reflected in the measurement yt arises from a low-dimensional set of processes, collected into a vector xt , which is not directly observed. [sent-18, score-0.869]
17 However, unlike the previous studies, we construct a recurrent model in which the hidden processes xt are driven directly and explicitly by the measured neural signals y1 . [sent-19, score-0.361]
18 We assume for simplicity that xt evolves with linear dynamics and affects the future state of the neural signal yt in a generalised-linear manner, although both assumptions may be relaxed. [sent-24, score-0.907]
19 As in the latent dynamical system, the resulting model enforces a “bottleneck”, whereby predictions of yt based on y1 . [sent-25, score-0.795]
20 1 State prediction in the RLM is related to the Kalman filter [7] and we show in the next section a formal equivalence between the likelihoods of the RLM and the latent dynamical model when observation noise is Gaussian distributed. [sent-29, score-0.217]
21 However, spiking data is not well modelled as Gaussian, and the generalisation of our approach to Poisson noise leads to a departure from the latent dynamical approach. [sent-30, score-0.336]
22 Unlike latent linear models with conditionally Poisson observations, the parameters of our model can be estimated efficiently and without approximation. [sent-31, score-0.093]
23 We show that, perhaps in consequence, the RLM can provide superior descriptions of neural population data. [sent-32, score-0.082]
24 2 From the Kalman filter to the recurrent linear model (RLM) Consider a latent linear dynamical system (LDS) model with linear-Gaussian observations. [sent-33, score-0.34]
25 The latent process is parametrised by a dynamics matrix A and innovations covariance Q that describe the evolution of the latent state xt : P (xt |xt−1 ) = N (xt |Axt−1 , Q) , where N (x|µ, Σ) represents a normal distribution on x with mean µ and (co)variance Σ. [sent-36, score-0.661]
26 The output distribution is determined by an observation loading matrix C and a noise covariance R often taken to be diagonal so that all covariance is modelled by the latent process: P (yt |xt ) = N (yt |Cxt , R) . [sent-38, score-0.363]
27 In the LDS, the joint likelihood of the observations {yt } can be written as the product: T P (yt |y1 . [sent-39, score-0.041]
28 yT ) = P (y1 ) t=2 and in the Gaussian case can be computed using the usual Kalman filter approach to find the conditional distributon at time t iteratively: P (yt+1 |y1 . [sent-45, score-0.023]
29 yt ) ˆ dxt+1 N (yt+1 |Cxt+1 , R) N (xt+1 |Axt , Vt+1 ) ˆ = N (yt+1 |CAxt , CVt+1 C + R) , ˆ where we have introduced the (filtered) state estimate xt = E [xt |y1 . [sent-51, score-0.824]
30 yt ] and (predictive) unˆ certainty Vt+1 = E (xt+1 − Axt )2 |y1 . [sent-54, score-0.601]
31 Both quantities are computed recursively using the Kalman gain Kt = Vt C (CVt C + R)−1 , giving the following recursive recipe to calculate the conditional likelihood of yt+1 : ˆ ˆ ˆ xt = Axt−1 + Kt (yt − yt ) Vt+1 = A(I − Kt C)Vt A + Q ˆ ˆ yt+1 = CAxt ˆ P (yt+1 |y1 . [sent-58, score-0.865]
32 yt ) = N (yt+1 |yt+1 , CVt+1 C + R) For the Gaussian LDS, the Kalman gain Kt and state uncertainty Vt+1 (and thus the output covariance CVt+1 C + R) depend on the model parameters (A, C, R, Q) and on the time step—although as time grows they both converge to stationary values. [sent-61, score-0.767]
33 Thus, we might consider a relaxation of the Gaussian LDS model in which these matrices are taken to be stationary from the outset, and are parametrised independently so that they are no longer constrained to take on the “correct” values as computed for Kalman inference. [sent-63, score-0.071]
34 Let us call this parametric form of the Kalman gain W and the parametric form of the output covariance S. [sent-64, score-0.121]
35 Then the conditional likelihood iteration becomes ˆ ˆ ˆ xt = Axt−1 + W (yt − yt ) ˆ ˆ yt+1 = CAxt ˆ P (yt+1 |y1 . [sent-65, score-0.824]
36 Shaded LDS variables are observed, unshaded circles are latent random variables and squares are variables that depend deterministically on their parents. [sent-70, score-0.121]
37 In B the LDS is redrawn in terms of the random innovations ηt = xt − Axt−1 , facilitating the transition towards the RLM. [sent-71, score-0.307]
38 The RLM RLM is then obtained by replacing ηt with a deterministically derived ˆ estimate W (yt − yt ). [sent-72, score-0.65]
39 This is a relaxation of the Gaussian latent LDS model because W has more degrees of freedom than Q, as does S than R (at least if R is constrained to be diagonal). [sent-74, score-0.093]
40 The new model has a recurrent linear structure in that the random ˆ observation yt is fed back linearly to perturb the otherwise deterministic evolution of the state xt . [sent-75, score-1.054]
41 The RLM can be viewed as replacing the random innovation variables ηt = xt − Axt−1 ˆ with data-derived estimates W (yt − yt ); estimates which are made possible by the fact that ηt ˆ contributes to the variability of yt around yt . [sent-78, score-2.026]
42 3 Recurrent linear models with Poisson observations The discussion above has transformed a stochastic-latent LDS model with Gaussian output to an RLM with deterministic latent, but still with Gaussian output. [sent-79, score-0.088]
43 Our goal, however, is to fit a model with an output distribution better suited to the binned point-processes that characterise neural spiking. [sent-80, score-0.143]
44 Both linear Kalman-filtering steps above and the eventual stationarity of the inference parameters depend on the joint Gaussian structure of the assumed LDS model. [sent-81, score-0.039]
45 They would not apply if we were to begin a similar derivation from an LDS with Poisson output. [sent-82, score-0.026]
46 However, a tractable approach to modelling point-process data with low-dimensional temporal structure may be provided by introducing a generalised-linear output stage directly to the RLM. [sent-83, score-0.076]
47 This model is given by: ˆ ˆ ˆ xt = Axt−1 + W (yt − yt ) ˆ ˆ g(yt+1 ) = CAxt ˆ P (yt+1 |y1 . [sent-84, score-0.803]
48 yt ) = ExpFam(yt+1 |yt+1 ) (1) where ExpFam is an exponential-family distribution such as Poisson, and the element-wise link ˆ function g allows for a nonlinear mapping from xt to the predicted mean yt+1 . [sent-87, score-0.849]
49 In the following, we ˆ ˆ will write f for the inverse-link as is more common for neural models, so that yt+1 = f(CAxt ). [sent-88, score-0.033]
50 The simplest Poisson-based generalised-linear RLM might take as its output distribution ˆ P (yt |yt ) = Poisson(yti |ˆti ); y ˆ ˆ yt = f(CAxt−1 )) , i where yti is the spike count of the ith cell in bin t and the function f is non-negative. [sent-89, score-0.813]
51 However, comparison with the output distribution derived for the Gaussian RLM suggests that this choice would fail to capture the instantaneous covariance that the LDS formulation transfers to the output distribution (and which appears in the low-rank structure of S above). [sent-90, score-0.204]
52 One option is to bin the data more finely, thus diminishing the influence of the instantaneous covariance. [sent-92, score-0.082]
53 The alternative is to replace the independent Poissons with a correlated output distribution on spike counts. [sent-93, score-0.118]
54 The cascaded generalised-linear model introduced below is a natural choice, and we will show that it captures instantaneous correlations faithfully with very few hidden dimensions. [sent-94, score-0.172]
55 3 In practice, we also sometimes add a fixed input µt to equation 1 that varies in time and determines the average behavior of the population or the peri-stimulus time histogram (PSTH). [sent-95, score-0.049]
56 The matrix A controls the evolution of the dynamical process xt . [sent-97, score-0.364]
57 The phenomenology of its dynamics is determined by the complex eigenvalues of A. [sent-98, score-0.111]
58 Eigenvalues with moduli close to 1 correspond to long timescales of fluctuation around the PSTH. [sent-99, score-0.042]
59 Eigenvalues with non-zero imaginary part correspond to oscillatory components. [sent-100, score-0.02]
60 Finally, the dynamics will be stable iff all the eigenvalues lie within the unit disc. [sent-101, score-0.088]
61 The matrix C describes the dependence of the high-dimensional neural signals on the lowdimensional latent processes xt . [sent-102, score-0.347]
62 This generalised-linear stage ensures that the firing rates are positive through the link function f, and the observation process is Poisson. [sent-104, score-0.096]
63 For other types of data, the generalised-linear stage might be replaced by other appropriate link functions and output distributions. [sent-105, score-0.122]
64 1 Relationship to other models RLMs are related to recurrent neural networks [8]. [sent-107, score-0.159]
65 The differences lie in the state evolution, which in the neural network is nonlinear: xt = h (Axt−1 + W yt−1 ); and in the recurrent term which depends on the observation rather than the prediction error. [sent-108, score-0.405]
66 On the data considered here, we found that using sigmoidal or threshold-linear functions h resulted in models comparable in likelihood to the RLM, and so we restricted our attention to simple linear dynamics. [sent-109, score-0.045]
67 We also found that ˆ using the prediction error term W (yt−1 − yt ) resulted in better models than the simple neural-net formulation, and we attribute this difference to the link between the RLM and Kalman inference. [sent-110, score-0.671]
68 It is also possible to work within the stochatic latent LDS framework, replacing the Gaussian output distribution with a generalised-linear Poisson output (e. [sent-111, score-0.212]
69 For an unobserved latent process xt , an inference procedure needs to be devised to estimate the posterior distribution on the entire sequence x1 . [sent-115, score-0.295]
70 However, with generalised-linear observations, inference becomes intractable and the necessary approximations [6] are computationally intense and can jeopardize the quality of the fitted models. [sent-120, score-0.022]
71 By contrast, in the RLM xt is a deterministic function of data. [sent-121, score-0.221]
72 In effect, the Kalman filter has been built into the model as the accurate estimation procedure, and efficient fitting is possible by direct gradient ascent on the log-likelihood. [sent-122, score-0.02]
73 Note that to estimate the matrices A and W the gradient must be backpropagated through successive iterations of equation 1. [sent-125, score-0.022]
74 This technique, known as backpropagation-through-time, was first described by [10] as a technique to fit recurrent neural network models. [sent-126, score-0.159]
75 Backpropagation-through-time is thought to be inherently unstable when propagated past many timesteps and often the gradient is truncated prematurely [11]. [sent-128, score-0.023]
76 We found that using large values of momentum in the gradient ascent alleviated these instabilities and allowed us to use backpropagation without the truncation. [sent-129, score-0.043]
77 4 The cascaded generalised-linear model (CGLM) The link between the RLM and the LDS raises the possibility that a model for simultaneouslyrecorded correlated spike counts might be derived in a similar way, starting from a non-dynamical, but low-dimensional, Gaussian model. [sent-130, score-0.24]
78 Stationary models of population activity have attracted recent interest for their own sake (e. [sent-131, score-0.068]
79 [1]), and would also provide a way model correlations introduced by common innovations that were neglected by the simple Poisson form of the RLM. [sent-133, score-0.101]
80 Thus, we consider vectors y of spike counts from N neurons, without explicit reference to the time at which they were collected. [sent-134, score-0.09]
81 A Gaussian model for y can certainly describe correlations between the cells, but is ill-matched to discrete count observations. [sent-135, score-0.064]
82 Thus, as with the derivation of the RLM from the Kalman filter, we derive here a new generalisation of a low-dimensional, structured Gaussian model to spike count data. [sent-136, score-0.185]
83 4 The distribution of any multivariate variable y can be factorized into a “cascaded” product of multiple one-dimensional distributions: N P (yn |y < n and ≤ n restrict the matrix to the first (n − 1) and n rows and/or columns respectively. [sent-137, score-0.019]
84 Thus, we might construct suitably structured linear weights for the CGLM by applying this result to the covariance matrix induced by the low-dimensional Gaussian model known as factor analysis [12]. [sent-138, score-0.069]
85 Factor analysis assumes that data are generated from a K-dimensional latent process x ∼ N (0, I), where I is the K×K identity matrix, and y has the conditional distribution P (y|x) = N (Λx, Ψ) with Ψ a diagonal matrix and Λ an N × K loading matrix. [sent-139, score-0.155]
86 This leads to a covariance of y given by Σ = Ψ + ΛΛT . [sent-140, score-0.05]
wordName wordTfidf (topN-words)
[('yt', 0.601), ('rlm', 0.446), ('lds', 0.328), ('xt', 0.202), ('caxt', 0.164), ('cglm', 0.16), ('axt', 0.155), ('kalman', 0.152), ('recurrent', 0.126), ('dynamical', 0.101), ('cvt', 0.094), ('latent', 0.093), ('poisson', 0.082), ('cascaded', 0.081), ('spike', 0.069), ('vt', 0.066), ('generalisation', 0.061), ('characterise', 0.061), ('instantaneous', 0.056), ('expfam', 0.053), ('rlms', 0.053), ('kt', 0.051), ('dynamics', 0.05), ('covariance', 0.05), ('output', 0.049), ('population', 0.049), ('neurons', 0.048), ('coordinated', 0.047), ('parametrised', 0.047), ('link', 0.046), ('gaussian', 0.046), ('lter', 0.045), ('innovations', 0.044), ('dxt', 0.043), ('redrawn', 0.043), ('loading', 0.043), ('evolution', 0.042), ('cxt', 0.041), ('yti', 0.039), ('recordings', 0.038), ('eigenvalues', 0.038), ('ca', 0.037), ('modelled', 0.036), ('correlations', 0.035), ('neural', 0.033), ('behaviour', 0.03), ('collective', 0.029), ('count', 0.029), ('populations', 0.029), ('deterministically', 0.028), ('stage', 0.027), ('spiking', 0.026), ('derivation', 0.026), ('bin', 0.026), ('stationary', 0.024), ('resulted', 0.024), ('sn', 0.024), ('simultaneouslyrecorded', 0.023), ('outset', 0.023), ('phenomenology', 0.023), ('prematurely', 0.023), ('moduli', 0.023), ('instabilities', 0.023), ('distributon', 0.023), ('concert', 0.023), ('pachitariu', 0.023), ('ring', 0.023), ('observation', 0.023), ('gain', 0.022), ('neglected', 0.022), ('intense', 0.022), ('backpropagated', 0.022), ('generalising', 0.022), ('marius', 0.022), ('poissons', 0.022), ('petreska', 0.022), ('state', 0.021), ('replacing', 0.021), ('counts', 0.021), ('likelihood', 0.021), ('perturb', 0.02), ('uctuation', 0.02), ('imaginary', 0.02), ('eventual', 0.02), ('extracellular', 0.02), ('ascent', 0.02), ('observations', 0.02), ('system', 0.02), ('timescales', 0.019), ('maneesh', 0.019), ('stationarity', 0.019), ('nely', 0.019), ('recipe', 0.019), ('departure', 0.019), ('matrix', 0.019), ('activity', 0.019), ('deterministic', 0.019), ('eeg', 0.018), ('facilitating', 0.018), ('sahani', 0.018)]
simIndex simValue paperId paperTitle
same-paper 1 1.0000001 266 nips-2013-Recurrent linear models of simultaneously-recorded neural populations
Author: Marius Pachitariu, Biljana Petreska, Maneesh Sahani
Abstract: Population neural recordings with long-range temporal structure are often best understood in terms of a common underlying low-dimensional dynamical process. Advances in recording technology provide access to an ever-larger fraction of the population, but the standard computational approaches available to identify the collective dynamics scale poorly with the size of the dataset. We describe a new, scalable approach to discovering low-dimensional dynamics that underlie simultaneously recorded spike trains from a neural population. We formulate the Recurrent Linear Model (RLM) by generalising the Kalman-filter-based likelihood calculation for latent linear dynamical systems to incorporate a generalised-linear observation process. We show that RLMs describe motor-cortical population data better than either directly-coupled generalised-linear models or latent linear dynamical system models with generalised-linear observations. We also introduce the cascaded generalised-linear model (CGLM) to capture low-dimensional instantaneous correlations in neural populations. The CGLM describes the cortical recordings better than either Ising or Gaussian models and, like the RLM, can be fit exactly and quickly. The CGLM can also be seen as a generalisation of a lowrank Gaussian model, in this case factor analysis. The computational tractability of the RLM and CGLM allow both to scale to very high-dimensional neural data. 1
2 0.29188111 286 nips-2013-Robust learning of low-dimensional dynamics from large neural ensembles
Author: David Pfau, Eftychios A. Pnevmatikakis, Liam Paninski
Abstract: Recordings from large populations of neurons make it possible to search for hypothesized low-dimensional dynamics. Finding these dynamics requires models that take into account biophysical constraints and can be fit efficiently and robustly. Here, we present an approach to dimensionality reduction for neural data that is convex, does not make strong assumptions about dynamics, does not require averaging over many trials and is extensible to more complex statistical models that combine local and global influences. The results can be combined with spectral methods to learn dynamical systems models. The basic method extends PCA to the exponential family using nuclear norm minimization. We evaluate the effectiveness of this method using an exact decomposition of the Bregman divergence that is analogous to variance explained for PCA. We show on model data that the parameters of latent linear dynamical systems can be recovered, and that even if the dynamics are not stationary we can still recover the true latent subspace. We also demonstrate an extension of nuclear norm minimization that can separate sparse local connections from global latent dynamics. Finally, we demonstrate improved prediction on real neural data from monkey motor cortex compared to fitting linear dynamical models without nuclear norm smoothing. 1
3 0.19481342 62 nips-2013-Causal Inference on Time Series using Restricted Structural Equation Models
Author: Jonas Peters, Dominik Janzing, Bernhard Schölkopf
Abstract: Causal inference uses observational data to infer the causal structure of the data generating system. We study a class of restricted Structural Equation Models for time series that we call Time Series Models with Independent Noise (TiMINo). These models require independent residual time series, whereas traditional methods like Granger causality exploit the variance of residuals. This work contains two main contributions: (1) Theoretical: By restricting the model class (e.g. to additive noise) we provide general identifiability results. They cover lagged and instantaneous effects that can be nonlinear and unfaithful, and non-instantaneous feedbacks between the time series. (2) Practical: If there are no feedback loops between time series, we propose an algorithm based on non-linear independence tests of time series. We show empirically that when the data are causally insufficient or the model is misspecified, the method avoids incorrect answers. We extend the theoretical and the algorithmic part to situations in which the time series have been measured with different time delays. TiMINo is applied to artificial and real data and code is provided. 1
4 0.15724128 146 nips-2013-Large Scale Distributed Sparse Precision Estimation
Author: Huahua Wang, Arindam Banerjee, Cho-Jui Hsieh, Pradeep Ravikumar, Inderjit Dhillon
Abstract: We consider the problem of sparse precision matrix estimation in high dimensions using the CLIME estimator, which has several desirable theoretical properties. We present an inexact alternating direction method of multiplier (ADMM) algorithm for CLIME, and establish rates of convergence for both the objective and optimality conditions. Further, we develop a large scale distributed framework for the computations, which scales to millions of dimensions and trillions of parameters, using hundreds of cores. The proposed framework solves CLIME in columnblocks and only involves elementwise operations and parallel matrix multiplications. We evaluate our algorithm on both shared-memory and distributed-memory architectures, which can use block cyclic distribution of data and parameters to achieve load balance and improve the efficiency in the use of memory hierarchies. Experimental results show that our algorithm is substantially more scalable than state-of-the-art methods and scales almost linearly with the number of cores. 1
5 0.14011398 17 nips-2013-A multi-agent control framework for co-adaptation in brain-computer interfaces
Author: Josh S. Merel, Roy Fox, Tony Jebara, Liam Paninski
Abstract: In a closed-loop brain-computer interface (BCI), adaptive decoders are used to learn parameters suited to decoding the user’s neural response. Feedback to the user provides information which permits the neural tuning to also adapt. We present an approach to model this process of co-adaptation between the encoding model of the neural signal and the decoding algorithm as a multi-agent formulation of the linear quadratic Gaussian (LQG) control problem. In simulation we characterize how decoding performance improves as the neural encoding and adaptive decoder optimize, qualitatively resembling experimentally demonstrated closed-loop improvement. We then propose a novel, modified decoder update rule which is aware of the fact that the encoder is also changing and show it can improve simulated co-adaptation dynamics. Our modeling approach offers promise for gaining insights into co-adaptation as well as improving user learning of BCI control in practical settings.
6 0.13917147 48 nips-2013-Bayesian Inference and Learning in Gaussian Process State-Space Models with Particle MCMC
7 0.13880499 240 nips-2013-Optimization, Learning, and Games with Predictable Sequences
8 0.13800612 230 nips-2013-Online Learning with Costly Features and Labels
9 0.11872735 203 nips-2013-Multilinear Dynamical Systems for Tensor Time Series
10 0.1154085 262 nips-2013-Real-Time Inference for a Gamma Process Model of Neural Spiking
11 0.11289948 89 nips-2013-Dimension-Free Exponentiated Gradient
12 0.10081164 141 nips-2013-Inferring neural population dynamics from multiple partial recordings of the same neural circuit
13 0.09766873 269 nips-2013-Regression-tree Tuning in a Streaming Setting
14 0.097563073 162 nips-2013-Learning Trajectory Preferences for Manipulators via Iterative Improvement
15 0.092171304 341 nips-2013-Universal models for binary spike patterns using centered Dirichlet processes
16 0.090122007 39 nips-2013-Approximate Gaussian process inference for the drift function in stochastic differential equations
17 0.089003786 6 nips-2013-A Determinantal Point Process Latent Variable Model for Inhibition in Neural Spiking Data
18 0.08343982 41 nips-2013-Approximate inference in latent Gaussian-Markov models from continuous time observations
19 0.081234321 24 nips-2013-Actor-Critic Algorithms for Risk-Sensitive MDPs
20 0.078255653 348 nips-2013-Variational Policy Search via Trajectory Optimization
topicId topicWeight
[(0, 0.16), (1, 0.024), (2, 0.028), (3, -0.066), (4, -0.223), (5, -0.018), (6, 0.015), (7, 0.098), (8, 0.044), (9, -0.15), (10, -0.094), (11, -0.238), (12, -0.038), (13, 0.071), (14, -0.131), (15, -0.016), (16, 0.032), (17, 0.106), (18, 0.063), (19, 0.078), (20, 0.077), (21, 0.042), (22, 0.126), (23, -0.054), (24, 0.005), (25, -0.076), (26, 0.053), (27, -0.034), (28, 0.034), (29, -0.006), (30, 0.055), (31, -0.063), (32, 0.04), (33, 0.049), (34, -0.071), (35, 0.072), (36, -0.089), (37, -0.023), (38, 0.082), (39, 0.007), (40, 0.041), (41, 0.032), (42, 0.019), (43, 0.031), (44, -0.008), (45, 0.022), (46, 0.127), (47, -0.069), (48, -0.007), (49, 0.003)]
simIndex simValue paperId paperTitle
same-paper 1 0.97299081 266 nips-2013-Recurrent linear models of simultaneously-recorded neural populations
Author: Marius Pachitariu, Biljana Petreska, Maneesh Sahani
Abstract: Population neural recordings with long-range temporal structure are often best understood in terms of a common underlying low-dimensional dynamical process. Advances in recording technology provide access to an ever-larger fraction of the population, but the standard computational approaches available to identify the collective dynamics scale poorly with the size of the dataset. We describe a new, scalable approach to discovering low-dimensional dynamics that underlie simultaneously recorded spike trains from a neural population. We formulate the Recurrent Linear Model (RLM) by generalising the Kalman-filter-based likelihood calculation for latent linear dynamical systems to incorporate a generalised-linear observation process. We show that RLMs describe motor-cortical population data better than either directly-coupled generalised-linear models or latent linear dynamical system models with generalised-linear observations. We also introduce the cascaded generalised-linear model (CGLM) to capture low-dimensional instantaneous correlations in neural populations. The CGLM describes the cortical recordings better than either Ising or Gaussian models and, like the RLM, can be fit exactly and quickly. The CGLM can also be seen as a generalisation of a lowrank Gaussian model, in this case factor analysis. The computational tractability of the RLM and CGLM allow both to scale to very high-dimensional neural data. 1
2 0.76151741 62 nips-2013-Causal Inference on Time Series using Restricted Structural Equation Models
Author: Jonas Peters, Dominik Janzing, Bernhard Schölkopf
Abstract: Causal inference uses observational data to infer the causal structure of the data generating system. We study a class of restricted Structural Equation Models for time series that we call Time Series Models with Independent Noise (TiMINo). These models require independent residual time series, whereas traditional methods like Granger causality exploit the variance of residuals. This work contains two main contributions: (1) Theoretical: By restricting the model class (e.g. to additive noise) we provide general identifiability results. They cover lagged and instantaneous effects that can be nonlinear and unfaithful, and non-instantaneous feedbacks between the time series. (2) Practical: If there are no feedback loops between time series, we propose an algorithm based on non-linear independence tests of time series. We show empirically that when the data are causally insufficient or the model is misspecified, the method avoids incorrect answers. We extend the theoretical and the algorithmic part to situations in which the time series have been measured with different time delays. TiMINo is applied to artificial and real data and code is provided. 1
3 0.70810622 17 nips-2013-A multi-agent control framework for co-adaptation in brain-computer interfaces
Author: Josh S. Merel, Roy Fox, Tony Jebara, Liam Paninski
Abstract: In a closed-loop brain-computer interface (BCI), adaptive decoders are used to learn parameters suited to decoding the user’s neural response. Feedback to the user provides information which permits the neural tuning to also adapt. We present an approach to model this process of co-adaptation between the encoding model of the neural signal and the decoding algorithm as a multi-agent formulation of the linear quadratic Gaussian (LQG) control problem. In simulation we characterize how decoding performance improves as the neural encoding and adaptive decoder optimize, qualitatively resembling experimentally demonstrated closed-loop improvement. We then propose a novel, modified decoder update rule which is aware of the fact that the encoder is also changing and show it can improve simulated co-adaptation dynamics. Our modeling approach offers promise for gaining insights into co-adaptation as well as improving user learning of BCI control in practical settings.
4 0.62133205 286 nips-2013-Robust learning of low-dimensional dynamics from large neural ensembles
Author: David Pfau, Eftychios A. Pnevmatikakis, Liam Paninski
Abstract: Recordings from large populations of neurons make it possible to search for hypothesized low-dimensional dynamics. Finding these dynamics requires models that take into account biophysical constraints and can be fit efficiently and robustly. Here, we present an approach to dimensionality reduction for neural data that is convex, does not make strong assumptions about dynamics, does not require averaging over many trials and is extensible to more complex statistical models that combine local and global influences. The results can be combined with spectral methods to learn dynamical systems models. The basic method extends PCA to the exponential family using nuclear norm minimization. We evaluate the effectiveness of this method using an exact decomposition of the Bregman divergence that is analogous to variance explained for PCA. We show on model data that the parameters of latent linear dynamical systems can be recovered, and that even if the dynamics are not stationary we can still recover the true latent subspace. We also demonstrate an extension of nuclear norm minimization that can separate sparse local connections from global latent dynamics. Finally, we demonstrate improved prediction on real neural data from monkey motor cortex compared to fitting linear dynamical models without nuclear norm smoothing. 1
5 0.61611652 41 nips-2013-Approximate inference in latent Gaussian-Markov models from continuous time observations
Author: Botond Cseke, Manfred Opper, Guido Sanguinetti
Abstract: We propose an approximate inference algorithm for continuous time Gaussian Markov process models with both discrete and continuous time likelihoods. We show that the continuous time limit of the expectation propagation algorithm exists and results in a hybrid fixed point iteration consisting of (1) expectation propagation updates for discrete time terms and (2) variational updates for the continuous time term. We introduce postinference corrections methods that improve on the marginals of the approximation. This approach extends the classical Kalman-Bucy smoothing procedure to non-Gaussian observations, enabling continuous-time inference in a variety of models, including spiking neuronal models (state-space models with point process observations) and box likelihood models. Experimental results on real and simulated data demonstrate high distributional accuracy and significant computational savings compared to discrete-time approaches in a neural application. 1
6 0.56342351 39 nips-2013-Approximate Gaussian process inference for the drift function in stochastic differential equations
8 0.50290132 146 nips-2013-Large Scale Distributed Sparse Precision Estimation
9 0.47765619 262 nips-2013-Real-Time Inference for a Gamma Process Model of Neural Spiking
10 0.4420318 89 nips-2013-Dimension-Free Exponentiated Gradient
11 0.44074732 269 nips-2013-Regression-tree Tuning in a Streaming Setting
12 0.43011779 348 nips-2013-Variational Policy Search via Trajectory Optimization
13 0.42781442 280 nips-2013-Robust Data-Driven Dynamic Programming
14 0.42271006 298 nips-2013-Small-Variance Asymptotics for Hidden Markov Models
15 0.41006637 24 nips-2013-Actor-Critic Algorithms for Risk-Sensitive MDPs
16 0.40674096 45 nips-2013-BIG & QUIC: Sparse Inverse Covariance Estimation for a Million Variables
17 0.40143552 121 nips-2013-Firing rate predictions in optimal balanced networks
18 0.39405757 6 nips-2013-A Determinantal Point Process Latent Variable Model for Inhibition in Neural Spiking Data
19 0.38815817 240 nips-2013-Optimization, Learning, and Games with Predictable Sequences
20 0.38599327 141 nips-2013-Inferring neural population dynamics from multiple partial recordings of the same neural circuit
topicId topicWeight
[(16, 0.038), (33, 0.101), (34, 0.085), (41, 0.024), (49, 0.46), (56, 0.086), (70, 0.03), (85, 0.015), (89, 0.024), (93, 0.03)]
simIndex simValue paperId paperTitle
1 0.91050082 323 nips-2013-Synthesizing Robust Plans under Incomplete Domain Models
Author: Tuan A. Nguyen, Subbarao Kambhampati, Minh Do
Abstract: Most current planners assume complete domain models and focus on generating correct plans. Unfortunately, domain modeling is a laborious and error-prone task, thus real world agents have to plan with incomplete domain models. While domain experts cannot guarantee completeness, often they are able to circumscribe the incompleteness of the model by providing annotations as to which parts of the domain model may be incomplete. In such cases, the goal should be to synthesize plans that are robust with respect to any known incompleteness of the domain. In this paper, we first introduce annotations expressing the knowledge of the domain incompleteness and formalize the notion of plan robustness with respect to an incomplete domain model. We then show an approach to compiling the problem of finding robust plans to the conformant probabilistic planning problem, and present experimental results with Probabilistic-FF planner. 1
2 0.89681029 6 nips-2013-A Determinantal Point Process Latent Variable Model for Inhibition in Neural Spiking Data
Author: Jasper Snoek, Richard Zemel, Ryan P. Adams
Abstract: Point processes are popular models of neural spiking behavior as they provide a statistical distribution over temporal sequences of spikes and help to reveal the complexities underlying a series of recorded action potentials. However, the most common neural point process models, the Poisson process and the gamma renewal process, do not capture interactions and correlations that are critical to modeling populations of neurons. We develop a novel model based on a determinantal point process over latent embeddings of neurons that effectively captures and helps visualize complex inhibitory and competitive interaction. We show that this model is a natural extension of the popular generalized linear model to sets of interacting neurons. The model is extended to incorporate gain control or divisive normalization, and the modulation of neural spiking based on periodic phenomena. Applied to neural spike recordings from the rat hippocampus, we see that the model captures inhibitory relationships, a dichotomy of classes of neurons, and a periodic modulation by the theta rhythm known to be present in the data. 1
3 0.85403818 274 nips-2013-Relevance Topic Model for Unstructured Social Group Activity Recognition
Author: Fang Zhao, Yongzhen Huang, Liang Wang, Tieniu Tan
Abstract: Unstructured social group activity recognition in web videos is a challenging task due to 1) the semantic gap between class labels and low-level visual features and 2) the lack of labeled training data. To tackle this problem, we propose a “relevance topic model” for jointly learning meaningful mid-level representations upon bagof-words (BoW) video representations and a classifier with sparse weights. In our approach, sparse Bayesian learning is incorporated into an undirected topic model (i.e., Replicated Softmax) to discover topics which are relevant to video classes and suitable for prediction. Rectified linear units are utilized to increase the expressive power of topics so as to explain better video data containing complex contents and make variational inference tractable for the proposed model. An efficient variational EM algorithm is presented for model parameter estimation and inference. Experimental results on the Unstructured Social Activity Attribute dataset show that our model achieves state of the art performance and outperforms other supervised topic model in terms of classification accuracy, particularly in the case of a very small number of labeled training videos. 1
same-paper 4 0.82940727 266 nips-2013-Recurrent linear models of simultaneously-recorded neural populations
Author: Marius Pachitariu, Biljana Petreska, Maneesh Sahani
Abstract: Population neural recordings with long-range temporal structure are often best understood in terms of a common underlying low-dimensional dynamical process. Advances in recording technology provide access to an ever-larger fraction of the population, but the standard computational approaches available to identify the collective dynamics scale poorly with the size of the dataset. We describe a new, scalable approach to discovering low-dimensional dynamics that underlie simultaneously recorded spike trains from a neural population. We formulate the Recurrent Linear Model (RLM) by generalising the Kalman-filter-based likelihood calculation for latent linear dynamical systems to incorporate a generalised-linear observation process. We show that RLMs describe motor-cortical population data better than either directly-coupled generalised-linear models or latent linear dynamical system models with generalised-linear observations. We also introduce the cascaded generalised-linear model (CGLM) to capture low-dimensional instantaneous correlations in neural populations. The CGLM describes the cortical recordings better than either Ising or Gaussian models and, like the RLM, can be fit exactly and quickly. The CGLM can also be seen as a generalisation of a lowrank Gaussian model, in this case factor analysis. The computational tractability of the RLM and CGLM allow both to scale to very high-dimensional neural data. 1
5 0.77939051 131 nips-2013-Geometric optimisation on positive definite matrices for elliptically contoured distributions
Author: Suvrit Sra, Reshad Hosseini
Abstract: Hermitian positive definite (hpd) matrices recur throughout machine learning, statistics, and optimisation. This paper develops (conic) geometric optimisation on the cone of hpd matrices, which allows us to globally optimise a large class of nonconvex functions of hpd matrices. Specifically, we first use the Riemannian manifold structure of the hpd cone for studying functions that are nonconvex in the Euclidean sense but are geodesically convex (g-convex), hence globally optimisable. We then go beyond g-convexity, and exploit the conic geometry of hpd matrices to identify another class of functions that remain amenable to global optimisation without requiring g-convexity. We present key results that help recognise g-convexity and also the additional structure alluded to above. We illustrate our ideas by applying them to likelihood maximisation for a broad family of elliptically contoured distributions: for this maximisation, we derive novel, parameter free fixed-point algorithms. To our knowledge, ours are the most general results on geometric optimisation of hpd matrices known so far. Experiments show that advantages of using our fixed-point algorithms. 1
6 0.7610324 70 nips-2013-Contrastive Learning Using Spectral Methods
7 0.73692948 221 nips-2013-On the Expressive Power of Restricted Boltzmann Machines
8 0.67801738 303 nips-2013-Sparse Overlapping Sets Lasso for Multitask Learning and its Application to fMRI Analysis
9 0.66080689 345 nips-2013-Variance Reduction for Stochastic Gradient Optimization
10 0.64107579 121 nips-2013-Firing rate predictions in optimal balanced networks
11 0.58746606 262 nips-2013-Real-Time Inference for a Gamma Process Model of Neural Spiking
12 0.54529023 141 nips-2013-Inferring neural population dynamics from multiple partial recordings of the same neural circuit
13 0.54379308 246 nips-2013-Perfect Associative Learning with Spike-Timing-Dependent Plasticity
14 0.53260154 64 nips-2013-Compete to Compute
15 0.49990547 236 nips-2013-Optimal Neural Population Codes for High-dimensional Stimulus Variables
16 0.49790543 148 nips-2013-Latent Maximum Margin Clustering
17 0.49630278 208 nips-2013-Neural representation of action sequences: how far can a simple snippet-matching model take us?
19 0.49071664 237 nips-2013-Optimal integration of visual speed across different spatiotemporal frequency channels
20 0.48990458 301 nips-2013-Sparse Additive Text Models with Low Rank Background