nips nips2007 nips2007-103 knowledge-graph by maker-knowledge-mining
Source: pdf
Author: Misha Ahrens, Maneesh Sahani
Abstract: Many perceptual processes and neural computations, such as speech recognition, motor control and learning, depend on the ability to measure and mark the passage of time. However, the processes that make such temporal judgements possible are unknown. A number of different hypothetical mechanisms have been advanced, all of which depend on the known, temporally predictable evolution of a neural or psychological state, possibly through oscillations or the gradual decay of a memory trace. Alternatively, judgements of elapsed time might be based on observations of temporally structured, but stochastic processes. Such processes need not be specific to the sense of time; typical neural and sensory processes contain at least some statistical structure across a range of time scales. Here, we investigate the statistical properties of an estimator of elapsed time which is based on a simple family of stochastic process. 1
Reference: text
sentIndex sentText sentNum sentScore
1 uk Abstract Many perceptual processes and neural computations, such as speech recognition, motor control and learning, depend on the ability to measure and mark the passage of time. [sent-5, score-0.423]
2 However, the processes that make such temporal judgements possible are unknown. [sent-6, score-0.462]
3 A number of different hypothetical mechanisms have been advanced, all of which depend on the known, temporally predictable evolution of a neural or psychological state, possibly through oscillations or the gradual decay of a memory trace. [sent-7, score-0.328]
4 Alternatively, judgements of elapsed time might be based on observations of temporally structured, but stochastic processes. [sent-8, score-0.452]
5 Such processes need not be specific to the sense of time; typical neural and sensory processes contain at least some statistical structure across a range of time scales. [sent-9, score-0.888]
6 Here, we investigate the statistical properties of an estimator of elapsed time which is based on a simple family of stochastic process. [sent-10, score-0.402]
7 1 Introduction The experience of the passage of time, as well as the timing of events and intervals, has long been of interest in psychology, and has more recently attracted attention in neuroscience as well. [sent-11, score-0.514]
8 Timing information is crucial for the correct functioning of a large number of processes, such as accurate limb movement, speech and the perception of speech (for example, the difference between “ba” and “pa” lies only in the relative timing of voice onsets), and causal learning. [sent-12, score-0.473]
9 Neuroscientific evidence that points to a specialized neural substrate for timing is very sparse, particularly when compared to the divergent set of specific mechanisms which have been theorized. [sent-13, score-0.689]
10 One of the most influential proposals, the scalar expectancy theory (SET) of timing [1], suggests that interval timing is based on the accumulation of activity from an internal oscillatory process. [sent-14, score-1.271]
11 An alternative, which we explore here, is to phrase time estimation as a statistical problem, in which the elapsed time ∆t is extracted from a collection of stochastic processes whose statistics are known. [sent-18, score-0.727]
12 Such models have 1 been related to recent psychological findings the show that the nature of the stimulus being timed affects judgments of duration [7]. [sent-20, score-0.198]
13 Here, by contrast, we consider the properties of duration estimators that are based on more general stochastic processes. [sent-21, score-0.221]
14 However, they may be seen as models both for internally-generated neural processes, such as (spontaneous) network activity and local field potentials, and for sensory processes, in the form of externally-driven neural activity, or (taking a functional view) in the form of the stimuli themselves. [sent-23, score-0.386]
15 Both neural activity and sensory input from the environment follow well-defined temporal statistical patterns, but the exploitation of these statistics has thus far not been studied as a potential substrate for timing judgements, despite being potentially attractive. [sent-24, score-0.94]
16 Such a basis for timing is consistent with recent studies that show that the statistics of external stimuli affect timing estimates [8, 7], a behavior not captured by the existing mechanistic models. [sent-25, score-1.1]
17 In addition, there is evidence that timing mechanisms are distributed [9] but subject to local (e. [sent-26, score-0.546]
18 Using the distributed time-varying processes which are already present in the brain is implementationally efficient, and lends itself straightforwardly to a distributed implementation. [sent-29, score-0.352]
19 At the same time, it suggests a possible origin for the modality-specificity and locality of the bias effects, as different sets of processes may be exploited for different timing purposes. [sent-30, score-0.837]
20 Likewise, sensory information too varies over a large range of temporal scales [13]. [sent-33, score-0.289]
21 The particular stochastic processes we use here are Gaussian Processes, whose power spectra are chosen to be broad and roughly similar to those seen in natural stimuli. [sent-34, score-0.559]
22 2 The framework To illustrate how random processes contain timing information, consider a random walk starting at the origin, and suppose that we see a snapshot of the random walk at another, unknown, point in time. [sent-35, score-0.95]
23 If the walk were to end up very far from the origin, and if some statistics of the random walk were known, we would expect that the time difference between the two observations, ∆t, must be reasonably long in comparison to the diffusion time of the process. [sent-36, score-0.291]
24 To formalize these ideas, we model the random processes as a family of independent stationary Gaussian Processes (GPs, [14]). [sent-42, score-0.311]
25 Although this is not a necessity, we let each process evolve independently according to the same stochastic dynamics; thus the process values differ only due to the random effects. [sent-48, score-0.176]
26 Mimicking the temporal statistics of natural scenes [15], we choose the dynamics to simultaneously contain multiple time scales — specifically, the power spectrum approximately follows a 1/f 2 power law, were f = frequency = 1/(time scale). [sent-49, score-0.888]
27 Some instances of such processes are shown in Figure 1. [sent-50, score-0.311]
28 Stationary Gaussian processes are fully described by the covariance function K(∆t): yi (t)yi (t + ∆t) = K(∆t) so that the probability of observing a sequence of values [yi (t1 ), yi (t2 ), . [sent-51, score-0.802]
29 , yi (tn )] is Gaussian distributed, with zero mean and covariance matrix Σn,n′ = K(tn′ − tn ). [sent-54, score-0.369]
30 2 y log power 0 −5 −10 −4 time −2 0 log frequency 2 4 Figure 1: Left: Two examples of the GPs used for inference of ∆t. [sent-55, score-0.222]
31 This is approximately a 1/f 2 spectrum, similar to the temporal power spectrum of visual scenes. [sent-57, score-0.491]
32 To 2 mimic a 1/f 2 spectrum, we choose the power of each component to be constant: αq = 1/Q. [sent-60, score-0.164]
33 Figure 2 1 shows that this choice does indeed quite accurately reproduce a 1/f power spectrum. [sent-61, score-0.164]
34 To illustrate how elapsed time is implicitly encoded in such stochastic processes, we infer the duration of an interval [t, t + ∆t] from two instantaneous observations of the processes, namely {yi (t)} and {yi (t+∆t)}. [sent-62, score-0.504]
35 For convenience, yi is used to denote the vector [yi (t), yi (t+∆t)]. [sent-63, score-0.438]
36 As we will see below (see Figure 2), this distribution tends to be centred on the true value of ∆t, showing that such random processes may indeed be exploited to obtain timing information. [sent-65, score-0.784]
37 In the following section, we explore the statistical properties of timing estimates based on Φ, and show that they correspond to several experimental findings. [sent-66, score-0.534]
38 Right: The Weber law of timing, σ ∝ ∆t, approximately holds true for this model. [sent-72, score-0.213]
39 1 Scaling laws and behaviour Empirical demonstration of Weber’s law Many behavioral studies have shown that the standard deviation of interval estimates is proportional to the interval being judged, σ ∝ ∆t, across a wide range of timescales and tasks (e. [sent-77, score-0.58]
40 To compare the behaviour of the model to experimental data, we must choose a mapping from the function Φ to a single scalar value, which will model the observer’s report. [sent-81, score-0.16]
41 To compare the statistics of this estimator to the experimental observation, we took samples {yi (t)} and {yi (t + ∆t)} from 50 GPs with identical 1/f 2 -like statistics containing time scales from 1 to 40 time units. [sent-83, score-0.386]
42 Thus, time estimation is possible using the stochastic process framework, and the Weber law of timing holds fairly accurately. [sent-89, score-0.842]
43 2 Fisher Information and Weber’s law A number of questions about this Weber-like result naturally arise: Does it still hold if one changes the power spectrum of the processes? [sent-91, score-0.512]
44 We 2 increased the noise scale σy , and found that the Weber law was still approximately satisfied. [sent-93, score-0.293]
45 When changing the power spectrum of the processes from a 1/f 2 -type spectrum to a 1/f 3 -type spectrum 2 2 (by letting αi ∝ li instead of αi ∝ 1), the Weber law was still approximately satisfied (Figure 3). [sent-94, score-1.284]
46 To find reasons for this behaviour, it would useful to have an analytical expression for the relationship between the variability of the estimated duration and the true duration. [sent-96, score-0.204]
47 This is a lower bound on the asymptotic variance of an unbiased Maximum Likelihood estimator of ∆t and is given by the inverse Fisher Information: 4 y standard deviation 2. [sent-98, score-0.169]
48 5 0 0 time 5 ∆t 10 15 2 Figure 3: Left: Two examples of GPs with a different power spectrum (αi ∝ li , for li ∝ i, which 3 approximates a 1/f power spectrum, resulting in much smoother dynamics). [sent-101, score-0.743]
49 The Weber law still approximately holds, even though the dynamics is different from the initial case. [sent-106, score-0.253]
50 (The power spectrum 2 is then shaped as exp(−f 2 li /2). [sent-115, score-0.426]
51 ) The likelihood of observing the processes at two instances is now P ({yi (t)}, {yi (t + ∆t)}|∆t) ∝ i 1 T |Ci |−1/2 exp − yi C−1 yi i 2 (2) This model shows very similar behaviour to the original model, but is somewhat less natural. [sent-116, score-0.918]
52 In particular, we tested both linear spacing of time scales (li ∝ i) and quadratic spacing (li ∝ i2 ), and we tested a constant power distribution 5 power ~ time scale lengthscales spaced linearly power ~ time scale lengthscales spaced quadratically l=0. [sent-118, score-1.299]
53 −Rao bound 0 10 ∆t 20 30 0 power = constant lengthscales spaced linearly 20 30 l=0. [sent-129, score-0.365]
54 −Rao bound F F I and (I )−1/2 −1/2 ∆t power = constant lengthscales spaced quadratically l=6 l=11,. [sent-136, score-0.399]
55 1, and the time scales are 2 either li = i, i = 1, 2, . [sent-143, score-0.25]
56 The power of each 2 2 process is either αi = 1 (constant) or αi = li . [sent-150, score-0.305]
57 The graphs show that each time scale contributes to the estimation of a wide range of ∆t, and that the Cramer-Rao bounds are all fairly linear, leading to a robust Weber-like behaviour of the estimator of elapsed time. [sent-151, score-0.544]
58 2 (αi = 1) and a power distribution where slower processes have more power (αi ∝ li ). [sent-152, score-0.734]
59 Next, we can evaluate the contribution of each time scale to the accuracy of estimates of ∆t, by inspecting the Fisher Information IF,i of a given process yi . [sent-154, score-0.431]
60 This lies at the heart of the robust Weber-like behaviour: the details of the distribution of time scales do not matter much, because each time scale contributes to the estimation of a wide range of ∆t. [sent-156, score-0.311]
61 For similar reasons, the distribution of power does not drastically affect the Cramer-Rao bound. [sent-157, score-0.164]
62 From the graphs of IF,i , it is evident that the Weber law arises from an accumulation of high values of Fisher Information at low values of ∆t. [sent-158, score-0.257]
63 Very small values of ∆t may be an exception, if the instantaneous noise dominates the subtle changes that the processes undergo during very short periods; for these ∆t, the standard deviation may rise. [sent-159, score-0.475]
64 However, it may be assumed that the shortest times that neural systems can evaluate are no shorter than the scale of the fastest process within the system, making these small ∆t’s irrelevant. [sent-161, score-0.164]
65 3 Dependence of timing variability on the number of processes Increasing the number of processes, say Nprocesses , will add more terms to the likelihood and make the estimated ∆t more accurate. [sent-163, score-0.826]
66 The Fisher Information (equation 1) scales with Nprocesses , which suggests that the standard deviation of ∆tMAP is proportional to 1/ Nprocesses ; this was confirmed empirically (data not shown). [sent-164, score-0.144]
67 6 Psychologically and neurally, increasing the number of processes would correspond to adding more perceptual processes, or expanding the size of the network that is being monitored for timing estimation. [sent-165, score-0.836]
68 Although experimental data on this issue is sparse, in [9], it is shown that unimanual rhythm tapping results in a higher variability of tapping times than bimanual rhythm tapping, and that tapping with two hands and a foot results in even lower variability. [sent-166, score-0.436]
69 This correlates well with the theoretical scaling behaviour of the estimator ∆tMAP . [sent-167, score-0.266]
70 Note that a similar scaling law is obtained from the Multiple Timer Model [16]. [sent-168, score-0.232]
71 √ Experimentally, a slower decrease in variability than a 1/ N law was observed. [sent-170, score-0.223]
72 This can be accounted for by assuming that the processes governing the right and left hands are dependent, so that the number of effectively independent processes grows more slowly than the number of effectors. [sent-171, score-0.656]
73 4 Conclusion We have shown that timing information is present in random processes, and can be extracted probabilistically if certain statistics of the processes are known. [sent-172, score-0.827]
74 A neural implementation of such a framework of time estimation could use both internally generated population activity as well as external stimuli to drive its processes. [sent-173, score-0.268]
75 The timing estimators considered were based on the full probability distribution of the process values at times t and t′ , but simpler estimators could also be constructed. [sent-174, score-0.669]
76 There are two reasons for considering simpler estimators: First, simpler estimators might be more easily implemented in neural systems. [sent-175, score-0.283]
77 Second, to calculate Φ(∆t), one needs all of {yi (t), yi (t′ )}, so that (at least) {yi (t)} has to be stored in memory. [sent-176, score-0.219]
78 One way to construct a simpler estimator might be to select a particular class (say, a linear function of {yi }) and optimize over its parameters. [sent-177, score-0.162]
79 Brownian motion, being translationally invariant, would require only {yi (t′ ) − yi (t)} instead of {yi (t), yi (t′ )}; we have not considered such processes because they are unbounded and therefore hard to associate with sensory or neural processes). [sent-181, score-0.923]
80 We have not addressed how a memory mechanism might be combined with the stochastic process framework; this will be explored in the future. [sent-182, score-0.21]
81 The intention of this paper is not to offer a complete theory of neural and psychological timing, but to examine the statistical properties of a hitherto neglected substrate for timing — stochastic processes that take place in the brain or in the sensory world. [sent-183, score-1.261]
82 It was demonstrated that estimators based on such processes replicate several important behaviors of humans and animals. [sent-184, score-0.369]
83 Full models might be based on the same substrate, thereby naturally incorporating the same behaviors, but contain more completely specified relations to external input, memory mechanisms, adaptive mechanisms, neural implementation, and importantly, (supervised) learning of the estimator. [sent-185, score-0.235]
84 The neural and sensory processes that we assume to form the basis of time estimation are, of course, not fully random. [sent-186, score-0.543]
85 But when the deterministic structure behind a process is unknown, they can still be treated as stochastic under certain statistical rules, and thus lead to a valid timing estimator. [sent-187, score-0.637]
86 Would the GP likelihood still apply to real neural processes or would the correct likelihood be completely different? [sent-188, score-0.382]
87 stochastic processes tend to Gaussian Processes — so that, when e. [sent-192, score-0.395]
88 monitoring average neuronal activity, the correct estimator may well be based on a GP likelihood. [sent-194, score-0.17]
89 there is evidence for a spatial 7 code for temporal frequency in V2 [17]), so that neural and stimulus fluctuations cannot always be treated on the same footing. [sent-199, score-0.209]
90 The framework presented here has some similarities with the very interesting and more explicitly physiological model proposed by Buonomano and colleagues [5, 18], in which time is implicitly encoded in deterministic2 neural networks through slow neuronal time constants. [sent-201, score-0.27]
91 However, temporal information in the network model is lost when there are stimulus-independent fluctuations in the network activity, and the network can only be used as a reliable timer when it starts from a fixed resting state, and if the stimulus is identical on every trial. [sent-202, score-0.384]
92 The difference in our scheme is that here timing estimates are based on statistics, rather than deterministic structure, so that it is fundamentally robust to noise, internal fluctuations, and stimulus changes. [sent-203, score-0.665]
93 The stochastic process framework is, however, more abstract and farther removed from physiology, and a neural implementation may well share some features of the network model of timing. [sent-204, score-0.253]
94 Scalar expectancy theory and Weber’s law in animal timing. [sent-207, score-0.233]
95 Judging the duration of time intervals: A process of remembering segments of experience. [sent-222, score-0.183]
96 Neural mechanisms for timing visual events are spatially selective in real-world coordinates. [sent-235, score-0.585]
97 Fractal characted of the neural spike train in the visual system of the cat. [sent-238, score-0.153]
98 Spatial and temporal frequency selectivity of neurones in visual cortical areas v1 and v2 of the macaque monkey. [sent-256, score-0.161]
99 Timing in the absence of clocks: encoding time in neural network states. [sent-259, score-0.181]
100 2 While this model and some other previous models might also contain neuronal noise, it is the deterministic (and known) element of their behaviour which encodes time. [sent-261, score-0.32]
wordName wordTfidf (topN-words)
[('timing', 0.473), ('processes', 0.311), ('yi', 0.219), ('weber', 0.203), ('law', 0.181), ('elapsed', 0.173), ('spectrum', 0.167), ('power', 0.164), ('fisher', 0.149), ('tmap', 0.129), ('behaviour', 0.128), ('sensory', 0.103), ('tn', 0.097), ('scales', 0.097), ('uctuations', 0.095), ('li', 0.095), ('tapping', 0.09), ('lengthscales', 0.09), ('timer', 0.09), ('temporal', 0.089), ('activity', 0.089), ('estimator', 0.087), ('gps', 0.086), ('stochastic', 0.084), ('neuronal', 0.083), ('duration', 0.079), ('buonomano', 0.078), ('nprocesses', 0.078), ('spaced', 0.076), ('mechanisms', 0.073), ('substrate', 0.072), ('neural', 0.071), ('psychological', 0.07), ('walk', 0.066), ('rao', 0.063), ('judgements', 0.062), ('estimates', 0.061), ('interval', 0.059), ('estimators', 0.058), ('time', 0.058), ('origin', 0.053), ('covariance', 0.053), ('network', 0.052), ('ahrens', 0.052), ('expectancy', 0.052), ('ivry', 0.052), ('evolving', 0.052), ('lj', 0.052), ('contributes', 0.051), ('instantaneous', 0.051), ('scaling', 0.051), ('external', 0.05), ('neurosci', 0.05), ('ci', 0.049), ('stimulus', 0.049), ('internal', 0.048), ('scale', 0.047), ('deviation', 0.047), ('process', 0.046), ('rhythm', 0.045), ('accumulation', 0.045), ('timescales', 0.045), ('reasons', 0.045), ('spike', 0.043), ('gp', 0.043), ('statistics', 0.043), ('variability', 0.042), ('lq', 0.041), ('oscillations', 0.041), ('psychol', 0.041), ('passage', 0.041), ('might', 0.041), ('exp', 0.041), ('brain', 0.041), ('dynamics', 0.04), ('visual', 0.039), ('memory', 0.039), ('spacing', 0.038), ('snapshots', 0.038), ('maneesh', 0.038), ('analytical', 0.038), ('neglected', 0.036), ('proposals', 0.036), ('blue', 0.035), ('bound', 0.035), ('hands', 0.034), ('quadratically', 0.034), ('temporally', 0.034), ('motion', 0.034), ('intervals', 0.034), ('deterministic', 0.034), ('simpler', 0.034), ('contain', 0.034), ('subtle', 0.033), ('macaque', 0.033), ('noise', 0.033), ('approximately', 0.032), ('scalar', 0.032), ('theories', 0.031), ('evident', 0.031)]
simIndex simValue paperId paperTitle
same-paper 1 1.0000002 103 nips-2007-Inferring Elapsed Time from Stochastic Neural Processes
Author: Misha Ahrens, Maneesh Sahani
Abstract: Many perceptual processes and neural computations, such as speech recognition, motor control and learning, depend on the ability to measure and mark the passage of time. However, the processes that make such temporal judgements possible are unknown. A number of different hypothetical mechanisms have been advanced, all of which depend on the known, temporally predictable evolution of a neural or psychological state, possibly through oscillations or the gradual decay of a memory trace. Alternatively, judgements of elapsed time might be based on observations of temporally structured, but stochastic processes. Such processes need not be specific to the sense of time; typical neural and sensory processes contain at least some statistical structure across a range of time scales. Here, we investigate the statistical properties of an estimator of elapsed time which is based on a simple family of stochastic process. 1
2 0.16334063 17 nips-2007-A neural network implementing optimal state estimation based on dynamic spike train decoding
Author: Omer Bobrowski, Ron Meir, Shy Shoham, Yonina Eldar
Abstract: It is becoming increasingly evident that organisms acting in uncertain dynamical environments often employ exact or approximate Bayesian statistical calculations in order to continuously estimate the environmental state, integrate information from multiple sensory modalities, form predictions and choose actions. What is less clear is how these putative computations are implemented by cortical neural networks. An additional level of complexity is introduced because these networks observe the world through spike trains received from primary sensory afferents, rather than directly. A recent line of research has described mechanisms by which such computations can be implemented using a network of neurons whose activity directly represents a probability distribution across the possible “world states”. Much of this work, however, uses various approximations, which severely restrict the domain of applicability of these implementations. Here we make use of rigorous mathematical results from the theory of continuous time point process filtering, and show how optimal real-time state estimation and prediction may be implemented in a general setting using linear neural networks. We demonstrate the applicability of the approach with several examples, and relate the required network properties to the statistical nature of the environment, thereby quantifying the compatibility of a given network with its environment. 1
3 0.16073851 133 nips-2007-Modelling motion primitives and their timing in biologically executed movements
Author: Ben Williams, Marc Toussaint, Amos J. Storkey
Abstract: Biological movement is built up of sub-blocks or motion primitives. Such primitives provide a compact representation of movement which is also desirable in robotic control applications. We analyse handwriting data to gain a better understanding of primitives and their timings in biological movements. Inference of the shape and the timing of primitives can be done using a factorial HMM based model, allowing the handwriting to be represented in primitive timing space. This representation provides a distribution of spikes corresponding to the primitive activations, which can also be modelled using HMM architectures. We show how the coupling of the low level primitive model, and the higher level timing model during inference can produce good reconstructions of handwriting, with shared primitives for all characters modelled. This coupled model also captures the variance profile of the dataset which is accounted for by spike timing jitter. The timing code provides a compact representation of the movement while generating a movement without an explicit timing model produces a scribbling style of output. 1
4 0.11491485 104 nips-2007-Inferring Neural Firing Rates from Spike Trains Using Gaussian Processes
Author: Maneesh Sahani, Byron M. Yu, John P. Cunningham, Krishna V. Shenoy
Abstract: Neural spike trains present challenges to analytical efforts due to their noisy, spiking nature. Many studies of neuroscientific and neural prosthetic importance rely on a smoothed, denoised estimate of the spike train’s underlying firing rate. Current techniques to find time-varying firing rates require ad hoc choices of parameters, offer no confidence intervals on their estimates, and can obscure potentially important single trial variability. We present a new method, based on a Gaussian Process prior, for inferring probabilistically optimal estimates of firing rate functions underlying single or multiple neural spike trains. We test the performance of the method on simulated data and experimentally gathered neural spike trains, and we demonstrate improvements over conventional estimators. 1
5 0.11166693 59 nips-2007-Continuous Time Particle Filtering for fMRI
Author: Lawrence Murray, Amos J. Storkey
Abstract: We construct a biologically motivated stochastic differential model of the neural and hemodynamic activity underlying the observed Blood Oxygen Level Dependent (BOLD) signal in Functional Magnetic Resonance Imaging (fMRI). The model poses a difficult parameter estimation problem, both theoretically due to the nonlinearity and divergence of the differential system, and computationally due to its time and space complexity. We adapt a particle filter and smoother to the task, and discuss some of the practical approaches used to tackle the difficulties, including use of sparse matrices and parallelisation. Results demonstrate the tractability of the approach in its application to an effective connectivity study. 1
6 0.098857462 164 nips-2007-Receptive Fields without Spike-Triggering
7 0.09359847 173 nips-2007-Second Order Bilinear Discriminant Analysis for single trial EEG analysis
9 0.090051576 192 nips-2007-Testing for Homogeneity with Kernel Fisher Discriminant Analysis
10 0.088802449 32 nips-2007-Bayesian Co-Training
11 0.087524325 3 nips-2007-A Bayesian Model of Conditioned Perception
12 0.082024358 33 nips-2007-Bayesian Inference for Spiking Neuron Models with a Sparsity Prior
13 0.081192903 140 nips-2007-Neural characterization in partially observed populations of spiking neurons
14 0.078087844 170 nips-2007-Robust Regression with Twinned Gaussian Processes
15 0.07495138 186 nips-2007-Statistical Analysis of Semi-Supervised Regression
16 0.0728423 94 nips-2007-Gaussian Process Models for Link Analysis and Transfer Learning
17 0.072179511 124 nips-2007-Managing Power Consumption and Performance of Computing Systems Using Reinforcement Learning
18 0.071569957 97 nips-2007-Hidden Common Cause Relations in Relational Learning
19 0.071216337 127 nips-2007-Measuring Neural Synchrony by Message Passing
20 0.068810321 135 nips-2007-Multi-task Gaussian Process Prediction
topicId topicWeight
[(0, -0.229), (1, 0.079), (2, 0.107), (3, 0.051), (4, -0.025), (5, 0.059), (6, -0.028), (7, -0.05), (8, -0.097), (9, -0.095), (10, -0.074), (11, -0.023), (12, -0.033), (13, 0.03), (14, 0.112), (15, -0.007), (16, 0.007), (17, 0.023), (18, 0.104), (19, 0.059), (20, 0.047), (21, 0.0), (22, 0.102), (23, -0.089), (24, 0.015), (25, -0.112), (26, 0.022), (27, -0.017), (28, 0.055), (29, -0.102), (30, 0.199), (31, 0.031), (32, 0.11), (33, 0.07), (34, -0.074), (35, -0.041), (36, -0.058), (37, -0.173), (38, 0.195), (39, -0.109), (40, 0.128), (41, 0.006), (42, -0.084), (43, 0.019), (44, -0.046), (45, -0.052), (46, -0.007), (47, 0.015), (48, 0.063), (49, -0.088)]
simIndex simValue paperId paperTitle
same-paper 1 0.97091013 103 nips-2007-Inferring Elapsed Time from Stochastic Neural Processes
Author: Misha Ahrens, Maneesh Sahani
Abstract: Many perceptual processes and neural computations, such as speech recognition, motor control and learning, depend on the ability to measure and mark the passage of time. However, the processes that make such temporal judgements possible are unknown. A number of different hypothetical mechanisms have been advanced, all of which depend on the known, temporally predictable evolution of a neural or psychological state, possibly through oscillations or the gradual decay of a memory trace. Alternatively, judgements of elapsed time might be based on observations of temporally structured, but stochastic processes. Such processes need not be specific to the sense of time; typical neural and sensory processes contain at least some statistical structure across a range of time scales. Here, we investigate the statistical properties of an estimator of elapsed time which is based on a simple family of stochastic process. 1
2 0.65973431 133 nips-2007-Modelling motion primitives and their timing in biologically executed movements
Author: Ben Williams, Marc Toussaint, Amos J. Storkey
Abstract: Biological movement is built up of sub-blocks or motion primitives. Such primitives provide a compact representation of movement which is also desirable in robotic control applications. We analyse handwriting data to gain a better understanding of primitives and their timings in biological movements. Inference of the shape and the timing of primitives can be done using a factorial HMM based model, allowing the handwriting to be represented in primitive timing space. This representation provides a distribution of spikes corresponding to the primitive activations, which can also be modelled using HMM architectures. We show how the coupling of the low level primitive model, and the higher level timing model during inference can produce good reconstructions of handwriting, with shared primitives for all characters modelled. This coupled model also captures the variance profile of the dataset which is accounted for by spike timing jitter. The timing code provides a compact representation of the movement while generating a movement without an explicit timing model produces a scribbling style of output. 1
3 0.60034484 59 nips-2007-Continuous Time Particle Filtering for fMRI
Author: Lawrence Murray, Amos J. Storkey
Abstract: We construct a biologically motivated stochastic differential model of the neural and hemodynamic activity underlying the observed Blood Oxygen Level Dependent (BOLD) signal in Functional Magnetic Resonance Imaging (fMRI). The model poses a difficult parameter estimation problem, both theoretically due to the nonlinearity and divergence of the differential system, and computationally due to its time and space complexity. We adapt a particle filter and smoother to the task, and discuss some of the practical approaches used to tackle the difficulties, including use of sparse matrices and parallelisation. Results demonstrate the tractability of the approach in its application to an effective connectivity study. 1
4 0.56394321 17 nips-2007-A neural network implementing optimal state estimation based on dynamic spike train decoding
Author: Omer Bobrowski, Ron Meir, Shy Shoham, Yonina Eldar
Abstract: It is becoming increasingly evident that organisms acting in uncertain dynamical environments often employ exact or approximate Bayesian statistical calculations in order to continuously estimate the environmental state, integrate information from multiple sensory modalities, form predictions and choose actions. What is less clear is how these putative computations are implemented by cortical neural networks. An additional level of complexity is introduced because these networks observe the world through spike trains received from primary sensory afferents, rather than directly. A recent line of research has described mechanisms by which such computations can be implemented using a network of neurons whose activity directly represents a probability distribution across the possible “world states”. Much of this work, however, uses various approximations, which severely restrict the domain of applicability of these implementations. Here we make use of rigorous mathematical results from the theory of continuous time point process filtering, and show how optimal real-time state estimation and prediction may be implemented in a general setting using linear neural networks. We demonstrate the applicability of the approach with several examples, and relate the required network properties to the statistical nature of the environment, thereby quantifying the compatibility of a given network with its environment. 1
5 0.470853 3 nips-2007-A Bayesian Model of Conditioned Perception
Author: Alan Stocker, Eero P. Simoncelli
Abstract: unkown-abstract
6 0.43918711 104 nips-2007-Inferring Neural Firing Rates from Spike Trains Using Gaussian Processes
7 0.43864012 25 nips-2007-An in-silico Neural Model of Dynamic Routing through Neuronal Coherence
8 0.42092985 28 nips-2007-Augmented Functional Time Series Representation and Forecasting with Gaussian Processes
9 0.4146384 187 nips-2007-Structured Learning with Approximate Inference
10 0.4140397 192 nips-2007-Testing for Homogeneity with Kernel Fisher Discriminant Analysis
11 0.39764002 138 nips-2007-Near-Maximum Entropy Models for Binary Neural Representations of Natural Images
12 0.39551085 88 nips-2007-Fast and Scalable Training of Semi-Supervised CRFs with Application to Activity Recognition
14 0.38276005 24 nips-2007-An Analysis of Inference with the Universum
15 0.381385 124 nips-2007-Managing Power Consumption and Performance of Computing Systems Using Reinforcement Learning
16 0.37328207 210 nips-2007-Unconstrained On-line Handwriting Recognition with Recurrent Neural Networks
17 0.37214607 97 nips-2007-Hidden Common Cause Relations in Relational Learning
18 0.35738489 186 nips-2007-Statistical Analysis of Semi-Supervised Regression
19 0.35108984 173 nips-2007-Second Order Bilinear Discriminant Analysis for single trial EEG analysis
20 0.34968752 214 nips-2007-Variational inference for Markov jump processes
topicId topicWeight
[(5, 0.026), (13, 0.021), (16, 0.034), (18, 0.026), (19, 0.018), (21, 0.536), (31, 0.012), (34, 0.025), (35, 0.021), (47, 0.084), (83, 0.069), (87, 0.018), (90, 0.042)]
simIndex simValue paperId paperTitle
same-paper 1 0.96034473 103 nips-2007-Inferring Elapsed Time from Stochastic Neural Processes
Author: Misha Ahrens, Maneesh Sahani
Abstract: Many perceptual processes and neural computations, such as speech recognition, motor control and learning, depend on the ability to measure and mark the passage of time. However, the processes that make such temporal judgements possible are unknown. A number of different hypothetical mechanisms have been advanced, all of which depend on the known, temporally predictable evolution of a neural or psychological state, possibly through oscillations or the gradual decay of a memory trace. Alternatively, judgements of elapsed time might be based on observations of temporally structured, but stochastic processes. Such processes need not be specific to the sense of time; typical neural and sensory processes contain at least some statistical structure across a range of time scales. Here, we investigate the statistical properties of an estimator of elapsed time which is based on a simple family of stochastic process. 1
2 0.9468658 32 nips-2007-Bayesian Co-Training
Author: Shipeng Yu, Balaji Krishnapuram, Harald Steck, R. B. Rao, Rómer Rosales
Abstract: We propose a Bayesian undirected graphical model for co-training, or more generally for semi-supervised multi-view learning. This makes explicit the previously unstated assumptions of a large class of co-training type algorithms, and also clarifies the circumstances under which these assumptions fail. Building upon new insights from this model, we propose an improved method for co-training, which is a novel co-training kernel for Gaussian process classifiers. The resulting approach is convex and avoids local-maxima problems, unlike some previous multi-view learning methods. Furthermore, it can automatically estimate how much each view should be trusted, and thus accommodate noisy or unreliable views. Experiments on toy data and real world data sets illustrate the benefits of this approach. 1
3 0.93673992 193 nips-2007-The Distribution Family of Similarity Distances
Author: Gertjan Burghouts, Arnold Smeulders, Jan-mark Geusebroek
Abstract: Assessing similarity between features is a key step in object recognition and scene categorization tasks. We argue that knowledge on the distribution of distances generated by similarity functions is crucial in deciding whether features are similar or not. Intuitively one would expect that similarities between features could arise from any distribution. In this paper, we will derive the contrary, and report the theoretical result that Lp -norms –a class of commonly applied distance metrics– from one feature vector to other vectors are Weibull-distributed if the feature values are correlated and non-identically distributed. Besides these assumptions being realistic for images, we experimentally show them to hold for various popular feature extraction algorithms, for a diverse range of images. This fundamental insight opens new directions in the assessment of feature similarity, with projected improvements in object and scene recognition algorithms. 1
4 0.90261489 19 nips-2007-Active Preference Learning with Discrete Choice Data
Author: Brochu Eric, Nando D. Freitas, Abhijeet Ghosh
Abstract: We propose an active learning algorithm that learns a continuous valuation model from discrete preferences. The algorithm automatically decides what items are best presented to an individual in order to find the item that they value highly in as few trials as possible, and exploits quirks of human psychology to minimize time and cognitive burden. To do this, our algorithm maximizes the expected improvement at each query without accurately modelling the entire valuation surface, which would be needlessly expensive. The problem is particularly difficult because the space of choices is infinite. We demonstrate the effectiveness of the new algorithm compared to related active learning methods. We also embed the algorithm within a decision making tool for assisting digital artists in rendering materials. The tool finds the best parameters while minimizing the number of queries. 1
5 0.61505312 172 nips-2007-Scene Segmentation with CRFs Learned from Partially Labeled Images
Author: Bill Triggs, Jakob J. Verbeek
Abstract: Conditional Random Fields (CRFs) are an effective tool for a variety of different data segmentation and labeling tasks including visual scene interpretation, which seeks to partition images into their constituent semantic-level regions and assign appropriate class labels to each region. For accurate labeling it is important to capture the global context of the image as well as local information. We introduce a CRF based scene labeling model that incorporates both local features and features aggregated over the whole image or large sections of it. Secondly, traditional CRF learning requires fully labeled datasets which can be costly and troublesome to produce. We introduce a method for learning CRFs from datasets with many unlabeled nodes by marginalizing out the unknown labels so that the log-likelihood of the known ones can be maximized by gradient ascent. Loopy Belief Propagation is used to approximate the marginals needed for the gradient and log-likelihood calculations and the Bethe free-energy approximation to the log-likelihood is monitored to control the step size. Our experimental results show that effective models can be learned from fragmentary labelings and that incorporating top-down aggregate features significantly improves the segmentations. The resulting segmentations are compared to the state-of-the-art on three different image datasets. 1
6 0.60150331 93 nips-2007-GRIFT: A graphical model for inferring visual classification features from human data
7 0.59625107 69 nips-2007-Discriminative Batch Mode Active Learning
8 0.58476704 56 nips-2007-Configuration Estimates Improve Pedestrian Finding
9 0.58082342 88 nips-2007-Fast and Scalable Training of Semi-Supervised CRFs with Application to Activity Recognition
10 0.58045125 28 nips-2007-Augmented Functional Time Series Representation and Forecasting with Gaussian Processes
11 0.56921756 212 nips-2007-Using Deep Belief Nets to Learn Covariance Kernels for Gaussian Processes
12 0.56514049 94 nips-2007-Gaussian Process Models for Link Analysis and Transfer Learning
13 0.5634281 136 nips-2007-Multiple-Instance Active Learning
14 0.55908769 97 nips-2007-Hidden Common Cause Relations in Relational Learning
15 0.55567122 100 nips-2007-Hippocampal Contributions to Control: The Third Way
16 0.5555163 175 nips-2007-Semi-Supervised Multitask Learning
17 0.55511016 157 nips-2007-Privacy-Preserving Belief Propagation and Sampling
18 0.55480444 104 nips-2007-Inferring Neural Firing Rates from Spike Trains Using Gaussian Processes
19 0.55240083 83 nips-2007-Evaluating Search Engines by Modeling the Relationship Between Relevance and Clicks
20 0.54792845 164 nips-2007-Receptive Fields without Spike-Triggering