nips nips2002 knowledge-graph by maker-knowledge-mining
1 nips-2002-"Name That Song!" A Probabilistic Approach to Querying on Music and Text
Author: Brochu Eric, Nando de Freitas
Abstract: We present a novel, flexible statistical approach for modelling music and text jointly. The approach is based on multi-modal mixture models and maximum a posteriori estimation using EM. The learned models can be used to browse databases with documents containing music and text, to search for music using queries consisting of music and text (lyrics and other contextual information), to annotate text documents with music, and to automatically recommend or identify similar songs.
2 nips-2002-A Bilinear Model for Sparse Coding
Author: David B. Grimes, Rajesh P. Rao
Abstract: Recent algorithms for sparse coding and independent component analysis (ICA) have demonstrated how localized features can be learned from natural images. However, these approaches do not take image transformations into account. As a result, they produce image codes that are redundant because the same feature is learned at multiple locations. We describe an algorithm for sparse coding based on a bilinear generative model of images. By explicitly modeling the interaction between image features and their transformations, the bilinear approach helps reduce redundancy in the image code and provides a basis for transformationinvariant vision. We present results demonstrating bilinear sparse coding of natural images. We also explore an extension of the model that can capture spatial relationships between the independent features of an object, thereby providing a new framework for parts-based object recognition.
3 nips-2002-A Convergent Form of Approximate Policy Iteration
Author: Theodore J. Perkins, Doina Precup
Abstract: We study a new, model-free form of approximate policy iteration which uses Sarsa updates with linear state-action value function approximation for policy evaluation, and a “policy improvement operator” to generate a new policy based on the learned state-action values. We prove that if the policy improvement operator produces -soft policies and is Lipschitz continuous in the action values, with a constant that is not too large, then the approximate policy iteration algorithm converges to a unique solution from any initial policy. To our knowledge, this is the first convergence result for any form of approximate policy iteration under similar computational-resource assumptions.
4 nips-2002-A Differential Semantics for Jointree Algorithms
Author: James D. Park, Adnan Darwiche
Abstract: A new approach to inference in belief networks has been recently proposed, which is based on an algebraic representation of belief networks using multi–linear functions. According to this approach, the key computational question is that of representing multi–linear functions compactly, since inference reduces to a simple process of ev aluating and differentiating such functions. W e show here that mainstream inference algorithms based on jointrees are a special case of this approach in a v ery precise sense. W e use this result to prov e new properties of jointree algorithms, and then discuss some of its practical and theoretical implications. 1
5 nips-2002-A Digital Antennal Lobe for Pattern Equalization: Analysis and Design
Author: Alex Holub, Gilles Laurent, Pietro Perona
Abstract: Re-mapping patterns in order to equalize their distribution may greatly simplify both the structure and the training of classifiers. Here, the properties of one such map obtained by running a few steps of discrete-time dynamical system are explored. The system is called 'Digital Antennal Lobe' (DAL) because it is inspired by recent studies of the antennallobe, a structure in the olfactory system of the grasshopper. The pattern-spreading properties of the DAL as well as its average behavior as a function of its (few) design parameters are analyzed by extending previous results of Van Vreeswijk and Sompolinsky. Furthermore, a technique for adapting the parameters of the initial design in order to obtain opportune noise-rejection behavior is suggested. Our results are demonstrated with a number of simulations. 1
6 nips-2002-A Formulation for Minimax Probability Machine Regression
Author: Thomas Strohmann, Gregory Z. Grudic
Abstract: We formulate the regression problem as one of maximizing the minimum probability, symbolized by Ω, that future predicted outputs of the regression model will be within some ±ε bound of the true regression function. Our formulation is unique in that we obtain a direct estimate of this lower probability bound Ω. The proposed framework, minimax probability machine regression (MPMR), is based on the recently described minimax probability machine classification algorithm [Lanckriet et al.] and uses Mercer Kernels to obtain nonlinear regression models. MPMR is tested on both toy and real world data, verifying the accuracy of the Ω bound, and the efficacy of the regression models. 1
7 nips-2002-A Hierarchical Bayesian Markovian Model for Motifs in Biopolymer Sequences
Author: Eric P. Xing, Michael I. Jordan, Richard M. Karp, Stuart Russell
Abstract: We propose a dynamic Bayesian model for motifs in biopolymer sequences which captures rich biological prior knowledge and positional dependencies in motif structure in a principled way. Our model posits that the position-specific multinomial parameters for monomer distribution are distributed as a latent Dirichlet-mixture random variable, and the position-specific Dirichlet component is determined by a hidden Markov process. Model parameters can be fit on training motifs using a variational EM algorithm within an empirical Bayesian framework. Variational inference is also used for detecting hidden motifs. Our model improves over previous models that ignore biological priors and positional dependence. It has much higher sensitivity to motifs during detection and a notable ability to distinguish genuine motifs from false recurring patterns.
Author: Dmitry Y. Pavlov, David M. Pennock
Abstract: We develop a maximum entropy (maxent) approach to generating recommendations in the context of a user’s current navigation stream, suitable for environments where data is sparse, high-dimensional, and dynamic— conditions typical of many recommendation applications. We address sparsity and dimensionality reduction by first clustering items based on user access patterns so as to attempt to minimize the apriori probability that recommendations will cross cluster boundaries and then recommending only within clusters. We address the inherent dynamic nature of the problem by explicitly modeling the data as a time series; we show how this representational expressivity fits naturally into a maxent framework. We conduct experiments on data from ResearchIndex, a popular online repository of over 470,000 computer science documents. We show that our maxent formulation outperforms several competing algorithms in offline tests simulating the recommendation of documents to ResearchIndex users.
9 nips-2002-A Minimal Intervention Principle for Coordinated Movement
Author: Emanuel Todorov, Michael I. Jordan
Abstract: Behavioral goals are achieved reliably and repeatedly with movements rarely reproducible in their detail. Here we offer an explanation: we show that not only are variability and goal achievement compatible, but indeed that allowing variability in redundant dimensions is the optimal control strategy in the face of uncertainty. The optimal feedback control laws for typical motor tasks obey a “minimal intervention” principle: deviations from the average trajectory are only corrected when they interfere with the task goals. The resulting behavior exhibits task-constrained variability, as well as synergetic coupling among actuators—which is another unexplained empirical phenomenon.
10 nips-2002-A Model for Learning Variance Components of Natural Images
Author: Yan Karklin, Michael S. Lewicki
Abstract: We present a hierarchical Bayesian model for learning efficient codes of higher-order structure in natural images. The model, a non-linear generalization of independent component analysis, replaces the standard assumption of independence for the joint distribution of coefficients with a distribution that is adapted to the variance structure of the coefficients of an efficient image basis. This offers a novel description of higherorder image structure and provides a way to learn coarse-coded, sparsedistributed representations of abstract image properties such as object location, scale, and texture.
11 nips-2002-A Model for Real-Time Computation in Generic Neural Microcircuits
Author: Wolfgang Maass, Thomas Natschläger, Henry Markram
Abstract: A key challenge for neural modeling is to explain how a continuous stream of multi-modal input from a rapidly changing environment can be processed by stereotypical recurrent circuits of integrate-and-fire neurons in real-time. We propose a new computational model that is based on principles of high dimensional dynamical systems in combination with statistical learning theory. It can be implemented on generic evolved or found recurrent circuitry.
12 nips-2002-A Neural Edge-Detection Model for Enhanced Auditory Sensitivity in Modulated Noise
Author: Alon Fishbach, Bradford J. May
Abstract: Psychophysical data suggest that temporal modulations of stimulus amplitude envelopes play a prominent role in the perceptual segregation of concurrent sounds. In particular, the detection of an unmodulated signal can be significantly improved by adding amplitude modulation to the spectral envelope of a competing masking noise. This perceptual phenomenon is known as “Comodulation Masking Release” (CMR). Despite the obvious influence of temporal structure on the perception of complex auditory scenes, the physiological mechanisms that contribute to CMR and auditory streaming are not well known. A recent physiological study by Nelken and colleagues has demonstrated an enhanced cortical representation of auditory signals in modulated noise. Our study evaluates these CMR-like response patterns from the perspective of a hypothetical auditory edge-detection neuron. It is shown that this simple neural model for the detection of amplitude transients can reproduce not only the physiological data of Nelken et al., but also, in light of previous results, a variety of physiological and psychoacoustical phenomena that are related to the perceptual segregation of concurrent sounds. 1 In t rod u ct i on The temporal structure of a complex sound exerts strong influences on auditory physiology (e.g. [10, 16]) and perception (e.g. [9, 19, 20]). In particular, studies of auditory scene analysis have demonstrated the importance of the temporal structure of amplitude envelopes in the perceptual segregation of concurrent sounds [2, 7]. Common amplitude transitions across frequency serve as salient cues for grouping sound energy into unified perceptual objects. Conversely, asynchronous amplitude transitions enhance the separation of competing acoustic events [3, 4]. These general principles are manifested in perceptual phenomena as diverse as comodulation masking release (CMR) [13], modulation detection interference [22] and synchronous onset grouping [8]. Despite the obvious importance of timing information in psychoacoustic studies of auditory masking, the way in which the CNS represents the temporal structure of an amplitude envelope is not well understood. Certainly many physiological studies have demonstrated neural sensitivities to envelope transitions, but this sensitivity is only beginning to be related to the variety of perceptual experiences that are evoked by signals in noise. Nelken et al. [15] have suggested a correspondence between neural responses to time-varying amplitude envelopes and psychoacoustic masking phenomena. In their study of neurons in primary auditory cortex (A1), adding temporal modulation to background noise lowered the detection thresholds of unmodulated tones. This enhanced signal detection is similar to the perceptual phenomenon that is known as comodulation masking release [13]. Fishbach et al. [11] have recently proposed a neural model for the detection of “auditory edges” (i.e., amplitude transients) that can account for numerous physiological [14, 17, 18] and psychoacoustical [3, 21] phenomena. The encompassing utility of this edge-detection model suggests a common mechanism that may link the auditory processing and perception of auditory signals in a complex auditory scene. Here, it is shown that the auditory edge detection model can accurately reproduce the cortical CMR-like responses previously described by Nelken and colleagues. 2 Th e M od el The model is described in detail elsewhere [11]. In short, the basic operation of the model is the calculation of the first-order time derivative of the log-compressed envelope of the stimulus. A computational model [23] is used to convert the acoustic waveform to a physiologically plausible auditory nerve representation (Fig 1a). The simulated neural response has a medium spontaneous rate and a characteristic frequency that is set to the frequency of the target tone. To allow computation of the time derivative of the stimulus envelope, we hypothesize the existence of a temporal delay dimension, along which the stimulus is progressively delayed. The intermediate delay layer (Fig 1b) is constructed from an array of neurons with ascending membrane time constants (τ); each neuron is modeled by a conventional integrate-and-fire model (I&F;, [12]). Higher membrane time constant induces greater delay in the neuron’s response [1]. The output of the delay layer converges to a single output neuron (Fig. 1c) via a set of connection with various efficacies that reflect a receptive field of a gaussian derivative. This combination of excitatory and inhibitory connections carries out the time-derivative computation. Implementation details and parameters are given in [11]. The model has 2 adjustable and 6 fixed parameters, the former were used to fit the responses of the model to single unit responses to variety of stimuli [11]. The results reported here are not sensitive to these parameters. (a) AN model (b) delay-layer (c) edge-detector neuron τ=6 ms I&F; Neuron τ=4 ms τ=3 ms bandpass log d dt RMS Figure 1: Schematic diagram of the model and a block diagram of the basic operation of each model component (shaded area). The stimulus is converted to a neural representation (a) that approximates the average firing rate of a medium spontaneous-rate AN fiber [23]. The operation of this stage can be roughly described as the log-compressed rms output of a bandpass filter. The neural representation is fed to a series of neurons with ascending membrane time constant (b). The kernel functions that are used to simulate these neurons are plotted for a few neurons along with the time constants used. The output of the delay-layer neurons converge to a single I&F; neuron (c) using a set of connections with weights that reflect a shape of a gaussian derivative. Solid arrows represent excitatory connections and white arrows represent inhibitory connections. The absolute efficacy is represented by the width of the arrows. 3 Resu lt s Nelken et al. [15] report that amplitude modulation can substantially modify the noise-driven discharge rates of A1 neurons in Halothane-anesthetized cats. Many cortical neurons show only a transient onset response to unmodulated noise but fire in synchrony (“lock”) to the envelope of modulated noise. A significant reduction in envelope-locked discharge rates is observed if an unmodulated tone is added to modulated noise. As summarized in Fig. 2, this suppression of envelope locking can reveal the presence of an auditory signal at sound pressure levels that are not detectable in unmodulated noise. It has been suggested that this pattern of neural responding may represent a physiological equivalent of CMR. Reproduction of CMR-like cortical activity can be illustrated by a simplified case in which the analytical amplitude envelope of the stimulus is used as the input to the edge-detector model. In keeping with the actual physiological approach of Nelken et al., the noise envelope is shaped by a trapezoid modulator for these simulations. Each cycle of modulation, E N(t), is given by: t 0≤t < 3D E N (t ) = P P − D (t − 3 D ) 3 D ≤ t < 4 D 0 4 D ≤ t < 8D £ P D ¢ ¡ where P is the peak pressure level and D is set to 12.5 ms. (b) Modulated noise 76 Spikes/sec Tone level (dB SPL) (a) Unmodulated noise 26 0 150 300 0 150 300 Time (ms) Figure 2: Responses of an A1 unit to a combination of noise and tone at many tone levels, replotted from Nelken et al. [15]. (a) Unmodulated noise and (b) modulated noise. The noise envelope is illustrated by the thick line above each figure. Each row shows the response of the neuron to the noise plus the tone at the level specified on the ordinate. The dashed line in (b) indicates the detection threshold level for the tone. The detection threshold (as defined and calculated by Nelken et al.) in the unmodulated noise was not reached. Since the basic operation of the model is the calculation of the rectified timederivative of the log-compressed envelope of the stimulus, the expected noisedriven rate of the model can be approximated by: ( ) ¢ E (t ) P0 d A ln 1 + dt ¡ M N ( t ) = max 0, ¥ ¤ £ where A=20/ln(10) and P0 =2e-5 Pa. The expected firing rate in response to the noise plus an unmodulated signal (tone) can be similarly approximated by: ) ¨ E ( t ) + PS P0 ¦ ( d A ln 1 + dt § M N + S ( t ) = max 0, © where PS is the peak pressure level of the tone. Clearly, both MN (t) and MN+S (t) are identically zero outside the interval [0 D]. Within this interval it holds that: M N (t ) = AP D P0 + P D t 0≤t < D Clearly, M N + S < M N for the interval [0 D] of each modulation cycle. That is, the addition of a tone reduces the responses of the model to the rising part of the modulated envelope. Higher tone levels (Ps ) cause greater reduction in the model’s firing rate. (c) (b) Level derivative (dB SPL/ms) Level (dB SPL) (a) (d) Time (ms) Figure 3: An illustration of the basic operation of the model on various amplitude envelopes. The simplified operation of the model includes log compression of the amplitude envelope (a and c) and rectified time-derivative of the log-compressed envelope (b and d). (a) A 30 dB SPL tone is added to a modulated envelope (peak level of 70 dB SPL) 300 ms after the beginning of the stimulus (as indicated by the horizontal line). The addition of the tone causes a great reduction in the time derivative of the log-compressed envelope (b). When the envelope of the noise is unmodulated (c), the time-derivative of the log-compressed envelope (d) shows a tiny spike when the tone is added (marked by the arrow). Fig. 3 demonstrates the effect of a low-level tone on the time-derivative of the logcompressed envelope of a noise. When the envelope is modulated (Fig. 3a) the addition of the tone greatly reduces the derivative of the rising part of the modulation (Fig. 3b). In the absence of modulations (Fig. 3c), the tone presentation produces a negligible effect on the level derivative (Fig. 3d). Model simulations of neural responses to the stimuli used by Nelken et al. are plotted in Fig. 4. As illustrated schematically in Fig 3 (d), the presence of the tone does not cause any significant change in the responses of the model to the unmodulated noise (Fig. 4a). In the modulated noise, however, tones of relatively low levels reduce the responses of the model to the rising part of the envelope modulations. (b) Modulated noise 76 Spikes/sec Tone level (dB SPL) (a) Unmodulated noise 26 0 150 300 0 Time (ms) 150 300 Figure 4: Simulated responses of the model to a combination of a tone and Unmodulated noise (a) and modulated noise (b). All conventions are as in Fig. 2. 4 Di scu ssi on This report uses an auditory edge-detection model to simulate the actual physiological consequences of amplitude modulation on neural sensitivity in cortical area A1. The basic computational operation of the model is the calculation of the smoothed time-derivative of the log-compressed stimulus envelope. The ability of the model to reproduce cortical response patterns in detail across a variety of stimulus conditions suggests similar time-sensitive mechanisms may contribute to the physiological correlates of CMR. These findings augment our previous observations that the simple edge-detection model can successfully predict a wide range of physiological and perceptual phenomena [11]. Former applications of the model to perceptual phenomena have been mainly related to auditory scene analysis, or more specifically the ability of the auditory system to distinguish multiple sound sources. In these cases, a sharp amplitude transition at stimulus onset (“auditory edge”) was critical for sound segregation. Here, it is shown that the detection of acoustic signals also may be enhanced through the suppression of ongoing responses to the concurrent modulations of competing background sounds. Interestingly, these temporal fluctuations appear to be a common property of natural soundscapes [15]. The model provides testable predictions regarding how signal detection may be influenced by the temporal shape of amplitude modulation. Carlyon et al. [6] measured CMR in human listeners using three types of noise modulation: squarewave, sine wave and multiplied noise. From the perspective of the edge-detection model, these psychoacoustic results are intriguing because the different modulator types represent manipulations of the time derivative of masker envelopes. Squarewave modulation had the most sharply edged time derivative and produced the greatest masking release. Fig. 5 plots the responses of the model to a pure-tone signal in square-wave and sine-wave modulated noise. As in the psychoacoustical data of Carlyon et al., the simulated detection threshold was lower in the context of square-wave modulation. Our modeling results suggest that the sharply edged square wave evoked higher levels of noise-driven activity and therefore created a sensitive background for the suppressing effects of the unmodulated tone. (b) 60 Spikes/sec Tone level (dB SPL) (a) 10 0 200 400 600 0 Time (ms) 200 400 600 Figure 5: Simulated responses of the model to a combination of a tone at various levels and a sine-wave modulated noise (a) or a square-wave modulated noise (b). Each row shows the response of the model to the noise plus the tone at the level specified on the abscissa. The shape of the noise modulator is illustrated above each figure. The 100 ms tone starts 250 ms after the noise onset. Note that the tone detection threshold (marked by the dashed line) is 10 dB lower for the square-wave modulator than for the sine-wave modulator, in accordance with the psychoacoustical data of Carlyon et al. [6]. Although the physiological basis of our model was derived from studies of neural responses in the cat auditory system, the key psychoacoustical observations of Carlyon et al. have been replicated in recent behavioral studies of cats (Budelis et al. [5]). These data support the generalization of human perceptual processing to other species and enhance the possible correspondence between the neuronal CMR-like effect and the psychoacoustical masking phenomena. Clearly, the auditory system relies on information other than the time derivative of the stimulus envelope for the detection of auditory signals in background noise. Further physiological and psychoacoustic assessments of CMR-like masking effects are needed not only to refine the predictive abilities of the edge-detection model but also to reveal the additional sources of acoustic information that influence signal detection in constantly changing natural environments. Ackn ow led g men t s This work was supported in part by a NIDCD grant R01 DC004841. Refe ren ces [1] Agmon-Snir H., Segev I. (1993). “Signal delay and input synchronization in passive dendritic structure”, J. Neurophysiol. 70, 2066-2085. [2] Bregman A.S. (1990). “Auditory scene analysis: The perceptual organization of sound”, MIT Press, Cambridge, MA. [3] Bregman A.S., Ahad P.A., Kim J., Melnerich L. (1994) “Resetting the pitch-analysis system. 1. Effects of rise times of tones in noise backgrounds or of harmonics in a complex tone”, Percept. Psychophys. 56 (2), 155-162. [4] Bregman A.S., Ahad P.A., Kim J. (1994) “Resetting the pitch-analysis system. 2. Role of sudden onsets and offsets in the perception of individual components in a cluster of overlapping tones”, J. Acoust. Soc. Am. 96 (5), 2694-2703. [5] Budelis J., Fishbach A., May B.J. (2002) “Behavioral assessments of comodulation masking release in cats”, Abst. Assoc. for Res. in Otolaryngol. 25. [6] Carlyon R.P., Buus S., Florentine M. (1989) “Comodulation masking release for three types of modulator as a function of modulation rate”, Hear. Res. 42, 37-46. [7] Darwin C.J. (1997) “Auditory grouping”, Trends in Cog. Sci. 1(9), 327-333. [8] Darwin C.J., Ciocca V. (1992) “Grouping in pitch perception: Effects of onset asynchrony and ear of presentation of a mistuned component”, J. Acoust. Soc. Am. 91 , 33813390. [9] Drullman R., Festen H.M., Plomp R. (1994) “Effect of temporal envelope smearing on speech reception”, J. Acoust. Soc. Am. 95 (2), 1053-1064. [10] Eggermont J J. (1994). “Temporal modulation transfer functions for AM and FM stimuli in cat auditory cortex. Effects of carrier type, modulating waveform and intensity”, Hear. Res. 74, 51-66. [11] Fishbach A., Nelken I., Yeshurun Y. (2001) “Auditory edge detection: a neural model for physiological and psychoacoustical responses to amplitude transients”, J. Neurophysiol. 85, 2303–2323. [12] Gerstner W. (1999) “Spiking neurons”, in Pulsed Neural Networks , edited by W. Maass, C. M. Bishop, (MIT Press, Cambridge, MA). [13] Hall J.W., Haggard M.P., Fernandes M.A. (1984) “Detection in noise by spectrotemporal pattern analysis”, J. Acoust. Soc. Am. 76, 50-56. [14] Heil P. (1997) “Auditory onset responses revisited. II. Response strength”, J. Neurophysiol. 77, 2642-2660. [15] Nelken I., Rotman Y., Bar-Yosef O. (1999) “Responses of auditory cortex neurons to structural features of natural sounds”, Nature 397, 154-157. [16] Phillips D.P. (1988). “Effect of Tone-Pulse Rise Time on Rate-Level Functions of Cat Auditory Cortex Neurons: Excitatory and Inhibitory Processes Shaping Responses to Tone Onset”, J. Neurophysiol. 59, 1524-1539. [17] Phillips D.P., Burkard R. (1999). “Response magnitude and timing of auditory response initiation in the inferior colliculus of the awake chinchilla”, J. Acoust. Soc. Am. 105, 27312737. [18] Phillips D.P., Semple M.N., Kitzes L.M. (1995). “Factors shaping the tone level sensitivity of single neurons in posterior field of cat auditory cortex”, J. Neurophysiol. 73, 674-686. [19] Rosen S. (1992) “Temporal information in speech: acoustic, auditory and linguistic aspects”, Phil. Trans. R. Soc. Lond. B 336, 367-373. [20] Shannon R.V., Zeng F.G., Kamath V., Wygonski J, Ekelid M. (1995) “Speech recognition with primarily temporal cues”, Science 270, 303-304. [21] Turner C.W., Relkin E.M., Doucet J. (1994). “Psychophysical and physiological forward masking studies: probe duration and rise-time effects”, J. Acoust. Soc. Am. 96 (2), 795-800. [22] Yost W.A., Sheft S. (1994) “Modulation detection interference – across-frequency processing and auditory grouping”, Hear. Res. 79, 48-58. [23] Zhang X., Heinz M.G., Bruce I.C., Carney L.H. (2001). “A phenomenological model for the responses of auditory-nerve fibers: I. Nonlinear tuning with compression and suppression”, J. Acoust. Soc. Am. 109 (2), 648-670.
Author: Eric Allender, Sanjeev Arora, Michael Kearns, Cristopher Moore, Alexander Russell
Abstract: We establish a new hardness result that shows that the difficulty of planning in factored Markov decision processes is representational rather than just computational. More precisely, we give a fixed family of factored MDPs with linear rewards whose optimal policies and value functions simply cannot be represented succinctly in any standard parametric form. Previous hardness results indicated that computing good policies from the MDP parameters was difficult, but left open the possibility of succinct function approximation for any fixed factored MDP. Our result applies even to policies which yield a polynomially poor approximation to the optimal value, and highlights interesting connections with the complexity class of Arthur-Merlin games.
14 nips-2002-A Probabilistic Approach to Single Channel Blind Signal Separation
Author: Gil-jin Jang, Te-Won Lee
Abstract: We present a new technique for achieving source separation when given only a single channel recording. The main idea is based on exploiting the inherent time structure of sound sources by learning a priori sets of basis filters in time domain that encode the sources in a statistically efficient manner. We derive a learning algorithm using a maximum likelihood approach given the observed single channel data and sets of basis filters. For each time point we infer the source signals and their contribution factors. This inference is possible due to the prior knowledge of the basis filters and the associated coefficient densities. A flexible model for density estimation allows accurate modeling of the observation and our experimental results exhibit a high level of separation performance for mixtures of two music signals as well as the separation of two voice signals.
15 nips-2002-A Probabilistic Model for Learning Concatenative Morphology
Author: Matthew G. Snover, Michael R. Brent
Abstract: This paper describes a system for the unsupervised learning of morphological suffixes and stems from word lists. The system is composed of a generative probability model and hill-climbing and directed search algorithms. By extracting and examining morphologically rich subsets of an input lexicon, the directed search identifies highly productive paradigms. The hill-climbing algorithm then further maximizes the probability of the hypothesis. Quantitative results are shown by measuring the accuracy of the morphological relations identified. Experiments in English and Polish, as well as comparisons with another recent unsupervised morphology learning algorithm demonstrate the effectiveness of this technique.
16 nips-2002-A Prototype for Automatic Recognition of Spontaneous Facial Actions
Author: M.S. Bartlett, G.C. Littlewort, T.J. Sejnowski, J.R. Movellan
Abstract: We present ongoing work on a project for automatic recognition of spontaneous facial actions. Spontaneous facial expressions differ substantially from posed expressions, similar to how continuous, spontaneous speech differs from isolated words produced on command. Previous methods for automatic facial expression recognition assumed images were collected in controlled environments in which the subjects deliberately faced the camera. Since people often nod or turn their heads, automatic recognition of spontaneous facial behavior requires methods for handling out-of-image-plane head rotations. Here we explore an approach based on 3-D warping of images into canonical views. We evaluated the performance of the approach as a front-end for a spontaneous expression recognition system using support vector machines and hidden Markov models. This system employed general purpose learning mechanisms that can be applied to recognition of any facial movement. The system was tested for recognition of a set of facial actions defined by the Facial Action Coding System (FACS). We showed that 3D tracking and warping followed by machine learning techniques directly applied to the warped images, is a viable and promising technology for automatic facial expression recognition. One exciting aspect of the approach presented here is that information about movement dynamics emerged out of filters which were derived from the statistics of images.
17 nips-2002-A Statistical Mechanics Approach to Approximate Analytical Bootstrap Averages
Author: Dörthe Malzahn, Manfred Opper
Abstract: We apply the replica method of Statistical Physics combined with a variational method to the approximate analytical computation of bootstrap averages for estimating the generalization error. We demonstrate our approach on regression with Gaussian processes and compare our results with averages obtained by Monte-Carlo sampling.
18 nips-2002-Adaptation and Unsupervised Learning
Author: Peter Dayan, Maneesh Sahani, Gregoire Deback
Abstract: Adaptation is a ubiquitous neural and psychological phenomenon, with a wealth of instantiations and implications. Although a basic form of plasticity, it has, bar some notable exceptions, attracted computational theory of only one main variety. In this paper, we study adaptation from the perspective of factor analysis, a paradigmatic technique of unsupervised learning. We use factor analysis to re-interpret a standard view of adaptation, and apply our new model to some recent data on adaptation in the domain of face discrimination.
19 nips-2002-Adapting Codes and Embeddings for Polychotomies
Author: Gunnar Rätsch, Sebastian Mika, Alex J. Smola
Abstract: In this paper we consider formulations of multi-class problems based on a generalized notion of a margin and using output coding. This includes, but is not restricted to, standard multi-class SVM formulations. Differently from many previous approaches we learn the code as well as the embedding function. We illustrate how this can lead to a formulation that allows for solving a wider range of problems with for instance many classes or even “missing classes”. To keep our optimization problems tractable we propose an algorithm capable of solving them using twoclass classifiers, similar in spirit to Boosting.
20 nips-2002-Adaptive Caching by Refetching
Author: Robert B. Gramacy, Manfred K. Warmuth, Scott A. Brandt, Ismail Ari
Abstract: We are constructing caching policies that have 13-20% lower miss rates than the best of twelve baseline policies over a large variety of request streams. This represents an improvement of 49–63% over Least Recently Used, the most commonly implemented policy. We achieve this not by designing a specific new policy but by using on-line Machine Learning algorithms to dynamically shift between the standard policies based on their observed miss rates. A thorough experimental evaluation of our techniques is given, as well as a discussion of what makes caching an interesting on-line learning problem.
21 nips-2002-Adaptive Classification by Variational Kalman Filtering
22 nips-2002-Adaptive Nonlinear System Identification with Echo State Networks
23 nips-2002-Adaptive Quantization and Density Estimation in Silicon
24 nips-2002-Adaptive Scaling for Feature Selection in SVMs
25 nips-2002-An Asynchronous Hidden Markov Model for Audio-Visual Speech Recognition
26 nips-2002-An Estimation-Theoretic Framework for the Presentation of Multiple Stimuli
27 nips-2002-An Impossibility Theorem for Clustering
28 nips-2002-An Information Theoretic Approach to the Functional Classification of Neurons
29 nips-2002-Analysis of Information in Speech Based on MANOVA
30 nips-2002-Annealing and the Rate Distortion Problem
31 nips-2002-Application of Variational Bayesian Approach to Speech Recognition
32 nips-2002-Approximate Inference and Protein-Folding
33 nips-2002-Approximate Linear Programming for Average-Cost Dynamic Programming
34 nips-2002-Artefactual Structure from Least-Squares Multidimensional Scaling
35 nips-2002-Automatic Acquisition and Efficient Representation of Syntactic Structures
36 nips-2002-Automatic Alignment of Local Representations
37 nips-2002-Automatic Derivation of Statistical Algorithms: The EM Family and Beyond
38 nips-2002-Bayesian Estimation of Time-Frequency Coefficients for Audio Signal Enhancement
39 nips-2002-Bayesian Image Super-Resolution
40 nips-2002-Bayesian Models of Inductive Generalization
41 nips-2002-Bayesian Monte Carlo
42 nips-2002-Bias-Optimal Incremental Problem Solving
43 nips-2002-Binary Coding in Auditory Cortex
44 nips-2002-Binary Tuning is Optimal for Neural Rate Coding with High Temporal Resolution
45 nips-2002-Boosted Dyadic Kernel Discriminants
46 nips-2002-Boosting Density Estimation
47 nips-2002-Branching Law for Axons
49 nips-2002-Charting a Manifold
50 nips-2002-Circuit Model of Short-Term Synaptic Dynamics
51 nips-2002-Classifying Patterns of Visual Motion - a Neuromorphic Approach
52 nips-2002-Cluster Kernels for Semi-Supervised Learning
53 nips-2002-Clustering with the Fisher Score
54 nips-2002-Combining Dimensions and Features in Similarity-Based Representations
55 nips-2002-Combining Features for BCI
56 nips-2002-Concentration Inequalities for the Missing Mass and for Histogram Rule Error
57 nips-2002-Concurrent Object Recognition and Segmentation by Graph Partitioning
58 nips-2002-Conditional Models on the Ranking Poset
59 nips-2002-Constraint Classification for Multiclass Classification and Ranking
60 nips-2002-Convergence Properties of Some Spike-Triggered Analysis Techniques
61 nips-2002-Convergent Combinations of Reinforcement Learning with Linear Function Approximation
63 nips-2002-Critical Lines in Symmetry of Mixture Models and its Application to Component Splitting
64 nips-2002-Data-Dependent Bounds for Bayesian Mixture Methods
65 nips-2002-Derivative Observations in Gaussian Process Models of Dynamic Systems
67 nips-2002-Discriminative Binaural Sound Localization
68 nips-2002-Discriminative Densities from Maximum Contrast Estimation
69 nips-2002-Discriminative Learning for Label Sequences via Boosting
70 nips-2002-Distance Metric Learning with Application to Clustering with Side-Information
71 nips-2002-Dopamine Induced Bistability Enhances Signal Processing in Spiny Neurons
72 nips-2002-Dyadic Classification Trees via Structural Risk Minimization
73 nips-2002-Dynamic Bayesian Networks with Deterministic Latent Tables
74 nips-2002-Dynamic Structure Super-Resolution
75 nips-2002-Dynamical Causal Learning
76 nips-2002-Dynamical Constraints on Computing with Spike Timing in the Cortex
77 nips-2002-Effective Dimension and Generalization of Kernel Learning
78 nips-2002-Efficient Learning Equilibrium
79 nips-2002-Evidence Optimization Techniques for Estimating Stimulus-Response Functions
80 nips-2002-Exact MAP Estimates by (Hyper)tree Agreement
81 nips-2002-Expected and Unexpected Uncertainty: ACh and NE in the Neocortex
82 nips-2002-Exponential Family PCA for Belief Compression in POMDPs
83 nips-2002-Extracting Relevant Structures with Side Information
84 nips-2002-Fast Exact Inference with a Factored Model for Natural Language Parsing
85 nips-2002-Fast Kernels for String and Tree Matching
86 nips-2002-Fast Sparse Gaussian Process Methods: The Informative Vector Machine
87 nips-2002-Fast Transformation-Invariant Factor Analysis
89 nips-2002-Feature Selection by Maximum Marginal Diversity
90 nips-2002-Feature Selection in Mixture-Based Clustering
91 nips-2002-Field-Programmable Learning Arrays
92 nips-2002-FloatBoost Learning for Classification
93 nips-2002-Forward-Decoding Kernel-Based Phone Recognition
94 nips-2002-Fractional Belief Propagation
96 nips-2002-Generalized² Linear² Models
97 nips-2002-Global Versus Local Methods in Nonlinear Dimensionality Reduction
98 nips-2002-Going Metric: Denoising Pairwise Data
100 nips-2002-Half-Lives of EigenFlows for Spectral Clustering
101 nips-2002-Handling Missing Data with Variational Bayesian Learning of ICA
102 nips-2002-Hidden Markov Model of Cortical Synaptic Plasticity: Derivation of the Learning Rule
103 nips-2002-How Linear are Auditory Cortical Responses?
104 nips-2002-How the Poverty of the Stimulus Solves the Poverty of the Stimulus
105 nips-2002-How to Combine Color and Shape Information for 3D Object Recognition: Kernels do the Trick
107 nips-2002-Identity Uncertainty and Citation Matching
108 nips-2002-Improving Transfer Rates in Brain Computer Interfacing: A Case Study
109 nips-2002-Improving a Page Classifier with Anchor Extraction and Link Analysis
110 nips-2002-Incremental Gaussian Processes
111 nips-2002-Independent Components Analysis through Product Density Estimation
112 nips-2002-Inferring a Semantic Representation of Text via Cross-Language Correlation Analysis
113 nips-2002-Information Diffusion Kernels
114 nips-2002-Information Regularization with Partially Labeled Data
115 nips-2002-Informed Projections
116 nips-2002-Interpreting Neural Response Variability as Monte Carlo Sampling of the Posterior
117 nips-2002-Intrinsic Dimension Estimation Using Packing Numbers
119 nips-2002-Kernel Dependency Estimation
120 nips-2002-Kernel Design Using Boosting
121 nips-2002-Knowledge-Based Support Vector Machine Classifiers
122 nips-2002-Learning About Multiple Objects in Images: Factorial Learning without Factorial Search
123 nips-2002-Learning Attractor Landscapes for Learning Motor Primitives
124 nips-2002-Learning Graphical Models with Mercer Kernels
125 nips-2002-Learning Semantic Similarity
126 nips-2002-Learning Sparse Multiscale Image Representations
127 nips-2002-Learning Sparse Topographic Representations with Products of Student-t Distributions
128 nips-2002-Learning a Forward Model of a Reflex
129 nips-2002-Learning in Spiking Neural Assemblies
130 nips-2002-Learning in Zero-Sum Team Markov Games Using Factored Value Functions
131 nips-2002-Learning to Classify Galaxy Shapes Using the EM Algorithm
132 nips-2002-Learning to Detect Natural Image Boundaries Using Brightness and Texture
133 nips-2002-Learning to Perceive Transparency from the Statistics of Natural Scenes
134 nips-2002-Learning to Take Concurrent Actions
135 nips-2002-Learning with Multiple Labels
137 nips-2002-Location Estimation with a Differential Update Network
138 nips-2002-Manifold Parzen Windows
139 nips-2002-Margin-Based Algorithms for Information Filtering
140 nips-2002-Margin Analysis of the LVQ Algorithm
141 nips-2002-Maximally Informative Dimensions: Analyzing Neural Responses to Natural Signals
142 nips-2002-Maximum Likelihood and the Information Bottleneck
143 nips-2002-Mean Field Approach to a Probabilistic Model in Information Retrieval
144 nips-2002-Minimax Differential Dynamic Programming: An Application to Robust Biped Walking
145 nips-2002-Mismatch String Kernels for SVM Protein Classification
146 nips-2002-Modeling Midazolam's Effect on the Hippocampus and Recognition Memory
147 nips-2002-Monaural Speech Separation
148 nips-2002-Morton-Style Factorial Coding of Color in Primary Visual Cortex
149 nips-2002-Multiclass Learning by Probabilistic Embeddings
150 nips-2002-Multiple Cause Vector Quantization
151 nips-2002-Multiplicative Updates for Nonnegative Quadratic Programming in Support Vector Machines
152 nips-2002-Nash Propagation for Loopy Graphical Games
153 nips-2002-Neural Decoding of Cursor Motion Using a Kalman Filter
154 nips-2002-Neuromorphic Bisable VLSI Synapses with Spike-Timing-Dependent Plasticity
155 nips-2002-Nonparametric Representation of Policies and Value Functions: A Trajectory-Based Approach
156 nips-2002-On the Complexity of Learning the Kernel Matrix
157 nips-2002-On the Dirichlet Prior and Bayesian Regularization
158 nips-2002-One-Class LP Classifiers for Dissimilarity Representations
159 nips-2002-Optimality of Reinforcement Learning Algorithms with Linear Function Approximation
160 nips-2002-Optoelectronic Implementation of a FitzHugh-Nagumo Neural Model
161 nips-2002-PAC-Bayes & Margins
162 nips-2002-Parametric Mixture Models for Multi-Labeled Text
163 nips-2002-Prediction and Semantic Association
164 nips-2002-Prediction of Protein Topologies Using Generalized IOHMMs and RNNs
165 nips-2002-Ranking with Large Margin Principle: Two Approaches
166 nips-2002-Rate Distortion Function in the Spin Glass State: A Toy Model
167 nips-2002-Rational Kernels
168 nips-2002-Real-Time Monitoring of Complex Industrial Processes with Particle Filters
169 nips-2002-Real-Time Particle Filters
171 nips-2002-Reconstructing Stimulus-Driven Neural Networks from Spike Times
172 nips-2002-Recovering Articulated Model Topology from Observed Rigid Motion
173 nips-2002-Recovering Intrinsic Images from a Single Image
174 nips-2002-Regularized Greedy Importance Sampling
175 nips-2002-Reinforcement Learning to Play an Optimal Nash Equilibrium in Team Markov Games
176 nips-2002-Replay, Repair and Consolidation
177 nips-2002-Retinal Processing Emulation in a Programmable 2-Layer Analog Array Processor CMOS Chip
178 nips-2002-Robust Novelty Detection with Single-Class MPM
179 nips-2002-Scaling of Probability-Based Optimization Algorithms
180 nips-2002-Selectivity and Metaplasticity in a Unified Calcium-Dependent Model
181 nips-2002-Self Supervised Boosting
182 nips-2002-Shape Recipes: Scene Representations that Refer to the Image
183 nips-2002-Source Separation with a Sensor Array using Graphical Models and Subband Filtering
184 nips-2002-Spectro-Temporal Receptive Fields of Subthreshold Responses in Auditory Cortex
185 nips-2002-Speeding up the Parti-Game Algorithm
186 nips-2002-Spike Timing-Dependent Plasticity in the Address Domain
187 nips-2002-Spikernels: Embedding Spiking Neurons in Inner-Product Spaces
188 nips-2002-Stability-Based Model Selection
189 nips-2002-Stable Fixed Points of Loopy Belief Propagation Are Local Minima of the Bethe Free Energy
190 nips-2002-Stochastic Neighbor Embedding
191 nips-2002-String Kernels, Fisher Kernels and Finite State Automata
192 nips-2002-Support Vector Machines for Multiple-Instance Learning
193 nips-2002-Temporal Coherence, Natural Image Sequences, and the Visual Cortex
194 nips-2002-The Decision List Machine
196 nips-2002-The RA Scanner: Prediction of Rheumatoid Joint Inflammation Based on Laser Imaging
198 nips-2002-Theory-Based Causal Inference
199 nips-2002-Timing and Partial Observability in the Dopamine System
200 nips-2002-Topographic Map Formation by Silicon Growth Cones
201 nips-2002-Transductive and Inductive Methods for Approximate Gaussian Process Regression
202 nips-2002-Unsupervised Color Constancy
203 nips-2002-Using Tarjan's Red Rule for Fast Dependency Tree Construction
204 nips-2002-VIBES: A Variational Inference Engine for Bayesian Networks
205 nips-2002-Value-Directed Compression of POMDPs
206 nips-2002-Visual Development Aids the Acquisition of Motion Velocity Sensitivities