nips nips2002 nips2002-55 knowledge-graph by maker-knowledge-mining
Source: pdf
Author: Guido Dornhege, Benjamin Blankertz, Gabriel Curio, Klaus-Robert Müller
Abstract: Recently, interest is growing to develop an effective communication interface connecting the human brain to a computer, the ’Brain-Computer Interface’ (BCI). One motivation of BCI research is to provide a new communication channel substituting normal motor output in patients with severe neuromuscular disabilities. In the last decade, various neurophysiological cortical processes, such as slow potential shifts, movement related potentials (MRPs) or event-related desynchronization (ERD) of spontaneous EEG rhythms, were shown to be suitable for BCI, and, consequently, different independent approaches of extracting BCI-relevant EEG-features for single-trial analysis are under investigation. Here, we present and systematically compare several concepts for combining such EEG-features to improve the single-trial classification. Feature combinations are evaluated on movement imagination experiments with 3 subjects where EEG-features are based on either MRPs or ERD, or both. Those combination methods that incorporate the assumption that the single EEG-features are physiologically mutually independent outperform the plain method of ’adding’ evidence where the single-feature vectors are simply concatenated. These results strengthen the hypothesis that MRP and ERD reflect at least partially independent aspects of cortical processes and open a new perspective to boost BCI effectiveness.
Reference: text
sentIndex sentText sentNum sentScore
1 ©§¢¤¢ Abstract Recently, interest is growing to develop an effective communication interface connecting the human brain to a computer, the ’Brain-Computer Interface’ (BCI). [sent-8, score-0.137]
2 One motivation of BCI research is to provide a new communication channel substituting normal motor output in patients with severe neuromuscular disabilities. [sent-9, score-0.159]
3 Feature combinations are evaluated on movement imagination experiments with 3 subjects where EEG-features are based on either MRPs or ERD, or both. [sent-12, score-0.275]
4 These results strengthen the hypothesis that MRP and ERD reflect at least partially independent aspects of cortical processes and open a new perspective to boost BCI effectiveness. [sent-14, score-0.157]
5 1 Introduction A brain-computer interface (BCI) is a system which translates a subject’s intentions into a control signal for a device, e. [sent-15, score-0.097]
6 When measuring non-invasively, brain activity is acquired by scalp-recorded electroencephalogram (EEG) from a subject that tries to convey its intentions by behaving according to well-defined paradigms, e. [sent-19, score-0.291]
7 ’Features’ (or feature vectors) are extracted from the digitized EEG-signals by signal processing methods. [sent-22, score-0.101]
8 These features are translated into a control signal, either (1) by simple equations or threshold criteria (with only a few free parameters that are estimated on training data), or (2) by machine learning algorithms that learn a more complex ∗ To whom correspondence should be addressed. [sent-23, score-0.136]
9 Concerning the pivotal step of feature extraction, neurophysiological a priori knowledge can aid to decide which EEG-feature is to be expected to hold the most discriminative information for the chosen paradigm. [sent-27, score-0.161]
10 Here, we present several methods for combining features to enhance single-trial EEG classification for BCI. [sent-32, score-0.159]
11 A special focus was placed on the question how to incorporate a priori knowledge about feature independence. [sent-33, score-0.101]
12 (1) Based on slow cortical potentials the Tübinger Thought Translation Device (TTD) [4] translates low-pass filtered brain activity from central scalp position into a vertical cursor movement on a computer screen. [sent-39, score-0.423]
13 This enables subjects to learn self-regulation of electrocortical positivity or negativity. [sent-40, score-0.074]
14 After some training, patients can generate binary decisions in a 4 seconds pace with an accuracies of up to 85 % and thereby handle a word processor or an internet browser. [sent-41, score-0.101]
15 (2) The Albany BCI system [2] allows the user to control cursor movement by oscillatory brain activity into one of two or four possible target areas on a computer screen. [sent-42, score-0.323]
16 In the first training sessions most subjects use some kind of motor imagery which is replaced by adapted strategies during further feedback sessions. [sent-43, score-0.291]
17 And (3), the Graz BCI system [5] is based on event-related modulations of the pericentral µ - and/or β -rhythms of sensorimotor cortices, with a focus on motor preparation and imagination. [sent-46, score-0.231]
18 Feature vectors calculated from spontaneous EEG signals by adaptive autoregressive modelling are used to train a classifier. [sent-47, score-0.133]
19 In a ternary classification task accuracies of over 96 % were obtained in an offline study with a trial duration of 8 seconds. [sent-48, score-0.115]
20 Most gain from a combination of different features is expected when the single features provide complementary information for the classification task. [sent-50, score-0.204]
21 , in primary (sensori-)motor cortex (M-1), supplementary motor area (SMA) and posterior parietal cortex (PP). [sent-53, score-0.19]
22 MRPs started over wide areas of the sensorimotor cortices (Bereitschaftspotential) and focalizes at the contralateral M-1 hand cortex with a steep negative slope prior to finger movement onset, reaching a negative peak approximately 100 ms after EMG onset (motor potential). [sent-55, score-0.481]
23 In contrast, a bilateral M-1 ERD just prior to movement onset appeared to reflect a more widespread cortical ’alerting’ function. [sent-56, score-0.222]
24 Note that these studies analyze movement preparation and execution only. [sent-58, score-0.137]
25 We presume a similar independence of MRP and ERD phenomena for imagined movements. [sent-59, score-0.104]
26 Apart from exploiting complementary information on cortical processes, combining MRP and ERD based features might give the benefit of being more robust against artifacts from non central nervous system (CNS) activity such as eye movement (EOG) or muscular artifacts (EMG). [sent-61, score-0.414]
27 MRPs, EMG activity is of more concern to oscillatory features, cf. [sent-64, score-0.141]
28 Accordingly, a classification method that is based on both features has better chance to handle trials that are contaminated by one kind of those artifacts. [sent-66, score-0.218]
29 On the other hand, it might increase the risk of using non-CNS activity for classification which would not be conform with the BCI idea, [1]. [sent-67, score-0.066]
30 In this paper we analyze EEG data from experiments with three subjects called aa, af and ak. [sent-71, score-0.144]
31 The subject sat in a normal chair, with arms lying relaxed on the table. [sent-72, score-0.111]
32 The subject was instructed to imagine performing left resp. [sent-76, score-0.111]
33 right hand finger movements as long as the symbol was visible. [sent-77, score-0.069]
34 200–300 trials were recorded for each class and each subject. [sent-78, score-0.148]
35 Brain activity was recorded with 28 (subject aa) resp. [sent-79, score-0.098]
36 52 (subjects af and ak) Ag/AgCl electrodes at 1000 Hz and downsampled to 100 Hz for the present offline study. [sent-80, score-0.07]
37 In these experiments the aim of classification is to discriminate ’left’ from ’right’ trials based on EEG-data during the whole period of imagination. [sent-84, score-0.116]
38 In the following we describe methods to derive feature vectors capturing MRP or ERD effects. [sent-90, score-0.135]
39 These values were used for both, classifying trials based on single-features and the combined classification. [sent-95, score-0.148]
40 Signals were baseline corrected on the interval 0–300 ms and downsampled by calculating five jumping means in several consecutive intervals beginning at 300 ms and ending between 1500–3500 ms. [sent-99, score-0.136]
41 5 Hz C3 lap C4 lap C3 lap C4 lap Figure 1: ERP and ERD (7–30 Hz) curves for subject aa in the time interval -500 ms to 3000 ms relative to stimulus. [sent-101, score-0.687]
42 To derive feature vectors for the ERD effects we use two different methods which may reflect different aspects of brain rhythm modulations. [sent-107, score-0.299]
43 The first (AR) reflects the spectral distribution of the most prominent brain rhythms whereas the second (CSP) reflects spatial patterns of most prominent power modulation in specifying frequency bands. [sent-108, score-0.227]
44 The feature vector of one trial is the concatenation of the AR coefficients plus the variance of each channel. [sent-112, score-0.16]
45 Accounting for this by adding the variance to the feature vector improves classification. [sent-114, score-0.101]
46 And to sharpen the spectral information to focal brain sources (spatial) Laplacian filters were applied. [sent-116, score-0.077]
47 The interval for estimating the AR parameters started at 500 ms and the end points were choosen between 2000 ms and 3500 ms. [sent-117, score-0.173]
48 This method was suggested for binary classification of EEG trials in [9]. [sent-119, score-0.116]
49 In features space projections on orientations with most differing power-ratios are used. [sent-120, score-0.102]
50 In weak features discriminative information is spread across many dimensions. [sent-138, score-0.102]
51 Classifying such features based on a small training set may lead to the well-known overfitting problem. [sent-139, score-0.102]
52 To avoid this, typically one of the following strategies is employed: (1) performing strong preprocessing to extract low dimensional feature vectors which are tractable for most classifiers. [sent-140, score-0.135]
53 Or (2) performing no or weak preprocessing and carefully regularizing the classifier such that high-dimensional features can be handled even with only a small training set. [sent-141, score-0.102]
54 A good introduction to regularized classification is [12] including regularized LDA which we used here. [sent-144, score-0.102]
55 The regularization coefficients were chosen by cross-validation together with the free parameters of the feature extraction methods, see section 2. [sent-147, score-0.205]
56 So in this off-line analysis where in each cross-validation procedure 100 different training sets are drawn randomly from the set of all trials one would have to do a cross-validation (for model selection, MS) within a cross-validation (for estimating the generalization error, GE). [sent-150, score-0.116]
57 On the other hand doing the model selection by cross-validation on all trials would could lead to overfitting and underestimating the generalization error. [sent-152, score-0.149]
58 As an intermediate way MS-cross-validation was performed on three subsets of all trials that were randomly drawn where the size of the subsets was the same as the size of the training sets in the GE-cross-validation, i. [sent-153, score-0.116]
59 Some differences in the quality of the features for classification are observable, but there is not one type of feature that is generally the best. [sent-162, score-0.203]
60 Only a small number of trials did fall in neither of those two categories (’ambivalent’) as could be expected due to the small standard deviation. [sent-167, score-0.116]
61 It is now interesting to see whether there are trials which are for one feature type in the well classified range and for the other feature in the badly classified part. [sent-168, score-0.318]
62 2 shows BP and CSP for subject af as example for each the part of the bad classified values which are good and bad classified in the other feature. [sent-170, score-0.363]
63 combined features this issue was not followed in further detail. [sent-176, score-0.134]
64 Typical approaches suggested are a winner-takes-all strategy, which cannot increase performance above the best single feature analysis, and concatenation of the single feature vectors, discussed as CONCAT below. [sent-185, score-0.202]
65 We propose two further methods that incorporate independence assumptions (PROB and to ak 17. [sent-187, score-0.111]
66 5 8% 82% 81% Figure 2: Left: Misclassification rates for single features classified with regularized LDA. [sent-205, score-0.153]
67 Free parameters of each feature extraction method were selected by cross-validation on subsets of all trials, see section 2. [sent-206, score-0.137]
68 Right: Pie charts show how ’MRP-bad’ and ’CSP-bad’ trials for subject af are classified based on the respective other feature: white is the portion of the trials which is ’good’ for the other feature, black marks ’bad’, and gray ’ambivalent’ trials for the other feature. [sent-208, score-0.529]
69 a smaller extend META) and allow individual decision boundary fitting to single features (META). [sent-210, score-0.141]
70 (CONCAT) In this simple approach of gathered evidence feature vectors are just concatenated. [sent-211, score-0.135]
71 Additionally, we tried classification with a linear programming machine (LPM), which is appealing for its sparse feature selection property, but it did not improve results compared to regularized LDA. [sent-213, score-0.152]
72 Here we derive the optimal classifier for combined feature vectors X = (X1 , . [sent-217, score-0.167]
73 , Xn ) under the additional assumption that individual features X1 , . [sent-220, score-0.102]
74 Denoting by Y (x) the decision function on feature space X ˆ Y (x) = ’R’ ⇔ P(Y = ’R’ | X = x) > P(Y = ’L’ | X = x) ⇔ fY =’R’ (x) P(Y = ’R’) > fY =’L’ (x) P(Y = ’L’), where Y is a random variable on the labels {’L’, ’R’} and f denotes densities. [sent-224, score-0.14]
75 There are two ways possible: Regularisation of the covariance matrices with one global parameter (PROBsame) or with three separately selected parameters corresponding to the single-type features (PROBdiff). [sent-230, score-0.102]
76 (META) In this approach a meta classifier is applied to the continuous output of individual classifiers that are trained on single features beforehand. [sent-231, score-0.316]
77 , if the decision boundary is linear for one feature and nonlinear for another. [sent-234, score-0.14]
78 Here we just use LDA for all features, but regularization coefficients are selected for each single feature individually. [sent-235, score-0.135]
79 Since the meta classifier acts on low (2 or 3) dimensional features further regularization is not needed, so we used unregularized LDA. [sent-236, score-0.35]
80 META extracts discriminative information from single features independently but the meta classification may exploit inter relations based on the output of the individual decision aa af ak mean Best Single 9. [sent-237, score-0.588]
81 of the means in 10×10-fold cross-validation for combined features compared to the most successful single-type feature. [sent-279, score-0.134]
82 3 Results Table 1 shows the results for the combined classification methods and for comparison the best result on single-type features (’Best Single’) from the table of Fig. [sent-283, score-0.134]
83 The CONCAT method performs only for subject ak better than the single feature methods. [sent-287, score-0.275]
84 And second, regularisation for the single features results in different regularisation parameters. [sent-290, score-0.324]
85 In CONCAT a single regularisation parameter has to be found. [sent-291, score-0.111]
86 In our case the regularisation parameters for subject aa for MRP are about 0. [sent-292, score-0.322]
87 From the other approaches the PROB methods are most successful, but META is very good, too, and better than the single feature results. [sent-295, score-0.101]
88 Concerning the results it is noteworthy that all subjects were BCI-untrained. [sent-297, score-0.074]
89 Only subject aa had experience as subject in EEG experiments. [sent-298, score-0.322]
90 The result obtained with single-features is in the range of the best results for untrained BCI performance with imagined movement paradigm, cf. [sent-299, score-0.193]
91 Whereas the result of less than 8 % error with our proposed combining approach for subject aa and af is better than for the 3 subjects in [17] in up to even 10 feedback sessions. [sent-301, score-0.459]
92 Additionally, it should be noted that the subject aa reported that he sometimes missed to react to the stimulus due to fatigue. [sent-303, score-0.211]
93 In contrast, the combination of features without any assumption of independence (CONCAT) did not improve accuracy in every case and always performs worse than PROB and META. [sent-308, score-0.15]
94 These results further support the hypothesis that MRP and ERD reflect independent aspects of brain activity. [sent-309, score-0.113]
95 Additionally, the combined approach has the practical advantage that no prior decision has to be made about what feature to use. [sent-311, score-0.172]
96 Combining features of different brain processes in feedback scenarios where the subject is trying to adapt to the feedback algorithm could in principle hold the risk of making the learning task too complex for the subject. [sent-312, score-0.416]
97 Finally, we would like to remark that the proposed feature combination principles can be used in other application areas where independent features can be obtained. [sent-314, score-0.203]
98 Hallett, “Event-related desynchronization and movement-related cortical potentials on the ECoG and EEG”, Electroencephalogr. [sent-381, score-0.238]
99 Pfurtscheller, “Optimal spatial filtering of single trial EEG during imagined hand movement”, IEEE Trans. [sent-401, score-0.148]
100 Pregenzer, “EEG-based discrimination between imagination of right and left hand movement”, Electroencephalogr. [sent-463, score-0.097]
wordName wordTfidf (topN-words)
[('erd', 0.342), ('mrp', 0.278), ('eeg', 0.269), ('bci', 0.254), ('meta', 0.214), ('csp', 0.204), ('classi', 0.198), ('mrps', 0.171), ('movement', 0.137), ('concat', 0.128), ('desynchronization', 0.128), ('ar', 0.117), ('trials', 0.116), ('motor', 0.114), ('regularisation', 0.111), ('subject', 0.111), ('prob', 0.111), ('emg', 0.107), ('features', 0.102), ('feature', 0.101), ('aa', 0.1), ('bad', 0.091), ('eog', 0.085), ('lap', 0.085), ('lda', 0.085), ('brain', 0.077), ('nger', 0.074), ('pfurtscheller', 0.074), ('rhythms', 0.074), ('sensorimotor', 0.074), ('cation', 0.074), ('subjects', 0.074), ('af', 0.07), ('ms', 0.068), ('hz', 0.066), ('activity', 0.066), ('ambivalent', 0.064), ('braincomputer', 0.064), ('contralateral', 0.064), ('imagination', 0.064), ('lateralized', 0.064), ('ak', 0.063), ('interface', 0.06), ('neurophysiological', 0.06), ('trial', 0.059), ('potentials', 0.058), ('combining', 0.057), ('accuracies', 0.056), ('imagined', 0.056), ('imagery', 0.056), ('signals', 0.054), ('cortical', 0.052), ('regularized', 0.051), ('rhythm', 0.051), ('ect', 0.05), ('independence', 0.048), ('feedback', 0.047), ('paradigms', 0.045), ('autoregressive', 0.045), ('patients', 0.045), ('babiloni', 0.043), ('fkz', 0.043), ('neuroimage', 0.043), ('pericentral', 0.043), ('probdiff', 0.043), ('probsame', 0.043), ('vaughan', 0.043), ('wn', 0.043), ('oscillatory', 0.043), ('berlin', 0.039), ('decision', 0.039), ('cortex', 0.038), ('prominent', 0.038), ('benjamin', 0.037), ('birbaumer', 0.037), ('choosen', 0.037), ('intentions', 0.037), ('mcfarland', 0.037), ('neuper', 0.037), ('strengthen', 0.037), ('wolpaw', 0.037), ('movements', 0.036), ('ller', 0.036), ('aspects', 0.036), ('extraction', 0.036), ('device', 0.035), ('vectors', 0.034), ('free', 0.034), ('cortices', 0.034), ('potsdam', 0.034), ('regularization', 0.034), ('onset', 0.033), ('slow', 0.033), ('hand', 0.033), ('xn', 0.033), ('combined', 0.032), ('cients', 0.032), ('processes', 0.032), ('recorded', 0.032), ('concern', 0.032)]
simIndex simValue paperId paperTitle
same-paper 1 1.0000004 55 nips-2002-Combining Features for BCI
Author: Guido Dornhege, Benjamin Blankertz, Gabriel Curio, Klaus-Robert Müller
Abstract: Recently, interest is growing to develop an effective communication interface connecting the human brain to a computer, the ’Brain-Computer Interface’ (BCI). One motivation of BCI research is to provide a new communication channel substituting normal motor output in patients with severe neuromuscular disabilities. In the last decade, various neurophysiological cortical processes, such as slow potential shifts, movement related potentials (MRPs) or event-related desynchronization (ERD) of spontaneous EEG rhythms, were shown to be suitable for BCI, and, consequently, different independent approaches of extracting BCI-relevant EEG-features for single-trial analysis are under investigation. Here, we present and systematically compare several concepts for combining such EEG-features to improve the single-trial classification. Feature combinations are evaluated on movement imagination experiments with 3 subjects where EEG-features are based on either MRPs or ERD, or both. Those combination methods that incorporate the assumption that the single EEG-features are physiologically mutually independent outperform the plain method of ’adding’ evidence where the single-feature vectors are simply concatenated. These results strengthen the hypothesis that MRP and ERD reflect at least partially independent aspects of cortical processes and open a new perspective to boost BCI effectiveness.
2 0.18494698 108 nips-2002-Improving Transfer Rates in Brain Computer Interfacing: A Case Study
Author: Peter Meinicke, Matthias Kaper, Florian Hoppe, Manfred Heumann, Helge Ritter
Abstract: In this paper we present results of a study on brain computer interfacing. We adopted an approach of Farwell & Donchin [4], which we tried to improve in several aspects. The main objective was to improve the transfer rates based on offline analysis of EEG-data but within a more realistic setup closer to an online realization than in the original studies. The objective was achieved along two different tracks: on the one hand we used state-of-the-art machine learning techniques for signal classification and on the other hand we augmented the data space by using more electrodes for the interface. For the classification task we utilized SVMs and, as motivated by recent findings on the learning of discriminative densities, we accumulated the values of the classification function in order to combine several classifications, which finally lead to significantly improved rates as compared with techniques applied in the original work. In combination with the data space augmentation, we achieved competitive transfer rates at an average of 50.5 bits/min and with a maximum of 84.7 bits/min.
3 0.14123635 88 nips-2002-Feature Selection and Classification on Matrix Data: From Large Margins to Small Covering Numbers
Author: Sepp Hochreiter, Klaus Obermayer
Abstract: We investigate the problem of learning a classification task for datasets which are described by matrices. Rows and columns of these matrices correspond to objects, where row and column objects may belong to different sets, and the entries in the matrix express the relationships between them. We interpret the matrix elements as being produced by an unknown kernel which operates on object pairs and we show that - under mild assumptions - these kernels correspond to dot products in some (unknown) feature space. Minimizing a bound for the generalization error of a linear classifier which has been obtained using covering numbers we derive an objective function for model selection according to the principle of structural risk minimization. The new objective function has the advantage that it allows the analysis of matrices which are not positive definite, and not even symmetric or square. We then consider the case that row objects are interpreted as features. We suggest an additional constraint, which imposes sparseness on the row objects and show, that the method can then be used for feature selection. Finally, we apply this method to data obtained from DNA microarrays, where “column” objects correspond to samples, “row” objects correspond to genes and matrix elements correspond to expression levels. Benchmarks are conducted using standard one-gene classification and support vector machines and K-nearest neighbors after standard feature selection. Our new method extracts a sparse set of genes and provides superior classification results. 1
4 0.13957152 187 nips-2002-Spikernels: Embedding Spiking Neurons in Inner-Product Spaces
Author: Lavi Shpigelman, Yoram Singer, Rony Paz, Eilon Vaadia
Abstract: Inner-product operators, often referred to as kernels in statistical learning, define a mapping from some input space into a feature space. The focus of this paper is the construction of biologically-motivated kernels for cortical activities. The kernels we derive, termed Spikernels, map spike count sequences into an abstract vector space in which we can perform various prediction tasks. We discuss in detail the derivation of Spikernels and describe an efficient algorithm for computing their value on any two sequences of neural population spike counts. We demonstrate the merits of our modeling approach using the Spikernel and various standard kernels for the task of predicting hand movement velocities from cortical recordings. In all of our experiments all the kernels we tested outperform the standard scalar product used in regression with the Spikernel consistently achieving the best performance. 1
5 0.13426951 21 nips-2002-Adaptive Classification by Variational Kalman Filtering
Author: Peter Sykacek, Stephen J. Roberts
Abstract: We propose in this paper a probabilistic approach for adaptive inference of generalized nonlinear classification that combines the computational advantage of a parametric solution with the flexibility of sequential sampling techniques. We regard the parameters of the classifier as latent states in a first order Markov process and propose an algorithm which can be regarded as variational generalization of standard Kalman filtering. The variational Kalman filter is based on two novel lower bounds that enable us to use a non-degenerate distribution over the adaptation rate. An extensive empirical evaluation demonstrates that the proposed method is capable of infering competitive classifiers both in stationary and non-stationary environments. Although we focus on classification, the algorithm is easily extended to other generalized nonlinear models.
6 0.11360613 59 nips-2002-Constraint Classification for Multiclass Classification and Ranking
7 0.1119616 196 nips-2002-The RA Scanner: Prediction of Rheumatoid Joint Inflammation Based on Laser Imaging
8 0.10749485 24 nips-2002-Adaptive Scaling for Feature Selection in SVMs
9 0.1069838 92 nips-2002-FloatBoost Learning for Classification
10 0.10399019 9 nips-2002-A Minimal Intervention Principle for Coordinated Movement
11 0.094803818 90 nips-2002-Feature Selection in Mixture-Based Clustering
12 0.090202764 45 nips-2002-Boosted Dyadic Kernel Discriminants
13 0.087057807 153 nips-2002-Neural Decoding of Cursor Motion Using a Kalman Filter
14 0.085069351 19 nips-2002-Adapting Codes and Embeddings for Polychotomies
15 0.085011475 68 nips-2002-Discriminative Densities from Maximum Contrast Estimation
16 0.077890255 89 nips-2002-Feature Selection by Maximum Marginal Diversity
17 0.074618421 132 nips-2002-Learning to Detect Natural Image Boundaries Using Brightness and Texture
18 0.072631128 72 nips-2002-Dyadic Classification Trees via Structural Risk Minimization
19 0.069887154 181 nips-2002-Self Supervised Boosting
20 0.069371924 86 nips-2002-Fast Sparse Gaussian Process Methods: The Informative Vector Machine
topicId topicWeight
[(0, -0.224), (1, -0.017), (2, 0.071), (3, -0.011), (4, 0.174), (5, -0.034), (6, -0.067), (7, -0.089), (8, 0.111), (9, 0.047), (10, -0.111), (11, 0.135), (12, 0.104), (13, 0.057), (14, 0.099), (15, -0.078), (16, -0.059), (17, -0.005), (18, -0.095), (19, -0.023), (20, -0.067), (21, -0.025), (22, 0.048), (23, -0.019), (24, -0.018), (25, 0.003), (26, 0.063), (27, -0.165), (28, 0.018), (29, -0.006), (30, -0.133), (31, -0.044), (32, 0.012), (33, 0.009), (34, 0.001), (35, -0.003), (36, 0.046), (37, -0.047), (38, -0.076), (39, 0.083), (40, 0.093), (41, 0.089), (42, 0.02), (43, -0.054), (44, -0.05), (45, -0.036), (46, 0.072), (47, 0.012), (48, 0.007), (49, -0.102)]
simIndex simValue paperId paperTitle
same-paper 1 0.92318529 55 nips-2002-Combining Features for BCI
Author: Guido Dornhege, Benjamin Blankertz, Gabriel Curio, Klaus-Robert Müller
Abstract: Recently, interest is growing to develop an effective communication interface connecting the human brain to a computer, the ’Brain-Computer Interface’ (BCI). One motivation of BCI research is to provide a new communication channel substituting normal motor output in patients with severe neuromuscular disabilities. In the last decade, various neurophysiological cortical processes, such as slow potential shifts, movement related potentials (MRPs) or event-related desynchronization (ERD) of spontaneous EEG rhythms, were shown to be suitable for BCI, and, consequently, different independent approaches of extracting BCI-relevant EEG-features for single-trial analysis are under investigation. Here, we present and systematically compare several concepts for combining such EEG-features to improve the single-trial classification. Feature combinations are evaluated on movement imagination experiments with 3 subjects where EEG-features are based on either MRPs or ERD, or both. Those combination methods that incorporate the assumption that the single EEG-features are physiologically mutually independent outperform the plain method of ’adding’ evidence where the single-feature vectors are simply concatenated. These results strengthen the hypothesis that MRP and ERD reflect at least partially independent aspects of cortical processes and open a new perspective to boost BCI effectiveness.
2 0.76560193 108 nips-2002-Improving Transfer Rates in Brain Computer Interfacing: A Case Study
Author: Peter Meinicke, Matthias Kaper, Florian Hoppe, Manfred Heumann, Helge Ritter
Abstract: In this paper we present results of a study on brain computer interfacing. We adopted an approach of Farwell & Donchin [4], which we tried to improve in several aspects. The main objective was to improve the transfer rates based on offline analysis of EEG-data but within a more realistic setup closer to an online realization than in the original studies. The objective was achieved along two different tracks: on the one hand we used state-of-the-art machine learning techniques for signal classification and on the other hand we augmented the data space by using more electrodes for the interface. For the classification task we utilized SVMs and, as motivated by recent findings on the learning of discriminative densities, we accumulated the values of the classification function in order to combine several classifications, which finally lead to significantly improved rates as compared with techniques applied in the original work. In combination with the data space augmentation, we achieved competitive transfer rates at an average of 50.5 bits/min and with a maximum of 84.7 bits/min.
3 0.6411413 196 nips-2002-The RA Scanner: Prediction of Rheumatoid Joint Inflammation Based on Laser Imaging
Author: Anton Schwaighofer, Volker Tresp, Peter Mayer, Alexander K. Scheel, Gerhard A. Müller
Abstract: We describe the RA scanner, a novel system for the examination of patients suffering from rheumatoid arthritis. The RA scanner is based on a novel laser-based imaging technique which is sensitive to the optical characteristics of finger joint tissue. Based on the laser images, finger joints are classified according to whether the inflammatory status has improved or worsened. To perform the classification task, various linear and kernel-based systems were implemented and their performances were compared. Special emphasis was put on measures to reliably perform parameter tuning and evaluation, since only a very small data set was available. Based on the results presented in this paper, it was concluded that the RA scanner permits a reliable classification of pathological finger joints, thus paving the way for a further development from prototype to product stage.
4 0.58804858 92 nips-2002-FloatBoost Learning for Classification
Author: Stan Z. Li, Zhenqiu Zhang, Heung-yeung Shum, Hongjiang Zhang
Abstract: AdaBoost [3] minimizes an upper error bound which is an exponential function of the margin on the training set [14]. However, the ultimate goal in applications of pattern classification is always minimum error rate. On the other hand, AdaBoost needs an effective procedure for learning weak classifiers, which by itself is difficult especially for high dimensional data. In this paper, we present a novel procedure, called FloatBoost, for learning a better boosted classifier. FloatBoost uses a backtrack mechanism after each iteration of AdaBoost to remove weak classifiers which cause higher error rates. The resulting float-boosted classifier consists of fewer weak classifiers yet achieves lower error rates than AdaBoost in both training and test. We also propose a statistical model for learning weak classifiers, based on a stagewise approximation of the posterior using an overcomplete set of scalar features. Experimental comparisons of FloatBoost and AdaBoost are provided through a difficult classification problem, face detection, where the goal is to learn from training examples a highly nonlinear classifier to differentiate between face and nonface patterns in a high dimensional space. The results clearly demonstrate the promises made by FloatBoost over AdaBoost.
5 0.5773052 153 nips-2002-Neural Decoding of Cursor Motion Using a Kalman Filter
Author: W Wu, M. J. Black, Y. Gao, M. Serruya, A. Shaikhouni, J. P. Donoghue, Elie Bienenstock
Abstract: The direct neural control of external devices such as computer displays or prosthetic limbs requires the accurate decoding of neural activity representing continuous movement. We develop a real-time control system using the spiking activity of approximately 40 neurons recorded with an electrode array implanted in the arm area of primary motor cortex. In contrast to previous work, we develop a control-theoretic approach that explicitly models the motion of the hand and the probabilistic relationship between this motion and the mean firing rates of the cells in 70 bins. We focus on a realistic cursor control task in which the subject must move a cursor to “hit” randomly placed targets on a computer monitor. Encoding and decoding of the neural data is achieved with a Kalman filter which has a number of advantages over previous linear filtering techniques. In particular, the Kalman filter reconstructions of hand trajectories in off-line experiments are more accurate than previously reported results and the model provides insights into the nature of the neural coding of movement. ¨ ©§
6 0.56361598 59 nips-2002-Constraint Classification for Multiclass Classification and Ranking
7 0.55597883 21 nips-2002-Adaptive Classification by Variational Kalman Filtering
8 0.54837847 88 nips-2002-Feature Selection and Classification on Matrix Data: From Large Margins to Small Covering Numbers
9 0.54462957 24 nips-2002-Adaptive Scaling for Feature Selection in SVMs
10 0.54012221 109 nips-2002-Improving a Page Classifier with Anchor Extraction and Link Analysis
11 0.52740324 68 nips-2002-Discriminative Densities from Maximum Contrast Estimation
12 0.5171833 187 nips-2002-Spikernels: Embedding Spiking Neurons in Inner-Product Spaces
13 0.50743866 9 nips-2002-A Minimal Intervention Principle for Coordinated Movement
14 0.48983112 62 nips-2002-Coulomb Classifiers: Generalizing Support Vector Machines via an Analogy to Electrostatic Systems
15 0.47608116 45 nips-2002-Boosted Dyadic Kernel Discriminants
16 0.46241102 132 nips-2002-Learning to Detect Natural Image Boundaries Using Brightness and Texture
17 0.45352533 89 nips-2002-Feature Selection by Maximum Marginal Diversity
18 0.44763404 86 nips-2002-Fast Sparse Gaussian Process Methods: The Informative Vector Machine
19 0.42422861 72 nips-2002-Dyadic Classification Trees via Structural Risk Minimization
20 0.42036325 81 nips-2002-Expected and Unexpected Uncertainty: ACh and NE in the Neocortex
topicId topicWeight
[(6, 0.302), (11, 0.016), (23, 0.101), (42, 0.045), (54, 0.096), (55, 0.057), (57, 0.019), (67, 0.023), (68, 0.032), (74, 0.098), (87, 0.02), (92, 0.029), (98, 0.093)]
simIndex simValue paperId paperTitle
1 0.84486735 177 nips-2002-Retinal Processing Emulation in a Programmable 2-Layer Analog Array Processor CMOS Chip
Author: R. Carmona, F. Jiménez-garrido, R. Dominguez-castro, S. Espejo, A. Rodriguez-vázquez
Abstract: A bio-inspired model for an analog programmable array processor (APAP), based on studies on the vertebrate retina, has permitted the realization of complex programmable spatio-temporal dynamics in VLSI. This model mimics the way in which images are processed in the visual pathway, rendering a feasible alternative for the implementation of early vision applications in standard technologies. A prototype chip has been designed and fabricated in a 0.5µm standard CMOS process. Computing power per area and power consumption is amongst the highest reported for a single chip. Design challenges, trade-offs and some experimental results are presented in this paper. 1
same-paper 2 0.78153586 55 nips-2002-Combining Features for BCI
Author: Guido Dornhege, Benjamin Blankertz, Gabriel Curio, Klaus-Robert Müller
Abstract: Recently, interest is growing to develop an effective communication interface connecting the human brain to a computer, the ’Brain-Computer Interface’ (BCI). One motivation of BCI research is to provide a new communication channel substituting normal motor output in patients with severe neuromuscular disabilities. In the last decade, various neurophysiological cortical processes, such as slow potential shifts, movement related potentials (MRPs) or event-related desynchronization (ERD) of spontaneous EEG rhythms, were shown to be suitable for BCI, and, consequently, different independent approaches of extracting BCI-relevant EEG-features for single-trial analysis are under investigation. Here, we present and systematically compare several concepts for combining such EEG-features to improve the single-trial classification. Feature combinations are evaluated on movement imagination experiments with 3 subjects where EEG-features are based on either MRPs or ERD, or both. Those combination methods that incorporate the assumption that the single EEG-features are physiologically mutually independent outperform the plain method of ’adding’ evidence where the single-feature vectors are simply concatenated. These results strengthen the hypothesis that MRP and ERD reflect at least partially independent aspects of cortical processes and open a new perspective to boost BCI effectiveness.
3 0.53957856 153 nips-2002-Neural Decoding of Cursor Motion Using a Kalman Filter
Author: W Wu, M. J. Black, Y. Gao, M. Serruya, A. Shaikhouni, J. P. Donoghue, Elie Bienenstock
Abstract: The direct neural control of external devices such as computer displays or prosthetic limbs requires the accurate decoding of neural activity representing continuous movement. We develop a real-time control system using the spiking activity of approximately 40 neurons recorded with an electrode array implanted in the arm area of primary motor cortex. In contrast to previous work, we develop a control-theoretic approach that explicitly models the motion of the hand and the probabilistic relationship between this motion and the mean firing rates of the cells in 70 bins. We focus on a realistic cursor control task in which the subject must move a cursor to “hit” randomly placed targets on a computer monitor. Encoding and decoding of the neural data is achieved with a Kalman filter which has a number of advantages over previous linear filtering techniques. In particular, the Kalman filter reconstructions of hand trajectories in off-line experiments are more accurate than previously reported results and the model provides insights into the nature of the neural coding of movement. ¨ ©§
4 0.53579706 108 nips-2002-Improving Transfer Rates in Brain Computer Interfacing: A Case Study
Author: Peter Meinicke, Matthias Kaper, Florian Hoppe, Manfred Heumann, Helge Ritter
Abstract: In this paper we present results of a study on brain computer interfacing. We adopted an approach of Farwell & Donchin [4], which we tried to improve in several aspects. The main objective was to improve the transfer rates based on offline analysis of EEG-data but within a more realistic setup closer to an online realization than in the original studies. The objective was achieved along two different tracks: on the one hand we used state-of-the-art machine learning techniques for signal classification and on the other hand we augmented the data space by using more electrodes for the interface. For the classification task we utilized SVMs and, as motivated by recent findings on the learning of discriminative densities, we accumulated the values of the classification function in order to combine several classifications, which finally lead to significantly improved rates as compared with techniques applied in the original work. In combination with the data space augmentation, we achieved competitive transfer rates at an average of 50.5 bits/min and with a maximum of 84.7 bits/min.
5 0.5213744 26 nips-2002-An Estimation-Theoretic Framework for the Presentation of Multiple Stimuli
Author: Christian W. Eurich
Abstract: A framework is introduced for assessing the encoding accuracy and the discriminational ability of a population of neurons upon simultaneous presentation of multiple stimuli. Minimal square estimation errors are obtained from a Fisher information analysis in an abstract compound space comprising the features of all stimuli. Even for the simplest case of linear superposition of responses and Gaussian tuning, the symmetries in the compound space are very different from those in the case of a single stimulus. The analysis allows for a quantitative description of attentional effects and can be extended to include neural nonlinearities such as nonclassical receptive fields. 1
6 0.51852393 33 nips-2002-Approximate Linear Programming for Average-Cost Dynamic Programming
7 0.51823902 44 nips-2002-Binary Tuning is Optimal for Neural Rate Coding with High Temporal Resolution
8 0.51519406 148 nips-2002-Morton-Style Factorial Coding of Color in Primary Visual Cortex
9 0.51384079 28 nips-2002-An Information Theoretic Approach to the Functional Classification of Neurons
10 0.51288176 10 nips-2002-A Model for Learning Variance Components of Natural Images
11 0.51273006 3 nips-2002-A Convergent Form of Approximate Policy Iteration
12 0.51241863 5 nips-2002-A Digital Antennal Lobe for Pattern Equalization: Analysis and Design
13 0.51104057 141 nips-2002-Maximally Informative Dimensions: Analyzing Neural Responses to Natural Signals
14 0.51069123 123 nips-2002-Learning Attractor Landscapes for Learning Motor Primitives
15 0.51043457 23 nips-2002-Adaptive Quantization and Density Estimation in Silicon
16 0.50934219 2 nips-2002-A Bilinear Model for Sparse Coding
17 0.50703168 93 nips-2002-Forward-Decoding Kernel-Based Phone Recognition
18 0.5056597 132 nips-2002-Learning to Detect Natural Image Boundaries Using Brightness and Texture
19 0.50352436 68 nips-2002-Discriminative Densities from Maximum Contrast Estimation
20 0.50331843 155 nips-2002-Nonparametric Representation of Policies and Value Functions: A Trajectory-Based Approach