nips nips2010 nips2010-121 knowledge-graph by maker-knowledge-mining

121 nips-2010-Improving Human Judgments by Decontaminating Sequential Dependencies


Source: pdf

Author: Harold Pashler, Matthew Wilder, Robert Lindsey, Matt Jones, Michael C. Mozer, Michael P. Holmes

Abstract: For over half a century, psychologists have been struck by how poor people are at expressing their internal sensations, impressions, and evaluations via rating scales. When individuals make judgments, they are incapable of using an absolute rating scale, and instead rely on reference points from recent experience. This relativity of judgment limits the usefulness of responses provided by individuals to surveys, questionnaires, and evaluation forms. Fortunately, the cognitive processes that transform internal states to responses are not simply noisy, but rather are influenced by recent experience in a lawful manner. We explore techniques to remove sequential dependencies, and thereby decontaminate a series of ratings to obtain more meaningful human judgments. In our formulation, decontamination is fundamentally a problem of inferring latent states (internal sensations) which, because of the relativity of judgment, have temporal dependencies. We propose a decontamination solution using a conditional random field with constraints motivated by psychological theories of relative judgment. Our exploration of decontamination models is supported by two experiments we conducted to obtain ground-truth rating data on a simple length estimation task. Our decontamination techniques yield an over 20% reduction in the error of human judgments. 1

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 of Psychological and Brain Sciences, Indiana University Abstract For over half a century, psychologists have been struck by how poor people are at expressing their internal sensations, impressions, and evaluations via rating scales. [sent-9, score-0.177]

2 When individuals make judgments, they are incapable of using an absolute rating scale, and instead rely on reference points from recent experience. [sent-10, score-0.296]

3 This relativity of judgment limits the usefulness of responses provided by individuals to surveys, questionnaires, and evaluation forms. [sent-11, score-0.345]

4 We explore techniques to remove sequential dependencies, and thereby decontaminate a series of ratings to obtain more meaningful human judgments. [sent-13, score-0.216]

5 In our formulation, decontamination is fundamentally a problem of inferring latent states (internal sensations) which, because of the relativity of judgment, have temporal dependencies. [sent-14, score-0.318]

6 We propose a decontamination solution using a conditional random field with constraints motivated by psychological theories of relative judgment. [sent-15, score-0.383]

7 Our exploration of decontamination models is supported by two experiments we conducted to obtain ground-truth rating data on a simple length estimation task. [sent-16, score-0.386]

8 Our decontamination techniques yield an over 20% reduction in the error of human judgments. [sent-17, score-0.296]

9 1 Introduction Suppose you are asked to make a series of moral judgments by rating, on a 1–10 scale, various actions, with a rating of 1 indicating ’not particularly bad or wrong’ and a rating of 10 indicating ’extremely evil. [sent-18, score-0.377]

10 Even though individuals are asked to make absolute judgments, the mean rating of statement (3) in the first context is reliably higher than the mean rating of the identical statement (3 ) in the second context (Parducci, 1968). [sent-21, score-0.417]

11 The classic explanation of this phenomenon is cast in terms of anchoring or primacy: information presented early in time serves as a basis for making judgments later in time (Tversky & Kahneman, 1974). [sent-22, score-0.149]

12 In the Netflix contest, significant attention was paid to anchoring effects by considering that an individual who gives high ratings early in a session is likely to be biased toward higher ratings later in a session (Koren, August 2009; Ellenberg, March 2008). [sent-23, score-0.301]

13 The need for anchors comes from the fact that individuals are poor at or incapable of making absolute judgments and instead must rely on reference points to make relative judgments (e. [sent-24, score-0.446]

14 There is a rich literature in experimental and theoretical psychology exploring sequential 1 dependencies suggesting that reference points change from one trial to the next in a systematic manner. [sent-28, score-0.302]

15 (We use the psychological jargon ‘trial’ to refer to a single judgment or rating in a series. [sent-29, score-0.301]

16 However, the most carefully controlled laboratory studies of sequential dependencies, dating back to the the 1950’s (discussed by Miller, 1956), involve the rating of unidimensional stimuli, such as the loudness of a tone or the length of a line. [sent-35, score-0.217]

17 Human performance at rating stimuli is surprisingly poor compared to an individual’s ability to discriminate the same stimuli. [sent-36, score-0.21]

18 Regardless of the domain, responses convey not much more than 2 bits of mutual information with the stimulus (Stewart et al. [sent-37, score-0.153]

19 Different types of judgment tasks have been studied including absolute identification, in which the individual’s task is to specify the distinct stimulus level (e. [sent-39, score-0.237]

20 , 10 levels of loudness), magnitude estimation, in which the task is to estimate the magnitude of a stimulus which may vary continuously along a dimension, and categorization which is a hybrid task requiring individuals to label stimuli by range. [sent-41, score-0.31]

21 Because the number of responses in absolute identification and categorization tasks is often quite large, and because individuals are often not aware of the discreteness of stimuli in absolute identification tasks, there isn’t a qualitative difference among tasks. [sent-42, score-0.379]

22 Without feedback, there are no explicit anchors against which stimuli can be assessed. [sent-44, score-0.144]

23 Typically, experimental trial t, trial t − 1 has a large influence on ratings, and trials t − 2, t − 3, etc. [sent-46, score-0.208]

24 The influence of recent trials is exerted by both the stimuli and responses, a fact which makes sense in light of the assumption that individuals form their response on the current trial by analogy to recent trials (i. [sent-48, score-0.428]

25 , they determine a response to the current stimulus that has the same relationship as the previous response had to the previous stimulus). [sent-50, score-0.261]

26 Both assimilation and contrast effects occur: an assimilative response on trial t occurs when the response moves in the direction of the stimulus or response on trial t − k; a contrastive response is one that moves away. [sent-51, score-0.755]

27 Interpreting recency effects in terms of assimilation and contrast is nontrivial and theory dependent (DeCarlo & Cross, 1990). [sent-52, score-0.157]

28 Many mathematical models have been developed to explain the phenomena of sequential effects in judgment tasks. [sent-53, score-0.255]

29 All adopt the assumption that the transduction of a stimulus to its internal representation is veridical. [sent-54, score-0.122]

30 ) Sequential dependencies and other corruptions of the representation occur in the mapping of the sensation to a response. [sent-57, score-0.262]

31 Other theories assume that multiple sensation-response anchors are required, one fixed and unchanging and another varying from trial to trial (e. [sent-62, score-0.29]

32 And in categorization and absolute identification tasks, some theories posit anchors for each distinct response, which are adjusted trial-to-trial (e. [sent-65, score-0.183]

33 Range-frequency theory (Parducci, 1965) claims that sequential effects arise because the sensation-response mapping is adjusted to utilize the full response range, and to produce roughly an equal number of responses of each type. [sent-68, score-0.282]

34 Because recent history interacts with the current stimulus to determine an individual’s response, responses have a complex relationship with the underlying sensation, and do not provide as much information about the internal state of the individual as one would hope. [sent-70, score-0.188]

35 In contrast, our approach to extracting more information from human judgments is to develop automatic techniques that recover the underlying sensation from a response that has been contaminated 2 by cognitive processes producing the response. [sent-72, score-0.437]

36 2 Experiments To collect ground-truth data for use in the design of decontamination techniques, we conducted two behavioral experiments using stimuli whose magnitudes could be objectively determined. [sent-77, score-0.339]

37 In both experiments, participants were asked to judge the horizontal gap between two vertically aligned dots on a computer monitor. [sent-78, score-0.195]

38 Participants were asked to respond to each dot pair using a 10-point rating scale, with 1 corresponding to the smallest gap they would see, and 10 corresponding to the largest. [sent-80, score-0.2]

39 They were not told that only 10 unique stimuli were presented, and were likely unaware of this fact (memory of exact absolute gaps is too poor), and thus the task is indistinguishable from a magnitude estimation or categorization task in which the gap varied continuously. [sent-83, score-0.27]

40 During the practice block, participants were shown every one of the ten gaps in random order, and simultaneous with the stimulus they were told—via text on the screen below the dots—the correct classification. [sent-85, score-0.231]

41 Although the psychology literature is replete with line-length judgment studies (two recent examples: Lacouture, 1997; Petrov & Anderson, 2005), the vast majority provide feedback to participants on at least some trials beyond the practice block. [sent-87, score-0.307]

42 Within a block, the trial sequence was arranged such that each gap was preceded exactly once by each other gap, with the exception that no repetitions occurred. [sent-94, score-0.139]

43 The main reason for conducting Experiment 2 was that we found the gaps used in Experiment 1 resulted in low error rates and few sequential effects for the smaller gaps. [sent-108, score-0.209]

44 Two participants in Experiment 1 and one participant in Experiment 2 were excluded from data analysis because their accuracy was below 20%. [sent-112, score-0.127]

45 3 error as a function of S(t−1) and S(t) error as a function of stimulus difference 1 error as a function of lagged stimulus −0. [sent-122, score-0.246]

46 The variation along the abscissa reflects sequential dependencies: assimilation is indicated by pairs of points with positive slopes (larger values of St−1 result in larger Rt ), and contrast is indicated by negative slopes. [sent-150, score-0.161]

47 The middle column shows another depiction of sequential dependencies by characterizing the distribution of errors (Rt − St ∈ {> 1, 1, 0, −1, < −1}) as a function of St − St−1 . [sent-152, score-0.12]

48 The predominance of assimilative responses is reflected in more Rt > St responses when St − St−1 < 0, and vice-versa. [sent-153, score-0.156]

49 The rightmost column presents the lag profile that characterizes how the stimulus on trial t − k for k = 1. [sent-154, score-0.202]

50 For the purpose of the current work, most relevant is that sequential dependencies in this task may stretch back two or three trials. [sent-159, score-0.12]

51 3 Approaches To Decontamination From a machine learning perspective, decontamination can be formulated in at least three different ways. [sent-160, score-0.247]

52 First, it could be considered an unsupervised infomax problem of determining a sensation associated with each distinct stimulus such that the sensation sequence has high mutual information with the response sequence. [sent-161, score-0.608]

53 Third, decontamination models could be built based on ground-truth data for one group of individuals and then tested on another group. [sent-164, score-0.371]

54 Formally, the decontamination problem involves inferring the sequence of (unobserved) sensations p given the complete response sequence. [sent-166, score-0.44]

55 To introduce some notation, let Rt1 ,t2 denote the sequence of responses made by participant p on trials t1 through t2 when shown a sequence of stimuli that 4 p evoke the sensation sequence St1 ,t2 . [sent-167, score-0.443]

56 Although psychological theories of human judgment address an altogether different problem—that p p p of predicting Rt , the response on trial t, given S1,t and R1,t−1 —they can inspire decontamination techniques. [sent-169, score-0.69]

57 Two classes of psychological theories correspond to two distinct function approximation techniques. [sent-170, score-0.136]

58 In contrast, other models favor highly flexible, nonlinear approaches that allow for similarity-based assimilation and contrast, and independent representations for each response label (e. [sent-172, score-0.17]

59 Given the discrete stimuli and responses, a lookup table seems the most general characterization of these models. [sent-175, score-0.226]

60 The first dimension of this space is the model class: regression, lookup table, or an additive hybrid. [sent-177, score-0.134]

61 Similarly, we define our lookup table LUTt (m, n) to produce an estimate of St by indexing over the m responses Rt−m+1,t and the n sensations St−n,t−1 . [sent-179, score-0.306]

62 Finally, we define an additive hybrid, REG⊕LUT(m, n) by first constructing a regression model, and then building a lookup table on the residual error, St − REGt (m, n). [sent-180, score-0.164]

63 The motivation for the hybrid is the complementarity of the two models, the regression model capturing linear regularities and the lookup table representing arbitrary nonlinear relationships. [sent-181, score-0.164]

64 The second dimension in our space of decontamination techniques specifies how inference is handled. [sent-182, score-0.247]

65 To utilize any of the models above for n > 0, sensations St−n,t−1 must be estimated. [sent-184, score-0.127]

66 As an alternative to the conditional random field (hereafter, CRF), we also consider a simple approach in which we simply set n = 0 and discard the sensation terms in our regression and lookup tables. [sent-187, score-0.381]

67 At the other extreme, we can assume an oracle that provides St−n,t−1 ; this oracle approach offers an upper bound on achievable performance. [sent-188, score-0.208]

68 The difference is minor because the stimulus and sensation are in one-to-one correspondence. [sent-201, score-0.304]

69 (Remember that lookup table values are indexed by St−1 , and therefore cannot be folded into the normalization constant. [sent-207, score-0.134]

70 Having now described a 3 × 3 space of decontamination approaches, we turn to the details of our decontamination experiments. [sent-209, score-0.494]

71 1 Debiasing and Decompressing Although our focus is on decontaminating sequential dependencies, or desequencing, the quality of human judgments can be reduced by at least three other factors. [sent-211, score-0.232]

72 Second, individuals may show compression, possibly nonlinear, of the response range. [sent-213, score-0.19]

73 For example, compression will be a natural consequence of assimilation because the endpoints of the response scale will move toward the center. [sent-216, score-0.19]

74 In the data from our two experiments, we found no evidence of drift, as determined by the fact that regression models with moving averages of the responses did not improve predictions. [sent-218, score-0.117]

75 For example, in Experiment 1, the shortest stimuli reported as G1 and G2 with high accuracy, but the longest stimuli tended to be underestimated by all participants. [sent-222, score-0.184]

76 The LUT(1, 0) compensates for this compression by associating responses G8 and G9 with higher sensation levels if the table entries are filled based on the training data according to: LUTt (1, 0) ≡ E[St |Rt ]. [sent-223, score-0.324]

77 All of the higher order lookup tables, LUT (m, n), for m ≥ 1 and n ≥ 0, will also perform nonlinear decompression in the same manner. [sent-224, score-0.216]

78 To debias the data, p ¯ we compute the mean response of a particular participant p, Rp ≡ 1/T Rt , and ensure the means p p ¯ p = St − S p . [sent-227, score-0.186]

79 Assuming that the mean sensation is ¯ are homogeneous via the constraint Rt − R identical for all participants—as it should be in our experiments—debiasing can be incorporated p ¯ into the lookup tables by storing not E[St |Rt . [sent-228, score-0.374]

80 ], and recovering the p ¯ sensation for a particular individual using LUT(m, n) − R . [sent-234, score-0.217]

81 (This trick is necessary to index into the lookup table with discrete response levels. [sent-235, score-0.221]

82 Note that this extra term—whether in the lookup table retrieval or the regression— ¯ results in additional features involving combinations of Rp and St , St−1 , and LUT(m, n) being added to the three CRF models. [sent-238, score-0.134]

83 The SIMPLE - REG⊕LUT and ORACLE - REG⊕LUT models are trained first by obtaining the regression coefficients, and then filling lookup table entries p with the expected residual, E[St − REGp |Rt , Rt−1 , . [sent-251, score-0.185]

84 99 1 sensation reconstruction error (RMSE) Figure 2: Results from Experiment 1 (left column) and Experiment 2 (right column). [sent-278, score-0.241]

85 The lookup tables used in the CRF - LUT and CRF - REG⊕LUT are the same as those in the ORACLE - LUT and ORACLE - REG⊕LUT models. [sent-282, score-0.157]

86 We tested models in which the sensation and/or response values are log transformed, because sensory transduction introduces logarithmic compression. [sent-289, score-0.325]

87 4 Results Figure 2 shows the root mean squared error (RMSE) between the ground-truth sensation and the model-estimated sensation over the set of validation subjects for 100 different splits of the data. [sent-300, score-0.458]

88 The difference between each pair of these results is highly reliable, indicating that bias, compression, and recency effects all contribute to the contamination of human judgments. [sent-303, score-0.12]

89 7 The reduction of error due to debiasing is 14. [sent-304, score-0.13]

90 Indeed models like CRF - REG⊕LUT perform nearly as well even without separate debiasing and decompression corrections. [sent-314, score-0.209]

91 The joint model REG⊕LUT that exploits both the regularity of the regression model and the flexibility of the lookup table clearly works better than either REG or LUT in isolation. [sent-316, score-0.164]

92 We do not have a good explanation for the advantage of SIMPLE - LUT over CRF - LUT in Experiment 1, although there are some minor differences in how the lookup tables for the two models are constructed, and we are investigating whether those differences might be responsible. [sent-318, score-0.178]

93 5 Discussion Psychologists have long been struck by the relativity of human judgments and have noted that relativity limits how well individuals can communicate their internal sensations, impressions, and evaluations via rating scales. [sent-320, score-0.555]

94 We’ve shown that decontamination techniques can improve the quality of judgments, reducing error by over 20% Is a 20% reduction significant? [sent-321, score-0.271]

95 Using the models we developed for this study, we can obtain a decontamination of the ratings and identify pairs of paintings where the participant’s ratings conflict with the decontaminated impressions. [sent-327, score-0.43]

96 Via a later session in which we ask participants for pairwise preferences, we can determine whether the decontaminator or the raw ratings are more reliable. [sent-328, score-0.19]

97 Indeed, it seems that if even responses to simple visual stimuli are contaminated, responses to more complex stimuli with a more complex judgment task will be even more vulnerable. [sent-330, score-0.421]

98 One such hint is the finding that systematic effects of sequences have been observed on response latencies in judgment tasks (Lacouture, 1997); therefore, latencies may prove useful for decontamination. [sent-336, score-0.246]

99 Bow, range, and sequential effects in absolute identification: A response-time analysis. [sent-373, score-0.174]

100 The dynamics of scaling: A memory-based anchor model of category rating and identification. [sent-414, score-0.118]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('lut', 0.554), ('reg', 0.363), ('crf', 0.258), ('decontamination', 0.247), ('st', 0.225), ('sensation', 0.217), ('lookup', 0.134), ('rating', 0.118), ('judgments', 0.108), ('rt', 0.107), ('debiasing', 0.106), ('sensations', 0.106), ('judgment', 0.105), ('oracle', 0.104), ('individuals', 0.103), ('stimuli', 0.092), ('trial', 0.09), ('stimulus', 0.087), ('participants', 0.087), ('response', 0.087), ('decompression', 0.082), ('lutt', 0.082), ('parducci', 0.082), ('ratings', 0.081), ('psychological', 0.078), ('sequential', 0.075), ('relativity', 0.071), ('responses', 0.066), ('ix', 0.063), ('psychology', 0.062), ('assimilation', 0.062), ('debias', 0.059), ('theories', 0.058), ('effects', 0.054), ('anchors', 0.052), ('gap', 0.049), ('experiment', 0.049), ('decarlo', 0.047), ('decompress', 0.047), ('stewart', 0.047), ('net', 0.047), ('dependencies', 0.045), ('absolute', 0.045), ('petrov', 0.041), ('anchoring', 0.041), ('recency', 0.041), ('compression', 0.041), ('participant', 0.04), ('mccallum', 0.037), ('decontaminate', 0.035), ('desequencing', 0.035), ('ellenberg', 0.035), ('lacouture', 0.035), ('mumma', 0.035), ('regt', 0.035), ('internal', 0.035), ('gaps', 0.034), ('rmse', 0.034), ('asked', 0.033), ('anderson', 0.031), ('regression', 0.03), ('reference', 0.03), ('brains', 0.029), ('categorization', 0.028), ('bar', 0.028), ('trials', 0.028), ('dots', 0.026), ('lag', 0.025), ('feedback', 0.025), ('human', 0.025), ('rp', 0.025), ('wilson', 0.024), ('error', 0.024), ('block', 0.024), ('abscissa', 0.024), ('assimilative', 0.024), ('barking', 0.024), ('chater', 0.024), ('decompressing', 0.024), ('decontaminating', 0.024), ('einhorn', 0.024), ('furnham', 0.024), ('hogarth', 0.024), ('impressions', 0.024), ('laming', 0.024), ('loudness', 0.024), ('outsmart', 0.024), ('poisoning', 0.024), ('portal', 0.024), ('psychologist', 0.024), ('struck', 0.024), ('wedell', 0.024), ('screen', 0.023), ('tables', 0.023), ('told', 0.022), ('conducting', 0.022), ('session', 0.022), ('models', 0.021), ('identi', 0.021), ('drift', 0.021)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000006 121 nips-2010-Improving Human Judgments by Decontaminating Sequential Dependencies

Author: Harold Pashler, Matthew Wilder, Robert Lindsey, Matt Jones, Michael C. Mozer, Michael P. Holmes

Abstract: For over half a century, psychologists have been struck by how poor people are at expressing their internal sensations, impressions, and evaluations via rating scales. When individuals make judgments, they are incapable of using an absolute rating scale, and instead rely on reference points from recent experience. This relativity of judgment limits the usefulness of responses provided by individuals to surveys, questionnaires, and evaluation forms. Fortunately, the cognitive processes that transform internal states to responses are not simply noisy, but rather are influenced by recent experience in a lawful manner. We explore techniques to remove sequential dependencies, and thereby decontaminate a series of ratings to obtain more meaningful human judgments. In our formulation, decontamination is fundamentally a problem of inferring latent states (internal sensations) which, because of the relativity of judgment, have temporal dependencies. We propose a decontamination solution using a conditional random field with constraints motivated by psychological theories of relative judgment. Our exploration of decontamination models is supported by two experiments we conducted to obtain ground-truth rating data on a simple length estimation task. Our decontamination techniques yield an over 20% reduction in the error of human judgments. 1

2 0.13967353 83 nips-2010-Evidence-Specific Structures for Rich Tractable CRFs

Author: Anton Chechetka, Carlos Guestrin

Abstract: We present a simple and effective approach to learning tractable conditional random fields with structure that depends on the evidence. Our approach retains the advantages of tractable discriminative models, namely efficient exact inference and arbitrarily accurate parameter learning in polynomial time. At the same time, our algorithm does not suffer a large expressive power penalty inherent to fixed tractable structures. On real-life relational datasets, our approach matches or exceeds state of the art accuracy of the dense models, and at the same time provides an order of magnitude speedup. 1

3 0.097974822 66 nips-2010-Double Q-learning

Author: Hado V. Hasselt

Abstract: In some stochastic environments the well-known reinforcement learning algorithm Q-learning performs very poorly. This poor performance is caused by large overestimations of action values. These overestimations result from a positive bias that is introduced because Q-learning uses the maximum action value as an approximation for the maximum expected action value. We introduce an alternative way to approximate the maximum expected value for any set of random variables. The obtained double estimator method is shown to sometimes underestimate rather than overestimate the maximum expected value. We apply the double estimator to Q-learning to construct Double Q-learning, a new off-policy reinforcement learning algorithm. We show the new algorithm converges to the optimal policy and that it performs well in some settings in which Q-learning performs poorly due to its overestimation. 1

4 0.095662802 254 nips-2010-Stability Approach to Regularization Selection (StARS) for High Dimensional Graphical Models

Author: Han Liu, Kathryn Roeder, Larry Wasserman

Abstract: A challenging problem in estimating high-dimensional graphical models is to choose the regularization parameter in a data-dependent way. The standard techniques include K-fold cross-validation (K-CV), Akaike information criterion (AIC), and Bayesian information criterion (BIC). Though these methods work well for low-dimensional problems, they are not suitable in high dimensional settings. In this paper, we present StARS: a new stability-based method for choosing the regularization parameter in high dimensional inference for undirected graphs. The method has a clear interpretation: we use the least amount of regularization that simultaneously makes a graph sparse and replicable under random sampling. This interpretation requires essentially no conditions. Under mild conditions, we show that StARS is partially sparsistent in terms of graph estimation: i.e. with high probability, all the true edges will be included in the selected model even when the graph size diverges with the sample size. Empirically, the performance of StARS is compared with the state-of-the-art model selection procedures, including K-CV, AIC, and BIC, on both synthetic data and a real microarray dataset. StARS outperforms all these competing procedures.

5 0.072686948 1 nips-2010-(RF)^2 -- Random Forest Random Field

Author: Nadia Payet, Sinisa Todorovic

Abstract: We combine random forest (RF) and conditional random field (CRF) into a new computational framework, called random forest random field (RF)2 . Inference of (RF)2 uses the Swendsen-Wang cut algorithm, characterized by MetropolisHastings jumps. A jump from one state to another depends on the ratio of the proposal distributions, and on the ratio of the posterior distributions of the two states. Prior work typically resorts to a parametric estimation of these four distributions, and then computes their ratio. Our key idea is to instead directly estimate these ratios using RF. RF collects in leaf nodes of each decision tree the class histograms of training examples. We use these class histograms for a nonparametric estimation of the distribution ratios. We derive the theoretical error bounds of a two-class (RF)2 . (RF)2 is applied to a challenging task of multiclass object recognition and segmentation over a random field of input image regions. In our empirical evaluation, we use only the visual information provided by image regions (e.g., color, texture, spatial layout), whereas the competing methods additionally use higher-level cues about the horizon location and 3D layout of surfaces in the scene. Nevertheless, (RF)2 outperforms the state of the art on benchmark datasets, in terms of accuracy and computation time.

6 0.063543409 155 nips-2010-Learning the context of a category

7 0.057633512 19 nips-2010-A rational decision making framework for inhibitory control

8 0.055836115 152 nips-2010-Learning from Logged Implicit Exploration Data

9 0.05084433 196 nips-2010-Online Markov Decision Processes under Bandit Feedback

10 0.050605789 130 nips-2010-Interval Estimation for Reinforcement-Learning Algorithms in Continuous-State Domains

11 0.050318923 98 nips-2010-Functional form of motion priors in human motion perception

12 0.05005626 21 nips-2010-Accounting for network effects in neuronal responses using L1 regularized point process models

13 0.047753971 127 nips-2010-Inferring Stimulus Selectivity from the Spatial Structure of Neural Network Dynamics

14 0.046719566 179 nips-2010-Natural Policy Gradient Methods with Parameter-based Exploration for Control Tasks

15 0.04583424 153 nips-2010-Learning invariant features using the Transformed Indian Buffet Process

16 0.045233257 114 nips-2010-Humans Learn Using Manifolds, Reluctantly

17 0.044520933 48 nips-2010-Collaborative Filtering in a Non-Uniform World: Learning with the Weighted Trace Norm

18 0.042575538 20 nips-2010-A unified model of short-range and long-range motion perception

19 0.042420346 268 nips-2010-The Neural Costs of Optimal Control

20 0.041736186 119 nips-2010-Implicit encoding of prior probabilities in optimal neural populations


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.112), (1, -0.008), (2, -0.05), (3, 0.032), (4, -0.017), (5, -0.003), (6, -0.045), (7, 0.006), (8, 0.008), (9, 0.033), (10, -0.04), (11, -0.034), (12, 0.017), (13, 0.004), (14, 0.058), (15, 0.012), (16, -0.044), (17, -0.038), (18, -0.039), (19, 0.101), (20, -0.092), (21, 0.026), (22, 0.1), (23, -0.012), (24, -0.007), (25, 0.036), (26, -0.033), (27, 0.04), (28, -0.047), (29, 0.052), (30, -0.06), (31, -0.11), (32, -0.162), (33, -0.028), (34, -0.078), (35, 0.117), (36, 0.016), (37, -0.07), (38, -0.007), (39, -0.04), (40, 0.119), (41, -0.003), (42, 0.005), (43, 0.066), (44, -0.014), (45, 0.047), (46, 0.041), (47, 0.043), (48, -0.001), (49, 0.034)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.93945271 121 nips-2010-Improving Human Judgments by Decontaminating Sequential Dependencies

Author: Harold Pashler, Matthew Wilder, Robert Lindsey, Matt Jones, Michael C. Mozer, Michael P. Holmes

Abstract: For over half a century, psychologists have been struck by how poor people are at expressing their internal sensations, impressions, and evaluations via rating scales. When individuals make judgments, they are incapable of using an absolute rating scale, and instead rely on reference points from recent experience. This relativity of judgment limits the usefulness of responses provided by individuals to surveys, questionnaires, and evaluation forms. Fortunately, the cognitive processes that transform internal states to responses are not simply noisy, but rather are influenced by recent experience in a lawful manner. We explore techniques to remove sequential dependencies, and thereby decontaminate a series of ratings to obtain more meaningful human judgments. In our formulation, decontamination is fundamentally a problem of inferring latent states (internal sensations) which, because of the relativity of judgment, have temporal dependencies. We propose a decontamination solution using a conditional random field with constraints motivated by psychological theories of relative judgment. Our exploration of decontamination models is supported by two experiments we conducted to obtain ground-truth rating data on a simple length estimation task. Our decontamination techniques yield an over 20% reduction in the error of human judgments. 1

2 0.59427053 83 nips-2010-Evidence-Specific Structures for Rich Tractable CRFs

Author: Anton Chechetka, Carlos Guestrin

Abstract: We present a simple and effective approach to learning tractable conditional random fields with structure that depends on the evidence. Our approach retains the advantages of tractable discriminative models, namely efficient exact inference and arbitrarily accurate parameter learning in polynomial time. At the same time, our algorithm does not suffer a large expressive power penalty inherent to fixed tractable structures. On real-life relational datasets, our approach matches or exceeds state of the art accuracy of the dense models, and at the same time provides an order of magnitude speedup. 1

3 0.52971172 71 nips-2010-Efficient Relational Learning with Hidden Variable Detection

Author: Ni Lao, Jun Zhu, Liu Xinwang, Yandong Liu, William W. Cohen

Abstract: Markov networks (MNs) can incorporate arbitrarily complex features in modeling relational data. However, this flexibility comes at a sharp price of training an exponentially complex model. To address this challenge, we propose a novel relational learning approach, which consists of a restricted class of relational MNs (RMNs) called relation tree-based RMN (treeRMN), and an efficient Hidden Variable Detection algorithm called Contrastive Variable Induction (CVI). On one hand, the restricted treeRMN only considers simple (e.g., unary and pairwise) features in relational data and thus achieves computational efficiency; and on the other hand, the CVI algorithm efficiently detects hidden variables which can capture long range dependencies. Therefore, the resultant approach is highly efficient yet does not sacrifice its expressive power. Empirical results on four real datasets show that the proposed relational learning method can achieve similar prediction quality as the state-of-the-art approaches, but is significantly more efficient in training; and the induced hidden variables are semantically meaningful and crucial to improve the training speed and prediction qualities of treeRMNs.

4 0.51620233 66 nips-2010-Double Q-learning

Author: Hado V. Hasselt

Abstract: In some stochastic environments the well-known reinforcement learning algorithm Q-learning performs very poorly. This poor performance is caused by large overestimations of action values. These overestimations result from a positive bias that is introduced because Q-learning uses the maximum action value as an approximation for the maximum expected action value. We introduce an alternative way to approximate the maximum expected value for any set of random variables. The obtained double estimator method is shown to sometimes underestimate rather than overestimate the maximum expected value. We apply the double estimator to Q-learning to construct Double Q-learning, a new off-policy reinforcement learning algorithm. We show the new algorithm converges to the optimal policy and that it performs well in some settings in which Q-learning performs poorly due to its overestimation. 1

5 0.48545155 19 nips-2010-A rational decision making framework for inhibitory control

Author: Pradeep Shenoy, Angela J. Yu, Rajesh P. Rao

Abstract: Intelligent agents are often faced with the need to choose actions with uncertain consequences, and to modify those actions according to ongoing sensory processing and changing task demands. The requisite ability to dynamically modify or cancel planned actions is known as inhibitory control in psychology. We formalize inhibitory control as a rational decision-making problem, and apply to it to the classical stop-signal task. Using Bayesian inference and stochastic control tools, we show that the optimal policy systematically depends on various parameters of the problem, such as the relative costs of different action choices, the noise level of sensory inputs, and the dynamics of changing environmental demands. Our normative model accounts for a range of behavioral data in humans and animals in the stop-signal task, suggesting that the brain implements statistically optimal, dynamically adaptive, and reward-sensitive decision-making in the context of inhibitory control problems. 1

6 0.44703826 122 nips-2010-Improving the Asymptotic Performance of Markov Chain Monte-Carlo by Inserting Vortices

7 0.41324154 1 nips-2010-(RF)^2 -- Random Forest Random Field

8 0.40792194 95 nips-2010-Feature Transitions with Saccadic Search: Size, Color, and Orientation Are Not Alike

9 0.40748093 81 nips-2010-Evaluating neuronal codes for inference using Fisher information

10 0.39875859 254 nips-2010-Stability Approach to Regularization Selection (StARS) for High Dimensional Graphical Models

11 0.394283 159 nips-2010-Lifted Inference Seen from the Other Side : The Tractable Features

12 0.38429946 119 nips-2010-Implicit encoding of prior probabilities in optimal neural populations

13 0.37460852 67 nips-2010-Dynamic Infinite Relational Model for Time-varying Relational Data Analysis

14 0.35589072 144 nips-2010-Learning Efficient Markov Networks

15 0.34758961 161 nips-2010-Linear readout from a neural population with partial correlation data

16 0.34435356 21 nips-2010-Accounting for network effects in neuronal responses using L1 regularized point process models

17 0.34131759 171 nips-2010-Movement extraction by detecting dynamics switches and repetitions

18 0.33690327 215 nips-2010-Probabilistic Deterministic Infinite Automata

19 0.33240134 162 nips-2010-Link Discovery using Graph Feature Tracking

20 0.32853475 203 nips-2010-Parametric Bandits: The Generalized Linear Case


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(13, 0.026), (27, 0.562), (30, 0.032), (35, 0.017), (44, 0.017), (45, 0.117), (50, 0.032), (52, 0.027), (60, 0.026), (77, 0.023), (90, 0.025)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.96025431 119 nips-2010-Implicit encoding of prior probabilities in optimal neural populations

Author: Deep Ganguli, Eero P. Simoncelli

Abstract: unkown-abstract

2 0.93767655 128 nips-2010-Infinite Relational Modeling of Functional Connectivity in Resting State fMRI

Author: Morten Mørup, Kristoffer Madsen, Anne-marie Dogonowski, Hartwig Siebner, Lars K. Hansen

Abstract: Functional magnetic resonance imaging (fMRI) can be applied to study the functional connectivity of the neural elements which form complex network at a whole brain level. Most analyses of functional resting state networks (RSN) have been based on the analysis of correlation between the temporal dynamics of various regions of the brain. While these models can identify coherently behaving groups in terms of correlation they give little insight into how these groups interact. In this paper we take a different view on the analysis of functional resting state networks. Starting from the definition of resting state as functional coherent groups we search for functional units of the brain that communicate with other parts of the brain in a coherent manner as measured by mutual information. We use the infinite relational model (IRM) to quantify functional coherent groups of resting state networks and demonstrate how the extracted component interactions can be used to discriminate between functional resting state activity in multiple sclerosis and normal subjects. 1

3 0.91753292 60 nips-2010-Deterministic Single-Pass Algorithm for LDA

Author: Issei Sato, Kenichi Kurihara, Hiroshi Nakagawa

Abstract: We develop a deterministic single-pass algorithm for latent Dirichlet allocation (LDA) in order to process received documents one at a time and then discard them in an excess text stream. Our algorithm does not need to store old statistics for all data. The proposed algorithm is much faster than a batch algorithm and is comparable to the batch algorithm in terms of perplexity in experiments.

same-paper 4 0.8885842 121 nips-2010-Improving Human Judgments by Decontaminating Sequential Dependencies

Author: Harold Pashler, Matthew Wilder, Robert Lindsey, Matt Jones, Michael C. Mozer, Michael P. Holmes

Abstract: For over half a century, psychologists have been struck by how poor people are at expressing their internal sensations, impressions, and evaluations via rating scales. When individuals make judgments, they are incapable of using an absolute rating scale, and instead rely on reference points from recent experience. This relativity of judgment limits the usefulness of responses provided by individuals to surveys, questionnaires, and evaluation forms. Fortunately, the cognitive processes that transform internal states to responses are not simply noisy, but rather are influenced by recent experience in a lawful manner. We explore techniques to remove sequential dependencies, and thereby decontaminate a series of ratings to obtain more meaningful human judgments. In our formulation, decontamination is fundamentally a problem of inferring latent states (internal sensations) which, because of the relativity of judgment, have temporal dependencies. We propose a decontamination solution using a conditional random field with constraints motivated by psychological theories of relative judgment. Our exploration of decontamination models is supported by two experiments we conducted to obtain ground-truth rating data on a simple length estimation task. Our decontamination techniques yield an over 20% reduction in the error of human judgments. 1

5 0.85416579 39 nips-2010-Bayesian Action-Graph Games

Author: Albert X. Jiang, Kevin Leyton-brown

Abstract: Games of incomplete information, or Bayesian games, are an important gametheoretic model and have many applications in economics. We propose Bayesian action-graph games (BAGGs), a novel graphical representation for Bayesian games. BAGGs can represent arbitrary Bayesian games, and furthermore can compactly express Bayesian games exhibiting commonly encountered types of structure including symmetry, action- and type-specific utility independence, and probabilistic independence of type distributions. We provide an algorithm for computing expected utility in BAGGs, and discuss conditions under which the algorithm runs in polynomial time. Bayes-Nash equilibria of BAGGs can be computed by adapting existing algorithms for complete-information normal form games and leveraging our expected utility algorithm. We show both theoretically and empirically that our approaches improve significantly on the state of the art. 1

6 0.8329795 266 nips-2010-The Maximal Causes of Natural Scenes are Edge Filters

7 0.76202422 81 nips-2010-Evaluating neuronal codes for inference using Fisher information

8 0.72617078 161 nips-2010-Linear readout from a neural population with partial correlation data

9 0.70263034 6 nips-2010-A Discriminative Latent Model of Image Region and Object Tag Correspondence

10 0.69018716 127 nips-2010-Inferring Stimulus Selectivity from the Spatial Structure of Neural Network Dynamics

11 0.66654831 21 nips-2010-Accounting for network effects in neuronal responses using L1 regularized point process models

12 0.66224831 97 nips-2010-Functional Geometry Alignment and Localization of Brain Areas

13 0.63708204 98 nips-2010-Functional form of motion priors in human motion perception

14 0.62625301 194 nips-2010-Online Learning for Latent Dirichlet Allocation

15 0.61917567 268 nips-2010-The Neural Costs of Optimal Control

16 0.61232114 19 nips-2010-A rational decision making framework for inhibitory control

17 0.6109035 44 nips-2010-Brain covariance selection: better individual functional connectivity models using population prior

18 0.60421848 123 nips-2010-Individualized ROI Optimization via Maximization of Group-wise Consistency of Structural and Functional Profiles

19 0.60208642 244 nips-2010-Sodium entry efficiency during action potentials: A novel single-parameter family of Hodgkin-Huxley models

20 0.59469539 17 nips-2010-A biologically plausible network for the computation of orientation dominance