andrew_gelman_stats andrew_gelman_stats-2013 andrew_gelman_stats-2013-1926 knowledge-graph by maker-knowledge-mining

1926 andrew gelman stats-2013-07-05-More plain old everyday Bayesianism


meta infos for this blog

Source: html

Introduction: Following up on this story , Bob Goodman writes: A most recent issue of the New England Journal of Medicine published a study entitled “Biventricular Pacing for Atrioventricular Block and Systolic Dysfunction,” (N Engl J Med 2013; 368:1585-1593), whereby “A hierarchical Bayesian proportional-hazards model was used for analysis of the primary outcome.” It is the first study I can recall in this journal that has reported on Table 2 (primary outcomes) “The Posterior Probability of Hazard Ratio < 1" (which in this case was .9978). This is ok, but to be really picky I will say that there’s typically not so much reason to care about the posterior probability that the effect is greater than 1; I’d rather have an estimate of the effect. Also we should be using informative priors.


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 ” It is the first study I can recall in this journal that has reported on Table 2 (primary outcomes) “The Posterior Probability of Hazard Ratio < 1" (which in this case was . [sent-2, score-0.71]

2 This is ok, but to be really picky I will say that there’s typically not so much reason to care about the posterior probability that the effect is greater than 1; I’d rather have an estimate of the effect. [sent-4, score-1.493]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('primary', 0.311), ('med', 0.269), ('goodman', 0.248), ('whereby', 0.241), ('picky', 0.234), ('posterior', 0.234), ('hazard', 0.224), ('england', 0.19), ('block', 0.188), ('journal', 0.183), ('probability', 0.18), ('entitled', 0.177), ('ratio', 0.171), ('medicine', 0.16), ('study', 0.159), ('greater', 0.159), ('bob', 0.156), ('table', 0.142), ('outcomes', 0.135), ('priors', 0.133), ('informative', 0.13), ('recall', 0.127), ('reported', 0.125), ('hierarchical', 0.122), ('care', 0.112), ('typically', 0.109), ('ok', 0.097), ('issue', 0.094), ('estimate', 0.089), ('reason', 0.087), ('effect', 0.084), ('published', 0.082), ('story', 0.079), ('following', 0.075), ('bayesian', 0.074), ('recent', 0.073), ('used', 0.069), ('case', 0.061), ('analysis', 0.061), ('using', 0.059), ('rather', 0.059), ('model', 0.056), ('first', 0.055), ('say', 0.054), ('new', 0.051), ('really', 0.048), ('much', 0.044), ('writes', 0.043), ('also', 0.037)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 1.0 1926 andrew gelman stats-2013-07-05-More plain old everyday Bayesianism

Introduction: Following up on this story , Bob Goodman writes: A most recent issue of the New England Journal of Medicine published a study entitled “Biventricular Pacing for Atrioventricular Block and Systolic Dysfunction,” (N Engl J Med 2013; 368:1585-1593), whereby “A hierarchical Bayesian proportional-hazards model was used for analysis of the primary outcome.” It is the first study I can recall in this journal that has reported on Table 2 (primary outcomes) “The Posterior Probability of Hazard Ratio < 1" (which in this case was .9978). This is ok, but to be really picky I will say that there’s typically not so much reason to care about the posterior probability that the effect is greater than 1; I’d rather have an estimate of the effect. Also we should be using informative priors.

2 0.26875666 1209 andrew gelman stats-2012-03-12-As a Bayesian I want scientists to report their data non-Bayesianly

Introduction: Philipp Doebler writes: I was quite happy that recently you shared some thoughts of yours and others on meta-analysis. I especially liked the slides by Chris Schmid that you linked from your blog. A large portion of my work deals with meta-analysis and I am also fond of using Bayesian methods (actually two of the projects I am working on are very Bayesian), though I can not say I have opinions with respect to the underlying philosophy. I would say though, that I do share your view that there are good reasons to use informative priors. The reason I am writing to you is that this leads to the following dilemma, which is puzzling me. Say a number of scientists conduct similar studies over the years and all of them did this in a Bayesian fashion. If each of the groups used informative priors based on the research of existing groups the priors could become more and more informative over the years, since more and more is known over the subject. At least in smallish studies these p

3 0.13748123 1713 andrew gelman stats-2013-02-08-P-values and statistical practice

Introduction: From my new article in the journal Epidemiology: Sander Greenland and Charles Poole accept that P values are here to stay but recognize that some of their most common interpretations have problems. The casual view of the P value as posterior probability of the truth of the null hypothesis is false and not even close to valid under any reasonable model, yet this misunderstanding persists even in high-stakes settings (as discussed, for example, by Greenland in 2011). The formal view of the P value as a probability conditional on the null is mathematically correct but typically irrelevant to research goals (hence, the popularity of alternative—if wrong—interpretations). A Bayesian interpretation based on a spike-and-slab model makes little sense in applied contexts in epidemiology, political science, and other fields in which true effects are typically nonzero and bounded (thus violating both the “spike” and the “slab” parts of the model). I find Greenland and Poole’s perspective t

4 0.12989603 1941 andrew gelman stats-2013-07-16-Priors

Introduction: Nick Firoozye writes: While I am absolutely sympathetic to the Bayesian agenda I am often troubled by the requirement of having priors. We must have priors on the parameter of an infinite number of model we have never seen before and I find this troubling. There is a similarly troubling problem in economics of utility theory. Utility is on consumables. To be complete a consumer must assign utility to all sorts of things they never would have encountered. More recent versions of utility theory instead make consumption goods a portfolio of attributes. Cadillacs are x many units of luxury y of transport etc etc. And we can automatically have personal utilities to all these attributes. I don’t ever see parameters. Some model have few and some have hundreds. Instead, I see data. So I don’t know how to have an opinion on parameters themselves. Rather I think it far more natural to have opinions on the behavior of models. The prior predictive density is a good and sensible notion. Also

5 0.12462186 775 andrew gelman stats-2011-06-21-Fundamental difficulty of inference for a ratio when the denominator could be positive or negative

Introduction: Ratio estimates are common in statistics. In survey sampling, the ratio estimate is when you use y/x to estimate Y/X (using the notation in which x,y are totals of sample measurements and X,Y are population totals). In textbook sampling examples, the denominator X will be an all-positive variable, something that is easy to measure and is, ideally, close to proportional to Y. For example, X is last year’s sales and Y is this year’s sales, or X is the number of people in a cluster and Y is some count. Ratio estimation doesn’t work so well if X can be either positive or negative. More generally we can consider any estimate of a ratio, with no need for a survey sampling context. The problem with estimating Y/X is that the very interpretation of Y/X can change completely if the sign of X changes. Everything is ok for a point estimate: you get X.hat and Y.hat, you can take the ratio Y.hat/X.hat, no problem. But the inference falls apart if you have enough uncertainty in X.hat th

6 0.11628205 2200 andrew gelman stats-2014-02-05-Prior distribution for a predicted probability

7 0.11388073 1865 andrew gelman stats-2013-05-20-What happened that the journal Psychological Science published a paper with no identifiable strengths?

8 0.11261594 1092 andrew gelman stats-2011-12-29-More by Berger and me on weakly informative priors

9 0.10651644 803 andrew gelman stats-2011-07-14-Subtleties with measurement-error models for the evaluation of wacky claims

10 0.10639571 1955 andrew gelman stats-2013-07-25-Bayes-respecting experimental design and other things

11 0.10529747 1087 andrew gelman stats-2011-12-27-“Keeping things unridiculous”: Berger, O’Hagan, and me on weakly informative priors

12 0.10324808 1208 andrew gelman stats-2012-03-11-Gelman on Hennig on Gelman on Bayes

13 0.10313188 1363 andrew gelman stats-2012-06-03-Question about predictive checks

14 0.10287468 1046 andrew gelman stats-2011-12-07-Neutral noninformative and informative conjugate beta and gamma prior distributions

15 0.10238912 1999 andrew gelman stats-2013-08-27-Bayesian model averaging or fitting a larger model

16 0.10028517 2035 andrew gelman stats-2013-09-23-Scalable Stan

17 0.10003968 1877 andrew gelman stats-2013-05-30-Infill asymptotics and sprawl asymptotics

18 0.099290848 2368 andrew gelman stats-2014-06-11-Bayes in the research conversation

19 0.097285964 1465 andrew gelman stats-2012-08-21-D. Buggin

20 0.096879959 1401 andrew gelman stats-2012-06-30-David Hogg on statistics


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.158), (1, 0.11), (2, 0.018), (3, -0.052), (4, -0.065), (5, -0.056), (6, 0.046), (7, -0.023), (8, -0.089), (9, -0.048), (10, 0.026), (11, -0.013), (12, 0.003), (13, 0.041), (14, 0.012), (15, -0.015), (16, 0.044), (17, 0.037), (18, -0.01), (19, 0.017), (20, -0.053), (21, 0.026), (22, 0.018), (23, -0.003), (24, -0.004), (25, 0.037), (26, -0.046), (27, 0.027), (28, -0.019), (29, -0.043), (30, -0.069), (31, -0.039), (32, -0.049), (33, 0.006), (34, -0.018), (35, 0.024), (36, -0.013), (37, -0.026), (38, 0.037), (39, 0.003), (40, -0.032), (41, -0.014), (42, 0.017), (43, -0.017), (44, 0.06), (45, -0.021), (46, -0.044), (47, 0.069), (48, -0.017), (49, 0.003)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.98164314 1926 andrew gelman stats-2013-07-05-More plain old everyday Bayesianism

Introduction: Following up on this story , Bob Goodman writes: A most recent issue of the New England Journal of Medicine published a study entitled “Biventricular Pacing for Atrioventricular Block and Systolic Dysfunction,” (N Engl J Med 2013; 368:1585-1593), whereby “A hierarchical Bayesian proportional-hazards model was used for analysis of the primary outcome.” It is the first study I can recall in this journal that has reported on Table 2 (primary outcomes) “The Posterior Probability of Hazard Ratio < 1" (which in this case was .9978). This is ok, but to be really picky I will say that there’s typically not so much reason to care about the posterior probability that the effect is greater than 1; I’d rather have an estimate of the effect. Also we should be using informative priors.

2 0.69488811 1209 andrew gelman stats-2012-03-12-As a Bayesian I want scientists to report their data non-Bayesianly

Introduction: Philipp Doebler writes: I was quite happy that recently you shared some thoughts of yours and others on meta-analysis. I especially liked the slides by Chris Schmid that you linked from your blog. A large portion of my work deals with meta-analysis and I am also fond of using Bayesian methods (actually two of the projects I am working on are very Bayesian), though I can not say I have opinions with respect to the underlying philosophy. I would say though, that I do share your view that there are good reasons to use informative priors. The reason I am writing to you is that this leads to the following dilemma, which is puzzling me. Say a number of scientists conduct similar studies over the years and all of them did this in a Bayesian fashion. If each of the groups used informative priors based on the research of existing groups the priors could become more and more informative over the years, since more and more is known over the subject. At least in smallish studies these p

3 0.67786717 1713 andrew gelman stats-2013-02-08-P-values and statistical practice

Introduction: From my new article in the journal Epidemiology: Sander Greenland and Charles Poole accept that P values are here to stay but recognize that some of their most common interpretations have problems. The casual view of the P value as posterior probability of the truth of the null hypothesis is false and not even close to valid under any reasonable model, yet this misunderstanding persists even in high-stakes settings (as discussed, for example, by Greenland in 2011). The formal view of the P value as a probability conditional on the null is mathematically correct but typically irrelevant to research goals (hence, the popularity of alternative—if wrong—interpretations). A Bayesian interpretation based on a spike-and-slab model makes little sense in applied contexts in epidemiology, political science, and other fields in which true effects are typically nonzero and bounded (thus violating both the “spike” and the “slab” parts of the model). I find Greenland and Poole’s perspective t

4 0.66942132 2099 andrew gelman stats-2013-11-13-“What are some situations in which the classical approach (or a naive implementation of it, based on cookbook recipes) gives worse results than a Bayesian approach, results that actually impeded the science?”

Introduction: Phil Nelson writes in the context of a biostatistics textbook he is writing, “Physical models of living systems”: There are a number of classic statistical problems that arise every day in the lab, and which are discussed in any book: 1. In a control group, M untreated rats out of 20 got a form of cancer. In a test group, N treated rats out of 20 got that cancer. Is this a significant difference? 2. In a control group of 20 untreated rates, their body weights at 2 weeks were w_1,…, w_20. In a test group of 20 treated rats, their body weights at 2 weeks were w’_1,…, w’_20. Are the means significantly different? 3. In a group of 20 rats, each given dose d_i of a drug, their body weights at 2 weeks were w_i. Is there a significant correlation between d and w? I would like to ask: What are some situations in which the classical approach (or a naive implementation of it, based on cookbook recipes) gives worse results than a Bayesian approach, results that actually impeded the scien

5 0.65380603 1046 andrew gelman stats-2011-12-07-Neutral noninformative and informative conjugate beta and gamma prior distributions

Introduction: Jouni Kerman did a cool bit of research justifying the Beta (1/3, 1/3) prior as noninformative for binomial data, and the Gamma (1/3, 0) prior for Poisson data. You probably thought that nothing new could be said about noninformative priors in such basic problems, but you were wrong! Here’s the story : The conjugate binomial and Poisson models are commonly used for estimating proportions or rates. However, it is not well known that the conventional noninformative conjugate priors tend to shrink the posterior quantiles toward the boundary or toward the middle of the parameter space, making them thus appear excessively informative. The shrinkage is always largest when the number of observed events is small. This behavior persists for all sample sizes and exposures. The effect of the prior is therefore most conspicuous and potentially controversial when analyzing rare events. As alternative default conjugate priors, I [Jouni] introduce Beta(1/3, 1/3) and Gamma(1/3, 0), which I cal

6 0.64684075 1877 andrew gelman stats-2013-05-30-Infill asymptotics and sprawl asymptotics

7 0.62919325 804 andrew gelman stats-2011-07-15-Static sensitivity analysis

8 0.62834144 1829 andrew gelman stats-2013-04-28-Plain old everyday Bayesianism!

9 0.62693489 2200 andrew gelman stats-2014-02-05-Prior distribution for a predicted probability

10 0.61703598 916 andrew gelman stats-2011-09-18-Multimodality in hierarchical models

11 0.61013216 846 andrew gelman stats-2011-08-09-Default priors update?

12 0.60762167 1941 andrew gelman stats-2013-07-16-Priors

13 0.60654658 2140 andrew gelman stats-2013-12-19-Revised evidence for statistical standards

14 0.60548127 801 andrew gelman stats-2011-07-13-On the half-Cauchy prior for a global scale parameter

15 0.59696704 2114 andrew gelman stats-2013-11-26-“Please make fun of this claim”

16 0.59463996 1858 andrew gelman stats-2013-05-15-Reputations changeable, situations tolerable

17 0.59268576 2027 andrew gelman stats-2013-09-17-Christian Robert on the Jeffreys-Lindley paradox; more generally, it’s good news when philosophical arguments can be transformed into technical modeling issues

18 0.59040838 2223 andrew gelman stats-2014-02-24-“Edlin’s rule” for routinely scaling down published estimates

19 0.59036011 2201 andrew gelman stats-2014-02-06-Bootstrap averaging: Examples where it works and where it doesn’t work

20 0.58526862 618 andrew gelman stats-2011-03-18-Prior information . . . about the likelihood


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(15, 0.028), (16, 0.083), (24, 0.172), (30, 0.033), (35, 0.187), (53, 0.048), (55, 0.023), (99, 0.297)]

similar blogs list:

simIndex simValue blogId blogTitle

1 0.95912695 473 andrew gelman stats-2010-12-17-Why a bonobo won’t play poker with you

Introduction: Sciencedaily has posted an article titled Apes Unwilling to Gamble When Odds Are Uncertain : The apes readily distinguished between the different probabilities of winning: they gambled a lot when there was a 100 percent chance, less when there was a 50 percent chance, and only rarely when there was no chance In some trials, however, the experimenter didn’t remove a lid from the bowl, so the apes couldn’t assess the likelihood of winning a banana The odds from the covered bowl were identical to those from the risky option: a 50 percent chance of getting the much sought-after banana. But apes of both species were less likely to choose this ambiguous option. Like humans, they showed “ambiguity aversion” — preferring to gamble more when they knew the odds than when they didn’t. Given some of the other differences between chimps and bonobos, Hare and Rosati had expected to find the bonobos to be more averse to ambiguity, but that didn’t turn out to be the case. Thanks to Sta

2 0.9521969 837 andrew gelman stats-2011-08-04-Is it rational to vote?

Introduction: Hear me interviewed on the topic here . P.S. The interview was fine but I don’t agree with everything on the linked website. For example, this bit: Global warming is not the first case of a widespread fear based on incomplete knowledge turned out to be false or at least greatly exaggerated. Global warming has many of the characteristics of a popular delusion, an irrational fear or cause that is embraced by millions of people because, well, it is believed by millions of people! All right, then.

3 0.94653118 2049 andrew gelman stats-2013-10-03-On house arrest for p-hacking

Introduction: People keep pointing me to this excellent news article by David Brown, about a scientist who was convicted of data manipulation: In all, 330 patients were randomly assigned to get either interferon gamma-1b or placebo injections. Disease progression or death occurred in 46 percent of those on the drug and 52 percent of those on placebo. That was not a significant difference, statistically speaking. When only survival was considered, however, the drug looked better: 10 percent of people getting the drug died, compared with 17 percent of those on placebo. However, that difference wasn’t “statistically significant,” either. Specifically, the so-called P value — a mathematical measure of the strength of the evidence that there’s a true difference between a treatment and placebo — was 0.08. . . . Technically, the study was a bust, although the results leaned toward a benefit from interferon gamma-1b. Was there a group of patients in which the results tipped? Harkonen asked the statis

same-blog 4 0.94533551 1926 andrew gelman stats-2013-07-05-More plain old everyday Bayesianism

Introduction: Following up on this story , Bob Goodman writes: A most recent issue of the New England Journal of Medicine published a study entitled “Biventricular Pacing for Atrioventricular Block and Systolic Dysfunction,” (N Engl J Med 2013; 368:1585-1593), whereby “A hierarchical Bayesian proportional-hazards model was used for analysis of the primary outcome.” It is the first study I can recall in this journal that has reported on Table 2 (primary outcomes) “The Posterior Probability of Hazard Ratio < 1" (which in this case was .9978). This is ok, but to be really picky I will say that there’s typically not so much reason to care about the posterior probability that the effect is greater than 1; I’d rather have an estimate of the effect. Also we should be using informative priors.

5 0.94496429 1443 andrew gelman stats-2012-08-04-Bayesian Learning via Stochastic Gradient Langevin Dynamics

Introduction: Burak Bayramli writes: In this paper by Sunjin Ahn, Anoop Korattikara, and Max Welling and this paper by Welling and Yee Whye The, there are some arguments on big data and the use of MCMC. Both papers have suggested improvements to speed up MCMC computations. I was wondering what your thoughts were, especially on this paragraph: When a dataset has a billion data-cases (as is not uncommon these days) MCMC algorithms will not even have generated a single (burn-in) sample when a clever learning algorithm based on stochastic gradients may already be making fairly good predictions. In fact, the intriguing results of Bottou and Bousquet (2008) seem to indicate that in terms of “number of bits learned per unit of computation”, an algorithm as simple as stochastic gradient descent is almost optimally efficient. We therefore argue that for Bayesian methods to remain useful in an age when the datasets grow at an exponential rate, they need to embrace the ideas of the stochastic optimiz

6 0.93655694 591 andrew gelman stats-2011-02-25-Quantitative Methods in the Social Sciences M.A.: Innovative, interdisciplinary social science research program for a data-rich world

7 0.93118978 942 andrew gelman stats-2011-10-04-45% hitting, 25% fielding, 25% pitching, and 100% not telling us how they did it

8 0.92747688 1516 andrew gelman stats-2012-09-30-Computational problems with glm etc.

9 0.92166877 881 andrew gelman stats-2011-08-30-Rickey Henderson and Peter Angelos, together again

10 0.9165684 296 andrew gelman stats-2010-09-26-A simple semigraphic display

11 0.91452324 388 andrew gelman stats-2010-11-01-The placebo effect in pharma

12 0.90607023 895 andrew gelman stats-2011-09-08-How to solve the Post Office’s problems?

13 0.90394342 392 andrew gelman stats-2010-11-03-Taleb + 3.5 years

14 0.89660937 1264 andrew gelman stats-2012-04-14-Learning from failure

15 0.89342558 1130 andrew gelman stats-2012-01-20-Prior beliefs about locations of decision boundaries

16 0.88488734 807 andrew gelman stats-2011-07-17-Macro causality

17 0.8819651 2274 andrew gelman stats-2014-03-30-Adjudicating between alternative interpretations of a statistical interaction?

18 0.87651002 1881 andrew gelman stats-2013-06-03-Boot

19 0.87605917 1799 andrew gelman stats-2013-04-12-Stan 1.3.0 and RStan 1.3.0 Ready for Action

20 0.87335026 687 andrew gelman stats-2011-04-29-Zero is zero