andrew_gelman_stats andrew_gelman_stats-2012 andrew_gelman_stats-2012-1562 knowledge-graph by maker-knowledge-mining

1562 andrew gelman stats-2012-11-05-Let’s try this: Instead of saying, “The probability is 75%,” say “There’s a 25% chance I’m wrong”


meta infos for this blog

Source: html

Introduction: I recently wrote about the difficulty people have with probabilities, in this case the probability that Obama wins the election. If the probability is reported as 70%, people think Obama is going to win. Actually, though, it just means that Obama is predicted to get about 50.8% of the two-party vote, with an uncertainty of something like 2 percentage points. So, as I wrote, the election really is too close to call in the sense that the predicted vote margin is less than its uncertainty. But . . . when people see a number such as 70%, they tend to attribute too much certainty to it. Especially when the estimated probability has increased from, say 60%. How to get the point across? Commenter HS had what seems like a good suggestion: Say that Obama will win, but there is 25% chance (or whatever) that this prediction is wrong? Same point, just slightly different framing, but somehow, this seems far less incendiary. I like that. Somehow a stated probability of 75% sounds a


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 I recently wrote about the difficulty people have with probabilities, in this case the probability that Obama wins the election. [sent-1, score-0.781]

2 If the probability is reported as 70%, people think Obama is going to win. [sent-2, score-0.455]

3 Actually, though, it just means that Obama is predicted to get about 50. [sent-3, score-0.368]

4 8% of the two-party vote, with an uncertainty of something like 2 percentage points. [sent-4, score-0.36]

5 So, as I wrote, the election really is too close to call in the sense that the predicted vote margin is less than its uncertainty. [sent-5, score-0.94]

6 when people see a number such as 70%, they tend to attribute too much certainty to it. [sent-9, score-0.532]

7 Especially when the estimated probability has increased from, say 60%. [sent-10, score-0.561]

8 Commenter HS had what seems like a good suggestion: Say that Obama will win, but there is 25% chance (or whatever) that this prediction is wrong? [sent-12, score-0.57]

9 Same point, just slightly different framing, but somehow, this seems far less incendiary. [sent-13, score-0.418]

10 Somehow a stated probability of 75% sounds all too sure, whereas saying that there’s a 25% chance of being wrong seems to better convey the uncertainty in the prediction. [sent-15, score-1.409]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('obama', 0.417), ('probability', 0.273), ('predicted', 0.23), ('somehow', 0.227), ('hs', 0.216), ('prediction', 0.207), ('vote', 0.19), ('uncertainty', 0.189), ('chance', 0.16), ('framing', 0.149), ('certainty', 0.147), ('margin', 0.147), ('attribute', 0.141), ('wins', 0.141), ('stated', 0.134), ('wrong', 0.133), ('seems', 0.133), ('convey', 0.128), ('suggestion', 0.119), ('commenter', 0.116), ('less', 0.112), ('increased', 0.111), ('win', 0.109), ('probabilities', 0.108), ('wrote', 0.108), ('slightly', 0.104), ('difficulty', 0.103), ('percentage', 0.101), ('whereas', 0.097), ('election', 0.097), ('sounds', 0.096), ('estimated', 0.095), ('reported', 0.094), ('tend', 0.093), ('people', 0.088), ('close', 0.083), ('say', 0.082), ('call', 0.081), ('point', 0.079), ('especially', 0.078), ('means', 0.077), ('whatever', 0.077), ('across', 0.076), ('like', 0.07), ('far', 0.069), ('recently', 0.068), ('saying', 0.066), ('number', 0.063), ('get', 0.061), ('though', 0.055)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 1.0 1562 andrew gelman stats-2012-11-05-Let’s try this: Instead of saying, “The probability is 75%,” say “There’s a 25% chance I’m wrong”

Introduction: I recently wrote about the difficulty people have with probabilities, in this case the probability that Obama wins the election. If the probability is reported as 70%, people think Obama is going to win. Actually, though, it just means that Obama is predicted to get about 50.8% of the two-party vote, with an uncertainty of something like 2 percentage points. So, as I wrote, the election really is too close to call in the sense that the predicted vote margin is less than its uncertainty. But . . . when people see a number such as 70%, they tend to attribute too much certainty to it. Especially when the estimated probability has increased from, say 60%. How to get the point across? Commenter HS had what seems like a good suggestion: Say that Obama will win, but there is 25% chance (or whatever) that this prediction is wrong? Same point, just slightly different framing, but somehow, this seems far less incendiary. I like that. Somehow a stated probability of 75% sounds a

2 0.34900174 1544 andrew gelman stats-2012-10-22-Is it meaningful to talk about a probability of “65.7%” that Obama will win the election?

Introduction: The other day we had a fun little discussion in the comments section of the sister blog about the appropriateness of stating forecast probabilities to the nearest tenth of a percentage point. It started when Josh Tucker posted this graph from Nate Silver : My first reaction was: this looks pretty but it’s hyper-precise. I’m a big fan of Nate’s work, but all those little wiggles on the graph can’t really mean anything. And what could it possibly mean to compute this probability to that level of precision? In the comments, people came at me from two directions. From one side, Jeffrey Friedman expressed a hard core attitude that it’s meaningless to give a probability forecast of a unique event: What could it possibly mean, period, given that this election will never be repeated? . . . I know there’s a vast literature on this, but I’m still curious, as a non-statistician, what it could mean for there to be a meaningful 65% probability (as opposed to a non-quantifiab

3 0.29557598 1574 andrew gelman stats-2012-11-12-How to Lie With Statistics example number 12,498,122

Introduction: This post is by Phil Price. Bill Kristol notes that “Four presidents in the last century have won more than 51 percent of the vote twice: Roosevelt, Eisenhower, Reagan and Obama”. I’m not sure why Kristol, a conservative, is promoting the idea that Obama has a mandate, but that’s up to him. I’m more interested in the remarkable bit of cherry-picking that led to this “only four presidents” statistic. There was one way in which Obama’s victory was large: he won the electoral college 332-206. That’s a thrashing. But if you want to claim that Obama has a “popular mandate” — which people seem to interpret as an overwhelming preference of The People such that the opposition is morally obligated to give way — you can’t make that argument based on the electoral college, you have to look at the popular vote. That presents you with a challenge for the 2012 election, since Obama’s 2.7-point margin in the popular vote was the 12th-smallest out of the 57 elections we’ve had. There’s a nice sor

4 0.2206625 2255 andrew gelman stats-2014-03-19-How Americans vote

Introduction: An interview with me from 2012 : You’re a statistician and wrote a book,  Red State, Blue State, Rich State, Poor State , looking at why Americans vote the way they do. In an election year I think it would be a good time to revisit that question, not just for people in the US, but anyone around the world who wants to understand the realities – rather than the stereotypes – of how Americans vote. I regret the title I gave my book. I was too greedy. I wanted it to be an airport bestseller because I figured there were millions of people who are interested in politics and some subset of them are always looking at the statistics. It’s got a very grabby title and as a result people underestimated the content. They thought it was a popularisation of my work, or, at best, an expansion of an article we’d written. But it had tons of original material. If I’d given it a more serious, political science-y title, then all sorts of people would have wanted to read it, because they would

5 0.19644256 1569 andrew gelman stats-2012-11-08-30-30-40 Nation

Introduction: Barack Obama’s win has a potentially huge effect on policy. The current budget negotiations will affect the level and direction of government spending and on the mix of taxes paid by different groups of Americans. We can guess that a President Romney would have fought hard against upper-income tax increases. Other areas of long-term impact include the government’s stance on global warming, foreign policy, and the likelihood that Obama will nominate new Supreme Court justices who will uphold the right to abortion announced in Roe v. Wade. When it comes to public opinion, the story is different. The Democrats may well benefit in 2014 and 2016 from the anticipated slow but steady recovery of the economy over the next few years—but, as of November 6, 2012, the parties are essentially tied, with Barack Obama receiving 51% of the two-party vote, compared to Mitt Romney’s 49%, a split comparable to Al Gore’s narrow victory in 2000, Richard Nixon’s in 1968, and John Kennedy’s in 1960.

6 0.19563037 1075 andrew gelman stats-2011-12-20-This guy has a regular column at Reuters

7 0.17640887 1540 andrew gelman stats-2012-10-18-“Intrade to the 57th power”

8 0.17384189 84 andrew gelman stats-2010-06-14-Is it 1930?

9 0.15511075 394 andrew gelman stats-2010-11-05-2010: What happened?

10 0.15199894 389 andrew gelman stats-2010-11-01-Why it can be rational to vote

11 0.15199894 1565 andrew gelman stats-2012-11-06-Why it can be rational to vote

12 0.13955022 551 andrew gelman stats-2011-02-02-Obama and Reagan, sitting in a tree, etc.

13 0.13426505 934 andrew gelman stats-2011-09-30-Nooooooooooooooooooo!

14 0.13309538 1564 andrew gelman stats-2012-11-06-Choose your default, or your default will choose you (election forecasting edition)

15 0.13131037 1567 andrew gelman stats-2012-11-07-Election reports

16 0.12704466 2226 andrew gelman stats-2014-02-26-Econometrics, political science, epidemiology, etc.: Don’t model the probability of a discrete outcome, model the underlying continuous variable

17 0.12403319 1650 andrew gelman stats-2013-01-03-Did Steven Levitt really believe in 2008 that Obama “would be the greatest president in history”?

18 0.12296125 959 andrew gelman stats-2011-10-14-The most clueless political column ever—I think this Easterbrook dude has the journalistic equivalent of “tenure”

19 0.12168244 2224 andrew gelman stats-2014-02-25-Basketball Stats: Don’t model the probability of win, model the expected score differential.

20 0.12066499 2200 andrew gelman stats-2014-02-05-Prior distribution for a predicted probability


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.171), (1, -0.043), (2, 0.158), (3, 0.109), (4, -0.102), (5, -0.028), (6, 0.014), (7, -0.012), (8, -0.003), (9, -0.13), (10, 0.021), (11, 0.038), (12, 0.019), (13, -0.13), (14, -0.123), (15, -0.016), (16, -0.003), (17, 0.022), (18, 0.007), (19, 0.034), (20, -0.007), (21, 0.1), (22, 0.07), (23, 0.055), (24, -0.045), (25, 0.035), (26, 0.017), (27, 0.054), (28, -0.048), (29, -0.07), (30, -0.014), (31, 0.083), (32, -0.049), (33, 0.031), (34, -0.065), (35, -0.004), (36, 0.065), (37, -0.006), (38, -0.062), (39, -0.095), (40, 0.0), (41, -0.093), (42, 0.085), (43, -0.052), (44, 0.029), (45, -0.017), (46, -0.02), (47, 0.105), (48, -0.057), (49, -0.09)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.96394396 1562 andrew gelman stats-2012-11-05-Let’s try this: Instead of saying, “The probability is 75%,” say “There’s a 25% chance I’m wrong”

Introduction: I recently wrote about the difficulty people have with probabilities, in this case the probability that Obama wins the election. If the probability is reported as 70%, people think Obama is going to win. Actually, though, it just means that Obama is predicted to get about 50.8% of the two-party vote, with an uncertainty of something like 2 percentage points. So, as I wrote, the election really is too close to call in the sense that the predicted vote margin is less than its uncertainty. But . . . when people see a number such as 70%, they tend to attribute too much certainty to it. Especially when the estimated probability has increased from, say 60%. How to get the point across? Commenter HS had what seems like a good suggestion: Say that Obama will win, but there is 25% chance (or whatever) that this prediction is wrong? Same point, just slightly different framing, but somehow, this seems far less incendiary. I like that. Somehow a stated probability of 75% sounds a

2 0.90202266 1544 andrew gelman stats-2012-10-22-Is it meaningful to talk about a probability of “65.7%” that Obama will win the election?

Introduction: The other day we had a fun little discussion in the comments section of the sister blog about the appropriateness of stating forecast probabilities to the nearest tenth of a percentage point. It started when Josh Tucker posted this graph from Nate Silver : My first reaction was: this looks pretty but it’s hyper-precise. I’m a big fan of Nate’s work, but all those little wiggles on the graph can’t really mean anything. And what could it possibly mean to compute this probability to that level of precision? In the comments, people came at me from two directions. From one side, Jeffrey Friedman expressed a hard core attitude that it’s meaningless to give a probability forecast of a unique event: What could it possibly mean, period, given that this election will never be repeated? . . . I know there’s a vast literature on this, but I’m still curious, as a non-statistician, what it could mean for there to be a meaningful 65% probability (as opposed to a non-quantifiab

3 0.87691015 1075 andrew gelman stats-2011-12-20-This guy has a regular column at Reuters

Introduction: Gregg Easterbrook : Gingrich is a wild card. He probably would end up a flaming wreckage in electoral terms, but there’s a chance he could become seen as the man unafraid to bring sweeping change to an ossified Washington, D.C. There’s perhaps a 90 percent likelihood Obama would wipe the floor with Gingrich, versus a 10 percent likelihood Gingrich would stage an historic upset. This is the dumbest thing I’ve seen since . . . ummm, I dunno, how bout this ? It actually gets worse because Easterbrook then invokes game theory. What next? Catastrophe theory? Intelligent design? P.S. Maybe I should explain for readers without an education in probability theory. Let’s suppose “wipe the floor” means that Obama gets 55%+ of the two-party vote, and let’s suppose that “an historic upset” means that Obama gets less than 50% of the vote. Now try to draw a forecast distribution that has 90% of its probability above 0.55 and 10% of it’s probability below 0.50. It’s a pretty weird-loo

4 0.87384957 1574 andrew gelman stats-2012-11-12-How to Lie With Statistics example number 12,498,122

Introduction: This post is by Phil Price. Bill Kristol notes that “Four presidents in the last century have won more than 51 percent of the vote twice: Roosevelt, Eisenhower, Reagan and Obama”. I’m not sure why Kristol, a conservative, is promoting the idea that Obama has a mandate, but that’s up to him. I’m more interested in the remarkable bit of cherry-picking that led to this “only four presidents” statistic. There was one way in which Obama’s victory was large: he won the electoral college 332-206. That’s a thrashing. But if you want to claim that Obama has a “popular mandate” — which people seem to interpret as an overwhelming preference of The People such that the opposition is morally obligated to give way — you can’t make that argument based on the electoral college, you have to look at the popular vote. That presents you with a challenge for the 2012 election, since Obama’s 2.7-point margin in the popular vote was the 12th-smallest out of the 57 elections we’ve had. There’s a nice sor

5 0.8368963 1556 andrew gelman stats-2012-11-01-Recently in the sister blogs: special pre-election edition!

Introduction: 1. I can simultaneously (a) accept that Obama has a 72 percent chance of winning and (b) say the election is too close to call 2. Michael’s a Republican, Susan’s a Democrat 3. The narcissism of the narcissism of small differences 4. Obamanomics: A Counter-counterhistory 5. Not a gaffe 6. Categories influence predictions about individual consistency

6 0.81270361 1540 andrew gelman stats-2012-10-18-“Intrade to the 57th power”

7 0.77819693 934 andrew gelman stats-2011-09-30-Nooooooooooooooooooo!

8 0.73684096 138 andrew gelman stats-2010-07-10-Creating a good wager based on probability estimates

9 0.72304595 1569 andrew gelman stats-2012-11-08-30-30-40 Nation

10 0.70559585 959 andrew gelman stats-2011-10-14-The most clueless political column ever—I think this Easterbrook dude has the journalistic equivalent of “tenure”

11 0.65184343 123 andrew gelman stats-2010-07-01-Truth in headlines

12 0.64108843 473 andrew gelman stats-2010-12-17-Why a bonobo won’t play poker with you

13 0.63574195 279 andrew gelman stats-2010-09-15-Electability and perception of electability

14 0.63543993 292 andrew gelman stats-2010-09-23-Doug Hibbs on the fundamentals in 2010

15 0.62375653 23 andrew gelman stats-2010-05-09-Popper’s great, but don’t bother with his theory of probability

16 0.61884588 1027 andrew gelman stats-2011-11-25-Note to student journalists: Google is your friend

17 0.61778557 1083 andrew gelman stats-2011-12-26-The quals and the quants

18 0.61417091 369 andrew gelman stats-2010-10-25-Misunderstanding of divided government

19 0.60700572 551 andrew gelman stats-2011-02-02-Obama and Reagan, sitting in a tree, etc.

20 0.59239227 341 andrew gelman stats-2010-10-14-Confusion about continuous probability densities


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(3, 0.045), (9, 0.026), (15, 0.061), (16, 0.042), (21, 0.053), (24, 0.151), (86, 0.038), (91, 0.018), (93, 0.055), (99, 0.394)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.98640043 1562 andrew gelman stats-2012-11-05-Let’s try this: Instead of saying, “The probability is 75%,” say “There’s a 25% chance I’m wrong”

Introduction: I recently wrote about the difficulty people have with probabilities, in this case the probability that Obama wins the election. If the probability is reported as 70%, people think Obama is going to win. Actually, though, it just means that Obama is predicted to get about 50.8% of the two-party vote, with an uncertainty of something like 2 percentage points. So, as I wrote, the election really is too close to call in the sense that the predicted vote margin is less than its uncertainty. But . . . when people see a number such as 70%, they tend to attribute too much certainty to it. Especially when the estimated probability has increased from, say 60%. How to get the point across? Commenter HS had what seems like a good suggestion: Say that Obama will win, but there is 25% chance (or whatever) that this prediction is wrong? Same point, just slightly different framing, but somehow, this seems far less incendiary. I like that. Somehow a stated probability of 75% sounds a

2 0.98142219 1968 andrew gelman stats-2013-08-05-Evidence on the impact of sustained use of polynomial regression on causal inference (a claim that coal heating is reducing lifespan by 5 years for half a billion people)

Introduction: Yu Xie thought I’d have something to say about this recent paper , “Evidence on the impact of sustained exposure to air pollution on life expectancy from China’s Huai River policy,” by Yuyu Chen, Avraham Ebenstein, Michael Greenstone, and Hongbin Li, which begins: This paper’s findings suggest that an arbitrary Chinese policy that greatly increases total suspended particulates (TSPs) air pollution is causing the 500 million residents of Northern China to lose more than 2.5 billion life years of life expectancy. The quasi-experimental empirical approach is based on China’s Huai River policy, which provided free winter heating via the provision of coal for boilers in cities north of the Huai River but denied heat to the south. Using a regression discontinuity design based on distance from the Huai River, we find that ambient concentrations of TSPs are about 184 μg/m3 [95% confidence interval (CI): 61, 307] or 55% higher in the north. Further, the results indicate that life expectanci

3 0.98128468 897 andrew gelman stats-2011-09-09-The difference between significant and not significant…

Introduction: E. J. Wagenmakers writes: You may be interested in a recent article [by Nieuwenhuis, Forstmann, and Wagenmakers] showing how often researchers draw conclusions by comparing p-values. As you and Hal Stern have pointed out, this is potentially misleading because the difference between significant and not significant is not necessarily significant. We were really suprised to see how often researchers in the neurosciences make this mistake. In the paper we speculate a little bit on the cause of the error. From their paper: In theory, a comparison of two experimental effects requires a statistical test on their difference. In practice, this comparison is often based on an incorrect procedure involving two separate tests in which researchers conclude that effects differ when one effect is significant (P < 0.05) but the other is not (P > 0.05). We reviewed 513 behavioral, systems and cognitive neuroscience articles in five top-ranking journals (Science, Nature, Nature Neuroscien

4 0.97941411 2170 andrew gelman stats-2014-01-13-Judea Pearl overview on causal inference, and more general thoughts on the reexpression of existing methods by considering their implicit assumptions

Introduction: This material should be familiar to many of you but could be helpful to newcomers. Pearl writes: ALL causal conclusions in nonexperimental settings must be based on untested, judgmental assumptions that investigators are prepared to defend on scientific grounds. . . . To understand what the world should be like for a given procedure to work is of no lesser scientific value than seeking evidence for how the world works . . . Assumptions are self-destructive in their honesty. The more explicit the assumption, the more criticism it invites . . . causal diagrams invite the harshest criticism because they make assumptions more explicit and more transparent than other representation schemes. As regular readers know (for example, search this blog for “Pearl”), I have not got much out of the causal-diagrams approach myself, but in general I think that when there are multiple, mathematically equivalent methods of getting the same answer, we tend to go with the framework we are used

5 0.97933006 1619 andrew gelman stats-2012-12-11-There are four ways to get fired from Caesars: (1) theft, (2) sexual harassment, (3) running an experiment without a control group, and (4) keeping a gambling addict away from the casino

Introduction: Ever since I got this new sound system for my bike, I’ve been listening to a lot of podcasts. This American Life is really good. I know, I know, everybody knows that, but it’s true. The only segments I don’t like are the ones that are too “writerly,” when they read a short story aloud. They don’t work for me. Most of the time, though, the show is as great as everyone says it is. Anyway, the other day I listened to program #466: Blackjack . It started with some items on card counting. That stuff is always fun. Then they get to the longer story, which is all about a moderately rich housewife from Iowa who, over a roughly ten-year period, lost her life savings, something like a million dollars, at Harrah’s casinos. Did you know they had casinos in Iowa and Indiana? I didn’t. Anyway, the lady was a gambling addict. That part’s pretty clear. You don’t lose your life savings at a casino by accident. The scary part, though, was how the casino company craftily enabled her to

6 0.97863537 2287 andrew gelman stats-2014-04-09-Advice: positive-sum, zero-sum, or negative-sum

7 0.9783991 395 andrew gelman stats-2010-11-05-Consulting: how do you figure out what to charge?

8 0.97808355 1848 andrew gelman stats-2013-05-09-A tale of two discussion papers

9 0.97767007 2236 andrew gelman stats-2014-03-07-Selection bias in the reporting of shaky research

10 0.97724551 506 andrew gelman stats-2011-01-06-That silly ESP paper and some silliness in a rebuttal as well

11 0.97710693 1395 andrew gelman stats-2012-06-27-Cross-validation (What is it good for?)

12 0.97671813 1435 andrew gelman stats-2012-07-30-Retracted articles and unethical behavior in economics journals?

13 0.97608197 2006 andrew gelman stats-2013-09-03-Evaluating evidence from published research

14 0.97570544 1650 andrew gelman stats-2013-01-03-Did Steven Levitt really believe in 2008 that Obama “would be the greatest president in history”?

15 0.97527981 797 andrew gelman stats-2011-07-11-How do we evaluate a new and wacky claim?

16 0.97527874 1630 andrew gelman stats-2012-12-18-Postdoc positions at Microsoft Research – NYC

17 0.97526711 957 andrew gelman stats-2011-10-14-Questions about a study of charter schools

18 0.97525871 1910 andrew gelman stats-2013-06-22-Struggles over the criticism of the “cannabis users and IQ change” paper

19 0.97524685 1228 andrew gelman stats-2012-03-25-Continuous variables in Bayesian networks

20 0.97494847 1722 andrew gelman stats-2013-02-14-Statistics for firefighters: update