andrew_gelman_stats andrew_gelman_stats-2010 andrew_gelman_stats-2010-20 knowledge-graph by maker-knowledge-mining

20 andrew gelman stats-2010-05-07-Bayesian hierarchical model for the prediction of soccer results


meta infos for this blog

Source: html

Introduction: Gianluca Baio sends along this article (coauthored with Marta Blangiardo): The problem of modelling football [soccer] data has become increasingly popular in the last few years and many different models have been proposed with the aim of estimating the characteristics that bring a team to lose or win a game, or to predict the score of a particular match. We propose a Bayesian hierarchical model to address both these aims and test its predictive strength on data about the Italian Serie A championship 1991-1992. To overcome the issue of overshrinkage produced by the Bayesian hierarchical model, we specify a more complex mixture model that results in better fit to the observed data. We test its performance using an example about the Italian Serie A championship 2007-2008. I like the use of the hierarchical model and the focus on prediction. I’m wondering, though, shouldn’t the model include a correlation between the “attack” and “defense” parameters? Or maybe that’s in the m


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 We propose a Bayesian hierarchical model to address both these aims and test its predictive strength on data about the Italian Serie A championship 1991-1992. [sent-2, score-0.937]

2 To overcome the issue of overshrinkage produced by the Bayesian hierarchical model, we specify a more complex mixture model that results in better fit to the observed data. [sent-3, score-0.755]

3 We test its performance using an example about the Italian Serie A championship 2007-2008. [sent-4, score-0.337]

4 I like the use of the hierarchical model and the focus on prediction. [sent-5, score-0.328]

5 I’m wondering, though, shouldn’t the model include a correlation between the “attack” and “defense” parameters? [sent-6, score-0.157]

6 Or maybe that’s in the model but I didn’t notice it. [sent-7, score-0.296]

7 Table 2 breaks every rule in the book–and I don’t mean that in a good way. [sent-9, score-0.167]

8 Too many significant digits (if the 95% interval is [-0. [sent-10, score-0.181]

9 06], then, no, you don’t need to report the posterior mean to four decimal places), no group-level predictors, and the teams are laid out in alphabetical order. [sent-12, score-0.582]

10 Table 3 is much better (although would be much better as a graph, I think). [sent-13, score-0.172]

11 The words on the graph are too tiny–they’re unreadable. [sent-15, score-0.091]

12 And it would be my preference to add a few sentences to the caption to explain what’s going on. [sent-16, score-0.275]

13 Figure 5 is looking pretty, but it reverts to the horrible, horrible alphabetical order–this time being made even worse by putting the alphabet in reverse. [sent-17, score-0.497]

14 And the Winbugs code has that discredited dgamma (epsilon, epsilon) model. [sent-18, score-0.101]

15 ) Not a huge deal, but something I notice because I’ve spent a lot of time thinking about it. [sent-20, score-0.139]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('serie', 0.291), ('italian', 0.25), ('epsilon', 0.25), ('championship', 0.239), ('alphabetical', 0.224), ('table', 0.198), ('hierarchical', 0.171), ('ugly', 0.158), ('model', 0.157), ('horrible', 0.141), ('notice', 0.139), ('alphabet', 0.132), ('oooh', 0.125), ('soccer', 0.119), ('winbugs', 0.112), ('caption', 0.109), ('decimal', 0.106), ('digits', 0.106), ('coauthored', 0.102), ('figure', 0.102), ('discredited', 0.101), ('modelling', 0.099), ('overcome', 0.099), ('test', 0.098), ('aims', 0.096), ('laid', 0.095), ('breaks', 0.092), ('graph', 0.091), ('specify', 0.09), ('propose', 0.089), ('aim', 0.088), ('strength', 0.087), ('better', 0.086), ('increasingly', 0.086), ('football', 0.085), ('tiny', 0.085), ('preference', 0.084), ('sentences', 0.082), ('teams', 0.082), ('characteristics', 0.082), ('flat', 0.081), ('defense', 0.08), ('attack', 0.079), ('produced', 0.076), ('mixture', 0.076), ('interval', 0.075), ('mean', 0.075), ('lose', 0.073), ('proposed', 0.072), ('score', 0.07)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.99999988 20 andrew gelman stats-2010-05-07-Bayesian hierarchical model for the prediction of soccer results

Introduction: Gianluca Baio sends along this article (coauthored with Marta Blangiardo): The problem of modelling football [soccer] data has become increasingly popular in the last few years and many different models have been proposed with the aim of estimating the characteristics that bring a team to lose or win a game, or to predict the score of a particular match. We propose a Bayesian hierarchical model to address both these aims and test its predictive strength on data about the Italian Serie A championship 1991-1992. To overcome the issue of overshrinkage produced by the Bayesian hierarchical model, we specify a more complex mixture model that results in better fit to the observed data. We test its performance using an example about the Italian Serie A championship 2007-2008. I like the use of the hierarchical model and the focus on prediction. I’m wondering, though, shouldn’t the model include a correlation between the “attack” and “defense” parameters? Or maybe that’s in the m

2 0.12615389 61 andrew gelman stats-2010-05-31-A data visualization manifesto

Introduction: Details matter (at least, they do for me), but we don’t yet have a systematic way of going back and forth between the structure of a graph, its details, and the underlying questions that motivate our visualizations. (Cleveland, Wilkinson, and others have written a bit on how to formalize these connections, and I’ve thought about it too, but we have a ways to go.) I was thinking about this difficulty after reading an article on graphics by some computer scientists that was well-written but to me lacked a feeling for the linkages between substantive/statistical goals and graphical details. I have problems with these issues too, and my point here is not to criticize but to move the discussion forward. When thinking about visualization, how important are the details? Aleks pointed me to this article by Jeffrey Heer, Michael Bostock, and Vadim Ogievetsky, “A Tour through the Visualization Zoo: A survey of powerful visualization techniques, from the obvious to the obscure.” Th

3 0.12118496 1765 andrew gelman stats-2013-03-16-Recently in the sister blog

Introduction: 1. New Italian production of Life on Mars . 2. Psychological essentialism in everyday thought .

4 0.1210143 1999 andrew gelman stats-2013-08-27-Bayesian model averaging or fitting a larger model

Introduction: Nick Firoozye writes: I had a question about BMA [Bayesian model averaging] and model combinations in general, and direct it to you since they are a basic form of hierarchical model, albeit in the simplest of forms. I wanted to ask what the underlying assumptions are that could lead to BMA improving on a larger model. I know model combination is a topic of interest in the (frequentist) econometrics community (e.g., Bates & Granger, http://www.jstor.org/discover/10.2307/3008764?uid=3738032&uid;=2&uid;=4&sid;=21101948653381) but at the time it was considered a bit of a puzzle. Perhaps small models combined outperform a big model due to standard errors, insufficient data, etc. But I haven’t seen much in way of Bayesian justification. In simplest terms, you might have a joint density P(Y,theta_1,theta_2) from which you could use the two marginals P(Y,theta_1) and P(Y,theta_2) to derive two separate forecasts. A BMA-er would do a weighted average of the two forecast densities, having p

5 0.1190399 1972 andrew gelman stats-2013-08-07-When you’re planning on fitting a model, build up to it by fitting simpler models first. Then, once you have a model you like, check the hell out of it

Introduction: In response to my remarks on his online book, Think Bayes, Allen Downey wrote: I [Downey] have a question about one of your comments: My [Gelman's] main criticism with both books is that they talk a lot about inference but not so much about model building or model checking (recall the three steps of Bayesian data analysis). I think it’s ok for an introductory book to focus on inference, which of course is central to the data-analytic process—but I’d like them to at least mention that Bayesian ideas arise in model building and model checking as well. This sounds like something I agree with, and one of the things I tried to do in the book is to put modeling decisions front and center. But the word “modeling” is used in lots of ways, so I want to see if we are talking about the same thing. For example, in many chapters, I start with a simple model of the scenario, do some analysis, then check whether the model is good enough, and iterate. Here’s the discussion of modeling

6 0.11686378 1425 andrew gelman stats-2012-07-23-Examples of the use of hierarchical modeling to generalize to new settings

7 0.11636017 1459 andrew gelman stats-2012-08-15-How I think about mixture models

8 0.11326838 2133 andrew gelman stats-2013-12-13-Flexibility is good

9 0.11111827 1403 andrew gelman stats-2012-07-02-Moving beyond hopeless graphics

10 0.11053169 754 andrew gelman stats-2011-06-09-Difficulties with Bayesian model averaging

11 0.10932391 1392 andrew gelman stats-2012-06-26-Occam

12 0.10763197 2224 andrew gelman stats-2014-02-25-Basketball Stats: Don’t model the probability of win, model the expected score differential.

13 0.10728092 870 andrew gelman stats-2011-08-25-Why it doesn’t make sense in general to form confidence intervals by inverting hypothesis tests

14 0.10725413 2109 andrew gelman stats-2013-11-21-Hidden dangers of noninformative priors

15 0.10639762 1913 andrew gelman stats-2013-06-24-Why it doesn’t make sense in general to form confidence intervals by inverting hypothesis tests

16 0.10273442 244 andrew gelman stats-2010-08-30-Useful models, model checking, and external validation: a mini-discussion

17 0.10243525 2299 andrew gelman stats-2014-04-21-Stan Model of the Week: Hierarchical Modeling of Supernovas

18 0.10197522 1695 andrew gelman stats-2013-01-28-Economists argue about Bayes

19 0.098117463 781 andrew gelman stats-2011-06-28-The holes in my philosophy of Bayesian data analysis

20 0.09678638 1247 andrew gelman stats-2012-04-05-More philosophy of Bayes


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.194), (1, 0.118), (2, 0.02), (3, 0.073), (4, 0.061), (5, -0.049), (6, 0.016), (7, -0.001), (8, 0.038), (9, 0.015), (10, 0.03), (11, 0.027), (12, -0.04), (13, -0.033), (14, -0.013), (15, 0.005), (16, 0.068), (17, -0.023), (18, 0.024), (19, -0.024), (20, 0.01), (21, 0.037), (22, 0.005), (23, -0.031), (24, 0.03), (25, -0.029), (26, -0.02), (27, -0.03), (28, 0.007), (29, -0.04), (30, -0.034), (31, -0.06), (32, -0.008), (33, -0.031), (34, 0.016), (35, 0.016), (36, -0.011), (37, -0.037), (38, 0.024), (39, 0.005), (40, 0.022), (41, -0.011), (42, 0.019), (43, 0.02), (44, -0.022), (45, -0.045), (46, -0.001), (47, 0.008), (48, 0.024), (49, -0.0)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.96070993 20 andrew gelman stats-2010-05-07-Bayesian hierarchical model for the prediction of soccer results

Introduction: Gianluca Baio sends along this article (coauthored with Marta Blangiardo): The problem of modelling football [soccer] data has become increasingly popular in the last few years and many different models have been proposed with the aim of estimating the characteristics that bring a team to lose or win a game, or to predict the score of a particular match. We propose a Bayesian hierarchical model to address both these aims and test its predictive strength on data about the Italian Serie A championship 1991-1992. To overcome the issue of overshrinkage produced by the Bayesian hierarchical model, we specify a more complex mixture model that results in better fit to the observed data. We test its performance using an example about the Italian Serie A championship 2007-2008. I like the use of the hierarchical model and the focus on prediction. I’m wondering, though, shouldn’t the model include a correlation between the “attack” and “defense” parameters? Or maybe that’s in the m

2 0.81023145 1723 andrew gelman stats-2013-02-15-Wacky priors can work well?

Introduction: Dave Judkins writes: I would love to see a blog entry on this article , Bayesian Model Selection in High-Dimensional Settings, by Valen Johnson and David Rossell. The simulation results are very encouraging although the choice of colors for some of the graphics is unfortunate. Unless I am colorblind in some way that I am unaware of, they have two thin charcoal lines that are indistinguishable. When Dave Judkins puts in a request, I’ll respond. Also, I’m always happy to see a new Val Johnson paper. Val and I are contemporaries—he and I got our PhD’s at around the same time, with both of us working on Bayesian image reconstruction, then in the early 1990s Val was part of the legendary group at Duke’s Institute of Statistics and Decision Sciences—a veritable ’27 Yankees featuring Mike West, Merlise Clyde, Michael Lavine, Dave Higdon, Peter Mueller, Val, and a bunch of others. I always thought it was too bad they all had to go their separate ways. Val also wrote two classic p

3 0.79648465 1999 andrew gelman stats-2013-08-27-Bayesian model averaging or fitting a larger model

Introduction: Nick Firoozye writes: I had a question about BMA [Bayesian model averaging] and model combinations in general, and direct it to you since they are a basic form of hierarchical model, albeit in the simplest of forms. I wanted to ask what the underlying assumptions are that could lead to BMA improving on a larger model. I know model combination is a topic of interest in the (frequentist) econometrics community (e.g., Bates & Granger, http://www.jstor.org/discover/10.2307/3008764?uid=3738032&uid;=2&uid;=4&sid;=21101948653381) but at the time it was considered a bit of a puzzle. Perhaps small models combined outperform a big model due to standard errors, insufficient data, etc. But I haven’t seen much in way of Bayesian justification. In simplest terms, you might have a joint density P(Y,theta_1,theta_2) from which you could use the two marginals P(Y,theta_1) and P(Y,theta_2) to derive two separate forecasts. A BMA-er would do a weighted average of the two forecast densities, having p

4 0.77687401 1374 andrew gelman stats-2012-06-11-Convergence Monitoring for Non-Identifiable and Non-Parametric Models

Introduction: Becky Passonneau and colleagues at the Center for Computational Learning Systems (CCLS) at Columbia have been working on a project for ConEd (New York’s major electric utility) to rank structures based on vulnerability to secondary events (e.g., transformer explosions, cable meltdowns, electrical fires). They’ve been using the R implementation BayesTree of Chipman, George and McCulloch’s Bayesian Additive Regression Trees (BART). BART is a Bayesian non-parametric method that is non-identifiable in two ways. Firstly, it is an additive tree model with a fixed number of trees, the indexes of which aren’t identified (you get the same predictions in a model swapping the order of the trees). This is the same kind of non-identifiability you get with any mixture model (additive or interpolated) with an exchangeable prior on the mixture components. Secondly, the trees themselves have varying structure over samples in terms of number of nodes and their topology (depth, branching, etc

5 0.77677739 2162 andrew gelman stats-2014-01-08-Belief aggregation

Introduction: Johannes Castner writes: Suppose there are k scientists, each with her own model (Bayesian Net) over m random variables. Then, because the space of Bayesian Nets over these m variables, with the square-root of the Jensen-Shannon Divergence as a distance metric is a closed and bounded space, there exists one unique Bayes Net that is a mixture of the k model joint-distributions which is at equal distance to each of the k models and may be called a “consensus graph.” This consensus graph is in turn a Bayes Net, which can be updated with evidence. The first question is: What are the conditions for which, given a new bit of evidence, the updated consensus graph is exactly the same graph as the consensus graph of the updated k Bayes Nets? In other words, if we arrive at a synthetic model from k models and then update this synthetic model, under what conditions is this the same thing as if we had first updated all k models and then build a synthesis. The second question is: If these ar

6 0.77429795 929 andrew gelman stats-2011-09-27-Visual diagnostics for discrete-data regressions

7 0.77229571 2176 andrew gelman stats-2014-01-19-Transformations for non-normal data

8 0.77226198 1041 andrew gelman stats-2011-12-04-David MacKay and Occam’s Razor

9 0.77156776 1392 andrew gelman stats-2012-06-26-Occam

10 0.7677514 1047 andrew gelman stats-2011-12-08-I Am Too Absolutely Heteroskedastic for This Probit Model

11 0.76596045 1287 andrew gelman stats-2012-04-28-Understanding simulations in terms of predictive inference?

12 0.76425672 1460 andrew gelman stats-2012-08-16-“Real data can be a pain”

13 0.76200223 1208 andrew gelman stats-2012-03-11-Gelman on Hennig on Gelman on Bayes

14 0.75740415 1141 andrew gelman stats-2012-01-28-Using predator-prey models on the Canadian lynx series

15 0.75561005 1395 andrew gelman stats-2012-06-27-Cross-validation (What is it good for?)

16 0.75343788 754 andrew gelman stats-2011-06-09-Difficulties with Bayesian model averaging

17 0.752289 810 andrew gelman stats-2011-07-20-Adding more information can make the variance go up (depending on your model)

18 0.74979752 1459 andrew gelman stats-2012-08-15-How I think about mixture models

19 0.74547619 1156 andrew gelman stats-2012-02-06-Bayesian model-building by pure thought: Some principles and examples

20 0.74065703 1431 andrew gelman stats-2012-07-27-Overfitting


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(0, 0.034), (16, 0.057), (21, 0.049), (24, 0.158), (43, 0.01), (44, 0.039), (53, 0.012), (55, 0.016), (58, 0.011), (63, 0.038), (79, 0.011), (86, 0.063), (92, 0.12), (95, 0.033), (99, 0.236)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.94421482 20 andrew gelman stats-2010-05-07-Bayesian hierarchical model for the prediction of soccer results

Introduction: Gianluca Baio sends along this article (coauthored with Marta Blangiardo): The problem of modelling football [soccer] data has become increasingly popular in the last few years and many different models have been proposed with the aim of estimating the characteristics that bring a team to lose or win a game, or to predict the score of a particular match. We propose a Bayesian hierarchical model to address both these aims and test its predictive strength on data about the Italian Serie A championship 1991-1992. To overcome the issue of overshrinkage produced by the Bayesian hierarchical model, we specify a more complex mixture model that results in better fit to the observed data. We test its performance using an example about the Italian Serie A championship 2007-2008. I like the use of the hierarchical model and the focus on prediction. I’m wondering, though, shouldn’t the model include a correlation between the “attack” and “defense” parameters? Or maybe that’s in the m

2 0.94253242 287 andrew gelman stats-2010-09-20-Paul Rosenbaum on those annoying pre-treatment variables that are sort-of instruments and sort-of covariates

Introduction: Last year we discussed an important challenge in causal inference: The standard advice (given in many books, including ours) for causal inference is to control for relevant pre-treatment variables as much as possible. But, as Judea Pearl has pointed out, instruments (as in “instrumental variables”) are pre-treatment variables that we would not want to “control for” in a matching or regression sense. At first, this seems like a minor modification, with the new recommendation being to apply instrumental variables estimation using all pre-treatment instruments, and to control for all other pre-treatment variables. But that can’t really work as general advice. What about weak instruments or covariates that have some instrumental aspects? I asked Paul Rosenbaum for his thoughts on the matter, and he wrote the following: In section 18.2 of Design of Observational Studies (DOS), I [Rosenbaum] discuss “seemingly innocuous confounding” defined to be a covariate that predicts a su

3 0.93037164 1720 andrew gelman stats-2013-02-12-That claim that Harvard admissions discriminate in favor of Jews? After seeing the statistics, I don’t see it.

Introduction: A few months ago we discussed Ron Unz’s claim that Jews are massively overrepresented in Ivy League college admissions, not just in comparison to the general population of college-age Americans, but even in comparison to other white kids with comparable academic ability and preparation. Most of Unz’s article concerns admissions of Asian-Americans, and he also has a proposal to admit certain students at random (see my discussion in the link above). In the present post, I concentrate on the statistics about Jewish students, because this is where I have learned that his statistics are particularly suspect, with various numbers being off by factors of 2 or 4 or more. Unz’s article was discussed, largely favorably, by academic bloggers Tyler Cowen , Steve Hsu , and . . . me! Hsu writes: “Don’t miss the statistical supplement.” But a lot of our trust in those statistics seems to be misplaced. Some people have sent me some information showing serious problems with Unz’s methods

4 0.92792708 1563 andrew gelman stats-2012-11-05-Someone is wrong on the internet, part 2

Introduction: My coblogger John Sides feeds a troll. It’s a tough call. Yesterday I gave my reasoning for ignoring these provocateurs, but in this case the troll in question is writing for a major newspaper so it makes sense for John to go to the trouble of shooting him down. Even though I suspect the columnist was trolling for no better reason than . . . he had a deadline and nothing to say so he thought he’d wade into a controversy. On the plus side, as a statistician I’m happy that statistics is considered important enough that it’s worth trolling! When they start attacking like this, they must feel a bit on the defensive. . . .

5 0.92011106 1004 andrew gelman stats-2011-11-11-Kaiser Fung on how not to critique models

Introduction: In the context of a debate between economists Brad DeLong and Tyler Cowen on the “IS-LM model” [no, I don't know what it is, either!], Kaiser writes : Since a model is an abstraction, a simplification of reality, no model is above critique. I [Kaiser] consider the following types of critique not deserving: 1) The critique that the modeler makes an assumption 2) The critique that the modeler makes an assumption for mathematical convenience 3) The critique that the model omits some feature 4) The critique that the model doesn’t fit one’s intuition 5) The critique that the model fails to make a specific prediction Above all, a serious critique must include an alternative model that is provably better than the one it criticises. It is not enough to show that the alternative solves the problems being pointed out; the alternative must do so while preserving the useful aspects of the model being criticized. I have mixed feelings about Kaiser’s rules. On one hand, I agree wit

6 0.90136594 2073 andrew gelman stats-2013-10-22-Ivy Jew update

7 0.90029228 1108 andrew gelman stats-2012-01-09-Blogging, polemical and otherwise

8 0.89545858 301 andrew gelman stats-2010-09-28-Correlation, prediction, variation, etc.

9 0.89485276 1743 andrew gelman stats-2013-02-28-Different modes of discourse

10 0.89312816 1785 andrew gelman stats-2013-04-02-So much artistic talent

11 0.89218092 1730 andrew gelman stats-2013-02-20-Unz on Unz

12 0.89081573 788 andrew gelman stats-2011-07-06-Early stopping and penalized likelihood

13 0.89052832 2112 andrew gelman stats-2013-11-25-An interesting but flawed attempt to apply general forecasting principles to contextualize attitudes toward risks of global warming

14 0.890288 1751 andrew gelman stats-2013-03-06-Janet Mertz’s response to “The Myth of American Meritocracy”

15 0.88961732 351 andrew gelman stats-2010-10-18-“I was finding the test so irritating and boring that I just started to click through as fast as I could”

16 0.88932765 2149 andrew gelman stats-2013-12-26-Statistical evidence for revised standards

17 0.88925338 1117 andrew gelman stats-2012-01-13-What are the important issues in ethics and statistics? I’m looking for your input!

18 0.88903928 2281 andrew gelman stats-2014-04-04-The Notorious N.H.S.T. presents: Mo P-values Mo Problems

19 0.88896924 502 andrew gelman stats-2011-01-04-Cash in, cash out graph

20 0.88882917 693 andrew gelman stats-2011-05-04-Don’t any statisticians work for the IRS?