andrew_gelman_stats andrew_gelman_stats-2014 andrew_gelman_stats-2014-2324 knowledge-graph by maker-knowledge-mining

2324 andrew gelman stats-2014-05-07-Once more on nonparametric measures of mutual information


meta infos for this blog

Source: html

Introduction: Ben Murell writes: Our reply to Kinney and Atwal has come out (http://www.pnas.org/content/early/2014/04/29/1403623111.full.pdf) along with their response (http://www.pnas.org/content/early/2014/04/29/1404661111.full.pdf). I feel like they somewhat missed the point. If you’re still interested in this line of discussion, feel free to post, and maybe the Murrells and Kinney can bash it out in your comments! Background: Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets? Heller, Heller, and Gorfine on univariate and multivariate information measures Kinney and Atwal on the maximal information coefficient Mr. Pearson, meet Mr. Mandelbrot: Detecting Novel Associations in Large Data Sets Gorfine, Heller, Heller, Simon, and Tibshirani don’t like MIC The fun thing is that all these people are sending me their papers, and I’m enough of an outsider in this field that each of the


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 Ben Murell writes: Our reply to Kinney and Atwal has come out (http://www. [sent-1, score-0.041]

2 If you’re still interested in this line of discussion, feel free to post, and maybe the Murrells and Kinney can bash it out in your comments! [sent-11, score-0.361]

3 Background: Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets? [sent-12, score-1.018]

4 Heller, Heller, and Gorfine on univariate and multivariate information measures Kinney and Atwal on the maximal information coefficient Mr. [sent-13, score-0.684]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('heller', 0.494), ('kinney', 0.384), ('atwal', 0.256), ('gorfine', 0.247), ('mic', 0.24), ('associations', 0.203), ('multivariate', 0.168), ('murell', 0.142), ('mc', 0.142), ('govern', 0.134), ('bash', 0.128), ('http', 0.128), ('mandelbrot', 0.11), ('maximal', 0.11), ('univariate', 0.11), ('bivariate', 0.106), ('tibshirani', 0.106), ('detecting', 0.104), ('simon', 0.103), ('pearson', 0.103), ('outsider', 0.099), ('papers', 0.096), ('attempts', 0.092), ('meet', 0.091), ('summarize', 0.088), ('ben', 0.088), ('feel', 0.088), ('sending', 0.088), ('novel', 0.087), ('large', 0.085), ('missed', 0.084), ('datasets', 0.083), ('coefficient', 0.076), ('information', 0.075), ('principles', 0.072), ('measures', 0.07), ('enough', 0.07), ('sets', 0.069), ('somewhat', 0.067), ('fun', 0.062), ('background', 0.061), ('field', 0.056), ('sent', 0.055), ('free', 0.052), ('line', 0.051), ('response', 0.048), ('along', 0.047), ('comments', 0.045), ('interested', 0.042), ('reply', 0.041)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 1.0 2324 andrew gelman stats-2014-05-07-Once more on nonparametric measures of mutual information

Introduction: Ben Murell writes: Our reply to Kinney and Atwal has come out (http://www.pnas.org/content/early/2014/04/29/1403623111.full.pdf) along with their response (http://www.pnas.org/content/early/2014/04/29/1404661111.full.pdf). I feel like they somewhat missed the point. If you’re still interested in this line of discussion, feel free to post, and maybe the Murrells and Kinney can bash it out in your comments! Background: Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets? Heller, Heller, and Gorfine on univariate and multivariate information measures Kinney and Atwal on the maximal information coefficient Mr. Pearson, meet Mr. Mandelbrot: Detecting Novel Associations in Large Data Sets Gorfine, Heller, Heller, Simon, and Tibshirani don’t like MIC The fun thing is that all these people are sending me their papers, and I’m enough of an outsider in this field that each of the

2 0.49725053 2314 andrew gelman stats-2014-05-01-Heller, Heller, and Gorfine on univariate and multivariate information measures

Introduction: Malka Gorfine writes: We noticed that the important topic of association measures and tests came up again in your blog, and we have few comments in this regard. It is useful to distinguish between the univariate and multivariate methods. A consistent multivariate method can recognise dependence between two vectors of random variables, while a univariate method can only loop over pairs of components and check for dependency between them. There are very few consistent multivariate methods. To the best of our knowledge there are three practical methods: 1) HSIC by Gretton et al. (http://www.gatsby.ucl.ac.uk/~gretton/papers/GreBouSmoSch05.pdf) 2) dcov by Szekely et al. (http://projecteuclid.org/euclid.aoas/1267453933) 3) A method we introduced in Heller et al (Biometrika, 2013, 503—510, http://biomet.oxfordjournals.org/content/early/2012/12/04/biomet.ass070.full.pdf+html, and an R package, HHG, is available as well http://cran.r-project.org/web/packages/HHG/index.html). A

3 0.43903872 2247 andrew gelman stats-2014-03-14-The maximal information coefficient

Introduction: Justin Kinney writes: I wanted to let you know that the critique Mickey Atwal and I wrote regarding equitability and the maximal information coefficient has just been published . We discussed this paper last year, under the heading, Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets? Kinney and Atwal’s paper is interesting, with my only criticism being that in some places they seem to aim for what might not be possible. For example, they write that “mutual information is already widely believed to quantify dependencies without bias for relationships of one type or another,” which seems a bit vague to me. And later they write, “How to compute such an estimate that does not bias the resulting mutual information value remains an open problem,” which seems to me to miss the point in that unbiased statistical estimates are not generally possible and indeed are often not desirable. Their

4 0.30710456 1706 andrew gelman stats-2013-02-04-Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets?

Introduction: Justin Kinney writes: Since your blog has discussed the “maximal information coefficient” (MIC) of Reshef et al., I figured you might want to see the critique that Gurinder Atwal and I have posted. In short, Reshef et al.’s central claim that MIC is “equitable” is incorrect. We [Kinney and Atwal] offer mathematical proof that the definition of “equitability” Reshef et al. propose is unsatisfiable—no nontrivial dependence measure, including MIC, has this property. Replicating the simulations in their paper with modestly larger data sets validates this finding. The heuristic notion of equitability, however, can be formalized instead as a self-consistency condition closely related to the Data Processing Inequality. Mutual information satisfies this new definition of equitability but MIC does not. We therefore propose that simply estimating mutual information will, in many cases, provide the sort of dependence measure Reshef et al. seek. For background, here are my two p

5 0.29578829 2310 andrew gelman stats-2014-04-28-On deck this week

Introduction: Mon : Crowdstorming a dataset Tues : Ken Rice presents a unifying approach to statistical inference and hypothesis testing Wed : The health policy innovation center: how best to move from pilot studies to large-scale practice? Thurs : Heller, Heller, and Gorfine on univariate and multivariate information measures Fri : Discovering general multidimensional associations Sat : “The graph clearly shows that mammography adds virtually nothing to survival and if anything, decreases survival (and increases cost and provides unnecessary treatment)” Sun : Honored oldsters write about statistics

6 0.28278843 1230 andrew gelman stats-2012-03-26-Further thoughts on nonparametric correlation measures

7 0.15816242 2264 andrew gelman stats-2014-03-24-On deck this month

8 0.12625396 2315 andrew gelman stats-2014-05-02-Discovering general multidimensional associations

9 0.086532034 1996 andrew gelman stats-2013-08-24-All inference is about generalizing from sample to population

10 0.082980797 1726 andrew gelman stats-2013-02-18-What to read to catch up on multivariate statistics?

11 0.068589367 1380 andrew gelman stats-2012-06-15-Coaching, teaching, and writing

12 0.052085999 1784 andrew gelman stats-2013-04-01-Wolfram on Mandelbrot

13 0.051369648 2245 andrew gelman stats-2014-03-12-More on publishing in journals

14 0.049189508 2240 andrew gelman stats-2014-03-10-On deck this week: Things people sent me

15 0.049018048 1878 andrew gelman stats-2013-05-31-How to fix the tabloids? Toward replicable social science research

16 0.04811278 1951 andrew gelman stats-2013-07-22-Top 5 stat papers since 2000?

17 0.046243314 109 andrew gelman stats-2010-06-25-Classics of statistics

18 0.045374319 1928 andrew gelman stats-2013-07-06-How to think about papers published in low-grade journals?

19 0.04525508 2199 andrew gelman stats-2014-02-04-Widening the goalposts in medical trials

20 0.045013454 1708 andrew gelman stats-2013-02-05-Wouldn’t it be cool if Glenn Hubbard were consulting for Herbalife and I were on the other side?


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.077), (1, 0.004), (2, -0.02), (3, -0.023), (4, 0.025), (5, -0.01), (6, -0.011), (7, -0.042), (8, -0.011), (9, 0.013), (10, 0.021), (11, 0.023), (12, 0.018), (13, 0.012), (14, 0.002), (15, 0.034), (16, 0.022), (17, 0.027), (18, -0.019), (19, -0.043), (20, 0.068), (21, 0.003), (22, 0.079), (23, -0.002), (24, 0.085), (25, 0.058), (26, 0.092), (27, 0.114), (28, 0.151), (29, 0.036), (30, 0.086), (31, 0.097), (32, 0.123), (33, -0.004), (34, 0.065), (35, -0.006), (36, 0.027), (37, 0.013), (38, -0.012), (39, -0.005), (40, 0.051), (41, 0.017), (42, 0.067), (43, 0.028), (44, -0.068), (45, -0.03), (46, 0.139), (47, -0.101), (48, -0.155), (49, 0.034)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.96118128 2324 andrew gelman stats-2014-05-07-Once more on nonparametric measures of mutual information

Introduction: Ben Murell writes: Our reply to Kinney and Atwal has come out (http://www.pnas.org/content/early/2014/04/29/1403623111.full.pdf) along with their response (http://www.pnas.org/content/early/2014/04/29/1404661111.full.pdf). I feel like they somewhat missed the point. If you’re still interested in this line of discussion, feel free to post, and maybe the Murrells and Kinney can bash it out in your comments! Background: Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets? Heller, Heller, and Gorfine on univariate and multivariate information measures Kinney and Atwal on the maximal information coefficient Mr. Pearson, meet Mr. Mandelbrot: Detecting Novel Associations in Large Data Sets Gorfine, Heller, Heller, Simon, and Tibshirani don’t like MIC The fun thing is that all these people are sending me their papers, and I’m enough of an outsider in this field that each of the

2 0.9152689 2314 andrew gelman stats-2014-05-01-Heller, Heller, and Gorfine on univariate and multivariate information measures

Introduction: Malka Gorfine writes: We noticed that the important topic of association measures and tests came up again in your blog, and we have few comments in this regard. It is useful to distinguish between the univariate and multivariate methods. A consistent multivariate method can recognise dependence between two vectors of random variables, while a univariate method can only loop over pairs of components and check for dependency between them. There are very few consistent multivariate methods. To the best of our knowledge there are three practical methods: 1) HSIC by Gretton et al. (http://www.gatsby.ucl.ac.uk/~gretton/papers/GreBouSmoSch05.pdf) 2) dcov by Szekely et al. (http://projecteuclid.org/euclid.aoas/1267453933) 3) A method we introduced in Heller et al (Biometrika, 2013, 503—510, http://biomet.oxfordjournals.org/content/early/2012/12/04/biomet.ass070.full.pdf+html, and an R package, HHG, is available as well http://cran.r-project.org/web/packages/HHG/index.html). A

3 0.85997957 2247 andrew gelman stats-2014-03-14-The maximal information coefficient

Introduction: Justin Kinney writes: I wanted to let you know that the critique Mickey Atwal and I wrote regarding equitability and the maximal information coefficient has just been published . We discussed this paper last year, under the heading, Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets? Kinney and Atwal’s paper is interesting, with my only criticism being that in some places they seem to aim for what might not be possible. For example, they write that “mutual information is already widely believed to quantify dependencies without bias for relationships of one type or another,” which seems a bit vague to me. And later they write, “How to compute such an estimate that does not bias the resulting mutual information value remains an open problem,” which seems to me to miss the point in that unbiased statistical estimates are not generally possible and indeed are often not desirable. Their

4 0.83433586 1230 andrew gelman stats-2012-03-26-Further thoughts on nonparametric correlation measures

Introduction: Malka Gorfine, Ruth Heller, and Yair Heller write a comment on the paper of Reshef et al. that we discussed a few months ago. Just to remind you what’s going on here, here’s my quick summary from December: Reshef et al. propose a new nonlinear R-squared-like measure. Unlike R-squared, this new method depends on a tuning parameter that controls the level of discretization, in a “How long is the coast of Britain” sort of way. The dependence on scale is inevitable for such a general method. Just consider: if you sample 1000 points from the unit bivariate normal distribution, (x,y) ~ N(0,I), you’ll be able to fit them perfectly by a 999-degree polynomial fit to the data. So the scale of the fit matters. The clever idea of the paper is that, instead of going for an absolute measure (which, as we’ve seen, will be scale-dependent), they focus on the problem of summarizing the grid of pairwise dependences in a large set of variables. As they put it: “Imagine a data set with hundreds

5 0.81498039 1706 andrew gelman stats-2013-02-04-Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets?

Introduction: Justin Kinney writes: Since your blog has discussed the “maximal information coefficient” (MIC) of Reshef et al., I figured you might want to see the critique that Gurinder Atwal and I have posted. In short, Reshef et al.’s central claim that MIC is “equitable” is incorrect. We [Kinney and Atwal] offer mathematical proof that the definition of “equitability” Reshef et al. propose is unsatisfiable—no nontrivial dependence measure, including MIC, has this property. Replicating the simulations in their paper with modestly larger data sets validates this finding. The heuristic notion of equitability, however, can be formalized instead as a self-consistency condition closely related to the Data Processing Inequality. Mutual information satisfies this new definition of equitability but MIC does not. We therefore propose that simply estimating mutual information will, in many cases, provide the sort of dependence measure Reshef et al. seek. For background, here are my two p

6 0.73053288 1062 andrew gelman stats-2011-12-16-Mr. Pearson, meet Mr. Mandelbrot: Detecting Novel Associations in Large Data Sets

7 0.61735237 2315 andrew gelman stats-2014-05-02-Discovering general multidimensional associations

8 0.58674693 1825 andrew gelman stats-2013-04-25-It’s binless! A program for computing normalizing functions

9 0.52145827 1828 andrew gelman stats-2013-04-27-Time-Sharing Experiments for the Social Sciences

10 0.50795949 1214 andrew gelman stats-2012-03-15-Of forecasts and graph theory and characterizing a statistical method by the information it uses

11 0.48934782 1380 andrew gelman stats-2012-06-15-Coaching, teaching, and writing

12 0.48371214 2250 andrew gelman stats-2014-03-16-“I have no idea who Catalina Garcia is, but she makes a decent ruler”

13 0.43598792 301 andrew gelman stats-2010-09-28-Correlation, prediction, variation, etc.

14 0.42527124 1107 andrew gelman stats-2012-01-08-More on essentialism

15 0.39638257 1975 andrew gelman stats-2013-08-09-Understanding predictive information criteria for Bayesian models

16 0.39424741 778 andrew gelman stats-2011-06-24-New ideas on DIC from Martyn Plummer and Sumio Watanabe

17 0.39383918 519 andrew gelman stats-2011-01-16-Update on the generalized method of moments

18 0.37892437 1424 andrew gelman stats-2012-07-22-Extreme events as evidence for differences in distributions

19 0.37828362 1490 andrew gelman stats-2012-09-09-I’m still wondering . . .

20 0.37487867 418 andrew gelman stats-2010-11-17-ff


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(16, 0.045), (17, 0.199), (20, 0.016), (24, 0.276), (30, 0.015), (31, 0.016), (40, 0.014), (69, 0.02), (76, 0.043), (77, 0.058), (81, 0.014), (99, 0.158)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.92012674 2324 andrew gelman stats-2014-05-07-Once more on nonparametric measures of mutual information

Introduction: Ben Murell writes: Our reply to Kinney and Atwal has come out (http://www.pnas.org/content/early/2014/04/29/1403623111.full.pdf) along with their response (http://www.pnas.org/content/early/2014/04/29/1404661111.full.pdf). I feel like they somewhat missed the point. If you’re still interested in this line of discussion, feel free to post, and maybe the Murrells and Kinney can bash it out in your comments! Background: Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets? Heller, Heller, and Gorfine on univariate and multivariate information measures Kinney and Atwal on the maximal information coefficient Mr. Pearson, meet Mr. Mandelbrot: Detecting Novel Associations in Large Data Sets Gorfine, Heller, Heller, Simon, and Tibshirani don’t like MIC The fun thing is that all these people are sending me their papers, and I’m enough of an outsider in this field that each of the

2 0.85210675 1706 andrew gelman stats-2013-02-04-Too many MC’s not enough MIC’s, or What principles should govern attempts to summarize bivariate associations in large multivariate datasets?

Introduction: Justin Kinney writes: Since your blog has discussed the “maximal information coefficient” (MIC) of Reshef et al., I figured you might want to see the critique that Gurinder Atwal and I have posted. In short, Reshef et al.’s central claim that MIC is “equitable” is incorrect. We [Kinney and Atwal] offer mathematical proof that the definition of “equitability” Reshef et al. propose is unsatisfiable—no nontrivial dependence measure, including MIC, has this property. Replicating the simulations in their paper with modestly larger data sets validates this finding. The heuristic notion of equitability, however, can be formalized instead as a self-consistency condition closely related to the Data Processing Inequality. Mutual information satisfies this new definition of equitability but MIC does not. We therefore propose that simply estimating mutual information will, in many cases, provide the sort of dependence measure Reshef et al. seek. For background, here are my two p

3 0.83432782 1230 andrew gelman stats-2012-03-26-Further thoughts on nonparametric correlation measures

Introduction: Malka Gorfine, Ruth Heller, and Yair Heller write a comment on the paper of Reshef et al. that we discussed a few months ago. Just to remind you what’s going on here, here’s my quick summary from December: Reshef et al. propose a new nonlinear R-squared-like measure. Unlike R-squared, this new method depends on a tuning parameter that controls the level of discretization, in a “How long is the coast of Britain” sort of way. The dependence on scale is inevitable for such a general method. Just consider: if you sample 1000 points from the unit bivariate normal distribution, (x,y) ~ N(0,I), you’ll be able to fit them perfectly by a 999-degree polynomial fit to the data. So the scale of the fit matters. The clever idea of the paper is that, instead of going for an absolute measure (which, as we’ve seen, will be scale-dependent), they focus on the problem of summarizing the grid of pairwise dependences in a large set of variables. As they put it: “Imagine a data set with hundreds

4 0.83014113 482 andrew gelman stats-2010-12-23-Capitalism as a form of voluntarism

Introduction: Interesting discussion by Alex Tabarrok (following up on an article by Rebecca Solnit) on the continuum between voluntarism (or, more generally, non-cash transactions) and markets with monetary exchange. I just have a few comments of my own: 1. Solnit writes of “the iceberg economy,” which she characterizes as “based on gift economies, barter, mutual aid, and giving without hope of return . . . the relations between friends, between family members, the activities of volunteers or those who have chosen their vocation on principle rather than for profit.” I just wonder whether “barter” completely fits in here. Maybe it depends on context. Sometimes barter is an informal way of keeping track (you help me and I help you), but in settings of low liquidity I could imagine barter being simply an inefficient way of performing an economic transaction. 2. I am no expert on capitalism but my impression is that it’s not just about “competition and selfishness” but also is related to the

5 0.82815665 1092 andrew gelman stats-2011-12-29-More by Berger and me on weakly informative priors

Introduction: A couple days ago we discussed some remarks by Tony O’Hagan and Jim Berger on weakly informative priors. Jim followed up on Deborah Mayo’s blog with this: Objective Bayesian priors are often improper (i.e., have infinite total mass), but this is not a problem when they are developed correctly. But not every improper prior is satisfactory. For instance, the constant prior is known to be unsatisfactory in many situations. The ‘solution’ pseudo-Bayesians often use is to choose a constant prior over a large but bounded set (a ‘weakly informative’ prior), saying it is now proper and so all is well. This is not true; if the constant prior on the whole parameter space is bad, so will be the constant prior over the bounded set. The problem is, in part, that some people confuse proper priors with subjective priors and, having learned that true subjective priors are fine, incorrectly presume that weakly informative proper priors are fine. I have a few reactions to this: 1. I agree

6 0.82814258 938 andrew gelman stats-2011-10-03-Comparing prediction errors

7 0.82598329 2247 andrew gelman stats-2014-03-14-The maximal information coefficient

8 0.82585955 1479 andrew gelman stats-2012-09-01-Mothers and Moms

9 0.82475746 38 andrew gelman stats-2010-05-18-Breastfeeding, infant hyperbilirubinemia, statistical graphics, and modern medicine

10 0.82414067 743 andrew gelman stats-2011-06-03-An argument that can’t possibly make sense

11 0.82407939 1787 andrew gelman stats-2013-04-04-Wanna be the next Tyler Cowen? It’s not as easy as you might think!

12 0.82391793 643 andrew gelman stats-2011-04-02-So-called Bayesian hypothesis testing is just as bad as regular hypothesis testing

13 0.82362998 241 andrew gelman stats-2010-08-29-Ethics and statistics in development research

14 0.82336092 1978 andrew gelman stats-2013-08-12-Fixing the race, ethnicity, and national origin questions on the U.S. Census

15 0.82073766 1376 andrew gelman stats-2012-06-12-Simple graph WIN: the example of birthday frequencies

16 0.81784004 1891 andrew gelman stats-2013-06-09-“Heterogeneity of variance in experimental studies: A challenge to conventional interpretations”

17 0.81399715 545 andrew gelman stats-2011-01-30-New innovations in spam

18 0.81381053 278 andrew gelman stats-2010-09-15-Advice that might make sense for individuals but is negative-sum overall

19 0.81303108 240 andrew gelman stats-2010-08-29-ARM solutions

20 0.81234336 1455 andrew gelman stats-2012-08-12-Probabilistic screening to get an approximate self-weighted sample