hunch_net hunch_net-2005 hunch_net-2005-86 knowledge-graph by maker-knowledge-mining

86 hunch net-2005-06-28-The cross validation problem: cash reward


meta infos for this blog

Source: html

Introduction: I just presented the cross validation problem at COLT . The problem now has a cash prize (up to $500) associated with it—see the presentation for details. The write-up for colt .


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 I just presented the cross validation problem at COLT . [sent-1, score-1.163]

2 The problem now has a cash prize (up to $500) associated with it—see the presentation for details. [sent-2, score-1.507]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('cash', 0.425), ('colt', 0.42), ('validation', 0.352), ('prize', 0.337), ('cross', 0.324), ('presented', 0.302), ('presentation', 0.28), ('associated', 0.28), ('problem', 0.185), ('see', 0.135)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 1.0 86 hunch net-2005-06-28-The cross validation problem: cash reward

Introduction: I just presented the cross validation problem at COLT . The problem now has a cash prize (up to $500) associated with it—see the presentation for details. The write-up for colt .

2 0.30457109 26 hunch net-2005-02-21-Problem: Cross Validation

Introduction: The essential problem here is the large gap between experimental observation and theoretical understanding. Method K-fold cross validation is a commonly used technique which takes a set of m examples and partitions them into K sets (“folds”) of size m/K . For each fold, a classifier is trained on the other folds and then test on the fold. Problem Assume only independent samples. Derive a classifier from the K classifiers with a small bound on the true error rate. Past Work (I’ll add more as I remember/learn.) Devroye , Rogers, and Wagner analyzed cross validation and found algorithm specific bounds. Not all of this is online, but here is one paper . Michael Kearns and Dana Ron analyzed cross validation and found that under additional stability assumptions the bound for the classifier which learns on all the data is not much worse than for a test set of size m/K . Avrim Blum, Adam Kalai , and myself analyzed cross validation and found tha

3 0.20005669 453 hunch net-2012-01-28-Why COLT?

Introduction: By Shie and Nati Following John’s advertisement for submitting to ICML, we thought it appropriate to highlight the advantages of COLT, and the reasons it is often the best place for theory papers. We would like to emphasize that we both respect ICML, and are active in ICML, both as authors and as area chairs, and certainly are not arguing that ICML is a bad place for your papers. For many papers, ICML is the best venue. But for many theory papers, COLT is a better and more appropriate place. Why should you submit to COLT? By-and-large, theory papers go to COLT. This is the tradition of the field and most theory papers are sent to COLT. This is the place to present your ground-breaking theorems and new models that will shape the theory of machine learning. COLT is more focused then ICML with a single track session. Unlike ICML, the norm in COLT is for people to sit through most sessions, and hear most of the talks presented. There is also often a lively discussion followi

4 0.18893574 89 hunch net-2005-07-04-The Health of COLT

Introduction: The health of COLT (Conference on Learning Theory or Computational Learning Theory depending on who you ask) has been questioned over the last few years. Low points for the conference occurred when EuroCOLT merged with COLT in 2001, and the attendance at the 2002 Sydney COLT fell to a new low. This occurred in the general context of machine learning conferences rising in both number and size over the last decade. Any discussion of why COLT has had difficulties is inherently controversial as is any story about well-intentioned people making the wrong decisions. Nevertheless, this may be worth discussing in the hope of avoiding problems in the future and general understanding. In any such discussion there is a strong tendency to identify with a conference/community in a patriotic manner that is detrimental to thinking. Keep in mind that conferences exist to further research. My understanding (I wasn’t around) is that COLT started as a subcommunity of the computer science

5 0.13673426 307 hunch net-2008-07-04-More Presentation Preparation

Introduction: We’ve discussed presentation preparation before , but I have one more thing to add: transitioning . For a research presentation, it is substantially helpful for the audience if transitions are clear. A common outline for a research presentation in machine leanring is: The problem . Presentations which don’t describe the problem almost immediately lose people, because the context is missing to understand the detail. Prior relevant work . In many cases, a paper builds on some previous bit of work which must be understood in order to understand what the paper does. A common failure mode seems to be spending too much time on prior work. Discuss just the relevant aspects of prior work in the language of your work. Sometimes this is missing when unneeded. What we did . For theory papers in particular, it is often not possible to really cover the details. Prioritizing what you present can be very important. How it worked . Many papers in Machine Learning have some sor

6 0.13425569 270 hunch net-2007-11-02-The Machine Learning Award goes to …

7 0.13142632 427 hunch net-2011-03-20-KDD Cup 2011

8 0.11756071 336 hunch net-2009-01-19-Netflix prize within epsilon

9 0.11244772 324 hunch net-2008-11-09-A Healthy COLT

10 0.11232358 458 hunch net-2012-03-06-COLT-ICML Open Questions and ICML Instructions

11 0.10435769 44 hunch net-2005-03-21-Research Styles in Machine Learning

12 0.10235791 394 hunch net-2010-04-24-COLT Treasurer is now Phil Long

13 0.090317301 19 hunch net-2005-02-14-Clever Methods of Overfitting

14 0.079708382 371 hunch net-2009-09-21-Netflix finishes (and starts)

15 0.076763928 41 hunch net-2005-03-15-The State of Tight Bounds

16 0.076426283 186 hunch net-2006-06-24-Online convex optimization at COLT

17 0.075936288 85 hunch net-2005-06-28-A COLT paper

18 0.074045911 430 hunch net-2011-04-11-The Heritage Health Prize

19 0.073386885 452 hunch net-2012-01-04-Why ICML? and the summer conferences

20 0.072477013 326 hunch net-2008-11-11-COLT CFP


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.087), (1, -0.053), (2, 0.03), (3, -0.064), (4, -0.012), (5, -0.064), (6, -0.016), (7, 0.002), (8, -0.043), (9, -0.031), (10, -0.037), (11, 0.24), (12, -0.058), (13, 0.114), (14, 0.205), (15, -0.098), (16, 0.209), (17, 0.049), (18, -0.16), (19, 0.0), (20, -0.119), (21, 0.101), (22, 0.121), (23, 0.043), (24, 0.04), (25, 0.117), (26, 0.001), (27, 0.086), (28, 0.012), (29, 0.076), (30, -0.024), (31, -0.042), (32, -0.026), (33, 0.054), (34, 0.021), (35, 0.207), (36, -0.035), (37, 0.049), (38, -0.061), (39, 0.001), (40, -0.058), (41, 0.028), (42, -0.086), (43, 0.008), (44, -0.011), (45, -0.029), (46, -0.071), (47, 0.011), (48, 0.096), (49, -0.013)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.98355228 86 hunch net-2005-06-28-The cross validation problem: cash reward

Introduction: I just presented the cross validation problem at COLT . The problem now has a cash prize (up to $500) associated with it—see the presentation for details. The write-up for colt .

2 0.6256451 324 hunch net-2008-11-09-A Healthy COLT

Introduction: A while ago , we discussed the health of COLT . COLT 2008 substantially addressed my concerns. The papers were diverse and several were interesting. Attendance was up, which is particularly notable in Europe. In my opinion, the colocation with UAI and ICML was the best colocation since 1998. And, perhaps best of all, registration ended up being free for all students due to various grants from the Academy of Finland , Google , IBM , and Yahoo . A basic question is: what went right? There seem to be several answers. Cost-wise, COLT had sufficient grants to alleviate the high cost of the Euro and location at a university substantially reduces the cost compared to a hotel. Organization-wise, the Finns were great with hordes of volunteers helping set everything up. Having too many volunteers is a good failure mode. Organization-wise, it was clear that all 3 program chairs were cooperating in designing the program. Facilities-wise, proximity in time and space made

3 0.55283052 26 hunch net-2005-02-21-Problem: Cross Validation

Introduction: The essential problem here is the large gap between experimental observation and theoretical understanding. Method K-fold cross validation is a commonly used technique which takes a set of m examples and partitions them into K sets (“folds”) of size m/K . For each fold, a classifier is trained on the other folds and then test on the fold. Problem Assume only independent samples. Derive a classifier from the K classifiers with a small bound on the true error rate. Past Work (I’ll add more as I remember/learn.) Devroye , Rogers, and Wagner analyzed cross validation and found algorithm specific bounds. Not all of this is online, but here is one paper . Michael Kearns and Dana Ron analyzed cross validation and found that under additional stability assumptions the bound for the classifier which learns on all the data is not much worse than for a test set of size m/K . Avrim Blum, Adam Kalai , and myself analyzed cross validation and found tha

4 0.55223769 89 hunch net-2005-07-04-The Health of COLT

Introduction: The health of COLT (Conference on Learning Theory or Computational Learning Theory depending on who you ask) has been questioned over the last few years. Low points for the conference occurred when EuroCOLT merged with COLT in 2001, and the attendance at the 2002 Sydney COLT fell to a new low. This occurred in the general context of machine learning conferences rising in both number and size over the last decade. Any discussion of why COLT has had difficulties is inherently controversial as is any story about well-intentioned people making the wrong decisions. Nevertheless, this may be worth discussing in the hope of avoiding problems in the future and general understanding. In any such discussion there is a strong tendency to identify with a conference/community in a patriotic manner that is detrimental to thinking. Keep in mind that conferences exist to further research. My understanding (I wasn’t around) is that COLT started as a subcommunity of the computer science

5 0.49956667 453 hunch net-2012-01-28-Why COLT?

Introduction: By Shie and Nati Following John’s advertisement for submitting to ICML, we thought it appropriate to highlight the advantages of COLT, and the reasons it is often the best place for theory papers. We would like to emphasize that we both respect ICML, and are active in ICML, both as authors and as area chairs, and certainly are not arguing that ICML is a bad place for your papers. For many papers, ICML is the best venue. But for many theory papers, COLT is a better and more appropriate place. Why should you submit to COLT? By-and-large, theory papers go to COLT. This is the tradition of the field and most theory papers are sent to COLT. This is the place to present your ground-breaking theorems and new models that will shape the theory of machine learning. COLT is more focused then ICML with a single track session. Unlike ICML, the norm in COLT is for people to sit through most sessions, and hear most of the talks presented. There is also often a lively discussion followi

6 0.49476627 458 hunch net-2012-03-06-COLT-ICML Open Questions and ICML Instructions

7 0.48244223 394 hunch net-2010-04-24-COLT Treasurer is now Phil Long

8 0.47588393 47 hunch net-2005-03-28-Open Problems for Colt

9 0.47014472 88 hunch net-2005-07-01-The Role of Impromptu Talks

10 0.40162951 292 hunch net-2008-03-15-COLT Open Problems

11 0.39460251 336 hunch net-2009-01-19-Netflix prize within epsilon

12 0.3906821 429 hunch net-2011-04-06-COLT open questions

13 0.37192354 44 hunch net-2005-03-21-Research Styles in Machine Learning

14 0.37005308 307 hunch net-2008-07-04-More Presentation Preparation

15 0.36312598 362 hunch net-2009-06-26-Netflix nearly done

16 0.35500717 326 hunch net-2008-11-11-COLT CFP

17 0.31858081 430 hunch net-2011-04-11-The Heritage Health Prize

18 0.31040433 270 hunch net-2007-11-02-The Machine Learning Award goes to …

19 0.30325833 427 hunch net-2011-03-20-KDD Cup 2011

20 0.29363087 371 hunch net-2009-09-21-Netflix finishes (and starts)


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(4, 0.755)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.95957065 86 hunch net-2005-06-28-The cross validation problem: cash reward

Introduction: I just presented the cross validation problem at COLT . The problem now has a cash prize (up to $500) associated with it—see the presentation for details. The write-up for colt .

2 0.88015831 108 hunch net-2005-09-06-A link

Introduction: I read through some of the essays of Michael Nielsen today, and recommend them. Principles of Effective Research and Extreme Thinking are both relevant to several discussions here.

3 0.59835654 88 hunch net-2005-07-01-The Role of Impromptu Talks

Introduction: COLT had an impromptu session which seemed as interesting or more interesting than any other single technical session (despite being only an hour long). There are several roles that an impromptu session can play including: Announcing new work since the paper deadline. Letting this happen now rather than later helps aid the process of research. Discussing a paper that was rejected. Reviewers err sometimes and an impromptu session provides a means to remedy that. Entertainment. We all like to have a bit of fun. For design, the following seem important: Impromptu speakers should not have much time. At COLT, it was 8 minutes, but I have seen even 5 work well. The entire impromptu session should not last too long because the format is dense and promotes restlessness. A half hour or hour can work well. Impromptu talks are a mechanism to let a little bit of chaos into the schedule. They will be chaotic in content, presentation, and usefulness. The fundamental adv

4 0.5370959 10 hunch net-2005-02-02-Kolmogorov Complexity and Googling

Introduction: Machine learning makes the New Scientist . From the article: COMPUTERS can learn the meaning of words simply by plugging into Google. The finding could bring forward the day that true artificial intelligence is developed‌. But Paul Vitanyi and Rudi Cilibrasi of the National Institute for Mathematics and Computer Science in Amsterdam, the Netherlands, realised that a Google search can be used to measure how closely two words relate to each other. For instance, imagine a computer needs to understand what a hat is. You can read the paper at KC Google . Hat tip: Kolmogorov Mailing List Any thoughts on the paper?

5 0.50880241 29 hunch net-2005-02-25-Solution: Reinforcement Learning with Classification

Introduction: I realized that the tools needed to solve the problem just posted were just created. I tried to sketch out the solution here (also in .lyx and .tex ). It is still quite sketchy (and probably only the few people who understand reductions well can follow). One of the reasons why I started this weblog was to experiment with “research in the open”, and this is an opportunity to do so. Over the next few days, I’ll be filling in details and trying to get things to make sense. If you have additions or ideas, please propose them.

6 0.34676418 129 hunch net-2005-11-07-Prediction Competitions

7 0.28896374 75 hunch net-2005-05-28-Running A Machine Learning Summer School

8 0.055006828 141 hunch net-2005-12-17-Workshops as Franchise Conferences

9 0.046930782 439 hunch net-2011-08-01-Interesting papers at COLT 2011

10 0.046810042 458 hunch net-2012-03-06-COLT-ICML Open Questions and ICML Instructions

11 0.039030582 204 hunch net-2006-08-28-Learning Theory standards for NIPS 2006

12 0.038251411 19 hunch net-2005-02-14-Clever Methods of Overfitting

13 0.036113676 94 hunch net-2005-07-13-Text Entailment at AAAI

14 0.036084965 143 hunch net-2005-12-27-Automated Labeling

15 0.034311518 186 hunch net-2006-06-24-Online convex optimization at COLT

16 0.03289257 203 hunch net-2006-08-18-Report of MLSS 2006 Taipei

17 0.02785209 51 hunch net-2005-04-01-The Producer-Consumer Model of Research

18 0.025272464 349 hunch net-2009-04-21-Interesting Presentations at Snowbird

19 0.025126314 487 hunch net-2013-07-24-ICML 2012 videos lost

20 0.0 1 hunch net-2005-01-19-Why I decided to run a weblog.