hunch_net hunch_net-2010 hunch_net-2010-418 knowledge-graph by maker-knowledge-mining
Source: html
Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.
sentIndex sentText sentNum sentScore
1 Slashdot points out the Traffic Prediction Challenge which looks pretty fun. [sent-1, score-0.696]
2 The temporal aspect seems to be very common in many real-world problems and somewhat understudied. [sent-2, score-1.369]
wordName wordTfidf (topN-words)
[('traffic', 0.476), ('temporal', 0.441), ('slashdot', 0.381), ('aspect', 0.296), ('looks', 0.286), ('challenge', 0.248), ('somewhat', 0.234), ('pretty', 0.217), ('points', 0.193), ('prediction', 0.147), ('common', 0.14), ('problems', 0.104), ('seems', 0.096), ('many', 0.058)]
simIndex simValue blogId blogTitle
same-blog 1 1.0 418 hunch net-2010-12-02-Traffic Prediction Problem
Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.
2 0.10928958 92 hunch net-2005-07-11-AAAI blog
Introduction: The AAAI conference is running a student blog which looks like a fun experiment.
3 0.10436411 271 hunch net-2007-11-05-CMU wins DARPA Urban Challenge
Introduction: The results have been posted , with CMU first , Stanford second , and Virginia Tech Third . Considering that this was an open event (at least for people in the US), this was a very strong showing for research at universities (instead of defense contractors, for example). Some details should become public at the NIPS workshops . Slashdot has a post with many comments.
4 0.10083219 239 hunch net-2007-04-18-$50K Spock Challenge
Introduction: Apparently, the company Spock is setting up a $50k entity resolution challenge . $50k is much less than the Netflix challenge, but it’s effectively the same as Netflix until someone reaches 10% . It’s also nice that the Spock challenge has a short duration. The (visible) test set is of size 25k and the training set has size 75k.
5 0.096672006 200 hunch net-2006-08-03-AOL’s data drop
Introduction: AOL has released several large search engine related datasets. This looks like a pretty impressive data release, and it is a big opportunity for people everywhere to worry about search engine related learning problems, if they want.
6 0.094991684 364 hunch net-2009-07-11-Interesting papers at KDD
7 0.091941655 399 hunch net-2010-05-20-Google Predict
8 0.082059875 446 hunch net-2011-10-03-Monday announcements
9 0.080537446 190 hunch net-2006-07-06-Branch Prediction Competition
10 0.06701488 129 hunch net-2005-11-07-Prediction Competitions
11 0.065326117 326 hunch net-2008-11-11-COLT CFP
12 0.062750414 275 hunch net-2007-11-29-The Netflix Crack
13 0.06119936 291 hunch net-2008-03-07-Spock Challenge Winners
14 0.060982488 384 hunch net-2009-12-24-Top graduates this season
15 0.05958163 151 hunch net-2006-01-25-1 year
16 0.053940065 433 hunch net-2011-04-23-ICML workshops due
17 0.052791297 332 hunch net-2008-12-23-Use of Learning Theory
18 0.052005105 211 hunch net-2006-10-02-$1M Netflix prediction contest
19 0.050370209 47 hunch net-2005-03-28-Open Problems for Colt
20 0.049578734 406 hunch net-2010-08-22-KDD 2010
topicId topicWeight
[(0, 0.071), (1, 0.002), (2, -0.023), (3, -0.014), (4, -0.009), (5, -0.003), (6, -0.036), (7, -0.023), (8, 0.024), (9, -0.045), (10, -0.034), (11, 0.127), (12, -0.074), (13, 0.015), (14, -0.037), (15, 0.015), (16, 0.004), (17, -0.02), (18, -0.002), (19, 0.014), (20, -0.041), (21, -0.041), (22, -0.037), (23, 0.004), (24, 0.011), (25, -0.025), (26, 0.061), (27, 0.023), (28, -0.113), (29, 0.037), (30, 0.099), (31, 0.066), (32, -0.05), (33, 0.059), (34, 0.018), (35, -0.129), (36, -0.023), (37, 0.07), (38, -0.008), (39, 0.036), (40, 0.016), (41, -0.014), (42, 0.026), (43, -0.023), (44, 0.092), (45, -0.021), (46, 0.025), (47, 0.037), (48, -0.072), (49, -0.041)]
simIndex simValue blogId blogTitle
same-blog 1 0.9712038 418 hunch net-2010-12-02-Traffic Prediction Problem
Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.
2 0.53369242 190 hunch net-2006-07-06-Branch Prediction Competition
Introduction: Alan Fern points out the second branch prediction challenge (due September 29) which is a follow up to the first branch prediction competition . Branch prediction is one of the fundamental learning problems of the computer age: without it our computers might run an order of magnitude slower. This is a tough problem since there are sharp constraints on time and space complexity in an online environment. For machine learning, the “idealistic track” may fit well. Essentially, they remove these constraints to gain a weak upper bound on what might be done.
3 0.52884352 427 hunch net-2011-03-20-KDD Cup 2011
Introduction: Yehuda points out KDD-Cup 2011 which Markus and Gideon helped setup. This is a prediction and recommendation contest for music. In addition to being a fun chance to show your expertise, there are cash prizes of $5K/$2K/$1K.
4 0.49368921 446 hunch net-2011-10-03-Monday announcements
Introduction: Various people want to use hunch.net to announce things. I’ve generally resisted this because I feared hunch becoming a pure announcement zone while I am much more interested contentful posts and discussion personally. Nevertheless there is clearly some value and announcements are easy, so I’m planning to summarize announcements on Mondays. D. Sculley points out an interesting Semisupervised feature learning competition, with a deadline of October 17. Lihong Li points out the webscope user interaction dataset which is the first high quality exploration dataset I’m aware of that is publicly available. Seth Rogers points out CrossValidated which looks similar in conception to metaoptimize , but directly using the stackoverflow interface and with a bit more of a statistics twist.
5 0.43867168 211 hunch net-2006-10-02-$1M Netflix prediction contest
Introduction: Netflix is running a contest to improve recommender prediction systems. A 10% improvement over their current system yields a $1M prize. Failing that, the best smaller improvement yields a smaller $50K prize. This contest looks quite real, and the $50K prize money is almost certainly achievable with a bit of thought. The contest also comes with a dataset which is apparently 2 orders of magnitude larger than any other public recommendation system datasets.
6 0.43248489 364 hunch net-2009-07-11-Interesting papers at KDD
7 0.42507496 94 hunch net-2005-07-13-Text Entailment at AAAI
8 0.41628519 129 hunch net-2005-11-07-Prediction Competitions
9 0.41556871 239 hunch net-2007-04-18-$50K Spock Challenge
10 0.41310352 275 hunch net-2007-11-29-The Netflix Crack
11 0.40879413 200 hunch net-2006-08-03-AOL’s data drop
12 0.3988111 399 hunch net-2010-05-20-Google Predict
13 0.398444 155 hunch net-2006-02-07-Pittsburgh Mind Reading Competition
14 0.38071641 349 hunch net-2009-04-21-Interesting Presentations at Snowbird
15 0.36577591 342 hunch net-2009-02-16-KDNuggets
16 0.35498941 291 hunch net-2008-03-07-Spock Challenge Winners
17 0.34051448 390 hunch net-2010-03-12-Netflix Challenge 2 Canceled
18 0.33957157 459 hunch net-2012-03-13-The Submodularity workshop and Lucca Professorship
19 0.33499727 47 hunch net-2005-03-28-Open Problems for Colt
20 0.32898033 92 hunch net-2005-07-11-AAAI blog
topicId topicWeight
[(27, 0.771)]
simIndex simValue blogId blogTitle
1 1.0 166 hunch net-2006-03-24-NLPers
Introduction: Hal Daume has started the NLPers blog to discuss learning for language problems.
2 1.0 246 hunch net-2007-06-13-Not Posting
Introduction: If you have been disappointed by the lack of a post for the last month, consider contributing your own (I’ve been busy+uninspired). Also, keep in mind that there is a community of machine learning blogs (see the sidebar).
same-blog 3 1.0 418 hunch net-2010-12-02-Traffic Prediction Problem
Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.
4 0.99896955 274 hunch net-2007-11-28-Computational Consequences of Classification
Introduction: In the regression vs classification debate , I’m adding a new “pro” to classification. It seems there are computational shortcuts available for classification which simply aren’t available for regression. This arises in several situations. In active learning it is sometimes possible to find an e error classifier with just log(e) labeled samples. Only much more modest improvements appear to be achievable for squared loss regression. The essential reason is that the loss function on many examples is flat with respect to large variations in the parameter spaces of a learned classifier, which implies that many of these classifiers do not need to be considered. In contrast, for squared loss regression, most substantial variations in the parameter space influence the loss at most points. In budgeted learning, where there is either a computational time constraint or a feature cost constraint, a classifier can sometimes be learned to very high accuracy under the constraints
5 0.99732149 247 hunch net-2007-06-14-Interesting Papers at COLT 2007
Introduction: Here are two papers that seem particularly interesting at this year’s COLT. Gilles Blanchard and François Fleuret , Occam’s Hammer . When we are interested in very tight bounds on the true error rate of a classifier, it is tempting to use a PAC-Bayes bound which can (empirically) be quite tight . A disadvantage of the PAC-Bayes bound is that it applies to a classifier which is randomized over a set of base classifiers rather than a single classifier. This paper shows that a similar bound can be proved which holds for a single classifier drawn from the set. The ability to safely use a single classifier is very nice. This technique applies generically to any base bound, so it has other applications covered in the paper. Adam Tauman Kalai . Learning Nested Halfspaces and Uphill Decision Trees . Classification PAC-learning, where you prove that any problem amongst some set is polytime learnable with respect to any distribution over the input X is extraordinarily ch
6 0.99507952 308 hunch net-2008-07-06-To Dual or Not
7 0.99310011 400 hunch net-2010-06-13-The Good News on Exploration and Learning
8 0.99267262 245 hunch net-2007-05-12-Loss Function Semantics
9 0.9924919 172 hunch net-2006-04-14-JMLR is a success
10 0.99159312 288 hunch net-2008-02-10-Complexity Illness
11 0.98636121 45 hunch net-2005-03-22-Active learning
12 0.97499895 9 hunch net-2005-02-01-Watchword: Loss
13 0.96963954 341 hunch net-2009-02-04-Optimal Proxy Loss for Classification
14 0.96788663 352 hunch net-2009-05-06-Machine Learning to AI
15 0.95890832 304 hunch net-2008-06-27-Reviewing Horror Stories
16 0.95277888 196 hunch net-2006-07-13-Regression vs. Classification as a Primitive
17 0.94472241 483 hunch net-2013-06-10-The Large Scale Learning class notes
18 0.94137728 244 hunch net-2007-05-09-The Missing Bound
19 0.93332106 293 hunch net-2008-03-23-Interactive Machine Learning
20 0.92975235 8 hunch net-2005-02-01-NIPS: Online Bayes