hunch_net hunch_net-2010 hunch_net-2010-418 knowledge-graph by maker-knowledge-mining

418 hunch net-2010-12-02-Traffic Prediction Problem


meta infos for this blog

Source: html

Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 Slashdot points out the Traffic Prediction Challenge which looks pretty fun. [sent-1, score-0.696]

2 The temporal aspect seems to be very common in many real-world problems and somewhat understudied. [sent-2, score-1.369]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('traffic', 0.476), ('temporal', 0.441), ('slashdot', 0.381), ('aspect', 0.296), ('looks', 0.286), ('challenge', 0.248), ('somewhat', 0.234), ('pretty', 0.217), ('points', 0.193), ('prediction', 0.147), ('common', 0.14), ('problems', 0.104), ('seems', 0.096), ('many', 0.058)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 1.0 418 hunch net-2010-12-02-Traffic Prediction Problem

Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.

2 0.10928958 92 hunch net-2005-07-11-AAAI blog

Introduction: The AAAI conference is running a student blog which looks like a fun experiment.

3 0.10436411 271 hunch net-2007-11-05-CMU wins DARPA Urban Challenge

Introduction: The results have been posted , with CMU first , Stanford second , and Virginia Tech Third . Considering that this was an open event (at least for people in the US), this was a very strong showing for research at universities (instead of defense contractors, for example). Some details should become public at the NIPS workshops . Slashdot has a post with many comments.

4 0.10083219 239 hunch net-2007-04-18-$50K Spock Challenge

Introduction: Apparently, the company Spock is setting up a $50k entity resolution challenge . $50k is much less than the Netflix challenge, but it’s effectively the same as Netflix until someone reaches 10% . It’s also nice that the Spock challenge has a short duration. The (visible) test set is of size 25k and the training set has size 75k.

5 0.096672006 200 hunch net-2006-08-03-AOL’s data drop

Introduction: AOL has released several large search engine related datasets. This looks like a pretty impressive data release, and it is a big opportunity for people everywhere to worry about search engine related learning problems, if they want.

6 0.094991684 364 hunch net-2009-07-11-Interesting papers at KDD

7 0.091941655 399 hunch net-2010-05-20-Google Predict

8 0.082059875 446 hunch net-2011-10-03-Monday announcements

9 0.080537446 190 hunch net-2006-07-06-Branch Prediction Competition

10 0.06701488 129 hunch net-2005-11-07-Prediction Competitions

11 0.065326117 326 hunch net-2008-11-11-COLT CFP

12 0.062750414 275 hunch net-2007-11-29-The Netflix Crack

13 0.06119936 291 hunch net-2008-03-07-Spock Challenge Winners

14 0.060982488 384 hunch net-2009-12-24-Top graduates this season

15 0.05958163 151 hunch net-2006-01-25-1 year

16 0.053940065 433 hunch net-2011-04-23-ICML workshops due

17 0.052791297 332 hunch net-2008-12-23-Use of Learning Theory

18 0.052005105 211 hunch net-2006-10-02-$1M Netflix prediction contest

19 0.050370209 47 hunch net-2005-03-28-Open Problems for Colt

20 0.049578734 406 hunch net-2010-08-22-KDD 2010


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.071), (1, 0.002), (2, -0.023), (3, -0.014), (4, -0.009), (5, -0.003), (6, -0.036), (7, -0.023), (8, 0.024), (9, -0.045), (10, -0.034), (11, 0.127), (12, -0.074), (13, 0.015), (14, -0.037), (15, 0.015), (16, 0.004), (17, -0.02), (18, -0.002), (19, 0.014), (20, -0.041), (21, -0.041), (22, -0.037), (23, 0.004), (24, 0.011), (25, -0.025), (26, 0.061), (27, 0.023), (28, -0.113), (29, 0.037), (30, 0.099), (31, 0.066), (32, -0.05), (33, 0.059), (34, 0.018), (35, -0.129), (36, -0.023), (37, 0.07), (38, -0.008), (39, 0.036), (40, 0.016), (41, -0.014), (42, 0.026), (43, -0.023), (44, 0.092), (45, -0.021), (46, 0.025), (47, 0.037), (48, -0.072), (49, -0.041)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.9712038 418 hunch net-2010-12-02-Traffic Prediction Problem

Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.

2 0.53369242 190 hunch net-2006-07-06-Branch Prediction Competition

Introduction: Alan Fern points out the second branch prediction challenge (due September 29) which is a follow up to the first branch prediction competition . Branch prediction is one of the fundamental learning problems of the computer age: without it our computers might run an order of magnitude slower. This is a tough problem since there are sharp constraints on time and space complexity in an online environment. For machine learning, the “idealistic track” may fit well. Essentially, they remove these constraints to gain a weak upper bound on what might be done.

3 0.52884352 427 hunch net-2011-03-20-KDD Cup 2011

Introduction: Yehuda points out KDD-Cup 2011 which Markus and Gideon helped setup. This is a prediction and recommendation contest for music. In addition to being a fun chance to show your expertise, there are cash prizes of $5K/$2K/$1K.

4 0.49368921 446 hunch net-2011-10-03-Monday announcements

Introduction: Various people want to use hunch.net to announce things. I’ve generally resisted this because I feared hunch becoming a pure announcement zone while I am much more interested contentful posts and discussion personally. Nevertheless there is clearly some value and announcements are easy, so I’m planning to summarize announcements on Mondays. D. Sculley points out an interesting Semisupervised feature learning competition, with a deadline of October 17. Lihong Li points out the webscope user interaction dataset which is the first high quality exploration dataset I’m aware of that is publicly available. Seth Rogers points out CrossValidated which looks similar in conception to metaoptimize , but directly using the stackoverflow interface and with a bit more of a statistics twist.

5 0.43867168 211 hunch net-2006-10-02-$1M Netflix prediction contest

Introduction: Netflix is running a contest to improve recommender prediction systems. A 10% improvement over their current system yields a $1M prize. Failing that, the best smaller improvement yields a smaller $50K prize. This contest looks quite real, and the $50K prize money is almost certainly achievable with a bit of thought. The contest also comes with a dataset which is apparently 2 orders of magnitude larger than any other public recommendation system datasets.

6 0.43248489 364 hunch net-2009-07-11-Interesting papers at KDD

7 0.42507496 94 hunch net-2005-07-13-Text Entailment at AAAI

8 0.41628519 129 hunch net-2005-11-07-Prediction Competitions

9 0.41556871 239 hunch net-2007-04-18-$50K Spock Challenge

10 0.41310352 275 hunch net-2007-11-29-The Netflix Crack

11 0.40879413 200 hunch net-2006-08-03-AOL’s data drop

12 0.3988111 399 hunch net-2010-05-20-Google Predict

13 0.398444 155 hunch net-2006-02-07-Pittsburgh Mind Reading Competition

14 0.38071641 349 hunch net-2009-04-21-Interesting Presentations at Snowbird

15 0.36577591 342 hunch net-2009-02-16-KDNuggets

16 0.35498941 291 hunch net-2008-03-07-Spock Challenge Winners

17 0.34051448 390 hunch net-2010-03-12-Netflix Challenge 2 Canceled

18 0.33957157 459 hunch net-2012-03-13-The Submodularity workshop and Lucca Professorship

19 0.33499727 47 hunch net-2005-03-28-Open Problems for Colt

20 0.32898033 92 hunch net-2005-07-11-AAAI blog


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(27, 0.771)]

similar blogs list:

simIndex simValue blogId blogTitle

1 1.0 166 hunch net-2006-03-24-NLPers

Introduction: Hal Daume has started the NLPers blog to discuss learning for language problems.

2 1.0 246 hunch net-2007-06-13-Not Posting

Introduction: If you have been disappointed by the lack of a post for the last month, consider contributing your own (I’ve been busy+uninspired). Also, keep in mind that there is a community of machine learning blogs (see the sidebar).

same-blog 3 1.0 418 hunch net-2010-12-02-Traffic Prediction Problem

Introduction: Slashdot points out the Traffic Prediction Challenge which looks pretty fun. The temporal aspect seems to be very common in many real-world problems and somewhat understudied.

4 0.99896955 274 hunch net-2007-11-28-Computational Consequences of Classification

Introduction: In the regression vs classification debate , I’m adding a new “pro” to classification. It seems there are computational shortcuts available for classification which simply aren’t available for regression. This arises in several situations. In active learning it is sometimes possible to find an e error classifier with just log(e) labeled samples. Only much more modest improvements appear to be achievable for squared loss regression. The essential reason is that the loss function on many examples is flat with respect to large variations in the parameter spaces of a learned classifier, which implies that many of these classifiers do not need to be considered. In contrast, for squared loss regression, most substantial variations in the parameter space influence the loss at most points. In budgeted learning, where there is either a computational time constraint or a feature cost constraint, a classifier can sometimes be learned to very high accuracy under the constraints

5 0.99732149 247 hunch net-2007-06-14-Interesting Papers at COLT 2007

Introduction: Here are two papers that seem particularly interesting at this year’s COLT. Gilles Blanchard and François Fleuret , Occam’s Hammer . When we are interested in very tight bounds on the true error rate of a classifier, it is tempting to use a PAC-Bayes bound which can (empirically) be quite tight . A disadvantage of the PAC-Bayes bound is that it applies to a classifier which is randomized over a set of base classifiers rather than a single classifier. This paper shows that a similar bound can be proved which holds for a single classifier drawn from the set. The ability to safely use a single classifier is very nice. This technique applies generically to any base bound, so it has other applications covered in the paper. Adam Tauman Kalai . Learning Nested Halfspaces and Uphill Decision Trees . Classification PAC-learning, where you prove that any problem amongst some set is polytime learnable with respect to any distribution over the input X is extraordinarily ch

6 0.99507952 308 hunch net-2008-07-06-To Dual or Not

7 0.99310011 400 hunch net-2010-06-13-The Good News on Exploration and Learning

8 0.99267262 245 hunch net-2007-05-12-Loss Function Semantics

9 0.9924919 172 hunch net-2006-04-14-JMLR is a success

10 0.99159312 288 hunch net-2008-02-10-Complexity Illness

11 0.98636121 45 hunch net-2005-03-22-Active learning

12 0.97499895 9 hunch net-2005-02-01-Watchword: Loss

13 0.96963954 341 hunch net-2009-02-04-Optimal Proxy Loss for Classification

14 0.96788663 352 hunch net-2009-05-06-Machine Learning to AI

15 0.95890832 304 hunch net-2008-06-27-Reviewing Horror Stories

16 0.95277888 196 hunch net-2006-07-13-Regression vs. Classification as a Primitive

17 0.94472241 483 hunch net-2013-06-10-The Large Scale Learning class notes

18 0.94137728 244 hunch net-2007-05-09-The Missing Bound

19 0.93332106 293 hunch net-2008-03-23-Interactive Machine Learning

20 0.92975235 8 hunch net-2005-02-01-NIPS: Online Bayes