hunch_net hunch_net-2006 hunch_net-2006-175 knowledge-graph by maker-knowledge-mining

175 hunch net-2006-04-30-John Langford –> Yahoo Research, NY


meta infos for this blog

Source: html

Introduction: I will join Yahoo Research (in New York) after my contract ends at TTI-Chicago . The deciding reasons are: Yahoo is running into many hard learning problems. This is precisely the situation where basic research might hope to have the greatest impact. Yahoo Research understands research including publishing, conferences, etc… Yahoo Research is growing, so there is a chance I can help it grow well. Yahoo understands the internet, including (but not at all limited to) experimenting with research blogs. In the end, Yahoo Research seems like the place where I might have a chance to make the greatest difference. Yahoo (as a company) has made a strong bet on Yahoo Research. We-the-researchers all hope that bet will pay off, and this seems plausible. I’ll certainly have fun trying.


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 I will join Yahoo Research (in New York) after my contract ends at TTI-Chicago . [sent-1, score-0.369]

2 The deciding reasons are: Yahoo is running into many hard learning problems. [sent-2, score-0.313]

3 This is precisely the situation where basic research might hope to have the greatest impact. [sent-3, score-0.873]

4 Yahoo Research understands research including publishing, conferences, etc… Yahoo Research is growing, so there is a chance I can help it grow well. [sent-4, score-0.931]

5 Yahoo understands the internet, including (but not at all limited to) experimenting with research blogs. [sent-5, score-0.801]

6 In the end, Yahoo Research seems like the place where I might have a chance to make the greatest difference. [sent-6, score-0.656]

7 Yahoo (as a company) has made a strong bet on Yahoo Research. [sent-7, score-0.369]

8 We-the-researchers all hope that bet will pay off, and this seems plausible. [sent-8, score-0.521]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('yahoo', 0.689), ('understands', 0.274), ('bet', 0.259), ('greatest', 0.259), ('research', 0.228), ('chance', 0.138), ('contract', 0.137), ('including', 0.12), ('ends', 0.118), ('join', 0.114), ('grow', 0.111), ('hope', 0.106), ('deciding', 0.105), ('experimenting', 0.102), ('company', 0.1), ('pay', 0.096), ('precisely', 0.092), ('fun', 0.092), ('publishing', 0.09), ('growing', 0.087), ('situation', 0.085), ('york', 0.081), ('internet', 0.078), ('limited', 0.077), ('certainly', 0.074), ('running', 0.073), ('place', 0.073), ('end', 0.065), ('might', 0.064), ('trying', 0.063), ('etc', 0.062), ('help', 0.06), ('seems', 0.06), ('reasons', 0.059), ('strong', 0.058), ('ll', 0.058), ('conferences', 0.057), ('made', 0.052), ('hard', 0.049), ('basic', 0.039), ('make', 0.033), ('like', 0.029), ('new', 0.026), ('many', 0.018), ('learning', 0.009)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 1.0 175 hunch net-2006-04-30-John Langford –> Yahoo Research, NY

Introduction: I will join Yahoo Research (in New York) after my contract ends at TTI-Chicago . The deciding reasons are: Yahoo is running into many hard learning problems. This is precisely the situation where basic research might hope to have the greatest impact. Yahoo Research understands research including publishing, conferences, etc… Yahoo Research is growing, so there is a chance I can help it grow well. Yahoo understands the internet, including (but not at all limited to) experimenting with research blogs. In the end, Yahoo Research seems like the place where I might have a chance to make the greatest difference. Yahoo (as a company) has made a strong bet on Yahoo Research. We-the-researchers all hope that bet will pay off, and this seems plausible. I’ll certainly have fun trying.

2 0.30142906 156 hunch net-2006-02-11-Yahoo’s Learning Problems.

Introduction: I just visited Yahoo Research which has several fundamental learning problems near to (or beyond) the set of problems we know how to solve well. Here are 3 of them. Ranking This is the canonical problem of all search engines. It is made extra difficult for several reasons. There is relatively little “good” supervised learning data and a great deal of data with some signal (such as click through rates). The learning must occur in a partially adversarial environment. Many people very actively attempt to place themselves at the top of rankings. It is not even quite clear whether the problem should be posed as ‘ranking’ or as ‘regression’ which is then used to produce a ranking. Collaborative filtering Yahoo has a large number of recommendation systems for music, movies, etc… In these sorts of systems, users specify how they liked a set of things, and then the system can (hopefully) find some more examples of things they might like by reasoning across multiple

3 0.26928869 464 hunch net-2012-05-03-Microsoft Research, New York City

Introduction: Yahoo! laid off people . Unlike every previous time there have been layoffs, this is serious for Yahoo! Research . We had advanced warning from Prabhakar through the simple act of leaving . Yahoo! Research was a world class organization that Prabhakar recruited much of personally, so it is deeply implausible that he would spontaneously decide to leave. My first thought when I saw the news was “Uhoh, Rob said that he knew it was serious when the head of ATnT Research left.” In this case it was even more significant, because Prabhakar recruited me on the premise that Y!R was an experiment in how research should be done: via a combination of high quality people and high engagement with the company. Prabhakar’s departure is a clear end to that experiment. The result is ambiguous from a business perspective. Y!R clearly was not capable of saving the company from its illnesses. I’m not privy to the internal accounting of impact and this is the kind of subject where there c

4 0.19744368 425 hunch net-2011-02-25-Yahoo! Machine Learning grant due March 11

Introduction: Yahoo!’s Key Scientific Challenges for Machine Learning grant applications are due March 11. If you are a student working on relevant research, please consider applying. It’s for $5K of unrestricted funding.

5 0.16417003 178 hunch net-2006-05-08-Big machine learning

Introduction: According to the New York Times , Yahoo is releasing Project Panama shortly . Project Panama is about better predicting which advertisements are relevant to a search, implying a higher click through rate, implying larger income for Yahoo . There are two things that seem interesting here: A significant portion of that improved accuracy is almost certainly machine learning at work. The quantitative effect is huge—the estimate in the article is $600*10 6 . Google already has such improvements and Microsoft Search is surely working on them, which suggest this is (perhaps) a $10 9 per year machine learning problem. The exact methodology under use is unlikely to be publicly discussed in the near future because of the competitive enivironment. Hopefully we’ll have some public “war stories” at some point in the future when this information becomes less sensitive. For now, it’s reassuring to simply note that machine learning is having a big impact.

6 0.13365734 457 hunch net-2012-02-29-Key Scientific Challenges and the Franklin Symposium

7 0.12766729 117 hunch net-2005-10-03-Not ICML

8 0.10688657 339 hunch net-2009-01-27-Key Scientific Challenges

9 0.10421078 132 hunch net-2005-11-26-The Design of an Optimal Research Environment

10 0.10342235 121 hunch net-2005-10-12-The unrealized potential of the research lab

11 0.10163535 36 hunch net-2005-03-05-Funding Research

12 0.095266514 344 hunch net-2009-02-22-Effective Research Funding

13 0.092388511 389 hunch net-2010-02-26-Yahoo! ML events

14 0.087890036 51 hunch net-2005-04-01-The Producer-Consumer Model of Research

15 0.08476571 423 hunch net-2011-02-02-User preferences for search engines

16 0.083051659 478 hunch net-2013-01-07-NYU Large Scale Machine Learning Class

17 0.07367368 400 hunch net-2010-06-13-The Good News on Exploration and Learning

18 0.073558405 449 hunch net-2011-11-26-Giving Thanks

19 0.070456088 142 hunch net-2005-12-22-Yes , I am applying

20 0.070389099 324 hunch net-2008-11-09-A Healthy COLT


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.114), (1, -0.063), (2, -0.131), (3, 0.106), (4, -0.134), (5, -0.032), (6, -0.002), (7, 0.06), (8, -0.127), (9, -0.023), (10, 0.075), (11, 0.055), (12, -0.028), (13, 0.01), (14, -0.029), (15, 0.167), (16, -0.126), (17, -0.031), (18, -0.068), (19, -0.194), (20, 0.121), (21, 0.043), (22, 0.111), (23, -0.023), (24, 0.087), (25, 0.144), (26, 0.027), (27, -0.079), (28, -0.072), (29, -0.008), (30, 0.015), (31, -0.017), (32, -0.082), (33, 0.045), (34, 0.048), (35, 0.104), (36, -0.001), (37, 0.031), (38, -0.089), (39, 0.02), (40, -0.0), (41, -0.058), (42, -0.024), (43, 0.031), (44, 0.009), (45, 0.007), (46, 0.103), (47, 0.124), (48, 0.014), (49, 0.054)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.96654671 175 hunch net-2006-04-30-John Langford –> Yahoo Research, NY

Introduction: I will join Yahoo Research (in New York) after my contract ends at TTI-Chicago . The deciding reasons are: Yahoo is running into many hard learning problems. This is precisely the situation where basic research might hope to have the greatest impact. Yahoo Research understands research including publishing, conferences, etc… Yahoo Research is growing, so there is a chance I can help it grow well. Yahoo understands the internet, including (but not at all limited to) experimenting with research blogs. In the end, Yahoo Research seems like the place where I might have a chance to make the greatest difference. Yahoo (as a company) has made a strong bet on Yahoo Research. We-the-researchers all hope that bet will pay off, and this seems plausible. I’ll certainly have fun trying.

2 0.7606135 464 hunch net-2012-05-03-Microsoft Research, New York City

Introduction: Yahoo! laid off people . Unlike every previous time there have been layoffs, this is serious for Yahoo! Research . We had advanced warning from Prabhakar through the simple act of leaving . Yahoo! Research was a world class organization that Prabhakar recruited much of personally, so it is deeply implausible that he would spontaneously decide to leave. My first thought when I saw the news was “Uhoh, Rob said that he knew it was serious when the head of ATnT Research left.” In this case it was even more significant, because Prabhakar recruited me on the premise that Y!R was an experiment in how research should be done: via a combination of high quality people and high engagement with the company. Prabhakar’s departure is a clear end to that experiment. The result is ambiguous from a business perspective. Y!R clearly was not capable of saving the company from its illnesses. I’m not privy to the internal accounting of impact and this is the kind of subject where there c

3 0.7382912 156 hunch net-2006-02-11-Yahoo’s Learning Problems.

Introduction: I just visited Yahoo Research which has several fundamental learning problems near to (or beyond) the set of problems we know how to solve well. Here are 3 of them. Ranking This is the canonical problem of all search engines. It is made extra difficult for several reasons. There is relatively little “good” supervised learning data and a great deal of data with some signal (such as click through rates). The learning must occur in a partially adversarial environment. Many people very actively attempt to place themselves at the top of rankings. It is not even quite clear whether the problem should be posed as ‘ranking’ or as ‘regression’ which is then used to produce a ranking. Collaborative filtering Yahoo has a large number of recommendation systems for music, movies, etc… In these sorts of systems, users specify how they liked a set of things, and then the system can (hopefully) find some more examples of things they might like by reasoning across multiple

4 0.69051582 178 hunch net-2006-05-08-Big machine learning

Introduction: According to the New York Times , Yahoo is releasing Project Panama shortly . Project Panama is about better predicting which advertisements are relevant to a search, implying a higher click through rate, implying larger income for Yahoo . There are two things that seem interesting here: A significant portion of that improved accuracy is almost certainly machine learning at work. The quantitative effect is huge—the estimate in the article is $600*10 6 . Google already has such improvements and Microsoft Search is surely working on them, which suggest this is (perhaps) a $10 9 per year machine learning problem. The exact methodology under use is unlikely to be publicly discussed in the near future because of the competitive enivironment. Hopefully we’ll have some public “war stories” at some point in the future when this information becomes less sensitive. For now, it’s reassuring to simply note that machine learning is having a big impact.

5 0.65273643 425 hunch net-2011-02-25-Yahoo! Machine Learning grant due March 11

Introduction: Yahoo!’s Key Scientific Challenges for Machine Learning grant applications are due March 11. If you are a student working on relevant research, please consider applying. It’s for $5K of unrestricted funding.

6 0.64163709 339 hunch net-2009-01-27-Key Scientific Challenges

7 0.59899819 121 hunch net-2005-10-12-The unrealized potential of the research lab

8 0.57493365 457 hunch net-2012-02-29-Key Scientific Challenges and the Franklin Symposium

9 0.47752479 132 hunch net-2005-11-26-The Design of an Optimal Research Environment

10 0.47312996 110 hunch net-2005-09-10-“Failure” is an option

11 0.47039357 142 hunch net-2005-12-22-Yes , I am applying

12 0.46844351 51 hunch net-2005-04-01-The Producer-Consumer Model of Research

13 0.45427507 36 hunch net-2005-03-05-Funding Research

14 0.4413 344 hunch net-2009-02-22-Effective Research Funding

15 0.44115242 255 hunch net-2007-07-13-The View From China

16 0.4202106 389 hunch net-2010-02-26-Yahoo! ML events

17 0.41834641 117 hunch net-2005-10-03-Not ICML

18 0.41220212 449 hunch net-2011-11-26-Giving Thanks

19 0.39497104 73 hunch net-2005-05-17-A Short Guide to PhD Graduate Study

20 0.38969931 423 hunch net-2011-02-02-User preferences for search engines


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(27, 0.174), (38, 0.051), (53, 0.055), (55, 0.122), (94, 0.021), (95, 0.063), (96, 0.351)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.91089463 175 hunch net-2006-04-30-John Langford –> Yahoo Research, NY

Introduction: I will join Yahoo Research (in New York) after my contract ends at TTI-Chicago . The deciding reasons are: Yahoo is running into many hard learning problems. This is precisely the situation where basic research might hope to have the greatest impact. Yahoo Research understands research including publishing, conferences, etc… Yahoo Research is growing, so there is a chance I can help it grow well. Yahoo understands the internet, including (but not at all limited to) experimenting with research blogs. In the end, Yahoo Research seems like the place where I might have a chance to make the greatest difference. Yahoo (as a company) has made a strong bet on Yahoo Research. We-the-researchers all hope that bet will pay off, and this seems plausible. I’ll certainly have fun trying.

2 0.81997424 53 hunch net-2005-04-06-Structured Regret Minimization

Introduction: Geoff Gordon made an interesting presentation at the snowbird learning workshop discussing the use of no-regret algorithms for the use of several robot-related learning problems. There seems to be a draft here . This seems interesting in two ways: Drawback Removal One of the significant problems with these online algorithms is that they can’t cope with structure very easily. This drawback is addressed for certain structures. Experiments One criticism of such algorithms is that they are too “worst case”. Several experiments suggest that protecting yourself against this worst case does not necessarily incur a great loss.

3 0.7971096 443 hunch net-2011-09-03-Fall Machine Learning Events

Introduction: Many Machine Learning related events are coming up this fall. September 9 , abstracts for the New York Machine Learning Symposium are due. Send a 2 page pdf, if interested, and note that we: widened submissions to be from anybody rather than students. set aside a larger fraction of time for contributed submissions. September 15 , there is a machine learning meetup , where I’ll be discussing terascale learning at AOL. September 16 , there is a CS&Econ; day at New York Academy of Sciences. This is not ML focused, but it’s easy to imagine interest. September 23 and later NIPS workshop submissions start coming due. As usual, there are too many good ones, so I won’t be able to attend all those that interest me. I do hope some workshop makers consider ICML this coming summer, as we are increasing to a 2 day format for you. Here are a few that interest me: Big Learning is about dealing with lots of data. Abstracts are due September 30 . The Bayes

4 0.76844376 104 hunch net-2005-08-22-Do you believe in induction?

Introduction: Foster Provost gave a talk at the ICML metalearning workshop on “metalearning” and the “no free lunch theorem” which seems worth summarizing. As a review: the no free lunch theorem is the most complicated way we know of to say that a bias is required in order to learn. The simplest way to see this is in a nonprobabilistic setting. If you are given examples of the form (x,y) and you wish to predict y from x then any prediction mechanism errs half the time in expectation over all sequences of examples. The proof of this is very simple: on every example a predictor must make some prediction and by symmetry over the set of sequences it will be wrong half the time and right half the time. The basic idea of this proof has been applied to many other settings. The simplistic interpretation of this theorem which many people jump to is “machine learning is dead” since there can be no single learning algorithm which can solve all learning problems. This is the wrong way to thi

5 0.65599829 105 hunch net-2005-08-23-(Dis)similarities between academia and open source programmers

Introduction: Martin Pool and I recently discussed the similarities and differences between academia and open source programming. Similarities: Cost profile Research and programming share approximately the same cost profile: A large upfront effort is required to produce something useful, and then “anyone” can use it. (The “anyone” is not quite right for either group because only sufficiently technical people could use it.) Wealth profile A “wealthy” academic or open source programmer is someone who has contributed a lot to other people in research or programs. Much of academia is a “gift culture”: whoever gives the most is most respected. Problems Both academia and open source programming suffer from similar problems. Whether or not (and which) open source program is used are perhaps too-often personality driven rather than driven by capability or usefulness. Similar phenomena can happen in academia with respect to directions of research. Funding is often a problem for

6 0.56518865 478 hunch net-2013-01-07-NYU Large Scale Machine Learning Class

7 0.53952265 466 hunch net-2012-06-05-ICML acceptance statistics

8 0.53460556 225 hunch net-2007-01-02-Retrospective

9 0.53163165 464 hunch net-2012-05-03-Microsoft Research, New York City

10 0.53101343 437 hunch net-2011-07-10-ICML 2011 and the future

11 0.52857721 89 hunch net-2005-07-04-The Health of COLT

12 0.52853811 452 hunch net-2012-01-04-Why ICML? and the summer conferences

13 0.52682167 343 hunch net-2009-02-18-Decision by Vetocracy

14 0.52539784 403 hunch net-2010-07-18-ICML & COLT 2010

15 0.5247491 194 hunch net-2006-07-11-New Models

16 0.52139848 406 hunch net-2010-08-22-KDD 2010

17 0.51992494 204 hunch net-2006-08-28-Learning Theory standards for NIPS 2006

18 0.51940644 454 hunch net-2012-01-30-ICML Posters and Scope

19 0.51834112 51 hunch net-2005-04-01-The Producer-Consumer Model of Research

20 0.51695681 484 hunch net-2013-06-16-Representative Reviewing