hunch_net hunch_net-2009 hunch_net-2009-376 knowledge-graph by maker-knowledge-mining

376 hunch net-2009-11-06-Yisong Yue on Self-improving Systems


meta infos for this blog

Source: html

Introduction: I’d like to point out Yisong Yue ‘s post on Self-improving systems , which is a nicely readable description of the necessity and potential of interactive learning to deal with the information overload problem that is endemic to the modern internet.


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 I’d like to point out Yisong Yue ‘s post on Self-improving systems , which is a nicely readable description of the necessity and potential of interactive learning to deal with the information overload problem that is endemic to the modern internet. [sent-1, score-3.204]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('readable', 0.338), ('nicely', 0.338), ('yisong', 0.319), ('yue', 0.319), ('overload', 0.292), ('endemic', 0.292), ('necessity', 0.273), ('interactive', 0.236), ('modern', 0.227), ('description', 0.223), ('internet', 0.193), ('potential', 0.179), ('deal', 0.148), ('systems', 0.147), ('post', 0.135), ('point', 0.109), ('information', 0.107), ('like', 0.072), ('problem', 0.066), ('learning', 0.022)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.99999994 376 hunch net-2009-11-06-Yisong Yue on Self-improving Systems

Introduction: I’d like to point out Yisong Yue ‘s post on Self-improving systems , which is a nicely readable description of the necessity and potential of interactive learning to deal with the information overload problem that is endemic to the modern internet.

2 0.15216003 384 hunch net-2009-12-24-Top graduates this season

Introduction: I would like to point out 3 graduates this season as having my confidence they are capable of doing great things. Daniel Hsu has diverse papers with diverse coauthors on {active learning, mulitlabeling, temporal learning, …} each covering new algorithms and methods of analysis. He is also a capable programmer, having helped me with some nitty-gritty details of cluster parallel Vowpal Wabbit this summer. He has an excellent tendency to just get things done. Nicolas Lambert doesn’t nominally work in machine learning, but I’ve found his work in elicitation relevant nevertheless. In essence, elicitable properties are closely related to learnable properties, and the elicitation complexity is related to a notion of learning complexity. See the Surrogate regret bounds paper for some related discussion. Few people successfully work at such a general level that it crosses fields, but he’s one of them. Yisong Yue is deeply focused on interactive learning, which he has a

3 0.1317035 293 hunch net-2008-03-23-Interactive Machine Learning

Introduction: A new direction of research seems to be arising in machine learning: Interactive Machine Learning. This isn’t a familiar term, although it does include some familiar subjects. What is Interactive Machine Learning? The fundamental requirement is (a) learning algorithms which interact with the world and (b) learn. For our purposes, let’s define learning as efficiently competing with a large set of possible predictors. Examples include: Online learning against an adversary ( Avrim’s Notes ). The interaction is almost trivial: the learning algorithm makes a prediction and then receives feedback. The learning is choosing based upon the advice of many experts. Active Learning . In active learning, the interaction is choosing which examples to label, and the learning is choosing from amongst a large set of hypotheses. Contextual Bandits . The interaction is choosing one of several actions and learning only the value of the chosen action (weaker than active learning

4 0.10516048 370 hunch net-2009-09-18-Necessary and Sufficient Research

Introduction: Researchers are typically confronted with big problems that they have no idea how to solve. In trying to come up with a solution, a natural approach is to decompose the big problem into a set of subproblems whose solution yields a solution to the larger problem. This approach can go wrong in several ways. Decomposition failure . The solution to the decomposition does not in fact yield a solution to the overall problem. Artificial hardness . The subproblems created are sufficient if solved to solve the overall problem, but they are harder than necessary. As you can see, computational complexity forms a relatively new (in research-history) razor by which to judge an approach sufficient but not necessary. In my experience, the artificial hardness problem is very common. Many researchers abdicate the responsibility of choosing a problem to work on to other people. This process starts very naturally as a graduate student, when an incoming student might have relatively l

5 0.09511286 386 hunch net-2010-01-13-Sam Roweis died

Introduction: and I can’t help but remember him. I first met Sam as an undergraduate at Caltech where he was TA for Hopfield ‘s class, and again when I visited Gatsby , when he invited me to visit Toronto , and at too many conferences to recount. His personality was a combination of enthusiastic and thoughtful, with a great ability to phrase a problem so it’s solution must be understood. With respect to my own work, Sam was the one who advised me to make my first tutorial , leading to others, and to other things, all of which I’m grateful to him for. In fact, my every interaction with Sam was positive, and that was his way. His death is being called a suicide which is so incompatible with my understanding of Sam that it strains my credibility. But we know that his many responsibilities were great, and it is well understood that basically all sane researchers have legions of inner doubts. Having been depressed now and then myself, it’s helpful to understand at least intellectually

6 0.077172957 69 hunch net-2005-05-11-Visa Casualties

7 0.076802664 469 hunch net-2012-07-09-Videolectures

8 0.066782936 265 hunch net-2007-10-14-NIPS workshp: Learning Problem Design

9 0.062435314 358 hunch net-2009-06-01-Multitask Poisoning

10 0.054660741 225 hunch net-2007-01-02-Retrospective

11 0.049422957 285 hunch net-2008-01-23-Why Workshop?

12 0.047268037 318 hunch net-2008-09-26-The SODA Program Committee

13 0.047067471 317 hunch net-2008-09-12-How do we get weak action dependence for learning with partial observations?

14 0.044469766 212 hunch net-2006-10-04-Health of Conferences Wiki

15 0.041286469 237 hunch net-2007-04-02-Contextual Scaling

16 0.041181386 375 hunch net-2009-10-26-NIPS workshops

17 0.040838104 385 hunch net-2009-12-27-Interesting things at NIPS 2009

18 0.040498994 334 hunch net-2009-01-07-Interesting Papers at SODA 2009

19 0.039498135 296 hunch net-2008-04-21-The Science 2.0 article

20 0.037270434 287 hunch net-2008-01-28-Sufficient Computation


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.064), (1, 0.007), (2, -0.04), (3, 0.031), (4, -0.002), (5, 0.002), (6, 0.016), (7, -0.029), (8, 0.03), (9, -0.001), (10, 0.01), (11, -0.001), (12, -0.014), (13, 0.073), (14, -0.05), (15, 0.011), (16, -0.03), (17, -0.022), (18, 0.018), (19, 0.072), (20, -0.013), (21, 0.023), (22, -0.003), (23, 0.036), (24, 0.01), (25, 0.055), (26, 0.008), (27, 0.055), (28, -0.006), (29, 0.046), (30, -0.031), (31, -0.029), (32, 0.025), (33, -0.021), (34, 0.013), (35, -0.035), (36, -0.023), (37, -0.115), (38, -0.064), (39, 0.088), (40, 0.064), (41, 0.028), (42, 0.012), (43, 0.048), (44, -0.1), (45, -0.019), (46, -0.15), (47, 0.074), (48, 0.094), (49, 0.049)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.9709965 376 hunch net-2009-11-06-Yisong Yue on Self-improving Systems

Introduction: I’d like to point out Yisong Yue ‘s post on Self-improving systems , which is a nicely readable description of the necessity and potential of interactive learning to deal with the information overload problem that is endemic to the modern internet.

2 0.64946914 386 hunch net-2010-01-13-Sam Roweis died

Introduction: and I can’t help but remember him. I first met Sam as an undergraduate at Caltech where he was TA for Hopfield ‘s class, and again when I visited Gatsby , when he invited me to visit Toronto , and at too many conferences to recount. His personality was a combination of enthusiastic and thoughtful, with a great ability to phrase a problem so it’s solution must be understood. With respect to my own work, Sam was the one who advised me to make my first tutorial , leading to others, and to other things, all of which I’m grateful to him for. In fact, my every interaction with Sam was positive, and that was his way. His death is being called a suicide which is so incompatible with my understanding of Sam that it strains my credibility. But we know that his many responsibilities were great, and it is well understood that basically all sane researchers have legions of inner doubts. Having been depressed now and then myself, it’s helpful to understand at least intellectually

3 0.53832418 384 hunch net-2009-12-24-Top graduates this season

Introduction: I would like to point out 3 graduates this season as having my confidence they are capable of doing great things. Daniel Hsu has diverse papers with diverse coauthors on {active learning, mulitlabeling, temporal learning, …} each covering new algorithms and methods of analysis. He is also a capable programmer, having helped me with some nitty-gritty details of cluster parallel Vowpal Wabbit this summer. He has an excellent tendency to just get things done. Nicolas Lambert doesn’t nominally work in machine learning, but I’ve found his work in elicitation relevant nevertheless. In essence, elicitable properties are closely related to learnable properties, and the elicitation complexity is related to a notion of learning complexity. See the Surrogate regret bounds paper for some related discussion. Few people successfully work at such a general level that it crosses fields, but he’s one of them. Yisong Yue is deeply focused on interactive learning, which he has a

4 0.53400218 212 hunch net-2006-10-04-Health of Conferences Wiki

Introduction: Aaron Hertzmann points out the health of conferences wiki , which has a great deal of information about how many different conferences function.

5 0.4724699 491 hunch net-2013-11-21-Ben Taskar is gone

Introduction: I was not as personally close to Ben as Sam , but the level of tragedy is similar and I can’t help but be greatly saddened by the loss. Various news stories have coverage, but the synopsis is that he had a heart attack on Sunday and is survived by his wife Anat and daughter Aviv. There is discussion of creating a memorial fund for them, which I hope comes to fruition, and plan to contribute to. I will remember Ben as someone who thought carefully and comprehensively about new ways to do things, then fought hard and successfully for what he believed in. It is an ideal we strive for, that Ben accomplished. Edit: donations go here , and more information is here .

6 0.4672412 293 hunch net-2008-03-23-Interactive Machine Learning

7 0.4293412 370 hunch net-2009-09-18-Necessary and Sufficient Research

8 0.38260186 269 hunch net-2007-10-24-Contextual Bandits

9 0.37911096 215 hunch net-2006-10-22-Exemplar programming

10 0.377455 307 hunch net-2008-07-04-More Presentation Preparation

11 0.37566274 237 hunch net-2007-04-02-Contextual Scaling

12 0.37255856 323 hunch net-2008-11-04-Rise of the Machines

13 0.35603747 264 hunch net-2007-09-30-NIPS workshops are out.

14 0.35234305 400 hunch net-2010-06-13-The Good News on Exploration and Learning

15 0.35186011 249 hunch net-2007-06-21-Presentation Preparation

16 0.33989295 312 hunch net-2008-08-04-Electoralmarkets.com

17 0.33803141 265 hunch net-2007-10-14-NIPS workshp: Learning Problem Design

18 0.337248 367 hunch net-2009-08-16-Centmail comments

19 0.33373007 217 hunch net-2006-11-06-Data Linkage Problems

20 0.33298761 412 hunch net-2010-09-28-Machined Learnings


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(27, 0.122), (54, 0.605), (94, 0.08)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.96504694 376 hunch net-2009-11-06-Yisong Yue on Self-improving Systems

Introduction: I’d like to point out Yisong Yue ‘s post on Self-improving systems , which is a nicely readable description of the necessity and potential of interactive learning to deal with the information overload problem that is endemic to the modern internet.

2 0.84099585 494 hunch net-2014-03-11-The New York ML Symposium, take 2

Introduction: The 20 13 14 is New York Machine Learning Symposium is finally happening on March 28th at the New York Academy of Science . Every invited speaker interests me personally. They are: Rayid Ghani (Chief Scientist at Obama 2012) Brian Kingsbury (Speech Recognition @ IBM) Jorge Nocedal (who did LBFGS) We’ve been somewhat disorganized in advertising this. As a consequence, anyone who has not submitted an abstract but would like to do so may send one directly to me (jl@hunch.net title NYASMLS) by Friday March 14. I will forward them to the rest of the committee for consideration.

3 0.67947602 336 hunch net-2009-01-19-Netflix prize within epsilon

Introduction: The competitors for the Netflix Prize are tantalizingly close winning the million dollar prize. This year, BellKor and Commendo Research sent a combined solution that won the progress prize . Reading the writeups 2 is instructive. Several aspects of solutions are taken for granted including stochastic gradient descent, ensemble prediction, and targeting residuals (a form of boosting). Relatively to last year, it appears that many approaches have added parameterizations, especially for the purpose of modeling through time. The big question is: will they make the big prize? At this point, the level of complexity in entering the competition is prohibitive, so perhaps only the existing competitors will continue to try. (This equation might change drastically if the teams open source their existing solutions, including parameter settings.) One fear is that the progress is asymptoting on the wrong side of the 10% threshold. In the first year, the teams progressed through

4 0.67856735 335 hunch net-2009-01-08-Predictive Analytics World

Introduction: Carla Vicens and Eric Siegel contacted me about Predictive Analytics World in San Francisco February 18&19, which I wasn’t familiar with. A quick look at the agenda reveals several people I know working on applications of machine learning in businesses, covering deployed applications topics. It’s interesting to see a business-focused machine learning conference, as it says that we are succeeding as a field. If you are interested in deployed applications, you might attend. Eric and I did a quick interview by email. John > I’ve mostly published and participated in academic machine learning conferences like ICML, COLT, and NIPS. When I look at the set of speakers and subjects for your conference I think “machine learning for business”. Is that your understanding of things? What I’m trying to ask is: what do you view as the primary goal for this conference? Eric > You got it. This is the business event focused on the commercial deployment of technology developed at

5 0.47977248 33 hunch net-2005-02-28-Regularization

Introduction: Yaroslav Bulatov says that we should think about regularization a bit. It’s a complex topic which I only partially understand, so I’ll try to explain from a couple viewpoints. Functionally . Regularization is optimizing some representation to fit the data and minimize some notion of predictor complexity. This notion of complexity is often the l 1 or l 2 norm on a set of parameters, but the term can be used much more generally. Empirically, this often works much better than simply fitting the data. Statistical Learning Viewpoint Regularization is about the failiure of statistical learning to adequately predict generalization error. Let e(c,D) be the expected error rate with respect to D of classifier c and e(c,S) the observed error rate on a sample S . There are numerous bounds of the form: assuming i.i.d. samples, with high probability over the drawn samples S , e(c,D) less than e(c,S) + f(complexity) where complexity is some measure of the size of a s

6 0.44536492 458 hunch net-2012-03-06-COLT-ICML Open Questions and ICML Instructions

7 0.311481 10 hunch net-2005-02-02-Kolmogorov Complexity and Googling

8 0.24910802 384 hunch net-2009-12-24-Top graduates this season

9 0.23900893 369 hunch net-2009-08-27-New York Area Machine Learning Events

10 0.22101511 276 hunch net-2007-12-10-Learning Track of International Planning Competition

11 0.21858346 252 hunch net-2007-07-01-Watchword: Online Learning

12 0.21749099 43 hunch net-2005-03-18-Binomial Weighting

13 0.21733683 229 hunch net-2007-01-26-Parallel Machine Learning Problems

14 0.21725692 120 hunch net-2005-10-10-Predictive Search is Coming

15 0.21664593 426 hunch net-2011-03-19-The Ideal Large Scale Learning Class

16 0.21531351 34 hunch net-2005-03-02-Prior, “Prior” and Bias

17 0.21530996 258 hunch net-2007-08-12-Exponentiated Gradient

18 0.21452866 475 hunch net-2012-10-26-ML Symposium and Strata-Hadoop World

19 0.21324143 286 hunch net-2008-01-25-Turing’s Club for Machine Learning

20 0.21301149 345 hunch net-2009-03-08-Prediction Science