fast_ml fast_ml-2013 fast_ml-2013-15 knowledge-graph by maker-knowledge-mining

15 fast ml-2013-01-07-Machine learning courses online


meta infos for this blog

Source: html

Introduction: How do you learn machine learning? A good way to begin is to take an online course. These courses started appearing towards the end of 2011, first from Stanford University, now from Coursera , Udacity , edX and other institutions. There are very many of them, including a few about machine learning. Here’s a list: Introduction to Artificial Intelligence by Sebastian Thrun and Peter Norvig. That was the first online class, and it contains two units on machine learning (units five and six). Both instructors work at Google. Sebastian Thrun is best known for building a self-driving car and Peter Norvig is a leading authority on AI, so they know what they are talking about. After the success of the class Sebastian Thrun quit Stanford to found Udacity, his online learning startup. Machine Learning by Andrew Ng. Again, one of the first classes, by Stanford professor who started Coursera, the best known online learning provider today. Andrew Ng is a world class authority on m


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 There are very many of them, including a few about machine learning. [sent-4, score-0.21]

2 That was the first online class, and it contains two units on machine learning (units five and six). [sent-6, score-0.452]

3 After the success of the class Sebastian Thrun quit Stanford to found Udacity, his online learning startup. [sent-9, score-0.451]

4 Again, one of the first classes, by Stanford professor who started Coursera, the best known online learning provider today. [sent-11, score-0.474]

5 Andrew Ng is a world class authority on machine learning, and this course is a good place to start. [sent-12, score-0.643]

6 It features well chosen topics (notably missing are trees and ensembles) and programming assignments (Matlab/Octave). [sent-13, score-0.434]

7 The course seems very good - it’s engaging and well-presented. [sent-19, score-0.349]

8 Yet by watching it you can get a clue why the craze these days is data science, big data, deep learning etc. [sent-21, score-0.26]

9 This course has a strong emphasis on theory of learning. [sent-30, score-0.224]

10 Originally it was broadcasted live from Caltech site , so you could watch the lecture and ask your question afterwards by means of an online chat. [sent-34, score-0.453]

11 Part two about unsupervised learning and part three about reinforcement learning are coming. [sent-37, score-0.312]

12 This is not a machine learning course, but rather a linear algebra course. [sent-43, score-0.47]

13 Linear algebra, that is matrix and vector operations, is the foundation of machine learning. [sent-44, score-0.217]

14 So if you find yourself struggling with math a little bit this is the course to take. [sent-45, score-0.224]

15 The presentation is straightforward and application-oriented and the programming is done in Python. [sent-46, score-0.276]

16 An edX course from University of Texas at Austin, programming in Python. [sent-48, score-0.5]

17 One example is Computing for Data Analysis by Roger Peng, a very good introduction to R programming language. [sent-50, score-0.484]

18 Be sure to browse Udacity courses , they are generally easier to digest than Coursera’s. [sent-54, score-0.208]

19 Here are a few non-interactive resources, mainly course video lectures: Machine Learning by Pedro Domingos. [sent-59, score-0.362]

20 The course has an interesting and pretty comprehensive choice of topics and the content is good, even though video quality is not quite up to par with other courses. [sent-61, score-0.527]


similar blogs computed by tfidf model

tfidf for this blog:

wordName wordTfidf (topN-words)

[('programming', 0.276), ('coursera', 0.26), ('course', 0.224), ('introduction', 0.208), ('sebastian', 0.188), ('thrun', 0.188), ('machine', 0.158), ('class', 0.157), ('algebra', 0.156), ('udacity', 0.156), ('university', 0.156), ('learning', 0.156), ('science', 0.153), ('online', 0.138), ('site', 0.138), ('video', 0.138), ('courses', 0.125), ('edx', 0.125), ('engaging', 0.125), ('lectures', 0.125), ('watch', 0.125), ('stanford', 0.125), ('topics', 0.106), ('authority', 0.104), ('de', 0.104), ('harvard', 0.104), ('professor', 0.104), ('andrew', 0.083), ('easier', 0.083), ('statistics', 0.083), ('classes', 0.083), ('started', 0.076), ('networks', 0.075), ('peter', 0.071), ('neural', 0.063), ('though', 0.059), ('matrix', 0.059), ('including', 0.052), ('pedro', 0.052), ('assignments', 0.052), ('geoffrey', 0.052), ('visualization', 0.052), ('afterwards', 0.052), ('bill', 0.052), ('caltech', 0.052), ('clue', 0.052), ('craze', 0.052), ('cs', 0.052), ('education', 0.052), ('focus', 0.052)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.99999988 15 fast ml-2013-01-07-Machine learning courses online

Introduction: How do you learn machine learning? A good way to begin is to take an online course. These courses started appearing towards the end of 2011, first from Stanford University, now from Coursera , Udacity , edX and other institutions. There are very many of them, including a few about machine learning. Here’s a list: Introduction to Artificial Intelligence by Sebastian Thrun and Peter Norvig. That was the first online class, and it contains two units on machine learning (units five and six). Both instructors work at Google. Sebastian Thrun is best known for building a self-driving car and Peter Norvig is a leading authority on AI, so they know what they are talking about. After the success of the class Sebastian Thrun quit Stanford to found Udacity, his online learning startup. Machine Learning by Andrew Ng. Again, one of the first classes, by Stanford professor who started Coursera, the best known online learning provider today. Andrew Ng is a world class authority on m

2 0.19533667 37 fast ml-2013-09-03-Our followers and who else they follow

Introduction: Recently we hit 400 followers mark on Twitter. To celebrate we decided to do some data mining on you , specifically to discover who our followers are and who else they follow. For your viewing pleasure we packaged the results nicely with Bootstrap. Here’s some data science in action. Our followers This table show our 20 most popular followers as measeared by their follower count. The occasional question marks stand for non-ASCII characters. Each link opens a new window.   Followers Screen name Name Description 8685 pankaj Pankaj Gupta I lead the Personalization and Recommender Systems group at Twitter. Founded two startups in the past.   5070 ogrisel Olivier Grisel Datageek, contributor to scikit-learn, works with Python / Java / Clojure / Pig, interested in Machine Learning, NLProc, {Big|Linked|Open} Data and braaains!   4582 thuske thuske & 4442 ram Ram Ravichandran So

3 0.15401696 62 fast ml-2014-05-26-Yann LeCun's answers from the Reddit AMA

Introduction: On May 15th Yann LeCun answered “ask me anything” questions on Reddit . We hand-picked some of his thoughts and grouped them by topic for your enjoyment. Toronto, Montreal and New York All three groups are strong and complementary. Geoff (who spends more time at Google than in Toronto now) and Russ Salakhutdinov like RBMs and deep Boltzmann machines. I like the idea of Boltzmann machines (it’s a beautifully simple concept) but it doesn’t scale well. Also, I totally hate sampling. Yoshua and his colleagues have focused a lot on various unsupervised learning, including denoising auto-encoders, contracting auto-encoders. They are not allergic to sampling like I am. On the application side, they have worked on text, not so much on images. In our lab at NYU (Rob Fergus, David Sontag, me and our students and postdocs), we have been focusing on sparse auto-encoders for unsupervised learning. They have the advantage of scaling well. We have also worked on applications, mostly to v

4 0.099016279 57 fast ml-2014-04-01-Exclusive Geoff Hinton interview

Introduction: Geoff Hinton is a living legend. He almost single-handedly invented backpropagation for training feed-forward neural networks. Despite in theory being universal function approximators, these networks turned out to be pretty much useless for more complex problems, like computer vision and speech recognition. Professor Hinton responded by creating deep networks and deep learning, an ultimate form of machine learning. Recently we’ve been fortunate to ask Geoff a few questions and have him answer them. Geoff, thanks so much for talking to us. You’ve had a long and fruitful career. What drives you these days? Well, after a man hits a certain age, his priorities change. Back in the 80s I was happy when I was able to train a network with eight hidden units. Now I can finally have thousands and possibly millions of them. So I guess the answer is scale. Apart from that, I like people at Google and I like making them a ton of money. They happen to pay me well, so it’s a win-win situ

5 0.096734762 42 fast ml-2013-10-28-How much data is enough?

Introduction: A Reddit reader asked how much data is needed for a machine learning project to get meaningful results. Prof. Yaser Abu-Mostafa from Caltech answered this very question in his online course . The answer is that as a rule of thumb, you need roughly 10 times as many examples as there are degrees of freedom in your model. In case of a linear model, degrees of freedom essentially equal data dimensionality (a number of columns). We find that thinking in terms of dimensionality vs number of examples is a convenient shortcut. The more powerful the model, the more it’s prone to overfitting and so the more examples you need. And of course the way of controlling this is through validation. Breaking the rules In practice you can get away with less than 10x, especially if your model is simple and uses regularization. In Kaggle competitions the ratio is often closer to 1:1, and sometimes dimensionality is far greater than a number of examples, depending on how you pre-process the data

6 0.095424481 29 fast ml-2013-05-25-More on sparse filtering and the Black Box competition

7 0.089695781 27 fast ml-2013-05-01-Deep learning made easy

8 0.085809767 41 fast ml-2013-10-09-Big data made easy

9 0.07297837 24 fast ml-2013-03-25-Dimensionality reduction for sparse binary data - an overview

10 0.071752332 46 fast ml-2013-12-07-13 NIPS papers that caught our eye

11 0.071282074 40 fast ml-2013-10-06-Pylearn2 in practice

12 0.065347031 58 fast ml-2014-04-12-Deep learning these days

13 0.063140601 45 fast ml-2013-11-27-Object recognition in images with cuda-convnet

14 0.063005053 12 fast ml-2012-12-21-Tuning hyperparams automatically with Spearmint

15 0.062244758 55 fast ml-2014-03-20-Good representations, distance, metric learning and supervised dimensionality reduction

16 0.061092407 16 fast ml-2013-01-12-Intro to random forests

17 0.058020711 21 fast ml-2013-02-27-Dimensionality reduction for sparse binary data

18 0.056589235 7 fast ml-2012-10-05-Predicting closed questions on Stack Overflow

19 0.055709563 54 fast ml-2014-03-06-PyBrain - a simple neural networks library in Python

20 0.049897201 2 fast ml-2012-08-27-Kaggle job recommendation challenge


similar blogs computed by lsi model

lsi for this blog:

topicId topicWeight

[(0, 0.234), (1, 0.158), (2, 0.245), (3, 0.007), (4, 0.135), (5, 0.149), (6, 0.018), (7, 0.133), (8, 0.319), (9, 0.133), (10, -0.139), (11, 0.156), (12, -0.054), (13, 0.07), (14, 0.159), (15, 0.075), (16, 0.035), (17, -0.034), (18, -0.011), (19, -0.097), (20, 0.034), (21, 0.088), (22, 0.015), (23, 0.008), (24, 0.017), (25, -0.16), (26, 0.134), (27, 0.259), (28, -0.046), (29, -0.01), (30, 0.113), (31, -0.08), (32, 0.082), (33, -0.175), (34, 0.009), (35, -0.09), (36, 0.189), (37, 0.111), (38, 0.064), (39, -0.148), (40, 0.148), (41, 0.242), (42, -0.028), (43, -0.157), (44, -0.044), (45, 0.14), (46, 0.145), (47, 0.065), (48, 0.121), (49, 0.124)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.97778517 15 fast ml-2013-01-07-Machine learning courses online

Introduction: How do you learn machine learning? A good way to begin is to take an online course. These courses started appearing towards the end of 2011, first from Stanford University, now from Coursera , Udacity , edX and other institutions. There are very many of them, including a few about machine learning. Here’s a list: Introduction to Artificial Intelligence by Sebastian Thrun and Peter Norvig. That was the first online class, and it contains two units on machine learning (units five and six). Both instructors work at Google. Sebastian Thrun is best known for building a self-driving car and Peter Norvig is a leading authority on AI, so they know what they are talking about. After the success of the class Sebastian Thrun quit Stanford to found Udacity, his online learning startup. Machine Learning by Andrew Ng. Again, one of the first classes, by Stanford professor who started Coursera, the best known online learning provider today. Andrew Ng is a world class authority on m

2 0.46645859 37 fast ml-2013-09-03-Our followers and who else they follow

Introduction: Recently we hit 400 followers mark on Twitter. To celebrate we decided to do some data mining on you , specifically to discover who our followers are and who else they follow. For your viewing pleasure we packaged the results nicely with Bootstrap. Here’s some data science in action. Our followers This table show our 20 most popular followers as measeared by their follower count. The occasional question marks stand for non-ASCII characters. Each link opens a new window.   Followers Screen name Name Description 8685 pankaj Pankaj Gupta I lead the Personalization and Recommender Systems group at Twitter. Founded two startups in the past.   5070 ogrisel Olivier Grisel Datageek, contributor to scikit-learn, works with Python / Java / Clojure / Pig, interested in Machine Learning, NLProc, {Big|Linked|Open} Data and braaains!   4582 thuske thuske & 4442 ram Ram Ravichandran So

3 0.37708527 62 fast ml-2014-05-26-Yann LeCun's answers from the Reddit AMA

Introduction: On May 15th Yann LeCun answered “ask me anything” questions on Reddit . We hand-picked some of his thoughts and grouped them by topic for your enjoyment. Toronto, Montreal and New York All three groups are strong and complementary. Geoff (who spends more time at Google than in Toronto now) and Russ Salakhutdinov like RBMs and deep Boltzmann machines. I like the idea of Boltzmann machines (it’s a beautifully simple concept) but it doesn’t scale well. Also, I totally hate sampling. Yoshua and his colleagues have focused a lot on various unsupervised learning, including denoising auto-encoders, contracting auto-encoders. They are not allergic to sampling like I am. On the application side, they have worked on text, not so much on images. In our lab at NYU (Rob Fergus, David Sontag, me and our students and postdocs), we have been focusing on sparse auto-encoders for unsupervised learning. They have the advantage of scaling well. We have also worked on applications, mostly to v

4 0.24626122 16 fast ml-2013-01-12-Intro to random forests

Introduction: Let’s step back from forays into cutting edge topics and look at a random forest, one of the most popular machine learning techniques today. Why is it so attractive? First of all, decision tree ensembles have been found by Caruana et al. as the best overall approach for a variety of problems. Random forests, specifically, perform well both in low dimensional and high dimensional tasks. There are basically two kinds of tree ensembles: bagged trees and boosted trees. Bagging means that when building each subsequent tree, we don’t look at the earlier trees, while in boosting we consider the earlier trees and strive to compensate for their weaknesses (which may lead to overfitting). Random forest is an example of the bagging approach, less prone to overfit. Gradient boosted trees (notably GBM package in R) represent the other one. Both are very successful in many applications. Trees are also relatively fast to train, compared to some more involved methods. Besides effectivnes

5 0.21906568 29 fast ml-2013-05-25-More on sparse filtering and the Black Box competition

Introduction: The Black Box challenge has just ended. We were thoroughly thrilled to learn that the winner, doubleshot , used sparse filtering, apparently following our cue. His score in terms of accuracy is 0.702, ours 0.645, and the best benchmark 0.525. We ranked 15th out of 217, a few places ahead of the Toronto team consisting of Charlie Tang and Nitish Srivastava . To their credit, Charlie has won the two remaining Challenges in Representation Learning . Not-so-deep learning The difference to our previous, beating-the-benchmark attempt is twofold: one layer instead of two for supervised learning, VW instead of a random forest Somewhat suprisingly, one layer works better than two. Even more surprisingly, with enough units you can get 0.634 using a linear model (Vowpal Wabbit, of course, One-Against-All). In our understanding, that’s the point of overcomplete representations*, which Stanford people seem to care much about. Recall The secret of the big guys and the pape

6 0.21589419 42 fast ml-2013-10-28-How much data is enough?

7 0.17173296 40 fast ml-2013-10-06-Pylearn2 in practice

8 0.16961624 27 fast ml-2013-05-01-Deep learning made easy

9 0.16913475 24 fast ml-2013-03-25-Dimensionality reduction for sparse binary data - an overview

10 0.16286442 55 fast ml-2014-03-20-Good representations, distance, metric learning and supervised dimensionality reduction

11 0.16161726 12 fast ml-2012-12-21-Tuning hyperparams automatically with Spearmint

12 0.15512481 46 fast ml-2013-12-07-13 NIPS papers that caught our eye

13 0.14075267 57 fast ml-2014-04-01-Exclusive Geoff Hinton interview

14 0.1387053 19 fast ml-2013-02-07-The secret of the big guys

15 0.13130891 58 fast ml-2014-04-12-Deep learning these days

16 0.13084696 28 fast ml-2013-05-12-And deliver us from Weka

17 0.12897721 21 fast ml-2013-02-27-Dimensionality reduction for sparse binary data

18 0.12184837 45 fast ml-2013-11-27-Object recognition in images with cuda-convnet

19 0.1143459 54 fast ml-2014-03-06-PyBrain - a simple neural networks library in Python

20 0.11429535 2 fast ml-2012-08-27-Kaggle job recommendation challenge


similar blogs computed by lda model

lda for this blog:

topicId topicWeight

[(6, 0.014), (26, 0.02), (30, 0.503), (31, 0.066), (35, 0.019), (55, 0.019), (69, 0.086), (71, 0.029), (73, 0.017), (78, 0.044), (79, 0.016), (84, 0.05), (99, 0.04)]

similar blogs list:

simIndex simValue blogId blogTitle

same-blog 1 0.91984606 15 fast ml-2013-01-07-Machine learning courses online

Introduction: How do you learn machine learning? A good way to begin is to take an online course. These courses started appearing towards the end of 2011, first from Stanford University, now from Coursera , Udacity , edX and other institutions. There are very many of them, including a few about machine learning. Here’s a list: Introduction to Artificial Intelligence by Sebastian Thrun and Peter Norvig. That was the first online class, and it contains two units on machine learning (units five and six). Both instructors work at Google. Sebastian Thrun is best known for building a self-driving car and Peter Norvig is a leading authority on AI, so they know what they are talking about. After the success of the class Sebastian Thrun quit Stanford to found Udacity, his online learning startup. Machine Learning by Andrew Ng. Again, one of the first classes, by Stanford professor who started Coursera, the best known online learning provider today. Andrew Ng is a world class authority on m

2 0.23211381 37 fast ml-2013-09-03-Our followers and who else they follow

Introduction: Recently we hit 400 followers mark on Twitter. To celebrate we decided to do some data mining on you , specifically to discover who our followers are and who else they follow. For your viewing pleasure we packaged the results nicely with Bootstrap. Here’s some data science in action. Our followers This table show our 20 most popular followers as measeared by their follower count. The occasional question marks stand for non-ASCII characters. Each link opens a new window.   Followers Screen name Name Description 8685 pankaj Pankaj Gupta I lead the Personalization and Recommender Systems group at Twitter. Founded two startups in the past.   5070 ogrisel Olivier Grisel Datageek, contributor to scikit-learn, works with Python / Java / Clojure / Pig, interested in Machine Learning, NLProc, {Big|Linked|Open} Data and braaains!   4582 thuske thuske & 4442 ram Ram Ravichandran So

3 0.2309612 19 fast ml-2013-02-07-The secret of the big guys

Introduction: Are you interested in linear models, or K-means clustering? Probably not much. These are very basic techniques with fancier alternatives. But here’s the bomb: when you combine those two methods for supervised learning, you can get better results than from a random forest. And maybe even faster. We have already written about Vowpal Wabbit , a fast linear learner from Yahoo/Microsoft. Google’s response (or at least, a Google’s guy response) seems to be Sofia-ML . The software consists of two parts: a linear learner and K-means clustering. We found Sofia a while ago and wondered about K-means: who needs K-means? Here’s a clue: This package can be used for learning cluster centers (…) and for mapping a given data set onto a new feature space based on the learned cluster centers. Our eyes only opened when we read a certain paper, namely An Analysis of Single-Layer Networks in Unsupervised Feature Learning ( PDF ). The paper, by Coates , Lee and Ng, is about object recogni

4 0.21644442 54 fast ml-2014-03-06-PyBrain - a simple neural networks library in Python

Introduction: We have already written a few articles about Pylearn2 . Today we’ll look at PyBrain. It is another Python neural networks library, and this is where similiarites end. They’re like day and night: Pylearn2 - Byzantinely complicated, PyBrain - simple. We attempted to train a regression model and succeeded at first take (more on this below). Try this with Pylearn2. While there are a few machine learning libraries out there, PyBrain aims to be a very easy-to-use modular library that can be used by entry-level students but still offers the flexibility and algorithms for state-of-the-art research. The library features classic perceptron as well as recurrent neural networks and other things, some of which, for example Evolino , would be hard to find elsewhere. On the downside, PyBrain feels unfinished, abandoned. It is no longer actively developed and the documentation is skimpy. There’s no modern gimmicks like dropout and rectified linear units - just good ol’ sigmoid and ta

5 0.21028081 62 fast ml-2014-05-26-Yann LeCun's answers from the Reddit AMA

Introduction: On May 15th Yann LeCun answered “ask me anything” questions on Reddit . We hand-picked some of his thoughts and grouped them by topic for your enjoyment. Toronto, Montreal and New York All three groups are strong and complementary. Geoff (who spends more time at Google than in Toronto now) and Russ Salakhutdinov like RBMs and deep Boltzmann machines. I like the idea of Boltzmann machines (it’s a beautifully simple concept) but it doesn’t scale well. Also, I totally hate sampling. Yoshua and his colleagues have focused a lot on various unsupervised learning, including denoising auto-encoders, contracting auto-encoders. They are not allergic to sampling like I am. On the application side, they have worked on text, not so much on images. In our lab at NYU (Rob Fergus, David Sontag, me and our students and postdocs), we have been focusing on sparse auto-encoders for unsupervised learning. They have the advantage of scaling well. We have also worked on applications, mostly to v

6 0.20710111 12 fast ml-2012-12-21-Tuning hyperparams automatically with Spearmint

7 0.20504837 48 fast ml-2013-12-28-Regularizing neural networks with dropout and with DropConnect

8 0.20294389 34 fast ml-2013-07-14-Running things on a GPU

9 0.20020236 29 fast ml-2013-05-25-More on sparse filtering and the Black Box competition

10 0.19881457 9 fast ml-2012-10-25-So you want to work for Facebook

11 0.19870691 18 fast ml-2013-01-17-A very fast denoising autoencoder

12 0.19714348 27 fast ml-2013-05-01-Deep learning made easy

13 0.19562724 11 fast ml-2012-12-07-Predicting wine quality

14 0.19385302 40 fast ml-2013-10-06-Pylearn2 in practice

15 0.19223441 13 fast ml-2012-12-27-Spearmint with a random forest

16 0.19171371 16 fast ml-2013-01-12-Intro to random forests

17 0.19102983 23 fast ml-2013-03-18-Large scale L1 feature selection with Vowpal Wabbit

18 0.19039014 55 fast ml-2014-03-20-Good representations, distance, metric learning and supervised dimensionality reduction

19 0.18959171 7 fast ml-2012-10-05-Predicting closed questions on Stack Overflow

20 0.18720768 43 fast ml-2013-11-02-Maxing out the digits