Daniel_Burfoot comments on Deepmind Plans for Rat-Level AI - Less Wrong

20 Post author: moridinamael 18 August 2016 04:26PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (9)

You are viewing a single comment's thread.

Comment author: Daniel_Burfoot 20 August 2016 02:30:57AM 6 points [-]

Note that DeepMind's two big successes (Atari and Go) come from scenarios that are perfectly simulable in a computer. That means they can generate an arbitrarily large number of data points to train their massive neural networks. Real world ML problems almost all have strict limitations on the amount of training data that is available.

Comment author: Houshalter 27 August 2016 01:00:53AM 3 points [-]

I'm not sure this is true. The internet contains billions of hours of video, trillions of images, and libraries worth of text. If they can use unsupervised, semi-supervised, or weakly-supervised learning, they could take advantage of nearly limitless data. And neural networks can do unsupervised learning well, by learning features for one task and then transferring those to another task.

Deepmind has also had a paper on approximate bayesian learning for neural net parameters. That would make them much more able to learn from limited amounts of data, instead of overfitting.

Anyway deep nets are not really going to take over traditional ML methods, but rather open up a whole new set of problems that traditional methods can't handle. Like processing audio and video data, or reinforcement learning.

Comment author: moridinamael 20 August 2016 01:10:55PM 3 points [-]

That is true. However, since they released those papers, they've published some results demonstrating learning from only a handful of samples in certain contexts by using specialized memory networks which seem to be more analogous to human memory.

Comment author: Gurkenglas 20 August 2016 10:19:07PM *  2 points [-]

On the other hand, it's simple to generate AI-complete problems where you can generate training data.