gwern comments on [Link]"Neural Turing Machines" - Less Wrong

16 Post author: Prankster 31 October 2014 08:54AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (21)

You are viewing a single comment's thread. Show more comments above.

Comment author: IlyaShpitser 31 October 2014 06:19:17PM 1 point [-]

A neural network is just a function of some sort (inputs to outputs). So is a finite state machine (which is what a regular Turing machine uses). Put that way, nothing conceptually interesting is going on. As usual, the devil is in the details here.

Comment author: gwern 31 October 2014 06:57:58PM *  4 points [-]

If you want to put it that way, nothing conceptually interesting is going on in any neural network paper - we already know they're universal.

My problem is the details: I visualize neural networks as like pachinko machines or Galton's quincunx - you drop a bunch of bits (many balls) into the top (bottom) layer of neurons (pins) and they cascade down to the bottom based on the activation functions (spacing of pins & how many balls hit a pin simultaneously), and at the bottom (top) is emitted a final smaller output (1 ball somewhere). I don't get the details of what it means to add a memory to this many-to-one function.

Comment author: Daniel_Burfoot 31 October 2014 07:06:35PM 6 points [-]

Recurrent Neural Networks feed their outputs back into their inputs, and run continuously. Instead of visualizing a river flowing directly downstream, visualize a river with a little cove off to one side, where vortices swirl around.

Comment author: V_V 31 October 2014 08:32:59PM 3 points [-]

The same way you add memory to many-to-one boolean functions: feedback loops.

Comment author: IlyaShpitser 01 November 2014 03:33:38PM *  2 points [-]

I think another way to look at neural networks is they are nested non-linear regression models.


I am probably in the minority here, but I don't think the stuff in the OP is that interesting.