Dr_Manhattan comments on Rationality is about pattern recognition, not reasoning - LessWrong

25 Post author: JonahSinick 26 May 2015 07:23PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (82)

You are viewing a single comment's thread.

Comment author: Dr_Manhattan 27 May 2015 04:33:42PM 1 point [-]

algorithms that people have been constructed (within the paradigm of deep learning) are highly nontransparent: nobody's been able to interpret their behavior in intelligible terms.

Not quite true Jonah: http://arxiv.org/pdf/1311.2901.pdf

Comment author: JonahSinick 27 May 2015 06:36:41PM 0 points [-]

Even if what I said isn't literally true, it's still true that the cutting edge research in pattern recognition is in deep learning, where the algorithms that are in some sense highly nontransparent.

Comment author: jacob_cannell 13 June 2015 08:56:16PM *  1 point [-]

Upon reading your comment about non-transparency in DL I thought of the exact same paper on visualizing ANN features that Dr_Manhattan posted. There was a recent post on the machine learning subreddit about using similar techniques to investigate the neural representations automatically learned in language model RNNs.

There is no intrinsic connection between transparency and automatic feature learning techniques. Consider the case of a large research team where the work in creating a visual system is divided amongst dozens of researchers, who each create specific features for certain layers/modules. The resulting features are not intrinsically opaque just because the lead researcher doesn't necessarily understand the details of each feature each engineer came up with. The lead researcher simply needs to invest the time in understanding those features (if so desired).

Deep learning simply automates the tedious feature engineering process. You can always investigate the features or specific circuits the machine came up with - if so desired. It is true that ML and DL optimization tools in particular are often used as black boxes where the researcher doesnt know or care about the details of the solution - but that does not imply that the algorithms themselves are intrinsically opaque.