You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Vaniver comments on Open thread, Nov. 16 - Nov. 22, 2015 - Less Wrong Discussion

7 Post author: MrMind 16 November 2015 08:03AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (185)

You are viewing a single comment's thread. Show more comments above.

Comment author: Vaniver 16 November 2015 10:45:02PM *  3 points [-]

Unfortunately I can't easily find a link to the presentation: it was a talk on Mondrian random forests by Yee Whye Teh back in 2014. I don't think it was necessarily anything special about the presentation, since I hadn't put much thought into them before then.

The very short version is it would be nice if classifiers had fuzzy boundaries--if you look at the optimization underlying things like logistic regression, it turns out that if the underlying data is linearly separable it'll make the boundary as sharp as possible, and put it in a basically arbitrary spot. Random forests will, by averaging many weak classifiers, create one 'fuzzy' classifier that gets the probabilities mostly right in a computationally cheap fashion.

(This comment is way more opaque than I'd like, but most of the ways I'd want to elaborate on it require a chalkboard.)