JoshuaZ comments on Bayes Slays Goodman's Grue - Less Wrong

0 Post author: potato 17 November 2011 10:45AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (120)

You are viewing a single comment's thread. Show more comments above.

Comment author: JoshuaZ 17 November 2011 04:51:15AM 2 points [-]

So does that mean that if an entity had a neuronal structure that intuited grue and bleen it would be justified in treating the hypothesis that way? I'd be willing to bite that bullet I think.

Comment author: moshez 17 November 2011 06:14:42PM 4 points [-]

It means that that entity's evolved instincts would be out-of-whack with the MML, so if that entity also got to the point where it invented Turing machines, it would see the flaw in its reasoning. This is no different than realizing that Maxwell's equations, though they look more complicated than "anger" to a human, are actually simpler. Sometimes, the intuition is wrong. In the blue/grue case, human intuition happens to not be wrong, but a hypothetical entity is -- and both humans and the entity, after understanding math and computer science, would agree that humans are wrong about anger, and hypothetical entities are wrong about grue. Why is that a problem?