MrMind comments on Noisy Reasoners - Less Wrong

11 Post author: lukeprog 13 December 2012 07:53AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (14)

You are viewing a single comment's thread. Show more comments above.

Comment author: MrMind 14 December 2012 03:18:44PM 0 points [-]

I'm not sure I've understood the sentence

Think of P(A) merely as the output of a noiseless version of the same algorithm.

because P(A) is the noiseless parameter.
Anyway, the entire paper is based on the counting algorithm to establish that random noise can give rise to structured bias, and that this is a problem for a bayesian AI.
But while the mechanism can be an interesting and maybe even correct way to unify the mentioned bias in human mind, it can hardly be posed as a problem for such an artificial intelligence. A counting algorithm for establishing probabilities basically denies everything bayesian update is designed for (the most trivial example: extraction from a finite urn).

Comment author: Kindly 14 December 2012 03:31:39PM 0 points [-]

Well, yes, the prior that yields counting algorithms is not universal. But in many cases it's good idea! And if you decide to use, for example, some rule-of-succession style modifications, the same situation appears.

In the case of a finite urn, you might see different biases (or none at all if your algorithm stubbornly refuses to update because you chose a silly prior).