wnoise comments on It's not like anything to be a bat - Less Wrong

15 Post author: Yvain 27 March 2010 02:32PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (189)

You are viewing a single comment's thread. Show more comments above.

Comment author: wnoise 30 March 2010 06:38:19PM 4 points [-]

Information is proportional not to the number of neurons, but to the number of patterns that can be stored in those neurons,

No, it's proportional to the log of the number of patterns that can be (semi-stably) stored. E.g. n bits can store 2^n patterns.

which is likely somewhere between N and N^2. I'm gonna call it NlogN.

I'd like to see a lot more justification for this. If each connection were binary (it's not), and connections were possible between all N neurons (they're not), than we would have N^2 bits.

Comment author: PhilGoetz 30 March 2010 07:11:55PM *  2 points [-]

No, it's proportional to the log of the number of patterns that can be (semi-stably) stored. E.g. n bits can store 2^n patterns.

Oops! Correct. That's what I was thinking, which is why I said info NlogN for N neurons. N neurons => max N^2 connections, 1 bit per connection, max N^2 bits, simplest model.

The math trying to estimate the number of patterns that can be stored in different neural networks is horrendous. I've seen "proofs" for Hopfield network capacity ranging from, I think, N/logN to NlogN.

Anyway, it's more-than-proportional to N, if for no other reason than that the number of connections per neuron is related to the number of neurons. A human neuron has about 10,000 connections to other neurons. Ant neurons don't.