Lumifer comments on How valuable is it to learn math deeply? - LessWrong

20 Post author: JonahSinick 02 September 2013 06:01PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (79)

You are viewing a single comment's thread. Show more comments above.

Comment author: Lumifer 11 October 2013 03:40:19PM *  0 points [-]

This should be straightforwardly testable by standard statistics.

Given the empirical distribution of IQ scores and given the estimated measurement error (which depends on the score -- scores in the tails are much less accurate) one should be able to come up with a probability that the empirical distribution was drawn from a particular normal.

Although I don't know if I'd want to include cases with clear brain damage (e.g. Downs) into the population for this purpose.

Comment author: Vaniver 11 October 2013 03:47:08PM *  1 point [-]

This should be straightforwardly testable by standard statistics.

Agreed.

Given the empirical distribution of IQ scores

If you have a source for one of these, I would love to see it. I haven't been able to find any, but I also haven't put on my "I'm affiliated with a research university" hat and emailed people asking for their data, so it might be available.

estimated measurement error (which depends on the score -- scores in the tails are much less accurate)

Agreed that this should be the case, but it's not clear to me how to estimate measurement error besides test-retest variability, which can be corrupted by learning effects unless you wait a significant time between tests. I think Project Talent only tested its subjects once, but unless you have something of that size which tests people during adulthood several times you're unlikely to get sufficient data to have a good estimate here.

Comment author: private_messaging 11 October 2013 04:28:37PM *  2 points [-]

This should be straightforwardly testable by standard statistics

Agreed.

That may require prohibitively large sample sizes, i.e. not be testable.

With regards to measuring g, and high IQs, you need to keep in mind regression towards the mean, which becomes fairly huge at the high range, even for fairly strongly correlated variables.

Other more subtle issue is that proxies generally fare even worse far from the mean than you'd expect from regression alone. I.e. if you use grip strength as a proxy for how quick someone runs a mile, that'll obviously work great for your average person, but at the very high range - professional athletes - you could obtain negative correlation because athletes with super strong grip - weightlifters maybe? - aren't very good runners, and very good runners do not have extreme grip strength. It's not very surprising that folks like Chris Langan are at very best mediocre crackpots rather than super-Einsteins.

Comment author: Lumifer 11 October 2013 04:33:11PM 0 points [-]

That may require prohibitively large sample sizes, i.e. not be testable.

At least for certain populations the sample sizes should be pretty large. Also a smaller-than-desired sample size doesn't mean it's not testable, all it means is that your confidence in the outcome will be lower.

proxies generally fare even worse far from the mean than you'd expect from regression alone

Yes, I agree. The tails are a problem in general, estimation in the tails gets very fuzzy very quickly.

Comment author: private_messaging 11 October 2013 04:39:35PM 1 point [-]

Yes, I agree. The tails are a problem in general, estimation in the tails gets very fuzzy very quickly.

And it seems to me that having studied math complete with boring exercises could help with understanding of that somewhat... all too often you see people not even ballpark by just how much necessary application of regression towards the mean affects the rarity.

Comment author: Lumifer 11 October 2013 04:12:37PM 1 point [-]

Now that I've started to think about it, the estimation of the measurement error might be a problem.

First we need to keep in mind the difference between precision and accuracy. Re-tests will only help with precision, obviously.

Moreover, given that we're trying to measure g, it happens to be unobservable. That makes estimates of accuracy somewhat iffy. Maybe it will help if you define g "originally", as the first principal component of a variety of IQ tests...

On the other hand, I think our measurement error estimates can afford to be guesstimates and as long as they are in the ballpark we shouldn't have too many problems.

As to the empirical datasets, I don't have time atm to go look for them, but didn't US Army and such ran large studies at some point? Theoretically the results should be in public domain. We can also look at proxies (of the SAT/GRE/GMAT/LSAT/etc.) kind, but, of course, these are only imperfect proxies.