Yvain comments on Looking for information on scoring calibration - Less Wrong

8 Post author: Yvain 29 January 2011 10:24PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (13)

You are viewing a single comment's thread. Show more comments above.

Comment author: Yvain 30 January 2011 01:37:52PM 1 point [-]

Do you have a more specific purpose in mind? I'm curious what spurred your question.

A prof doing an experiment gave me a bunch of data from calibration tests with demographic identifiers, and I'd like to be able to analyze it to say things like "Old people have better calibration than young people" or "Training in finance improves your calibration".

Comment author: datadataeverywhere 30 January 2011 04:33:53PM 1 point [-]

Oh, excellent. I do love data. What is the format (what is the maximum amount of information you have about each individual)?

Given that you already have the data, (and you probably have reason to suspect that individuals were not trying to game the test?), I suspect the best way is to graph both accuracy and anticipated accuracy against the chosen demographic, and then for all your readers who want numbers, compute either the ratio or the difference of those two and publish the PMCC of that against the demographic (it's Frequentist, but it's also standard practice, and I've had papers rejected that don't follow it...).

Comment author: Cyan 31 January 2011 03:04:11AM 0 points [-]

...PMCC...

I'm not sure what the Pacific Mennonite Children's Choir has to do with it... oh wait, nevermind.

Comment author: b1shop 31 January 2011 03:14:06AM 0 points [-]

Leaving them with two separate metrics would allow you to make interesting statements like "financial training increased accuracy, but it also decreased calibration. Subjects overestimated their ability."