Morendil comments on [LINK] Get paid to train your rationality - Less Wrong

27 Post author: XFrequentist 03 August 2011 03:01PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (55)

You are viewing a single comment's thread. Show more comments above.

Comment author: Morendil 04 August 2011 11:43:04AM 3 points [-]

Apparently the only way to know is to try. It seems likely that there is such a restriction. I'd estimate a better than 70% chance that I get turned down. :)

Comment author: gwern 05 August 2011 08:17:34PM 5 points [-]

I got an email an hour ago from the study saying I was accepted and taking me to the initial survey (a long one, covering calibration on geopolitics, finance, and religion; personality surveys with a lot of fox/hedgehog questions; basic probability; a critical thinking test, the CRT; and then what looked like a full matrix IQ test). The message at the end of all the questions:

Congratulations! You’ve completed the survey. Sometime later this year, we’ll post information on the distribution of answers among those participating in this study.

What comes next? Some of you (by random assignment) will receive an e-mail with a link to a training exercise. Again, we ask you to complete that exercise before forecasting begins on September 1st. That’s the big day for the entire team – the official start of forecasting on 9/1/2011.

Be sure to watch your e-mail for a personalized link to “your” forecasting website. We hope you’re as eager as we are for the tournament to begin.

So I'm marking me as accepted, anyway.

Comment author: Morendil 07 September 2011 10:05:03AM 2 points [-]

And the "tournament" is now begun. Just got email with login instructions.

Looks somewhat similar to PredictionBook, actually. :)

Comment author: gwern 07 September 2011 12:49:06PM 3 points [-]

I did all my predictions last night immediately after the email showed up, so that meant I got to place a lot of bets at 50/50 odds :)

(Then I recorded everything privately in PredictionBook. No point in leaving my predictions trapped on their site.)

Interface-wise, I don't like it at all. I'm still not sure what exactly I am betting at or with, compared to PB with straight probabilities or Intrade with share prices.

Comment author: Morendil 07 September 2011 02:51:23PM 2 points [-]

Did you take the "training refresher"? That includes a general-knowledge test at the end which scores you on both calibration and resolution. My results were pretty poor (but not abysmal):

You got 63% of the items correct, and your average confidence rating over all of the items was 74.33%. (...) In this exercise, your calibration is 11.00 (average confidence minus percent correct). (...) Your confidence when you were correct was 75.26%, and your confidence when you were incorrect was 72.73%. The difference is 2.53%.

I'd be curious to compare with yours if you'd care to share.

Comment author: gwern 07 September 2011 03:00:53PM 2 points [-]

Without actually going through the whole refresher, it seems to be the same; when I did the training, I don't remember that calibration/resolution test. Perhaps that is one of the experimental differences.

Comment author: Morendil 07 September 2011 03:05:02PM 2 points [-]

I didn't remember that test from earlier, either. Worth checking out? I don't mind accidentally unblinding a little if it is an experimental/control difference - curious folks will be curious.

Comment author: gwern 07 September 2011 03:16:52PM 2 points [-]

I just went through the whole thing again; there was no test of that kind at the end. (What there was was the previous multiple-choice quiz about some example forecasts and how they went wrong.) Looks like this is an experimental/control difference. I'd rather not discuss that bit further - this isn't about possibly life-or-death drugs, after all, and I already know where I can find calibration tests like that.

Comment author: Morendil 07 September 2011 03:19:57PM 1 point [-]

Fine with me. :)

BTW, look what I found. Did you know about this one?

Comment author: gwern 07 September 2011 03:34:07PM 1 point [-]

Looks like someone is being very naughty. I've asked him on Twitter.

Comment author: Morendil 08 September 2011 01:38:09PM 1 point [-]

Have you entered any comments on your predictions at the GJ site? (You're supposed to enter a minimum number of comments over one year, and also a minimum number of responses to others' comments. My understanding is that this will in time be run as a team game, with team play conventions.)

From my first experiences, I'm assuming the scoring will be pretty much as with PB.com - based on probability. Their model seems to be calibration/resolution rather than the visual "slope" representation.

Comment author: gwern 08 September 2011 02:21:31PM 1 point [-]

Comments? I don't see any relevant fields for that, checking right now, nor does my 'About' include the substring "comment". Another experimental difference, I guess...

Comment author: Morendil 09 September 2011 07:00:16AM 1 point [-]

The "Why did you answer the way you did" field. I've been assuming we're both using the same underlying app, i.e. Crowdcast. But perhaps we're not...

Comment author: gwern 04 August 2011 05:31:21PM 4 points [-]
Comment author: Morendil 05 August 2011 08:17:03PM 2 points [-]

I'm in; pleasantly surprised.

This bit from the final registration page is interesting - "But one other requirement for forecasters has changed. We can welcome those who are not US citizens." Implying that at some prior point non-US citizens were not accepted.

Comment author: XFrequentist 04 August 2011 06:00:25PM 1 point [-]

That is awesome!

Comment author: Morendil 04 August 2011 08:47:11PM 2 points [-]

Especially (mischievous mode ON) as I've only implied, not outright stated, that I've applied.

Mischievous mode OFF - that's a problem in arbitrating predictions, btw - the potential for ambiguity inherent in all human languages. If I hadn't in fact applied (I have), how should the prediction that I am "turned down" be judged?

I should use PredictionBook more often but I don't, partly due this kind of thing, also due to the trivial-inconvenience effort of having to come up with my own predictions to assess and the general uselessness for that purpose of the stream of other users' predictions.

Other than Tricycle folks, is anyone here on LW officially (or unofficially) "in charge" of maintaining and enhancing PredictionBook?

Comment author: gwern 04 August 2011 11:09:17PM *  3 points [-]

Other than Tricycle folks, is anyone here on LW officially (or unofficially) "in charge" of maintaining and enhancing PredictionBook?

I have some sort of moderator power; I am de facto in charge of the content house-keeping - editing bad due-dates, making private bad or long-overdue-unjudged predictions, criticizing predictions, etc. I also make and register hundreds of predictions, obviously.

(In addition, I have commit access to the codebase on GitHub, but I don't know Ruby, so I will probably never make use of said commit-bit.)

Comment author: Morendil 04 September 2011 03:53:59PM 1 point [-]

One thing that would probably greatly improve PB for my purposes is a tagging / filtering system, so that you could for instance pick out predictions about consumer devices or predictions about politics; or conversely leave out some uninteresting categories (e.g. predictions about the private lives of particular PB users, which I interpret as pure noise).

Comment author: gwern 04 September 2011 03:57:35PM 0 points [-]

Google is not sufficient, I take it?

Comment author: Morendil 04 September 2011 04:38:01PM 3 points [-]

No; I just tried the query "consumer electronics site:predictionbook.com", and that only returned 1 hit; I know there are more (including one I just made and another I just voted on). It really is the lack of user-supplied meta-information that prevents useful querying, not the lack of a UI for doing so. The UI encourages predictions to be written very tersely, and doesn't supply an extended-info field when you make a prediction.

PB.com is quite possibly the least well executed idea out there that I keep not giving up on. :)

Comment author: gwern 04 September 2011 05:37:13PM *  1 point [-]

Ah, that's what you meant by tags. Yes, that would be nice. On the other hand, I rather doubt that tags would instantly create massive demand for PB's services - other places like Intrade have well-categorized predictions/bets, and none of them have seen traffic explode the moment they implemented that feature.

If you really found tags all that valuable, you could start doing them inside comments. Go over the 969 upcoming predictions and add comments like 'tags: personal, exercise' or 'tags: America, politics'. Later, it'd be even easier to turn them into some real software-supported tags/categories, and in the meantime, you can query using Google. This wouldn't even take very long - at 30 predictions a day, which ought to take 10 minutes max, you'd be done in a month.

(I doubt you will adopt my suggestion and tag even 500 predictions (10%). This seems to be common to suggestions for PB: 'I'd use and really find PB useful if only it were executed better in this way', which of course never happens. It's starting to remind me of cryonics.)

Comment author: Morendil 04 September 2011 07:28:48PM *  1 point [-]

If you really found tags all that valuable, you could start doing them inside comments.

Preliminary report: this isn't going to work, not without drastic contortions in the choice of tags (which IMO kills the effectiveness of the tactic). For instance, from my first set of 30 I tagged a number with the tag "personal", predictions which only concern one user (or two acquainted with each other) and that I don't want to see because I can't effectively assess them. The Google query including "personal" returns close to 30 spurious results: for instance those containing "personal computer" or "personal transportation". (A temporary workaround is to include the term "tags" in the query, but this will cease to work once a greater fraction of predictions have been tagged.)

I doubt you will adopt my suggestion

You are correct about the likely outcome, but I think I've just proven your model of the underlying reasons wrong: I won't do it because it won't work, not because I lack the conscientiousness to do so, or because I'm too selfish to take on an effort that will benefit all users.

Comment author: Morendil 04 September 2011 07:06:22PM 1 point [-]

This seems to be common to suggestions for PB: 'I'd use and really find PB useful if only it were executed better in this way', which of course never happens.

How many times was a new feature implemented as a test of such a hypothesis?

PB.com seems like it would be a great place for things like A/B testing and other tactics in the "Lean startup" repertoire, but what actually seems to be the case is that the site isn't under active development any more; no one is apparently trying to develop traffic or usage by improving the user experience there. (This isn't to slight your own efforts or your obvious enthusiasm; merely my best current hypothesis.)

(I'm finding the comparison with cryonics ironically apt, as a fence-straddling LW user who's not giving up on the entire project despite facing, as a non-US citizen, a battery of obstacles that I suspect also apply in the US, where they're just less obvious and as a result people take it for granted that things will "just work". Though it's more likely that the comparison is entirely besides the point and just a red herring for the purposes of present discussion.)

If you really found tags all that valuable, you could start doing them inside comments.

I'll try that, for a minimum of 30 predictions.

Comment author: JoshuaZ 04 September 2011 05:54:40PM 0 points [-]

Hmm, I think this is a good idea. When I make a prediction or comment on it I will add tag remarks. It is non-ideal hack but should help a little bit.