Singularity Institute desperately needs someone who is not me who can write cognitive-science-based material. Someone smart, energetic, able to speak to popular audiences, and with an excellent command of the science. If you’ve been reading Less Wrong for the last few months, you probably just thought the same thing I did: “SIAI should hire Lukeprog!” To support Luke Muelhauser becoming a full-time Singularity Institute employee, please donate and mention Luke (e.g. “Yay for Luke!”) in the check memo or the comment field of your donation - or if you donate by a method that doesn’t allow you to leave a comment, tell Louie Helm (louie@intelligence.org) your donation was to help fund Luke.
Note that the Summer Challenge that doubles all donations will run until August 31st. (We're currently at $31,000 of $125,000.)
During his stint as a Singularity Institute Visiting Fellow, Luke has already:
- Co-organized and taught sessions for a well-received one-week Rationality Minicamp, and taught sessions for the nine-week Rationality Boot Camp.
- Written many helpful and well-researched articles for Less Wrong on metaethics, rationality theory, and rationality practice, including the 20-page tutorial A Crash Course in the Neuroscience of Human Motivation.
- Written a new Singularity FAQ.
- Published an intelligence explosion website for academics.
- ...and completed many smaller projects.
As a full-time Singularity Institute employee, Luke could:
- Author and co-author research papers and outreach papers, including
- A chapter already accepted to Springer’s The Singularity Hypothesis volume (co-authored with Louie Helm).
- A paper on existential risk and optimal philanthropy, co-authored with a Columbia University researcher.
- Continue to write articles for Less Wrong on the theory and practice of rationality.
- Write a report that summarizes unsolved problems related to Friendly AI.
- Continue to develop his metaethics sequence, the conclusion of which will be a sort of Polymath Project for collaboratively solving open problems in metaethics relevant to FAI development.
- Teach courses on rationality and social effectiveness, as he has been doing for the Singularity Institute’s Rationality Minicamp and Rationality Boot Camp.
- Produce introductory materials to help bridge inferential gaps, as he did with the Singularity FAQ.
- Raise awareness of AI risk and the uses of rationality by giving talks at universities and technology companies, as he recently did at Halcyon Molecular.
If you’d like to help us fund Luke Muehlhauser to do all that and probably more, please donate now and include the word “Luke” in the comment field. And if you donate before August 31st, your donation will be doubled as part of the 2011 Summer Singularity Challenge.
I mean 'evidence' in the Bayesian sense, not the scientific sense. I have significant Bayesian evidence that minicamp was a success on several measures, but I can't know more until we collect more data.
Thanks for providing a list of information we could post. One reason for not posting more information is that doing so requires lots of staff hours, and we don't have enough of those available. We're also trying to, for example, develop a rationality curriculum and write a document of open problems in FAI theory.
If you're anxious to learn more about the rationality camps before SI has time to publish about that data, you're welcome to contact the people who attended; many of them have identified themselves on Less Wrong.
I'm fairly confident that campers got more out of my fashion sessions than what they can learn only from looking at a few fashion magazines.
Cheers,
Luke
This comes off very strongly as the typical bureaucratic protectiveness - a business doesn't want to share raw data, because raw data is a valuable resource. If you came out and said this was the reason, I'd be more understanding, but it would still feel like a major violation of community norms to be so secretive.
If simple secrecy is indeed the case, I would urge you, please, be honest about this motive and say so explicitly! At least then we are having an honest discussion, and the rest of this comment can be disregarded.
... (read more)