Open Thread: May 2010

3 Post author: Jack 01 May 2010 05:29AM

You know what to do.

This thread is for the discussion of Less Wrong topics that have not appeared in recent posts. If a discussion gets unwieldy, celebrate by turning it into a top-level post.

Comments (543)

Comment author: Jack 01 May 2010 08:47:23AM *  24 points [-]

He who controls the karma controls the world.

Less Wrong dystopian speculative fiction: An excerpt.

JulXutil sat, legs crossed in the lotus position, at the center of the Less Wrong hedonist-utilitarian subreddit. Above him, in a foot-long green oval, was his karma total: 230450036. The subreddit was a giant room with a wooden floor and rice paper walls. In the middle the floor was raised, and then raised again to form a shallow step pyramid with bamboo staircases linking the levels. The subreddit was well lit. Soft light emanated from the rice paper walls as if they were lit from behind and Japanese lanterns hung from the ceiling.

Foot soldiers, users JerelYu and Maxine stood at the top of each staircase to deal with the newbies who wanted to bother the world famous JulXutil and to spot and downvote trolls before they did much damage. They also kept their eyes out for members of rival factions because while /lw/hedutil was officially public, every Less Wrong user knew this subreddit was Wireheader territory and had been since shortly after Lewis had published his famous Impossibility Proof for Friendliness. The stitched image of an envelope on JulXutil’s right sleeve turned red. He tapped it twice and the dojo disappeared and was replaced by his inbox. He tapped on the new message and its sender appeared before him.

Henry_Neil: Jul, I just heard from my source at Alcor. The procedure was successful. He’s been admitted. It'll go public in the morning.

JulXutil: Exciting, terrifying news. What will happen to his account?

Henry_Neil: It won't go anywhere. But users who haven’t logged in for thirty days don’t get counted when the server computes controlling karma. That leaves his 40% up for grabs.

JulXutil: How much support we end up with will depend on how organized the opposition is. We need full admin powers and enough backing to amend the constitution. Henry, I need you to take care of a few high karma players. They'd interfere with our plans. I’ll tell you whom. It’ll have to be timed just right. Contact me again when you've selected your men.

Henry_Neil: If the Blindsighters have heard the news they'll try the same thing. Your karmic reputation is in danger. Take precautions, stay out of the main subreddits, especially EvPsych. You’ll hear from me soon.

To be continued...

Comment author: Kutta 01 May 2010 09:52:56AM *  5 points [-]

This is golden. I demand continuation.

Comment author: Thomas 01 May 2010 03:42:03PM 2 points [-]

It's a real question where to, the Karma system leads. In a long run, we might see quite unexpected and unwanted results. But there is probably no other way to see that, than to wait where to it will actually go. I guess, a kind of conformism will prevail, if it hasn't already.

Comment author: Alexandros 12 May 2010 06:52:15AM *  11 points [-]

I have an idea that may create a (small) revenue stream for LW/SIAI. There are a lot of book recommendations, with links to amazon, going around in LW, and many of them do not use an affiliate code. Having a script add a LessWrong affiliate code to those links that don't already have one may lead to some income, especially given that affiliate codes persist and may get credited for unrelated purchases later in the day.

I believe Posterous did this, and there was a minor PR hubbub about it, but the main issue was that they did not communicate the change properly (or at all). Also, given that LW/SIAI are not-for-profit endeavours, this is much easier to swallow. In fact, if it can be done in an easy-to-implement way, I think quite a few members with popular blogs may be tempted to apply this modification to their own blogs.

Does this sound viable?

Comment author: RobinZ 12 May 2010 11:24:50AM 2 points [-]

Yes, under two conditions:

  1. It is announced in advance and properly implemented.

  2. It does not delete other affiliate codes if links are posted with affiliate codes.

Breaking both these rules is one of the many things which Livejournal has done wrong in the last few years, which is why I mention them.

Comment author: MartinB 01 May 2010 01:03:29PM *  11 points [-]

Question: Which strongly held opinion did you change in a notable way, since learning more about rationality/thinking/biases?

Comment author: [deleted] 02 May 2010 12:56:37PM 13 points [-]

Theism. Couldn't keep it. In the end, it wasn't so much that the evidence was good -- it had always been good -- as that I lost the conviction that "holding out" or "staying strong" against atheism was a virtue.

Standard liberal politics, of the sort that involved designing a utopia and giving it to people who didn't want it. I had to learn, by hearing stories, some of them terrible, that you have no choice but to respect and listen to other people, if you want to avoid hurting them in ways you really don't want to hurt them.

Comment author: Liron 03 May 2010 03:21:19AM 2 points [-]

I just listened to UC Berkeley's "Physics for Future Presidents" course on iTunes U (highly recommended) and I thought, "Surely no one can take theism seriously after experiencing what it's like to have real knowledge about the universe."

Comment author: MartinB 03 May 2010 03:48:27PM 6 points [-]

Disagreed. My current opinion is that you can be a theist and combine that with pretty much any other knowledge. Eliezer points to Robert Aumann as an example. For someone that has theism hardcoded into their brain and treats it as a different kind of knowledge than physics there can be virtually no visible difference in everyday life from a normal a-theist. I think the problem is not so much the theism, but that people use it to base decisions on it.

Comment author: [deleted] 03 May 2010 04:01:07PM 3 points [-]

oh it's true. I know deeply religious scientists. Some of them are great scientists. Let's not get unduly snide about this.

Comment author: Tyrrell_McAllister 03 May 2010 07:12:19PM 3 points [-]

There seems to be a common thought-pattern among intelligent theists. When they learn a lot about the physics of the Universe, they don't think "I should only be satisfied with beliefs in things that I understand in this deep way." Instead, they think, "As smart as I am, I have only this dim understanding of the universe. Imagine how smart I would have to be to create it! Truly, God is wonderful beyond comprehension."

Comment author: JoshuaZ 19 May 2010 03:29:31AM *  6 points [-]

I stopped being a theist a few years ago. That was due more to what Less Wrong people would call "traditional rationalism" than the sort often advocated here (I actually identify as closer to a traditionalist rationalist than a strict Bayesianism but I suspect that the level of disagreement is smaller than Eliezer makes it out to be). And part of this was certainly also emotional reactions to having the theodicy problem thrown in my face rather than direct logic.

One major update that occurred when I first took intro psych was realizing how profoundly irrational the default human thinking processes were. Before then, my general attitude was very close to humans as the rational animal. I'm not sure how relevant that is, since that's saying something like "learning about biases taught me that we are biased." I don't know if that's very helpful.

My political views have updated a lot on a variety of different issues. But I suspect that some of those are due to spending time with people who have those views rather than actually getting relevant evidence.

I've updated on how dangerous extreme theism is. It may sound strange, but this didn't arise as much out of things like terrorism, but rather becoming more aware of how many strongly held beliefs about the nature of the world there were out there that were motivated by religion and utterly at odds with reality. This was not about evolution which even in my religious phases I understood and was annoyed at by the failure of religious compatriots to understand. Rather this has included geocentrism among the Abrahamic religions, flat-Earthism among some Islamic extremists, spontaneous generation among ultra-Orthodox Jews (no really. Not a joke. And not even microscopic spontaneous generation but spontaneous generation of mice), belief among some ultra-Orthodox Jews that the kidneys are the source of moral guidance (which they use as an argument against kidney transplants).

My three most recent major updates (last six months or so) are 1) Thinking that cryonics has a substantial success probability (although I still think it is very low). This came not from actually learning more about rationality, but rather after reading some of the stuff here going back and trying to find out more about cryonics. Learning that the ice formation problem is close to completely solved substantially changed my attitude. 2) Deciding that there's a high chance that we'll have space elevators before we have practical fusion power. (This is a less trivial observation than one might think since once one has a decent space elevator it becomes pretty cheap to put up solar power satelites). This is to some extent a reevaluation based primarily on time-frames given by relevant experts. 3) Deciding that there's a substantial chance that P=NP may undecidable in ZFC. This update occurred because I was reading about how complexity results can be connected to provability of certain classes of statements in weakened forms of the Peano axioms. That makes this sound more potentially like it might be in a class of problems that have decent reasons for being undecidable.

Comment author: MartinB 19 May 2010 09:08:56AM 2 points [-]

I'm not sure how relevant that is, since that's saying something like "learning about biases taught me that we are biased." I don't know if that's very helpful.

It is! I am repeatedly surprised about a) basic level insights that are not wide spread and b) insights that other people consider basic that I do not have c) applications of an idea i understand in an area I did not think of applying it too

To list a few: People are biased => I am biased! Change is possible Understanding is possible I am a brain in a vat. Real life rocks :-)

Even after learning about cached thought, happy death and many others I still managed to fall into the traps of those.

So i consider it helpful to see where someone applies biases.

My political views have updated a lot on a variety of different issues. But I suspect that some of those are due to spending time with people who have those views rather than actually getting relevant evidence.

That statement in itself looks like a warning sign.

Comment author: JoshuaZ 19 May 2010 02:13:51PM 1 point [-]

That statement in itself looks like a warning sign.

Yeah, being aware that there are biases at play doesn't always mean I'm at all sure I'm able to correct for all of them. The problem is made more complicated by the fact that for each of the views in questions, I can point to new information leading to the updates. But I don't know if in general that's the actual cause of the updates.

Comment author: Matt_Simpson 01 May 2010 08:51:36PM 6 points [-]

I'm no longer a propertarian/Lockean/natural rights libertarian. Learning about rationality essentially made me feel comfortable letting go of a position that I honestly didn't have a good argument for (and I knew it). The ev-psych stuff scared the living hell out of me (and the libertarianism* apparently).

*At least that sort of libertarianism

Comment author: gelisam 01 May 2010 01:35:45PM 4 points [-]

I started to believe in the Big Bang here. I was convinced by the evidence, but as this comment indicates, not by the strongest evidence I was given; rather, it was necessary to contradict the specific reasoning I used to disbelieve the Big Bang in the first place.

Is this typical? I think it would be very helpful if, in addition to stating which opinion you have changed, you stated whether the evidence convinced you because it was strong or because it broke the chain of thought which led to your pre-change opinion.

Comment author: MartinB 01 May 2010 09:09:54PM *  3 points [-]

To answer my own question:

  • changed political and economic views (similar to Matt).

  • changed views on the effects of Nutrition and activity on health (including the actions that follow from that)

  • changed view on the dangers of GMO (yet again)

  • I became aware of areas where I am very ignorant of opposing arguments, and try to counterbalance

  • I finally understand the criticisms about the skeptics movement

  • I repeatedly underestimated the amount of ignorance in the world, and got shocked when discovering that

And on the funnier side. Last week I found out that i learned a minor physics fact wrong. That was not a strongly held opinion, just a fact i never looked up again till now. For some reason i always was convinced that the volume increase in freshly frozen water is 10x, while its actually more like 9%

Comment author: Vladimir_Golovin 02 May 2010 09:15:40AM 2 points [-]

As a result of reading this post, I uninstalled a 10-year old habit -- drinking a cup of strong coffee every morning. Now I drink coffee only when I feel that I need a short-term boost.

Comment author: NancyLebovitz 02 May 2010 10:10:18AM *  3 points [-]

Coffee and concentration experiment

Article about self-measurement

A few months ago, Barooah began to wean himself from coffee. His method was precise. He made a large cup of coffee and removed 20 milliliters weekly. This went on for more than four months, until barely a sip remained in the cup. He drank it and called himself cured. Unlike his previous attempts to quit, this time there were no headaches, no extreme cravings. Still, he was tempted, and on Oct. 12 last year, while distracted at his desk, he told himself that he could probably concentrate better if he had a cup. Coffee may have been bad for his health, he thought, but perhaps it was good for his concentration.

Barooah wasn’t about to try to answer a question like this with guesswork. He had a good data set that showed how many minutes he spent each day in focused work. With this, he could do an objective analysis. Barooah made a chart with dates on the bottom and his work time along the side. Running down the middle was a big black line labeled “Stopped drinking coffee.” On the left side of the line, low spikes and narrow columns. On the right side, high spikes and thick columns. The data had delivered their verdict, and coffee lost.

This doesn't mean you don't get a boost, but it might be worth checking.

Comment author: Vladimir_Golovin 02 May 2010 11:55:04AM 2 points [-]

My experience is quite similar to what is described in the first article -- no coffee leads to better concentration for me. The caffeine 'boost' I was talking about reduces my concentration but makes me more inclined to action -- I found it useful for breaking through procrastination periods. The effect of Red Bull on me is similar but more pronounced.

The effect seems to be physical, but I don't rule out placebo (and frankly, it's fine with me either way.)

Comment author: mattnewport 01 May 2010 09:20:49PM 2 points [-]

For some reason i always was convinced that the volume increase in freshly frozen water is 10x, while its actually more like 9%

Have you never made ice cubes?

Comment author: TraditionalRationali 19 May 2010 02:43:00AM 1 point [-]

Very interesting. If you find time, could you elaborate on these. I am particularly interested in hearing more on the criticism of the skeptics movement.

Comment author: MartinB 19 May 2010 08:56:41AM *  2 points [-]

I think it was mentioned here before. Skeptics do a decent job of raising the sanity waterline and thats all nice and important.

I watched all of Randis U-tube videos, Penn&Teller Bullshit, Dawkins, Derren Brown and what ever else looked interesting. But as some keep pointing out. Randi is not a scientist! He talks about stuff that should be obvious to elementary school kids. P&T get stuff wrong on their show. (I identified 2 topics so far). And they use a style of edutainment that might make you think a bit, or move in-groups. But you dont learn more about reasoning from it. I am not sure, but you might be able to compare it to any standard theist shoutout show. (To be fair, they generally do a decent job of representing oposing views. But might have learned some Tricks from a certain Michael Moore.)

All those skeptics push saner beliefs into the public and make it cool to have those in their respective subculture. As a fellow Rationalist i feel sometimes smug listening to them. But telling me stuff i already know is not too effective, while i dont have any indicators if they reach a target audience where an opinion shift is really mandated.

And: skeptics are not particularly rational. (I don't think they are even into the concept of learning more about thought processes or how science works.)

Wenn you spend your time battling idiots you might not remark, when you are wrong yourself.

Find a skeptic that will listen to your criticism of the traditional scientific method, and/or about how awesome baysianism is :-)

On a personal note: there is a distinct line of highly accidental circumstances that lead me to become involved in this particular group here. Each step involved people i learned from, and that knew more than my general surrounding. But each of those people got stuck in their personal level of thought (and field of interest respectively), and didn't follow me any further. Becoming an Atheist, and reading sceptics stuff was one of the steps. But i am very glad i didn't get stuck there. I gave a few lectures on scepticism and atheism in one of my peer groups, was highly surprised how difficult it is to bring the most basic points across, and now basically gave up on that, and concentrate on my own education.

Comment author: gwern 04 May 2010 09:17:37PM *  9 points [-]

I have a (short) essay, 'Drug heuristics' in which I take a crack at combining Bostrom's evolutionary heuristics and nootropics - both topics I consider to be quite LW-germane but underdiscussed.

I'm not sure, though, that it's worth pursuing in any greater depth and would appreciate feedback.

Comment author: cwillu 01 May 2010 09:21:31PM *  9 points [-]

Has anybody considered starting a folding@home team for lesswrong? Seems like it would be a fairly cheap way of increasing our visibility.

<30 seconds later>

After a brief 10 word discussion on #lesswrong, I've made a lesswrong team :p

Our team number is 186453; enter this into the folding@home client, and your completed work units will be credited.

Comment author: nhamann 02 May 2010 12:55:07AM 2 points [-]

Does anyone know the relative merits of folding@home and rosetta@home, which I currently run? I don't understand enough of the science involved to compare them, yet I would like to contribute to the project which is likely to be more important. I found this page, which explains the differences between the projects (and has some information about other distributed computing projects), but I'm still not sure what to think about which project I should prefer to run.

Comment author: JamesPfeiffer 05 May 2010 05:38:32AM *  8 points [-]

I noticed something recently which might be a positive aspect of akrasia, and a reason for its existence.

Background: I am generally bad at getting things done. For instance, I might put off paying a bill for a long time, which seems strange considering the whole process would take < 5 minutes.

A while back, I read about a solution: when you happen to remember a small task, if you are capable of doing it right then, then do it right then. I found this easy to follow, and quickly got a lot better at keeping up with small things.

A week or two into it, I thought of something evil to do, and following my pattern, quickly did it. Within a few minutes, I regretted it and thankfully, was able to undo it. But it scared me, and I discontinued my habit.

I'm not sure how general a conclusion I can draw from this; perhaps I am unusually prone to these mistakes. But since then I've considered akrasia as a sort of warning: "Some part of you doesn't want to do this. How about doing something else?"

Now when the part of you protesting is the non-exercising part or the ice-cream eating part, then akrasia isn't being helpful. But... it's worth listening to that feeling and seeing why you are avoiding the action.

Comment author: NancyLebovitz 05 May 2010 10:52:36AM *  8 points [-]

the most extreme example is depressed people having an increased risk of suicide if an antidepressant lifts their akrasia before it improves their mood.

Comment author: Nisan 10 May 2010 03:35:50PM 2 points [-]

Interesting. Are you sure that is going on when antidepressants have paradoxical effects?

Comment author: NancyLebovitz 10 May 2010 04:36:33PM 3 points [-]

Not absolutely certain. It's an impression I've picked up from mass media accounts, and it seems reasonable to me.

It would be good to have both more science and more personal accounts.

Thanks for asking.

Comment author: MineCanary 14 May 2010 05:17:50PM 1 point [-]

I've also read that people with bipolar disorder are more likely to commit suicide as their depression lifts.

But antidepressant effects can be very complicated. I know someone who says one med made her really really want to sleep with her feet where her head normally went. I once reacted to an antidepressant by spending three days cycling through the thoughts, "I should cut off a finger" (I explained to myself why that was a bad idea) "I should cut off a toe" (ditto) "I should cut all the flesh from my ribs" (explain myself out of it again), then back to the start.

The akrasia-lifting explanation certainly seems plausible to me (although "mood" may not be the other relevant variable--it may be worldview and plans; I've never attempted suicide, but certainly when I've self-harmed or sabotaged my own life it's often been on "autopilot", carrying out something I've been thinking about a lot, not directly related to mood--mood and beliefs are related, but I've noticed a lag between one changing and the other changing to catch up to it; someone might no longer be severely depressed but still believe that killing themself is a good course of action). Still, I would also believe an explanation that certain meds cause suicidal impulses in some people, just as they can cause other weird impulses.

Comment author: Morendil 05 May 2010 08:46:10AM 5 points [-]

Good observations.

Sometimes I procrastinate for weeks about doing something, generally non-urgent, only to have something happen that would have made the doing of it unnecessary. (For instance, I procrastinate about getting train tickets for a short trip to visit a client, and the day before the visit is due the client rings me to call it off.)

The useful notion here is that it generally pays to defer action or decision until "the last responsible moment"; it is the consequence of applying the theory of options valuation, specifically real options, to everyday decisions.

A top-level post about this would probably be relevant to the LW readership, as real options are a non-trivial instance of a procedure for decision under uncertainty. I'm not entirely sure I'm qualified to write it, but if no one else steps up I'll volunteer to do the research and write it up.

Comment author: ig0r 08 May 2010 05:07:58PM 2 points [-]

I work in finance (trading) and go through my daily life quantifying everything in terms of EV.

I would just caution in saying that, yes procrastinating provides you with some real option value as you mentioned but you need to weigh this against the probability of you exercising that option value as well as the other obvious costs of delaying the task.

Certain tasks are inherently valuable to delay as long as possible and can be identified as such beforehand. As an example, work related emails that require me to make a decison or choice I put off as long as is politely possible in case new information comes in which would influence my decision.

On the other hand, certain tasks can be identified as possessing little or no option value when weighted with the appropriate probabilities. What is the probability that delaying the payment of your cable bill will have value to you? Perhaps if you experience an emergency cash crunch. Or the off chance that your cable stops working and you decide to try to withhold payment (not that this will necessarily do you any good).

Comment author: Leafy 06 May 2010 08:05:49AM 3 points [-]

Continuing on the "last responsible moment" comment from one of the other responders - would it not be helpful to consider the putting off of a task until the last moment as an attempt to gather the largest amount of information persuant to the task without incurring any penalty?

Having poor focus and attention span I use an online todo-list for work and home life where I list every task as soon as I think of it, whether it is to be done within the next hour or year. The list soon mounts up, occassionally causing me anxiety, and I regularly have cause to carry a task over to the next day for weeks at a time - but what I have found is that a large number of tasks get removed because a change makes the task no longer necessary and a small proportion get notes added to them while they stay on the list so that the by the time the task gets actioned it has been enhanced by the extra information.

By having everything captured I can be sure no task will be lost, but by procrastinating I can ensure the highest level of efficiency in the tasks that I do eventually perform.

Thoughts?

Comment author: bogdanb 05 May 2010 12:34:36PM 3 points [-]

I suspect it’s just a figure of speech, but can you elaborate on what you meant by “evil” above?

Comment author: CronoDAS 28 May 2010 06:18:22AM *  7 points [-]

I am thinking of making a top-level post criticizing libertarianism, in spite of the current norm against discussing politics. Would you prefer that I write the post, or not write it?

Comment author: Blueberry 28 May 2010 06:51:28AM 3 points [-]

I'd love to read it, though I may well disagree with a lot of it. I'd prefer it if it were kept more abstract and philosophical, as opposed to discussing current political parties and laws and so forth: I think that would increase the light-to-heat ratio.

Comment author: cupholder 29 May 2010 07:14:15AM *  2 points [-]

Upvoted your comment for asking in the first place.

If your post was a novel explanation of some aspect of rationality, and wasn't just about landing punches on libertarianism, I'd want to see it. If it was pretty much just about criticizing libertarianism, I wouldn't.

I say this as someone very unsympathetic to libertarianism (or at least what contemporary Americans usually mean by 'libertarianism') - I'm motivated by a feeling that LW ought to be about rationality and things that touch on it directly, and I set the bar high for mind-killy topics, though I know others disagree with me about that, and that's OK. So, though I personally would want to downvote a top-level post only about libertarianism, I likely wouldn't, unless it were obnoxiously bare-faced libertarian baiting.

Comment author: ata 29 May 2010 07:30:03AM *  3 points [-]

I agree on most counts.

However, I'd also enjoy reading it if it were just a critique of libertarianism but done in an exceptionally rational way, such that if it is flawed, it will be very clear why. At minimum, I'd want it to explicitly state what terminal values or top-level goals it is assuming we want a political system to maximize, consider only the least convenient possible interpretation of libertarianism, avoid talking about libertarians too much (i.e. avoid speculating on their motives and their psychology; focus as much as possible on the policies themselves), separate it from discussion of alternatives (except insofar as is necessary to demonstrate that there is at least one system from which we can expect better outcomes than libertarianism), not appear one-sided, avoid considering it as a package deal whenever possible, etc.

Comment author: Alicorn 28 May 2010 06:18:58AM 2 points [-]

I'm interested.

Comment author: tut 28 May 2010 10:53:22AM *  3 points [-]

I will vote it down unless you say something that I have not seen before. I think that it was a good idea to not make LW a site for rehearsing political arguments, but if you have thought of something that hasn't been said before and if you can explain how you came up with it then it might be a good reasoning lesson.

Comment author: NancyLebovitz 28 May 2010 12:11:52PM 3 points [-]

I will only vote it up if there's something I haven't seen before, but will only vote it down if I think it's dreadful.

We may not be ready for it yet, but at some point we need to be able to pass the big test of addressing hard topics.

Comment author: Blueberry 28 May 2010 02:22:48PM 0 points [-]

I will vote it up to cancel the above downvote, to encourage you to make the post in case the threat of downvoting scares you off.

Comment author: cousin_it 28 May 2010 11:59:37AM *  1 point [-]

Not enough information to answer. I will upvote your post if I find it novel and convincing by rationalist lights. Try sending draft versions to other contributors that you trust and incorporate their advice before going public. I can offer my help, if being outside of American politics doesn't disqualify me from that.

Comment author: khafra 03 May 2010 06:02:01AM *  19 points [-]

Ask A Rationalist--choosing a cryonics provider:

I'm sold on the concept. We live in a world beyond the reach of god; if I want to experience anything beyond my allotted threescore and ten, I need a friendly singularity before my metabolic processes cease; or information-theoretic preservation from that cessation onward.

But when one gets down to brass tacks, the situation becomes murkier. Alcor whole body suspension is nowhere near as cheap as numbers that get thrown around in discussions on cryonics--if you want to be prepared for senescence as well as accidents, a 20 year payoff on whole life insurance and Alcor dues runs near $200/month; painful but not impossible for me.

The other primary option, Cryonics Institute, is 1/5th the price; but the future availability--even at additional cost--of timely suspension is called into question by their own site.

Alcor shares case reports, but no numbers for average time between death and deep freeze, which seems to stymie any easy comparison on effectiveness. I have little experience reading balance sheets, but both companies seem reasonably stable. What's a prospective immortal on a budget to do?

Comment author: Jack 03 May 2010 06:27:59AM 3 points [-]

Alcor whole body suspension

Why not save some money and lose what's below the neck?

Comment author: NancyLebovitz 08 May 2010 03:11:17AM 6 points [-]

Self-forgiveness limits procrastination

Wohl's team followed 134 first year undergrads through their first mid-term exams to just after their second lot of mid-terms. Before the initial exams, the students reported how much they'd procrastinated with their revision and how much they'd forgiven themselves. Next, midway between these exams and the second lot, the students reported how positive or negative they were feeling. Finally, just before the second round of mid-terms, the students once more reported how much they had procrastinated in their exam preparations.

The key finding was that students who'd forgiven themselves for their initial bout of procrastination subsequently showed less negative affect in the intermediate period between exams and were less likely to procrastinate before the second round of exams. Crucially, self-forgiveness wasn't related to performance in the first set of exams but it did predict better performance in the second set.

Comment author: SilasBarta 03 May 2010 03:36:17PM 6 points [-]

I recalled the strangest thing an AI could tell you thread, and I came up with another one in a dream. Tell me how plausible you think this one is:

Claim: "Many intelligent mammals (e.g. dogs, cats, elephants, cetaceans, and apes) act just as intelligently as feral humans, and would be capable of human-level intelligence with the right enculturation."

That is, if we did to pet mammals something analogous to what we do to feral humans when discovered, we could assimilate them; their deficiencies are the result of a) not knowing what assimilation regimen is necessary for pets/zoo mammals; and b) mammals in the wild being currently at a lower level of cultural development, but which humans at one time passed through.

Thoughts?

Comment author: wiresnips 04 May 2010 05:07:30PM *  5 points [-]

I don't know that we've ever successfully assimilated a feral human either.

Comment author: [deleted] 01 May 2010 06:25:05PM *  38 points [-]

Is anyone else here disturbed over the recent Harvard incident where Stephanie Grace's perfectly reasonable email where she merley expreses agnosticism over the posiblity that the well documented IQ differences between groups are partially genetic is worthy of harsh and inaccurate condemnation from the Harvard Law school dean?

I feel sorry for the girl since she trusted the wrong people (the email was alegedly leaked by one of her girlfriends who got into a dispute with her over a man). We need to be extra carefull to selfcensure any rationalist discusions about cows "everyone" agrees are holy. These are things I don't feel comfortable even discussing here since they have ruined many carrers and lives due to relentless persecution. Even recanting dosen't help at the end of the day, since you are a google away and people who may not even understand the argument will hate you intensly. Scary.

I mean surley everyone here agrees that the only way to discover truth is to allow all the hypothesies to stand on their own without giving them the privilige of supressing competition to a few. Why is our society so insane that this regurarly happens even concerning views that many relevant academics hold in private (or even the majority of if in certain fields if the polling is anon)?

PS Also why does the Dean equate inteligence with genetic superiority and imlicitly even worth as a person? This is a disturbing view since half by definition will always be below average. And we're all going to be terribly stupid compared to AIs in the near future, such implicit values are dangerus in the context of the time we may be living in.

Comment author: Nick_Tarleton 02 May 2010 11:55:52PM *  12 points [-]

PS Also why does the Dean equate inteligence with genetic superiority and imlicitly even worth as a person?

See Michael Vassar's discussion of this phenomenon. Also, I think that people discussing statements they see as dangerous often implicitly (and unconsciously) adopt the frames that make those statements dangerous, which they (correctly) believe many people unreflectively hold and can't easily be talked out of, and treat those frames as simple reality, in order to more simply and credibly call the statement and the person who made it dangerous and Bad.

Comment author: [deleted] 02 May 2010 12:39:50PM 24 points [-]

I'm a bit upset.

In my world, that's dinner-table conversation. If it's wrong, you argue with it. If it upsets you, you are more praiseworthy the more you control your anger. If your anti-racism is so fragile that it'll crumble if you don't shut students up -- if you think that is the best use of your efforts to help people, or to help the cause of equality -- then something has gone a little screwy in your mind.

The idea that students -- students! -- are at risk if they write about ideas in emails is damn frightening to me. I spent my childhood in a university town. This means that political correctness -- that is, not being rude on the basis of race or ethnicity -- is as deep in my bones as "please" and "thank you." I generally think it's a good thing to treat everyone with respect. But the other thing I got from my "university values" is that freedom to look for the truth is sacrosanct. And if it's tempting to shut someone up, take a few deep cleansing breaths and remember your Voltaire.

My own beef with those studies is that you cannot (to my knowledge) isolate the genetics of race from the experience of race. Every single black subject whose IQ is tested has also lived his whole life as black. And we have a history and culture that makes race matter. You can control for income and education level, because there are a variety of incomes and education levels among all races. You can control for home environment with adoption and twin studies, I guess. But you can't control for what it's like to live as a black person in a society where race matters, because all black people do. So I can't see how such a study can really ever isolate genetics alone. (But correct me if I'm missing something.)

Comment author: Jack 02 May 2010 06:56:50PM 5 points [-]

Since mixed racial background should make a difference in genes but makes only a small difference in the way our culture treats a person, if the IQ gap is the result of genetics we should see find that the those with mixed race backgrounds have higher IQs than those of mostly or exclusively African descent. This has been approximated with skin tone studies in the past and my recollection is that one study showed a slight correlation between lighter skin tone and IQ and the other study showed no correlation. There just hasn't been much research done and I doubt there will ever be much research (which is fine by me).

Comment author: NancyLebovitz 02 May 2010 07:05:49PM 4 points [-]

Afaik, skin tone, hair texture, and facial features make a large difference in how African Americans treat each other.

White people, in my experience, are apt to think of race in binary terms, but this might imply that skin tone affects how African Americans actually get treated.

Comment author: [deleted] 03 May 2010 02:39:27PM *  3 points [-]

I'm still not confident because we're not, as Nancy mentioned, completely binary about race even in the US.

What you'd really need to do is a comparative study between the US and somewhere like Brazil or Cuba, which had a different history regarding mixed race. (The US worked by the one-drop-of-blood rule; Spanish and Portuguese colonies had an elaborate caste system where more white blood meant more legal rights.) If it's mainly a cultural distinction, we ought to see a major difference between the two countries -- the light/dark gap should be larger in the former Spanish colony than it is in the US. If culture doesn't matter much, and the gap is purely genetic, it should be the same all around the world.

The other thing I would add, which is easy to lose track of, is that this is not research that should be done exclusively by whites, and especially not exclusively by whites who have an axe to grind about race. Bias can go in that direction as well, and a subject like this demands extraordinary care in controlling for it. Coming out with a bad, politically motivated IQ study could be extremely harmful.

Comment author: Jack 03 May 2010 04:53:48PM -1 points [-]

The other thing I would add, which is easy to lose track of, is that this is not research that should be done exclusively by whites, and especially not exclusively by whites who have an axe to grind about race.

Frankly, I'm not sure why the research should be done at all.

Comment author: timtyler 02 May 2010 12:17:43PM 8 points [-]

The Harvard incident is business as usual: http://timtyler.org/political_correctness/

Comment author: Rain 01 May 2010 06:54:28PM 6 points [-]

Undiscriminating skepticism strikes again: here's the thread on the very topic of genetic IQ differences.

Comment author: Jack 01 May 2010 07:12:13PM 8 points [-]

Oh good. Make it convenient for the guys running background searches.

Comment author: [deleted] 01 May 2010 07:06:06PM 2 points [-]

Thanks for the link! I'm new here and really appreciate stuff to read up on since its mostly new to me. :)

Comment author: [deleted] 01 May 2010 07:37:32PM *  21 points [-]

Here is the leaked email by Stephanie Grace if anyone is interested.

… I just hate leaving things where I feel I misstated my position.

I absolutely do not rule out the possibility that African Americans are, on average, genetically predisposed to be less intelligent. I could also obviously be convinced that by controlling for the right variables, we would see that they are, in fact, as intelligent as white people under the same circumstances. The fact is, some things are genetic. African Americans tend to have darker skin. Irish people are more likely to have red hair. (Now on to the more controversial:)

Women tend to perform less well in math due at least in part to prenatal levels of testosterone, which also account for variations in mathematics performance within genders. This suggests to me that some part of intelligence is genetic, just like identical twins raised apart tend to have very similar IQs and just like I think my babies will be geniuses and beautiful individuals whether I raise them or give them to an orphanage in Nigeria. I don’t think it is that controversial of an opinion to say I think it is at least possible that African Americans are less intelligent on a genetic level, and I didn’t mean to shy away from that opinion at dinner.

I also don’t think that there are no cultural differences or that cultural differences are not likely the most important sources of disparate test scores (statistically, the measurable ones like income do account for some raw differences). I would just like some scientific data to disprove the genetic position, and it is often hard given difficult to quantify cultural aspects. One example (courtesy of Randall Kennedy) is that some people, based on crime statistics, might think African Americans are genetically more likely to be violent, since income and other statistics cannot close the racial gap. In the slavery era, however, the stereotype was of a docile, childlike, African American, and they were, in fact, responsible for very little violence (which was why the handful of rebellions seriously shook white people up). Obviously group wide rates of violence could not fluctuate so dramatically in ten generations if the cause was genetic, and so although there are no quantifiable data currently available to “explain” away the racial discrepancy in violent crimes, it must be some nongenetic cultural shift. Of course, there are pro-genetic counterarguments, but if we assume we can control for all variables in the given time periods, the form of the argument is compelling.

In conclusion, I think it is bad science to disagree with a conclusion in your heart, and then try (unsuccessfully, so far at least) to find data that will confirm what you want to be true. Everyone wants someone to take 100 white infants and 100 African American ones and raise them in Disney utopia and prove once and for all that we are all equal on every dimension, or at least the really important ones like intelligence. I am merely not 100% convinced that this is the case.

Please don’t pull a Larry Summers on me,

A few minor fallacies but overall quite respectable and even stimulating conversation nothing any reasonable person would consider should warrant ostracism. Note the reference to "disscused over Dinner". She was betrayed by someone she socialised with.

And yes I am violating my own advice by boldening that one sentence. ;) I just wanted to drive home how close she may be to a well meaning if perhaps a bit untactfull poster on Less Wrong. Again, we need to be carefull. What society considers taboo changes over time as well, so one must get a feel for where on the scale of forbidden a subject is at any time and where the winds of change are blowing before deciding whether to discuss it online. Something inoccus could cost you your job a decade or so in the future.

Edit: For anyone wondering what a "Larry Summers" is.

Comment author: arundelo 01 May 2010 08:02:59PM 13 points [-]
Comment author: CronoDAS 01 May 2010 08:55:26PM 6 points [-]

One of the people criticizing the letter accused the letter writer of privileging the hypothesis - that it's only because of historical contingency (i.e. racism) that someone would decide to carve reality between "African-Americans" and "whites" instead of, say, "people with brown eyes" and "people with blue eyes". (She didn't use that exact phrase, but it's what she meant.)

Comment author: [deleted] 01 May 2010 10:01:20PM *  11 points [-]

Isn't nearly everything a social construct though? We can divide people based into two groups, those with university degrees and those without. People with them may tend to live longer or die earlier, they may earn more money or earn less, ect. We may also divide people into groups based on self identification, do blondes really have more fun than brunettes or do hipsters really feel superior to nonhipsters or do religious people have lower IQs than self-identified atheists ect Concepts like species, subspecies and family are also constructs that are just about as arbitrary as race.

I dosen't really matter in the end. Regardless of how we carve up reality, we can then proceed to ask questions and get answers. Suppose we decided to in 1900 take a global test to see whether blue eyed or brown eyed people have higher IQs. Lo and behold we see brown eyed people have higher IQs. But in 2050 the reverse is true. What happened? The population with brown eyes was heterogeneous and its demographics changed! However if we took skin cancer rates we would still see people with blue eyes have higher rates of skin cancer in both periods.

So why should we bother carving up reality on this racial metric and ask questions about it? For the same reason we bother to carve up reality on the family or gender metric. We base policy on it. If society was colour blind, there would be no need for this. But I hope everyone here can see that society isn't colour blind.

For example Affirmative action's ethical status (which is currently framed as a nesecary adjustment against biases and not reparations for past wrongs) depends on what the data has to about say about group differences.

If the data shows we people with blue eyes in our country have lower mean IQs when controlling for socioeconomic status and such, we shouldn't be accusing racism for their higher college drop out rates if the rates are what is to be expected when controlling for IQs. To keep this policy would mean to discriminate against competent brown eyed people. But if there are no difference well then the policy is justified unless it turns out there is another reason that has nothing to do with discrimination behind it.

I hope that you however agree that (regardless of what the truth of this particular matter is) someone should not be vilified for asking questions or proposing hypothesises regarding social constructs we have in place, regularly operate with and even make quantifiable claims about.

Comment author: Jack 01 May 2010 10:22:47PM 1 point [-]

Concepts like species, subspecies and faimily are also constructs that are just about as arbitrary as race.

This is a matter of much dispute and a lot of confusion. See here.

Comment author: kim0 02 May 2010 12:52:36PM 3 points [-]

I wondered how humans are grouped, so I got some genes from the world, and did an eigenvalue analysis, and this is what i found:

http://kim.oyhus.no/EigenGenes.html

As you can see, humans are indeed clustered in subspecies.

Comment deleted 02 May 2010 11:56:56PM [-]
Comment author: Matt_Simpson 01 May 2010 08:46:36PM 2 points [-]

A few minor fallacies

Care to point them out?

Comment author: [deleted] 01 May 2010 09:31:26PM 2 points [-]

Most escape me right now but I do recall something that bothered me... She implicity uses stereotypes of African American behvariour and how they change over time as an indicator of the actuall change in violent behaviour.

I'm sure it correlates somewhat, but considering how much stronger changes in wider society where and how much people's interests regarding what it was best to have other people belive about Black behaviour changed over time I don't think you can base an argument on this either way.

Comment author: CronoDAS 03 May 2010 09:01:28PM *  1 point [-]

Here's a bit more on the "privileging the hypothesis" bit, taken from here:

UPDATE: A lot of commenters have said that not addressing the substance of Stephanie’s email — the contention that it’s possible that black people are genetically inferior to white people — is a mistake, and weakens my point. So, why am I not addressing the idea that maybe black people are intellectually inferior, even if just to quickly debunk Stephanie’s argument? Because if I did that, the comments to this post would turn into a referendum on the genetics of intelligence, and there are always going to be a few very vocal people who have a lot invested in the falsity that black people are genetically inferior, and those people are not going to be convinced by any amount of evidence. It’s also impossible to prove, beyond any scientific doubt at all, that there is no genetic differentiation between racial and ethnic groups. That is, basically, how science works — it’s the reason that people who have some political or religious or personal investment in the idea that evolution is a crock will fall back on the “well evolution is only a theory!” line. Yes, it is “only” a theory, but it’s a theory that has a whole mountain of evidence behind it; and it’s called a theory because scientists are awfully hedgy, for good reason, about calling anything The Absolute Proven Truth. I’ll quote commenter MJ, who makes this point well:

One hears this kind of statement often from advocates of quasi-racist positions. “Oh, of course I could be convinced of perfect equality, if only someone could show me a study the proves that no differences exist!” It’s an extremely disingenuous argument and reflects a fundamental (deliberate?) misunderstanding about statistics.

No study can ever “prove” that no difference between two groups exists—a study can only fail to detect a difference of a certain size with a certain confidence level. Any experiment with enough statistical power will be able to find differences between any two groups, even two flasks of genetically identical bacteria, if you try hard enough.

My point is that asking for a study that demonstrates equality may sound reasonable, but is in fact just a rhetorical technique that can never be satisfied and serves as a shield for racist ideas.

Intelligence, too, is impossible to separate from environment and socialization, again making it impossible for anyone to say with absolute certainty that there is absolutely no biological or genetic difference at all ever between racial and ethnic groups. Intelligence is also incredibly difficult to accurately measure. But for all intents and purposes, the evidence is pretty clear that there aren’t discernible genetic differences when it comes to intelligence. But it’s always possible to make the argument that “We haven’t proven that there are no differences.” That argument tells you a lot more about the person making it than it does about any scientific fact.

I take people who argue that maybe there are race-based genetic differences that determine intelligence about as seriously as I take people who argue that maybe God did create the earth in 7 days with all humans and animals in the exact same form as we find them today. And you know, opening up a free-for-all discussion about race-based genetic difference will be about as fruitful as opening up a discussion about Creationism vs. Evolution. Discussing why Creationists are wrong and trying to convince anyone to switch “sides” in that debate (if you can even call it that) is pointless; if you really feel the need to repeat, “But evolution is just a theory and it doesn’t explain everything, so Creationism can’t be totally ruled out”, then you have some personal or religious or political or cultural investment in that idea, which won’t be toppled by evidence or rationality. Similarly, if you feel the need to repeat, “But it can’t be totally ruled out that there may be genetic differences between the races which make black people intellectually inferior,” you have some personal or political or cultural investment in that idea, which probably won’t be toppled by evidence or rationality. There is no “winning” in this debate.

And the greater danger of even opening up the debate is that, unlike creationism vs. evolution, the question of “are some people genetically inferior to others?” has been used in the service of injustice great and small. Even if we put aside the point that the genetics question has been used to justify slavery, mass sterilization, genocide, incarceration and violence — not a small point to put aside, certainly — the fact remains that the continued asking such an absurd, disproven question does harm. I can understand, for people who are not in the group that has been deemed potentially genetically inferior, that just raising the issue may not feel harmful. But for the people who are in that group? Who know the history? Who are routinely treated to questions like this under the guise of “I’m only asking the question!” but who know quite well that “only asking the question” is, itself, a way of suggesting that the answer to the question just might be yes? Who, by having to respond to the question over and over are basically being told, “You may just not be as intelligent as white people, genetically; you, as you were born, are just less”? It is harmful. It is part of a generations-long continuum of harm. It is a kind of psychological warfare that white people have waged on people of color and other less “fit” populations for centuries, which has augmented, supported and justified physical warfare, slavery, colonialism and genocide.

So no, I am not going to open up a discussion as to why Stephanie Grace’s suggestion that black people may be genetically intellectually inferior to white people is wrong. The asking of the question, and taking the question seriously, suggests that the answer just might be yes, no matter how many times the evidence points in the opposite direction. That does real harm to members of our community; it has done real harm historically to huge numbers of people, and will continue to do real harm in the future. There are many, many places on the internet where you all can talk about this to your heart’s content. I feel no obligation to provide a forum for such a pointless, hateful and harmful debate.

Comment author: Emile 04 May 2010 12:19:14PM *  9 points [-]

My "wrong-headed thinking" radar is picking up more bleeps from this than from the incriminating email:

  • "There are people with vested interests" is basically unverifiable, she's basically assuming anybody who disagrees is a fundamentally racist mutant
  • "People won't change their mind anyway, the discussion will be pointless" can be said of any controversial subject
  • The comparison to creationists can also be used to tar any opponent, there should be some version of Godwin's law for that
  • The argument that "one can always find a difference if one looks hard enough"
  • "No study can ever “prove” that no difference between two groups exists" seems to be besides the point - the question isn't whether any difference exists, but whether this specific difference exists, something that can be proved or disproved by experiment. (Well, more exactly, the topic would be what the cause of the difference is)
Comment author: steven0461 02 May 2010 09:26:13PM *  5 points [-]

I think there's something to be said for not posting opinions such that 1) LW is likely to agree with the opinion, and 2) sites perceived as agreeing with the opinion are likely to be the target of hate campaigns.

Comment author: Nick_Tarleton 02 May 2010 11:48:03PM *  5 points [-]

I share your concern. Literal hate campaigns seem unlikely to me, but such opinions probably do repulse some people, and make it considerably easier for us to lose credibility in some circles, that we might (or might not) care about. On the other hand, we pretty strongly want rationalists to be able to discuss, and if necessary slay, sacred cows, for which purpose leading by example might be really valuable.

Comment author: mattnewport 02 May 2010 10:35:02PM *  8 points [-]

This is the best exposition I have seen so far of why I believe strongly that you are very wrong.

On a Bus in Kiev

I remember very little about my childhood in the Soviet Union; I was only seven when I left. But one memory I have is being on a bus with one of my parents, and asking something about a conversation we had had at home, in which Stalin and possibly Lenin were mentioned as examples of dictators. My parent took me off the bus at the next stop, even though it wasn’t the place we were originally going.

Please read the whole thing and remember that this is where the road inevitably leads.

Comment author: Nick_Tarleton 02 May 2010 11:40:50PM *  12 points [-]

Yes, self-censorship is Prisoner's Dilemma defection, but unilaterally cooperating has costs (in terms of LW's nominal purpose) which may outweigh that (and which may in turn be outweighed by considerations having nothing to do with this particular PD).

Also, I think that's an overly dramatic choice of example, especially in conjunction with the word "inevitably".

Comment author: mattnewport 02 May 2010 11:53:42PM -1 points [-]

Also, I think that's an overly dramatic choice of example, especially in conjunction with the word "inevitably".

I don't, which is why I posted it.

In the end the Party would announce that two and two made five, and you would have to believe it. It was inevitable that they should make that claim sooner or later: the logic of their position demanded it. Not merely the validity of experience, but the very existence of external reality was tacitly denied by their philosophy. The heresy of heresies was common sense. And what was terrifying was not that they would kill you for thinking otherwise, but that they might be right. For, after all how do we know that two and two make four? Or that the force of gravity works? Or that the past is unchangeable? If both the past and the external world exist only in the mind, and if the mind itself is controllable – what then?

  • Winston Smith in George Orwell’s 1984
Comment author: JoshuaZ 05 May 2010 01:30:52AM 3 points [-]

It isn't inevitable. There's a trivial demonstration that censorship self-censorship don't form necessarily form a collective downward spiral: There are societies that at one point had much heavier censorship and now don't. That's not easily made consistent with your claim.

Censorship is bad. Self-censorship is very bad. Especially on a website devoted to improving rationality we shouldn't censor what we have to say. But the notion that small bits of self-censorship will eventually lead to believing that 2+2=5 if the Party says so is simply not called for. This is a classic example where a strong argument can be made for a claim but that claim is being inherently undermined by the use of a very weak argument for the claim instead of the strong one.

(Incidentally, generalization from fictional evidence also comes up here).

Comment author: mattnewport 05 May 2010 02:27:09AM 3 points [-]

It isn't inevitable. There's a trivial demonstration that censorship self-censorship don't form necessarily form a collective downward spiral: There are societies that at one point had much heavier censorship and now don't. That's not easily made consistent with your claim.

I am claiming that this road leads to totalitarianism. That is not the same as claiming that the road is one way with no exits and no U-turns. If I thought otherwise there would be little point in me expressing my concerns. As long as society keeps its foot on the pedal and fails to realize it is heading in the wrong direction however that is where it will end up. Inevitably.

(Incidentally, generalization from fictional evidence also comes up here).

This is not generalizing from fictional evidence. It is using a literary quote to express an idea more eloquently than I can myself. Since the book can be seen as a parable illustrating the same concerns I am emphasizing I believe it is quite appropriate to quote from it. I am not using the fictional story as proof of my claim, I am quoting it to elaborate on what it is I am claiming.

Comment author: steven0461 03 May 2010 03:15:21AM 8 points [-]

I'm sympathetic to this as a general principle, but it's not clear to me that LW doesn't have specific battles to fight that are more important than the general principle.

Comment author: jimmy 04 May 2010 11:33:58PM 3 points [-]

Perhaps there should be a "secret underground members only" section where we can discuss these things?

Comment author: RobinZ 05 May 2010 01:19:15AM 5 points [-]

Logic would suggest that such a section would be secret, if it existed. It would be simple enough to send private messages to trusted members alerting them to the existence of a private invitation-only forum on another website where such discussions could be held.

Naturally, I would say none of this if I knew of such a forum, or had any intention of creating such. And I would not appreciate any messages informing me of the existence of such a forum - if for no other reason than that I am the worst keeper of secrets I have ever known.

Comment author: mattnewport 05 May 2010 02:31:10AM 3 points [-]

The first rule of rationality club is: you do not talk about rationality club.

Comment author: jimmy 05 May 2010 02:22:20AM 3 points [-]

There could still be a lower level of 'secrecy' where it wont show up on google and you cant actually read it unless you have the minimum karma, but its existence is acknowledged.

It's not where you'd plan to take over the world, but I'd hope it'd be sufficient for talking about race/intelligence issues

Comment author: Morendil 02 May 2010 11:54:23PM -1 points [-]

I'm more directly disturbed by the bias present in your exposition: "perfectly reasonable", "merely expresses agnosticism", "well documented", "harsh and inaccurate".

Starting off a discussion with assorted applause lights and boo lights strikes me as unlikely to lead to much insight.

What would be likely to lead to useful insight? Making use of the tools LessWrong's mission is to introduce us to, such as the applications of Bayesian reasoning.

"Intelligence has a genetic component" strikes me as a causal statement. If it is, we ought to be able to represent it formally as such, tabooing the terms that give rise to cognitive muddles, until we can tell precisely what kind of data would advance our knowledge on that topic.

I've only just cracked open Pearl's Causality, and started playing with the math, so am still very much an apprentice at such things. (I have my own reasons to be fooling with that math, which are not related to the race-IQ discussion.) But it has already convinced me that probability and causality are deep topics which it's very easy to draw mistaken conclusions about if you rely solely on a layman's intuition.

For instance, "the well documented IQ differences between groups" are purely probabilistic data, which tell us very little about causal pathways generating the data, until and unless we have either controlled experiments, or further data sets which do discriminate between the competing causal models (only very grossly distinguished into "nature" and "nurture").

I don't know if the email you quoted (thanks for that, BTW, it's a treat to have access to a primary source without needing to chase it down) is racist, but it does sound very ignorant to me. It makes unwarranted inferential leaps, e.g. from "skin and hair color are definitely genetic" to "some part of intelligence is genetic", omitting the very different length of developmental chains leading from genes to pigmentation on the one hand, and intelligence on the other. It comes across as arrogant and elitist as well as ignorant when saying "I think my babies will be geniuses and beautiful individuals whether I raise them or give them to an orphanage in Nigeria".

It is not bad science to be on the lookout specifically for data that claims to be "scientific proof" of some old and demonstrably harmful prejudices, and to hold such claims to a higher standard. Just as we do hold claims of "scientific proof of ESP" to a higher standard - at least of scrutiny and replicability - than, say, claims of a correlation between apparel color and competitive performance. We have more reason to suspect ulterior motives in the former case than in the latter.

Comment author: Jack 03 May 2010 12:54:30AM *  19 points [-]

Dinnertime conversations between regular, even educated people do not contain probabilistic causal analyses. In the email Grace claimed something was a live possibility and gave some reasons why. Her argument was not of the quality we expect comments to have here at Less Wrong. And frankly, she does sound kind of annoying.

But that all strikes me as irrelevant compared to being made into a news story and attacked on all sides, by her dean, her classmates and dozens of anonymous bloggers. By the standards of normal, loose social conversation she did nothing deserving of this reaction.

I feel a chilling effect and I've only ever argued against the genetic hypothesis. Frankly, you should too since in your comment you quite clearly imply that you don't know for sure there is no genetic component. My take from the reaction to the email is that the only socially acceptable response to encountering the hypothesis is to shout "RACIST! RACIST!" at the top of your lungs. If you think we'd be spared because we're more deliberate and careful when considering the hypothesis you're kidding yourself.

Comment author: Morendil 03 May 2010 01:39:49AM 1 point [-]

By the standards of normal, loose social conversation she did nothing deserving of this reaction.

Sure. What I do find disturbing is how, knowing what she was doing (and who she was sending it to), the "friend" who leaked that email went ahead and did it anyway. That's positively Machiavellian, especially six months after the fact.

However, I do not feel a need to censure myself when discussing the race-IQ hypothesis. If intelligence has a genetic component, I want to see the evidence and understand how the evidence rules out alternatives. I would feel comfortable laying out the case for and against in an argument map, more or less as I feel comfortable laying out my current state of uncertainty regarding cryonics in the same format.

Neither do I feel a need to shout at the top of my lungs, but it does seem clear to me that racism was a strong enough factor in human civilization that it is necessary, for the time being, to systematically compensate, even at the risk of over-compensating.

"I absolutely do not rule out the possibility [of X]" can be a less than open-minded, even-handed stance, depending on what X you declare it about. (Consider "I absolutely do not rule of the possibility that I will wake up tomorrow with my left arm replaced by a blue tentacle.") Saying this and mistaking it for an "agnostic" stance is kidding oneself.

Comment author: [deleted] 03 May 2010 04:36:18PM *  4 points [-]

Since people are discussing group differences anyway. I would just like people to be a bit clearer in their phrasing.

Inteligence does have a genetic component. I hope no one argues that the cognitive difference between the average Chimpanzee and Resus monkey are result of nurture. The question is if there is any variation in the genetic component in Humans.

Studies have shown a high heritability for IQ, this dosen't nesecarily mean much of it is genetic but it does seem a strong position to take, especially considering results from twin studies. A good alternative explanation I can think of, that could be considered equivalent in explanatory power, would be differences in prenatal environment beyond those controled in previous studies (which could get sticky since such differences may also show group genetic variation ! for example the average lenght of pregnancy and risks associated with postterm complications does vary slightly between races).

The question disscused here however is whether there are any meaningfull differences between human groups regarding their genetic predispositions towards mental faculties.

We know quite a bit from genetic analysis about where people with certain markers have spread and which groups have been isolated. Therefore the real question we face is twofold:

  1. Just how really evolutionary recent is abstract thinking and other mental tricks the IQ test measures? The late advent of behavioral modernity compared vs. the early evidence of anatomically nearly modern could be considered for example. Some claim it was an evolutionary change following the well documented recent bottleneck of the Human species others say the advent of modern behaviour was a radical cultural adaptation to a abrupt environmental change or just part of a long and slow progress of rising population density and material culture complexity we haven't yet spotted. Considering how sketchy the archeological record is we can't be suprised at all if it turns out we've been wrong for decades and modern behvaiour isn't recent at all.

  2. Is the selective value of inteligence compared to other traits identical in all environments econuntered by Homo Sapiens? Remember we may already have some evidence that sometimes inteligence may not be that usefull for hominids depending on how we interpret the fossiles of Homo Floresiensis. Could this also be true of Homo Sapiens population as well?

The answers to these two questions would tell us how likley it would be to see these differences appear and how noticeable they may be in the time window current biology estimates we have for differences between populations to occur.

Note: This from Razib Khan's site (Gene Expression), I'm reposting it here so you don't need to hunt it down in my other post. http://www.gnxp.com/wp/wp-content/uploads/2010/02/PIIS096098220902065X.gr2_.lrg_.jpg

Comment author: Tyrrell_McAllister 03 May 2010 01:14:52AM *  8 points [-]

I don't know if the email you quoted (thanks for that, BTW, it's a treat to have access to a primary source without needing to chase it down) is racist, but it does sound very ignorant to me. It makes unwarranted inferential leaps, e.g. from "skin and hair color are definitely genetic" to "some part of intelligence is genetic", omitting the very different length of developmental chains leading from genes to pigmentation on the one hand, and intelligence on the other.

Let's be careful here. The letter does not assert baldly that "some part of intelligence is genetic". Rather, the letter asserts that some evidence "suggests to me that some part of intelligence is genetic".

Furthermore, that particular inferential leap does not begin with the observation that "skin and hair color are definitely genetic". Rather, the inferential leap begins with the claim that "Women tend to perform less well in math due at least in part to prenatal levels of testosterone, which also account for variations in mathematics performance within genders." Therefore, at least with regards to that particular inference, it is not fair to criticize the author for "omitting the very different length of developmental chains leading from genes to pigmentation on the one hand, and intelligence on the other."

[ETA: Of course, the inference that the author did make is itself open to criticism, just not the criticism that you made.]

I say all this as someone who considers Occam to be pretty firmly on the side of nongenetic explanations for the racial IQ gaps. But no progress in these kinds of discussions is possible without assiduous effort to avoid misrepresenting the other side's reasoning.

Comment author: ciphergoth 06 May 2010 04:27:14PM 5 points [-]

The Cognitive Bias song:

http://www.youtube.com/watch?v=3RsbmjNLQkc

Not very good, but, you know, it's a song about cognitive bias, how cool is that?

Comment author: Alexandros 20 May 2010 10:08:58AM *  4 points [-]

I remember hearing a few anecdotes about abstaining for food for a period of time (fasting) and improved brain performance. I also seem to recall some pop-sci explanation involving detoxification of the body and the like. Today something triggered interest in this topic again, but a quick Google search did not return much on the topic (fasting is drowned in religious references).

I figure this is well within LW scope, so does anyone have any knowledge or links that offer more concrete insight into (or rebuttal of) this notion?

Comment author: RichardKennaway 19 May 2010 08:08:07AM 4 points [-]

In another comment I coined (although not for the first time, it turns out) the expression "Friendly Human Intelligence". Which is simply geekspeak for how to bring up your kids right and not make druggie losers, wholesale killers, or other sorts of paperclipper. I don't recall seeing this discussed on LessWrong. Maybe most of us don't have children, and Eliezer has said somewhere that he doesn't consider himself ready to create new people, but as the saying is, if not now, when, and if not this, what?

I don't have children and don't intend to. I have two nephews and a niece, but have not had much to do with their lives, beyond sending them improving books for birthdays and Christmas. I wonder if LessWrongers, with or without children, have anything to say on how to raise children to be rational non-paperclippers?

Comment author: JanetK 22 May 2010 09:08:12AM 1 point [-]

I think that question is a conversation stopper because those who do not have children who not feel qualify and those that do have children know what a complex and tricky question it is. Personally I don't think there is a method that fits all children and all relationships with them. But... You might try activities rather than presents. 'Oh cool, uncle is gone to make a video with us and we're going to do it at the zoo.' If you get the right activity (depends on child), they will remember it and what you did and said for years. I had a uncle that I only saw a few times but he showed me how to make and throw a bomerang. He explained why it returned. I have thanked him for that day for 60 years.

Comment author: Mitchell_Porter 09 May 2010 06:47:47AM 4 points [-]

Entertainment for out-of-work Judea Pearl fans: go to your local job site and search on the word "causal", and then imagine that all those ads aren't just mis-spelling the word "casual"...

Comment author: SilasBarta 06 May 2010 10:40:58PM *  4 points [-]

Tough financial question about cryonics: I've been looking into the infinite banking idea, which actually has credible supporters, and basically involves using a mutual whole life insurance policy as a tax shelter for your earnings, allow you to accumulate dividends thereon tax free ("'cause it's to provide for the spouse and kids"), and withdraw from your premiums and borrow against yourself (and pay yourself back).

Would having one mutual whole life insurance policy keep you from having a separate policy of the kind of life insurance needed to fund a cryonic self-preservation project? Would the mutual whole life policy itself be a way to fund cryopreservation?

Comment author: Kevin 06 May 2010 08:12:21PM 4 points [-]
Comment author: [deleted] 07 May 2010 12:15:52AM *  2 points [-]

Whoooohooo! Awsomest thing in the last ten years of genetic news for me! YAAY! WHO HOO!!! /does a little dance / I want to munch on that delicious data!

Ahem.

Sorry about that.

But people 1 to 4% admixture! This is big! This gets an emotional response from me!That survived more than a thousand generations of selection, the bulk of it is probably neutral but think about how many perfectly usefull and working allels we may have today (since the Neanderthalls where close to us to start with). 600 000 or something years of speration these guys evolved sperate from us for nearly as long as the fictional Vampires in Blindsight.

It seems some of us are have a bit our ancestors picked of another species in our genes! Could this have anything to do with behavioural modernity that started off at about the same time the populations crossbred in the middle east ~100 000 years ago? Which adaptations did we pick up? Think of the possiblities!

Ok I'll stop the torrent of downvote magnet words and get back to reading about this. And then everything else my grubby little paws can get on Neanderthals, I need to brush up!

Edit: I just realized part of the reason why I got so excited is because it shows I may have a bit of exotic ancestry. Considering how much people, all else being equal, like to play up their "foreign" or "unusual" semimythical ancestors or even roots in conversation, national myths or on the census instead of the ethnicity of the majority of their ancestors this may be a more general bias, that I could of course quickly justify with a evo psych "just so" story but I'll refrain from that to search for what studies have to say about this.

Comment author: simplicio 03 May 2010 11:19:15PM 4 points [-]

I have a request. My training is in science & engineering, but I am totally ignorant of basic economics. I have come to see this as a huge blind spot. I feel my views on social issues are fairly well-reasoned, but when it comes to anything fiscal, it's all very touchy-feely at present.

Can anyone recommend intro material on economics (books, tutorials)? I ask on LW because I have no idea where to start and who to trust. If you offer a recommendation of a book pushing some particular economic "school of thought," that's fine, but I'd like to know what that school is.

Thanks!

Comment author: mattnewport 03 May 2010 11:35:36PM *  3 points [-]

Economics in One Lesson by Henry Hazlitt is a good slim introduction to the economic mindset. For a different approach focused on the application of economic thinking to everyday life The Logic of Life by Tim Harford is worth a look. Neither book covers much of the math of economics but I think that is a good thing since most of the math heavy parts of economics are the least useful and relevant.

ETA: Economics In One Lesson is a heavily free market / free trade 'classical' economic slant.

Comment author: CronoDAS 04 May 2010 09:09:56AM 2 points [-]

The book I used in my college Econ 101 class was this one.

Comment author: steven0461 04 May 2010 12:25:56AM 2 points [-]

MIT OpenCourseWare has a lot of material. I also like Bryan Caplan's lecture notes (these sometimes have a libertarian slant).

Comment author: [deleted] 02 May 2010 12:55:13AM 9 points [-]

Today, while I was attending an honors banquet, a girl in my class and her boyfriend were arguing over whether or not black was a color. When she had somewhat convinced him that it wasn't (I say somewhat because the argument was more-or-less ending and he didn't have a rebuttal), I asked "Wait, are you saying I can't paint with black paint?" She conceded that, of course black paint can be used to paint with, but that black wasn't technically a color. At which point I explained that we were likely using two different definitions of color, and that we should explain what we mean. I gave two definitions: 1] The various shade which a human eye was seeing and the brain was processing. 2] The specific wavelength of light that a human eye can pick up. The boyfriend and I were using definition 1, where as she was using definition 2. And with that cleared up, the debate ended.

Note: Both definitions aren't word for word, but somewhat close. I was simply making the distinction between the wavelength itself and the process of seeing something and placing it in a certain color category.

Comment author: cousin_it 02 May 2010 01:42:43PM *  7 points [-]

One could argue that definition 2 is Just Wrong, because it implies that purple isn't a color (purple doesn't have a wavelength, it is non-spectral).

Comment author: sketerpot 03 May 2010 07:44:45PM 2 points [-]

By her definition, the yellow color you see on a computer screen is not a color at all, since it's made up of two wavelengths of light which happen to stimulate the red and green cone cells in your retina in approximately the same way that yellow light would.

Comment author: Liron 03 May 2010 03:18:18AM 2 points [-]

This will replace Eliezer's tree falling in a forest sound as my go-to example of how an algorithm feels on the inside about wrong questions.

Comment author: cousin_it 19 May 2010 10:17:27AM *  3 points [-]

Rolf Nelson's AI deterrence doesn't work for Schellingian reasons: the Rogue AI has incentive to modify itself to not understand such threats before it first looks at the outside world. This makes you unable to threaten, because when you simulate the Rogue AI you will see its precommitment first. So the Rogue AI negates your "first mover advantage" by becoming the first mover in your simulation :-) Discuss.

Comment author: rolf_nelson 20 May 2010 02:14:56AM 1 point [-]

I agree that AI deterrence will necessarily fail if:

  1. All AI's modify themselves to ignore threats from all agents (including ones it considers irrational), and

  2. any deterrence simulation counts as a threat.

Why do you believe that both or either of these statements are true? Do you have some concrete definition of 'threat' in mind?

Comment author: Kevin 16 May 2010 11:05:00AM 3 points [-]
Comment author: arundelo 13 May 2010 04:35:49AM *  3 points [-]

Kaj_Sotala is doing a series of interviews with people in the SIAI house. The first is with Alicorn.

Edit: They are tagged as "siai interviews".

Comment author: NancyLebovitz 10 May 2010 12:22:44PM 3 points [-]

Most people's intuition is that assassination is worse than war, but simple utilitarianism suggests that war is much worse.

I have some ideas about why assassination isn't a tool for getting reliable outcomes-- leaders are sufficiently entangled in the groups that they lead that removing a leader isn't like removing a counter from a game, it's like cutting a piece out of a web which is going to rebuild itself in not quite the same shape-- but this doesn't add up to why assassination could be worse than war.

Is there any reason to think the common intuition is right?

Comment author: bogdanb 21 May 2010 02:27:56PM *  5 points [-]

TLDR: “War” is the inter-group version of “duel” (ie, lawful conflict). “Assassination” is the inter-group version of “murder” (ie, unlawful conflict).

My first “intuition about the intuition” is that it’s a historical consequence: During most history, things like freedom, and power and responsibility for enforcement of rules when conflicts (freedom vs. freedom) occur, were stratified. Conflicts between individuals in a family are resolved by the family (e.g. by the head thereof), conflicts between families (or individuals in different families) by tribal leaders or the kind. During feudalism the “scale” was formalized, but even before we had a large series of family → group → tribe → city → barony → kingdom → empire.

The key about this system is that attempts to “cross the borders” in this system, for instance punishing someone from a different group directly rather than invoking punishment from that group’s leadership is seen as an intrusion in that group’s affairs.

So assassination becomes seen as the between-group version of murder: going around the established rules of society. That’s something that is selected against in social environments (and has been discussed elsewhere).

By contrast, war is the “normal” result when there is no higher authority to recurse to, in a conflict of groups. Note that, analogously, for much of history duels were considered correct methods of conflict resolution between some individuals, as long as they respected some rules. So as long as, at least in theory, there are laws of war, war is considered a direct extension of that instinct. Assassination is seen as breaking rules, so it’s seen differently.

A few other points:

  • war is very visible, so you can expend a lot of signaling to dehumanize the adversary.
  • but assassination is supposed to be done in secret, so you can’t use propaganda as well (assassinating opposing leadership during a war is not seen as that much of a big problem; they’re all infidels/drug lords/terrorists anyway!)
  • assassination was a bit harder (even now, drones are expensive), and failed assassination attempts would lead to escalation to war often, anyway
  • assassination is oriented towards leaders, who have an interest to discourage, as much as they can, the concept. You can do that, e.g., via the meme that conflict is only honorable when it’s between armored knights on horses and the like. (For best results, add another meme which implies that observing that peasants are not allowed to own armor and horses is “dissent”.)
Comment author: JanetK 22 May 2010 09:30:54AM 2 points [-]

What an excellent analysis. I voted up. The only thing I can think of that could be added is that making a martyr can backfire.

Comment author: PhilGoetz 15 May 2010 04:13:37AM *  3 points [-]

Who thinks assassination is worse than war?

I could make an argument for it, though: If countries engaged regularly in assassination, it would never come to a conclusion, and would not reduce (and might increase) the incidence of war. Phrasing it as "which is worse" makes it sound like we can choose one or the other. This assumes that an assassination can prevent a war (and doesn't count the cases where it starts a war).

Comment author: Nick_Tarleton 10 May 2010 06:25:57PM *  3 points [-]

I've always assumed that the norm against assassination, causally speaking, exists mostly due to historical promotion by leaders who wanted to maintain a low-assassination equilibrium, now maintained largely by inertia. (Of course, it could be normatively supported by other considerations.)

It makes sense to me that people would oversimplify the effect of assassination in basically the way you describe, overestimating the indispensability of leaders. I know I've seen a study on the effects of assassination on terrorist groups, but can't find a link or remember the conclusions.

Comment author: PhilGoetz 08 May 2010 05:55:28PM 3 points [-]

If we get forums, I'd like a projects section. A person could create a project, which is a form centered around a problem to work on with other people over an extended period of time.

Comment author: NihilCredo 17 May 2010 03:57:52PM *  2 points [-]

This seems like the sort of activity Google Wave is (was?) meant for.

Comment author: AllanCrossman 04 May 2010 09:18:29PM 3 points [-]

Is Eliezer alive and well? He's not said anything here (or on Hacker News, for that matter) for a month...

Comment author: TraditionalRationali 16 May 2010 01:48:16AM *  5 points [-]

Eliezer Yudkowsky and Massimo Pigliucci just recently had a dialogue on Bloggingheads.tv. The title is The Great Singularity Debate.

After Yudkowsky at the beginning gives three different definitions of "the singularity" they discuss strong artificial intelligence and consciousness. Pigliucci is the one who quite quickly takes the discussion from intelligence to consciousness. Just before that they discuss whether simulated intelligence is actually intelligence. Yudkowsky made an argument (something like) if the AI can solve problems over a sufficiently broad range of areas and give answers then that is what we mean by intelligence, so if it manages to do that then it has intelligence. I.e., it is then not "just simulating to have intelligence" but is actually intelligent. Pigliucci however seems to want to distinguish between those and say that "well it may then just simulate intelligence, but maybe it is not actually having it". (Too difficult for me to summarize it very well, you have too look for yourself if you want it more accurately.)

There it seemed to me (but I am certainly not an expert in the field) that Yudkowsky's definition looked reasonable. It would have been interesting to have that point elaborated in more detail though.

Pigliucci's point seemed to be something like that for the only intelligence that we know so far (humans (and to lesser extent other higher animals)) intelligence comes together with consciousness. And for consciousness we know less, maybe only that the human biological brain somehow manages to have it, and therefore we of course do not know whether or not e.g. a computer simulating the brain on a different substrate will also be conscious. Yudkowsky seemed to think this very likely while Pigliucci seemed to think that very unlikely. But what I lacked in that discussion is what do we know (or reasonable conjecture) about the connection between intelligence and consciousness? Of course Pigliucci is right in that for the only intelligence we know of so so far (the human brain) intelligence and consciousness comes together. But for me (who do not know much about this subject matter) that seems not a strong argument for discussing them so closely together when it comes to artificial intelligence. Maybe someone here on Less Wrong who knows more about connection or not between intelligence and consciousness? For a naive non-expert like me intelligence seems (rather) easy to test if anything has: just test how good it is to solve general problems? While to test if anything has consciousness I would guess that a working theory of consciousness would have to be developed before a test could be designed?

This was the second recent BHTV dialogue where Pigliucci discussed singularity/transhumanism related questions. The previous I mentioned here. As mentioned there it seems to have started with a blogg-post of Pigliucci's where he criticized transhumanism. I think it interesting that Pigliucci continues his interest in the topic. I personally see it as a very positive establishing of contact between "traditional rationalist/skeptic/(cis-)humanist"-community and "LessWrong-style rationalist/trans-humanist".community. Massimo Pigliucci very much gave the impression of enjoying the discussion with Elizer Yudkowsky! I am also pleased to have noticed that recently Pigliucci's blog has now and then linked to LessWrong/ElizerYudkowsky (mostly Julias Galef if I remember correctly (too lazy to locate the exact links right now)). I would very much like to see this continue (e.g. Yudkowsky discussing with people like e.g. Paul Kurtz, Michael Shermer, Richard Dawkins, Sean Carroll, Steven Weinberg, Victor Stenger (realizing of course that they are probably too busy for it to happen)).

Previous BHTV dialogues with Elizer Yudkowsky I have seen noticed here on LessWrong but not this one (hope it is not that I have just missed that post). Therefore I posted this here, I did not find a perfect place for it, this was the least-bad I noticed. Although my post here is only partly about "Is Elizer alive and well" (he surely looked so on BHTV), I hope it is not considered too much off-topic.

Comment author: kodos96 20 May 2010 09:28:30PM *  5 points [-]

I found this diavlog entertaining, but not particularly enlightening - the two of them seemed to mostly just be talking past each other. Pigliucci kept on conflating intelligence and consciousness, continually repeating his photosynthesis analogy, which makes sense in the context of consciousness, but not intelligence, and Eliezer would respond by explaining why that doesn't make sense in the context of intelligence, and then they'd just go in circles. I wish Eliezer had been more strict about forcing him to explicitly differentiate between intelligence/consciousness. Frustrating.... but worth watching regardless.

Note that I'm not saying I agree with Pigliucci's photosynthesis analogy, even when applied to consciousness, just that it seems at least to be coherent in that context, unlike in the context of intelligence, in which case it's just silly. Personally, I don't see any reason for consciousness to be substrate-dependant, but I feel much less confident in asserting that it isn't, just because I don't really know what consciousness is, so it seems more arrogant to make any definitive pronouncement about it.

Comment author: Christian_Szegedy 23 May 2010 08:21:17AM 5 points [-]

That diavlog was a total shocker!

Pigliucci is not a nobody: he is a university professor, authored several books, holds 3 PhD's.

Still, he made an utterly confused impression on me. I don't think people must agree on everything, especially when it comes to hard questions like consciousness,but his views were so weak and incoherent that it was just too painful to watch. My head still aches... :(

Comment author: Jack 16 May 2010 01:57:58AM 3 points [-]

SIAI may have built an automaton to keep donors from panicking

Comment author: Zack_M_Davis 20 May 2010 11:15:28PM 2 points [-]

I personally see it as a very positive establishing of contact between "traditional rationalist/skeptic/(cis-)humanist"-community

I'm going to have to remember to use the word cishumanism more often.

Comment author: komponisto 21 May 2010 01:12:35AM 1 point [-]
Comment author: gwern 05 May 2010 01:47:57AM 5 points [-]

You can tell he's alive and well because he's posted several chapters in his Harry Potter fanfiction in that time; his author's notes lead me to believe that, as he stated long ago, he's letting LW drift so he has time to write his book.

Comment author: Mass_Driver 05 May 2010 02:22:07AM 4 points [-]

Anyway, he can't be hurt; "Somebody would have noticed."

Comment author: CarlShulman 04 May 2010 09:25:35PM 4 points [-]

He's writing his book.

Comment author: JaapSuter 04 May 2010 07:13:13PM *  3 points [-]

Recycling an email I wrote in a Existential Risk Reduction Career Network discussion. The topic looked at various career options, specifically with an eye towards accumulating wealth - the two major fields recognized being finance and software development.

Frank Adamek enquired as to my (flippant) vanilla latte comments, which revealed a personal blind-spot. Namely, that my default assumption for people with an interest in accumulating wealth is that they're motivated by an interest in improving the quality of their own life (e.g., expensive gadgets, etc.).

I should know -- especially in X-Risk Network context -- that wealth accumulation is not necessarily predominantly selfish, and that instead wealth can be an effective multiplier to benefit positive futures. Thanks for mentioning this Frank.

The motivation for copying this email here is two-fold.

  • One, what else can further rational critic of my own rants teach me?

  • Two, I've lurked in this community for a long time, but can't muster gusto to contribute. The quality-bar for top-level posts is well beyond my thinking and writing skills. Which is great, because it means I get to learn and grow. But there's a flip-side, which is the gap between Less Wrong discourse and that of my day-to-day interaction with friends, family, and coworkers. I don't have a solution to this, but perhaps an increase in open-thread comment mediocrity helps close the gap.

Ugh, probably not. Alas, here goes - posted as a reply to myself, because of comment-length limits.

Comment author: JaapSuter 04 May 2010 07:14:59PM *  2 points [-]

In a thread called Acturial vs. Software Engineering - what pays best?, somebody wrote:

Do any of you know how much of an option it is to start a software engineering career with a math/science but not CS background?

My response...

I encourage most people to pursue a math or science degree, rather than comp.sci., even if their long term goals are in the field of software engineering. My opinion is based on personal hindsight (having majored in computer science, I often wish my ability to absorb and apply fundamental math or hard physics was stronger) and on eleven years industry experience (where I've noticed an inverse correlation between the amount of formal comp.sci. training a person's had and his or her strength as a software engineer).

In regards to my personal hindsight; it could well be that had I studied math or physics, that I'd feel my comp.sci. expertise would need brushing up. That's probably true to some extent, but there's another factor; namely that many comp.sci. programs are a less-than-ideal blend of theoretical math (better obtained through a dedicated programs[1]) and practical engineering (most definitely useful[2], but because of its nature easily accessible in your spare time). That last point is critical; anybody who can afford university education, has access to a computer and a compiler. So why not tinker at home - you're passionate, right? Compare with programs like mechanical engineering, chemistry, and most hard physics programs - you probably don't have access to a particle accelerator or DNA extraction lab at home.

Not yet anyway... :-)

That brings me to my observation from industry experience, namely that the best programmers I've worked with often hadn't majored in comp.sci. The point of course not that a comp.sci. education makes for worse programmers. Rather, that people with the audacity and discipline to pursue hard physics or math who also have a passion for programming have a leg-up on those who are only passionate about programming.

I'm sure there's the occasional failed particle physicist applying for a hundred programming gigs without success, but that person would've been just as unskilled as a programmer had he or she majored in comp.sci.

Comment author: JaapSuter 04 May 2010 07:15:41PM 4 points [-]

Having shared my view on comp.sci. education, I do wish to throw in a recommendation for pursuing a career in software development (beyond the years of formal education). Specifically in contrast to one alternative discussed earlier in this thread, namely a career in finance.

Full disclaimer, my perspective on "jobs that involve working with money" stems mostly from how the mainstream portrays it and is likely to be extremely naive. Despite what I'm about to say, I actually have a great deal of respect for money-savy people. Considering my personal financial situation is a constant source of akrasia, I'm often envious of people who are able to wield money itself as a tool to generate more of it.

I'm realistic enough to admit that income potential is a valid factor in deciding what kind of career to pursue - like most of us, I enjoy food, shelter, and expensive gadgets. Meanwhile, I also believe nobody treats money as the only factor in choosing a caree - we all rather work in fields we're passionate about.

So really, we have a realistic assessment of various career options - all of whom promise at least a decent living. Even agreeing with comments made earlier, that programming is prole and finance has higher likelihood of fast-tracking prestige (and as a programmer, I actually must admit there's some truth to this sentiment), my gut says that your passion and interest far outweighs these observations. I mean, we're not talking about whether you'll become a high-school janitor versus United States president. If you like money and you have knack for using it for growth and your benefit, go to Wall Street. If you like computers and have a knack for using them for innovation, go to Silicon Valley. In both cases you'll be able to afford a grande sugar-free vanilla low-fat soy latte every morning - if that's your cup of tea.

Now all of this is fairly generic advice, nothing you weren't told already by your parents. My reason for chiming in on this discussion has (obviously) to do with how the above is affected by accelerating chance. That's something most parents or advisors haven't really clued into yet, and I felt it worth pointing out.

The question is, assuming the kind of consequences from accelerating change that are commonly accepted in singularity circles; what type of careers promise the most leverage in the future? In other words, what skill set guarantees you can maintain or expand the amount of control you have over the reality that surrounds and affects you?

Presumably there won't be much contention over why leverage is an important metric. Now imagine the world one, two, or three decades from now - and ask yourself; what can I offer that is of value? Value comes in many forms, we can roughly categorize these as: money, ideas (and secrets), goods, labor (and skill). Of these, money and ideas are the ones with the most long term potential. The value of manual labor will dissappear rapidly, even skilled labor (biological enhancement notwithstanding). The value of goods will diminish when life moves from its reliance on matter to information, and our ability to transform and distribute matter improves. The value of secrets is likely to exist for eternity, but those who consider this a worthy pursuit should read Snowcrash, not this email.

It's my belief the only types of leverage with future potential are money and ideas, some conditions apply.

In the case of money, the assumption is that there'll exist a legal system to assure the continuous promise of value in tender. Considering the alternative is impractical barter - or worse - all-out chaos, I believe money will stick around for a long time. In the case of ideas, the assumption is that you can turn them into reality. An idea stuck in your head is useless, so you'll need money, skill, or both to make things happen.

But wait, didn't I just say that skilled labor is a dead-end path? Yes, when speaking of the mechanical kind (i.e., the things you can do by moving your limbs around, such as playing the piano). But when it comes to ideas (and the direction our society is heading) - the kind of skill I'm referring to is of the information-theoretic kind. Future creativity will occur primarily in a universe of bits and bytes, and the more adept you are at wielding these bits and bytes, the more leverage your ideas will have.

There is one more assumption in this, namely that creative information-based skill is of a different nature than biological mechanical skill. It may be that strong AI will leapfrog well past our human ability to merge and enhance, in which case both creative skill and mechanical skill will be displaced. If that's the case, I don't expect money will be much value to humans very long either, and we'll be on a short-lived dead end path.

I'm hoping for a more optimistic future, where intellectual enhancement permits us to remain competitively creative.

So unless you have money, and use it to make more money (e.g., pursue a financial career - a valid option), I recommend people become creative experts in a digital universe. That is, study theoretical computer science (through formal math education, in your spare time, or through a career), familiarize yourself breath-first with the entire hardware and software stack that permits the digital universe (from primitive electronics to silicon to computer architectures to machine language to assembly to compilers to higher level languages to creative tools for both art and process improvement), and pick two or three comp.sci. specialties in which you become a depth-first expert. Ideally, you do this alongside a grounding in a hard physical science, to keep you in touch with the universe you currently embed (it'll be around for a while to come).

That's what you'll need to escape from the consumer end of information, and become a creative source of information - which in turn is your future leverage and source of income. Those with the ability to command, influence, and transform the growing stream bits and bytes will have the most value to offer (and be able to afford two sugar free vanilla soy lattes).

On a bit of a tangential note, this is why I advocate the introduction of a mandatory comp.sci. component from kindergarten all the way up to university - on par with traditional components like math or phys-ed. To verbalize this as: "...our society relies increasingly on computers" is to state the obvious, and the point is not that everybody should become a software developer. The critical point is to raise a generation that understands the notion of algorithmic computation well enough to believe they can (in principle) be in control of a computing device, rather than it controlling them. Computers are not magic, and one day present-day humans won't be either.

Then again, even basic schooling in math and physics fails to teach many people they can (in principle) be in control of their own life. But alas, I digress - lest this become political... :-)

Long post, little value - time to return to my computer and become a better programmer. Gotta make a living...

Two cents,

Jaap Suter - http://jaapsuter.com

[1] To be clear, I love the fundamentals of computer science. It's a great passion of mine. But I believe its place in education is by and large a sub-field of math. I suspect that'll change over time, but I'm not yet sure in which direction (math absorbing computer science, or theoretical computer science growing enough meat to justify recognition as being a field on its own.)

[2] With the additional remark that the fundamental habits of good engineering are timeless and emerge from developing your expertise in the humanities (both in one's ability to interact and cooperate with other people to achieve your goals, and the study of interactions between man, his environment, and the fruits of your labor). The tools we use along the way are fleeting - software and hardware is commonly outdated by the time you've become an expert - better to recognize the underlying patterns.

Comment author: clay 04 May 2010 01:33:33AM 3 points [-]

Would it be reasonable to request a LW open thread digest to accompany these posts? A simple bullet list of most of the topics covered would be nice.

Comment author: Thomas 01 May 2010 05:55:06PM 3 points [-]

Question: How many of you, readers and contributers here on this site, actually do work on some (nontrivial) AI project?

Or have an intention to do that in the future?

Comment author: ata 03 May 2010 06:14:38AM 2 points [-]

Yes, I have an intention to do so, because I'm convinced that it is very important to the future of humanity. I don't quite know how I'll be able to contribute yet, but I think I'm smart and creative enough that I'll be able to acquire the necessary knowledge and thinking habits (that's the part I'm working on these days) and eventually contribute something novel, if I can do all that soon enough for it to matter.

Comment author: Baughn 01 May 2010 06:52:27PM *  2 points [-]

I'm working on one as part of a game, where I'm knocking off just about every concept I've run into - goal systems, eurisko-type self-modifying code, AIXI, etc. I'll claim it's nontrivial because the game is, and I very much intend to make it unusually smart by game standards.

But that's not really true AI. It's for fun, as much as anything else. I'm not going to claim it works very well, if at all; it's just interesting to see what kind of code is involved.

(I have, nevertheless, considered FAI. There's no room to implement it, which was an interesting thing to discover in itself. Clearly my design is insufficiently advanced.)

Comment author: PhilGoetz 01 May 2010 03:40:23PM 3 points [-]

I was going thru the rationality quotes, and noticed that I always glanced at the current point score before voting. I wasn't able to not do that.

It might be useful to have a setting under which the points on a comment, and maybe also on a post, would be hidden until after you voted on it.

Comment author: Rain 01 May 2010 04:15:00PM *  4 points [-]

Marcello posted an anti-kibitzer Greasemonkey script which does that. It'd be nice to have it as core functionality of the site though, yeah.

Comment author: Morendil 01 May 2010 04:46:14PM 6 points [-]

Been working on it - it's actually committed to the LW codebase - but not released yet due to browser issues. Finding a design that avoids those is more work, not sure when I can commit to taking it on.

Comment author: mattnewport 20 May 2010 07:01:55PM 2 points [-]

This seems like a potentially significant milestone: 'Artificial life' breakthrough announced by scientists

Scientists in the US have succeeded in developing the first synthetic living cell.

The researchers constructed a bacterium's "genetic software" and transplanted it into a host cell.

The resulting microbe then looked and behaved like the species "dictated" by the synthetic DNA.

Comment author: retiredurologist 21 May 2010 03:15:39PM 2 points [-]

Given that this now opens the door for artificially designed and deployed harmful viruses, perhaps unfriendly AI falls a few notches on existentialist risk ladder.

Comment author: Kevin 14 May 2010 04:18:53AM *  2 points [-]

"Effects of nutritional supplements on aggression, rule-breaking, and psychopathology among young adult prisoners"

Likely the effects were due to the fish oil. This study was replicating similar results seen in a UK youth prison.

http://www3.interscience.wiley.com/journal/123213582/abstract?CRETRY=1&SRETRY=0

Also see this other study of the use of fish oil to present the onset of schizophrenia in a population of youth that had had one psychotic episode or similar reason to seek treatment. The p-values they got are ridiculous -- fish oil appears to be way more effective in reality than I would have expected.

http://archpsyc.ama-assn.org/cgi/content/short/67/2/146

Take your fish oil, people.

Comment author: Will_Newsome 10 May 2010 05:11:47AM 2 points [-]

You know, lots of people claim to be good cooks, or know good cooks, or have an amazing recipe for this or that. But Alicorn's cauliflower soup... it's the first food that, upon sneakily shoveling a fourth helping into my bowl, made me cackle maniacally like an insane evil sorcerer high on magic potions of incredible power, unable to keep myself from alerting three other soup-enjoying people to my glorious triumph. It's that good.

Comment author: Alicorn 10 May 2010 05:37:19AM *  3 points [-]

Awwwww :D

PS: If this endorsement of house food quality encourages anyone to apply for an SIAI fellowship, note your inspiration in the e-mail! We receive referral rewards!

Comment author: alexflint 05 May 2010 10:41:50PM 2 points [-]

Apparently it is all too easy to draw neat little circles around concepts like "science" or "math" or "rationality" and forget the awesome complexity and terrifying beauty of what is inside the circles. I certainly did. I recommend all 1400 pages of "Molecular Biology Of The Cell" (well, at least the first 600 pages) as an antidote. A more spectacularly extensive, accessible, or beautifully illustrated textbook I have never seen.

Comment author: ata 05 May 2010 08:47:44AM 2 points [-]

Is it possible to change the time zone in which LW displays dates/times?

Comment author: Jack 04 May 2010 09:50:07PM 2 points [-]
Comment author: RobinZ 04 May 2010 07:38:08PM *  2 points [-]

By the way: getting crashes on the comments page again. Prior to 1yp8 works and subsequent to 1yp8 works; I haven't found the thread with the broken comment.

Edit: It's not any of the posts after 23andme genome analysis - $99 today only in Recent Posts, I believe.

Edit 2: Recent Comments still broken for me, but ?before=t1_1yp8 is no longer showing the most recent comments to me - ?before=t1_1yqo continues where the other is leaving off.

Edit 3: Recent Comments has now recovered for me.

Comment author: eugman 03 May 2010 02:59:30AM 2 points [-]

Has anyone read The Integral Trees by Larry Niven? Something I always wonder about people supporting cryonics is why do they assume that the future will be a good place to live in? Why do they assume they will have any rights? Or do they figure that if they are revived, FAI has most likely come to pass?

Comment author: JoshuaZ 03 May 2010 06:07:32AM 3 points [-]

A dystopian society is unlikely to thaw out and revive people in cryostasis. Cryostasis revival makes sense for societies that are benevolent and have a lot of free resources. Also, be careful not to try to generalize from fictional examples. They are not evidence. That's all the more the case here because science fiction is in general a highly reactionary genre that even as it uses advance technology either warns about the perils or uses it as an excuse to hearken back to a more romantic era. For example look how many science fiction stories and universes have feudal systems of government.

Comment author: eugman 03 May 2010 02:13:24PM 4 points [-]

Now that's a reasonable argument: benevolent, resource rich societies are more likely to thaw people. Thanks.

And yes, that's true, science fiction does often look at what could go really wrong.

Comment author: Jack 04 May 2010 07:05:04AM *  2 points [-]

That's all the more the case here because science fiction is in general a highly reactionary genre that even as it uses advance technology either warns about the perils or uses it as an excuse to hearken back to a more romantic era. For example look how many science fiction stories and universes have feudal systems of government.

This is a little too broad for me to be comfortable with. There are certainly subgenres and authors who are reactionary but then there are those that are quite the opposite. Military SF and space opera (which, frankly, is just fantasy with lasers) are usually quite reactionary. Cyberpunk is cautionary but not so much about technology as about capitalism. Post-apocalyptic sf is sometimes about technology getting to great for us to handle but the jewel of the genre, A Canticle for Leibowitz is about the tragedy of a nationwide book burning. Post-cyberpunk is characterized by it's relative optimism. Hard sf varies in its political sensibilities (there seem to be a lot of libertarians) but it's almost always pro-tech for obvious reasons.

I'm having a hard time coming up with authors that fit the reactionary bill, but that might be because I read the wrong subgenres. And the libertarians are hard to classify. Michael Crichton is the obvious one that occurs to me. Larry Niven, I suppose. Card and Heinlein could be put there though both are more complicated than that. Herbert. In the other camp: Brin, Kim Stanley Robinson, LeGuin, Dick, Neil Stephenson, Gibson, Vonnegut, Orwell, Doctorow, Bradbury. Asimov and Clark probably fall in the second camp...

Am I just missing the reactionary stuff?

Comment author: Mass_Driver 04 May 2010 04:21:22AM 2 points [-]

There certainly is a large chunk of science fiction that could be accurately described as medieval fantasy moved to a superficially futuristic setting.

There is also the legitimate question of how fragile our liberal norms and economy are -- do they depend on population density? on the ratio between the reach of weapons and the reach of communications? on the dominance of a particular set of subcultures that attained to industrial hegemony through what amounts to chance and might not be repeated?

If egalitarianism is not robust to changes in the sociological environment, then there might simply be many more possible futures with feudal regimes than with capitalist or democratic regimes.

Comment author: JoshuaZ 04 May 2010 04:35:44AM 4 points [-]

Yes, but how often do they bother to explain this rise other than in some very vague way? And it isn't just feudalism. Look for example at Dune where not only is there a feudal system but the technology conveniently makes sword fighting once again a reasonable melee tactic. Additional evidence for the romantic nature is that almost invariably the stories are about people who happen to be nobles. So there's less thinking and focusing on how unpleasant feudalism is for the lower classes.

The only individual I've ever seen give a plausible set of explanations for the presence of feudal cultures is Bujold in her Vorkosigan books. But it is important to note that there there are many different governmental systems including dictatorships and anarcho-capitalist worlds and lots of other things. And she's very aware that feudalism absolutely sucks for the serfs.

I don't think that most of these writers are arriving at their societies by probabilistic extrapolation. Rather, they are just writing what they want their societies to have. (Incidentally, I suspect that many of these cultural and political norms are much more fragile than we like to think. There are likely large swaths of the space of political systems that we haven't even thought about. There might well be very stable systems that we haven't conceived of yet. Or there might be Markov chains of what systems are likely to transfer to other systems).

Comment author: Emile 04 May 2010 01:31:37PM 3 points [-]

I don't think that most of these writers are arriving at their societies by probabilistic extrapolation. Rather, they are just writing what they want their societies to have.

Those aren't the only possibilities - much more likely is the Rule of Cool. Wielding a sword is cooler than wielding a gun, and swordfights are more interesting than gunfights.

Comment author: Mass_Driver 04 May 2010 05:22:31AM 3 points [-]

I don't think that most of these writers are arriving at their societies by probabilistic extrapolation.

Granted. Some are, though. Two more counter-examples, besides Bujold:

  • Asimov's Foundation, e.g. the planet of Anacreon. Feudalism is portrayed as the result of a security dilemma and the stagnation of science, as reducing the access of ordinary people to effective medicine and nuclear power, and as producing a variety of sham nobles who deserve mockery.

  • Brave New World. Feudalism is portrayed as a logical outgrowth of an endless drive toward bureaucratic/administrative efficiency in a world where personal freedom has been subordinated to personal pleasure. Regionally-based bureaucrat-lords with concentrically overlapping territories 'earn' their authority not by protecting ordinary serfs from the danger of death but from the danger of momentary boredom or discomfort. Huxler doesn't seem overly fond of this feudalism; the question of whether a romantic would prefer this sort of system is, at worst, left as an exercise for the reader.

Comment author: NancyLebovitz 03 May 2010 06:03:35AM 3 points [-]

Science fiction has a bias towards things going wrong.

In the particular case of cryonics, if there's a dystopian future where the majority of people have few or no rights, it's a disaster all around, but as ata says, you can presumably commit suicide. There's a chance that even that will be unfeasible-- for example if brains are used, while conscious, for their processing power. This doesn't seem likely, but I don't know how to evaluate it in detail.

The other case-- people in general have rights, but thawed people, or thawed people from before a certain point in time, do not-- requires that thawed people do not have a constituency. This doesn't seem terribly likely, though as I recall, Niven has it that it takes a very long time for thawing to be developed.

Normally, I would expect for there to be commercial and legal pressures for thawed people to be treated decently. (I've never seen an sf story in which thawed people are a political football, but it's an interesting premise.)

I think the trend is towards better futures (including richer, with less reason to enslave people), but there's no guarantee. I think it's much more likely that frozen people won't be revived than that they'll be revived into a bad situation.

Comment author: ata 03 May 2010 06:26:15AM 4 points [-]

Science fiction has a bias towards things going wrong.

All fiction has a bias towards things going wrong. Need some kind of conflict.

(Reality also has a bias towards things going wrong, but if Fun Theory is correct, then unlike with fiction, we can change that condition without reducing the demand for reality.)

Comment author: NancyLebovitz 03 May 2010 06:43:53AM 3 points [-]

Science fiction has a stronger bias towards things going wrong on a grand scale than most fiction does.

Comment author: Lightwave 02 May 2010 11:43:33PM 2 points [-]

Here's my question to everyone:

What do you think are the benefits of reading fiction (all kinds, not just science fiction) apart from the entertainment value? Whatever you're learning about the real world from fiction, wouldn't it be more effective to read a textbook instead or something? Is fiction mostly about entertainment rather than learning and improvement? Any thoughts?

Comment author: [deleted] 03 May 2010 02:32:11PM 5 points [-]

We are wired for individual rather than general insights. Stories are much more effective at communicating certain things than treatises are. I would never have believed, in theory, that a man who enjoyed killing could be worthy of respect; only a story could convince me. To use Robin Hanson's terminology, narrative can bring near mode and far mode together.

Why not true stories? I think there you get into Aristotle and why versimilitude can be more effective than mere reality. True stories are good too, but life is disorderly and not necessarily narrative. It's a truism of writing workshops and creative writing classes that whenever you see a particularly unrealistic event in a story, the author will protest "But that really happened!" It doesn't matter; it's still unrealistic. Narrative is, I think, a particular kind of brain function that humans are good at, and it's a painting, not a photograph. To tap into our ability to understand each other through narrative, we usually need to fictionalize the world, apply some masks and filters.

Comment author: Morendil 03 May 2010 12:04:42AM 5 points [-]

A possible benefit of fiction is that it leads you to experience emotions vicariously that it would be much more expensive to experience for real, yet the vicarious experience is realistic enough that it serves as useful practice, a way of "taming" the emotions. Textbooks don't convey emotions.

I seem to recall this argument from a review of Cloverfield, or possibly the director's commentary. Broadcast images such as from the 9/11 aftermath generated lots of anxiety, and seeing similar images - the amateurish, jerky camcorder type - reframed in a fictional setting which is "obviously" over the top helps you, the audience, come to terms with the reality.

Comment author: Nisan 03 May 2010 06:50:11AM *  4 points [-]

It was not until I read Three Worlds Collide that I began to embrace moral consequentialism. I would not have found an essay or real-life case study nearly as convincing.

ETA: I didn't change my mind just because I liked the story. The story made me realize that in a particular situation, I would be a moral consequentialist.

Comment author: Academian 03 May 2010 07:09:14AM *  6 points [-]

My take on works of fiction, especially written fiction, is that they're thought experiments for your emotional intelligence. The best ones are the ones written for that purpose, since I think they tend to better optimize the net value of entertainment and personal growth.

Morality in particular usually stems from some sort of emotional intelligence, like empathy, so it makes sense to me that written fiction could help especially with that.

Comment author: Jack 02 May 2010 11:53:32PM 2 points [-]

Fiction is good for teasing out possibilities and counterfactuals, experimenting with different attitudes toward the world (as opposed to learning facts about the world), and learning to be cool.

Comment author: NancyLebovitz 03 May 2010 12:28:36AM *  3 points [-]

On the other hand (and I speak as a person who really likes fiction), it's possible that you learn more about the human range by reading letters and diaries-- whatever is true in fiction may be distorted to make good stories.

Comment author: MartinB 01 May 2010 01:02:29PM 2 points [-]

Question: How do you apply the rationalist ideas you learned on lesswrong in your own (professional and/or private) life?

Comment author: Bo102010 01 May 2010 02:52:08PM 4 points [-]

I remind myself of Conservation of Expected Evidence most days I'm at work.

I'm an engineer, and it helps remind me that a data point can either support a hypothesis or that hypothesis's opposite, but not both at once. This is especially useful for explaining things to non-technical people.

Comment author: ShardPhoenix 01 May 2010 03:40:01PM 2 points [-]

I think learning more about rationalization, akrasia, and so forth, has made it easier for me to keep regularly going to the gym, by noticing when I'm just making excuses for being lazy, etc.

Comment author: Matt_Stein 03 May 2010 04:32:54AM 5 points [-]

So, I'm somewhat new to this whole rationality/Bayesianism/(nice label that would describe what we do here on LessWrong). Are there any podcasts or good audiobooks that you'd recommend on the subjects of LessWrong? I have a large amount of time at work that I can listen to audio, but I'm not able to read during this time. Does anyone have any suggestions for essential listening/reading on subjects similar to the ones covered here?

Comment author: [deleted] 03 May 2010 07:07:04PM *  5 points [-]

I know you said you don't have a ton of time to read but Gary Drescher's Good and Real has been called Less Wrong in book form on occasion. If nothing else, I found it an enjoyable read that gives a good start to getting into the mindset people have in this community.

Comment author: gwern 03 May 2010 04:55:05PM 3 points [-]

Some people have curious ideas about what LW is; from http://www.fanfiction.net/r/5782108/18/1/ :

"HO-ley **! That was awesome! You might also be interested to know that my brother, my father and I all had a wonderful evening reading that wikipedia blog on rationality that you are named for. Thank you for this, most dearly and truly."

Comment author: thomblake 03 May 2010 05:07:39PM 2 points [-]

I'm not sure I even know how to parse "wikipedia blog on rationality". But at least in some sense, we apparently are Wikipedia. Congrats.

Comment author: CronoDAS 01 May 2010 09:54:20PM 3 points [-]
Comment author: PhilGoetz 15 May 2010 04:06:27AM 2 points [-]

Today the Pope finally admitted there has been a problem with child sex abuse by Catholic priests. He blamed it on sin.

What a great answer! It covers any imaginable situation. Sin could be the greatest tool for bad managers everywhere since Total Quality Management.

"Sir, your company, British Petroleum, is responsible for the biggest environmental disaster in America this decade. How did this happen, and what is being done to prevent it happening again?"

"Senator, I've made a thorough investigation, and I'm afraid there has been sin in the ranks of British Petroleum. BP has a deep need to re-learn penance, to accept purification, to learn on one hand forgiveness but also the need for justice."

"Thank you, Mr. Hayward. I'm glad you're on top of the situation."

I wonder if I can use this at work.

Comment author: Morendil 15 May 2010 08:04:06AM *  1 point [-]

Sin could be the greatest tool for bad managers everywhere since Total Quality Management.

That sounds like the kind of remark that goes out of its way to offend several categories of people at once. :)

But in that category the gold standard remains Evelyn Waugh's “now that they no longer defrock priests for sexual perversities, one can no longer get any decent proofreading.”

Comment author: Kevin 08 May 2010 11:24:32PM 2 points [-]
Comment author: mattnewport 08 May 2010 11:32:08PM *  2 points [-]

Related, Obama authorizes assassination of US citizen. I'm amazed how little anybody seems to care.

Comment author: Jack 10 May 2010 09:58:48PM 3 points [-]

Something problematic: if you're a cosmopolitan, as I assume most people here are, can you consistently object to assassinations of citizens if you don't object to assassinations of non-citizens?

Comment author: mattnewport 10 May 2010 10:49:52PM 1 point [-]

Probably not, though you might be able to make a case that if a particular non-citizen is a significant perceived threat but there is no legal mechanism for prosecuting them then different rules apply. Most people are not cosmopolitan however and so I am more surprised at the lack of outrage over ordering the assassination of a US citizen than by the lack of outrage over the assassination of non-US citizens.

Comment author: PhilGoetz 15 May 2010 04:21:10AM *  2 points [-]

I care, and approve, provided that Al-Awlaki can forestall it if he chooses by coming to the US to face charges.

I don't believe in treating everything with the slippery-slope argument. That way lies the madness I saw at the patent office, where every decision had to be made following precedent and procedure with syntactic regularity, without any contaminating element of human judgement.

Comment author: steven0461 09 May 2010 03:08:16AM 1 point [-]

Is there a consensus on whether or not it's OK to discuss not-specifically-rationality-related politics on LW?

Comment author: ata 13 May 2010 08:47:02AM 1 point [-]

Rationality comix!

Hover over the red button at the bottom (to the left of the RSS button and social bookmarking links) for a bonus panel.

Edit: "Whoever did the duplication" would be a better answer than "The guy who came first", admittedly. The duplicate and original would both believe themselves to be the original, or, if they are a rationalist, would probably withhold judgment.

Comment author: RobinZ 13 May 2010 12:26:22PM 3 points [-]

Speaking as an engineer, I'd think he wasn't talking about subjective aspects: "The guy who came first" is the one which was copied (perfectly) to make the clone, and therefore existed before the clone existed.

Comment author: vinayak 01 May 2010 11:23:11AM 1 point [-]

I want to understand Bayesian reasoning in detail, in the sense that, I want to take up a statement that is relevant to our daily life and then try to find exactly how much should I believe in it based on the the beliefs that I already have. I think this might be a good exercise for the LW community? If yes, then let's take up a statement, for example, "The whole world is going to be nuked before 2020." And now, based on whatever you know right now, you should form some percentage of belief in this statement. Can someone please show me exactly how to do that?

Comment author: Jack 01 May 2010 06:30:42PM *  4 points [-]

Well to begin with we need a prior. You can choose one of two wagers. In the first, 1,000,000 blue marble and one red marble are put in a bag. You get to remove one marble, if it is the red one you win a million dollars. Blue you get nothing. In the second wager, you win a million dollars if a a nuclear weapon is detonated under non-testing and non-accidental conditions before 2020. Otherwise, nothing. In both cases you don't get the money until January 1st 2021. Which wager do you prefer?

If you prefer the nuke bet, repeat with 100,000 blue marbles, if you prefer the marbles try 100,000,000. Repeat until you get wagers that are approximately equal in their estimated value to you.

Edit: Commenters other than vinayak should do this too so that he has someone to exchange information with. I think I stop at maybe 200:1 against nuking.

Comment author: Morendil 02 May 2010 09:17:00AM *  3 points [-]

The interesting question isn't so much "how do I convert a degree of belief into a number", but "how do I reconcile my degrees of beliefs in various propositions so that they are more consistent and make me less vulnerable to Dutch books".

One way to do that is to formalize what you take that statement to mean, so that its relationships to "other beliefs" becomes clearer. It's what, in the example you suggest, the Doomsday clock scientists have done. So you can look at whatever data has been used by the Doomsday Clock people, and if you have reason to believe they got the data wrong (say, about international agreements), then your estimate would have to be different from theirs. Or you could figure out they forgot to include some evidence that is relevant (say, about peak uranium), or that they included evidence you disagree is relevant. In each of these cases Bayes' theorem would probably tell you at the very least in what direction you should update your degree of belief, if not the exact amount.

Or, finally, you could disagree with them about the structural relationships between bits of evidence. That case pretty much amounts to making up your own causal model of the situation. As other commenters have noted it's fantastically hard to apply Bayes rigorously to even a moderately sophisticated causal model, especially one that involves such an intricately interconnected system as human society. But you can always simplify, and end up with something you know is strictly wrong, but has enough correspondence with reality to be less wrong than a more naive model.

In practice, it's worth noting that only very seldom does science tackle a statement like this one head-on; as a reductionist approach science generally tries to explicate causal relationships in much smaller portions of the whole situation, treating each such portion as a "black box" module, and hoping that once this module's workings are formalized it can be plugged back into a more general model without threatening the overall model's validity too much.

The word "complex" is appropriate to refer precisely to situations where this approach fails, IMHO.

Comment author: Matt_Duing 27 May 2010 03:25:01AM 1 point [-]

Has anyone read "Games and Decisions: Introduction and Critical Survey" by R. Duncan Luce and Howard Raiffa? Any thoughts on its quality?

Comment author: eugman 25 May 2010 06:19:52PM 1 point [-]

I have a cognitive problem and I figured someone might be able to help with it.

I think I might have trouble filtering stimuli, or something similar. A dog barking, an ear ache, loud people, or a really long day can break me down. I start to have difficulty focusing. I can't hold complex concepts in my head. I'll often start a task, and quit in the middle because it feels too difficult and try to switch to something else, ultimately getting nothing done. I'll have difficulty deciding what to work on. I'll start to panic or get intimidated. It's really an issue.

I've found two things that help:

Music is good at filtering out noise and helping me focus. However, sometimes i can't listen to it or it is not enough.

The other thing is to make a extremely granular tasklist and then follow it without question. The tasks have to be really small and seem manageable.

Anyone have any suggestions? I'm not neurotypical in the broader sense, but I don't believe I fall on the autism spectrum.

Comment author: Alicorn 25 May 2010 06:26:29PM 1 point [-]

I have similar sensory issues on occasion and believe them to be a component of my autism, but if you don't have other features of an ASD then this could just be a sensory integration disorder. When it's an auditory processing issue, I find that listening to loud techno or other music with a strong beat helps more than other types of music, and ear-covering headphones help filter out other noise. I'm more often assaulted by textures, which I have to deal with by avoiding contact with the item(s).

As for the long day, that sounds like a matter of running out of (metaphorical) spoons. Paying attention to what activities drain or replenish said spoons, and choosing spoon-neutral or spoon-positive activities whenever they're viable options, is the way to manage this.

Comment author: NancyLebovitz 15 May 2010 11:59:25PM *  1 point [-]

Crinimal profiling, good and bad

Article discusses the shift from impressive-looking guesswork to use of statistics. Also has an egregious example of the guesswork approach privileging the hypothesis.

Comment author: JoshuaZ 13 May 2010 04:08:29AM *  1 point [-]

There an article in this month's Nature examining the statistical evidence for universal common descent. This is the first time someone has taken the massive amounts of genetic data and applied a Bayesian analysis to determine whether the existence of a universal common ancestor is the best model. Most of what we generally think of as evidence for evolution and shared ancestry is evidence for shared ancestry of large collections, such as mammals or birds, or for smaller groups. Some of the evidence is for common ancestry for a phylum. There is prior evidence for their shared ancestry based on primitive fossils and on the shared genetic code and extreme similarity of genomes across very different species. This is the first paper to make that last argument mathematically rigorous. When taken in this fashion, the paper more or less concludes that a Bayesian analysis using just the genetic and phylogenetic known data puts the universal common ancestor model as overwhelmingly more likely than other models. (The article is behind a paywall so until I get back to the university tomorrow I won't be able to comment on this in any substantial detail but this looks pretty cool and a good example how careful Bayesianism can help make something more precise).

Comment author: JoshuaZ 13 May 2010 08:13:48PM 2 points [-]

Ok. Reading the paper now. Some aspects are bit technical and so I don't follow all of the arguments or genetic claims other than at a broad level. However, the money quote is "Therefore, UCA is at least 10^2,860 times more probable than the closest competing hypothesis." (I've replaced the superscript with a ^ becaause I don't know how to format superscripts). 10^2860 is a very big number.

Comment author: Psy-Kosh 13 May 2010 08:32:27PM 1 point [-]

What were they using for prior probabilities for the various candidate hypotheses? Uniform? Some form of complexity weighting? Other?

Comment author: JoshuaZ 23 May 2010 06:04:05PM *  1 point [-]

They have hypotheses concerning whether Eukarya, Archaea and Bacteria share a common ancestor or not, or possibly in pairs. All hypotheses were given equal prior likelyhood.

Comment author: Seth_Goldin 10 May 2010 03:18:45PM 1 point [-]

Cool paper: When Did Bayesian Inference Become “Bayesian”?

http://ba.stat.cmu.edu/journal/2006/vol01/issue01/fienberg.pdf