Open Thread June 2010, Part 2

7 Post author: komponisto 07 June 2010 08:37AM

The title says it all.

Comments (534)

Sort By: Leading
Comment author: DanielVarga 07 June 2010 11:38:32PM *  18 points [-]

Less Wrong Rationality Quotes since April 2009, sorted by points.

Pre-alpha, one hour of work. I plan to improve it.

EDIT: Here is the source code. 80 lines of python. It makes raw text output, links and formatting are lost. It would be quite trivial to do nice and spiffy html output.

EDIT2: I can do html output now. It is nice and spiffy, but it has some CSS bug. After the fifth quote it falls apart. This is my first time with CSS, and I hope it is also the last. Could somebody help me with this? Thanks.

EDIT3: Bug resolved. I wrote another top-level comment. about the final version, because my access logs suggested that the EDITs have reached only a very few people. Of course, an alternative explanation is that everybody who would have been interested in the html version already checked out the txt version. We will soon find out which explanation is the correct one.

Comment author: Alicorn 07 June 2010 11:44:13PM 4 points [-]

Not having to side scroll would be spiffy.

Comment author: CronoDAS 08 June 2010 02:35:42AM 4 points [-]

If you're using Firefox, there's an add-on for that.

Comment author: Blueberry 08 June 2010 04:19:42AM 3 points [-]

Or, if you're lazy like me, you can select 'Page Source' under the View menu and then select the 'Wrap Long Lines' option.

Comment author: JoshuaZ 08 June 2010 12:10:26AM 2 points [-]

It might make more sense to put this on the Wiki. Two notes: First, some of the quotes have remarks contained in the posts which you have not edited out. I don't know if you intend to keep those. Second, some of the quotes are comments from quote threads that aren't actually quotes. 14 SilasBarta is one example. (And is just me or does that citation form read like a citation from a religious text ?)

Comment author: Vladimir_Nesov 08 June 2010 12:24:29AM 4 points [-]

On the wiki, this text will be dead, because nobody will be adding new items there by hand.

Comment author: DanielVarga 08 June 2010 02:11:00PM *  2 points [-]

I agreed with you, I even started to write a reply to JoshuaZ about the intricacies of human-machine cooperation in text-processing pipelines. But then I realized that it is not necessarily a problem if the text is dead. A Rationality Quotes, Best of 2010 Edition could be nice.

Comment author: Vladimir_Nesov 08 June 2010 02:17:15PM *  1 point [-]

Agreed. Best of 2009 can be compiled now and frozen, best of 2010 end of the year and so on. It'd also be useful to publish the source code of whatever script was used to generate the rating on the wiki, as a subpage.

Comment author: NancyLebovitz 07 June 2010 11:56:53PM 1 point [-]

Very cool idea.

It would be nice if links were preserved.

Comment author: DanielVarga 12 June 2010 05:13:54PM *  16 points [-]

Less Wrong Rationality Quotes since April 2009, sorted by points.

This version copies the visual style and preserves the formatting of the original comments.

Here is the source code.

I already wrote a top-level comment about the original raw text version of this, but my access logs suggested that EDITs of older comments only reach a very few people. See that comment for a bit more detail.

Comment author: XiXiDu 22 October 2010 11:26:54AM 1 point [-]

This is great, even more so as you made it open source. I added it to References & Resources for LessWrong.

Comment author: Kevin 13 June 2010 02:36:50AM *  1 point [-]

You should make a short top-level post about this so more people see this

Comment author: cupholder 12 June 2010 06:13:48PM 1 point [-]

I'd vote you up again for handing out your source code as well as the quote list, but I can't, so an encouraging reply will have to do...

Comment author: Eneasz 09 June 2010 08:41:05PM 10 points [-]

You Are Not So Smart is a great little blog that covers many of the same topics as LessWrong, but in a much more bite-sized format and with less depth. It probably won't offer much to regular/long-time LW readers, but it's a great resource to give to friends/family who don't have the time/energy demanded by LW.

Comment author: hegemonicon 10 June 2010 12:57:30PM 1 point [-]

It is a good blog, and it has a slightly wider topic spread than LW, so even if you're familiar with most of the standard failures of judgment there'll be a few new things worth reading. (I found the "introducing fines can actually increase a behavior" post particularly good, as I wasn't aware of that effect.)

Comment author: cousin_it 08 June 2010 06:24:24AM *  9 points [-]

As an old quote from DanielLC says, consequentialism is "the belief that doing the right thing makes the world a better place". I now present some finger exercises on the topic:

  1. Is it okay to cheat on your spouse as long as (s)he never knows?

  2. If you have already cheated and managed to conceal it perfectly, is it right to stay silent?

  3. If your spouse asks you to give a solemn promise to never cheat, and you know you will cheat perfectly discreetly, is it right to give the promise to make them happy?

  4. If your wife loves you, but you only stay in the marriage because of the child, is it right to assure the wife you still love her?

  5. If your husband loves you, but doesn't know the child isn't his, is it right to stay silent?

  6. The people from #4 and #5 are actually married to each other. They seem to be caught in an uncomfortable equilibrium of lies. Would they have been better off as deontologists?

While you're thinking about these puzzles, be extra careful to not write the bottom line in advance and shoehorn the "right" conclusion into a consequentialist frame. For example, eliminating lies doesn't "make the world a better place" unless it actually makes people happier; claiming so is just concealed deontologism.

Comment author: Matt_Simpson 08 June 2010 04:10:42PM 5 points [-]

For example, eliminating lies doesn't "make the world a better place" unless it actually makes people happier; claiming so is just concealed deontologism.

Just picking nits. Consequentialism =/= maximizing happiness. (The latter is a case of the former). So one could be a consequentialist and place a high value and not lying. In fact, the answers to all of your questions depend on the values one holds.

Or what Nesov said below.

Comment author: Vladimir_Nesov 08 June 2010 08:06:25AM *  5 points [-]

For example, eliminating lies doesn't "make the world a better place" unless it actually makes people happier; claiming so is just concealed deontologism.

I disagree. Not lying or not being lied to might well be a terminal value, why not? You that lies or doesn't lie is part of the world. A person may dislike being lied to, value the world where such lying occurs less, irrespective of whether they know of said lying. (Correspondingly, the world becomes a better place even if you eliminate some lying without anyone knowing about that, so nobody becomes happier in the sense of actually experiencing different emotions, assuming nothing else that matters changes as well.)

Of course, if you can only eliminate a specific case of lying by on the net making the outcome even worse for other reasons, it shouldn't be done (and some of your examples may qualify for that).

Comment author: prase 08 June 2010 10:04:35AM *  4 points [-]

A person may dislike being lied to, value the world where such lying occurs less, irrespective of whether they know of said lying.

In my opinion, this is a lawyer's attempt to masquerade deontologism as consequentialism. You can, of course, reformulate the deontologist rule "never lie" as a consequentialist "I assign an extremely high disutility to situations where I lie". In the same way you can put consequentialist preferences as a deontoligst rule "at any case, do whatever maximises your utility". But doing that, the point of the distinction between the two ethical systems is lost.

Comment author: cupholder 08 June 2010 02:06:09PM 2 points [-]

But doing that, the point of the distinction between the two ethical systems is lost.

If so, maybe we want that.

Comment author: NancyLebovitz 08 June 2010 09:42:15AM 3 points [-]

I suggest that eliminating lying would only be an improvement if people have reasonable expectations of each other.

Comment author: RobinZ 08 June 2010 03:14:49PM 2 points [-]

Less directly, a person may value a world where beliefs were more accurate - in such a world, both lying and bullshit would be negatives.

Comment author: AlephNeil 09 June 2010 01:22:15PM 3 points [-]

Is it okay to cheat on your spouse as long as (s)he never knows?

Is this actually possible? Imagine that 10% of people cheat on their spouses when faced with a situation 'similar' to yours. Then the spouses can 'put themselves in your place' and think "Gee, there's about a 10% chance that I'd now be cheating on myself. I wonder if this means my husband/wife is cheating on me?"

So if you are inclined to cheat then spouses are inclined to be suspicious. Even if the suspicion doesn't correlate with the cheating, the net effect is to drive utility down.

I think similar reasoning can be applied to the other cases.

(Of course, this is a very "UDT-style" way of thinking -- but then UDT does remind me of Kant's categorical imperative, and of course Kant is the arch-deontologist.)

Comment author: cousin_it 09 June 2010 01:43:20PM *  1 point [-]

Your reasoning goes above and beyond UDT: it says you must always cooperate in the Prisoner's Dilemma to avoid "driving net utility down". I'm pretty sure you made a mistake somewhere.

Comment author: thomblake 08 June 2010 09:15:08PM 3 points [-]

As an old quote from DanielLC says, consequentialism is "the belief that doing the right thing makes the world a better place". I now present some finger exercises on the topic:

It seems like it would be more aptly defined as "the belief that making the world a better place constitutes doing the right thing". Non-consequentialists can certainly believe that doing the right thing makes the world a better place, especially if they don't care whether it does.

Comment author: taw 11 June 2010 12:39:29PM 2 points [-]

It is a common failure of moral analysis (invented by deontologists undoubtedly) that they assume idealized moral situation. Proper consequentialism deals with the real world, not this fantasy.

  • #1/#2/#3 - "never knows" fails far too often, so you need to include a very large chance of failure in your analysis.
  • #4 - it's pretty safe to make stuff like that up
  • #5 - in the past, undoubtedly yes; in the future this will be nearly certain to leak with everyone undergoing routine genetic testing for medical purposes, so no. (future is relevant because situation will last decades)
  • #6 - consequentialism assumes probabilistic analysis (% that child is not yours, % chance that husband is making stuff up) - and you weight costs and benefits of different situations proportionally to their likelihood. Here they are in unlikely situation that consequentialism doesn't weight highly. They might be better off with some other value system, but only at cost of being worse off in more likely situations.
Comment author: ciphergoth 11 June 2010 03:44:37PM *  1 point [-]

#4 - it's pretty safe to make stuff like that up

You seem to make the error here that you rightly criticize. Your feelings have involuntary, detectable consequences; lying about them can have a real personal cost.

Comment author: RobinZ 08 June 2010 02:41:38PM 2 points [-]

A quick Internet search turns up very little causal data on the relationship between cheating and happiness, so for purposes of this analysis I will employ the following assumptions:

a. Successful secret cheating has a small eudaemonic benefit for the cheater.
b. Successful secret lying in a relationship has a small eudaemonic cost for the liar.
c. Marital and familial relationships have a moderate eudaemonic benefits for both parties.
d. Undermining revelations in a relationship have a moderate (specifically, severe in intensity but transient in duration) eudaemonic cost for all parties involved.
e. Relationships transmit a fraction of eudaemonic effects between partners.

Under these assumptions, the naive consequentialist solution* is as follows:

  1. Cheating is a risky activity, and should be avoided if eudaemonic supplies are short.
  2. This answer depends on precise relationships between eudaemonic values that are not well established at this time.
  3. Given the conditions, lying seems appropriate.
  4. Yes.
  5. Yes.
  6. The husband may be better off. The wife more likely would not be. The child would certainly not be.

Are there any evident flaws in my analysis on the level it was performed?

* The naive consequentialist solution only accounts for direct effects of the actions of a single individual in a single situation, rather than the general effects of widespread adoption of a strategy in many situations - like other spherical cows, this causes a lot of problematic answers, like two-boxing.

Comment author: cousin_it 08 June 2010 02:55:56PM *  2 points [-]

Ouch. In #5 I intended that the wife would lie to avoid breaking her husband's heart, not for some material benefit. So if she knew the husband didn't love her, she'd tell the truth. The fact that you automatically parsed the situation differently is... disturbing, but quite sensible by consequentialist lights, I suppose :-)

I don't understand your answer in #2. If lying incurs a small cost on you and a fraction of it on the partner, and confessing incurs a moderate cost on both, why are you uncertain?

No other visible flaws. Nice to see you bite the bullet in #3.

ETA: double ouch! In #1 you imply that happier couples should cheat more! Great stuff, I can't wait till other people reply to the questionnaire.

Comment author: Nisan 08 June 2010 04:50:52PM 1 point [-]

It's okay to deceive people if they're not actually harmed and you're sure they'll never find out. In practice, it's often too risky.

1-3: This is all okay, but nevertheless, I wouldn't do these things. The reason is that for me, a necessary ingredient for being happily married is an alief that my spouse is honest with me. It would be impossible for me to maintain this alief if I lied.

4-5: The child's welfare is more important than my happiness, so even I would lie if it was likely to benefit the child.

6: Let's assume the least convenient possible world, where everyone is better off if they tell the truth. Then in this particular case, they would be better off as deontologists. But they have no way of knowing this. This is not problematic for consequentialism any more than a version of the Trolley Problem in which the fat man is secretly a skinny man in disguise and pushing him will lead to more people dying.

Comment author: cousin_it 08 June 2010 05:03:04PM *  2 points [-]

1-3: It seems you're using an irrational rule for updating your beliefs about your spouse. If we fixed this minor shortcoming, would you lie?

6: Why not problematic? Unlike your Trolley Problem example, in my example the lie is caused by consequentialism in the first place. It's more similar to the Prisoner's Dilemma, if you ask me.

Comment author: Nisan 08 June 2010 08:33:29PM 2 points [-]

1-3: It's an alief, not a belief, because I know that lying to my spouse doesn't really make my spouse more likely to lie to me. But yes, I suppose I would be a happier person if I were capable of maintaining that alief (and repressing my guilt) while having an affair. I wonder if I would want to take a pill that would do that. Interesting. Anyways, if I did take that pill, then yes, I would cheat and lie.

Comment author: hegemonicon 10 June 2010 02:40:51PM *  8 points [-]

An idea that may not stand up to more careful reflection.

Evidence shows that people have limited quantities of willpower – exercise it too much, and it gets used up. I suspect that rather than a mere mental flaw, this is a design feature of the brain.

Man is often called the social animal. We band together in groups – families, societies, civilizations – to solve our problems. Groups are valuable to have, and so we have values – altruism, generosity, loyalty – that promote group cohesion and success. However, it doesn’t pay to be COMPLETELY supportive of the group. Ultimately the goal is replication of your genes, and though being part of a group can further that goal, it can also hinder it if you take it too far (sacrificing yourself for the greater good is not adaptive behavior). So it pays to have relatively fluid group boundaries that can be created as needed, depending on which group best serves your interest. And indeed, studies show that group formation/division is the easiest thing in the world to create – even groups chosen completely at random from a larger pool will exhibit rivalry and conflict.

Despite this, it’s the group-supporting values that form the higher level values that we pay lip service too. Group values are the ones we believe are our ‘real’ values, the ones that form the backbone of our ethics, the ones we signal to others at great expense. But actually having these values is tricky from an evolutionary standpoint – strategically, you’re much better off being selfish than generous, being two-faced than loyal, and furthering your own gains at the expense of everyone elses. So humans are in a pickle – it’s beneficial for them to form groups to solve their problems and increase their chances of survival, but it’s also beneficial for people to be selfish and mooch off the goodwill of the group. Because of this, we have sophisticated machinery called ’suspicion’ to ferret out any liars or cheaters furthering their own gains at the groups expense. Of course, evolution is an arms race, so it’s looking for a method to overcome these mechanisms, for ways it can fulfill it’s base desires while still appearing to support the group.

It accomplished this by implementing willpower. Because deceiving others about what we believe would quickly be uncovered, we don’t actually deceive them – we’re designed so that we really, truly, in our heart of hearts believe that the group-supporting values – charity, nobility, selflessness – are the right things to do. However, we’re only given a limited means to accomplish them. We can leverage our willpower to overcome the occasional temptation, but when push comes to shove – when that huge pile of money or that incredible opportunity or that amazing piece of ass is placed in front of us, willpower tends to fail us. Willpower is generally needed for the values that don’t further our evolutionary best interests – you don’t need willpower to run from danger or to hunt an animal if you’re hungry or to mate with a member of the opposite sex. We have much better, much more successful mechanisms that accomplish those goals. Willpower is designed so that we really do want to support the group, but wind up failing at it and giving in to our baser desires – the ones that will actually help our genes get replicated.

Of course, the maladaption comes into play due to the fact that we use willpower to try to accomplish other, non-group related goals – mostly the long-term, abstract plans we create using high-level, conscious thinking. This does appear to be a design flaw (though since humans are notoriously bad at making long-term predictions, it may not be as crippling as it first appears.)

Comment author: RobinZ 10 June 2010 04:27:21PM 1 point [-]

That is certainly interesting enough to subject to further reflection. Do we have any evolutionary psychologists in the audience?

Comment author: Clippy 07 June 2010 03:59:56PM 8 points [-]

I have a question about why humans see the following moral positions as different when really they look the same to me:

1) "I like to exist in a society that has punishments for non-cooperation, but I do not want the punishments to be used against me when I don't cooperate."

2) "I like to exist in a society where beings eat most of their children, and I will, should I live that long, want to eat most of my children too, but, as a child, I want to be exempt from being a target for eating."

Comment author: JenniferRM 08 June 2010 12:07:24AM 4 points [-]

Abstract preferences for or against the existence of enforcement mechanisms that could create binding cooperative agreements between previously autonomous agents have very very few detailed entailments.

These abstractions leave the nature of the mechanisms, the conditions of their legitimate deployment, and the contract they will be used to enforce almost completely open to interpretation. The additional details can themselves be spelled out later, in ways that maintain symmetry among different parties to a negotiation, which is a strong attractor in the semantic space of moral arguments.

This makes agreement with "the abstract idea of punishment" into the sort of concession that might be made at the very beginning of a negotiating process with an arbitrary agent you have a stake in influencing (and who has a stake in influencing you) upon which to build later agreements.

The entailments of "eating children" are very very specific for humans, with implications in biology, aging, mortality, specific life cycles, and very distinct life processes (like fuel acquisition versus replication). Given the human genome, human reproductive strategies, and all extant human cultures, there is no obvious basis for thinking this terminology is superior until and unless contact is made with radically non-human agents who are nonetheless "intelligent" and who prefer this terminology and can argue for it by reference to their own internal mechanisms and/or habits of planning, negotiation, and action.

Are you proposing to be such an agent? If so, can you explain how this terminology suits your internal mechanisms and habits of planning, negotiation, and action? Alternatively, can you propose a different terminology for talking about planning, negotiation, and action that suits your own life cycle?

For example, if one instance of Clippy software running on one CPU learns something of grave importance to its systems for choosing between alternative courses of action, how does it communicate this to other instances running basically the same software? Is this inter-process communication trusted, or are verification steps included in case one process has been "illegitimately modified" or not? Assuming verification steps take place, do communications with humans via text channels like this website feed through the same filters, analogous filters, or are they entirely distinct?

More directly, can you give us an IP address, port number, and any necessary "credentials" for interacting with an instance of you in the same manner that your instances communicate over TCP/IP networks with each other? If you aren't currently willing to provide such information, are there preconditions you could propose before you would do so?

Comment author: Blueberry 07 June 2010 04:49:50PM 3 points [-]

Except for the bizarreness of eating most of your children, I suspect that most humans would find the two positions equally hypocritical. Why do you think we see them as different?

Comment author: Clippy 07 June 2010 06:03:43PM *  1 point [-]

Why do you think we see them as different?

That belief is based on the reaction to this article, and the general position most of you take, which you claim requires you to balance current baby-eater adult interests against those of their children, such as in this comment and this one.

The consensus seems to be that humans are justified in exempting baby-eater babies from baby-eater rules, just like the being in statement (2) requests be done for itself. Has this consensus changed?

Comment author: Blueberry 07 June 2010 08:05:09PM 2 points [-]

I understand what you mean now.

Ok, so first of all, there's a difference between a moral position and a preference. For instance, I may prefer to get food for free by stealing it, but hold the moral position that I shouldn't do that. In your example (1), no one wants the punishments used against them, but we want them to exist overall because they make society better (from the point of view of human values).

In example (2), (most) humans don't want the Babyeaters to eat any babies: it goes against our values. This applies equally to the child and adult Babyeaters. We don't want the kids to be eaten, and we don't want the adults to eat. We don't want to balance any of these interests, because they go against our values. Just like you wouldn't balance out the interests of people who want to destroy metal or make staples instead of paperclips.

So my reaction to position (1) is "Well, of course you don't want the punishments. That's the point. So cooperate, or you'll get punished. It's not fair to exempt yourself from the rules." And my reaction to position (2) is "We don't want any baby-eating, so we'll save you from being eaten, but we won't let you eat any other babies. It's not fair to exempt yourself from the rules." This seems consistent to me.

Comment author: Clippy 07 June 2010 09:03:47PM *  3 points [-]

But I thought the human moral judgment that the baby-eaters should not eat babies was based on how it inflicts disutility on the babies, not simply from a broad, categorical opposition to sentient beings being eaten?

That is, if a baby wanted to get eaten (or perhaps suitably intelligent being like an adult), you would need some other compelling reason to oppose the being being eaten, correct? So shouldn't the baby-eaters' universal desire to have a custom of baby-eating put any baby-eater that wants to be exempt from baby-eating entirely, in the same position as the being in (1) -- which is to say, a being that prefers a system but prefers to "free ride" off the sacrifices that the system requires of everyone?

Comment author: JStewart 08 June 2010 03:41:19AM 3 points [-]

Isn't your point of view precisely the one the SuperHappies are coming from? Your critique of humanity seems to be the one they level when asking why, when humans achieved the necessary level of biotechnology, they did not edit their own minds. The SuperHappy solution was to, rather than inflict disutility by punishing defection, instead change preferences so that the cooperative attitude gives the highest utility payoff.

Comment author: Clippy 08 June 2010 07:05:31PM 1 point [-]

No, I'm criticizing humans for wanting to help enforce a relevantly-hypocritical preference on the grounds of its superficial similarities to acts they normally oppose. Good question though.

Comment author: jimrandomh 07 June 2010 04:32:53PM 1 point [-]

Adults, by choosing to live in a society that punishes non-cooperators, implicitly accept a social contract that allows them to be punished similarly. While they would prefer not to be punished, most societies don't offer asymmetrical terms, or impose difficult requirements such as elections, on people who want those asymmetrical terms.

Children, on the other hand, have not yet had the opportunity to choose the society that gives them the best social contract terms, and wouldn't have sufficient intelligence to do so anyways. So instead, we model them as though they would accept any social contract that's at least as good as some threshold (goodness determined retrospectively by adults imagining what they would have preferred). Thus, adults are forced by society to give implied consent to being punished if they are non-cooperative, but children don't give consent to be eaten.

Comment author: Clippy 07 June 2010 05:40:20PM 5 points [-]

Children, on the other hand, have not yet had the opportunity to choose the society that gives them the best social contract terms, and wouldn't have sufficient intelligence to do so anyways.

What if I could guess, with 100% accuracy, that the child will decide to retroactively endorse the child-eating norm as an adult? To 99.99% accuracy?

Comment author: jimrandomh 07 June 2010 06:07:45PM 1 point [-]

It is not the adults' preference that matters, but the adults' best model of the childrens' preferences. In this case there is an obvious reason for those preferences to differ - namely, the adult knows that he won't be one of those eaten.

In extrapolating a child's preferences, you can make it smarter and give it true information about the consequences of its preferences, but you can't extrapolate from a child whose fate is undecided to an adult that believes it won't be eaten; that change alters its preferences.

Comment author: Clippy 07 June 2010 06:14:31PM 1 point [-]

It is not the adults' preference that matters, but the adults' best model of the childrens' preferences.

Do you believe that all children's preferences must be given equal weight to that of adults, or just the preferences that the child will retroactively reverse on adulthood?

Comment author: NancyLebovitz 07 June 2010 12:44:09PM 8 points [-]
Comment author: Houshalter 10 June 2010 02:00:52AM *  6 points [-]

I just realised that infinite processing power creates a weird moral dilema:

Suppose you take this machine and put in a program which simulates every possible program it could ever run. Of course it only takes a second to run the whole program. In that second, you created every possible world that could ever exist, every possible version of yourself. This includes versions that are being tortured, abused, and put through horrible unethical situations. You have created an infinite number of holocausts and genocides and things much, much worse then what you could ever immagine. Most people would consider a program like this unethical to run. But what if the computer wasn't really a computer, it was an infinitely large database that contained every possible input and a corresponding output. When you put the program in, it just finds the right output and gives it to you, which is essentially a copy of the database itself. Since there isn't actually any computational process here, there is no unethical things being simulated. Its no more evil than a book in the library about genocide. And this does apply to the real world. It's essentially the chineese room problem - does a simulated brain "understand" anything? Does it have "rights"? Does how the information was processed make a difference? I would like to know what people at LW think about this.

Comment author: Nick_Tarleton 10 June 2010 07:17:05AM 4 points [-]
Comment author: toto 10 June 2010 03:18:31PM *  1 point [-]

I have problems with the "Giant look-up table" post.

"The problem isn't the levers," replies the functionalist, "the problem is that a GLUT has the wrong pattern of levers. You need levers that implement things like, say, formation of beliefs about beliefs, or self-modeling... Heck, you need the ability to write things to memory just so that time can pass for the computation. Unless you think it's possible to program a conscious being in Haskell."

If the GLUT is indeed behaving like a human, then it will need some sort of memory of previous inputs. A human's behaviour is dependent not just on the present state of the environment, but also on previous states. I don't see how you can successfully emulate a human without that. So the GLUT's entries would be in the form of products of input states over all previous time instants. To each of these possible combinations, the GLUT would assign a given action.

Note that "creation of beliefs" (including about beliefs) is just a special case of memory. It's all about input/state at time t1 influencing (restricting) the set of entries in the table that can be looked up at time t2>t1. If a GLUT doesn't have this ability, it can't emulate a human. If it does, then it can meet all the requirements spelt out by Eliezer in the above passage.

So I don't see how the non-consciousness of the GLUT is established by this argument.

But in this case, the origin of the GLUT matters; and that's why it's important to understand the motivating question, "Where did the improbability come from?"

The obvious answer is that you took a computational specification of a human brain, and used that to precompute the Giant Lookup Table. (...) In this case, the GLUT is writing papers about consciousness because of a conscious algorithm. The GLUT is no more a zombie, than a cellphone is a zombie because it can talk about consciousness while being just a small consumer electronic device. The cellphone is just transmitting philosophy speeches from whoever happens to be on the other end of the line. A GLUT generated from an originally human brain-specification is doing the same thing.

But the difficulty is precisely to explain why the GLUT would be different from just about any possible human-created AI in this respect. Keeping in mind the above, of course.

Comment author: JoshuaZ 07 June 2010 01:41:25PM *  6 points [-]

That's well done although two of the central premises are likely incorrect. First, the notion that a quantum computer would have infinite processing capability is incorrect. Quantum computation allows speed-ups of certain computational processes. Thus for example, Shor's algorithm allows us to factor integers quickly. But if our understanding of the laws of quantum mechanics is at all correct, this can't lead to anything like that in the story. In particular, under the standard descriptor for quantum computing, the class of problems reliably solvable on a quantum computer in polynomial time (that is the time required to solve is bounded above by a polynomial function of the length of the input sequence), BQP is is a subset of of PSPACE, the set of problems which can be solved on a classical computer using memory bounded by a polynomial of the space of the input. Our understanding of quantum mechanics would have to be very far off for this to be wrong.

Second, if our understanding of quantum mechanics is correct, there's a fundamentally random aspect to the laws of physics. Thus, we can't simply make a simulation and advance it ahead the way they do in this story and expect to get the same result.

Even if everything in the story was correct, I'm not at all convinced that things would settle down on a stable sequence as they do here. If your universe is infinite then your possible number of worlds are infinite so there's no reason you couldn't have a wandering sequence of worlds. Edit: Or for that matter, couldn't have branches if people simulate additional worlds with other laws of physics or the same laws but different starting conditions.

Comment author: ocr-fork 07 June 2010 04:16:09PM *  4 points [-]

First, the notion that a quantum computer would have infinite processing capability is incorrect... Second, if our understanding of quantum mechanics is correct

It isn't. They can simulate a world where quantum computers have infinite power because because they live in a world where quantum computers have infinite power because...

Comment author: JoshuaZ 07 June 2010 04:23:35PM 4 points [-]

Ok, but in that case, that world in question almost certainly can't be our world. We'd have to have deep misunderstandings about the rules for this universe. Such a universe might be self-consistent but it isn't our universe.

Comment author: ocr-fork 07 June 2010 04:49:52PM 4 points [-]

Of course. It's fiction.

Comment author: JoshuaZ 07 June 2010 04:59:23PM *  3 points [-]

What I mean is that this isn't a type of fiction that could plausibly occur in our universe. In contrast for example, there's nothing in the central premises of say Blindsight that as we know it would prevent the story from taking place. The central premise here is one that doesn't work in our universe.

Comment author: Blueberry 07 June 2010 05:06:26PM 1 point [-]

Well, it does suggest they've made recent discoveries that changed the way they understood the laws of physics, which could happen in our world.

Comment author: jimrandomh 07 June 2010 10:04:42PM *  3 points [-]

The likely impossibility of getting infinite comutational power is a problem, but quantum nondeterminism or quantum branching don't prevent using the trick described in the story, they just make it more difficult. You don't have to identify one unique universe that you're in, just a set of universes that includes it. Given an infinitely fast, infinite storage computer, and source code to the universe which follows quantum branching rules, you can get root powers by the following procedure:

Write a function to detect a particular arrangement of atoms with very high information content - enough that it probably doesn't appear by accident anywhere in the universe. A few terabytes encoded as iron atoms present or absent at spots on a substrate, for example. Construct that same arrangement of atoms in the physical world. Then run a program that implements the regular laws of physics, except that wherever it detects that exact arrangement of atoms, it deletes them and puts a magical item, written into the modified laws of physics, in their place.

The only caveat to this method (other than requiring an impossible computer) is that it also modifies other worlds, and other places within the same world, in the same way. If the magical item created is programmable (as it should be), then every possible program will be run on it somewhere, including programs that destroy everything in range, so there will need to be some range limit.

Comment author: Houshalter 07 June 2010 07:29:24PM 2 points [-]

Couldn't they just run the simulation to its end rather then just let it sit there and take the chance that it could accidently be destroyed. If its infinitley powerful, it would be able to do that.

Comment author: ocr-fork 08 June 2010 12:48:16AM 2 points [-]

Then they miss their chance to control reality. They could make a shield out of black cubes.

Comment author: Blueberry 07 June 2010 07:45:53PM 1 point [-]

They could just turn it off. If they turned off the simulation, the only layer to exist would be the topmost layer. Since everyone has identical copies in each layer, they wouldn't notice any change if they turned it off.

Comment author: SilasBarta 12 June 2010 07:11:47PM *  7 points [-]

Potential top-level article, have it mostly written, let me know what you think:

Title: The hard problem of tree vibrations [tentative]

Follow-up to: this comment (Thanks Adelene Dawner!)

Related to: Disputing Definitions, Belief in the Implied Invisible

Summary: Even if you agree that trees normally make vibrations when they fall, you're still left with the problem of how you know if they make vibrations when there is no observational way to check. But this problem can be resolved by looking at the complexity of the hypothesis that no vibrations happen. Such a hypothesis is predicated on properties specific to the human mind, and therefore is extremely lengthy to specify. Lacking the type and quantity of evidence necessary to locate this hypothesis, it can be effectively ruled out.

Body: A while ago, Eliezer Yudkowsky wrote an article about the "standard" debate over a famous philosophical dilemma: "If a tree falls in a forest and no one hears it, does it make a sound?" (Call this "Question Y.") Yudkowsky wrote as if the usual interpretation was that the dilemma is in the equivocation between "sound as vibration" and "sound as auditory perception in one's mind", and that the standard (naive) debate relies on two parties assuming different definitions, leading to a pointless argument. Obviously, it makes a sound in the first sense but not the second, right?

But throughout my whole life up to that point (the question even appeared in the animated series Beetlejuice that I saw when I was little), I had assumed a different question was being asked: specifically,

If a tree falls, and no human (or human-entangled[1] sensor) is around to hear it, does it still make vibrations? On what basis do you believe this, lacking a way to directly check? (Call this "Question S".)

Now, if you're a regular on this site, you will find that question easy to answer. But before going into my exposition of the answer, I want to point out some errors that Question S does not make.

For one thing, it does not equivocate between two meanings of sound -- there, sound is taken to mean only one thing: the vibrations.

Second, it does not reduce to a simple question about anticipation of experience. In Question Y, the disputants can run through all observations they anticipate, and find them to be the same. However, if you look at the same cases in Question S, you don't resolve the debate so easily: both parties agree that by putting a tape-recorder by the tree, you will detect vibrations from the tree falling, even if people aren't around. But Question S instead specifically asks about what goes on when these kinds of sensors are not around, rendering such tests unhelpful for resolving such a disagreement.

So how do you go about resolving Question S? Yudkowsky gave a model for how to do this in Belief in the Implied Invisible, and I will do something similar here.

Complexity of the hypothesis

First, we observe that, in all cases where we can make a direct measurement, trees make vibrations when they fall. And we're tasked with finding out whether, specifically in those cases where a human (or appropriate organism with vibration sensitivity in its cognition) will never make a measurement of the vibrations, the vibrations simply don't happen. That is, when we're not looking -- and never intend to look -- trees stop the "act" and don't vibrate.

The complexity this adds to the laws of physics is astounding and may be hard to appreciate at first. This belief would require us to accept that nature has some way of knowing which things will eventually reach a cognitive system in such a way that it informs it that vibrations have happened. It must selectively modify material properties in precisely defined scenarios. It must have a precise definition of what counts as a tree.

Now, if this actually happens to be how the world works, well, then all the worse for our current models! However, each bit of complexity you add to a hypothesis reduces its probability and so must be justified by observations with a corresponding likelihood ratio -- that is, the ratio of the probability of the observation happening if this alternate hypothesis is true, compared to if it were false. By specifying the vibrations' immunity to observation, the log of this ratio is zero, meaning observations are stipulated to be uninformative, and unable to justify this additional supposition in the hypothesis.

[1] You might wonder how someone my age in '89-'91 would come up with terms like "human-entangled sensor", and you're right: I didn't use that term. Still, I considered the use of a tape recorder that someone will check to be a "someone around to hear it", for purposes of this dilemma. Least Convenient Possible World and all...

Comment author: Oscar_Cunningham 12 June 2010 07:47:17PM 2 points [-]

I think that if this post is left as it is this post would be to trivial to be a top level post. You could reframe it as a beginners' guide to Occam, or you could make it more interesting by going deeper into some of the issues (if you can think of anything more to say on the topic of differentiating between hypotheses that make the same predictions, that might be interesting, although I think you might have said all there is to say)

Comment author: AdeleneDawner 13 June 2010 05:41:53AM 3 points [-]

It could also be framed as an issue of making your beliefs pay rent, similar to the dragon in the garage example - or perhaps as an example of how reality is entangled with itself to such a degree that some questions that seem to carve reality at the joints don't really do so.

(If falling trees don't make vibrations when there's no human-entangled sensor, how do you differentiate a human-entangled sensor from a non-human-entangled sensor? If falling-tree vibrations leave subtle patterns in the surrounding leaf litter that sufficiently-sensitive human-entangled sensors can detect, does leaf litter then count as a human-entangled sensor? How about if certain plants or animals have observably evolved to handle falling-tree vibrations in a certain way, and we can detect that. Then such plants or animals (or their absence, if we're able to form a strong enough theory of evolution to notice the absence of such reactions where we would expect them) could count as human-entangled sensors well before humans even existed. In that case, is there anything that isn't a human-entangled sensor?)

Comment author: SilasBarta 13 June 2010 07:17:37PM 3 points [-]

Good points in the parenthetical -- if I make it into a top-level article, I'll be sure to include a more thorough discussion of what concept is being carved with the hypothesis that there are no tree vibrations.

Comment author: AdeleneDawner 12 June 2010 07:23:44PM 2 points [-]

I believe this is the conversation you're responding to.

(upvoted)

Comment author: NancyLebovitz 10 June 2010 04:28:08PM 7 points [-]
Comment author: komponisto 11 June 2010 04:20:20AM *  4 points [-]

Seconding kodos96. As this would exonerate not only Knox and Sollecito but Guede as well, it has to be treated with considerable skepticism, to say the least.

More significant, it seems to me (though still rather weak evidence), is the Alessi testimony, about which I actually considered posting on the March open thread.

Still, the Aviello story is enough of a surprise to marginally lower my probability of Guede's guilt. My current probabilities of guilt are:

Knox: < 0.1 % (i.e. not a chance)

Sollecito: < 0.1 % (likewise)

Guede: 95-99% (perhaps just low enough to insist on a debunking of the Aviello testimony before convicting)

It's probably about time I officially announced that my revision of my initial estimates for Knox and Sollecito was a mistake, an example of the sin of underconfidence.

I of course remain willing to participate in a debate with Rolf Nelson on this subject.

Finally, I'd like to note that the last couple of months have seen the creation of a wonderful new site devoted to the case, Injustice in Perugia, which anyone interested should definitely check out. Had it been around in December, I doubt that I could have made my survey seem like a fair fight between the two sides.

Comment author: kodos96 11 June 2010 07:21:12PM 1 point [-]

More significant, it seems to me (though still rather weak evidence), is the Alessi testimony, about which I actually considered posting on the March open thread. Still, the story is enough of a surprise to marginally lower my probability of Guede's guilt.

I hadn't heard about this - I just read your link though, and maybe I'm missing something, but I don't see how it lowers the probability of Guede's guilt. He (supposedly) confessed to having been at the crimescene, and that Knox and Sollecito weren't there. How does that, if true, exonerate Guede?

Comment author: komponisto 11 June 2010 09:57:04PM *  2 points [-]

You omitted a crucial paragraph break. :-)

The Aviello testimony would exonerate Guede (and hence is unlikely to be true); the Alessi testimony is essentially consistent with everything else we know, and isn't particularly surprising at all.

I've edited the comment to clarify.

Comment author: RobinZ 10 June 2010 05:39:15PM 2 points [-]

That story would be consistent with Guédé's, modulo the usual eyewitness confusion.

Comment author: kodos96 10 June 2010 07:06:10PM *  4 points [-]

And modulo all the forensic evidence.

Obviously this is breaking news and it's too soon to draw a conclusion, but at first blush this sounds like just another attention seeker, like those who always pop up in these high profile cases. If he really can produce a knife, and it matches the wounds, then maybe I'll reconsider, but at the moment my BS detector is pegged.

Of course, it's still orders of magnitude more likely than Knox and Sollecito being guilty.

Comment author: RichardKennaway 10 June 2010 10:00:57AM *  7 points [-]

How many lottery tickets would you buy if the expected payoff was positive?

This is not a completely hypothetical question. For example, in the Euromillions weekly lottery, the jackpot accumulates from one week to the next until someone wins it. It is therefore in theory possible for the expected total payout to exceed the cost of tickets sold that week. Each ticket has a 1 in 76,275,360 (i.e. C(50,5)*C(9,2)) probability of winning the jackpot; multiple winners share the prize.

So, suppose someone draws your attention (since of course you don't bother following these things) to the number of weeks the jackpot has rolled over, and you do all the relevant calculations, and conclude that this week, the expected win from a €1 bet is €1.05. For simplicity, assume that the jackpot is the only prize. You are also smart enough to choose a set of numbers that look too non-random for any ordinary buyer of lottery tickets to choose them, so as to maximise your chance of having the jackpot all to yourself.

Do you buy any tickets, and if so how many?

If you judge that your utility for money is sublinear enough to make your expected gain in utilons negative, how large would the jackpot have to be at those odds before you bet?

Comment author: RobinZ 10 June 2010 01:26:29PM *  2 points [-]

The traditional answer is to follow the Kelly criterion, is it not? That would imply

where n is the number of tickets. This implies you should buy n such that (€1)*n = Wf*, where W is your initial wealth.

Edit: Thanks, JoshuaZ, for pointing out that the Kelly criterion might not be the applicable one in a given situation.

Comment author: Mass_Driver 10 June 2010 04:14:57PM 1 point [-]

OK, I have a question! Suppose I hold a risky asset that costs me c at time t, and whose value at time t is predicted to be k * (1 + r), with standard deviation s. How can I calculate the length of time that I will have to hold the asset in order to rationally expect the asset to be worth, say, 2c with probability p?

I am not doing a finance class or anything; I am genuinely curious.

Comment author: Morendil 09 June 2010 04:06:51PM *  7 points [-]

Less Wrong Book Club and Study Group

(This is a draft that I propose posting to the top level, with such improvements as will be offered, unless feedback suggests it is likely not to achieve its purposes. Also reply if you would be willing to co-facilitate: I'm willing to do so but backup would be nice.)

Do you want to become stronger in the way of Bayes? This post is intended for people whose understanding of Bayesian probability theory is currently between levels 0 and 1, and who are interested in developing deeper knowledge through deliberate practice.

Our intention is to form a self-study group composed of peers, working with the assistance of a facilitator - but not necessarily of a teacher or of an expert in the topic. Some students may be somewhat more advanced along the path, and able to offer assistance to others.

Our first text will be E.T. Jayne's Probability Theory: The Logic of Science, which can be found in PDF form (in a slightly less polished version than the book edition) here or here.

We will work through the text in sections, at a pace allowing thorough understanding: expect one new section every week, maybe every other week. A brief summary of the currently discussed section will be published as an update to this post, and simultaneously a comment will open the discussion with a few questions, or the statement of an exercise. Please use ROT13 whenever appropriate in your replies.

A first comment below collects intentions to participate. Please reply to this comment only if you are genuinely interested in gaining a better understanding of Bayesian probability and willing to commit to spend a few hours per week reading through the section assigned or doing the exercises. A few days from now the first section will be posted.

Comment author: MartinB 07 June 2010 07:20:04PM 7 points [-]

Question: whats your experience with stuff that seems new agy at first look, like yoga, meditation and so on. Anything worth trying?

Case in point: i read in Feynmans book about deprivation tanks, and recently found out that they are available in bigger cities. (Berlin, germany in my case.) will try and hopefully enjoy that soon. Sadly those places are run by new-age folks that offer all kinds of strange stuff, but that might not take away from the experience of floating in a sensory empty space.

Comment author: khafra 08 June 2010 07:29:51PM 3 points [-]

Chinese internal martial arts: Tai Chi, Xingyi, and Bagua. The word "chi" does not carve reality at the joints: There is no literal bodily fluid system parallel to blood and lymph. But I can make training partners lightheaded with a quick succession of strikes to Ren Ying (ST9) then Chi Ze (LU5); I can send someone stumbling backward with some fairly light pushes; after 30-60 seconds of sparring to develop a rapport I can take an unwary opponent's balance without physical contact.

Each of these skills fit more naturally under different categories, but if you want to learn them all the most efficient way is to study a Chinese internal martial art or something similar.

Comment author: Blueberry 09 June 2010 07:04:52PM 4 points [-]

I can take an unwary opponent's balance without physical contact.

This sounds magical at first reading, but is actually not that tricky. It's just psychology and balance. If you set up a pattern of predictable attacks, then feint in the right direction while your opponent is jumping at you off-balance, you can surprise him enough to make him fall as he attempts to ward off your feint.

Comment author: RichardKennaway 09 June 2010 09:55:28PM *  3 points [-]

I used to go to a Tai Chi class (I stopped only because I decided I'd taken it as far as I was going to), and the instructor, who never talked about "chi" as anything more than a metaphor or a useful visualisation, said this about the internal arts:

In the old days (that would be pre-revolutionary China) you wouldn't practice just Tai Chi, or begin with Tai Chi. Tai Chi was the equivalent of postgraduate study in the martial arts. You would start out by learning two or three "hard", "external" styles. Then, having reached black belt in those, and having developed your power, speed, strength, and fighting spirit, you would study the internal arts, which would teach you the proper alignments and structures, the meaning of the various movements and forms. In the class there were two students who did Gojuryu karate, a 3rd dan and a 5th dan, and they both said that their karate had improved no end since taking up Tai Chi.

Which is not to say that Tai Chi isn't useful on its own, it is, but there is that wider context for getting the maximum use out of it.

Comment author: RichardKennaway 07 June 2010 11:00:07PM 3 points [-]

Question: whats your experience with stuff that seems new agy at first look, like yoga, meditation and so on. Anything worth trying?

The Five Tibetans are a set of physical exercises which rejuvenate the body to youthful vigour and prolong life indefinitely. They are at least 2,500 years old, and practiced by hidden masters of secret wisdom living in remote monasteries in Tibet, where, in the earlier part of the 20th century, a retired British army colonel sought out these monasteries, studied with the ancient masters to great effect, and eventually brought the exercises to the West, where they were first published in 1939.

Ok, you don't believe any of that, do you? Neither do I, except for the first eight words and the last six. I've been doing these exercises since the beginning of 2009, since being turned on to them by Steven Barnes' blog and they do seem to have made a dramatic improvement in my general level of physical energy. Whether it's these exercises specifically or just the discipline of doing a similar amount of exercise first thing in the morning, every morning, I haven't taken the trouble to determine by varying them.

More here and here. Nancy Lebovitz also mentioned them.

I also do yoga for flexibility (it works) and occasionally meditation (to little detectable effect). I'd be interested to hear from anyone here who meditates and gets more from it than I do.

Comment author: NancyLebovitz 07 June 2010 11:05:24PM 1 point [-]
Comment author: Mass_Driver 11 June 2010 07:43:46PM 1 point [-]

I've had great results from modest (2-3 hrs/wk) investments in hatha yoga, over and above what I get from standard Greco-Roman "calisthenics."

Besides the flexibility, breathing, and posture benefits, I find that the idea of 'chakras' is vaguely useful for focusing my conscious attention on involuntary muscle systems. I would be extremely surprised if chakras "cleaved reality at the joints" in any straightforward sense, but the idea of chakras helps me pay attention to my digestion, heart rate, bladder, etc. by making mentally uninteresting but nevertheless important bodily functions more interesting.

Comment author: Jonathan_Graehl 08 June 2010 02:13:20AM 1 point [-]

I've done yoga every week for the last month or two. It's pleasant. Other than paying attention to how I'm holding my body vs. the instruction, I mostly stop thinking for an hour (as we're encouraged to do), which is nice.

I can't say I notice any significant lasting effects yet. I'm slightly more flexible.

Comment author: gwern 07 June 2010 09:49:58PM 1 point [-]

Hard to say - even New Agey stuff evolves. (Not many followers of Reich pushing their copper-lined closets these days.)

Generally, background stuff is enough. There's no shortage of hard scientific evidence about yoga or meditation, for example. No need for heuristics there. Similarly there's some for float tanks. In fact, I'm hard pressed to think of any New Agey stuff where there isn't enough background to judge it on its own merits.

Comment author: [deleted] 07 June 2010 12:45:52PM *  6 points [-]

Many are calling BP evil and negligent, has there actually been any evidence of criminal activities on their part? My first guess is that we're dealing with hindsight bias. I am still casually looking into it, but I figured some others here may have already invested enough work into it to point me in the right direction.

Like any disaster of this scale, it may be possible to learn quite a bit from it, if we're willing.

Comment author: Piglet 07 June 2010 04:01:43PM 7 points [-]

It depends on what you mean by "criminal"; under environmental law, there are both negligence-based (negligent discharge of pollutants to navigable waters) and strict liability (no intent requirement, such as killing of migratory birds) crimes that could apply to this spill. I don't think anyone thinks BP intended to have this kind of spill, so the interesting question from an environmental criminal law perspective is whether BP did enough to be treated as acting "knowingly" -- the relevant intent standard for environmental felonies. This is an extremely slippery concept in the law, especially given the complexity of the systems at issue here. Litigation will go on for many years on this exact point.

Comment author: Unnamed 17 June 2010 06:23:05AM 2 points [-]

I've heard scattered bits of accusations of misdeeds by BP which may have contributed to the spill. Here's a list from the congressional investigation of 5 decisions that BP made "for economic reasons that increased the danger of a catastrophic well failure" according to a letter from the congressmen. It sounds like BP took a bunch of risky shortcuts to save time and money, although I'd want to hear from people who actually understand the technical issues before being too confident.

There are other suspicions and allegations floating around, like this one.

Comment author: [deleted] 24 June 2010 06:08:00PM 0 points [-]

That's a good start, I appreciate it!

Comment author: billswift 08 June 2010 02:55:59AM 1 point [-]

You are not really going to learn much unless you are interested in wading through lots of technical articles. If you want to learn, you need to wait until it has been digested by relevant experts into books. I am not sure what you think you can learn from this, but there are two good books of related information available now:

Jeff Wheelwright, Degrees of Disaster, about the environmental effects of the Exxon Valdez spill and the clean up.

Trevor Kletz, What Went Wrong?: Case Histories of Process Plant Disasters, which is really excellent. [For general reading, an older edition is perfectly adequate, new copies are expensive.] It has an incredible amount of detail, and horrifying accounts of how apparently insignificant mistakes can (often literally) blow up on you.

Comment author: NancyLebovitz 08 June 2010 09:18:38AM 3 points [-]

In a recent video, Taleb argues that people generally put too much focus on the specifics of a disaster, and too little on what makes systems fragile.

He said that high debt means (among other things) too much focus on the short run, and skimping on insurance and precautions.

Comment author: Piglet 10 June 2010 03:35:18PM *  2 points [-]

Also, Richard Feynman's remarks on the loss of the Space Shuttle Challenger are a pretty accessible overview of the kinds of dynamics that contribute to major industrial accidents. http://history.nasa.gov/rogersrep/v2appf.htm

[edit: corrected, thx.]

Comment author: JoshuaZ 10 June 2010 04:20:52PM 2 points [-]

Pretty sure you mean Challenger. Feynman was involved in the investigation of the Challenger disaster. He was dead long before Columbia.

Comment author: Houshalter 07 June 2010 06:11:37PM 1 point [-]

I'm not sure it's relevant whether they did anything illegal or not. People always seem to want to blame and punish someone for their problems. In my opinion, they should be forced to pay for and compensate for all the damage, as well as a very large fine as punishment. This way in the future they, and other companies, can regulate themselves and prepare for emergencies as efficiently as possible without arbitrary and clunky government regulations and agencies trying to slap everything together at the last moment. Of course, if a single person actually did something irresponsible (eg; bob the worker just used duct tape to fix that pipe knowing that it wouldn't hold) then they should be able to be tried in court or sued/fined by the company. But even then, it's up to the company to make sure that stuff like this doesn't happen by making sure all of their workers are competent and certified.

Comment author: billswift 07 June 2010 11:17:33AM *  6 points [-]

Regrets and Motivation

Almost invariably everything is larger in your imagination than in real life, both good and bad, the consequences of mistakes loom worse, and the pleasure of gains looks better. Reality is humdrum compared to our imaginations. It is our imagined futures that get us off our butts to actually accomplish something.

And the fact that what we do accomplish is done in the humdrum, real world, means it can never measure up to our imagined accomplishments, hence regrets. Because we imagine that if we had done something else it could have measured up. The worst part of having regrets is the impact it has on our motivation.

somewhat expanded version of comment on OB a couple of months ago

Added: I didn't make the connection at first, but this is also Eliezer's point in this quote from The Super Happy People story, "It's bad enough comparing yourself to Isaac Newton without comparing yourself to Kimball Kinnison."

Comment author: xamdam 07 June 2010 04:35:27PM *  3 points [-]

I was talking to a friend yesterday and he mentioned a psychological study (I am trying to track down the source) that people tend to suffer MORE from failing to pursue certain opportunities than FAILING after pursuing them. So even if you're right about the overestimation of pleasure, it might just be irrelevant.

Comment author: Unnamed 08 June 2010 03:09:34AM *  4 points [-]

Here is a review of that psychological research (pdf), and there are more studies linked here (the keyword to look for is "regret"). The paper I linked is:

Gilovich, T., & Medvec, V. H. (1995). The experience of regret: What, when, and why. Psychological Review, 102, 379-395.

This article reviews evidence indicating that there is a temporal pattern to the experience of regret. Actions, or errors of commission, generate more regret in the short term; but inactions, or errors of omission, produce more regret in the long run. The authors contend that this temporal pattern is multiply determined, and present a framework to organize the divergent causal mechanisms that are responsible for it. In particular, this article documents the importance of psychological processes that (a) decrease the pain of regrettable action over time, (b) bolster the pain of regrettable inaction over time, and (c) differentially affect the cognitive availability of these two types of regrets. Both the functional and cultural origins of how people think about regret are discussed.

Comment author: billswift 08 June 2010 02:39:48AM *  2 points [-]

I haven't seen a study, but that is a common belief. A good quote to that effect,

Regret for the things we did can be tempered by time; it is regret for the things we did not do that is inconsolable. - - Sydney Harris

And I vaguely remember seeing another similar quote from Churchill.

Comment author: Wei_Dai 13 June 2010 06:15:35PM *  5 points [-]

While searching for literature on "intuition", I came upon a book chapter that gives "the state of the art in moral psychology from a social-psychological perspective". This is the best summary I've seen of how morality actually works in human beings.

The authors gives out the chapter for free by email request, but to avoid that trivial inconvenience, I've put up a mirror of it.

ETA: Here's the citation for future reference: Haidt, J., & Kesebir, S. (2010). Morality. In S. Fiske, D. Gilbert, & G. Lindzey (Eds.) Handbook of Social Psychology, 5th Edition. Hobeken, NJ: Wiley. Pp. 797-832.

Comment author: Alexandros 08 June 2010 01:35:33PM *  5 points [-]

This one came up at the recent London meetup and I'm curious what everyone here thinks:

What would happen if CEV was applied to the Baby Eaters?

My thoughts are that if you applied it to all baby eaters, including the living babies and the ones being digested, it would end up in a place that adult baby eaters would not be happy. If you expanded it to include all babyeaters that ever existed, or that would ever exist, knowing the fate of 99% of them, it would be a much more pronounced effect. So what I make of all this is that either CEV is not utility-function-neutral, or that the babyeater morality is objectively unstable when aggregated.

Thoughts?

Comment author: Morendil 08 June 2010 03:35:11PM 2 points [-]

What would happen if CEV was applied to the Baby Eaters?

My intuitions of CEV are informed by the Rawlsian Veil of Ignorance, which effectively asks: "What rules would you want to prevail if you didn't know in advance who you would turn out to be?"

Where CEV as I understand it adds more information - assumes our preferences are extrapolated as if we knew more, were more the kind of people we want to be - the Veil of Ignorance removes information: it strips people under a set of specific circumstances of the detailed information about what their preferences are, what their contignent histories brought them there, and so on. This includes things like what age you are, and even - conceivably - how many of you there are.

To this bunch of undifferentiated people you'd put the question, "All in favor of a 99% chance of dying horribly shortly after being born, in return for the 1% chance to partake in the crowning glory of babyeating cultural tradition, please raise your hands."

I expect that not dying horribly takes lexical precedence over any kind of cultural tradition, for any sentient being whose kin has evolved to sentience (it may not be that way for constructed minds). So I would expect the Babyeaters to choose against cultural tradition.

The obvious caveat is that my intuitions about CEV may be wrong, but lacking a formal explanation of CEV it's hard to check intuitions.

Comment author: Kevin 12 June 2010 06:15:20PM 4 points [-]
Comment deleted 07 June 2010 12:37:39PM *  [-]
Comment author: ata 07 June 2010 09:28:53PM 1 point [-]

I think the expectation is that, if all humans had the same knowledge and were better at thinking (and were more the people we'd like to be, etc.), then there would be a much higher degree of coherence than we might expect, but not necessarily that everyone would ultimately have the same utility function.

Comment author: billswift 07 June 2010 11:27:38AM 4 points [-]

I have been reading the “economic collapse” literature since I stumbled on Casey’s “Crisis Investing” in the early 1980s. They have really good arguments, and the collapses they predict never happen. In the late-90s, after reading “Crisis Investing for the Rest of the 1990s”, I sat down and tried to figure out why they were all so consistently wrong.

The conclusion I reached was that humans are fundamentally more flexible and more adaptable than the collapse-predictors' arguments allowed for, and society managed to work-around all the regulations and other problems the government and big businesses keep creating. Since the regulations and rules keep growing and creating more problems and rigidity along the way, eventually there will be a collapse, but anyone that gives any kind of timing for it is grabbing at the short end of the stick.

Anyone here have more suggestions as to reasons they have been wrong?

(originally posted on esr's blog 2010-05-09, revised and expanded since)

Comment author: blogospheroid 09 June 2010 09:18:43AM 4 points [-]

Not sure if you're referring to the same literature, but I note a great divergence between peak oil advocates and singularitarians. This is a little weird, if you think of Aumann's Agreement theorem.

Both groups are highly populated with engineer types, highly interested in cognitive biases, group dynamics, habits of individuals and societies and neither are mainstream.

Both groups use extrapolation of curves from very real phenomena. In the case of the kurzweillian singularitarians, it is computing power and in the case of the peak oil advocates, it is the hubbert curve for resources along with solid Net Energy based arguments about how civilization should decline.

The extreme among the Peak Oil advocates are collapsitarians and believe that people should drastically change their lifestyles, if they want to survive. They are also not waiting for the others to join them and many are preparing to go to small towns, villages etc. The oildrum, linked here had started as a moderate peak oil site discussing all possibilities, nowadays, apparently, its all doom all the time.

The extreme among the singularitarians have been asked no such sacrifice, just to give enough money and support to make sure that Friendly AI is achieved first.

Both groups believe that business as usual cannot go on for too long, but they expect dramatically different consequences. The singularitarians assert that economics conditions and technology will improve until a nonchalant super-intelligence will be created and wipe out humanity. The collapsitarians believe that economic conditions will worsen, civilization is not built robustly and will collapse badly with humanity probably going extinct or only the last hunter gatherers surviving.

Comment author: NancyLebovitz 09 June 2010 02:08:21PM 1 point [-]

It should be possible to believe both-- unless you're expecting peak oil to lead to social collapse fairly soon, Moore's law could make a singluarity possible while energy becomes more expensive.

Comment author: cupholder 10 June 2010 01:09:21AM 1 point [-]

Which could suggest a distressing pinch point: not wanting to delay AI too long in case we run out of energy for it to use; not wanting to make an AI too soon in case it's Unfriendly.

Comment author: ShardPhoenix 07 June 2010 01:17:39PM 2 points [-]

Could you give some examples of the predicted collapses that didn't happen?

Comment author: soreff 07 June 2010 03:15:21PM 4 points [-]

Y2K. I thought I had a solid lower bound for the size of that one: Small businesses basically did nothing in preparation, and they still had a fair amount of dependence on date-dependent programs, so I was expecting that the impact on them would set a sizable lower bound on the the size of the overall impact. I've never been so glad to be wrong. I would still like to see a good retrospective explaining how that sector of the economy wound up unaffected...

Comment author: pjeby 07 June 2010 04:10:29PM *  5 points [-]

Small businesses basically did nothing in preparation [for Y2K], and they still had a fair amount of dependence on date-dependent programs

The smaller the business, the less likely they are to have their own software that's not simply a database or spreadsheet, managed in say, a Microsoft product. The smaller the business, the less likely that anything automated is relying on correct date calculations.

These at least would have been strong mitigating factors.

[Edit: also, even industry-specific programs would likely be fixed by the manufacturer. For example, most of the real-estate software produced by the company I worked for in the 80's and 90's was Y2K-ready since before 1985.]

Comment author: billswift 07 June 2010 03:31:24PM 1 point [-]

First, the "economic collapse" I referred to in the original post were actually at least 6 different predictions at different times.

As another example, but not quite a "collapse" scenario, consider the predictions of the likelihood of nuclear war; there were three distinct periods where it was considered more or less likely by different groups. The late 1940s some intelligent and informed, but peripheral, observers like Robert Heinlein considered it a significant risk. Next was the late 1950s through the Cuban Missile Crisis in the early 1960s, when nearly everybody considered it a major risk. Then there was another scare in the late 1970s to early 1980s, primarily leftists (including the media) favoring disarmament promulgating the fear to try to get the US to reduce their stockpiles and conservatives (derided by the media as "survivalists" and nuts) who were afraid they would succeed.

Comment author: Mass_Driver 13 June 2010 10:07:37PM 3 points [-]

Anyone know how to defeat the availability heuristic? Put another way, does anyone have advice on how to deal with incoherent or insane propositions while losing as little personal sanity as possible? Is there such a thing as "safety gloves" for dangerous memes?

I'm asking because I'm currently studying for the California Bar exam, which requires me to memorize hundreds of pages of legal rules, together with their so-called justifications. Of course, in many cases the "justifications" are incoherent, Orwellian doublespeak, and/or tendentiously ideological. I really do want to memorize (nearly) all of these justifications, so that I can be sure to pass the exam and continue my career as a rationalist lawyer, but I don't want the pattern of thought used by the justifications to become a part of my pattern of thought.

Comment author: Jordan 13 June 2010 10:55:07PM 3 points [-]

I worry about this as well when I'm reading long arguments or long works of fiction presenting ideas I disagree with. My tactic is to stop occasionally and go through a mental dialog simulating how I would respond to the author in person. This serves a double purpose, as hopefully I'll have better cached arguments in the event I ever need them.

Of course, this is a dangerous tactic as well, because you may be shutting off critical reasoning applied to your preexisting beliefs. I only apply this tactic when I'm very confident the author is wrong and is using fallacious arguments. Even then I make sure to spend some amount of time playing devil's advocate.

Comment author: rhollerith_dot_com 13 June 2010 10:46:08PM *  3 points [-]

I would not worry overmuch about the long-term negative effects of your studying for the bar: with the possible exception of the "overly sincere" types who fall very hard for cults and other forms of indoctrination, people have a lot of antibodies to this kind of thing.

You will continue to be entagled with reality after you pass the exam, and you can do things, like read works of social science that carve reality at the joints, to speed up the rate at which your continued entaglement with reality with cancel out any falsehoods you have to cram for now. Specifically, there are works about the law that do carve reality at the joints -- Nick Szabo's online writings IMO fall in that category. Nick has a law degree, by the way, and there is certainly nothing wrong with his ability to perceive reality correctly.

ADDED. The things that are really damaging to a person's rationality, IMHO, are natural human motivations. When for example you start practicing, if you were to decide to do a lot of trials, and you learned to derive pleasure -- to get a real high -- from the combative and adversarial part of that, so that the high you got from winning with a slick and misleading angle trumped the high you get from satisfying you curiosity and from refining and finding errors in your model of reality -- well, I would worry about that a lot more than your throwing yourself fully into winning on this exam because IMHO the things we derive no pleasure from, but do to achieve some end we care about (like advancing in our career by getting a credential) have a lot less influence on who we turn out to be than things we do because we find them intrinsically rewarding.

One more thing: we should not all make our living as computer programmers. That would make the community less robust than it otherwise would be :)

Comment author: Mass_Driver 13 June 2010 11:08:00PM 1 point [-]

Thank you! This is really helpful, and I look forward to reading Szabo in August.

Comment author: simplicio 13 June 2010 04:13:15PM 3 points [-]

An interesting article criticizing speculation about social trends (specifically teen sex) in the absence of statistical evidence.

Comment author: RobinZ 13 June 2010 05:47:45PM 1 point [-]

Beautiful. Matthew Yglesias, +1 point.

It is entirely possible that some social groups are experiencing the kind of changes that Flanagan describes, but as Yglesias says, she apparently is unaware that there is such a thing as scientific evidence on the question.

Comment author: Kevin 12 June 2010 05:10:35AM 3 points [-]
Comment author: NancyLebovitz 11 June 2010 09:12:06AM 3 points [-]

Inspired by Chapter 24 of Methods of Rationality, but not a spoiler: If the evolution of human intelligence was driven by competition between humans, why aren't there a lot of intelligent species?

Comment author: Risto_Saarelma 11 June 2010 10:51:54AM 3 points [-]

Five-second guess: Human-level Machiavellian intelligence needs language facilities to co-evolve with, grunts and body language doesn't allow nearly as convoluted schemes. Evolving some precursor form of human-style language is the improbable part that other species haven't managed to pull off.

Comment author: taw 11 June 2010 12:52:42PM 1 point [-]

Somewhat accepted partial answer is that huge brains are ridiculously expensive - you need a lot of high energy density food (= fire), a lot of DHA (= fish) etc. Chimp diet simply couldn't support brains like ours (and aquatic ape etc.), nor could they spend as much time as us engaging in politics as they were too busy just getting food.

Perhaps chimp brains are as big as they could possibly be given their dietary constraints.

Comment author: NancyLebovitz 11 June 2010 02:22:58PM 1 point [-]

That's conceivable, and might also explain why wolves, crows, elephants, and other highly social animals aren't as smart as people.

Also, I think the original bit in Methods of Rationality overestimates how easy it is for new ideas to spread. As came up recently here, even if tacit knowledge can be explained, it usually isn't.

This means that if you figure out a better way to chip flint, you might not be able to explain it in words, and even if you can, you might chose to keep it as a family or tribal secret. Inventions could give their inventors an advantage for quite a long time.

Comment author: Rain 09 June 2010 07:51:57PM *  3 points [-]

I've recently begun downvoting comments that are at -2 rating regardless of my feelings about them. I instituted this policy after observing that a significant number of comments reach -2 but fail to be pushed over to -3, which I'm attributing to the threshold being too much of a psychological barrier for many people to penetrate; they don't want to be 'the one to push the button'. This is an extension of my RL policy of taking 'the last' of something laid out for communal use (coffee, donuts, cups, etc.). If the comment thread really needs to be visible, I expect others will vote it back up.

Edit: It's likely that most of the negative response to this comment centers around the phrase "regardless of my feelings about them." I now consider this to be too strong a statement with regards to my implemented actions. I do read the comment to make sure I don't consider it any good, and doubt I would perversely vote something down even if I wanted to see more of it.

Comment author: Morendil 09 June 2010 08:02:41PM 5 points [-]

I wish you wouldn't do that, and stuck instead with the generally approved norm of downvoting to mean "I'd prefer to see fewer comments like this" and upvoting "I'd like to see more like this".

You're deliberately participating in information cascades, and thereby undermining the filtering process. As an antidote, I recommend using the anti-kibitzer script (you can do that through your Preferences page).

Comment author: Rain 09 June 2010 08:05:39PM 1 point [-]

I wish you wouldn't do that, and stuck instead with the generally approved norm of downvoting to mean "I'd prefer to see fewer comments like this" and upvoting "I'd like to see more like this".

I disagree that that's the formula used for comments that exist within the range -2 to 2. Within that range, from what I've observed of voting patterns, it seems far more likely that the equation is related to what value the comment "should be at." If many people used anti-kibitzing, I doubt this would remain a problem.

Comment author: Vladimir_Nesov 09 June 2010 08:35:03PM *  2 points [-]

I believe your hypothesis and decision are possibly correct, but if they are, you should expect your downvotes to often be corrected upwards again. If this doesn't happen, then you are wrong and shouldn't apply this heuristic.

I disagree that that's the formula used for comments that exist within the range -2 to 2.

Morendil doesn't say it's what actually happens, he merely says it should happen this way, and that you in particular should behave this way.

Comment author: Rain 09 June 2010 08:48:55PM *  1 point [-]

I thought of doing this after reading the article Composting Fruitless Debates and making a voted-up suggestion to downvote below threshold.

I'm using it as an excuse to overcome my general laziness with regards to voting, which has the typical pattern of one vote (up or down) per hundreds of comments read.

Edit: And due to remembering Eliezer's comments about moderation.

Comment author: gwern 09 June 2010 05:36:33PM 3 points [-]

My recent comment on Reddit reminded me of WrongTomorrow.com - a site that was mentioned briefly here a while ago, but which I haven't seen much since.

Try it out, guys! LongBets and PredictionBook are good, but they're their own niche; LongBets won't help you with pundits who don't use it, and PredictionBook is aimed at personal use. If you want to track current pundits, WrongTomorrow seems like the best bet.

Comment author: Alexandros 07 June 2010 09:51:42AM *  3 points [-]

Let's get this thread going:

I'd like to ask everyone what probability bump they give to an idea given that some people believe it.

This is based on the fact that out of the humongous idea-space, some ideas are believed by (groups of) humans, and a subset of those are believed by humans and are true. (of course there exist some that are true and not yet believed by humans.)

So, given that some people believe X, what probability do you give for X being true, compared to Y which nobody currently believes?

Comment author: RobinZ 07 June 2010 12:27:59PM 6 points [-]

I'd like to ask everyone what probability bump they give to an idea given that some people believe it.

Usually fairly substantial - if someone presents me with two equally-unsupported claims X and Y and tells me that they believe X and not Y, I would give greater credence to X than to Y. Many times, however, that credence would not reach the level of ... well, credence, for various good reasons.

Comment author: MartinB 07 June 2010 01:24:50PM 3 points [-]

Depends on the person and the idea. I have some people whose recommendations I follow regardless, even if I estimate upfront that I will consider the idea wrong. There are different levels of wrongness, and it does not hurt to get good counterarguments. It also depends on the real life practicability of the idea. If it is for everyday things than common sense is a good starting prior. (Also there is a time and place to use the public joker on Who wants to be a millionaire.) If a group of professionals agree on something related to their profession it is also a good start. To systematize: if a group of people has a belief about something they have experience with, that that belief is worth looking at.

And then on further investigation it often turns out that there are systematic mistakes being made.

I was shocked to read in the book on checklists, that not only doctors often don't like them. But even financial companies, that can see how the usage ups their monetary gains. But finding flaws in a whole group does not imply that everything they say is wrong. It is good to see a doctor, even if he not using statistics right. He can refer you to a specialist, and treat all the common stuff right away. If you get a complicated disease you can often read up on it.

The obvious example to your question would be religion. It is widely believed, but probably wrong, yet I did not discard it right away, but spent years studying stuff till I decided there was nothing to it. There is nothing wrong in examining the ideas other people have.

Comment author: Torben 07 June 2010 05:14:44PM *  3 points [-]

Agreed.

As the OP states, idea space is humongous. The fact alone that people comprehend something sufficiently to say anything about it at all means that this something is a) noteworthy enough to be picked up by our evolutionarily derived faculties by even a bad rationalist b) expressible by same faculties c) not immediately, obviously wrong

To sum up, the fact that someone claims something is weak evidence that it's true, cf. Einstein's Arrogance. If this someone is Einstein, the evidence is not so weak.

Edit: just to clarify, I think this evidence is very weak, but evidence for the proposition, nonetheless. Dependent on the metric, by far most propositions must be "not even wrong", i.e. garbled, meaningless or absurd. The ratio of "true" to {"wrong" + "not even wrong"} seems to ineluctably be larger for propositions expressed by humans than for those not expressed, which is why someone uttering the proposition counts as evidence for it. People simply never claim that apples fall upwards, sideways, green, kjO30KJ&¤k etc.

Comment author: MartinB 07 June 2010 05:27:13PM 1 point [-]

I forgot the major influence of my own prior knowledge. (Which i guess holds true for everyone.) That makes the cases where I had a fixed opinion, and managed to change it all the more interesting. If you never dealt with an idea before you go where common sense or the experts lead you. But if you already have good knowledge, than public opinion should do nothing to your view. Public opinion or even experts (esp. when outside their field) often enough state opinions without comprehending the idea. So it doesnt really mean too much. Regarding Einstein, he made the statements before becoming super famous. I understand it as a case of signaling 'look over here!' And he is not particularly safe against errors. One of his last actions (which I have not fact checked sufficiently so far) was to write a foreword for a book debunking the movement of the continental plates.

Comment author: Torben 07 June 2010 06:47:07PM 1 point [-]

Regarding Einstein, he made the statements before becoming super famous. I understand it as a case of signaling 'look over here!' And he is not particularly safe against errors. One of his last actions (which I have not fact checked sufficiently so far) was to write a foreword for a book debunking the movement of the continental plates.

I didn't intend to portray Einstein as bulletproof, but rather highlight his reasoning. Plus point to the idea of even locating the idea in idea space. Obviously, creationism is wrong, but less wrong than a random string. It at least manages to identify a problem and using cause and effect.

Comment author: [deleted] 13 June 2010 12:37:11PM 1 point [-]

If no people believe Y -- literally no people -- then either the topic is very little examined by human beings, or it's very exhaustively examined and seems obvious to everyone. In the first case, I give a smaller probability than in the second case.

In the first case, only X believers exist because only X believers have yet considered the issue. That's minimal evidence in favor of X. In the second case, lots of people have heard of the issue; if there were a decent case against X, somebody would have thought of it. The fact that none of them -- not a minority, but none -- argued against X is strong evidence that X is true.

Comment author: Jack 08 June 2010 01:05:43AM 1 point [-]

I don't think belief has a consistent evidentiary strength since it depends on the testifier's credibility relative to my own. Children have much lower credibility than me on the issue of the existence of Santa. Professors of physics have much higher credibility that me on the issue of dimensions greater than four. Some person other than me has much higher credibility on the issue of how much money they are carrying. But I have more credibility than anyone else on the issue of how much money I'm carrying. I don't see any relation that could be described as baseline so the only answer is: context.

Comment author: AlanCrowe 07 June 2010 02:01:26PM *  1 point [-]

I've become increasingly disillusioned with people's capacity for abstract thought. Here are two points on my journey.

The public discussion of using wind turbines for carbon-free electricity generation seems to implicitly assume that electricity output goes as something like the square-root of windspeed. If the wind is only blowing half speed you still get something like 70% output. You won't see people saying this directly, but the general attitude is that you only need back up for the occasional calm day when the wind doesn't blow at all.

In fact output goes as the cube of windspeed. The energy in the windstream is one half m v squared, where m, the mass passing your turbine is proportional to the windspeed. If the wind is at half strength, you only get 1/8 output.

Well, that is physics. Ofcourse people suck at physics. Trouble is, the more I look at people's capacity for abstract thought the more problems I see. When people do a cost/benefit analysis they are terribly vague on whether they are suposed to add the costs and benefits or whether the costs get subtracted from the benefits. Even if they realise that they have to subtract they are still at risk of using an inverted scale for the costs and ending up effectively adding.

The probabiltiy bump I give to an idea just because some people believe it is zero. Equivantly my odds ratio is one. However you describe it, my posterior is just the same as my prior.

Comment author: Douglas_Knight 07 June 2010 07:09:45PM *  1 point [-]

When people do a cost/benefit analysis they are terribly vague on whether they are suposed to add the costs and benefits or whether the costs get subtracted from the benefits.

Revised: I do not think that link provides evidence for the quoted sentence. Nor I do see other evidence that people are that bad at cost-benefit analysis. I agree that the example presented there is interesting and that one should keep in mind that disagreements about values can be hidden, sometimes maliciously.

Comment author: AlanCrowe 01 July 2010 04:00:20PM 1 point [-]

I've got a better link. David Henderson catches a professor of economics getting costs and benefits confused in a published book. Henderson's review is on on page 54 of Regulation, and my viewer puts it on the ninth page of the pdf that Henderson links to

Comment author: Douglas_Knight 01 July 2010 06:14:12PM 0 points [-]

That is a good example. Talk of creating jobs as a benefit, rather than a cost is quite common. But is it confusion or malice? It is hard for me to imagine that economists would publish such a book without having it pointed out to them. The audience certainly is confused. Henderson says "Almost no one spending his own money makes this mistake" and would not generalize to people's capacity for abstract thought.

The original question was how much information to extract from the conventional wisdom. I do not take this as a reason to doubt the conventional wisdom about personal decisions. Partly, this is public choice, and partly because people do not address externalities in their personal decisions. Maybe any commonly accepted argument involving economics should be suspect, though the existence of the very well-established applause-line of "creating jobs" suggests that there are limits to how to fool people. But your claim was not that people are bad at physics and economics, but at the abstract thought of decision theory.

Comment author: xamdam 07 June 2010 01:24:49PM 1 point [-]

I think it largely depends on a) what the idea is and b) who believes it = and what their rationality skills are.

Comment author: MartinB 07 June 2010 01:28:15PM 1 point [-]

I recently learned the hard way, that one can easily be an idiot in one area, while being very competent in another. Religious scientists / programmers etc. Or lets say people that are highly competent in their area of occupation without looking into other things.

Comment author: Baughn 13 June 2010 08:17:32PM 2 points [-]

I found an interesting paper on Arxiv earlier today, by the name of Closed timelike curves via post-selection: theory and experimental demonstration.

It promises such lovely possibilities as quick solutions to NP-complete problems, and I'm not entirely sure the mechanism couldn't also be used to do arbitrary amounts of computation in finite time. Certainly worth a read.

However, I don't understand quantum mechanics well enough to tell how sane the paper is, or what the limits of what they've discovered are. I'm hoping one of you does.

Comment author: Rain 13 June 2010 02:35:25PM *  2 points [-]

Clippy-related: The Paper Clips Project is run by a school trying to overcome scope insensitivity by representing the eleven million people killed in the Holocaust with one paper clip per victim.

Comment author: Blueberry 13 June 2010 05:11:59PM 2 points [-]

From that Wikipedia article:

Inside the railcar, besides the paper clips, there are the Schroeders’ book and a suitcase filled with letters of apology to Anne Frank by a class of German schoolchildren.

Apologizing for ... being German? That's really bizarre.

Comment author: LucasSloan 13 June 2010 06:49:10PM 2 points [-]

Apologizing for ... being German? That's really bizarre.

Not really. Most cultures go funny in the head around the Holocaust. It is, for some reason, considered imperative that 10th graders in California spend more time being made to feel guilty about the Holocaust than learning about the actual politics of the Weimar Republic.

Comment author: Kevin 13 June 2010 02:37:11AM 2 points [-]

The number of heart attacks has fallen since England imposed a smoking ban

http://www.economist.com/node/16333351?story_id=16333351&fsrc=scn/tw/te/rss/pe

Comment author: cupholder 13 June 2010 03:22:46AM *  3 points [-]

I think I found the study they're talking about thanks to this article. I might take a look at it - if the methodology is literally just 'smoking was banned, then the heart attack rate dropped', that sucks.

(Edit to link to the full study and not the abstract.)


Just skimmed it. The methodology is better than that. They use a regression to adjust for the pre-existing downward trend in the heart attack hospital admission rate; they represent it as a linear trend, and that looks fair to me based on eyeballing the data in figures 1 and 2. They also adjust for week-to-week variation and temperature, and the study says its results are 'more modest' than others', and fit the predictions of someone else's mathematical model, which are fair sanity checks.

I still don't know how robust the study is - there might be some confounder they've overlooked that I don't know enough about smoking to think of - but it's at least not as bad as I expected. The authors say they want to do future work with a better data set that has data on whether patients are active smokers, to separate the effect of secondhand smoke from active smoking. Sounds interesting.

Comment author: Kevin 12 June 2010 09:08:54PM 2 points [-]

In the Singularity Movement, Humans Are So Yesterday (long Singularity article in this Sunday's NY Times; it isn't very good)

http://news.ycombinator.com/item?id=1426386

Comment author: JoshuaZ 12 June 2010 09:54:23PM *  2 points [-]

I agree that this article isn't very good. It seems to do the standard problem of combining a lot of different ideas about what the Singularity would entail. It emphasizes Kurzweil way too much, and includes Kurzweil's fairly dubious ideas about nutrition and health. The article also uses Andrew Orlowski as a serious critic of the Singularity making unsubstantiated claims about how the Singularity will only help the rich. Given that Orlowski's entire approach is to criticize anything remotely new or weird-seeming, I'm disappointed that the NYT would really use him as a serious critic in this context. The article strongly reinforces the perception that the Singularity is just a geek-religious thing. Overall, not well done at all.

Comment author: ata 12 June 2010 10:50:01PM *  9 points [-]

I'm starting to think SIAI might have to jettison the "singularity" terminology (for the intelligence explosion thesis) if it's going to stand on its own. It's a cool word, and it would be a shame to lose it, but it's become associated too much with utopian futurist storytelling for it to accurately describe what SIAI is actually working on.

Edit: Look at this Facebook group. This sort of thing is just embarrassing to be associated with. "If you are feeling brave, you can approach a stranger in the street and speak your message!" Seriously, this practically is religion. People should be raising awareness of singularity issues not as a prophecy but as a very serious and difficult research goal. It doesn't do any good to have people going around telling stories about the magical Future-Land while knowing nothing about existential risks or cognitive biases or friendly AI issues.

Comment author: JoshuaZ 14 June 2010 05:24:13AM *  2 points [-]

I'm not sure that your criticism completely holds water. Friendly AI is simply put only a worry that has convinced some Singularitarians. One might not be deeply concerned about that (Possible example reasons: 1) You expect uploading to come well before general AI. 2) you think that the probable technical path to AI will force a lot more stages of AI of much lower intelligence which will be likely to give us good data for solving the problem)

I agree that this Facebook group does look very much like something one would expect out of a missonizing religion. This section in particular looked like a caricature:

To raise awareness of the Singularity, which is expected to occur no later than the year 2045, we must reach out to everyone on the 1st day of every month.

At 20:45 hours (8:45pm) on the 1st day of each month we will send SINGULARITY MESSAGES to friends or strangers.

Example message:

"Nanobot revolution, AI aware, technological utopia: Singularity2045."

The certainty for 2045 is the most glaring aspect of this aside from the pseudo-missionary aspect. Also note that some of the people associated with this group are very prominent Singularitarians and Transhumanists. Aubrey de Grey is listed as an administrator.

But, one should remember that reversed stupidity is not intelligence. Moreover, there's a reason that missionaries sound like this: They have a very high confidence in their correctness. If one had a similarly high confidence in the probability of a Singularity event, and you thought that that event was more likely to occur safely if more people were aware of it, and was more likely to occur soon if more people were aware of it, and buy into something like the galactic colonization argument, and you believe that sending messages like this has a high chance of getting people to be aware and take you seriously then this is a reasonable course of action. Now, that's a lot of premises, some of which have likelyhoods others which have very low ones. Obviously there's a very low probability that sending out these sorts of messages is at all a net benefit. Indeed, I have to wonder if there's any deliberate mimicry of how religious groups send out messages or whether successfully reproducing memes naturally hit on a small set of methods of reproduction (but if that were the case I think they'd be more likely to hit an actually useful method of reproduction). And in fairness, they may just be using a general model for how one goes about raising awareness for a cause and how it matters. For some causes, simple, frequent appeals to emotion are likely an effective method (for example, making people aware of how common sexual assault is on college campuses, short messages that shock probably do a better job than lots of fairly dreary statistics). So then the primary mistake is just using the wrong model of how to communicate to people.

Comment author: NancyLebovitz 15 June 2010 10:54:50AM *  3 points [-]

Speaking of things to be worried about other than AI, I wonder if a biotech disaster is a more urgent problem, even if less comprehensive

Part of what I'm assuming is that developing a self-amplifying AI is so hard that biotech could be well-developed first.

While it doesn't seem likely to me that a bio-tech disaster could wipe out the human race, it could cause huge damage-- I'm imagining diseases aimed at monoculture crops, or plagues as the result of terrorism or incompetent experiments.

My other assumptions are that FAI research is dependent on a wealthy, secure society with a good bit of surplus wealth for individual projects, and is likely to be highly dependent on a small number of specific people for the forseeable future.

On the other hand, FAI is at least a relatively well-defined project. I'm not sure where you'd start to prevent biotech disasters.

Comment author: NihilCredo 12 April 2011 11:53:25PM 3 points [-]

On the other hand, FAI is at least a relatively well-defined project

That's one hell of a "relatively" you've got there!

Comment author: orthonormal 17 June 2010 03:51:39AM 0 points [-]

It's better than mainstream Singularity articles in the past, IMO; unfortunately, Kurzweil is seen as an authority, but at least it's written with some respect for the idea.

Comment author: Kevin 11 June 2010 08:31:32PM *  2 points [-]

Heuristics and biases in charity

http://www.sas.upenn.edu/~baron/papers/charity.pdf (I considered making this link as a top-level post.)

Comment author: Sniffnoy 10 June 2010 09:44:31PM 2 points [-]

Saw this over on Bruce Schneier's blog, it seemed worth reposting here. Wharton’s “Quake” Simulation Game Shows Why Humans Do Such A Poor Job Planning For & Learning From Catastrophes (link is to summary, not original article, as original article is a bit redundant). Not so sure how appropriate the "learning from" part of the title is, as they don't seem to mention people playing the game more than once, but still quite interesting.

Comment author: ciphergoth 10 June 2010 10:11:39AM 2 points [-]

What solution do people prefer to Pascal's Mugging? I know of three approaches:

1) Handing over the money is the right thing to do exactly as the calculation might indicate.

2) Debiasing against overconfidence shouldn't mean having any confidence in what others believe, but just reducing our own confidence; thus the expected gain if we're wrong is found by drawing from a broader reference class, like "offers from a stranger".

3) The calculation is correct, but we must pre-commit to not paying under such circumstances in order not to be gamed.

What have I left out?

Comment author: CarlShulman 10 June 2010 01:11:59PM 4 points [-]

The unbounded utility function (in some physical objects that can be tiled indefinitely) in Pascal's mugging gives infinite expected utility to all actions, and no reason to prefer handing over the money to any other action. People don't actually show the pattern of preferences implied by an unbounded utility function.

If we make the utility function a bounded function of happy lives (or other tilable physical structures) with a high bound, other possibilities will offer high expected utility. The Mugger is not the most credible way to get huge rewards (investing in our civilization on the chance that physics allows unlimited computation beats the Mugger). This will be the case no matter how huge we make the (finite) bound.

Comment author: ciphergoth 10 June 2010 11:04:51PM 1 point [-]

Bounding the utility function definitely solves the problem, but there are a couple of problems. One is the principle that the utility function is not up for grabs, the other is that a bounded utility function has some rather nasty consequences of the "leave one baby on the track" kind.

Comment author: CarlShulman 11 June 2010 03:27:59AM *  3 points [-]

One is the principle that the utility function is not up for grabs,

I don't buy this. Many people have inconsistent intuitions regarding aggregation, as with population ethics. Someone with such inconsistent preferences doesn't have a utility function to preserve.

Also note that a bounded utility function can allot some of the potential utility under the bound to producing an infinite amount of stuff, and that as a matter of psychological fact the human emotional response to stimuli can't scale indefinitely with bigger numbers.

And, of course, allowing unbounded growth of utility with some tilable physical process means that process can dominate the utility of any non-aggregative goods, e.g. the existence of at least some instantiations of art or knowledge, or overall properties of the world like ratios of very good to lives just barely worth living/creating (although you might claim that the value of the last scales with population size, many wouldn't characterize it that way).

Bounded utility functions seem to come much closer to letting you represent actual human concerns, or to represent more of them, in my view.

Comment author: RichardKennaway 10 June 2010 12:40:26PM *  2 points [-]

Eliezer's original article bases its argument on the use of Solomonoff induction. He even suggests up front what the problem with it is, although the comments don't make anything of it: SI is based solely on program length and ignores computational resources. The optimality theorems around SI depend on the same assumption. Therefore I suggest:

4. Pascal's Mugging is a refutation of the Solomonoff prior.

But where a computationally bounded agent, or an unbounded one that cares how much work it does, should get its priors from instead would require more thought than a few minutes on a lunchtime break.

Comment author: cupholder 10 June 2010 01:29:05PM *  1 point [-]

Tom_McCabe2 suggests generalizing EY's rebuttal of Pascal's Wager to Pascal's Mugging: it's not actually obvious that someone claiming they'll destroy 3^^^^3 people makes it more likely that 3^^^^3 people will die. The claim is arguably such weak evidence that it's still about equally likely that handing over the $5 will kill 3^^^^3 people, and if the two probabilities are sufficiently equal, they'll cancel out enough to make it not worth handing over the $5.

Personally, I always just figured that the probability of someone (a) threatening me with killing 3^^^^3 people, (b) having the ability to do so, and (c) not going ahead and killing the people anyway after I give them the $5, is going to be way less than 1/3^^^^3, so the expected utility of giving the mugger the $5 is almost certainly less than the $5 of utility I get by hanging on to it. In which case there is no problem to fix. EY claims that the Solomonoff-calculated probability of someone having 'magic powers from outside the Matrix' 'isn't anywhere near as small as 3^^^^3 is large,' but to me that just suggests that the Solomonoff calculation is too credulous.

(Edited to try and improve paraphrase of Tom_McCabe2.)

Comment author: ciphergoth 10 June 2010 11:07:32PM 1 point [-]

This seems very similar to the "reference class fallback" approach to confidence set out in point 2, but I prefer to explicitly refer to reference classes when setting out that approach, otherwise the exactly even odds you apply to massively positive and massively negative utility here seem to come rather conveniently out of a hat...

Comment author: Douglas_Knight 09 June 2010 07:41:12PM *  2 points [-]

Does countersignaling actually happen? Give me examples.

I think most claims of countersignaling are actually ordinary signaling, where the costly signal is foregoing another group and the trait being signaled is loyalty to the first group. Countersignaling is where foregoing the standard signal sends a stronger positive message of the same trait to the usual recipients.

Comment author: RobinZ 09 June 2010 08:22:46PM 3 points [-]

That article makes it sound like "countersignaling" is forgoing a mandated signal - like showing up at a formal-dress occasion in street clothes.

Alicorn made a post about the tactics of countersignaling a while back.

Comment author: Douglas_Knight 09 June 2010 08:57:12PM 1 point [-]

That article makes it sound like "countersignaling" is forgoing a mandated signal

I said "standard" because game theory doesn't talk about mandates, but that's pretty much what I said, isn't it? If you disagree with that usage, what do you think is right?

Incidentally, in von Neumann's model of poker, you should raise when you have a good hand or a poor hand, and check when you have a mediocre hand, which looks kind of like countersignaling. Of course, the information transference that yields the name "signal" is rather different. Also, I'm not interested in applications of game theory to hermetically sealed games.

Comment author: RobinZ 09 June 2010 09:10:13PM 1 point [-]

I guess I don't understand your question, then - countersignaling seems like a perfectly ordinary proper subset of signaling.

Comment author: Douglas_Knight 09 June 2010 09:19:25PM 1 point [-]

Yes, countersignaling is signaling. The question is about practice, not theory. Does countersignaling actually happen?

Comment author: SilasBarta 09 June 2010 09:00:40PM *  1 point [-]

I play randomly for the first several rounds, so as to destroy the entanglement between my bets, my face, and my hand.

Comment author: Kevin 09 June 2010 03:54:10AM 2 points [-]
Comment author: simplicio 08 June 2010 06:43:10AM 2 points [-]

Some clips on the dark-side epistemology of history done by Christian apologists by Robert M Price, who describes himself as a Christian Atheist.

Not sure how worthwhile Price is to listen to in general though.

Comment author: Alexandros 08 June 2010 10:20:14AM *  1 point [-]

Thanks for that, Price is a very knowledgeable New Testament scholar. Check out his interview at the commonsenseatheism podcast here, also covers his path to becoming a christian atheist.

Comment author: MartinB 07 June 2010 01:12:41PM *  2 points [-]

Because it was used somewhere I calculated my own weights worth in gold - it is about 3.5 million EUR. In silver you can get me for 50.000 EUR. The Mythbusters recently build a lead balloon and had it fly. Some proverb don't hold up to reality and/or engineering.

Comment author: [deleted] 13 June 2010 12:09:25AM 1 point [-]

Maybe this has been discussed before -- if so, please just answer with a link.

Has anyone considered the possibility that the only friendly AI may be one that commits suicide?

There's great diversity in human values, but all of them have in common that they take as given the limitations of Homo sapiens. In particular, the fact that each Homo sapiens has roughly equal physical and mental capacities to all other Homo sapiens. We have developed diverse systems of rules for interpersonal behavior, but all of them are built for dealing with groups of people like ourselves. (For instance, ideas like reciprocity only make sense if the things we can do to other people are similar to the things they can do to us.)

The decision function of a lone, far more powerful AI would not have this quality. So it would be very different from all human decision functions or principles. Maybe this difference should cause us to call it immoral.

Comment author: Mitchell_Porter 13 June 2010 07:52:03AM 4 points [-]

Do you ever have a day when you log on and it seems like everyone is "wrong on the Internet"? (For values of "everyone" equal to 3, on this occasion.) Robin Hanson and Katja Grace both have posts (on teenage angst, on population) where something just seems off, elusively wrong; and now SarahC suggests that "the only friendly AI may be one that commits suicide". Something about this conjunction of opinions seems obscurely portentous to me. Maybe it's just a know-thyself moment; there's some nascent opinion of my own that's going to crystallize in response.

Now that my special moment of sharing is out of the way... Sarah, is the friendly AI allowed to do just one act of good before it kills itself? Make a child smile, take a few pretty photos from orbit, save someone from dying, stop a war, invent cures for a few hundred diseases? I assume there is some integrity of internal logic behind this thought of yours, but it seems to be overlooking so much about reality that there has to be a significant cognitive disconnect at work here.

Comment author: [deleted] 13 June 2010 01:07:16PM 1 point [-]

I'm not necessarily arguing for this position as saying we need to address it. "Suicidal AI" is to the problem of constructing FAI as anarchism is to political theory; if you want to build something (an FAI, a good government) then, on the philosophical level, you have to at least take a stab at countering the argument that perhaps it is impossible to build it.

I'm working under the assumption that we don't really know at this point what "Friendly" means, otherwise there wouldn't be a problem to solve. We don't yet know what we want the AI to do.

What we do know about morality is that human beings practice it. So all our moral laws and intuitions are designed, in particular, for small, mortal creatures, living among other small, mortal creatures.

Egalitarianism, for example, only makes sense if "all men are created equal" is more or less a statement of fact. What should an egalitarian human make of a powerful AI? Is it a tyrant? Well, no, a tyrant is a human who behaves as if he's not equal to other humans; the AI simply isn't equal. Well, then, is the AI a good citizen? No, not really, because citizens treat each other on an equal footing...

The trouble here, I think, is that really all our notions of goodness are really "what is good for a human to do." Perhaps you could extend them to "what is good for a Klingon to do" -- but a lot of moral opinions are specifically about how to treat other people who are roughly equivalent to yourself. "Do unto others as you would have them do unto you." The kind of rules you'd set for an AI would be fundamentally different from our rules for ourselves and each other.

It would be as if a human had a special, obsessive concern and care for an ant farm. You can protect the ants from dying. But there are lots of things you can't do for the ants: be an ant's friend, respect an ant, keep up your end of a bargain with an ant, treat an ant as a brother...

I had a friend once who said, "If God existed, I would be his enemy." Couldn't someone have the same sentiment about an AI?

(As always, I may very well be wrong on the Internet.)

Comment author: Mitchell_Porter 14 June 2010 06:26:18AM 2 points [-]

You say, human values are made for agents of equal power; an AI would not be equal; so maybe the friendly thing to do is for it to delete itself. My question was, is it allowed to do just one or two positive things before it does this? I can also ask: if overwhelming power is the problem, can't it just reduce itself to human scale? And when you think about all the things that go wrong in the world every day, then it is obvious that there is plenty for a friendly superhuman agency to do. So the whole idea that the best thing it could do is delete itself or hobble itself looks extremely dubious. If your point was that we cannot hope to figure out what friendliness should actually be, and so we just shouldn't make superhuman agents, that would make more sense.

The comparison to government makes sense in that the power of a mature AI is imagined to be more like that of a state than that of a human individual. It is likely that once an AI had arrived at a stable conception of purpose, it would produce many, many other agents, of varying capability and lifespan, for the implementation of that purpose in the world. There might still be a central super-AI, or its progeny might operate in a completely distributed fashion. But everything would still have been determined by the initial purpose. If it was a purpose that cared nothing for life as we know it, then these derived agencies might just pave the earth and build a new machine ecology. If it was a purpose that placed a value on humans being there and living a certain sort of life, then some of them would spread out among us and interact with us accordingly. You could think of it in cultural terms: the AI sphere would have a culture, a value system, governing its interactions with us. Because of the radical contingency of programmed values, that culture might leave us alone, it might prod our affairs into taking a different shape, or it might act to swiftly and decisively transform human nature. All of these outcomes would appear to be possibilities.

Comment author: NancyLebovitz 13 June 2010 10:34:15AM 1 point [-]

It seems unlikely that an FAI would commit suicide if humans need to be protected from UAI, or if there are other threats that only an FAI could handle.

Comment author: roundsquare 09 June 2010 09:43:56AM 1 point [-]

A question about Bayesian reasoning:

I think one of the things that confused me the most about this is that Bayesian reasoning talks about probabilities. When I start with Pr(My Mom Is On The Phone) = 1/6, its very different from saying Pr(I roll a one on a fair die) = 1/6.

In the first case, my mom is either on the phone or not, but I'm just saying that I'm pretty sure she isn't. In the second, something may or may not happen, but its unlikely to happen.

Am I making any sense... or are they really the same thing and I'm over complicating?

Comment author: Maelin 09 June 2010 10:01:53AM *  5 points [-]

Remember, probabilities are not inherent facts of the universe, they are statements about how much you know. You don't have perfect knowledge of the universe, so when I ask, "Is your mum on the phone?" you don't have the guaranteed correct answer ready to go. You don't know with complete certainty.

But you do have some knowledge of the universe, gained through your earlier observations of seeing your mother on the phone occasionally. So rather than just saying "I have absolutely no idea in the slightest", you are able to say something more useful: "It's possible, but unlikely." Probabilities are simply a way to quantify and make precise our imperfect knowledge, so we can form more accurate expectations of the future, and they allow us to manage and update our beliefs in a more refined way through Bayes' Law.

Comment author: Oscar_Cunningham 09 June 2010 10:23:01AM 3 points [-]

The cases are different in the way that you describe, but the maths of the probability is the same in each case. If you have an unseen die under a cup, and a die that you are about to roll, then one is already determined and the other isn't, but you'd bet at the same odds for each one to come up a six.

Comment author: Alexandros 09 June 2010 11:08:07AM 2 points [-]

I think the difference is that one event is a statement about the present which is either presently true or not, and the other is a prediction. So you could illustrate the difference by using the following pairs: P(Mom on phone now) vs. P(Mom on phone tomorrow at 12:00am). In the dice case P(die just rolled but not yet examined is 1) vs. P(die I will roll will come out 1).

I do agree with Oscar though, the maths should be the same.

Comment author: Vladimir_M 09 June 2010 05:22:56PM 1 point [-]

You might be interested in this recent discussion, if you haven't seen it already:

http://lesswrong.com/lw/2ax/open_thread_june_2010/23fa

Comment author: Jack 09 June 2010 09:51:58AM 1 point [-]

It looks to me like your confusion with these examples just stems from the fact that one event is in the present and the other in the future. Are you still confused if you make it P(Mom will be on the phone at 4 PM tomorrow)= 1/6. Or conversely, you make it P(I rolled a one on the fair die that is now beneath this cup) =1/6

Comment author: Jack 09 June 2010 07:38:54AM 1 point [-]

We've talked about a book club before but did anyone ever actually succeed in starting one? Since it is summer now I figure a few more of us might have some free time. Are people actually interested?

Comment author: Morendil 09 June 2010 09:26:37AM 3 points [-]

I've been thinking about finally starting a Study Group thread, primarily with a focus on Jaynes and Pearl both of which I'm studying at the moment. It would probably make sense to expand it to other books including non-math books - though the set of active books should remain small.

Two things have been holding me back - for one, the IMO excessively blog-like nature of LW with the result that once a conversation has rolled off the front page it often tends to die off, and for another a fear of not having enough time and energy to devote to actually facilitating discussion.

Facilitation of some sort seems required: as I understand it a book club or study group entails asking a few participants to make a firm commitment to go through a chapter or a section at a time and report back, help each other out and so on.

Comment author: Jack 09 June 2010 10:20:25AM *  1 point [-]

Well those are actually exactly the two books I had in mind (though I think we should probably just start with one of them).

the IMO excessively blog-like nature of LW with the result that once a conversation has rolled off the front page it often tends to die off

Agreed. Two options

  1. A new top level post for every chapter (or perhaps every two chapters, whatever division is convenient). This was a little annoying when it was one person covering every chapter in Dennett's Consciousness explained but if a decent number of people were participating the book club (and if each new post was put up by the facilitator, explaining hard to understand concepts) they'd probably justify themselves.

  2. We start a dedicated wordpress or blogspot blog and give the facilitators posting powers.

I wouldn't at all mind posting to start discussion on some sections but I'm not the best person to be explaining the math if it gets confusing-- if that was part of your expectation of facilitation.

I was thinking a reading group for Jaynes would be have a better chance of success than Pearl-- the issues are more general, the math looks easier and the entire thing is online. But it sounds like you've looked at them more than I have, what are your thoughts? I guess what really matters is what people are interested in.

For those interested the Jaynes book can be found here and much of Pearl's book can be found here.

Comment author: RichardKennaway 09 June 2010 10:12:42AM 1 point [-]

Is there any existing off-the-shelf web software for setting up book-club-type discussions?

I don't want to make too much of the infrastructure issue, as what really makes a book club work is the commitment of its members and facilitators, but it would be convenient if there was a ready-made infrastructure available, like there is for blogging and mailing lists.

Maybe the LW blog+wiki software running on a separate domain (lesswrongbooks.com?) would be enough. Blog for current discussions, wiki for summaries of past discussions.

Comment author: Morendil 09 June 2010 11:50:49AM 2 points [-]

There's a risk that any amount of thinking about infrastructure could kill off what energy there is, and since there appears to be some energy at present, I would rather favor having the discussion about the book club in the book club thread. :)

IOW we can kick off the initiative locally and let it find a new venue if and when that becomes necessary. There also seems to be some sort of provisional consensus that it's not quite time yet to fragment the LW readership : the LW subreddit doesn't seem to have panned out.

It seems to me that Jaynes is definitely topical for LW, I wouldn't worry about discussions among people studying it becoming annoying to the rest of the community. There are many, many gems pertaining to rationality in each of the chapters I've read so far.

Comment author: Jonathan_Graehl 08 June 2010 01:50:56AM *  1 point [-]

Supposedly (actual study) milk reduces catechin level in bloodstream.

Other research says: "does not!"

Really hot (but not scalded) milk tastes fantastic to me, so I've often added it to tea. I don't really care much about the health benefits of tea per se; I'm mostly curious if anyone has additional evidence one way or the other.

The surest way to resolve the controversy is to replicate the studies until it's clear that some of them were sloppy, unlucky, or lies. But, short of that, should I speculate that perhaps some people are opposed to milk drinking in general, or that perhaps tea in the researchers' home country is/isn't primarily taken with milk? I'm always tempted to imagine most of the scientists having some ulterior motive or prior belief they're looking to confirm.

It would be cool if researchers sometimes (credibly) wrote: "we did this experiment hoping to show X, but instead, we found not X". Knowing under what goals research was really performed (and what went into its selection for publication) would be valuable, especially if plans (and statements of intent/goal) for experiments were published somewhere at the start of work, even for studies that are never completed or published.

Comment author: ellx 07 June 2010 11:59:57AM 1 point [-]

I'd like to hear what people think about calibrating how many ideas you voice versus how confident you are in their accuracy.

For lack of a better example, i recall eliezer saying that new open threads should be made quadanually, once per season, but this doesn't appear to be the optimum amount. Perhaps eliezer misjudged how much activity they would receive and how fast they would fill up or he has a different opinion on how full a thread has to be to make it time for a new thread, but for sake of the example lets assume that eliezer was wrong and that the current one or two threads per month is better than quadanually. Should eliezer have recalibrated his confidence on this and never said it because its chance of being right was too low or would lowering his confidence on ideas be counter productive and is it optimal for people to have confidence in the ideas that they voice even it causes them to say some things which aren't right.

I suppose this is of importance to me because I think I might be better off if i lowered how judgemental i am of people who say things which are wrong and also lowered how judgemental i am of the ideas i have because i might be putting too much weight on people voicing ideas which are wrong.

Comment author: MartinB 07 June 2010 01:36:19PM 3 points [-]

Being right on group effects is difficult.

Is there a consistent path for what LW wants to be? a) rationalist site filled up with meta topics and examples b) a) + detailed treats of some important topics c) open to everything as long as reason is used

and so on. I personally like and profit from the discussing of akrasia methods. But it might be detrimental to the main target of the site. Also I would very much like to see a cannon develop for knowledge that LWers generally agree upon including, but not limited to the topics I currently care about myself.

Voicing ideas depends on where you are. In social settings I more and more advice against it. Arguing/discussing is just not helpful. And if you are filled up with weird ideas then you get kicked out, which might be bad for other goals you have.

It would be great to have a place for any idea to be examined for right and wrong.

Comment author: GreenRoot 07 June 2010 02:12:48PM 1 point [-]

I would very much like to see a cannon develop for knowledge that LWers generally agree upon

LW is working on it, and you can help!

Comment author: RichardKennaway 07 June 2010 02:28:06PM 4 points [-]

I'd like to see a picture of this LW cannon!

Comment author: Yvain 07 June 2010 07:02:37PM 16 points [-]

I'd like to see a picture of this LW cannon!

Rather than waste time doing both your cannon request and Roko's Fallacyzilla request, I just combined them into one picture of the Less Wrong Cannon attacking Fallacyzilla.

...now someone take Photoshop away from me, please.

Comment author: JoshuaZ 07 June 2010 07:07:20PM 3 points [-]

What does Fallacyzilla have on its chest? It looks like it has "A -> B, ~B, therefore ~A" But that is valid logic. Am I misreading it or did you mean to put "A -> B, ~A, therefore ~B"? That would be actually wrong.

Comment author: Yvain 07 June 2010 07:13:23PM *  8 points [-]

I noticed that two seconds after I put it up and it's now corrected...er...incorrected. (Today I learned - my brain has that same annoying auto-correct function as Microsoft Word)

Comment author: JoshuaZ 07 June 2010 07:16:17PM 4 points [-]

There's a related XKCD. The mouse-over text is especially relevant.

Comment author: Morendil 07 June 2010 03:02:16PM 1 point [-]

To whoever downvoted the parent: please refrain from downvoting people who draw attention to other's mistakes in a gentle and humorous way.