Open Thread June 2010, Part 4

5 Post author: Will_Newsome 19 June 2010 04:34AM

This thread is for the discussion of Less Wrong topics that have not appeared in recent posts. If a discussion gets unwieldy, celebrate by turning it into a top-level post.

This thread brought to you by quantum immortality.

Comments (325)

Comment author: Kevin 19 June 2010 06:06:09AM *  5 points [-]

Strange occurrence in US South Carolina Democratic primary.

The only explanation, Mr. Rawl’s representatives told the committee, was faulty voting machines — not chance, name order on the ballot, or Republicans crossing over to vote for the weaker Democrat. With testimony dominated by talk of standard variances, preference theories and voting machine software, the hearing took on the spirit of a political science seminar.

The Washington Post profiled Alvin Greene last week

10 minute video interview with Greene

What happened here?

Wikipedia has a list of possible explanations.

Fivethirtyeight lists possible explanations and analysis.

Rawl and co presented five hours of testimony that the results could only be attributed to a problem with the voting machines.

What is your probability estimate for Alvin Greene's win in this election being legitimate (Greene getting lucky as a result of aggregate voter intent+indifference+confusion, as opposed to voting machine malfunction or some sort of active conspiracy)? What evidence do you need in order to update your estimate?

Comment deleted 19 June 2010 10:47:42AM *  [-]
Comment author: LucasSloan 20 June 2010 03:22:56AM *  2 points [-]

Surely the "voters aren't actually paying attention" hypothesis deserves more than 1% probability.

Comment author: h-H 20 June 2010 02:10:01PM 0 points [-]

that could fall under 'error of some kind'.

Comment author: wedrifid 19 June 2010 11:18:18AM 2 points [-]

Probability that this person would have a worse influence on the senate than a more standard politician: 5%.

Comment author: Larks 19 June 2010 11:49:05AM 0 points [-]

Conditional on their winning the election, presumably.

Comment author: wedrifid 19 June 2010 01:13:20PM 0 points [-]

I'm not sure that is technically necessary given the precise phrasing.

Comment author: Larks 19 June 2010 03:18:26PM 0 points [-]

Because, unless he is a politician, the sentence fails to make sense, because 'more standard politician' requires him to be one? If so, I think being selected as a candidate makes you a politician.

Comment author: wedrifid 19 June 2010 04:26:22PM 0 points [-]

It seems to make sense without any fancy interpretation.

Comment author: Kevin 19 June 2010 02:38:05PM 0 points [-]

I would give it lower than that, US Senators have surprisingly little power.

Comment author: Daniel_Burfoot 19 June 2010 02:45:33PM 0 points [-]

US Senators have surprisingly little power.

???

I mean, in the sense that the US government is like a massive Ouija board that is not really controlled by anyone, then sure. But the senators seem to have a particularly heavy hand on the board.

Comment author: Kevin 19 June 2010 02:50:00PM 0 points [-]

Sorry, I meant "influence", not "power".

Comment author: AlexMennen 19 June 2010 03:41:29PM 1 point [-]

That is not important when considering the probability that Alvin Greene would have a worse influence on the Senate than the average politician if he got elected. It is only important when considering the probability that he would have a much worse influence on the Senate than average.

Comment author: jimrandomh 19 June 2010 01:08:13PM *  3 points [-]

Here is my probability distribution:

  • Voters actually voted for him: 0.1
  • Someone tampered with the voting machines or memory cards to make Alvin Greene win: 0.4
  • ...and that person did it because they wanted Alvin Greene to win: 0.1
  • ...and that person did it for kicks: 0.1
  • ...and that person did it because they wanted to expose the insecure voting machines: 0.2
  • Someone meant to tamper with a different election on the same ballot, but accidentally altered the democratic primary additionally or instead: 0.1
  • The votes were altered by leftover malware from a previous election which was also hacked: 0.2
  • There was a legitimate error in setting up or managing the voting machines altered the vote totals: 0.2

Note that I started researching this topic with an atypically high prior probability for voting machine fraud, and believe that it is very likely that major US elections in the past were altered this way. The strongest direct evidence I see for fraud having occurred is that there were "three counties with more votes cast in Republican governor's race than reported turnout in the Republican primary" FiveThirtyEight. Note that this means botched vote fraud, not correctly-implemented vote fraud, since correctly implemented vote fraud, using a strategy such as the Hursti hack, would have changed the votes but not the turnout numbers.

The Benford's Law analysis on FiveThirtyEight, on the other hand, I find very unconvincing - first because it has a low p-value, and second because it doesn't represent the way voting machine fraud really works; it can only detect if someone makes up vote totals from scratch, rather than adding to or subtracting from real vote totals.

Comment author: SilasBarta 19 June 2010 01:23:30PM *  8 points [-]

Not ready to answer the rationalist questions, but why is it that, as soon as elections don't go toward someone who played the standard political game, suddenly, "it must be a mistake somehow"? You guys set the terms of the primaries, you pick the voting machines. If you're not ready to trust them before the election, the time to contest them was back then, not when you don't like the result.

Where was Rawl on the important issue of voting machine reliability when they did "what they're supposed to"?

I understand that elections are evidence, and given the prior on Greene, this particular election may be insufficient to justify a posterior that Greene has the most "support", however defined. But elections also serve as a bright line to settle an issue. We could argue forever about who "really" has the most votes, but eventually we have to say who won, and elections are just as much about finality on that issue as they are as an evidential test of fact.

To an extent, then, it doesn't matter that Greene didn't "really" get the most votes. If you allow every election to be indefinitely contested until you're convinced there's no reason the loser really should have won, elections never settle anything. The price for indifference to voting procedure reliability (in this case, the machines) should be acceptance of a bad outcome for that time, to be corrected for the next election, or through the recall process.

Frankly, if Greene had lost but could present evidence of the strength Rawl presented, we wouldn't even be having this conversation.

ETA: Oh, and you gotta love this:

On election night, I was among the first reporters to speak with Greene after his victory was announced. His verbal tics and strange affect were immediately apparent: he frequently repeats and interrupts himself, speaks haltingly, and sometimes descends into incoherent rambling, as subsequent video and audio interviews have made all the more obvious.

Damn those candidates with autism symptoms! Only manipulative people like us deserve to win elections!

Comment author: jimrandomh 19 June 2010 01:27:57PM *  6 points [-]

I should point out that most of the people who ought to know about the issue, have been screaming bloody murder about electronic voting machines for some time now. Politicians and the general public just haven't been listening. This issue is surfacing now, not because it wasn't an issue before, but because having a specific election to point to makes it easier to get people to listen. It also helps that the election wasn't an important one (it was a Democratic primary for a safe Republican seat), and the candidates involved don't have the resources to influence the discussion like they normally would.

Comment author: Blueberry 19 June 2010 04:23:18PM *  1 point [-]

But elections also serve as a bright line to settle an issue. We could argue forever about who "really" has the most votes, but eventually we have to say who won, and elections are just as much about finality on that issue as they are as an evidential test of fact.

To an extent, then, it doesn't matter that Greene didn't "really" get the most votes. If you allow every election to be indefinitely contested until you're convinced there's no reason the loser really should have won, elections never settle anything.

Yes. Exactly. This is true for lawsuits as well: getting a final answer is more important than getting the "right" answer, which is why finality is an important judicial value that courts balance.

Comment author: JoshuaZ 19 June 2010 04:35:59PM 6 points [-]

This doesn't sound like autism to me. It sounds more like a neurotypical individual who is dealing with a very unexpected and stressful set of events and having to talk about them.

Comment author: SilasBarta 19 June 2010 11:35:33PM *  3 points [-]

Be that as it may, those are typical characteristics of high-functioning autistics, and I'm more than a little bothered that they view this as justification for reversing his victory.

Take the part I bolded and remove the "incoherent rambling" bit, and you could be describing me! Well, at least my normal mode of speech without deliberate self-adjustment.

And my lack of incoherent rambling is a judgment call ;-)

Comment author: wedrifid 19 June 2010 04:42:57PM 3 points [-]

Damn those candidates with autism symptoms!

Well... knowing that someone is autistic is some inferential evidence in favor of them being a good hacker.

Comment author: Morendil 19 June 2010 01:30:00PM 5 points [-]

My most likely explanations would be 1) software bug(s) 2) voter whim or confusion 3) odd hypothesis no one has thought of yet. Active intent to steal the nomination a distant fourth. Make it 60/30 among the first two.

Evidence? Well, anything credible, but how likely is that. :)

Comment author: JoshuaZ 19 June 2010 01:49:46PM *  2 points [-]

I put a very high probability that some form of tampering occurred primarily due to the failure of the data to obey a generalized Benford's law. Although a large amount of noise has been made about the the fact that some counties had more votes cast in the Republic governor's race than reported turnout, I don't see that as strong evidence of fraud since turnout levels in local elections are often based on the counting ability of the election volunteers who often aren't very competent.

I'd give probability estimates very similar to those of Jim's but with a slightly higher percentage for people actually voting for him. I'd do that I think by moving most of the probability mass from the idea of someone tampering with the election to expose the insecure voting machines which implies a very strange set of ethical thought processes. I've also had enough experience in local elections to know that sometimes very weird things happen for reasons that no one can explain (and that this occurs even with systems that are difficult to tamper with). So using the primary breakdown given by Jim I'd put it as follows:

* Voters actually voted for him: 0.25
* Someone tampered with the voting machines or memory cards to make Alvin Greene win: 0.25
* ...and that person did it because they wanted Alvin Greene to win: 0.1
* ...and that person did it for kicks: 0.1
* ...and that person did it because they wanted to expose the insecure voting machines: 0.05
* Someone meant to tamper with a different election on the same ballot, but accidentally altered the democratic primary additionally or instead: 0.1
* The votes were altered by leftover malware from a previous election which was also hacked: 0.2
* There was a legitimate error in setting up or managing the voting machines altered the vote totals: 0.2

Edit: Thinking this through another possibility that should be listed is deliberate Republican cross-over (since it is an open primary) but given the evidence that seems of negligible probability at this point (< .01)).

Comment author: jimrandomh 19 June 2010 01:59:04PM 0 points [-]

Edit: Thinking this through another possibility that should be listed is deliberate Republican cross-over (since it is an open primary) but given the evidence that seems of negligible probability at this point (< .01)).

I would count that under "voters actually voted for him"

Comment author: JoshuaZ 19 June 2010 02:06:00PM 1 point [-]

Ok. Yeah, so that should probably be a subcategory of that in that it explains the weird results in a sensible fashion.

Comment author: prase 19 June 2010 04:41:00PM *  2 points [-]

I don't know the details about the American voting system, but (or maybe therefore) I am surprised how low estimates all people give to the possibility that the result is genuine. My estimate (without much research, I've just read the links) is

  • 0.5 voters actually voted for Greene
  • 0.3 error of some kind
  • 0.2 conspiracy

In order to update, any evidence is accepted, of course. What I would most like to see: results of some statistical survey, conducted either before or better after the election, historical data concerning performance of black candidates, historical data from elections with big difference between the intensity of the campaign between the competing candidates, a lot of independent testimonies of trustworthy voters reporting non-standard behaviour of the voting machines, description of how can the results be altered (and what is normally done to avoid that).

Comment author: wedrifid 19 June 2010 04:46:07PM 1 point [-]

What evidence do you need in order to update your estimate?

The next election being won by a ficus would boost my estimate. Or, you know, something else ridiculous like an action hero actor.

Comment author: LucasSloan 20 June 2010 03:20:59AM *  2 points [-]

an action hero actor.

Why is this at all ridiculous? Is there any reason to believe Arnold Schwarzenegger has done a significantly worse job than other governors, controlling for ability of the legislature to agree on anything and the health of the economy?

Comment author: wedrifid 20 June 2010 03:41:16AM *  3 points [-]

Why is this at all ridiculous?

It merely serves to illustrate what politics is really about. It certainly isn't about voting for people who are the best suited for making and implementing the decisions that are best for the country, planet or species. I actually would have voted for him unless he had a particularly remarkable opponent. All else being equal I take a contribution in another field that is popular and that I appreciate is a more important signal to me than success as a pure courtier. It is unfortunate that I do not have reason to consider consider political popularity as a stronger signal of country-leading competence than creating 'kindergarten cop'.

Is there any reason to believe Arnold Schwarzenegger has done a significantly poorer job than other governors, controlling for ability of the legislature to agree on anything and the health of the economy?

I've already assigned a low probability to Alvin being at all worse than the alternatives. I expect Arnold would be 'even' better.

(Oh, and I do think that one liner is sub par. It would be better to stick to actual ridiculous rather than superficially ridiculous.)

Comment author: Liron 19 June 2010 07:17:09PM 1 point [-]

I think voters were clueless about both candidates, but they like to fill in all the boxes on the ballot, so they chose the name that has the higher positive affect by far: "Alvin Greene".

To me that would be sufficient to explain the entire anomaly, if not for the mysterious origin of Greene's $10,000 filing fee.

Comment author: Kevin 20 June 2010 01:05:34AM 3 points [-]

Also the possible "Al Green" effect -- voters may have thought they were voting for the famous soul singer.

Comment author: ideclarecrockerrules 19 June 2010 07:58:44PM *  4 points [-]

Here is some javascript to help follow LW comments. It only works if your browser supports offline storage. You can check that here.

To use it, follow the pastebin link, select all that text and make a bookmark out of it. Then, when reading a LW page, just click the bookmark. Unread comments will be highlighted, and you can jump to next unread comment by clicking on that new thing in the top left corner. The script looks up every (new) comment on the page and stores its ID in the local database.

Edit: to be more specific, all comments are marked as read as soon as the script is run. I could come up with a version that only marks them as read once you click that thing in upper left corner. Let me know if you're using it or if you'd like anything changed/added.

Comment author: W-Shadow 19 June 2010 11:13:24PM 1 point [-]

I made a similar Greasemonkey script some time ago.

Comment author: multifoliaterose 19 June 2010 08:11:03PM 4 points [-]

I remember a post by Eliezer in which he was talking about how a lot of people who believe in evolution are actually exhibiting the same thinking styles that creationists use when they justify their belief in evolution (using buzz words like "evidence" and "natural selection" without having a deep understanding of what they're talking about, having Guessed the Teacher's Password ). I can't remember what this post was called - does anybody remember? I remember it being good and wanted to refer people to it.

Comment author: wedrifid 19 June 2010 08:36:38PM 3 points [-]

I don't remember a post by Eliezer on the subject but it is oh so true. I often feel a 'cringe' reaction when I hear 'evidence' being used as religious symbol. It is the same cringe reaction I get when I hear people say "God Says" on something that I know isn't even covered in their bible. In both cases something BAD is going on that has nothing to do with whether or not there is a God.

Comment author: Vladimir_M 19 June 2010 11:01:24PM *  7 points [-]

I remember reading a post titled "Science as Attire," which struck me as making a very good point along these lines. It could be what you're looking for.

As a related point, it seems to me that people who do understand evolution (and generally have a strong background in math and natural sciences) are on average heavily biased in their treatment of creationism, in at least two important ways. First, as per the point made in the above linked post, they don't stop to think that the great majority of folks who do believe in evolution don't actually have any better understanding of it than creationists. (In fact, I would say that the best informed creationists I've read, despite the biases that lead them towards their ultimate conclusions, have a much better understanding of evolution than, say, a typical journalist who will attack them as ignorant.) Second, they tend to way overestimate the significance of the phenomenon. Honestly, if I were to write down a list of widespread delusions sorted by the practical dangers they pose, creationism probably wouldn't make the top fifty.

Comment author: multifoliaterose 20 June 2010 01:37:22AM 2 points [-]

Thanks, this is what I had in mind.

Comment author: JoshuaZ 20 June 2010 02:24:18AM 3 points [-]

If you said that it wouldn't make the top 10, I'd find that not implausible. Claiming it wouldn't make the top 50 seems implausible. Actual dangers posed by creationism:1) It makes people have a general more anti-science attitude and makes children less likely to become scientists 2) it takes up large sets of resources that would be spent usefully otherwise 3) it actively includes the spreading of a lot of misinformation 4) it snags otherwise bright minds who might otherwise becomes productive individuals (Jonathan Sarfati for example is a chess master, unambiguously quite bright, and had multiple good scientific papers before getting roped into YECism. Michael Behe is in a similar situation although for ID rather than young earth creationism). 5) The young earth variants encourage a narrow time outlook which is not helpful for long-term planning about the world or appreciation of serious existential threats (although honestly so few people pay attention to existential risks this is probably a minor issue) 6) It causes actual scientists and teachers to lose their jobs or have their work restricted (admittedly this isn't common but that's partially because creationism doesn't have much ground). 7) It encourages general extremist religious attitudes.

So not in the top 10? I'd agree with that. But I have trouble seeing it not in the top 50 most dangerous widespread delusions.

Comment author: Mass_Driver 20 June 2010 02:35:45AM 6 points [-]

I'm extremely curious to hear both your list and JoshuaZ's list of the top 20 or so most harmful delusions. Feel free to sort by category (1-4, 5-10, 11-20, etc.) rather than rank in individual order.

Comment author: LucasSloan 20 June 2010 03:18:15AM 3 points [-]

If I might jump in on the listing of delusions, I think that perhaps one of the most important things to understand about widespread delusions is who, in fact, holds them. A bunch of rednecks in Louisiana not believing in evolution isn't important, because even if they did, it wouldn't inform other parts of their worldview. In general, the specific delusions of ordinary people (IQ < 120) aren't important, because they aren't the ones who are actually affecting anything. Even improving the rationality and general problem awareness of smart people (120 < IQ < 135) doesn't really help, because then you get people who will expend enormous effort doing things like evangelizing atheism to the ordinary people and fighting global warming and the like. Raising the sanity waterline is important, but effort should be focused on people with the ability to actually use true beliefs.

Comment author: Mass_Driver 20 June 2010 03:52:53AM 3 points [-]

I sincerely hope that you are using IQ as only the crudest shorthand for "ability to actually use true beliefs," but your point in general is very well taken. Please do jump in if you have a listing of the most harmful delusions. :-)

Comment author: wedrifid 20 June 2010 03:58:11AM 2 points [-]

IQ >= 120 is a fairly low bar. IQ is also a strong indicator for the potential for someone's behavior to be influenced by delusions (rather than near mode thinking + social pressure being the dominant adaptation.)

Comment author: Mass_Driver 20 June 2010 04:03:51AM 2 points [-]

Do you mean do say that people of ordinary intelligence, as a general rule, don't actually believe whatever it is they say they believe, but instead just parrot what those around them say? You might be right. I think I need to find a way to re-immerse myself in a crowd of people of average intelligence; it's been far too long, and my predictive/descriptive powers for such people are fraying.

Note that none of this is sarcasm; this comment is entirely sincere.

Comment author: LucasSloan 20 June 2010 04:07:14AM 0 points [-]

I would say that people of ordinary intelligence don't actually have anything that I would identify as a non-trivial belief. They might say they believe in god, but they don't actually expect to get the pony they prayed for (even if they say that they do). However, they do have accurate beliefs regarding, say, how to cook food, or whether jumping off a building is a healthy idea, because they actually have to use such beliefs.

Comment author: wedrifid 20 June 2010 04:21:34AM 2 points [-]

I differentiate between 'actually believe' and 'act as if they are an agent with the belief that'. All people mostly do the latter but high IQ people are somewhat more likely to let 'actual beliefs' interfere with their lives.

Comment author: Douglas_Knight 20 June 2010 04:24:15AM 4 points [-]

Wedrifid only said "potential"; most people, smart or not, behave as you say. And I would expand "delusion" to 'belief": being smart is correlated with being influenced by beliefs, true or false.

That people act on beliefs or have at all coherent world-views is the most dangerous widespread delusion. ("The world is mad.") Immersing yourself in a crowd of average intelligence might help you see this, but I rather doubt that your associates act on their beliefs.

Comment author: wedrifid 20 June 2010 04:34:13AM 3 points [-]

Another thing that is dangerous is the people that actually act on their beliefs. They are much harder to control. People 'acting as if' pragmatically don't do things that we strongly socially penalize.

Comment author: Mass_Driver 20 June 2010 04:44:37AM 1 point [-]

I rather doubt that your associates act on their beliefs.

Not on their stated beliefs; surely, but don't most people have a set of actual beliefs? Can't these actual beliefs, at least in some contexts, be nudged so as to influence the level and direction of cognitive dissonance, which in turn can influence actions?

Comment author: JoshuaZ 20 June 2010 04:28:01AM *  3 points [-]

There's certainly evidence that intelligent people are more likely to have more coherent worldviews. For example, the GSS data shows that higher vocabulary is associated with more extreme political views to either end of the traditional political spectrum. There's similar research for IQ scores but I don't have a citation for that.

Comment author: Mass_Driver 20 June 2010 04:42:57AM 1 point [-]

There's certainly evidence that intelligence people

You really should watch your grammar, syntax, and spelling while commenting on intelligence. The irony is distracting, otherwise. Unless you were referring to the CIA and FBI?

Comment author: JoshuaZ 20 June 2010 04:50:32AM 1 point [-]

It might be more generally a sign that I shouldn't comment when it is late at night in my timezone. Also, it should constitute evidence that we need better spellcheckers that don't just catch non-words but also words that are clearly wrong from minimal context (although in this particular case catching that that was the wrong word would almost seem to require solving the natural language problem unless one had very good statistical methods).

Comment author: Blueberry 20 June 2010 06:54:54AM 1 point [-]

Are you saying more extreme political views are more coherent? I'm not following this.

Comment author: LucasSloan 20 June 2010 06:59:48AM 2 points [-]

Typically, yes. People with extreme views typically don't fail to make inferences from their beliefs along the lines of "X is good, so doing Y, which creates even more of X's goodness, would be even better!" Y might in fact be utterly stupid and evil and wrong, and a moderate with less extreme views might be against it, but the moderate and the extremist might both agree with X, even though the failure of logic that leads the extremist to endorse the evil Y is the belief that X is good.

Comment author: Vladimir_M 20 June 2010 09:58:39AM *  5 points [-]

Blueberry:

Are you saying more extreme political views are more coherent?

That seems like an almost self-evident observation to me. I have never seen anyone state clearly any political or ideological principles, of whatever sort and from whatever position, whose straightforward application wouldn't lead to positions that are utterly extremist by the standards of the present centrist opinion.

Getting people with regular respectable opinions to contradict themselves by asking a few Socratic questions is a trivial exercise (though not one that's likely to endear you to them!). The same is not necessarily true for certain extremist positions.

Comment author: cupholder 20 June 2010 07:00:26AM *  0 points [-]

Do more extreme political views signify more coherent worldviews?

Comment author: cupholder 20 June 2010 07:12:48AM 3 points [-]

In general, the specific delusions of ordinary people (IQ < 120) aren't important, because they aren't the ones who are actually affecting anything.

I'm less sure. I would have thought that they affect things indirectly at least through social transmission of beliefs, what they choose to spend their money on, and the demands they make of politicians.

Even improving the rationality and general problem awareness of smart people (120 < IQ < 135) doesn't really help, because then you get people who will expend enormous effort doing things like evangelizing atheism to the ordinary people and fighting global warming and the like.

Arguably, one should expect it to help less than improving the rationality and awareness of people with IQ < 120, just because there are 11 times as many people with IQ < 120 than there are with 120 < IQ < 135.

Comment author: PhilGoetz 24 June 2010 03:15:02AM 0 points [-]

In a democracy, specific delusions of ordinary people are important.

Comment author: LucasSloan 25 June 2010 05:52:07AM 1 point [-]

In a representative democracy, the specific delusions of the elected and unelected officials are important.

Comment author: wedrifid 20 June 2010 03:52:35AM 2 points [-]
  1. The creation of an FAI is not the most important thing the species could be doing.
  2. The best way to create an FAI is not...
Comment author: LucasSloan 20 June 2010 04:03:43AM 2 points [-]

Taking into account what I already said about needing to influence people who can actually use beliefs (thus controlling for things like atheism, evolution, etc.)...

  1. FAI and related.
  2. Inability to do math.
  3. Failures around believing the state of the world is good (thinking aging is a good thing and the like).
  4. Believing that politics is the best way to influence the world.
Comment author: wedrifid 20 June 2010 04:36:24AM 2 points [-]

Inability to do math? Really? Are you talking 'disinclination to shut up and multiply' or actual ability to do math?

I love math but don't really think most people need it.

Comment author: PhilGoetz 20 June 2010 04:47:17AM 2 points [-]

Most people don't need to understand evolution. Maybe we should distinguish between "harmful to self", "harmful to society", and "harmful to a democratic society".

If you can't do math at a fairly advanced level - at least having competence with information theory, probability, statistics, and calculus - you can't understand the world beyond what's visible on its (metaphorical) surface.

Comment author: JoshuaZ 20 June 2010 04:59:48AM *  7 points [-]

If you can't do math at a fairly advanced level - at least having competence with information theory, probability, statistics, and calculus - you can't understand the world beyond what's visible on its (metaphorical) surface.

While as a mathematician I find that claim touching, I can't really agree with it. To use the example that was one of the starting points of this conversation, how much math do you need to understand evolution? Sure, if you want to really understand the modern synthesis in detail you need math. And if you want to make specific predictions about what will happen to allele frequencies you'll need math. But in those cases it is very basic probability and maybe a tiny bit of calculus (and even then, more often than not you can use the formulas without actually knowing why they work beyond a very rough idea).

Similar remarks apply to other areas. I don't need a deep understanding of any of those subjects to have a basic idea about atoms, although again I will need some of them if I want to actually make useful predictions (say for Brownian motion).

Similarly, I don't need any of those subjects to understand the Keplerian model of orbits, and I'll only need one of those four (calculus) if I want to make more precise estimations for orbits (using Newtonian laws).

The amount of actual math needed to understand the physical world is pretty minimal unless one is doing hard core physics or chemistry.

Comment author: wedrifid 20 June 2010 08:20:29AM 1 point [-]

The amount of actual math needed to understand the physical world is pretty minimal unless one is doing hard core physics or chemistry.

For example... trying to work out what happens when I shoot a neverending stream of electrons at a black hole. The related theories were more or less incomprehensible to me at first glance. Not being able to do off the wall theorizing on everything at the drop of the hat has to at least make 49!

Comment author: PhilGoetz 24 June 2010 03:42:02AM 0 points [-]

The human-scale physical world is relatively easy to understand, and we may have evolved or learned to perform the trickier computations using specialized modules, such as perhaps recognizing parabolas to predict where a thrown object will land. You get far with linear models, for instance, assuming that the distance something will move is proportional to the force that you hit it with, or that the damage done is proportional to the size of the rock you hit something with. You rarely come across any trajectory where the second derivative changes sign.

The social world, the economic world, ecology, game theory, predicting the future, and politics are harder to understand. There are a lot of nonlinear and even non-differentiable interactions. To understanding a phenomenon qualitatively, it's helpful to perform a stability analysis, and recognize likely stable areas, and also unstable regions where you have phase transitions, period doublings, and other catastrophes, You usually can't do the math and solve one of these systems; but if you've worked with a lot of toy systems mathematically, you'll understand the kind of behaviors you might see, and know something about how the number of variables and the correlations between them affect the limits of linear extrapolation. So you won't assume that a global warming rate of .017C/year will lead to a temperature increase of 1.7C in 100 years.

I'm making this up as I go; I don't have any good evidence at hand. I have the impression that I use math a lot to understand the world (but not the "physical" world of kinematics). I haven't observed myself and counted how often it happens.

Comment author: Will_Newsome 20 June 2010 06:48:48AM 2 points [-]

I'd like this to be true, as I want the time I spend learning math in the future to be as useful as you say, but I seem to have come rather far by knowing the superficial version of a lot of things. Knowing the actual math from something like PT:LOS would be great, and I plan on reaching at least that level in the Bayesian conspiracy, but I can currently talk about things like quantum physics and UDT and speed priors and turn this into changes in expected anticipation. I don't know what Kolmogorov complexity is, really, in a strictly formal from-the-axioms sense, nor Solomonoff induction, but I reference it or things related to it about 10 times a day in conversations at SIAI house, and people who know a lot more than I do mostly don't laugh at my postulations. Perhaps you mean a deeper level of understanding? I'd like to achieve that, but my current level seems to be doing me well. Perhaps I'm an outlier. (I flunked out of high school calculus and 'Algebra 2' and haven't learned any math since. I know the Wikipedia/Scholarpedia versions of a whole bunch of things, including information theory, computer science, algorithmic probability, set theory, etc., but I gloss over the fancy Greek letters and weird symbols and pretend I know the terms anyway.)

Comment author: LucasSloan 20 June 2010 06:55:01AM 2 points [-]

I'm pretty sure that most people around lesswrong have about the same level of familiarity with most subjects (outside whatever field they actually specialize). Although I do think that you are relatively weak in mathematics, but advanced math just really isn't that important, vis-a-vis being generally well educated and rational.

Comment author: Will_Newsome 20 June 2010 07:34:33AM 3 points [-]

A public reminder to myself so as to make use of consistency pressure: I shouldn't write comments like the one I wrote above. It lingers too long on a specific argument that is not particularly strong and was probably subconsciously fueled by a desire to talk about myself and perhaps countersignal to someone whose writing I respect (Phil Goetz).

Comment author: CronoDAS 21 June 2010 09:22:11AM 3 points [-]

I flunked out of high school calculus and 'Algebra 2' and haven't learned any math since.

I have a belief that I can fix things like this, having spent time working with other students in high school. If I ever meet you in person, will you assist me in testing that belief? ;)

Comment author: sketerpot 21 June 2010 02:34:04AM *  10 points [-]

I've got a tangential question: what math, if learned by more people, would give the biggest improvement in understanding for the effort put into learning it?

Take calculus, for example. It's great stuff if you want to talk about rates of change, or understand anything involving physics. There's the benefit; how about the cost? Most people who learn it have a very hard time doing so, and they're already well above average in mathematical ability. So, the benefit mostly relates to understanding physics, and the cost is fairly high for most people.

Compare this with learning basic probability and statistical thinking. I'm not necessarily talking about learning anything in depth, but people should have at least some exposure to ideas like probability distributions, variance, normal distributions and how they arise, and basic design of experiments -- blinding, controlling for variables, and so on. This should be a lot easier to learn than calculus, and it would give insight into things that apply to more people.

I'll give a concrete example: racism. Typical racist statements, like "black people are lazy and untrustworthy," couldn't possibly be true in more than a statistical sense, and obviously a statistical statement about a large group doesn't apply to every member of that group -- there's plenty of variance to take into account. Basic statistical thinking makes racist bigotry sound preposterously silly, like someone claiming that the earth is flat. This also applies to every other form of irrational bigotry that I can think of off the top of my head.

Remember when Larry Summers suggested that maybe part of the reason for the underrepresentation of women in Harvard's science faculty was that women may have lower variance in intelligence than men, and so are underrepresented in the highest part of the intelligence bell curve? What almost everybody heard was "Women can't be scientists because they're stupid." People heard a statistical statement and had no idea how to understand it.

There are important, relevant subjects that people just can not understand without basic statistical thinking. I would like to see most people exposed to basic statistical thinking.

Are there any other kinds of math that offer high bang-for-the-buck, as far as learning difficulty goes? (I've always thought that the math behind computer programming was damn useful stuff, but the engineering students I've talked with usually find it harder than calculus, so maybe that's not the best idea.)

Comment author: taiyo 21 June 2010 03:54:44AM 2 points [-]

Probability theory as extended logic.

I think it can be presented in a manner accessible to many (Jaynes PT:LOS is not accessible to many).

Comment author: nhamann 21 June 2010 04:13:05AM 2 points [-]

(I've always thought that the math behind computer programming was damn useful stuff, but the engineering students I've talked with usually find it harder than calculus, so maybe that's not the best idea.)

Tangential question to your tangential question: I'm puzzled, which math are you talking about here? The only math relevant to programming that I can think of that engineering students would also learn would be discrete math, but the extent needed for good programming competency is pretty small and easy to pick up.

Are we talking numerical computing instead, with optimization problems and approximating solutions to DE's? That's the only thing I can think of relevant to engineering for which the math background might be more difficult than calculus.

Comment author: sketerpot 21 June 2010 04:50:23AM 2 points [-]

I was thinking more basic: induction, recursion, reasoning about trees. Understanding those things on an intuitive level is one of the main barriers that people face when they learn to program. It's one thing to be able to solve problems out of a textbook involving induction or recursion, but another thing to learn them so well that they become obvious -- and it's that higher level of understanding that's important if you want to actually use these concepts.

Comment author: taiyo 21 June 2010 06:04:57AM 0 points [-]

I'm not sure about all the details, but I believe that there was a small kerfuffle a few decades ago over a suggestion to change the apex of U.S. ``school mathematics'' from calculus to a sort of discrete math for programming course. I cannot remember what sort of topics were suggested though. I do remember having the impression that the debate was won by the pro-calculus camp fairly decisively -- of course, we all see that school mathematics hasn't changed much.

Comment author: wedrifid 21 June 2010 04:16:54AM *  0 points [-]

Most people who learn it have a very hard time doing so, and they're already well above average in mathematical ability.

Well above average mathematical ability and cannot do calculus to the extent of understanding rates of change? For crying out loud. You multiply by the number up to the top right of the letter then reduce that number by 1. Or you do the reverse in the reverse order. You know, like you put on your socks then your shoes but have to take off your shoes then take off your socks.

Sometimes drawing a picture helps prime an intuitive understanding of the physics. You start with a graph of velocity vs time. That is the 'acceleration'. See... it is getting faster each second. Now, use a pencil and progressively color in under the line. that's the distance that is getting covered. See how later on more when it is going faster more distance is being traveled at one time and we have to shade in more area? Now, remember how we can find the area of a triangle? Well, will you look at that... the maths came out the same!

Comment author: sketerpot 21 June 2010 05:02:11AM *  3 points [-]

Well above average mathematical ability and cannot do calculus to the extent of understanding rates of change? For crying out loud.

People get the simple concepts mixed together with a bunch of mathy-looking symbols and equations, and it all congeals into an undifferentiated mass of confusing math. Yes, I know calculus is actually pretty straightforward, but we're probably not a representative sample. Talk with random bewildered college freshmen to combat sample bias. I did this, and what I learned is that most people have serious trouble learning calculus.

Now, if you want to be able to partially understand a bunch of physics stuff but you don't necessarily need to be able to do the math, you could probably get away with a small subset of what people learn in calculus classes. If you learned about integration and differentiation (but not how to do them symbolically), as well as vectors, vector fields, and divergence and curl, then you could probably get more benefit-per-hour-of-study than if you went and learned calculus properly. It leaves a bad taste in my mouth, though.

Comment author: wedrifid 21 June 2010 06:52:59AM 2 points [-]

what I learned is that most people have serious trouble learning calculus.

When taught well the calculus required for the sort of applications you mentioned is not something that causes significant trouble, certainly not compared to vector fields, divergence or curl. By taught well, if you will excuse my lack of seemly modesty, is how I taught it in my (extremely brief - don't let me get started on what I think of western school systems) stint teaching high school physics. The biggest problem for people learning basic calculus is that people teaching it try to convey that it is hard.

I'm only talking here about the level of stuff required for everyday physics. Definitely not for the vast majority of calculus that we try to teach them.

Comment author: taiyo 21 June 2010 06:24:06AM *  5 points [-]

I teach calculus often. Students don't get hung up on mechanical things like (x^3)' = 3x^2. They instead get hung up on what

has to do with the derivative as a rate of change or as a slope of a tangent line. And from the perspective of a calculus student who has gone through the standard run of American school math, I can understand. It does require a level up in mathematical sophistication.

Comment author: wedrifid 21 June 2010 06:50:56AM *  0 points [-]

from the perspective of a calculus student who has gone through the standard run of American school math,

That's the problem. See that bunch of symbols? That isn't the best way to teach stuff. It is like trying to teach them math while speaking a foreign language (even if technically we are saving the greek till next month). To teach that concept you start with the kind of picture I was previously describing, have them practice that till they get it then progress to diagrams that change once in the middle, etc.

Perhaps the students here were prepared differently but the average student started getting problems with calculus when it reached a point slightly beyond what you require for the basic physics we were talking about here. ie. they would be able to do 1. and but have no chance at all with 2:

Comment author: Nisan 25 June 2010 12:26:01PM 0 points [-]

Calculus might not be the best example of a skill with relatively low payoff, because you need some calculus to understand what a continuous probability distribution is.

Comment author: wedrifid 25 June 2010 01:10:40PM 0 points [-]

I do? I thought I understood both calculus and continuous probability but I didn't know one relied on the other. You are probably right, sometimes things that are 'obvious' just don't get remembered.

Comment author: Nisan 25 June 2010 02:57:18PM 1 point [-]

For example, suppose you have a biased coin which lands heads up with probability p. A probability distribution that represents your belief about p is usually a non-negative real function f on the unit interval whose integral is 1. Your credence in the proposition that p lies between 1/3 and 1/2 is the integral of f from 1/3 to 1/2.

Comment author: LucasSloan 20 June 2010 05:21:52AM 0 points [-]

Are you talking 'disinclination to shut up and multiply'

This one.

Comment author: JoshuaZ 20 June 2010 05:25:26AM 1 point [-]

Are you then asserting that non-utilitarian views constitute a delusion?

Comment author: LucasSloan 20 June 2010 06:43:56AM *  2 points [-]

I'm asserting that saying "We must do X, because it produces good effect Y", when there is option Z, which delivers the same Y for half the cost, is a delusion.

Comment author: JoshuaZ 20 June 2010 03:09:08PM 1 point [-]

I'm asserting that saying "We must do X, because it produces good effect Y", when there is option Z, which delivers the same Y for half the cost, is a delusion.

This seems more like a common cognitive error than a delusion. How are you defining delusion? It seems like I am using a more narrow definition of delusion. I'm using something like "statement or class of statements about the physical world that are demonstrably extremely unlikely to be true." What definition are you using?

Comment author: wedrifid 20 June 2010 03:40:12PM 1 point [-]

I'm using something like "statement or class of statements about the physical world that are demonstrably extremely unlikely to be true." What definition are you using?

Lucas's statement fits this definition. This may me be clearer if you consider just "we must do X", which is a claim about the physical world. The because part does not happen to change this.

If you don't agree that the truncated claim fits the criteria I infer that the most likely difference in definitions between you and Lucas is in not so much around 'delusion' but rather about what 'must' means in relation to the physical world. This would make what you say true even if it isn't grounded in my preferred ontology.

Comment author: RichardKennaway 21 June 2010 10:11:48AM 2 points [-]

Dredging this up from deep nesting, because I think it's important: wedrifid says

The biggest problem for people learning basic calculus is that people teaching it try to convey that it is hard.

Yes. Never tell anyone that what you're teaching them is hard. When you do that, you're telling them they'll fail, telling them to fail.

Comment author: Alicorn 21 June 2010 05:39:12PM 3 points [-]

But if you tell them it's easy, then they will be embarrassed for failing at something easy, or can't be proud of succeeding at something easy.

Comment author: RichardKennaway 21 June 2010 06:48:59PM 2 points [-]

Telling them it's easy is also a bad idea.

Comment author: Alicorn 21 June 2010 08:18:27PM *  1 point [-]

It strikes me that giving no information about the general difficulty of the subject is also a bad idea. (I imagined myself struggling with a topic where I had no information on how hard others found it, and my hypothetical self was ashamed, because clearly if it were something everyone found hard, they'd warn people and teach it more slowly, so it must be easy for everybody else but me.)

Comment author: Blueberry 21 June 2010 08:25:32PM 3 points [-]

Ideally, you'd teach the student not to be concerned with how well or how quickly they learn compared to others, which is a general learning technique that can apply to any field.

Comment author: RichardKennaway 22 June 2010 07:38:29AM 0 points [-]

When I teach, I don't say anything about "easy" or "difficult". I just teach the material. What is this "easy", this "difficult"? There is no "easy" or "difficult" for a Jedi -- there is only the work to be done and the effort it takes. "Difficult" means "I will fail". "Effort" means "I will succeed".

(I imagined myself

You are torturing yourself by inventing fictional evidence. You have an entire imaginary scenario there, shadows and fog conjured from thin air.

Comment author: SilasBarta 21 June 2010 08:57:00PM *  2 points [-]

Right, and there's the issue of whose fault the difficulty is. Sure, the student might not really be trying. But also, the teacher may not be explaining in a way that speaks to the learner's natural fluency. A method that works for the geeky types won't work work for more neurotypical types.

For my part, I never have trouble explaining high school math to those who haven't completed it, even if they're told that trig, calculus, etc. is hard. It's because I first focus on finding out where exactly their knowledge deficit is and why the subject matter is useful. Of course, teachers don't have the luxury of one-on-one instruction, but yes, how you present the material matters greatly.

Comment author: JoshuaZ 20 June 2010 04:43:09AM 2 points [-]

FAI and related.

What is the delusion here?

Inability to do math.

What is the delusion here? Do you mean people convincing themselves that they can't do math?

Failures around believing the state of the world is good

This seems too subjective to label a delusion.

Believing that politics is the best way to influence the world.

What do you mean by best and by influence?

Comment author: JoshuaZ 20 June 2010 04:46:47AM *  6 points [-]

I've separated some forms of alternative medicine out when one might arguably put them closer together. Also, I'm including Young Earth Creationism, but not creationism as a whole. Where that goes might be a bit more complicated. There's some overlap between some of these (such as young earth creationism and religion). The list also does not include any beliefs that have a fundamentally moral component. I've tried to not include beliefs which are stupid but hard to deal with empirically (say that there's something morally inferior about specific racial groups). Finally, when compiling this list I've tried to avoid thinking too much about the overall balance that the delusion provides. So for example, religion is listed where it is based on the harm it does, without taking into account the societal benefits that it also produces.

1-4: Religion, Ayurveda, Homeopathy, Traditional Chinese medicine (as standardized post 1950s)

5-10 The belief that intelligence differences have no strong genetic component. The belief that intelligence differences have no strong environmental component. The belief that there are no serious existential threats to humans. The belief that external cosmetic features or national allegiances are strong indicators of mental superiority or inferiority. That human females have fundamentally less mental capacity and that this difference is enough to be a useful data point when evaluating humans. The belief that the Chinese government can be trusted to benefit its people or decide what information they should or should not have access to. (The primary reason this gets on the list is the sheer size of China. There are other governments which are much, much worse and have similar delusions by the people. But the damage level done is frequently much smaller.)

11-20 Vaccines cause autism. Young Earth Creationism. Invisible Hand of the Market solves everything. Government solves everything. Providence. That there are not fundamental limits on certain natural resources. That nuclear power is intrinsically worse than other forms of energy. The belief that large segments of the population are fundamentally not good at math or science. Astrology. The belief that antibiotics can deal with viral infections.

There were a few that I wanted to stick on for essentially emotional reasons. So for example Holocaust Denial almost got on the list and when I tried to justify it I saw myself engaging in what was clearly motivated cognition.

This list is very preliminary. The grouping is also very tentative and could likely be easily subject to change.

Comment author: JanetK 20 June 2010 09:12:43AM 2 points [-]

Where would you put 'belief in free will' and 'belief in determinism'?

Comment author: JoshuaZ 20 June 2010 01:37:48PM 3 points [-]

They probably wouldn't get anywhere on the list for the reason that a) I'm not convinced that either determinism or free will as often given are actually well-defined notions and b) I don't see either belief as causing much harm in practice.

Comment author: Emile 21 June 2010 08:21:40AM 2 points [-]

Lists like that are good !

The belief that the Chinese government can be trusted to benefit its people or decide what information they should or should not have access to. (The primary reason this gets on the list is the sheer size of China. There are other governments which are much, much worse and have similar delusions by the people. But the damage level done is frequently much smaller.)

I'm a bit surprised at that one - the current Chinese government seems pretty rational and efficient to me, and I'd be hard-pressed to say what I would do differently in it's place or rather - there are things I would do differently, but I'm not sure I'd get better results).

Control of information by the government should be seen mostly as a way of preserving it's own power. So I'm not really sure of how to interpret "The belief that the Chinese government can be trusted to [...] decide what information they should or should not have access to." - could you rephrase that belief so that it's irrationality becomes more apparent, maybe tabooing "can be trusted to" ? If you mean "Chinese people wrongly believe that the government is restricting information access for their own good", then I'm not sure that a lot of people actually believe that, and for those that do, that believing it does any harm.

Comment author: JoshuaZ 21 June 2010 03:42:33PM *  1 point [-]

If you mean "Chinese people wrongly believe that the government is restricting information access for their own good", then I'm not sure that a lot of people actually believe that, and for those that do, that believing it does any harm.

Ok. My impression is that that is a common belief in China and is connected to the belief that the government doesn't actively lie. I don't have a very good citation for this other than general impressions so I'm going to point to a relevant blog entry by a friend who spent a few years in China where she discusses this with examples. There are of course even limits to how far that will go. This is also complicated by the fact that many of the really serious harm in China (detainment of citizens for questioning policies, beatings and torture, ignoring of basic environmental and safety issues) stem from the local governments rather than the central government, and the relationship between Beijing and the local governments is very complicated. See also my remarks above to wedrifid which touch on these issues also. So yeah, it may make sense to take this off the list given the lack of harm directly coming from this issue.

Comment author: Douglas_Knight 21 June 2010 09:09:29PM 2 points [-]

I don't interpret the story in that blog post that way at all. People repeating nationalist lies doesn't mean they've been fooled.

I highly recommend these posts about the psychology of mass lies. I don't recommend the third part.

Comment author: Emile 21 June 2010 08:33:04AM 1 point [-]

That there are not fundamental limits on certain natural resources.

Does anybody actually claim to believe that ?

Comment author: JoshuaZ 21 June 2010 03:28:51PM 3 points [-]

This view is surprisingly common. I don't want to move to much to a potentially mind-killing subject, but the idea isn't uncommon among certain groups in US politics. Indeed, they think it so strongly about some resources that they take it almost as an ideological point. This occurs when discussing oil most frequently. Emphasis is placed on things like the Eugene Island field and abiotic oil which they argue shows we won't run out of oil. The second is particularly galling because even if the abiotic oil hypotheses were correct the level of oil production would still be orders of magnitudes below the consumption rate. I'd point more generally to followers of Julian Simon (not Simon himself per se. His own arguments were generally more nuanced and subtle than what many people seem to get out of them).

Comment author: wedrifid 21 June 2010 08:53:52AM 2 points [-]

The belief that the Chinese government can be trusted to benefit its people or decide what information they should or should not have access to.

Is it trust or fear that is the real problem in that case? What would you do as an average Chinese citizen who wanted to change the policy? (Then, the same question assuming you were an actual Chinese citizen who didn't have your philosophical mind, intelligence, idealism and resourcefulness.)

Comment author: JoshuaZ 21 June 2010 03:16:35PM *  3 points [-]

Is it trust or fear that is the real problem in that case?

It seems like it is a mix. From people I've spoken to in China and the impression I get from what I've read about the Chinese censorship, the majority of people are generally ok with letting the government control things and think that that's really for the best. This seems to be changing slightly with the younger generation but it is hard to tell.

What would you do as an average Chinese citizen who wanted to change the policy? (Then, the same question assuming you were an actual Chinese citizen who didn't have your philosophical mind, intelligence, idealism and resourcefulness.)

Good points certainly. I'm not sure any average Chinese citizen alone can do anything. If I were an actual Chinese citizen alone given my "philosophical mind, intelligence, idealism and resourcefulness," I'm not sure I'd do anything either, not because I can't, but because the risk would be high. It is easy to say "oh, people in X situation should do Y because that's morally better or better for everyone overall" when one isn't in that situation. When one's life, family, or livelihood is the one being threatened then it is obviously going to be a lot more difficult. It isn't that I'm a coward (although I might be) it is just that standing up to the government in that sort of situation takes a lot of courage that I'm pretty sure I (and most people) don't have. But if the general population took an attitude that was more willing to do minor things (spread things like TOR or other methods of getting around the Great Firewall for example), then things might be different. But even that might not have a large impact.

So yeah, I may need to take this off the list.

Comment author: Emile 21 June 2010 05:19:19PM 1 point [-]

From people I've spoken to in China and the impression I get from what I've read about the Chinese censorship, the majority of people are generally ok with letting the government control things and think that that's really for the best. This seems to be changing slightly with the younger generation but it is hard to tell.

I get the impression that overall, the younger generation is more apathetic about politics than the older one.

(Though there is also the relatively recent phenomenon of "angry youths" (fenqing), who rant on forums and such.)

Comment author: Risto_Saarelma 21 June 2010 12:51:22PM 1 point [-]

The belief that large segments of the population are fundamentally not good at math or science.

This one caught my eye, I don't think I've seen this listed as an obvious delusion before. Can you maybe expand more on this? I guess the idea is that a much larger number of people could make use of math or science if they weren't predisposed to think that they belong in an incapable segment?

I'm thinking of something like picking the quarter of population that scores in the bottom at a standard IQ test or the local SAT-equivalent as the "large segment of population" though. A test for basic science and mathematics skills could be being able to successfully figure out solutions for some introductionary exercises from a freshman university course in mathematics or science, given the exercise, relevant textbooks and prerequisite materials, and, say, up to a week to work out things from the textbook.

It doesn't seem obvious to me that such a test would end up with results that would make the original assertion go straight into 'delusion' status. My suspicions are somewhat based on the article from a couple of years back, which claimed that many freshman computer science students seem to simple lack the basic mental model building ability needed to start comprehending programming.

Comment author: JoshuaZ 21 June 2010 03:58:22PM 2 points [-]

I guess the idea is that a much larger number of people could make use of math or science if they weren't predisposed to think that they belong in an incapable segment?

Yes. And more people would go into math and science.

My suspicions are somewhat based on the article from a couple of years back, which claimed that many freshman computer science students seem to simple lack the basic mental model building ability needed to start comprehending programming.

That's a very interesting article. I think that the level of, and type of abstraction necessary to program is already orders of magnitude beyond where most people stop being willing to do math. My own experience in regards to tutoring students who aren't doing well in math is that one of the primary issues is one of confidence: students of all types think they aren't good at math and thus freeze up when they see something that is slightly different from what they've done before. If they understand that they aren't bad at math or that they don't need to be bad at math, they are much more likely to be willing to try to play around with a problem a bit rather than just panic.

I was an undergraduate at Yale which is generally considered to be a decent school that admits people who are by and large not dumb. And one thing that struck me was that even in that sort of setting, many people minimized the amount of math and science they took. When asked about it the most common claim was that they weren't good at it. Some of those people are going to end up as future senators and congressman and have close to zero idea of how science works or how statistics work other than at the level they got from high school. If we're lucky, they know the difference between a median and a mean.

Comment author: Vladimir_M 20 June 2010 09:37:25AM *  4 points [-]

Mass_Driver:

I'm extremely curious to hear both your list and JoshuaZ's list of the top 20 or so most harmful delusions.

I'm not sure if that would be a smart move, since it would mean an extremely high concentration of unsupported controversial claims in a single post. Many of my opinions on these matters would require non-obvious lengthy justifications, and just dumping them into a list would likely leave most readers scratching their heads. If you're really curious, you can read the comment threads I've participated in for a sample, in particular those in which I argue against beliefs that aren't specific to my interlocutors.

Also, it should be noted that the exact composition of the list would depend on the granularity of individual entries. If each entry covered a relatively wide class of beliefs, creationism might find itself among the top fifty (though probably nowhere near the top ten).

Comment author: wedrifid 20 June 2010 10:01:49AM *  5 points [-]

I'm not sure if that would be a smart move, since it would mean an extremely high concentration of unsupported controversial claims in a single post.

In this format that sounds like a good thing! At worst it would spark curiosity and provoke discussion. At best people would encounter a startling opinion that they had never seriously considered, think about for 60 seconds then form an understanding that either agrees with yours or disagrees, for a considered reason.

Comment author: h-H 20 June 2010 12:35:23PM *  1 point [-]

seconded, but a list of 20 seems too long/too much work, no?

Comment author: wedrifid 20 June 2010 12:48:29PM 1 point [-]

I'd be thinking 5. :)

Comment author: CronoDAS 21 June 2010 08:11:50AM 5 points [-]

I'll give you a big one: Dying a martyr's death gives you a one-way ticket to Paradise.

Comment author: Kaj_Sotala 19 June 2010 09:38:58PM *  5 points [-]

Deus Ex: Human Revolution

IGN Preview

It has been a while since I needed to buy a new computer to play a game.

In addition to being a sequel to Deus Ex and looking generally bad-ass, transhumanism is explicitly mentioned. From the FAQ:

Essentially, DX: HR explores the beginnings of human augmentation and the transhumanism movement is a major influence in the game. There are people who think it's "playing God" to modify the body whatsoever and there are people (Transhumanists) who think it's the natural evolution of the human species to utilise technology. You're caught in the middle of this storm and must decide which path you take. The visual stigma augmentated people bear adds fuel to the huge societal rift between them and natural humans that's at the centre of Deus Ex: HR's vision of the future.

Comment author: Kaj_Sotala 20 June 2010 03:27:52AM 10 points [-]

A visual study guide to 105 types of cognitive biases

"The Royal Society of Account Planning created this visual study guide to cognitive biases (defined as "psychological tendencies that cause the human brain to draw incorrect conclusions). It includes descriptions of 19 social biases, 8 memory biases, 42 decision-making biases, and 36 probability / belief biases."

Comment author: knb 20 June 2010 04:41:46AM *  10 points [-]

Some random thoughts about thinking, based mostly on my own experience.

I've been playing minesweeper lately (and I've never played before). For the uninitiated, minesweeper is a game that involves using deductive reasoning (and rarely, guessing) to locate the "mines" in a grid of identical boxes. For such an abstract puzzle, it really does a good job of working the nerves, since one bad click can spoil several minutes' effort.

I was surprised to find that even when I could be logically certain about the state of a box, I felt afraid that I was incorrect (before I clicked), and (mildly) amazed when I turned out to be correct. It felt like some kind of low level psychic power or something. So it seems that our brains don't exactly "trust" deductive reasoning. Maybe because problems in the ancestral environment didn't have clean, logical solutions?

I also find that when I'm stymied by a puzzle, if I turn my attention to something else for a while, when I come back, I can easily find some way forward. The effect is stunning, an unsolvable problem becomes trivial five minutes later. I'm pretty sure there is a name for this phenomenon, but I don't know what it is. In any case, it's jarring.

Another random thought. When I'm sad about something in my life, I usually can make myself feel much better by simply saying, in a sentence, why I'm sad. I don't know why this works, but it seems to make the emotion abstract, as though it happened to somebody else.

Comment author: Alicorn 20 June 2010 05:17:38AM *  6 points [-]

When I'm sad about something in my life, I usually can make myself feel much better by simply saying, in a sentence, why I'm sad.

Explicitly acknowledging emotions as things with causes is a huge chunk of managing them deliberately. (I have a post in the works on this, but I'm not sure when I'll pull it together.)

Comment author: Will_Newsome 20 June 2010 07:14:03AM *  0 points [-]

Lots of references to the CBT literature would be nice... no need to reinvent the wheel; CBT has a lot of useful things to say about NATs, and strategies to take care of them. (Then again this applies mostly to negative emotions, and deliberately managing positive emotions seems like a cool thing to do too.) That said, more instrumental rationality posts would be great.

Comment author: CronoDAS 21 June 2010 09:08:45PM 0 points [-]

What does NAT stand for?

Comment author: h-H 20 June 2010 01:14:28PM 0 points [-]

It felt like some kind of low level psychic power or something. So it seems that our brains don't exactly "trust" deductive reasoning. Maybe because problems in the ancestral environment didn't have clean, logical solutions?

I'm nitpicking, but maybe it was simple pleasure at getting the game?

Comment author: CronoDAS 21 June 2010 09:10:55PM 2 points [-]

Another random thought. When I'm sad about something in my life, I usually can make myself feel much better by simply saying, in a sentence, why I'm sad. I don't know why this works, but it seems to make the emotion abstract, as though it happened to somebody else.

I don't think that works for me. I often can't identify a specific cause of my sad feeling, and when I can, thinking about it often makes me feel worse rather than better.

Comment author: SilasBarta 21 June 2010 09:25:50PM *  3 points [-]

Same here. I also found that often there's not any cause in the sense of something specific upsetting me; it's just an automatic reaction to not getting enough social interaction.

Comment author: knb 25 June 2010 12:27:21AM 2 points [-]

Well I don't mean ruminating about the cause of the sad feeling. That is probably one of the worst things you can do. Rather I meant just identifying it.

For example, when a girlfriend and I broke up (this was a couple years ago) I spent maybe two days feeling really depressed. Eventually, I thought to myself, "You're sad because you broke up with your girlfriend."

That really put it in perspective for me. It made me think of all the cheesy teen movies where kids breakup with their sweethearts and act like it's the end of the world, when in the viewer sees it as a normal, even banal rite of passage to adulthood. I had always thought people who reacted like that were ridiculous. In other words, it feels like that thought put the issue in "far mode" for me.

Comment author: Blueberry 25 June 2010 01:13:52AM 0 points [-]

That works if there is a specific cause, but like some other people have said, my sad feelings aren't caused by external events.

Comment author: Daniel_Burfoot 24 June 2010 11:37:33PM 0 points [-]

So it seems that our brains don't exactly "trust" deductive reasoning. Maybe because problems in the ancestral environment didn't have clean, logical solutions?

Arguably, problems in the modern environment don't have clean, logical solutions either! Note also that people get good at games like minesweeper and chess through learning. If the brain was primarily a big deductive logic machine, it would become good at these games immediately upon understanding the rules; no learning would be necessary.

Comment author: Alicorn 20 June 2010 05:45:52AM 12 points [-]
Comment author: JoshuaZ 20 June 2010 05:52:04AM 0 points [-]

That's one of the funniest things I've seen in a while. I wish I could upvote that more.

Comment author: Unnamed 20 June 2010 06:30:45AM 4 points [-]
Comment author: h-H 20 June 2010 11:47:39AM 0 points [-]

upvoted-both-I remember yours this from a couple of years ago

Comment author: h-H 20 June 2010 11:41:51AM *  7 points [-]

genes, memes and parasites?

tl:dr:"People who suffer from schizophrenia are, in fact, three times more likely to carry T. gondii than those who do not."

"Over the last five years or so, evidence has been building that some human cultural shifts might be influenced, or even caused, by the spread of Toxoplasma gondii."

"In the United States, 12.3 percent of women tested carried the parasite, and in the United Kingdom only 6.6 percent were infected. But in some countries, statistics were much higher. 45 percent of those tested in France were infected, and in Yugoslavia 66.8 percent were infected!"

Comment author: wedrifid 20 June 2010 11:57:51AM 1 point [-]

"In the United States, 12.3 percent of women tested carried the parasite

Wow. How is this parasite spread? Could those 'girly germs' that I avoided in primary school actually reduce my chances of getting schizophrenia?

Comment author: h-H 20 June 2010 12:31:37PM 1 point [-]

wait, what's a girly germ? I googled it and it game me a link about a Micronesian island :/

Comment deleted 20 June 2010 12:38:20PM [-]
Comment author: wedrifid 20 June 2010 05:55:52PM 0 points [-]

Ick. My double posting browser bug again.

Comment author: JoshuaZ 20 June 2010 05:58:07PM 1 point [-]

Have you tried using another browser? That might help you figure out if the problem is actually on the browser end and not something weird with the LW software.

Comment author: wedrifid 20 June 2010 06:05:44PM 1 point [-]

I'm using a different browser (different computer same browser by name) now and it is working fine. My other browser seems to work fine for a while after I restart it until some event causes it to thereafter double post every time. My hunch is that I could identify the triggering of one of the plugins as the cause. Even then the symptom is outright bizarre. What kind of bug would make the browser double send all post requests?

Perhaps a failed attempt at spyware!

No matter. I don't like my other computer anyway.

Comment author: wedrifid 20 June 2010 12:38:21PM 2 points [-]

Do young kids where you are come tease each other about the other sex? 'Cooties?' Whatever they call it.

My question is how the parasite is spread. What does that 12.3% mean for the rest of the population? Why did they only test women?

Comment author: Morendil 20 June 2010 05:03:07PM 3 points [-]

Why did they only test women?

It's a major pregnancy risk.

Comment author: SilasBarta 20 June 2010 07:36:30PM 0 points [-]

Gravitomagnetism -- what's up with that?

It's an phrasing of how gravity works with equations that have the same form as Maxwell's equations. And frankly, it's pretty neat: writing the laws for gravity this way gets you mechanics while approximately accounting for general relativity (how approximate and what it leaves off, I'm not sure of).

When I first found out about this, it blew my mind to know that gravity acts just like electromagnetism, but for different properties. We all know about the parallel between Coulomb's law and Newton's law of gravitation, but the gravitoelectromagnetism (GEM) equations show that it goes a lot deeper.

Besides being a good way to ease into an intuitive understanding of the Einstein field equations, to me, it's basically saying that gravity and EM are both obeying some more general law. Anyone know if work has been done in unifying gravity and EM this way? All I hear about is that it's easy to unify strong, weak, and EM forces, but gravity is the stumbling block, so this should be something they'd want to explore more.

Yet when you go investigate "gravitational induction" to find out how the gravitic parallel to magnetic fields works, you find that this gravitomagnetic field is called the torsion field, and its existence is (at least approximately) implied by general relativity, but then the Wikipedia page says that the torsion field is a pseudoscientific concept. Hm...

So, anyone have an understanding of the GEM analogy and can make sense of this? Does it suggest a way to unify gravity and EM? Or how to create a coil of mass flow that can "gravitize" a region (as a coil of current magnitizes a metal bar)?

Comment author: wedrifid 20 June 2010 08:24:33PM 0 points [-]

I'm intrigued by the notion and would like to hear more from someone who can tell me whether I can take this seriously. That 'approximately accounting for' part scares me. Is that just word chioce that makes it sound scary? Or perhaps an approximation in the way that Newtonian physics is an approximation? Or maybe it is only an approximation is as much as it suffers the same problem all our theories do of being unable to unify all of our physics at once... I'd need someone several levels ahead of me to figure that out.

Comment author: SilasBarta 20 June 2010 08:31:59PM 1 point [-]

It's definitely better of an approximation than Newtonian physics. This paper might help, as it derives the GEM equations from GR and specifically states what simplifying assumptions it uses, which look to be basically "for greater-than-subatomic distances". And that's exactly where you care about gravity anyway. (At subatomic distances, the other three forces dominate.)

Comment author: red75 20 June 2010 08:24:57PM 0 points [-]

Be careful, you are near fringe science domain.

Comment author: mindviews 21 June 2010 08:16:06AM 5 points [-]

it's basically saying that gravity and EM are both obeying some more general law

No, what's happening is that under certain approximations the two are described by similar math. The trick is to know when the approximations break down and what the math actually translates to physically.

Does it suggest a way to unify gravity and EM?

No.

Keep in mind that for EM there are 2 charges while gravity has only 1. Also, like electric charges repel while like gravitic charges attract. This messes with your expectations about the sign of an interaction when you go from one to the other. That means your intuitive understanding of EM doesn't map well to understanding gravity.

Comment author: SilasBarta 21 June 2010 01:02:01PM 0 points [-]

True, but what got me the most interested is the gravitic analog of magnetic fields. It shows that masses can produce something analogous to magnetism by their rotation. Rotate one way, you drag the object closer; rotate the other way, you push it away. This allows both attraction and repulsion in the equations for gravity, and suggests something similar is going on that generates magnetism.

Comment author: cousin_it 21 June 2010 10:47:49AM *  1 point [-]

Your link to "torsion field" talks about a completely different concept than the one in GEM. That concept is indeed a notorious example of pseudoscience here in Russia.

Comment author: Sniffnoy 21 June 2010 05:16:21PM 0 points [-]

I'd mostly like to echo what mindviews said - similar math is not unification - and point out that there was an actual attempt at unification in Kaluza-Klein theory. But I don't actually know anything about that, I should note...

Comment author: NancyLebovitz 21 June 2010 01:10:07PM 5 points [-]

On not being able to cut reality at the joints because you don't even know what a joint is: diagnosing schizophrenia

If you gave Aristotle ten thousand unplugged computers of different makes and models, no matter how systematically he analyzed them he'd not only be wrong, he'd be misleadingly wrong. He would find that they were related by shape-- rectangles/squares; by color-- black, white, or tan. Size/weight; material.

Aristotle was smart, but there is nothing he could ever learn about computers from his investigations. His science is all wrong for what he was doing. But Aristotle would think he knew a terrible amount about computers from his studies. In fact, he'd probably be considered an expert. "To fix this computer, we need to make it more rectangular. Get chopping, malaka."

Comment author: GuySrinivasan 21 June 2010 04:18:28PM 2 points [-]

Statistical Analysis Overflow is trying to start up. If you'd be a regular contributor go over and commit, if enough commit it'll go into beta.

It's a "Proposed Q&A site for statistics, data analysis, data mining and data visualization", like Stack Overflow or Math Overflow.

Comment author: Kevin 21 June 2010 05:36:15PM 5 points [-]

Part one of a five part series on the Dunning-Kruger effect, by Errol Morris.

http://opinionator.blogs.nytimes.com/2010/06/20/the-anosognosics-dilemma-1/

Also note that Oscar winning director Morris's next project is a dark comedy that is a fictionalized version of the founding of Alcor!

Comment author: gwern 21 June 2010 06:26:04PM 1 point [-]

Also note that Oscar winning director Morris's next project is a dark comedy that is a fictionalized version of the founding of Alcor!

Isn't that a bad thing? I suspect a major source will be that recent book...

Comment author: Kevin 21 June 2010 09:04:00PM 0 points [-]
Comment author: arundelo 21 June 2010 10:32:45PM 1 point [-]

Ooh, it's nice to see more details on the lemon juice bank robber. When I first heard about him I thought he was probably schizophrenic. Maybe he was, but the details make it sound like he may indeed have been just really stupid.

Comment author: Randaly 21 June 2010 10:39:40PM 4 points [-]

A recent study found that one effective way to resist procrastination in future tasks is to forgive previous procrastination- because the negative emotions that would otherwise remain create an ugh field around that task.

I found the study recently, but I've personally found this to be effective previously. Forcing your way through an ugh field isn't sustainable due to our limited supply of willpower (this is hardly a new idea, but I haven't seen it referenced in my limited readings on LW.)

Comment author: wedrifid 22 June 2010 07:48:38AM 0 points [-]

(this is hardly a new idea, but I haven't seen it referenced in my limited readings on LW.)

Some people have tried to emphasize that point but it isn't always universally understood.

Comment author: [deleted] 22 June 2010 06:51:37AM *  3 points [-]

For those of you who don't want to register to fanfic.com to receive notifications of new chapters to Harry Potter and the methods of rationality, I have added a Mailinglist. You can add yourself here: http://felix-benner.com/cgi-bin/mailman/listinfo/fanfic It is still untested so I don't know it will work, but I assume so.

Comment author: cousin_it 22 June 2010 09:00:18AM *  1 point [-]

A recent comment about Descartes inspired this thought: the simplest possible utility function for an agent is one that only values survival of mind, as in "I think therefore I am". This function also seems to be immune to the wireheading problem because it's optimizing something directly perceivable by the agent, rather than some proxy indicator.

But when I started thinking about an AI with this utility function, I became very confused. How exactly do you express this concept of "me" in the code of a utility-maximizing agent? The problem sounds easy enough: it doesn't refer to any mystical human qualities like "consciousness", it's purely a question about programming tricks, but still it looks quite impossible to solve. Any thoughts?

Comment author: wedrifid 22 June 2010 09:04:49AM 0 points [-]

But when I started thinking about an AI with this utility function, I became very confused. How exactly do you express this concept of "me" in the code of a utility-maximizing agent? The problem sounds easy enough: it doesn't refer to any mystical human qualities like "consciousness", it's purely a question about programming tricks, but still it looks quite impossible to solve. Any thoughts?

It refers to mystical human qualities like "me" and "think". Basically I put it in the exact same category as 'consciousness'.

Comment author: cousin_it 22 June 2010 09:23:20AM *  0 points [-]

No it doesn't. I'm not interested in replicating the inner experience of humans. I'm interested in something that can be easily noticed and tested from the outside: a program that chooses the actions that allow the program to keep running. It just looks like a trickier version of the quine problem, do you think that one's impossible as well?

Comment author: Vladimir_Nesov 22 June 2010 10:11:46AM 0 points [-]

something that can be easily noticed and tested from the outside

This concept is extremely complex (for example, which "outside" are you talking about?).

Comment author: cousin_it 22 June 2010 10:34:20AM 0 points [-]

You seem to be reading more than I intended into my original question. If the program is running in a simulated world, we're on the outside.

Comment author: Vladimir_Nesov 22 June 2010 10:52:40AM 0 points [-]

Yes, using a formal world simplifies this a lot.

Comment author: Mitchell_Porter 22 June 2010 10:34:42AM 0 points [-]

It's not hard to design a program with a model of the world that includes itself (though actually coding it requires more effort). The first step is to forget about self-modeling, and just ask, how can I model a world with programs? Then later on you put that model in a program, and then you add a few variables or data structures which represent properties of that program itself.

None of this solves problems about consciousness, objective referential meaning of data structures, and so on. But it's not hard to design a program which will make choices according to a utility function which refers in turn to the program itself.

Comment author: cousin_it 22 June 2010 10:37:37AM *  0 points [-]

Well, I don't want to solve the problem of consciousness right now. You seem to be thinking along correct lines, but I'd appreciate it if you gave a more fleshed out example - not necessarily working code, but an unambiguous spec would be nice.

Comment author: Mitchell_Porter 23 June 2010 01:40:42AM 1 point [-]

Getting a program to represent aspects of itself is a well-studied topic. As for representing its relationship to a larger environment, two simple examples:

1) It would be easy to write a program whose "goal" is to always be the biggest memory hog. All it has to do is constantly run a background calculation of adjustable computational intensity, periodically consult its place in the rankings, and if it's not number one, increase its demand on CPU resources.

2) Any nonplayer character in a game which fights to preserve itself is also engaged in a limited form of self-preservation. And the computational mechanisms for this example should be directly transposable to a physical situation, like robots in a gladiator arena.

All these examples work through indirect self-reference. The program or robot doesn't know that it is representing itself. This is why I said that self-modeling is not the challenge. If you want your program to engage in sophisticated feats of self-analysis and self-preservation - e.g. figuring out ways to prevent its mainframe from being switched off, asking itself whether a particular port to another platform would still preserve its identity, and so on - the hard part is not the self part. The hard part is to create a program that can reason about such topics at all, whether or not they apply to itself. If you can create an AI which could solve such problems (keeping the power on, protecting core identity) for another AI, you are more than 99% of the way to having an AI that can solve those problems for itself.

Comment author: wedrifid 22 June 2010 03:27:50PM 0 points [-]

My observation is just that the process you're going through here in taking the "I think therefore I am" and making it into the descriptive and testable system is similar to the process others may go through to find the simplest way to have a 'conscious' system. In fact many people would resolve 'conscious' to a very similar kind of system!

I do not think either are impossible to do once you make, shall we say, appropriate executive decisions regarding resolving the ambiguity in "me" or "conscious" into something useful. In fact, I think both are useful problems to look at.

Comment author: Nick_Tarleton 22 June 2010 06:30:10PM *  0 points [-]

If you want this to work in the real world, not a just much simpler computational environment, then for starters: what counts as a "program" "running"? And what distinguishes "the" program from other possible programs? These seem likely to be in the same category as (not to mention subproblems of) consciousness, whatever that category is.

Comment author: cousin_it 22 June 2010 08:19:05PM *  0 points [-]

Right now I'd be content with an answer in some simple computational environment. Let's solve the easy problem before attempting the hard one.

Comment author: red75 22 June 2010 09:36:48AM *  0 points [-]

Program must have something to preserve. My first thought is preservation of declarative memory: ensure that future contain chain of systems, implementing same goal, with overlapping declarative memory.

I haven't made an analysis, just first thought.

Comment author: RichardKennaway 22 June 2010 09:55:30AM *  1 point [-]

An agent's "me" is its model of itself. This is already a fairly complicated thing for an agent to have, and it need not have one.

Why do you say that an agent can "directly perceive" its own mind? Or anything else? A perception is just a signal somewhere inside the agent: a voltage, a train of neural firings, or whatever. It can never be identical to the thing that caused it, the thing that it is a perception of. People can very easily have mistaken ideas of who they are.

Comment author: Vladimir_Nesov 22 June 2010 10:05:59AM *  2 points [-]

You want the program to keep running in the context of the world. To specify what that means, you need to build on top of an ontology that refers to the world. But figuring out such ontology is a very difficult problem and you can't even in principle refer to the whole world as it really is: you'll always have uncertainty left, even in a general ontological model.

The program will have to know what tradeoffs to make, for example whether it's important to survive in most possible worlds with fair probability, or in at least one possible world with high probability. These would lead to very different behavior, and the possibility of such tradeoffs exemplifies how much data such preference would require. If additionally you want to keep most of the world as it would be if the AI was never created, that's another complex counterfactual for you to bake in into its preference.

It's a very difficult problem, probably more difficult that FAI, since for FAI we at least have some hope of cheating and copying formal preference from an existing blueprint, and here you have to build that from scratch, translating your requirements from human-speak to formal specification.

Comment author: CronoDAS 23 June 2010 05:34:28PM 2 points [-]
Comment author: Matt_Simpson 23 June 2010 06:09:45PM *  1 point [-]

There's an interesting article in the New York Times on warfare among chimpanzees. One problem, though, is that they attempt to explain the level of coordination necessary in warfare with group selection. This, of course, will not do. I'm under-read in evolutionary biology, but it seems like kin selection accounts for this phenomenon just fine. You are more likely to be related to members of your group than an opposing group, so taking territory from a rival group doesn't just increase your fitness directly, but indirectly through your shared genes among group members.

What do you think, LessWrong?

edit: Some commentary on the article.

Comment author: PhilGoetz 23 June 2010 06:56:03PM *  0 points [-]

they attempt to explain the level of coordination necessary in warfare with group selection. This, of course, will not do.

Group selection has been vilified; but irrationally so. Group selection has been observed many times in human groups, so dismissing it is silly.

Comment author: Matt_Simpson 23 June 2010 07:42:41PM 2 points [-]

From the LWwiki:

To the best of this editor's knowledge, no definite example of a group-level adaptation has ever been observed in a mammalian species. Ever.

So, can you point to one of these observations? (and if so, update the wiki!)

Comment author: PhilGoetz 24 June 2010 03:52:06AM *  3 points [-]

I can point to observations of groups being eliminated, and in some of these cases, it seems obvious that elimination was attributable to a behavior, a biological phenotype, or a social phenotype. For instance, there was a group of related tribes in South America, described IIRC in "Life among the Yanomamo", who were very aggressive and kept killing and raping members of neighboring tribes. Eventually, the neighboring tribes got together and killed every last man of the aggressive tribe that they could find. The book "Black Robe" fictionalizes a real-life account of another group selection incident, in which one North American tribe adopted Christianity, and (the book implies) as a result became less violent and were wiped out by neighboring non-Christian tribes. The villages of the Christianized natives of Papua New Guinea are at this moment being razed by the (Muslim) Indonesian army (not that you'll hear anything about it in the news), which you could relate to either the religious or the technological difference between the groups.

I don't know what counts as an "adaptation". When Spanish genes spread rapidly among the natives of central America due to the superior technology of Spain, was that an adaptation?

What I do know is that social norms lead to differential reproductive success. There is obvious group selection going on in the world right now that favors culture that place a high value on high birth rate, or that prohibit birth control.

Comment author: Gabriel 24 June 2010 01:30:46PM *  2 points [-]

What I do know is that social norms lead to differential reproductive success. There is obvious group selection going on in the world right now that favors culture that place a high value on high birth rate, or that prohibit birth control.

But group selection is a more specific idea, the idea that a trait can become widespread due to it's positive effects on group success, regardless of the effects on individual fitness. An example of group selection would be a trait such that: (1) groups in which it is widespread win, (2) lacking the trait doesn't lower the reproductive success of an individual member of such a group. While your examples show (1) it is not clear that they satisfy (2).

Comment author: SilasBarta 24 June 2010 02:11:21PM *  1 point [-]

Then I must admit confusion here: when human groups have norms that punish "defectors", genes that predispose someone to play a "tit for tat" strategy (or, to some extent, altruism) rather than defection are rewarded and spread through the gene pool faster. Is that not a case where group-favoring genes become widespread? To the extent it diverges from the definition you gave, that's because of pretty arbitrary caveats.

I thought that counted as group selection, but was regarded as a "special case" because it requires enforcement of norms to an extent that has only been observed in humans.

Edit: And what other species has anything like China's one-child policy?

Comment author: Matt_Simpson 25 June 2010 05:03:11AM *  2 points [-]

Then I must admit confusion here: when human groups have norms that punish "defectors", genes that predispose someone to play a "tit for tat" strategy (or, to some extent, altruism) rather than defection are rewarded and spread through the gene pool faster. Is that not a case where group-favoring genes become widespread? To the extent it diverges from the definition you gave, that's because of pretty arbitrary caveats.

The definition of group selection, from Wikipedia:

In evolutionary biology, group selection refers to the idea that alleles can become fixed or spread in a population because of the benefits they bestow on groups, regardless of the alleles' effect on the fitness of individuals within that group.

The key is that the benefit to the group is at least part of what is driving the adaptation. Now an adaptation (like tit-for-tat) can certainly benefit the group, but that doesn't mean there is group selection going on - the benefit to the group has to be part of the cause for the trait's spread, apart from the benefit from the individual.

Tit-for-tat is individually fitness maximizing in many situations. In fact, it's an Evolutionary Stable Strategy. So in a population of tit-for-tat players, it's fitness maximizing to play tit-for-tat. So tit-for-tat is not an example of group selection, or at least it's existence doesn't imply group selection has occurred.

Edit: And what other species has anything like China's one-child policy?

That's a decision of a small group of people imposed on a much larger group of people. If each person was individually choosing to have only one child, then it might be group selection. With that being said, the changing birth patterns of developed countries is an interesting phenomena to consider. It's probably just a case of external conditions changing faster than evolution changes us though.

Comment author: PhilGoetz 23 June 2010 06:54:46PM 0 points [-]

Max More writes about biases that treat natural chemicals as safer than man-made chemicals, natural hazards as safer than man-made hazards, and the status quo as preferable to possible futures, in The proactionary principle.

Comment author: Alexandros 24 June 2010 06:16:49AM 0 points [-]

There's a skepticism stack overflow site proposed. If enough follow it, it will go into beta. So if that's your thing, go here

Comment author: red75 24 June 2010 10:15:33AM *  0 points [-]

Schroedinger Cat is dead. Maybe it's time to update plausibility of classic many worlds interpretation is spite of "Einselection, Envariance, Quantum Darwinism".

I am not sufficiently competent to analyze work of W.H. Zurek, but I think that work can be a great source of insights.

Edit: Abstract. Zurek derived Born's rule.

Comment author: Mitchell_Porter 24 June 2010 12:18:59PM 3 points [-]

The "derivation" is on page 12.

The repeated problem for many worlds is that if the quantum state is 1/2 |dead cat> + sqrt(3)/2 |live cat>, then (squaring the coefficients) the probability of dead cat is 1/4, the probability of live cat is 3/4, and so there should be three times as many live cats compared to dead cats (for such a wavefunction); but the decomposition into wavefunction components just produces one dead-cat world and one live-cat world, which naively suggests equal probabilities. The problem is, how do you interpret a superposition like that, in terms of coexisting, equally real worlds, so as to give the right probabilities.

It looks like part of what Zurek does is to pick a basis (Schmidt decomposition) where the components all have the same amplitude - which means they all have the same probability, so the naive branch-counting method works! A potential problem with this way of proceeding is that, expressed in the position basis, the branches end up being complicated superpositions of spatial configurations. (The space of quantum states, the Hilbert space, is a large abstract vector space with a coordinate basis formally labeled by spatial configurations, so the basis vectors of a different basis will be sums of those position-basis vectors.) Explaining complicated superpositions which don't look like reality by positing the existence of many worlds, each of which is itself a complicated superposition that doesn't look like reality, is not very promising. It's sort of okay to do this for microscopic entities because we don't have apriori knowledge about what their reality is like, and we might suppose that the abstract Hilbert-space vector is the actual reality; but somewhere between microscopic and macroscopic, you have to produce an actual live cat, and not just a live cat summed with an epsilon-amplitude dead cat. I have no idea how Zurek deals with this.

Actually, Zurek has a lot of background assumptions which make his reasoning obscure to me and I really don't expect it to make sense in the end, though it's impossible to be sure until you have decoded his outlook. His philosophy is a weird mixture of Bohr's antirealism and Everett's multirealism, and in other papers he says things like

Quantum states acquire objective existence when reproduced in many copies. Individual states -- one might say with Bohr -- are mostly information, too fragile for objective existence.

(thanks to DZS for the quote). And of course it's nonsense to say that something doesn't exist until there are multiple copies of it (how many is the magic number? how can you make an existing copy of a nonexistent original?). Zurek is using the words "objective existence" in some twisted way. I'm sure the reason is that he doesn't have the answer to QM, but he wants to believe he does; that is how smart people end up writing nonsense. But I would have to understand his system to offer a more precise diagnosis.

Comment author: Alexandros 24 June 2010 01:52:35PM *  1 point [-]

If a being presented itself to you and claimed to be omni(potent/scient/present/benevolent), what evidence would you require to accept its claim?

(EDIT: On a second reading, this sounds like a typical theist opening a conversation. I assure you, this is not the case. I am genuinely interested in the range of possible answers to this question.)

Comment author: [deleted] 24 June 2010 11:08:48PM 3 points [-]

Has anybody looked into OpenCog? And why is it that the wiki doesn't include much in the way of references to previous AI projects?

Comment author: Mitchell_Porter 25 June 2010 03:46:59AM 1 point [-]

If making a Friendly AI is compared to landing on the moon, I'd say OpenCog is something like the scaffolding for a backyard rocket. It still needs something extra - the rocket - and even then it won't achieve escape velocity. But a radically scaled-up version of OpenCog - with a lot more theory behind it, and tailored to run at the level of a whole data center rather than on a single PC - is the sort of toolset that could make a singularity.

Comment author: Christian_Szegedy 25 June 2010 06:42:23PM *  0 points [-]
Comment author: Alexandros 26 June 2010 07:09:40AM 1 point [-]
Comment author: cupholder 27 June 2010 12:25:42AM *  2 points [-]

Statisticians Andrew Gelman and Cosma Shalizi have a new preprint out, 'Philosophy and the practice of Bayesian statistics.' The abstract:

A substantial school in the philosophy of science identifies Bayesian inference with inductive inference and even rationality as such, and seems to be strengthened by the rise and practical success of Bayesian statistics. We argue that the most successful forms of Bayesian statistics do not actually support that particular philosophy but rather accord much better with sophisticated forms of hypothetico-deductivism. We examine the actual role played by prior distributions in Bayesian models, and the crucial aspects of model checking and model revision, which fall outside the scope of Bayesian confirmation theory. We draw on the literature on the consistency of Bayesian updating and also on our experience of applied work in social science.

Clarity about these matters should benefit not just philosophy of science, but also statistical practice. At best, the inductivist view has encouraged researchers to fit and compare models without checking them; at worst, theorists have actively discouraged practitioners from performing model checking because it does not fit into their framework.

Comment author: Kevin 27 June 2010 07:27:16AM 1 point [-]
Comment author: Alexandros 27 June 2010 09:20:30AM *  2 points [-]

I started writing something but it came up short for an article, so I'm posting it here:

Title: On the unprovability of the omni*

Our hero is walking down the street, thinking about proofs and disproofs of the existence of a god. This is no big coincidence as our hero does this often. Suddenly, between one step and the next, the world around her fades out, and she finds herself standing on thin air, surrounded by empty space. Then she hears a voice. "I am Omega. The all-powerful, all-knowing, all-good, ever-present being. I see you have been debating my existence with true purity of heart, so I have decided to provide you with any evidence you request". Once the shock wears off, our hero runs through the list of possible requests she could make. Healing the sick? Perhaps the reanimation of a dead person? Some time-travel? Maybe this could still be doubted. How about creation of a solar system? Or a universe? Maybe a proof of P vs. NP? Alas, our hero realises that any evidence she could request would only be proof of the power of Omega to produce just that thing, not an inclusive proof.

What's more, our hero knows that her thinking is subject to the operation of her mind and the readings of her senses, something she cannot trust in the presence of a vastly overpowering entity. The lower bound of power required of Omega to produce any experience for our hero is much lower than the power to create universes. It is the ability to control only the senses of our hero, become a kind of hypervisor, and simulate all requests. While this is great power indeed, the distance from there to omnipotence is great indeed. Similarly for omniscience, omnipresence, and omnibenevolence.

Our hero does not ask anything of Omega, and their meeting ends uneventfully, at least in terms of new universes being created, or problems thought unsolvable being solved. She does realise though, that omnipotence, omniscience, omnipresence, and omnibenevolence are not properties that can be verified by a human. If this is the definition of a god that theists are working with, then it is not only undisprovable, it is also unprovable. Taking knowledge to be 'justified true belief', a belief in an omni* god can never be justified, putting if firmly in the territory of the unknowable. The strongest claims that can be reasonably made are that of a being that is very powerful, very knowledgeable, etc. But that is not nearly as interesting.


Now, I have posted a question along those lines in this thread before, with little response. What I would like your feedback on is whether this is a reasonable argument, whether I've gotten something completely wrong in my epistemology, and whether there have been similar arguments made by others. All help appreciated, cheers.

Comment author: red75 27 June 2010 11:10:07AM 0 points [-]

Your logic is ok. By the way Thomas Aquinas thought along this lines, but in different direction. However discussing scholastic here doesn't make much sense (if it can make sense at all).

Comment author: NancyLebovitz 27 June 2010 01:25:47PM 2 points [-]

What if your hero asks to be made omniscient, including the capacity to still be able to think well in the face of all that knowledge?

Throw in omnibenevolence if you like, but I think you get some contradictions if you ask omnipotence. Either that, or you and Omega coalesce.

How could you test your omniscience to be sure it's the real thing?

Comment author: Alexandros 27 June 2010 03:33:41PM 0 points [-]

Asking to modify yourself may be a useful strategy, (or maybe not, as you note) but it's not something that's available to philosophers trying to prove the existence of a god. As far as we know that is :)

Comment author: Alicorn 27 June 2010 06:53:32PM *  5 points [-]

Wait... a being which, while possibly not omni-anything, is likely very powerful, offers to provide her any evidence she likes, and she considers and rejects the "healing the sick" and "resurrecting the dead" plans?

Comment author: Blueberry 27 June 2010 08:30:32PM 1 point [-]

Not to mention a solution to the P=NP problem (or the Riemann Hypothesis)?

Comment author: Oscar_Cunningham 27 June 2010 07:01:37PM 1 point [-]

Nothing is provable to the level you demand (well, pretty much nothing, cogito ergo sum and all that). Given that none of the omni* are well defined, the question doesn't mean much either.

Comment author: NancyLebovitz 27 June 2010 02:59:48PM 4 points [-]

News and mental focus

RIKI OTT: Exxon never said it in a press conference. Just when the media started to ask questions, where did that 10.8 million gallons come from, has it been independently verified, Frank Iarossi, the owner of Exxon Shipping, at a press conference said, alcohol may be involved. And I kid you not, I witnessed the entire international media just switch tracks, and that was how we got 10.8 million gallons, rounded up to 11.

A couple years later, when I saw the movie Wag the Dog, I saw that scene where the president was just about to get nailed, and a plant in the audience says, well, what about the bombs in Albania? And the whole media switched to bombs in Albania. And I rose up out of my seat, and I said, that is how we got 11 million gallons. And my two friends each grabbed a wrist and pulled me back down into my chair. And I just swore that I would never forget 38 million gallons.

Comment author: xamdam 27 June 2010 04:27:13PM *  1 point [-]

I think Derren Brown uses this as a mind hack a lot.: http://www.youtube.com/watch?v=3Vz_YTNLn6w (notice specific diversion into spatial memory, it's probably been tried and tested as the best distraction from the color of money in hand)

I feel that mental focus if VERY weak and very exploitable.

As a side note, I think there is another, less obious, mental hack going on, on the audience. Derren claims (in the intro to this TV series) that there is no acting here, but a lot of misdirection. I believe it. I think when he shows this trick work 2 out of 3 times, it's probably more like 2 out of 30. My guess is that he biases the sample quite cleverly, showing 3 cases is exactly the minimum that you can show giving the impression that a) reporting is honest (see - I showed a failure!) and b) the 'magic' works in most cases. Also I think getting caught/embarrassed by a hot dog vendor evokes certain associations that yeah, he can be beat which prevent you from thinking how much he can be beat.

Here is to you Derren, Master of Dark Arts.

Comment author: Morendil 27 June 2010 08:06:59PM 0 points [-]

diversion into spatial memory

Oh. I thought the point of the subway anecdote jnf gb unir na rkphfr gb fyvc va n cerfhccbfvgvba, va gur sbez "Gnxr vg [gur zbarl], vg'f svar".

Comment author: xamdam 27 June 2010 08:32:56PM *  0 points [-]

Yes, I missed it, largely due to lack of knowledge of NLP. I wouldn't be surprised if the spatial thing is true also, (and possibly intended) making people picture something is supposed to make them look up IIRC.

Comment author: Vladimir_M 27 June 2010 08:16:08PM 4 points [-]

Note however that Derren Brown's tricks have turned out to be staged in at least one instance. This makes me extremely skeptical towards the rest of them too.