Open Thread: December 2009

3 Post author: CannibalSmith 01 December 2009 04:25PM

ITT we talk about whatever.

Comments (263)

Sort By: Controversial
Comment author: byrnema 02 December 2009 06:59:37AM *  0 points [-]

A poem, not a post:

Intelligence is not computation.

As you know.

Yet the converse bears … contemplation, reputation. Only then refutation.

We are irritated by our fellows that observe that A mostly implies B, and B mostly implies C, but they will not, will not concede that A implies C, to any extent.

We consider this; an error in logic, an error in logic.

Even though! we know: intelligence is not computation.

Intelligence is finding the solution in the space of the impossible. I don’t mean luck At all. I mean: while mathematical proofs are formal, absolute, without question, convincing, final,

We have no Method, no method for their generation. As well we know:

No computation can possibly be found to generate, not possibly. Not systematically, not even with ingenuity. Yet, how and why do we know this -- this Impossibility?

Intelligence is leaping, guessing, placing the foot unexpectedly yet correctly. Which you find verified always afterwards, not before.

Of course that’s why humans don’t calculate correctly.

But we knew that.

You and I, being too logical about it, pretending that computation is intelligence.

But we know that; already, everything. That pretending is the part of intelligence not found in the Computating. Yet, so? We’ll pretend that intelligence is computing and we’ll see where the computation fails! Telling us what we already knew but a little better.

Than before, we’ll see afterwards. How ingenuous, us.

The computation will tell us, finally so, we'll pretend.

Comment author: gwern 12 December 2009 03:05:46AM *  2 points [-]

While reading a collection of Tom Wayman's poetry, suddenly a poem came to me about Hal Finney ("Dying Outside"); since we're contributing poems, I don't feel quite so self-conscious. Here goes:

He will die outside, he says.
Flawed flesh betrayed him,
it has divorced him -
for the brain was left him,
but not the silverware
nor the limbs nor the car.
So he will take up
a brazen hussy,
tomorrow's eve,
a breather-for-him,
a pisser-for-him.
He will be letters,
delivered slowly;
deliberation
his future watch-word.
He would not leave until he left this world.
I try not to see his mobile flesh,
how it will sag into eternal rest,
but what he will see:
symbol and symbol, in their endless braids,
and with them, spread over strange seas of thought
mind (not body), forever voyaging.

http://www.gwern.net/fiction/Dying%20Outside

Comment author: byrnema 02 December 2009 05:58:54PM *  2 points [-]

I wrote this poem yesterday in an unusual mood. I don't entirely agree with it today. Or at least, I would qualify it.

What is meant by computation? When I wrote that intelligence is not computation, I must have meant a certain sort of computation because of course all thought is some kind of computation.

To what extent has distinction been made between systematic/linear/deductive thought (which I am criticizing as obviously limited in the poem) and intelligent pattern-based thought? Has there been any progress in characterizing the latter?

For example, consider the canonical story about Gauss. To keep him busy with a computation, his math teacher told him to add all the numbers from 1 to 100. Instead, according to the story, Gauss added the first number and the last number, multiplied by 100 and divided by 2. Obviously, this is a computation. But yet a different sort. To what extent do you suppose he logically deduced the pattern of the lowest number and highest number combining always to single value or just guessed/observed it was a pattern that might work? And then found that it did work inductively?

I'm very interested in characterizing the difference between these kinds of computation. Intelligent thinking seems to really be guesses followed by verification, not steady linear deduction.

Comment author: JamesAndrix 03 December 2009 03:12:12AM *  1 point [-]

What is meant by computation? When I wrote that intelligence is not computation, I must have meant a certain sort of computation because of course all thought is some kind of computation.

Gah, Thank You. Saves me the trouble of a long reply. I'll upvote for a change-of-mind disclaimer in the original.

Intelligent thinking seems to really be guesses followed by verification, not steady linear deduction.

My recent thoughts have been along these lines, but this is also what evolution does. At some point, the general things learned by guessing have to be incorporated into the guess-generating process.

Comment author: bgrah449 02 December 2009 04:33:57PM 1 point [-]

I've read Newcomb's problem (Omega, two boxes, etc.), but I was wondering if, shortly, "Newcomb's problem is when someone reliably wins as a result of acting on wrong beliefs." Is Peter walking on water a special case of Newcomb? Is the story from Count of Monte Cristo, about Napoleon attempting suicide with too much poison and therefore surviving, a special case of Newcomb?

Comment author: Tyrrell_McAllister 02 December 2009 06:33:08PM *  3 points [-]

No, that's not the case. A one-boxer in Newcomb's problems is acting with entirely correct beliefs. All agree that the one-boxer will get more money than the two-boxer. That correct belief is what motivates the one-boxer.

The scenarios that you describe sound somewhat (but not exactly) like Gettier problems to me.

(I wasn't the downvoter.)

Comment author: bgrah449 02 December 2009 05:44:44PM *  6 points [-]

I am completely baffled why this would be downvoted. I guess asking a question in genuine pursuit of knowledge, in an open thread, is wasting someone's time, or is offensive.

I like to think someone didn't have the time to write "No, that's not the case," and wished, before dashing off, leaving nothing but a silhouette of dust, that their cryptic, curt signal would be received as intended; that as they hurry down the underground tunnel past red, rotating spotlights, they hoped against hope that their seed of truth landed in fertile ground -- godspeed, downvote.

Comment author: Morendil 02 December 2009 06:50:10PM 0 points [-]

The community would benefit from a convention of "no downvotes in Open Thread".

However, I did find your question cryptic; you're dragging into a decision theory problem historical and religious referents that seem to have little to do with it. You need to say more if you really want an answer to the question.

Comment author: bgrah449 02 December 2009 07:31:15PM 1 point [-]

Sure, that's fair.

Peter walked on water out to Jesus because he thought he could; when he looked down and saw the sea, the fell in. As long as he believed Jesus instead of his experience with the sea, he could walk on water.

I don't think the Napoleon story is true, but that's beside the point. He thought he was so tough that an ordinary dose of poison wouldn't kill him, so he took six times the normal dosage. This much gave his system such a shock that the poison was rejected and he lived, thinking to himself, "Damn, I underestimated how incredibly fantastic I am." As long as he (wrongly) believed in his own exceptionalism instead of his experience with poison on other men, he was immune to the poison.

My train of thought was, you have a predictor and a chooser, but that's just getting you to a point where you choose either "trust the proposed worldview" or "trust my experience to date" - do you go for the option that your prior experience tells you shouldn't work (and hope your prior experience was wrong) or do you go with your prior experience (and hope the proposed worldview is wrong)?

I understand that in Newcomb's, that what Omega says is true. But change it up to "is true way more than 99% of the time but less than 100% of the time" and start working your way down that until you get to "is false way more than 99% of the time but less than 100% of the time" and at some point, not that long after you start, you get into situations very close to reality (I think, if I'm understanding it right).

This basically started from trying to think about who, or what, in real life takes on the Predictor role, who takes on the Belief-holder role, who takes on the Chooser role, and who receives the money, and seeing if anything familiar starts falling out if I spread those roles out to more than 2 people or shrink them down to a single person whose instincts implore them to do something against the conclusion to which their logical thought process is leading them.

Comment author: Morendil 02 December 2009 07:47:15PM -2 points [-]

You seem to be generalizing from fictional evidence, which is frowned upon here, and may explain the downvote (assuming people inferred the longer version from your initial question).

Comment author: bgrah449 02 December 2009 08:24:07PM 2 points [-]

That post (which was interesting and informative - thanks for the link) was about using stories as evidence for use in predicting the actual future, whereas my question is about whether these fictional stories are examples of a general conceptual framework. If I asked if Prisoner's Dilemma was a special case of Newcomb's, I don't think you'd say, "We don't like generalizing from fictional evidence."

Which leads, ironically, to the conclusion that my error was generalizing from evidence which wasn't sufficiently fictional.

Comment author: Morendil 02 December 2009 09:04:54PM 1 point [-]

Perhaps I jumped to conclusions. Downvotes aren't accompanied with explanations, and groping for one that might fit I happened to remember the linked post. More PC than supposing you were dinged just for a religious allusion. (The Peter reference at least required no further effort on my part to classify as fictional; I had to fact-check the Napoleon story, which was an annoyance.)

It still seems the stories you're evoking bear no close relation to Newcomb's as I understand it.

Comment author: gwern 06 December 2009 02:22:08AM 1 point [-]

I upvoted you because your second sentence painted a story that deeply amused me.

/pats seed of truth, and pours a little water on it

Comment author: Zack_M_Davis 04 December 2009 01:33:47AM 1 point [-]

With Channukah right around the corner, it occurs to me that "Light One Candle" becomes a transhumanist/existential-risk-reduction song with just a few line edits.

Light one candle for all humankind's children
With thanks that their light didn't die
Light one candle for the pain they endured
When the end of their world was surmised
Light one candle for the terrible sacrifice
Justice and freedom demand
But light one candle for the wisdom to know
When the singleton's time is at hand

Don't let the light go out! (&c.)

Comment author: Jordan 04 December 2009 02:34:37AM 3 points [-]

Whether or not a singleton is the best outcome or not, I'm way too uncomfortable with the idea to be singing songs of praise about it.

Comment author: Zack_M_Davis 04 December 2009 03:13:20AM 1 point [-]

Upvoted. I'm actually really uncomfortable with the idea, too. My comment above is meant in a silly and irreverent manner (cf. last month), and the substitution for "peacemaker" was too obvious to pass up.

Comment deleted 05 December 2009 01:00:21PM [-]
Comment author: wedrifid 05 December 2009 01:18:09PM 1 point [-]

I strongly suspect that lesswrong.com has an ideological bias in favor of "morality."

It does.

There has been no proof that rationality requires morality. Yet I suspect that posts coming from a position of moral nihilism would not be welcomed.

In general they are not. But I find that a high quality clearly rational reply that doesn't adopt the politically correct morality will hover around 0 instead of (say) 8. You can then post a couple of quotes to get your karma fix if desired.

Comment author: Alicorn 01 December 2009 05:55:11PM 6 points [-]

I wrote a short story with something of a transhumanism theme. People can read it here. Actionable feedback welcome; it's still subject to revision.

Note: The protagonist's name is "Key". Key, and one other character, receive Spivak pronouns, which can make either Key's name or eir pronouns look like some kind of typo or formatting error if you don't know it's coming. If this annoys enough people, I may change Key's name or switch to a different genderless pronoun system. I'm curious if anyone finds that they think of Key and the other Spivak character as having a particular gender in the story; I tried to write them neither, but may have failed (I made errors in the pronouns in the first draft, and they all went in one direction).

Comment author: [deleted] 02 December 2009 07:05:42AM 1 point [-]

Gur qrngu qvqa'g srry irel qrngu-yvxr. Vg frrzrq yvxr gur rzbgvba fheebhaqvat vg jnf xvaq bs pbzcerffrq vagb bar yvggyr ahttrg gung V cerggl zhpu fxvzzrq bire. V jnf nyfb rkcrpgvat n jbeyq jvgubhg qrngu, juvpu yrsg zr fhecevfrq. Va gur erny jbeyq, qrngu vf bsgra n fhecevfr, ohg yvxr va gur erny jbeyq, fhqqra qrngu va svpgvba yrnirf hf jvgu n srryvat bs qvforyvrs. Lbh pbhyq unir orra uvagvat ng gung qrngu sebz gur svefg yvar.

Nyfb, lbh xvaq bs qevir evtug cnfg gur cneg nobhg birecbchyngvba. V guvax birecbchyngvba vf zl zbgure'f znva bowrpgvba gb pelbavpf.

Comment author: gwern 11 December 2009 06:52:51PM 0 points [-]

Alicorn goes right past it probably because she's read a fair bit of cyronics literature herself and has seen the many suggestions (hence the librarian's invitation to think of 'a dozen solutions'), and it's not the major issue anyway.

Comment author: Emily 03 December 2009 04:42:45AM 0 points [-]

I enjoyed it. I made an effort to read Key genderlessly. This didn't work at first, probably because I found the Spivak pronouns quite hard to get used to, and "ey" came out as quite male to me, then fairly suddenly flipped to female somewhere around the point where ey was playing on the swing with Trellis. I think this may have been because Trellis came out a little more strongly male to me by comparison (although I was also making a conscious effort to read ey genderlessly). But as the story wore on, I improved at getting rid of the gender and by the end I no longer felt sure of either Key or Trellis.

Point of criticism: I didn't find the shift between what was (to me) rather obviously the two halves of the story very smooth. The narrative form appeared to take a big step backwards from Key after the words "haze of flour" and never quite get back into eir shoes. Perhaps that was intentional, because there's obviously a huge mood shift, but it left me somewhat dissatisfied about the resolution of the story. I felt as though I still didn't know what had actually happened to the original Key character.

Comment author: rwallace 01 December 2009 07:23:34PM 2 points [-]

You traded off a lot of readability for the device of making the protagonist's gender indeterminate. Was this intended to serve some literary purpose that I'm missing? On the whole the story didn't seem to be about gender.

I also have to second DanArmak's comment that if there was an overall point, I'm missing that also.

Comment author: Alicorn 01 December 2009 07:28:44PM 2 points [-]

Key's gender is not indeterminate. Ey is actually genderless. I'm sorry if I didn't make that clear - there's a bit about it in eir second conversation with Trellis.

Comment author: gwern 11 December 2009 06:54:46PM 0 points [-]

I'm sorry if I didn't make that clear - there's a bit about it in eir second conversation with Trellis.

I thought it was pretty clear. The paragraph about 'boy or girl' make it screamingly obvious to me, even if the Spivak or general gender-indeterminacy of the kids hadn't suggested it.

Comment author: Liron 03 December 2009 06:11:08PM 2 points [-]

Your gender pronouns just sapped 1% of my daily focusing ability.

Comment author: Blueberry 04 December 2009 04:34:10AM 1 point [-]

Looks like I'm in the minority for reading Key as slightly male. I didn't get a gender for Trellis. I also read the librarian as female, which I'm kind of sad about.

I loved the story, found it very touching, and would like to know more about the world it's in. One thing that confused me: the librarian's comments to Key suggested that some actual information was withheld from even the highest levels available to "civilians". So has someone discovered immortality, but some ruling council is keeping it hidden? Or is it just that they're blocking research into it, but not hiding any actual information? Are they hiding the very idea of it? And what's the librarian really up to?

Were you inspired by Nick Bostrom's "Fable of the Dragon"? It also reminded me a little of Lois Lowry's "The Giver".

Thanks so much for sharing it with us!

Comment author: Alicorn 04 December 2009 04:44:31AM 0 points [-]

I also read the librarian as female, which I'm kind of sad about.

Lace is female - why are you sad about reading her that way?

I loved the story, found it very touching, and would like to know more about the world it's in.

Yaaaay! I'll answer any setting questions you care to pose :)

So has someone discovered immortality, but some ruling council is keeping it hidden? Or is it just that they're blocking research into it, but not hiding any actual information? Are they hiding the very idea of it? And what's the librarian really up to?

Nobody has discovered it yet. The communities in which Key's ilk live suppress the notion of even looking for it; in the rest of the world they're working on it in a few places but aren't making much progress. The librarian isn't up to a whole lot; if she were very dedicated to finding out how to be immortal she'd have ditched the community years ago - she just has a few ideas that aren't like what the community leaders would like her to have and took enough of a shine to Key that she wanted to share them with em. I have read both "Fable of the Dragon" and "The Giver" - the former I loved, the latter I loved until I re-read it with a more mature understanding of worldbuilding, but I didn't think of either consciously when writing.

You are most welcome for the sharing of the story. Have a look at my other stuff, if you are so inclined :)

Comment author: LucasSloan 03 December 2009 04:54:32AM 1 point [-]

For me, both of the characters appeared female.

Sbe zr gur fgbel fbeg bs oebxr qbja whfg nf Xrl'f sevraq jnf xvyyrq. Vg frrzrq gbb fbba vagb gur aneengvir gb znxr fhpu n znwbe punatr. Nyfb, jvgu erfcrpg gb gur zbeny, vg frrzrq vafhssvpvragyl fubja gung lbh ernyyl ertneq cnva nf haqrfvenoyr - vg frrzrq nf gubhtu lbh pbhyq or fnlvat fbzrguvat nybat gur yvarf bs "gurl whfg qba'g haqrefgnaq." Orpnhfr bs gung nf jryy nf gur ehfurq srry bs gur raqvat, vg fbeg bs pnzr bss nf yrff rzbgvbanyyl rssrpgvir guna vg pbhyq.

Comment author: Zack_M_Davis 02 December 2009 07:41:46PM 3 points [-]

I love the new gloss on "What do you want to be when you grow up?"

or switch to a different genderless pronoun system.

Don't. Spivak is easy to remember because it's just they/them/their with the ths lopped off. Nonstandard pronouns are difficult enough already without trying to get people to remember sie and hir.

Comment author: ChrisPine 02 December 2009 12:20:08PM 1 point [-]

I liked it. :)

Part of the problem that I had, though, was the believability of the kids: kids don't really talk like that: "which was kind of not helpful in the not confusing me department, so anyway"... or, in an emotionally painful situation:

Key looked suspiciously at the librarian. "You sound like you're trying not to say something."

Improbably astute, followed by not seeming to get the kind of obvious moral of the story. At times it felt like it was trying to be a story for older kids, and at other times like it was for adults.

The gender issue didn't seem to add anything to the story, but it only bothered me at the beginning of the story. Then I got used to it. (But if it doesn't add to the story, and takes getting used to... perhaps it shouldn't be there.)

Anyway, I enjoyed it, and thought it was a solid draft.

Comment author: Alicorn 02 December 2009 02:47:23PM 0 points [-]

You've hit on one of my writing weaknesses: I have a ton of trouble writing people who are just plain not very bright or not very mature. I have a number of characters through whom I work on this weakness in (unpublished portions of) Elcenia, but I decided to let Key be as smart I'm inclined to write normally for someone of eir age - my top priority here was finishing the darn thing, since this is only the third short story I can actually claim to have completed and I consider that a bigger problem.

Comment author: Blueberry 04 December 2009 04:36:31AM 1 point [-]

I actually have to disagree with this. I didn't think Key was "improbably astute". Key is pretty clearly an unusual child (at least, that's how I read em). Also, the librarian was pretty clearly being elliptical and a little patronizing, and in my experience kids are pretty sensitive to being patronized. So it didn't strike me as unbelievable that Key would call the librarian out like that.

Comment author: Kaj_Sotala 20 December 2009 04:29:58PM 0 points [-]

Finally got around reading the story. I liked it, and finishing it gave me a wild version of that "whoa" reaction you get when you've doing something emotionally immersive and then switch to some entirely different activity.

I read Key as mostly genderless, possibly a bit female because the name sounded feminine to me. Trellis, maybe slightly male, though that may also have been from me afterwards reading the comments about Trellis feeling slightly male and those contaminating the memory.

I do have to admit that the genderless pronouns were a bit distracting. I think it was the very fact that they were shortened version of "real" pronouns that felt so distracting - my mind kept assuming that it had misread them and tried to reread. In contrast, I never had an issue with Egan's use of ve / ver / vis / vis / verself.

Comment author: Larks 04 December 2009 11:18:56PM 0 points [-]

I got used to the Spivak after a while, and while it'd be optimal for an audience used to it, it did detract a little at first. On the whole I'd say it's necessary though (if you were going to use a gender'd pronoun, I'd use female ones)

I read Key as mainly female, and Trellis as more male- it would be interesting to know how readers' perceptions correlated with their own gender.

The children seemed a little mature, but I thought they'd had a lot better education, or genetic enhancement or something. I think spending a few more sentences on the important events would be good though- otherwise one can simply miss them.

I think you were right to just hint at the backstory- guessing is always fun, and my impression of the world was very similar to that which you gave in one of the comments.

Comment author: arundelo 04 December 2009 10:13:30PM 0 points [-]

Great story!

I kept thinking of Key as female. This may be because I saw some comments here that saw em as female, or because I know that you're female.

The other character I didn't assign a sex to.

Comment author: NancyLebovitz 04 December 2009 03:05:50PM 0 points [-]

I enjoyed the story-- it was an interesting world. By the end of the story, you were preaching to a choir I'm in.

None of the characters seemed strongly gendered to me.

I was expecting opposition to anesthesia to include religiously based opposition to anesthesia for childbirth, and for the whole idea of religion to come as a shock. On the other hand, this might be cliched thinking on my part. Do they have religion?

The neuro couldn't be limited to considered reactions-- what about the very useful fast reflexive reaction to pain?

Your other two story links didn't open.

Comment author: Alicorn 04 December 2009 03:09:29PM *  0 points [-]

Religion hasn't died out in this setting, although it's uncommon in Key's society specifically. Religion was a factor in historical opposition to anesthesia (I'm not sure of the role it plays in modern leeriness about painkillers during childbirth) but bringing it up in more detail would have added a dimension to the story I didn't think it needed.

Reflexes are intact. The neuro just translates the qualium into a bare awareness that damage has occurred. (I don't know about everyone, but if I accidentally poke a hot burner on the stove, my hand is a foot away before I consciously register any pain. The neuro doesn't interfere with that.)

I will check the links and see about fixing them; if necessary, I'll HTMLify those stories too. ETA: Fixed; they should be downloadable now.

Comment author: Jack 26 December 2009 12:02:51AM 8 points [-]

Sharing my Christmas (totally non-supernatural) miracle:

My theist girlfriend on Christmas Eve: "For the first time ever I went to mass and thought it was ridiculous. I was just like, this is nuts. The priest was like 'oh, we have to convert the rest of the world, the baby Jesus spoke to us as an infant without speaking, etc.' I almost laughed."

Comment author: Eliezer_Yudkowsky 26 December 2009 12:12:19AM 2 points [-]

This made me say "Awwwwwwww..."

Did LW play a part, or was she just browsing the Internet?

Comment author: Jack 26 December 2009 12:45:23AM 2 points [-]

Well I think I was the first vocal atheist she had ever met so arguments with me and me making fun of superstition while not being a bad person were probably crucial. Some Less Wrong stuff probably got to her through me, though. I should find something to introduce the site to her, though I doubt she would ever spend a lot of time here.

Comment author: timtyler 15 December 2010 07:11:36PM 4 points [-]
Comment author: ata 15 December 2010 07:20:59PM *  3 points [-]

Thanks for posting this (I didn't know the videos were up), though you've posted it in the December 2009 open thread.

(The current open thread is in the Discussion section, though this may be worth its own Discussion post.)

Comment author: timtyler 15 December 2010 07:40:40PM 1 point [-]

Oops - perhaps someone else can harvest the karma for spreading it, then...

Comment author: MBlume 15 December 2010 07:19:42PM 1 point [-]

Awesome, thanks ^_^

Comment author: Johnicholas 30 December 2009 01:26:21PM *  1 point [-]

Hmm, this "mentat wiki" seems to have some reasonably practical intelligence (and maybe rationality) techniques.

Comment author: Cyan 03 January 2010 02:24:13AM 0 points [-]

Huh. Looking at that wiki drove home how much I ReallyHate CamelCase.

Comment author: Eliezer_Yudkowsky 30 December 2009 07:00:23PM 0 points [-]

I'd say go ahead and post a link to this - if they've got any especially good sample techniques, be sure to include a pointer there.

Comment author: Corey_Newsome 28 December 2009 08:24:50AM 2 points [-]

I'm going to repeat my request (for the last time) that the most recent Open Thread have a link in the bar up top, between 'Top' and 'Comments', so that people can reach it a tad easier. (Possible downside: people could amble onto the site and more easily post time-wasting nonsense.)

Comment author: wedrifid 28 December 2009 01:16:58PM 0 points [-]

If I recall correctly this request is already on the implementation queue.

Comment author: Psy-Kosh 27 December 2009 03:42:14PM 3 points [-]

If, say, I have a basic question, is it appropriate to post it to open thread, to a top level post, or what? ie, say if I'm working through Pearl's Causality and am having trouble deriving something... or say I've stared at the wikipedia pages for ages and STILL don't get the difference between Minimum Description Length and Minimum Message Length... is LW an appropriate place to go "please help me understand this", and if so, should I request it in a top level post or in an open thread or...

More generally: LW is about developing human rationality, but is it appropriate for questions about already solved aspects of rationality? like "please help me understand the math for this aspect of reasoning" or even "I'm currently facing this question in my life or such, help me reason through this please?"

Thanks.

Comment author: Kaj_Sotala 27 December 2009 03:57:56PM 1 point [-]

I think that kind of a question is fine in the Open Thread.

Comment author: Psy-Kosh 27 December 2009 05:00:31PM 0 points [-]

Okay, thanks. In that case, I am asking indeed about the difference between MML and MDL. I've stared at the wikipedia pages, including the bits that supposedly explain the difference, and I'm still going "huh?"

Comment author: DanArmak 27 December 2009 03:46:51PM 2 points [-]

More generally: LW is about developing human rationality, but is it appropriate for questions about already solved aspects of rationality?

Most posts here are written by someone who understands an aspect of rationality, to explain it to those who don't. I see no reason not to ask questions in the open thread. I think they should be top-level posts only if you anticipate a productive discussion around them; most already-solved questions can be answered with a single comment and that would be that, so no need for a separate post.

Comment author: Psy-Kosh 27 December 2009 05:01:07PM 0 points [-]

Okay, thanks. In that case, as I replied to Kaj Sotala, I am indeed asking about the difference between MML and MDL

Comment author: byrnema 22 December 2009 10:28:52PM *  1 point [-]

I think this is close to the question that has been lurking in my mind for some time: Why optimize our strategies to achieve what we happen to want, instead of just modifying what we want?

Suppose, for my next question, that it was trivial to modify what we want. Is there some objective meta-goal we really do need to pay attention to?

Comment author: EStokes 30 December 2009 01:08:05PM *  2 points [-]

Well, if we modified what we wanted, we wouldn't get what we originally wanted because we wouldn't want to...

Can't think of anything else off the top of my head.

Comment author: MrHen 30 December 2009 10:36:35PM 0 points [-]

tut's quip holds the key:

But in general the answer is another question: Why would you want to do that?

But to expound on it a bit further, if I want to drive to Dallas to see a band play I can (a) figure out a strategy to get there or (b) stop wanting to go. Assuming that (b) is even possible, it isn't actually a solution to the problem of how to get to Dallas. Applying the same principle to all Wants does not provide you with a way to always get what you want. Instead, it helps you avoid not getting what you want.

If you wanted nothing more than to avoid disappointment by not getting what you want than the safest route is to never desire anything that isn't a sure thing. Or simply not want anything at all. But a simpler route to this whole process is to ditch that particular Want first. The summation is a bit wordy and annoying, but it ends like this:

Instead of wanting to avoid not getting what you want by not wanting anything else, simply forgo the want that is pushing you to avoid not getting what you want.

With parens:

Instead of (wanting to avoid {not getting what you want} by not wanting anything else), simply forgo (the want that is pushing you to avoid {not getting what you want}).

In other words, you can achieve the same result that modifying your wants would create by not getting too disappointed if you don't get what you want. Or, don't take it so hard when things don't go your way.

Hopefully that made some sense (and I got it right.)

Comment author: byrnema 30 December 2009 10:46:46PM *  0 points [-]

Thank you for your responses, but I guess my question wasn't clear. I was asking about purpose. If there's no point in going to Dallas, why care about wanting to go to Dallas?

This is my problem if there's no objective value (that I tried to address more directly here). If there's no value to anything I might want, why care about what I want, much less strive for what I want?

I don't know if there's anything to be done. Whining about it is pointless. If anyone has a constructive direction, please let me know. I picked up a Sartre's , "Truth and Existence" rather randomly, maybe it will lead in a different (hopefully more interesting) direction.

Comment author: orthonormal 31 December 2009 04:55:29AM *  5 points [-]

I second the comments above. The answer Alicorn and Furcas give sounds really shallow compared to a Framework Of Objective Value; but when I become convinced that there really is no FOOV, I was relieved to find that I still, you know, wanted things, and these included not just self-serving wants, but things like "I want my friends and family to be happy, even in circumstances where I couldn't share or even know of their happiness", and "I want the world to become (for example) more rational, less violent, and happier, even if I wouldn't be around to see it— although if I had the chance, I'd rather be around to see it, of course".

It doesn't sound as dramatic or idealistic as a FOOV, but the values and desires encoded in my brain and the brains of others have the virtue of actually existing; and realizing that these values aren't written on a stone tablet in the heart of the universe doesn't rob them of their importance to the life I live.

Comment author: Alicorn 30 December 2009 11:41:01PM *  4 points [-]

If there's no value to anything I might want, why care about what I want, much less strive for what I want?

Because in spite of everything, you still want it.

Or: You can create value by wanting things. If things have value, it's because they matter to people - and you're one of those, aren't you? Want things, make them important - you have that power.

Comment author: byrnema 31 December 2009 05:30:07AM *  0 points [-]

Because in spite of everything, you still want it.

Maybe I wouldn't. There have been times in my life when I've had to struggle to feel attached to reality, because it didn't feel objectively real. Now if value isn't objectively real, I might find myself again feeling indifferent, like one part of myself is carrying on eating and driving to work, perhaps socially moral, perhaps not, while another part of myself is aware that nothing actually matters. I definitely wouldn't feel integrated.

I don't want to burden anyone with what might be idiosyncratic sanity issues, but I do mention them because I don't think they're actually all that idiosyncratic.

Comment author: Alicorn 31 December 2009 06:01:16AM 1 point [-]

Can you pick apart what you mean by "objectively"? It seems to be a very load-bearing word here.

Comment author: byrnema 31 December 2009 01:52:15PM 0 points [-]

I thought this was a good question, so I took some time to think about it. I am better at recognizing good definitions than generating them, but here goes:

'Objective' and 'subjective' are about the relevance of something across contexts.

Suppose that there is some closed system X. The objective value of X is its value outside X. The subjective value of X is its value inside X.

For example, if I go to a party and we play a game with play money, then the play money has no objective value. I might care about the game, and have fun playing it with my friends, but it would be a choice whether or not to place any subjective attachment to the money; I think that I wouldn't and would be rather equanimous about how much money I had in any moment. If I went home and looked carefully at the money to discover that it was actually a foreign currency, then it turns out that the money had objective value after all.

Regarding my value dilemma, the system X is myself. I attach value to many things in X. Some of this attachment feels like a choice, but I hazard that some of this attachment is not really voluntary. (For example, I have mirror neurons.) I would call these attachments 'intellectual' and 'visceral' respectively.

Generally, I do not have much value for subjective experience. If something only has value in 'X', then I have a tendency to negate that as a motivation. I'm not altruistic, I just don't feel like subjective experience is very important. Upon reflection, I realize that re: social norms, I actually act rather selfishly when I think I'm pursuing something with objective value.

If there's no objective value, then at the very least I need to do a lot of goal reorganization; losing my intellectual attachments unless they can be recovered as visceral attachments. At the worst, I might feel increasingly like I'm a meaningless closed system of self-generated values. At this point, though, I doubt I'm capable of assimilating an absence of objective value on all levels -- my brain might be too old -- and for now I'm just academically interested in how self-validation of value works without feeling like its an illusion.

Comment author: AdeleneDawner 31 December 2009 02:48:10PM 1 point [-]

I know this wasn't your main point, but money doesn't have objective value, either, by that definition. It only has value in situations where you can trade it for other things. It's extremely common to encounter such situations, so the limitation is pretty ignorable, but I suspect you're at least as likely to encounter situations where money isn't tradeable for goods as you are to encounter situations where your own preferences and values aren't part of the context.

Comment author: byrnema 31 December 2009 03:16:10PM *  0 points [-]

I used the money analogy because it has a convenient idea of value.

While debating about the use of that analogy, I had already considered it ironic that the US dollar hasn't had "objective" value since it was disconnected from the value of gold in 1933. Not that gold has objective value unless you use it to make a conductor. But at the level, I start losing track of what I mean by 'value'. Anyway, it is interesting that the value of the US dollar is exactly an example of humans creating value, echoing Alicorn's comment.

Real money does have objective value relative to the party, since you can buy things on your way home, but no objective value outside contexts where the money can be exchanged for goods.

Comment author: Alicorn 31 December 2009 03:00:37PM 0 points [-]

If you are a closed system X, and something within system X only has objective value inasmuch as something outside X values it, then does the fact that other people care about you and your ability to achieve your goals help? They are outside X, and while their first-order interests probably never match yours perfectly, there is a general human tendency to care about others' goals qua others' goals.

Comment author: byrnema 31 December 2009 03:54:01PM *  0 points [-]

then does the fact that other people care about you and your ability to achieve your goals help?

If you mean that I might value myself and my ability to achieve my goals more because I value other people valuing that, then it does not help. My valuation of their caring is just as subjective as any other value I would have.

On the other hand, perhaps you were suggesting that this mutual caring could be a mechanism for creating objective value, which is kind of in line with what I think. For that matter, I think that my own valuation of something, even without the valuation of others, does create objective value -- but that's a FOOM. I'm trying to imagine reality without that.

Comment author: Alicorn 31 December 2009 04:18:32PM 1 point [-]

If you mean that I might value myself and my ability to achieve my goals more because I value other people valuing that, then it does not help. My valuation of their caring is just as subjective as any other value I would have.

That's not what I mean. I don't mean that their caring about you/your goals makes things matter because you care if they care. I mean that if you're a closed system, and you're looking for a way outside of yourself to find value in your interests, other people are outside you and may value your interests (directly or indirectly). They would carry on doing this, and this would carry on conferring external value to you and your interests, even if you didn't give a crap or didn't know anybody else besides you existed - how objective can you get?

On the other hand, perhaps you were suggesting that this mutual caring could be a mechanism for creating objective value

I don't think it's necessary - I think even if you were the only person in the universe, you'd matter, assuming you cared about yourself - and I certainly don't think it has to be really mutual. Some people can be "free riders" or even altruistic, self-abnegating victims of the scheme without the system ceasing to function. So this is a FOOV? So now it looks like we don't disagree at all - what was I trying to convince you of, again?

Comment author: MrHen 31 December 2009 03:54:17PM *  0 points [-]

Maybe I wouldn't. There have been times in my life when I've had to struggle to feel attached to reality, because it didn't feel objectively real. Now if value isn't objectively real, I might find myself again feeling indifferent, like one part of myself is carrying on eating and driving to work, perhaps socially moral, perhaps not, while another part of myself is aware that nothing actually matters. I definitely wouldn't feel integrated.

Yeah, I think I can relate to that. This edges very close to an affective death spiral, however, so watch the feedback loops.

The way I argued myself out of mine was somewhat arbitrary and I don't have it written up yet. The basic idea was taking the concepts that I exist and that at least one other thing exists and, generally speaking, existence is preferred over non-existence. So, given that two things exist and can interact and both would rather be here than not be here, it is Good to learn the interactions between the two so they can both continue to exist. This let me back into accepting general sensory data as useful and it has been a slow road out of the deep.

I have no idea if this is relevant to your questions, but since my original response was a little off maybe this is closer?

Comment author: byrnema 02 January 2010 12:57:15AM 0 points [-]

The way I argued myself out of mine was somewhat arbitrary and I don't have it written up yet.

This paragraph (showing how you argued yourself out of some kind of nihilism) is completely relevant, thanks. This is exactly what I'm looking for.

The basic idea was taking the concepts that I exist and that at least one other thing exists and, generally speaking, existence is preferred over non-existence.

What do you mean by, "existence is preferred over non-existence"? Does this mean that in the vacuum of nihilism, you found something that you preferred, or that it’s better in some objective sense?

My situation is that if I try to assimilate the hypothesis that there is no objective value (or, rather, I anticipate trying to do so), then immediately I see that all of my preferences are illusions. It's not actually any better if I exist or don't exist, or if the child is saved from the tracks or left to die. It's also not better if I choose to care subjectively about these things (and be human) or just embrace nihilism, if that choice is real. I understand that caring about certain sorts of these things is the product of evolution, but without any objective value, I also have no loyalty to evolution and its goals -- what do I care about the values and preferences it instilled in me?

The question is; how has evolution actually designed my brain; in the state 'nihilism' does my brain (a) abort intellectual thinking (there's no objective value to truth anyway) and enter a default mode of material hedonism that acts based on preferences and impulses just because they exist and that’s what I’m programmed to do or (b) does it cling to its ability to think beyond that level of programming, and develop this separate identity as a thing that knows that nothing matters?

Perhaps I’m wrong, but your decision to care about the preference of existence over non-existence and moving on from there appears to be an example of (a). Or perhaps a component (b) did develop and maintain awareness of nihilism, but obviously that component couldn’t be bothered posting on LW, so I heard a reply from the part of you that is attached to your subjective preferences (and simply exists).

Comment author: MrHen 06 January 2010 04:48:46PM 0 points [-]

Perhaps I’m wrong, but your decision to care about the preference of existence over non-existence and moving on from there appears to be an example of (a). Or perhaps a component (b) did develop and maintain awareness of nihilism, but obviously that component couldn’t be bothered posting on LW, so I heard a reply from the part of you that is attached to your subjective preferences (and simply exists).

Well, my bit about existence and non-existence stemmed from a struggle with believing that things did or did not exist. I have never considered nihilism to be a relevant proposal: It doesn't tell me how to act or what to do. It also doesn't care if I act as if there is an objective value attached to something. So... what is the point in nihilism?

To me, nihilism seems like a trap for other philosophical arguments. If those arguments and moral ways lead them to a logical conclusion of nihilism, than they cannot escape. They are still clinging to whatever led them there but say they are nihilists. This is the death spiral: Believing that nothing matters but acting as if something does.

If I were to actually stop and throw away all objective morality, value, etc than I would except a realization that any belief in nihilism would have to go away too. At this point I my presuppositions about the world reset and... what? It is this behavior that is similar to my struggles with existence.

The easiest summation of my belief that existence is preferred over non-existence is that existence can be undone and non-existence is permanent. If you want more I can type it up. I don't know how helpful it will be against nihilism, however.

Comment author: byrnema 02 January 2010 12:23:53AM 0 points [-]

This edges very close to an affective death spiral,

Agreed. I find that often it isn't so much that I find the thought process intrinsically pleasurable (affective), but that in thinking about it too much, I over-stimulate the trace of the argument so that after a while I can't recall the subtleties and can't locate the support. After about 7 comments back and forth, I feel like a champion for a cause (no objective values RESULTS IN NIHILISM!!) that I can't relate to anymore. Then I need to step back and not care about it for a while, and maybe the cause will spontaneously generate again, or perhaps I'll have learned enough weighting in another direction that the cause never takes off again.

Comment author: AdeleneDawner 31 December 2009 09:22:41AM 0 points [-]

Feel free to tell me to mind my own business, but I'm curious. That other part: If you gave it access to resources (time, money, permission), what do you expect that it would do? Is there anything about your life that it would change?

Comment author: byrnema 31 December 2009 02:40:10PM *  0 points [-]

Jack also wrote, "The next question is obviously "are you depressed?" But that also isn't any of my business so don't feel obligated to answer."

I appreciate this sensitivity, and see where it comes from and hy its justified, but I also find it interesting that interrogating personal states is perceived as invasive, even as this is the topic at hand.

However, I don't feel like its so personal and I will explain why. My goals here are to understand how the value validation system works outside FOOM. I come from the point of view that I can't do this very naturally, and most people I know also could not. I try to identify where thought gets stuck and try to find general descriptions of it that aren't so personal. I think feeling like I have inconsistent pieces (i.e., like I'm going insane) would be a common response to the anticipation of a non-FOOM world.

That other part: If you gave it access to resources (time, money, permission), what do you expect that it would do? Is there anything about your life that it would change?

To answer your question, a while ago, I thought my answer to your question would be a definitive, "no, this awareness wouldn't feel any motivation to change anything". I had written in my journal that even if there was a child laying on the tracks, this part of myself would just look on analytically. However, I felt guilty about this after a while and I've seen repressed the experience of this hypothetical awareness so that its more difficult to recall.

But recalling, it felt like this: it would be "horrible" for the child to die on the tracks. However, what is "horrible" about horrible? There's nothing actually horrible about it. Without some terminal value behind the value (for example, I don't think I ever thought a child dieing on the tracks was objectively horrible, but that it might be objectively horrible for me not to feel like horrible was horrible at some level of recursion) it seems that the value buck doesn't get passed, it doesn't stop, it just disappears.

Comment author: AdeleneDawner 31 December 2009 04:02:26PM 1 point [-]

I appreciate this sensitivity, and see where it comes from and why its justified, but I also find it interesting that interrogating personal states is perceived as invasive, even as this is the topic at hand.

Actually, I practically never see it as invasive; I'm just aware that other people sometimes do, and try to act accordingly. I think this is a common mindset, actually: It's easier to put up a disclaimer that will be ignored 90-99% of the time than it is to deal with someone who's offended 1-10% of the time, and generally not worth the effort of trying to guess whether any given person will be offended by any given question.

I think feeling like I have inconsistent pieces (i.e., like I'm going insane) would be a common response to the anticipation of a non-FOOM world.

I'm not sure how you came to that conclusion - the other sentences in that paragraph didn't make much sense to me. (For one thing, one of us doesn't understand what 'FOOM' means. I'm not certain it's you, though.) I think I know what you're describing, though, and it doesn't appear to be a common response to becoming an atheist or embracing rationality (I'd appreciate if others could chime in on this). It also doesn't necessarily mean you're going insane - my normal brain-function tends in that direction, and I've never seen any disadvantage to it. (This old log of mine might be useful, on the topic of insanity in general. Context available on request; I'm not at the machine that has that day's logs in it at the moment. Also, disregard the username, it's ooooold.)

But recalling, it felt like this: it would be "horrible" for the child to die on the tracks. However, what is "horrible" about horrible? There's nothing actually horrible about it. Without some terminal value behind the value (for example, I don't think I ever thought a child dieing on the tracks was objectively horrible, but that it might be objectively horrible for me not to feel like horrible was horrible at some level of recursion) it seems that the value buck doesn't get passed, it doesn't stop, it just disappears.

My Buddhist friends would agree with that. Actually, I pretty much agree with it myself (and I'm not depressed, and I don't think it's horrible that I don't see death as horrible, at any level of recursion). What most people seem to forget, though, is that the absence of a reason to do something isn't the same as the presence of a reason not to do that thing. People who've accepted that there's no objective value in things still experience emotions, and impulses to do various things including acting compassionately, and generally have no reason not to act on such things. We also experience the same positive feedback from most actions that theists do - note how often 'fuzzies' are explicitly talked about here, for example. It does all add back up to normality, basically.

Comment author: byrnema 04 January 2010 06:53:53AM 0 points [-]

Thank you. So maybe I can look towards Buddhist philosophy to resolve some of my questions. In any case, it's really reassuring that others can form these beliefs about reality, and retain things that I think are important (like sanity and moral responsibility.)

Comment author: byrnema 31 December 2009 04:07:09PM 0 points [-]

Sorry! FOOV: Framework Of Objective Value!

Comment author: AdeleneDawner 31 December 2009 04:57:26PM 1 point [-]

Okay, I went back and re-read that bit with the proper concept in place. I'm still not sure why you think that non-FOOV value systems would lead to mental problems, and would like to hear more about that line of reasoning.

As to how non-FOOV value systems work, there seems to be a fair amount of variance. As you may've inferred, I tend to take a more nihilistic route than most, assigning value to relatively few things, and I depend on impulses to an unusual degree. I'm satisfied with the results of this system: I have a lifestyle that suits my real preferences (resources on hand to satisfy most impulses that arise often enough to be predictable, plus enough freedom and resources to pursue most unpredictable impulses), projects to work on (mostly based on the few things that I do see as intrinsically valuable), and very few problems. It appears that I can pull this off mostly because I'm relatively resistant to existential angst, though. Most value systems that I've seen discussed here are more complex, and often very other-oriented. Eliezer is an example of this, with his concept of coherent extrapolated value. I've also seen at least one case of a person latching on to one particular selfish goal and pursuing that goal exclusively.

Comment author: AdeleneDawner 31 December 2009 04:12:10PM 0 points [-]

Ah. That makes more sense.

Also, a point I forgot to add in my above post: Some (probably the vast majority of) atheists do see death as horrible; they just have definitions of 'horrible' that don't depend on objective value.

Comment author: Furcas 30 December 2009 11:40:26PM 1 point [-]

There is objective value, you know. It is an objective fact about reality that you care about and value some things and people, as do all other minds.

The point of going to Dallas is a function of your values, not the other way around.

Comment author: byrnema 31 December 2009 04:00:23PM 0 points [-]

There is objective value, you know. It is an objective fact about reality that you care about and value some things and people, as do all other minds.

I'm not sure this question will make sense, but do you place any value on that objective value?

Comment author: tut 30 December 2009 01:32:42PM *  0 points [-]

Why optimize our strategies to achieve what we happen to want, instead of just modifying what we want?

For some things it is probably wise to change your desires to something you can actually do. But in general the answer is another question: Why would you want to do that?

Comment author: Zack_M_Davis 20 December 2009 11:20:05AM 6 points [-]

Does anyone here think they're particularly good at introspection or modeling themselves, or have a method for training up these skills? It seems like it would be really useful to understand more about the true causes of my behavior, so I can figure out what conditions lead to me being good and what conditions lead to me behaving poorly, and then deliberately set up good conditions. But whenever I try to analyze my behavior, I just hit a brick wall---it all just feels like I chose to do what I did out of my magical free will. Which doesn't explain anything.

If you know what you want, and then you choose actions that will help you get it, then that's simple enough to analyze: you're just rational, that's all. But when you would swear with all your heart that you want some simple thing, but are continually breaking down and acting dysfunctionally---well, clearly something has gone horribly wrong with your brain, and you should figure out the problem and fix it. But if you can't tell what's wrong because your decision algorithm is utterly opaque, then what do you do?

Comment author: orthonormal 31 December 2009 06:18:53AM *  1 point [-]

One thing that has worked for me lately is the following: whenever I do something and don't really know why I did it (or am uncomfortable with the validity of my rationalizations), I try and think of the action in Outside View terms. I think of (or better, write out) a short external description of what I did, in its most basic form, and its probable consequences. Then I ask what goal this action looks optimized for; it's usually something pretty simple, but which I might not be happy consciously acknowledging (more selfish than usual, etc).

That being said, even more helpful than this has been discussing my actions with a fairly rational friend who has my permission to analyze it and hypothesize freely. When they come up with a hypothesis that I don't like, but which I have no good counterarguments against, we've usually hit paydirt.

well, clearly something has gone horribly wrong with your brain, and you should figure out the problem and fix it.

I don't think of this as something wrong with my brain, so much as it functioning properly in maintaining a conscious/unconscious firewall, even though this isn't as adaptive in today's world as it once was. It's really helped me in introspection to not judge myself, to not get angry with my revealed preferences.

Comment author: Alicorn 20 December 2009 03:37:09PM 3 points [-]

People who watch me talking about myself sometimes say I'm good at introspection, but I think about half of what I do is making up superstitions so I have something doable to trick myself into making some other thing, previously undoable, doable. ("Clearly, the only reason I haven't written my paper is that I haven't had a glass of hot chocolate, when I'm cold and thirsty and want refined sugar." Then I go get a cup of cocoa. Then I write my paper. I have to wrap up the need for cocoa in a fair amount of pseudoscience for this to work.) This is very effective at mood maintenance for me - I was on antidepressants and in therapy for a decade as a child, and quit both cold turkey in favor of methods like this and am fine - but I don't know which (if, heck, any) of my conclusions that I come to this way are "really true" (that is, if the hot chocolate is a placebo or not). They're just things that pop into my head when I think about what my brain might need from me before it will give back in the form of behaving itself.

You have to take care of you brain for it to be able to take care of you. If it won't tell you what it wants, you have to guess. (Or have your iron levels checked :P)

Comment author: whpearson 20 December 2009 02:39:37PM *  3 points [-]

I tend to think of my brain as a thing with certain needs. Companionship, recognition, physical contact, novelty, etc. Activities that provide these tend to persist. Figure out what your dysfunctional actions provide you in terms of your needs. Then try and find activities that provide these but aren't so bad and try and replace the dysfunctional bits. Also change the situation you are in so that the dysfunctional default actions don't automatically trigger.

My dream is to find a group of like minded people that I can socialise and work with. SIAI is very tempting in that regard.

Comment author: wedrifid 20 December 2009 02:29:26PM 5 points [-]

Does anyone here think they're particularly good at introspection or modeling themselves, or have a method for training up these skills? It seems like it would be really useful to understand more about the true causes of my behavior, so I can figure out what conditions lead to me being good and what conditions lead to me behaving poorly, and then deliberately set up good conditions. But whenever I try to analyze my behavior, I just hit a brick wall---it all just feels like I chose to do what I did out of my magical free will. Which doesn't explain anything.

My suggestion is focussing your introspection on working out what you really want. That is, keep investigating what you really want until such time as the phrase 'me behaving poorly' and 'being good' sound like something that is in a foreign language, that you can understand only by translating.

You may be thinking "clearly something has gone horribly wrong with my brain" but your brain is thinking "Something is clearly wrong with my consciousness. It is trying to make me do all this crazy shit. Like the sort of stuff we're supposed to pretend we want because that is what people 'Should' want. Consciousnesses are the kind of things that go around believing in God and sexual fidelity. That's why I'm in charge, not him. But now he's thinking he's clever and is going to find ways to manipulate me into compliance. F@#@ that s#!$. Who does he think he is?"

When trying to work effectively with people empathy is critical. You need to be able to understand what they want and be able to work with each other for mutual benefit. Use the same principle with yourself. Once your brain believes you actually know what it (ie. you) want and are on approximately the same page it may well start trusting you and not feel obliged to thwart your influence. Then you can find a compromise that allows you to get that 'simple thing' you want without your instincts feeling that some other priority has been threatened.

Comment author: timtyler 19 December 2009 09:38:46PM 1 point [-]
Comment author: MrHen 18 December 2009 02:37:56PM 2 points [-]

I am posting this in the open thread because I assume that somewhere in the depths of posts and comments there is an answer to the question:

If someone thought we lived in an internally consistent simulation that is undetectable and inescapable, is it even worth discussing? Wouldn't the practical implications of such a simulation imply the same things as the material world/reality/whatever you call it?

Would it matter if we dropped "undetectable" from the proposed simulation? At what point would it begin to matter?

Comment author: MrHen 17 December 2009 05:26:09PM 1 point [-]

It has been awhile since I have been around, so please ignore if this has been brought up before.

I would appreciate it if offsite links were a different color. The main reason for this is because of the way I skim online articles. Links are generally more important text and I if I see a link for [interesting topic] it helps me to know at a glance that there will be a good read with a LessWrong discussion at the end as opposed to a link to Amazon where I get to see the cover of a book.

Comment author: Alicorn 17 December 2009 05:40:10PM 0 points [-]

Firefox (or maybe one of the million extensions that I've downloaded and forgotten about) has a feature where, if you mouseover a link, the URL linked to will appear in the lower bar of the window. A different color would be easier, though.

Comment author: Wei_Dai 17 December 2009 12:11:46PM 7 points [-]

What are the implications to FAI theory of Robin's claim that most of what we do is really status-seeking? If an FAI were to try to extract or extrapolate our values, would it mostly end up with "status" as the answer and see our detailed interests, such as charity or curiosity about decision theory, as mere instrumental values?

Comment author: wedrifid 18 December 2009 02:59:15PM 3 points [-]

What are the implications to FAI theory of Robin's claim that most of what we do is really status-seeking? If an FAI were to try to extract or extrapolate our values, would it mostly end up with "status" as the answer and see our detailed interests, such as charity or curiosity about decision theory, as mere instrumental values?

If it went that far it would also go the next step. It would end up with "getting laid".

Comment author: EStokes 18 December 2009 02:43:54PM 3 points [-]

I think it's kinda like inclusive genetic fitness: It's the reason you do things, but you're (usually) not conciously striving for an increased amount of it. So I don't think it could be called a terminal value, as such...

Comment author: Wei_Dai 20 December 2009 08:49:13AM 3 points [-]

I had thought of that, but, if you consider a typical human mind as a whole instead of just the conscious part, it seems clear that it is striving for increased status. The same cannot be said for inclusive fitness, or at least the number of people who do not care about having higher status seems much lower than the number of people who do not care about having more offspring.

I think one of Robin's ideas is that unconscious preferences, not just conscious ones, should matter in ethical considerations. Even if you disagrees with that, how do you tell an FAI how to distinguish between conscious preferences and unconscious ones?

Comment author: Johnicholas 15 December 2009 04:44:35PM 1 point [-]

Repost from Bruce Schneier's CRYPTO-GRAM:

The Psychology of Being Scammed

This is a very interesting paper: "Understanding scam victims: seven principles for systems security, by Frank Stajano and Paul Wilson." Paul Wilson produces and stars in the British television show The Real Hustle, which does hidden camera demonstrations of con games. (There's no DVD of the show available, but there are bits of it on YouTube.) Frank Stajano is at the Computer Laboratory of the University of Cambridge.

The paper describes a dozen different con scenarios -- entertaining in itself -- and then lists and explains six general psychological principles that con artists use:

  1. The distraction principle. While you are distracted by what retains your interest, hustlers can do anything to you and you won't notice.

  2. The social compliance principle. Society trains people not to question authority. Hustlers exploit this "suspension of suspiciousness" to make you do what they want.

  3. The herd principle. Even suspicious marks will let their guard down when everyone next to them appears to share the same risks. Safety in numbers? Not if they're all conspiring against you.

  4. The dishonesty principle. Anything illegal you do will be used against you by the fraudster, making it harder for you to seek help once you realize you've been had.

  5. The deception principle. Things and people are not what they seem. Hustlers know how to manipulate you to make you believe that they are.

  6. The need and greed principle. Your needs and desires make you vulnerable. Once hustlers know what you really want, they can easily manipulate you.

It all makes for very good reading.

The paper The Real Hustle

Comment author: Cyan 13 December 2009 06:38:19PM *  2 points [-]

In two recent comments [1][2], it has been suggested that to combine ostensibly Bayesian probability assessments, it is appropriate to take the mean on the log-odds scale. But Bayes' Theorem already tells us how we should combine information. Given two probability assessments, we treat one as the prior, sort out the redundant information in the second, and update based on the likelihood of the non-redundant information. This is practically infeasible, so we have to do something else, but whatever else it is we choose to do, we need to justify it as an approximation to the infeasible but correct procedure. So, what is the justification for taking the mean on the log-odds scale? Is there a better but still feasible procedure?

Comment author: Douglas_Knight 13 December 2009 07:43:25PM 1 point [-]

An independent piece of evidence moves the log-odds a constant additive amount regardless of the prior. Averaging log-odds amounts to moving 2/3 of that distance if 2/3 of the people have the particular piece of evidence. It may behave badly if the evidence is not independent, but if all you have are posteriors, I think it's the best you can do.

Comment author: Johnicholas 11 December 2009 08:16:05PM *  2 points [-]

Ivan Sutherland (inventor of Sketchpad - the first computer-aided drawing program) wrote about how "courage" feels, internally, when doing research or technological projects.

"[...] When I get bogged down in a project, the failure of my courage to go on never feels to me like a failure of courage, but always feels like something entirely dif- ferent. One such feeling is that my research isn't going anywhere anyhow, it isn't that important. Another feeling involves the urgency of something else. I have come to recognize these feelings as “who cares” and “the urgent drives out the important.” [...]"

Comment author: Eliezer_Yudkowsky 11 December 2009 07:55:24AM 4 points [-]

Tags now sort chronologically oldest-to-newest by default, making them much more useful for reading posts in order.

Comment author: Vladimir_Nesov 08 December 2009 11:16:13PM *  3 points [-]

David Chalmers surveys the kinds of crazy believed by modern philosophers, as well as their own predictions of the results of the survey.

56% of target faculty responding favor (i.e. accept or lean toward) physicalism, while 27% favor nonphysicalism (for respondents as a whole, the figure is 54:29). A priori knowledge is favored by 71-18%, an analytic-synthetic distinction is favored by 65-27%, Millianism is favored over Fregeanism by 34-29%, while the view that zombies are conceivable but not metaphysically possible is favored over metaphysical possibility and conceivability respectively by 35-23-16% respectively.

Comment author: [deleted] 07 December 2009 05:43:06AM 3 points [-]

Just thought I'd mention this: as a child, I detested praise. (I'm guessing it was too strong a stimulus, along with such things as asymmetry, time being a factor in anything, and a mildly loud noise ceasing.) I wonder how it's affected my overall development.

Incidentally, my childhood dislike of asymmetry led me to invent the Thue-Morse sequence, on the grounds that every pattern ought to be followed by a reversal of that pattern.

Comment author: anonym 07 December 2009 09:31:00AM 1 point [-]

Fascinating. As a child, I also detested praise, and I have always had something bordering on an obsession for symmetry and an aversion to asymmetry.

I hadn't heard of the Thue-Morse sequence until now, but it is quite similar to a sequence I came up with as a child and have tapped out (0 for left hand/foot/leg, 1 for right hand/foot/leg) or silently hummed (or just thought) whenever I was bored or was nervous.

My sequence is:

[0, 1, 1, 0] [1001, 0110, 0110, 1001] [0110 1001 1001 0110, 1001 0110 0110 1001, 1001 0110 0110 1001, 0110 1001 1001 0110] ...

[commas and brackets added to make the pattern obvious]

As a kid, I would routinely get the pattern up into the thousands as I passed the time imagining sounds or lights very quickly going off on either the left (0) or right (1) side.

Comment author: [deleted] 07 December 2009 09:10:40PM 2 points [-]

Every finite subsequence of your sequence is also a subsequence of the Thue-Morse sequence and vice versa. So in a sense, each is a shifted version of the other; it's just that they're shifted infinitely much in a way that's difficult to define.

Comment author: Zack_M_Davis 07 December 2009 05:47:45AM 4 points [-]

Incidentally, my childhood dislike of asymmetry led me to invent the Thue-Morse sequence

I love this community.

Comment author: [deleted] 07 December 2009 06:00:40AM 1 point [-]

Can I interpret that as an invitation to send you a friend request on Facebook? >.>

Comment author: Zack_M_Davis 07 December 2009 06:04:57AM 1 point [-]

Um, sure?

Comment author: Yorick_Newsome 08 December 2009 04:39:32AM 0 points [-]

I spent much of my childhood obsessing over symmetry. At one point I wanted to be a millionaire solely so I could buy a mansion, because I had never seen a symmetrical suburban house.

Comment author: AdeleneDawner 07 December 2009 12:02:27PM 0 points [-]

Just thought I'd mention this: as a child, I detested praise.

++MeToo;

Comment author: Mario 05 December 2009 08:33:14PM 7 points [-]

I'm looking for a particular fallacy or bias that I can't find on any list.

Specifically, this is when people say "one more can't hurt;" like a person throwing an extra piece of garbage on an already littered sidewalk, a gambler who has lost nearly everything deciding to bet away the rest, a person in bad health continuing the behavior that caused the problem, etc. I can think of dozens of examples, but I can't find a name. I would expect it to be called the "Lost Cause Fallacy" or the "Fallacy of Futility" or something, but neither seems to be recognized anywhere. Does this have a standard name that I don't know, or is it so obvious that no one ever bothered to name it?

Comment author: AdeleneDawner 31 December 2009 09:46:12AM 1 point [-]

Your first example sounds related to the broken window theory, but I've never seen a name for the underlying bias. (The broken window fallacy is something else altogether.)

Comment author: timtyler 31 December 2009 09:02:02AM -1 points [-]

It seems like a type of apathy.

Comment author: Eliezer_Yudkowsky 31 December 2009 06:46:02AM 1 point [-]

Bee-sting theory of poverty is the closest I've heard. You're right, this is real and deserves a name, but I don't know what it would be.

Comment author: Sniffnoy 31 December 2009 06:47:11AM 0 points [-]

This seems like a special case of the more general "just one can't hurt" (whatever the current level) way of thinking. I don't know any name for this but I guess you could call it something like the "non-Archimedean bias"?

Comment author: Zack_M_Davis 31 December 2009 06:39:37AM 0 points [-]

"Lost Cause Fallacy" or the "Fallacy of Futility" [...] Does this have a standard name

"Sunk cost fallacy"

Comment author: Eliezer_Yudkowsky 31 December 2009 06:45:38AM 3 points [-]

No, that's different. That's pursuing a reward so as to not acknowledge a loss. This is ignoring a penalty because of previous losses.

Comment author: orthonormal 31 December 2009 06:07:38AM *  0 points [-]

Informally, "throwing good money after bad"? I agree that this is a real and interesting phenomenon.

Comment author: CronoDAS 05 December 2009 06:45:51AM 2 points [-]

I'm looking for a certain quote I think I may have read on either this blog or Overcoming Bias before the split. It goes something like this: "You can't really be sure evolution is true until you've listened to a creationist for five minutes."

Ah, never mind, I found it.

"In a way, no one can really trust the theory of natural selection until after they have listened to creationists for five minutes; and then they know it's solid."

I'd like a pithier way of phrasing it, though, than the original quote.

Comment author: JamesAndrix 04 December 2009 07:11:16AM 2 points [-]

http://scicom.ucsc.edu/SciNotes/0901/pages/geeks/geeks.html

" They told them that half the test generally showed gender differences (though they didn't mention which gender it favored), and the other half didn't.

Women and men did equally well on the supposedly gender-neutral half. But on the sexist section, women flopped. They scored significantly lower than on the portion they thought was gender-blind."

Comment author: Matt_Simpson 03 December 2009 08:44:10PM 1 point [-]

Is there a proof anywhere that occam's razor is correct? More specifically, that occam priors are the correct priors. Going from the conjunction rule to P(A) >= P(B & C) when A and B&C are equally favored by the evidence seems simple enough (and A, B, and C are atomic propositions), but I don't (immediately) see how to get from here to an actual number that you can plug into Baye's rule. Is this just something that is buried in textbook on information theory?

On that note, assuming someone had a strong background in statistics (phd level) and little to no background in computer science outside of a stat computing course or two, how much computer science/other fields would they have to learn to be able to learn information theory?

Thanks to anyone who bites

Comment author: wedrifid 03 December 2009 10:05:53PM 0 points [-]

actual number that you can plug into Baye's rule

Bayes' rule.

Comment author: Zack_M_Davis 03 December 2009 09:31:43PM 1 point [-]

I found Rob Zhara's comment helpful.

Comment author: Matt_Simpson 06 December 2009 07:02:13AM 0 points [-]

thanks. I suppose a mathematical proof doesn't exist, then.

Comment author: Unknowns 08 January 2010 11:13:26AM 0 points [-]
Comment author: whpearson 08 January 2010 11:09:13AM 0 points [-]
Comment author: ciphergoth 08 January 2010 11:00:32AM 0 points [-]

Not only is there no proof, there isn't even any evidence for it. Any effort to collect evidence for it leaves you assuming what you're trying to prove. This is the "problem of induction" and there is no solution; however, since you are built to be incapable of not applying induction and you couldn't possibly make any decisions without it.

Comment author: timtyler 09 December 2009 06:15:51AM 0 points [-]

Occam's razor is dependent on a descriptive language / complexity metric (so there are multiple flavours of the razor).

Unless a complexity metric is specified, the first question seems rather vague.

Comment author: Jayson_Virissimo 08 January 2010 10:40:40AM 0 points [-]

Occam's razor is dependent on a descriptive language / complexity metric (so there are multiple flavours of the razor).

I think you might be making this sound easier than it is. If there are an infinite number of possible descriptive languages (or of ways of measuring complexity) aren't there an infinite number of "flavours of the razor"?

Comment author: timtyler 09 January 2010 12:14:09AM 0 points [-]

Yes, but not all languages are equal - and some are much better than others - so people use the "good" ones on applications which are sensitive to this issue.

Comment author: ciphergoth 08 January 2010 11:01:52AM 0 points [-]

There's a proof that any two (Turing-complete) metrics can only differ by at most a constant amount, which is the message length it takes to encode one metric in the other.

Comment author: timtyler 09 January 2010 12:12:02AM 0 points [-]

Of course, the constant can be arbitrarily large.

However, there are a number of domains for which this issue is no big deal.

Comment author: Jonii 03 December 2009 04:10:08PM 1 point [-]

I'm interested in values. Rationality is usually defined as something like an agent that tries to maximize its own utility function. But, humans, as far as I can tell, don't really have anything like "values" beside "stay alive, get immediatly satisfying sensory input".

This, afaict, results to lip servive to "greater good", when people just select some nice values that they signal they want to promote, when in reality they haven't done the math by which these selected "values" derive from these "stay alive"-like values. And so, their actions seem irrational, but only because they signal of having values they don't actually have or care about.

This probably boils down to finding something to protect, but overall this issue is really confusing.

Comment author: Jack 03 December 2009 08:29:09PM 2 points [-]

I'm confused. Have you never seen long-term goal directed behavior?

Comment author: Jonii 03 December 2009 09:02:33PM 2 points [-]

I'm not sure, maybe, but more of a problem here is to select your goals. The choice seems to be arbitrary, and as far as I can tell, human psychology doesn't really even support having value systems that go deeper than that "lip service" + conformism state.

But I'm really confused when it comes to this, so I thought I could try describing my confusion here :)

Comment author: Jack 03 December 2009 09:52:57PM *  0 points [-]

I think you need to meet better humans. Or just read about some.

John Brown, Martin Luther, Galileo Galilei, Abraham Lincoln, Charles Darwin, Mohandas Gandhi

Can you make sense of those biographies without going deeper than "lip service" + conformism state?

Edit: And I don't even necessarily mean that these people were supremely altruistic or anything. You can add Adolph Hitler to the list too.

Comment author: Jonii 03 December 2009 10:19:38PM 4 points [-]

Can you make sense of those biographies without going deeper than "lip service" + conformism state?

Dunno, haven't read any of those. But if you're sure that something like that exists, I'd like to hear how is it achievable on human psychology.

I mean, paperclip maximizer is seriously ready to do anything to maximize paperclips. It really takes the paperclips seriously.

On the other hand, there are no humans that seem to care about anything in particular that's going on in the world. People are suffering and dying, misfortune happens, animals go extinct, and relatively few do anything about it. Many claim they're concerned and that they value human life and happiness, but if doing something requires going beyond the safe zone of conformism, people just don't do it. The best way I've figured out to overcome this is to manipulate that safe zone to allow more actions, but it would seem that many people think they know better. I just don't understand what.

Comment author: Jonii 03 December 2009 11:11:58PM 1 point [-]

I could go on and state that I'm well aware that the world is complicated. It's difficult to estimate where our choices do lead us to, since net of causes and effects is complex and requires a lot of thinking to grasp. Heuristics human brain uses exist pretty much because of that. This means that it's difficult to figure out how to do something beside staying in the safe zone that you know to work at least somehow.

However, I still think there's something missing here. This just doesn't look like a world where people do particularly care about anything at all. Even if it was often useful to stay in a safe zone, there doesn't really seem to be any easy way to snap them out of it. No magic word, no violation of any sort of values makes anyone stand up and fight. I could literally tell people that millions are dying in vain(aging) or that the whole world is at stake(existential risks), and most people simply don't care.

At least, that's how I see it. I figure that rare exceptions to the rule there can be explained as a cost of signalling something, requirements of the spot in the conformist space you happen to be occupy or something like that.

I'm not particularly fond of this position, but I'm simply lacking a better alternative.

Comment author: Jack 03 December 2009 11:09:49PM 0 points [-]

Dunno, haven't read any of those. But if you're sure that something like that exists, I'd like to hear how is it achievable on human psychology. Biographies as in the stories of their lives not as in books about those stories. Try wikipedia, these aren't obscure figures.

On the other hand, there are no humans that seem to care about anything in particular that's going on in the world.

This is way too strong, isn't it? I also don't think the reason a lot of people ignore these tragedies has as much to do with conformism as it does self-interest. People don't want to give up their vacation money. If anything there is social pressure in favor of sacrificing for moral causes. As for values, I think most people would say that the fact they don't do more is a flaw. "If I was a better person I would do x" or "Wow, I respect you so much for doing x" or "I should do x but I want y so much." I think it is fair to interpret these statements as second order desires that represent values.

Comment author: Jonii 03 December 2009 11:33:11PM 2 points [-]

Remember what I said about "lip service"?

If they want to care about stuff, that's kinda implying that they don't actually care about stuff (yet). Also, based on simple psychology, someone who chooses a spot in the conformist zone that requires giving lip service to something creates cognitive dissonance which easily produces second order desire to want what you claim you want. But what is frightening here is how this choice of values is arbitrary to the ultimate. If you'd judged another spot to be cheaper, you'd need to modify your values in a different way.

On both cases though, it seems that people really rarely move any bit towards actually caring about something.

Comment author: Jack 04 December 2009 12:08:08AM 0 points [-]

What is a conformist zone and why is it spotted?

Remember what I said about "lip service"?

Lip service is "Oh, what is happening in Darfur is so terrible!". That is different from "If I was a better person I'd help the people of Darfur" or "I'm such a bad person, I bought a t.v. instead of giving to charity". The first signals empathy the second and third signal laziness or selfishness (and honesty I guess).

If they want to care about stuff, that's kinda implying that they don't actually care about stuff (yet).

Why do values have to produce first order desires? For that matter, why can't they be socially constructed norms which people are rewarded for buying into? When people do have first order desires that match these values we name those people heroes. Actually sacrificing for moral causes doesn't get you ostracized it gets you canonized.

But what is frightening here is how this choice of values is arbitrary to the ultimate.

Not true. The range of values in the human community is quite limited.

On both cases though, it seems that people really rarely move any bit towards actually caring about something.

People are rarely complete altruists. But that doesn't mean that they don't care about anything. The world is full of broke artists who could pay for more food, drugs and sex with a real job. These people value art.

Comment author: Jonii 04 December 2009 11:18:58AM 0 points [-]

Lip service is "Oh, what is happening in Darfur is so terrible!". That is different from "If I was a better person I'd help the people of Darfur" or "I'm such a bad person, I bought a t.v. instead of giving to charity". The first signals empathy the second and third signal laziness or selfishness (and honesty I guess).

Both are hollow words anyway. Both imply that you care, when you really don't. There are no real actions.

Why do values have to produce first order desires?

Because, uhm, if you really value something, you'd probably want to do something? Not "want to want", or anything, but really care about that stuff which you value. Right?

For that matter, why can't they be socially constructed norms which people are rewarded for buying into?

Sure they can. I expressed this as safe zone manipulation, attempting to modify your envinroment so that your conformist behavior leads to working for some value.

The point here is that actually caring about something and working towards something due to arbitrary choice and social pressure are quite different things. Since you seem to advocate the latter, I'm assuming that we both agree that people rarely care about anything and most actions are result of social pressure and stuff not directly related to actually caring or valuing anything.

Which brings me back to my first point: Why does it seem that many people here actually care about the world? Like, as in paperclip maximizer cares about paperclips. Just optical illusion and conscious effort to appear as a rational agent valuing the world, or something else?

Comment author: Jonii 07 December 2009 01:42:11PM 0 points [-]

So, I've been thinking about this for some time now, and here's what I've got:

First, the point here is to self-reflect to want what you really want. This presumably converges to some specific set of first degree desires for each one of us. However, now I'm a bit lost on what do we call "values", are they the set of first degree desires we have(not?), set of first degree desires we would reach after infinity of self-reflection, or set of first degree desires we know we want to have at any given time?

As far as I can tell, akrasia would be a subproblem of this.

So, this should be about right. However, I think it's weird that here people talk a bit about akrasia, and how to achieve those n-degree desires, but I haven't seen anything about actually reflecting and updating what you want. Seems to me that people trust a tiny bit too much to the power of cognitive dissonance fixing the problem between wanting to want and actually wanting, this resulting to the lack of actual desire in achieving what you know you should want(akrasia).

I really dunno how to overcome this, but this gap seems worth discussing.

Also, since we need eternity of self-reflection to reach what we really want, this looks kinda bad for FAI: Figuring out where our self-reflection would converge in infinity seems pretty much impossible to compute, and so, we're left with compromises that can and probably will eventually lead to something we really don't want.

Comment author: Wei_Dai 03 December 2009 10:42:20AM 4 points [-]

This blog comment describes what seems to me the obvious default scenario for an unFriendly AI takeoff. I'd be interested to see more discussion of it.

Comment author: whpearson 03 December 2009 11:36:17AM *  2 points [-]

The problem with the specific scenario given, with experimental modification/duplication, rather than careful proof based modification, is that is liable to have the same problem that we have with creating systems this way. The copies might not do what the agent that created them want.

Which could lead to a splintering of the AI, and in-fighting over computational resources.

It also makes the standard assumptions that AI will be implemented on and stable on the von Neumann style computing architecture.

Comment author: Johnicholas 03 December 2009 06:45:42PM *  0 points [-]

Let me point out that we (humanity) does actually have some experience with this scenario. Right now, mobile code that spreads across a network without effective controls on the bounds of its expansion by the author is worms. If we have experience, we should mine it for concrete predictions and countermeasures.

General techniques against worms might include: isolated networks, host diversity, rate-limiting, and traffic anomaly detection.

Are these low-cost/high-return existential reduction techniques?

Comment author: Wei_Dai 03 December 2009 11:46:53PM 6 points [-]

No, these are high-cost/low-return existential risk reduction techniques. Major corporations and governments already have very high incentive to protect their networks, but despite spending billions of dollars, they're still being frequently penetrated by human attackers, who are not even necessarily professionals. Not to mention the hundreds of millions of computers on the Internet that are unprotected because their owners have no idea how to do so, or they don't contain information that their owners consider especially valuable.

I got into cryptography partly because I thought it would help reduce the risk of a bad Singularity. But while cryptography turned out to work relatively well (against humans anyway), the rest of the field of computer security is in terrible shape, and I see little hope that the situation would improve substantially in the next few decades.

Comment author: whpearson 08 January 2010 08:04:53AM 0 points [-]

What do you think of the object-capability model? And removing ambient authority in general.

Comment author: Wei_Dai 08 January 2010 08:59:06AM 0 points [-]

That's outside my specialization of cryptography, so I don't have too much to say about it. I do remember reading about the object-capability model and the E language years ago, and thinking that it sounded like a good idea, but I don't know why it hasn't been widely adopted yet. I don't know if it's just inertia, or whether there are some downsides that its proponents tend not to publicize.

In any case, it seems unlikely that any security solution can improve the situation enough to substantially reduce the risk of a bad Singularity at this point, without a huge cost. If the cause of existential-risk reduction had sufficient resources, one project ought to be to determine the actual costs and benefits of approaches like this and whether it would be feasible to implement (i.e., convince society to pay whatever costs are necessary to make our networks more secure), but given the current reality I think the priority of this is pretty low.

Comment author: whpearson 08 January 2010 09:29:16AM 0 points [-]

Thanks. I just wanted to know if this was the sort of thing you had in mind, and whether you knew any technical reasons why it wouldn't do what you want.

This is one thing I keep a close-ish eye on. One of the major proponents of this sort of security has recently gone to work for Microsoft on their research operating systems. So it might come a long in a bit.

As to why it hasn't caught on, it is partially inertia and partially it requires more user interaction/understanding of the systems than ambient authority. Good UI and metaphors can decrease that cost though.

The ideal would be to have a self-maintaining computer system with this sort of security system. However a good self-maintaining system might be dangerously close to a self-modifying AI.

Comment author: ChrisHibbert 10 January 2010 09:09:45PM 2 points [-]

There's also a group of proponents of this style working on Caja at Google, including Mark Miller, the designer of E. And some people at HP.

Actually, all these people talk to one another regularly. They don't have a unified plan or a single goal, but they collaborate with one another frequently. I've left out several other people who are also trying to find ways to push in the same direction. Just enough names and references to give a hint. There are several mailing lists where these issues are discussed. If you're interested, this is probably the one to start with.

Comment author: ciphergoth 08 January 2010 09:43:45AM 0 points [-]

One of the major proponents of this sort of security has recently gone to work for Microsoft on their research operating systems.

Sadly, I suspect this moves things backwards rather than forwards. I was really hoping that we'd see Coyotos one day, which now seems very unlikely.

Comment author: whpearson 08 January 2010 11:01:27AM 1 point [-]

I meant it more as an indication that Microsoft are working in the direction of better secured OSes already, rather than his being a pivotal move. Coyotos might get revived when the open source world sees what MS produces and need to play catch up.

Comment author: Cyan 03 December 2009 12:52:25AM 4 points [-]

Henceforth, I am Dr. Cyan.

Comment author: Eliezer_Yudkowsky 03 December 2009 03:46:42PM 2 points [-]

Congratulations! I guess people will believe everything you say now.

Comment author: Cyan 03 December 2009 04:10:20PM 1 point [-]

I certainly hope so!

Comment author: CannibalSmith 04 December 2009 02:26:04PM 0 points [-]

Wear a lab coat for extra credibility.

Comment author: Cyan 04 December 2009 03:16:35PM 3 points [-]

I was thinking I'd wear a stethoscope and announce, "Trust me! I'm a doctor! (sotto voce)... of philosophy."

Comment author: Aurini 05 December 2009 11:54:47PM 0 points [-]

Congrats! My friend recently got his Master's in History, and has been informing every telemarketer who calls that "Listen cupcake, it's not Dave - I'm not going to hang at your crib and drink forties; listen here, pal, I have my own office! Can you say that? To you I'm Masters Smith."

I certainly hope you wear your new title with a similar air of pretention, Doctor Cyan. :)

Comment author: Cyan 06 December 2009 12:12:39AM *  1 point [-]

I'll do my best!

Sincerely,
Cyan, Ph.D.

Comment author: gwern 11 December 2009 07:09:28PM 0 points [-]

Is 'Masters' actually a proper prefix (akin to the postfix Ph.D) for people with a Master's degree? I don't think I've ever seen that before.

Comment author: Daniel_Burfoot 07 December 2009 04:57:33AM 0 points [-]

Congratulations!

Why not post an introduction to your thesis research on LW?

Comment author: Cyan 07 December 2009 02:44:31PM 1 point [-]

Because I'd need to preface it with a small deluge of information about protein chemistry, liquid chromatography, and mass spectrometry. I think I'd irritate folks if I did that.

Comment author: SilasBarta 03 December 2009 08:32:42PM 0 points [-]

With a doctorate in ...?

Comment author: Cyan 03 December 2009 08:38:20PM 2 points [-]

Biomedical engineering. My thesis concerned the analysis of proteomics data by Bayesian methods.

Comment author: SilasBarta 03 December 2009 11:03:16PM 0 points [-]

Isn't that what they normally use to analyze proteomics data? <\naive>

Comment author: Cyan 04 December 2009 12:16:41AM 1 point [-]

Not always, or even usually. It seems to me that by and large, scientists invent ad hoc methods for their particular problems, and that applies in proteomics as well as other fields.

Comment author: PeerInfinity 02 December 2009 06:06:33PM 7 points [-]

(reposted from last month's open thread)

An interesting site I recently stumbled upon:

http://changingminds.org/

They have huge lists of biases, techniques, explanations, and other stuff, with short summaries and longer articles.

Here's the results from typing in "bias" into their search bar.

A quick search for "changingminds" in LW's search bar shows that noone has mentioned this site before on LW.

Is this site of any use to anyone here?

Comment author: Tom_Talbot 02 December 2009 08:54:35PM 2 points [-]

The conversion techniques page is fascinating. I'll put this to use good in further spreading the word of Bayes.

Comment author: Yorick_Newsome 02 December 2009 11:06:32AM *  2 points [-]

Big Edit: Jack formulated my ideas better, so see his comment.
This was the original: The fact that the universe hasn't been noticeably paperclipped has got to be evidence for a) the unlikelihood of superintelligences, b) quantum immortality, c) our universe being the result of a non-obvious paperclipping (the theists were right after all, and the fine-tuned universe argument is valid), d) the non-existence of intelligent aliens, or e) that superintelligences tend not to optimize things that are astronomically visible (related to c). Which of these scenarios is most likely? Related question: If we built a superintelligence without worrying about friendliness or morality at all, what kind of things would it optimize? Can we even make a guess? Would it be satisfied to be a dormant Laplace's Demon?

Comment author: wuwei 03 December 2009 04:14:43AM 2 points [-]

d) should be changed to the sparseness of intelligent aliens and limits to how fast even a superintelligence can extend its sphere of influence.

Comment author: Jack 02 December 2009 11:37:30AM *  4 points [-]

Restructuring since the fact that the universe hasn't been noticeably paperclipped can't possible be considered evidence for (c).

The universe has either been paperclipped (1) or it hasn't been (2).

If (1):

(A) we have observed paperclipping and not realized it (someone was really into stars, galaxies and dark matter)

(B) Our universe is the result of paperclipping (theists were right, sort of)

(C) Superintelligences tend not to optimize things that are astronomically visible.

If (2)

(D) Super-intelligences are impossible.

(E) Quantum immortality true.

(F) No intelligent aliens.

(G) Some variety of simulation hypothesis is true.

(H) Galactic aliens exist but have never constructed a super-intelligence do to a well enforced prohibition on AI construction/research, an evolved deficiency in thinking about minds as a physical object (substance dualism is far more difficult for them to avoid than it is for us), or some other reason that we can't fathom.

(I) Friendliness is easy + Alien ethics doesn't include any values that lead to us noticing them.

Comment author: bgrah449 01 December 2009 11:06:44PM 2 points [-]

I like the color red. When people around me wear red, it makes me happy - when they wear any other color, it makes me sad. I crunch some numbers and tell myself, "People wear red about 15% of the time, but they wear blue 40% of the time." I campaign for increasing the amount that people wear red, but my campaign fails miserably.

"It'd be great if I could like blue instead of red," I tell myself. So I start trying to get myself to like blue - I choose blue over red whenever possible, surround myself in blue, start trying to put blue in places where I experience other happinesses so I associate blue with those things, etc.

What just happened? Did a belief or a preference change?

Comment author: Alicorn 01 December 2009 11:25:45PM 8 points [-]

You acquired a second-order desire, which is a preference about preferences.

Comment author: SilasBarta 01 December 2009 07:45:46PM *  1 point [-]

By coincidence, two blog posts went up today that should be of interest to people here.

Gene Callahan argues that Bayesianism lacks the ability to smoothly update beliefs as new evidence arrives, forcing the Bayesian to irrationally reset priors.

Tyler Cowen offering a reason why the CRU hacked emails should raise our confidence in AGW. An excellent exercise in framing an issue in Bayesian terms. Also discusses metaethical issues related to bending rules.

(Needless to say, I don't agree with either of these arguments, but they're great for application of your own rationality.)

Comment author: Matt_Simpson 02 December 2009 05:24:21AM 2 points [-]

Tyler Cowen offering a reason why the CRU hacked emails should raise our confidence in AGW...

That's not what he is saying. His argument is not that the hacked emails actually should raise our confidence in AGW. His argument is that there is a possible scenario under which this should happen, and the probability that this scenario is true is not infinitesimal. The alternative possibility - that the scientists really are smearing the opposition with no good reason - is far more likely, and thus the net effect on our posteriors is to reduce them - or at least keep them the same if you agree with Robin Hanson.

Here's (part of) what Tyler actually said:

Another response, not entirely out of the ballpark, is: 2. "These people behaved dishonorably. They must have thought this issue was really important, worth risking their scientific reputations for. I will revise upward my estimate of the seriousness of the problem."

I am not saying that #2 is correct, I am only saying that #2 deserves more than p = 0. Yet I have not seen anyone raise the possibility of #2.

Comment author: SilasBarta 02 December 2009 05:39:09PM *  -1 points [-]

Tyler Cowen offering a reason why the CRU hacked emails should raise our confidence in AGW...

That's not what he is saying. His argument is not that the hacked emails actually should raise our confidence in AGW. His argument is that there is a possible scenario under which this should happen, and the probability that this scenario is true is not infinitesimal

Right -- that is what I called "giving a reason why the hacked emails..." and I believe that characterization is accurate: he's described a reason why they would raise our confidence in AGW.

The alternative possibility - that the scientists really are smearing the opposition with no good reason - is far more likely, and thus the net effect on our posteriors is to reduce them

This is reason why Tyler's argument for a positive Bayes factor is in error, not a reason why my characterization was inaccurate.

I think we agree on the substance.

Comment author: Matt_Simpson 02 December 2009 10:12:49PM *  1 point [-]

The alternative possibility - that the scientists really are smearing the opposition with no good reason - is far more likely, and thus the net effect on our posteriors is to reduce them

This is reason why Tyler's argument for a positive Bayes factor is in error, not a reason why my characterization was inaccurate.

Tyler isn't arguing for a positive Bayes factor. (I assume that by "Bayes factor" you mean the net effect on the posterior probability). He posted a followup because many people misunderstood him. Excerpt:

I did not try to justify any absolute level of belief in AGW, or government spending for that matter. I'll repeat my main point about our broader Bayesian calculations:

I am only saying that #2 [scientists behaving badly because they think the future of the world is at stake] deserves more than p = 0.

edited to add:

I'm not sure I understand you're criticism, so here's how I understood his argument. There are two major possibilities worth considering:

1) "These people behaved dishonorably.

and

2) "These people behaved dishonorably. They must have thought this issue was really important, worth risking their scientific reputations for.

Then the argument goes that the net effect of 1 is to lower our posteriors for AGW while the net effect of 2 is to raise them.

Finally, p(2 is true) != 0.

This doesn't tell us the net effect of the event on our posteriors - for that we need p(1), p(2) and p(anything else). Presumably, Tyler thinks p(anything else) ~ 0, but that's a side issue.

Is this how you read him? If so, which part do you disagree with?

Comment author: SilasBarta 02 December 2009 11:04:11PM *  1 point [-]

I assume that by "Bayes factor" you mean the net effect on the posterior probability).

I'm using the standard meaning: for a hypothesis H and evidence E, the bayes factor is p(E|H)/p(E|~H). It's easiest to think of it as the factor you mutiply your prior odds to get posterior odds. (Odds, not probabilities.) Which means I goofed and said "positive" when I meant "above unity" :-/

I read Tyler as not knowing what he's talking about. For one thing, do you notice how he's trying to justify why something should have p>0 under a Bayesian analysis ... when Bayesian inference already requires p's to be greater than zero?

In his original post, he was explaining a scenario under which seeing fraud should make you raise your p(AGW). Though he's not thinking clearly enough to say it, this is equivalent to describing a scenario under which the Bayes factor is greater than unity. (I admit I probably shouldn't have said "argument for >1 Bayes factor", but rather, "suggestion of plausibility of >1 Bayes factor")

That's the charitable interpretation of what he said. If he didn't mean that, as you seem to think, then he's presenting metrics that aren't helpful, and this is clear when he think's its some profound insight to put p(fraud due to importance of issue) greater than zero. Yes, there are cases where AGW is true despite this evidence -- but what's the impact on the Bayes factor?

Why should we care about arbitrarily small probabilities?

Tyler was not misunderstood: he used probability and Bayesian inference incorrectly and vacuously, then tried to backpedal. ( My comment on page 2.)

Anyway, I think we agree on the substance:

  • The fact that the p Tyler referred to is greater than zero is insufficient information to know how to update.
  • The scenario Tyler described is insufficient to give Climategate a Bayes factor above 1.

(I was going to the drop the issue, but you seem serious about de-Aumanning this, so I gave a full reply.)

Comment author: Matt_Simpson 03 December 2009 05:40:18AM *  1 point [-]

I think we are arguing past each other, but it's about interpreting someone else so I'm not that worried about it. I'll add one more bullet to your list to clarify what I think Tyler is saying. If that doesn't resolve it, oh well.

  • If we know with certainty that the secenario that Tyler described is true, that is if we know that the scientists fudged things because they knew that AGW was real and that the consequences were worth risking their reputations on, then Climategate has a Bayes factor above 1.

I don't think Tyler was saying anything more than that. (Well, and P(his scenario) is non-negligible)

Comment author: DanArmak 01 December 2009 07:53:54PM 2 points [-]

The second link doesn't load; should be this.

Comment author: CannibalSmith 01 December 2009 05:04:53PM 4 points [-]

I intend to participate in the StarCraft AI Competition. I figured there are lots of AI buffs here that could toss some pieces of wisdom at me. Shower me with links you deem relevant and recommend books to read.

Generally, what approaches should I explore and what dead ends should I avoid? Essentially, tell me how to discard large portions of porential-starcraft-AI thingspace quickly.

Specifically, the two hardest problems that I see are:
1. Writing an AI that can learn how to move units efficiently on its own. Either by playing against itself or just searching the game tree. And I'm not just looking for what the best StarCraft players do - I'm searching for the optimum.
2. The exact rules of the game are not known. By exact I mean Laplace's Demon exact. It would take me way too long to discover them through experimentation and disassembly of the StarCraft executable. So, I either have to somehow automate this discovery or base my AI on a technique that doesn't need that.

Comment author: Johnicholas 01 December 2009 06:47:11PM *  10 points [-]

I have some advice.

  1. Pay attention to the timing of your edit/compile/test cycle time. Efforts to get this shorter pay off both in more iterations and in your personal motivation (interacting with a more-responsive system is more rewarding). Definitely try to get it under a minute.

  2. A good dataset is incredibly valuable. When starting to attack a problem - both the whole thing, and subproblems that will arise - build a dataset first. This would be necessary if you are doing any machine learning, but it is still incredibly helpful even if you personally are doing the learning.

  3. Succeed "instantaneously" - and don't break it. Make getting to "victory" - a complete entry - your first priority and aim to be done with it in a day or a week. Often, there's temptation to do a lot of "foundational" work before getting something complete working, or a "big refactoring" that will break lots of things for a while. Do something (continuous integration or nightly build-and-test) to make sure that you're not breaking it.

Comment author: rwallace 01 December 2009 05:44:17PM 3 points [-]

Great! That competition looks like a lot of fun, and I wish you the best of luck with it.

As for advice, perhaps the best I can give you is to explain the characteristics the winning program will have.

It will make no, or minimal, use of game tree search. It will make no, or minimal, use of machine learning (at best it will do something like tuning a handful of scalar parameters with a support vector machine). It will use pathfinding, but not full pathfinding; corners will be cut to save CPU time. It will not know the rules of the game. Its programmer will probably not know the exact rules either, just an approximation discovered by trial and error. In short, it will not contain very much AI.

One reason for this is that it will not be running on a supercomputer, or even on serious commercial hardware; it will have to run in real time on a dinky beige box PC with no more than a handful of CPU cores and a few gigabytes of RAM. Even more importantly, only a year of calendar time is allowed. That is barely enough time for nontrivial development. It is not really enough time for nontrivial research, let alone research and development.

In short, you have to decide whether your priority is Starcraft or AI. I think it should be the latter, because that's what has actual value at the end of the day, but it's a choice you have to make. You just need to understand that the reward from the latter choice will be in long-term utility, not in winning this competition.

Comment author: RobinZ 01 December 2009 06:17:17PM 2 points [-]

Strictly speaking, this reads a lot like advice to sell nonapples. I'll grant you that it's probably mostly true, but more specific advice might be helpful.

Comment author: CannibalSmith 01 December 2009 06:09:03PM *  3 points [-]

That's disheartening, but do give more evidence. To counter: participants of DARPA's Grand Challenge had just a year too, and their task was a notch harder. And they did use machine learning and other fun stuff.

Also, I think a modern gaming PC packs a hell of a punch. Especially with the new graphics cards that can run arbitrary code. But good catch - I'll inquire about the specs of the machines the competition will be held on.

Comment author: rwallace 01 December 2009 07:20:37PM 4 points [-]

The Grand Challenge teams didn't go from zero to victory in one year. They also weren't one-man efforts.

That having been said, and this is a reply to RobinZ also, for more specifics you really want to talk to someone who has written a real-time strategy game AI, or at least worked in the games industry. I recommend doing a search for articles or blog posts written by people with such experience. I also recommend getting hold of some existing game AI code to look at. (You won't be copying the code, but just to get a feel for how things are done.) Not chess or Go, those use completely different techniques. Real-time strategy games would be ideal, but failing that, first-person shooters or turn-based strategy games - I know there are several of the latter at least available as open source.

Oh, and Johnicholas gives good advice, it's worth following.

Comment author: CannibalSmith 01 December 2009 08:57:59PM *  1 point [-]

The Grand Challenge teams didn't go from zero to victory in one year.

Stanford's team did.

They also weren't one-man efforts.

Neither is mine.

I do not believe I can learn much from existing RTS AIs because their goal is entertaining the player instead of winning. In fact, I've never met an AI that I can't beat after a few days of practice. They're all the same: build a base and repeatedly throw groups of units at the enemy's defensive line until run out of resources, mindlessly following the same predictable route each time. This is true for all of Command & Conquer series, all of Age of Empires series, all of Warcraft series, and StarCraft too. And those are the best RTS games in the world with the biggest budgets and development teams.

But I will search around.

Comment author: DanArmak 01 December 2009 09:29:49PM 1 point [-]

Was these games' development objective to make the best AI they could that would win in all scenarios? I doubt that would be the most fun for human players to play against. Maybe humans wanted a predictable opponent.

Comment author: ChrisPine 02 December 2009 11:40:22AM 2 points [-]

They want a fun opponent.

In games with many players (where alliances are allowed), you could make the AI's more likely to ally with each other and to gang up on the human player. This could make an 8-player game nearly impossible. But the goal is not to beat the human. The goal is for the AI to feel real (human), and be fun.

As you point out, the goal in this contest is very different.

Comment author: ShardPhoenix 04 December 2009 08:58:25AM 0 points [-]

There's some discussion and early examples here: http://www.teamliquid.net/forum/viewmessage.php?topic_id=105570

You might also look at some of the custom AIs for Total Annihilation and/or Supreme Commander, which are reputed to be quite good.

Ultimately though the winner will probably come down to someone who knows Starcraft well enough to thoroughly script a bot, rather than more advanced AI techniques. It might be easier to use proper AI in the restricted tournaments, though.

Comment author: whpearson 19 December 2009 11:58:08AM 0 points [-]

Is the status obsession that Robin Hanson finds all around him partially due to the fact that we live in a part of a world where our immediate needs are easily met? So we have a lot of time and resources to devote to signaling compared to times past.

Comment author: wedrifid 19 December 2009 01:28:29PM 1 point [-]

The manner of status obsession that Robin Hanson finds all around him is definitely due to the fact that we live in a part of a world where our immediate needs are easily met. Particularly if you are considering signalling.

I think you are probably right in general too. Although a lot of the status obsession remains even in resources scarce environments, it is less about signalling your ability to conspicuously consume or do irrational costly things. It's more being obsessed with having enough status that the other tribe members don't kill you to take your food (for example).

Comment author: whpearson 14 December 2009 10:59:19AM 0 points [-]

Are people interested in discussing bounded memory rationality? I see a fair number of people talking about solomonov type systems, but not much about what a finite system should do.

Comment author: wedrifid 14 December 2009 11:15:05AM 0 points [-]

Are people interested in discussing bounded memory rationality?

Sure. What about it in particular? Care to post some insights?

Comment author: whpearson 17 December 2009 12:28:27PM 0 points [-]

Would my other reply to you be an interesting/valid way of thinking about the problem. If not what were you looking for?

Comment author: wedrifid 17 December 2009 12:50:54PM 0 points [-]

Would my other reply to you be an interesting/valid way of thinking about the problem. If not what were you looking for?

Pardon me. Missed the reply. Yes, I'd certainly engage with that subject if you fleshed it out a bit.

Comment author: whpearson 14 December 2009 04:17:13PM 0 points [-]

I was thinking about starting with very simple agents. Things like 1 input, 1 output with 1 bit of memory and looking at them from a decision theory point of view. Asking questions like "Would we view it as having a goal/decision theory?" If not what is the minimal agent that we would and does it make any trade offs for having a decision theory module in terms of the complexity of the function it can represent.

Comment author: wedrifid 17 December 2009 12:41:56PM 0 points [-]

Things like 1 input, 1 output with 1 bit of memory and looking at them from a decision theory point of view. Asking questions like "Would we view it as having a goal/decision theory?"

I tend to let other people draw those lines up. It just seems like defining words and doesn't tend to spark my interest.

If not what is the minimal agent that we would and does it make any trade offs for having a decision theory module in terms of the complexity of the function it can represent.

I would be interested to see where you went with your answer to that one.