It is because a mirror has no commitment to any image that it can clearly and accurately reflect any image before it. The mind of a warrior is like a mirror in that it has no commitment to any outcome and is free to let form and purpose result on the spot, according to the situation.
—Yagyū Munenori, The Life-Giving Sword
This reminds me of how I felt when I learned that a third of the passengers of the Hindenburg survived. Went something like this, if I recall:
Apparently if you drop people out of the sky in a ball of fire, that's not enough to kill all of them, or even 90% of them.
Actually, according to Wikipedia, only 35 out of the 97 people aboard were killed. Not enough to kill even 50% of them.
I have become 30% confident that my comments here are a net harm, which is too much to bear and so I am discontinuing my comments here unless someone cares to convince me otherwise.
Edit: Good-bye.
What the hell? Did you catch Konkvistador disease or something? What is up with high-quality contributors deciding to up and leave?
I, for one, think that your posts are valuable to the community.
(I'm assuming you mean that you suspect your comments cause harm to others, obviously if you think you're spending to much time procrastinating on LessWrong then leaving is fine.)
Which is not the same thing as expecting a project to take much less time than it actually will.
Edit: I reveal my ignorance. Mea culpa.
Parts of this I think are brilliant, other parts I think are absolute nonsense. Not sure how I want to vote on this.
there is no way for an AI employing computational epistemology to bootstrap to a deeper ontology.
This strikes me as probably true but unproven.
My own investigations suggest that the tradition of thought which made the most progress in this direction was the philosophical school known as transcendental phenomenology.
You are anthropomorphizing the universe.
That isn't the planning fallacy.
This is a better explanation than I could have given for my intuition that physicalism (i.e. "the universe is made out of physics") is a category error.
Whether or not a non-self-modifying planning Oracle is the best solution in the end, it's not such an obvious privileged-point-in-solution-space that someone should be alarmed at SIAI not discussing it. This is empirically verifiable in the sense that 'tool AI' wasn't the obvious solution to e.g. John McCarthy, Marvin Minsky, I. J. Good, Peter Norvig, Vernor Vinge, or for that matter Isaac Asimov.
Nonsense. The problem has posed has always been around, and the solution is just to avoid repeating the same state twice, because that results in a draw.
I like this. I was going to say something like,
"Suppose , what does that say about your solutions designed for real life?" and screw you I hate when people do this and think it is clever. Utility monster is another example of this sort of nonsense.
but you said the same thing, and less rudely, so upvoted.
Perhaps it was imprudent, but I assumed that someone trying to promote rationality would herself be rational enough to overcome this parochialism bias.
"Any and all cost" would subsume low probabilities if it were true (which, of course, it is not).
The poll results are intriguing. 35% would want cybernetic immortality at any and all cost! And yet I don't see 35% of people who can afford it signed up for cryonics.
35% of people who read the article and took the quiz right after reading. Admittedly the numbers are larger than I naively expected to get past each stage of filtering. I was thinking maybe 300 people taking the poll total and its over 30k now, with over 10k answering "at any cost". (I would love to see the log files for that page... referring URLs, IP addresses, etc.)
Because cryonics is not even close to "certain immortality now".
...A failure or so, in itself, would not matter, if it did not incur a loss of self-esteem and of self-confidence. But just as nothing succeeds like success, so nothing fails like failure. Most people who are ruined are ruined by attempting too much. Therefore, in setting out on the immense enterprise of living fully and comfortably within the narrow limits of twenty-four hours a day, let us avoid at any cost the risk of an early failure. I will not agree that, in this business at any rate, a glorious failure is better than a petty success. I am all for the
I personally found the research in Influence rather lacking and thought Cialdini speculated too much. But chapter 3 of the book is dead on.
Do people think superrationality, TDT, and UDT are supposed to be useable by humans?
I had always assumed that these things were created as sort of abstract ideals, things you could program an AI to use (I find it no coincidence that all three of these concepts come from AI researchers/theorists to some degree) or something you could compare humans to, but not something that humans can actually use in real life.
But having read the original superrationality essays, I realize that Hofstadter makes no mention of using this in an AI framework and instead thinks...
An expert on political ruthlessness, not an expert at political ruthlessness!
My current guess is that having the knows-the-solution property puts them in a different reference class. But if even a tiny fraction deletes this knowledge...
The part you highlight about shminux's comment is correct, but this part:
this would define "looks attractive to a certain subset of humans"
is wrong; attractiveness is psychological reactions to things, not the things themselves. Theoretically you could alter the things and still produce the attractiveness response; not to mention the empirical observation that for any given thing, you can find humans attracted to it. Since that part of the comment is wrong but the rest of it is correct, I can't vote on it; the forces cancel out. But anyway I ...
For a while, I assumed that I would never understand UDT. I kept getting confused trying to understand why an agent wouldn't want or need to act on all available information and stuff. I also assumed that this intuition must simply be wrong because Vladimir Goddamned Nesov and Wei Motherfucking Dai created it or whatever and they are both straight upgrades from Grognor.
Yesterday, I saw an exchange involving Mitchell Porter, Vladimir Nesov, and Dmytry_messaging. The latter of these insisted that one-boxing in transparent Newcomb's (when the box is empty) wa...
Both of the studies linked to at the top of this post, on which the entire post is based, have been discredited. Even if they were true, I think it was a stretch to go from those to postulating a generalized verbal overshadowing bias.
With the benefit of hindsight I can say that this post was probably a mistake, which leaves me a bit dumbfounded at its karma score of 61 and endorsement by Newsome. When I scrolled down to the bottom I saw that I had already downvoted it, which made me even more confused.
I am no more qualified to disprove a religious belief than I would be to perform surgery... on anything.
I disagree with this claim. If you are capable of understanding concepts like the Generalized Anti-Zombie Principle, you are more than capable of recognizing that there is no god and that that hypothesis wouldn't even be noticeable for a bounded intelligence unless a bunch of other people had already privileged it thanks to anthropomorphism.
Also, please don't call what we do here, "rationalism". Call it "rationality".
On reflection, I agree with you and will be downvoting all of Clippy's comments and those of all other abusive sockpuppets I'm aware of.
I really wish you would have put a disclaimer on these posts the likes of:
One of the assumptions The Art of Strategy makes is that rational agents use causal decision theory. This is not actually true, but I'll be using their incorrect use of "rationality" in order to make you uncomfortable.
Anyway,
Nick successfully meta-games the game by transforming it from the Prisoner's Dilemma (where defection is rational) [...]
this is the problem with writing out your whole sequence before submitting even the first post. You make the later posts insu...
I initially had the parent upvoted, but I retracted it on learning that the grandparent comment is speaking from experience, and since I have the same experience, it's difficult not to believe.
Could someone please explain to me exactly, precisely, what a utility function is? I have seen it called a perfectly well-defined mathematical object as well as not-vague, but as far as I can tell, no one has ever explained what one is, ever.
The words "positive affine transformation" have been used, but they fly over my head. So the For Dummies version, please.
The first half of the first sentence of your comment is incomprehensible. "If you throw enough money into the hole that you spend less" is somewhere between gibberish and severely sleep-deprived mumblings.
A -2 score means almost nothing. Two people downvoted your comment, so what?
Calling the community "pathetic" in response to downvotes is hypocrisy.
In all likelihood, the downvoters did not even recognize your "Keynesian slant" and were downvoting because signaling. Oftentimes people will downvote a -1 comment just by t
I believe this term is used solely to countersignal and has no more technical meaning than "guy I don't like who defends females".
Hello, friend, and welcome to Less Wrong.
I do think you should start a discussion post, as this seems clearly important to you.
My advice to you at the moment is to brush up on Less Wrong's own atheism sequence. If you find that insufficient, then I suggest reading some of Paul Almond's (and I quote):
If you find that insufficient, then it is time for the big guy, Richard Dawkins:
If you are somehow still unsatisfied after all this, lukeprog's new website should direct you to some other resources, ...
It occurred to me that although I agree with Statement #5 - "People are morally responsible for the opportunity costs of their actions," I do not think it is a claim actually being made by the optimal philanthropy zeitgeist. I think the actual claim is "Your actions have opportunity costs and you should probably think about that," which should be uncontroversial.
After the addition I just made, my list contains 58 items, and yours contains 47. If you feel like updating. (Also, I removed one, because William Eden deleted one of his accounts.)
Comments section makes for interesting reading.
No it doesn't; it is the standard cloud of thought-failure.
Do you see that as a good thing?
So, let's take this hypothetical (harrumph) youth. They see irrationality around them, obvious and immense, they see the waste and the pain it causes. They'd like to do something about it. How would you advise them to go about it?
Donate to CFAR. There's no good reason to demand a local increase in rationality.
[...]should we try to distance ourselves from atheism and anti-religiousness as such? Is this baggage too inconvenient, or is it too much a part of what we stand for?
We don't stand for atheism; we stand by atheism, prepared to walk away at any ...
The title of this post tempted me to make another article called "Eliezer apparently right about just about everything else" but I already tried that and it was a bad idea.
My sense is that this assertion can be empirically falsified for all levels of abstraction below "Do what is right."
Indeed, this is one of many reasons why I am starting to think "go meta" is really, really good advice.
Edit: Clarification, what I mean is that I think virtue ethics, deontology, utilitarianism, and the less popular ethical theories agree way more than their proponents think they do. At this point this is still a guess.
[...]and related-to-rationality enough to deserve its own thread.
I've gotten to thinking that morality and rationality are very, very isomorphic. The former seems to require the latter, and in my experience the latter gives rise to the former. So they may not even be completely distinguishable. We've got lots of commonalities between the two, noting that both are very difficult for humans due to our haphazard makeup, and both have imaginary Ideal versions (respectively: God, and the agent who only has true beliefs and optimal decisions and infinite comp...
Irrationality game comment
The correct way to handle Pascal's Mugging and other utilitarian mathematical difficulties is to use a bounded utility function. I'm very metauncertain about this; my actual probability could be anywhere from 10% to 90%. But I guess that my probability is 70% or so.
On the People page, the picture next to Richard Carrier's name is the same as the picture next to Richard Boyd's.
and that's because science.
That made me smile, [edit] especially because I had just recently read the Language Log article Because NOUN.
The rest of the video made me kind of uncomfortable, though, because it felt like (and I guess sort of was) an advertisement, and you keep saying "worldview naturalism" where anyone else would have said "the naturalistic worldview" or just "naturalism".
(And this is just a person...
Signing up was a waste of time. This course is for people who don't know anything at all.
Much to everyone else's chagrin.
Nope, I'm talking about the humans' in questions subjective "nows", not their futures. Although if a person isn't particularly rational and has never heard of rationality and if you mentioned it to him he wouldn't feel particularly motivated to become more rational has a pretty irrational-looking future, and in such case there's no choice to make, no will, only a default path.
I suggest a new rule: the source of the quote should be at least three months old. It's too easy to get excited about the latest blog post that made the rounds on Facebook.