pjeby comments on Raising the Sanity Waterline - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (207)
Unless Eliezer has some supernatural entity to do his thinking for him, his ethics and meta-ethics require some physical implementation. Where else are you proposing that he store and process them, besides physical reality?
I think you're shifting between 'ethics' and 'what Eliezer thinks about ethics'. While it's possible that ideas are not real save via some implementation, I don't think it would therefore have to be in a particular human; systems know things too.
You seem to frequently shift the focus of conversation as it happens, hurting the potential for rational discourse in favor of making emotively positive statements that loosely correlate with the topic at hand. Would you be the same pjeby that writes those reprehensible self-help books?
That seemed a bit ad hominem. The commenter pjeby (I know nothing else about him) seems like someone who might be unfamiliar with part of the LW/OB background corpus but is reasoning pretty well under those conditions.
Actually, I'm quite familiar with a large segment of the OB corpus -- it's been highly influential on my work. However, I also see what appear to be a few holes or incoherencies within the OB corpus... some of which appear to stem from precisely the issue I've been asking you about in this thread. (i.e. the role of negative utilities in creating bias)
In my personal experience, negative utilities create bias because they cut off consideration of possibilities. This is useful in an emergency -- but not much anywhere else. If human beings had platonically perfect minds, there would be no difference between a uniform utility scale and a dual positive/negative one... but as far as I can tell (and research strongly suggests) we do have two different systems.
So, although you're wary of Robin's "cynicism" and my "psychological explanations", this is inconsistent with your own statements, such as:
See, I'm as puzzled by your ability to write something like that, and then turn around and argue an absolute utility for unhappiness, as you are puzzled by that Nobel-winning Bayesian dude who still believes in God. From my POV, it's just as inconsistent.
There must be some psychology that creates your position, but if your position is "truly" valid (assuming there were such a thing), then the psychology wouldn't matter. You should be able to destroy the position, and then reconstruct it from more basic principles, once the original influence is removed, no? (This idea is also part of the corpus.)
pjeby,
Are you familiar with Eliezer's take on naturalistic meta-ethics in particular, or just with other large segments of the OB corpus? If the former, maybe you could take more care to spell out that you get the difference between "achieving one's original goals" and "hacking one's goal-system so that the goal-system thinks one has acheived one's goals (e.g., by wireheading)".
I like your writing, but in this particular thread, my impression is that you're "rounding to the nearest cliche" -- interpreting Eliezer and others as saying the nearest mistake that you've heard your students or others make, rather than making an effort to understand where people are coming from. My impression may be false, but it sounds like I'm not the only one who has it, and it's distracting, so maybe take more care to spell out in visible terms a summary of peoples' main points, so we know you're disagreeing with what they're saying and not with some other view.
More generally, you've joined a community that has been thinking awhile and has some unusual concepts. I'm glad you've joined the commenters, because we badly need the best techniques we can get for changing our own thinking habits and for teaching the same to others -- we need techniques for learning and teaching rationality -- and I find your website helpful here, and your actual thinking on the subject, in context, can probably become better still. But I wonder if you could maybe take a bit more care in general to hear the threads you're responding to. I've felt like you were "rounding to the nearest cliche" in your thread with me as well (I wasn't going off the Lisa Simpson happiness theory), and it might be nice if you could take the stance of a co-participant in the conversation, who is interested in both learning and teaching, instead of repeating the (good) points on your website in response to all comments, whatever the comments' subject matter.
First, yes, I do understand the the difference between goal-achievement and wireheading. I'm drawing a much finer distinction about the means by which you set up a system to achieve your goals, as well as the means by which you choose those goals in the first place.
It is possible in some cases that I've "rounded to the nearest cliche" as you put it. But I'm pretty confident that I'm not doing that with Eliezer's points, precisely because I've read so much of his work... but also because the mistake I believe he is making (or at least, the thing he appears to not be noticing) is a perfect example of a point that I was trying to make in another thread... about why you can't just put one new, correct belief in someone's head, and have it magically fix every broken belief they already have.
I'm a little confused about the rest of your statement; it doesn't seem to me that I'm repeating the same points, so much as that I've been struggling to deal with the fact that so many of the threads I've become involved in, boil down (AFAICT) to the same issues -- and trying NOT to have too much duplication in my responses, while also not wanting to create a bunch of inter-comment links. (Another fine example of how avoiding negatives leads to bad decisions... ;-) )
Now, whether that's a case of me having only a hammer, or whether it's simply because everything really is made out of ones and zeros, I'm not sure. It has been seeming to me for a bit now, that what I really need to do is write an LW article about positive/negative utility and abstract/concrete thinking, as these are the main concepts I work with that clash with some portions of the OB corpus (and some of the more vocal LW commenters). Putting that stuff in one place would certainly help reduce duplication.
Meanwhile, it's not my intention to reduce anyone to cliche, or to presume that I understand something I don't. If I were, I wouldn't spend so much time in so many of my comments, asking so many questions. They are not rhetorical; they represent genuine curiosity. And I've actually learned quite a lot from the process of asking and commenting in the last few days; many things I've written here are NOT concepts I previously had.
This is especially true for the two comments that were replies to you; they were my musings on the ideas I got from your statements, more than critique or commentary of anything you said. I can see how that might make you feel not understood, however. (Also, the "Lisa Simpson theory" part of that one comment was actually directed to the comment you were replying to, not your comment in that thread, btw. I was trying to avoid writing two replies there.)
I also get the sense that you're trying to say something off-the-cuff in your replies that would be better done as a specific LW post.
Thanks for the thoughtful reply. It's quite possible I misinterpreted. Also, re: the Lisa Simpson thing, I'll be more careful to look at other nearby posts people might be replying to instead of reading comments so much from the new comments page.
It seems slightly odd to me that you say you're "pretty confident" you're not rounding Eliezer's point to the nearest cliche in part because the mistake you think he's making "is a perfect example of a point [you] were trying to make in another thread". Isn't that what it feels like when one rounds someone's response to a pre-existing image of "oh, the such-and-such mistake"?
A LW article about how people think about positive/negative utility, and another about abstract/concrete thinking, sounds wonderful. Then we can sift through your concepts as a community, air confusions or objections in a coherent manner, etc.; and you can reference it and it'll be part of our shared corpus. Both topics sound useful.
So, how would you distinguish that, from the case where their response is making the such-and-such mistake?
The way I'd distinguish it, is to ask questions that would have different answers, depending on whether the person is making that mistake or not. I asked Eliezer those questions, and of the ones he answered, the answers were consistent with my model of the mistake.
Of course, there's always the possibility of confirmation bias... except that I also know what answers I'd have taken as disconfirming my hypothesis, which makes it at least a little less likely. (But I do know of more than one mechanism by which beliefs and behaviors are formed and maintained, and it would've been plausible -- albeit less probable -- that his evaluation could've been formed another way. And I'd have been perfectly okay with my hypothesis being wrong.)
See, I'm not pointing out what I believe to be a mistake because I think I'm smarter than Eliezer... it's because I'm constantly making the same mistake. We all do, because it's utterly trivial to make it, and really non-trivial to spot it. And if you haven't gotten an intuitive grasp of why and how that mistake comes into being (for example, if you insist it doesn't exist in the first place!), then it's hard to see why there's "no silver bullet" for reducing the complexity of developing "rationality" in people.
If my interlocutor is someone who might well have thoughts that don't fit into my schemas, I might be suspicious enough of my impression that they were making one of my standard cached example-mistakes that I'd:
Make a serious effort at original seeing, and make sure my model of the such-and-such mistake is really the best way to organically understand the situation in front of me; and then
Describe my schema for the such-and-such mistake (in general), and see if the person agrees that such-and-such is a mistake; and then
Describe the instance of the such-and-such mistake that the person seems to be making, and ask if they agree or if there's a kind of reasoning going into their claim that doesn't fit into my schema.
Or maybe this is just the pain-in-the-neck method one should use if one's original communication attempt stalls somewhere. Truth be told, I'm at this point rather confused about which aspects of meta-ethics under dispute, and I can't easily scan back through the conversation to find the quotes of yours that made me think you misunderstood because our conversation has overflowed LW's conversation-display settings. And you've made some good points, and I'm convinced I misunderstood you in at least some cases. I'm going to bow out of this conversation for now and wait to discuss values and their origins properly, in response to your own post. (Or if you'd rather, I'd love to discuss by email; I'm annasalamon at gmail.)
Yes, the comment system here is really not suited to the kind of conversation I've been trying to have... not that I'm sure what system would work for it. ;-)
As far as meta-ethics goes, the short summary is:
"Avoiding badness" and "seeking goodness" are not interchangeable when you experience them concretely on human hardware,
It is therefore a reasoning error to treat them as if they were interchangeable in your abstract moral calculations (as they will not work the same way in practice),
Due to the specific nature of the human hardware biases involved (i.e., the respective emotional, chemical, and neurological responses to pain vs. pleasure) , badness-avoidance values are highly likely to be found irrational upon detailed examination... and thus they are always the ones worth examining first.
Badness-avoidance values are a disproportionately high (if not exclusive!) source of "motivated reasoning". i.e., we don't so much rationalize to paint pretty pictures, as to hide the ugly ones. (Which makes rooting them out of critical importance to rationalists.)
This summary is more to clarify my thoughts for the eventual post, than an attempt to continue the discussion. (To me, these things are so obvious and so much a part of my day-to-day experience that I often forget the inferential distance involved for most people.)
These ideas are all capable of experimental verification; the first one has certainly been written about in the literature. None are particularly unorthodox or controversial in and of themselves, as far as I'm aware.
However, there are common arguments against some of these ideas that my own students bring up, so in my (eventual) post I'll need to bring them up and refute them as well.
For example, a common argument against positively-motivated goodness is that feeling good about being generous means you're "really" being selfish... and thus bad! So, the person advancing this argument is motivated to rationalize the "virtue" of being dutiful -- i.e., doing something you don't want to, but nonetheless "should" -- because it would be bad not to.
Strangely, most people have these judgments only in relation to their self... They see no problem with someone else doing good out of generosity or kindness, with no pain or duty involved. It's only themselves they sentence to this "virtue" of suffering to achieve goodness. (Which is sort of like "fighting for peace" or "f*ing for virginity", but I digress.)
Whether this is something inbuilt, cultural, or selection bias of people I work with, I have no idea. But it's damn common... and Eliezer's making a virtue out of unhappiness (beyond the bare minimums demanded by safety, etc.) fits smack dab in the middle of this territory.
Whew. Okay, I'm going to stop writing this now... this really needs to be a post. Or several. The more I think about how to get here, starting from only the OB corpus and without recapitulating my own, the bigger I realize the inferential gap is.
It was deliberately ad hominem, of course - just not the fallacious kind. We seriously need profile pages of some sort. Wish I had the stomach for Python.
I don't expect anyone to be familiar with the LW/OB background corpus - I expect my education and training is quite different from yours, for example. However, I still expect one to follow rules of conduct with respect to reasonable discourse, for example avoiding equivocation and its related vices.
Or maybe I'm just viscerally angered by the winky smileys. Who knows.
I don't see how I can separate "ethics" from "what Eliezer thinks about ethics" and still have a meaningful conversation with him on the topic.
Meanwhile, reading back through the thread, the only digressions I see in my comments are those made in response to those raised by you or Eliezer. Perhaps you could point to some specific examples of these shifted foci and emotively positive statements? I do not see them.
As for my "reprehensible" books, I trust you formed that judgment by actually reading them, yes? If so, then yes, I'm that person. But if you didn't read them, then clearly your judgment isn't about the books I actually wrote... and thus, I could not have been the person who wrote the (imaginary) ones you'd therefore be talking about. ;-)
I was not referring only to this thread, but to several ongoing discussions. If you'd like clear examples, feel free to contact me via http://thomblake.com or http://thomblake.mp
As Eliezer has kindof pointed out, I'm weary enough from this discussion to be on the verge of irrationality, so I shall retire from it (if only because this forum is devoted to rationality!).