You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Eliezer_Yudkowsky comments on How SIAI could publish in mainstream cognitive science journals - Less Wrong Discussion

64 Post author: lukeprog 09 March 2011 09:17PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (76)

You are viewing a single comment's thread.

Comment author: Eliezer_Yudkowsky 10 March 2011 08:03:35PM 8 points [-]

Agree or disagree with the following statement?

"After publishing the paper in a philosophy journal so that academics would be allowed to talk about it without losing face, you would have to write a separate essay to explain the ideas to anyone who actually wanted to know them, including those philosophers."

Comment author: lukeprog 11 March 2011 01:07:42AM *  16 points [-]

Do you think Nick Bostrom's journal-published work on very similar subjects needs to be rewritten in different language to be understood? I don't, anyway. I personally find the style of mainstream philosophy and science much easier to understand than, say, your CEV paper. But that might be because mainstream philosophy and science is what I spend most of my time reading.

Frankly, I think your arguments can be made more clear and persuasive to a greater number of intelligent people if phrased in the common language.

Just because most philosophy is bad doesn't mean that when you write mainstream philosophy, you have to write badly.

Comment author: Vladimir_M 11 March 2011 05:46:40AM *  10 points [-]

lukeprog:

I personally find the style of mainstream philosophy and science much easier to understand than, say, your CEV paper. But that might be because mainstream philosophy and science is what I spend most of my time reading.

Seconded. I haven't read that many academic philosophy papers, but what I have seen has almost always been remarkably clear and understandable. I'm baffled that Eliezer would make such an extreme statement and actually mean it seriously (and get upvoted for it?!), considering how often he's cited academic philosophers like e.g. Chalmers, Bostrom, Dennett, or Parfit.

(Here of course I have in mind the Anglospheric analytic philosophy; continental philosophy is a horrible mess in comparison.)

Comment author: lukeprog 11 March 2011 06:20:56AM 1 point [-]

Yeah, don't get me started on continental philosophy.

BTW, one of my favorite takedowns of postmodernism is this one.

Comment author: Vladimir_M 11 March 2011 06:51:17AM *  0 points [-]

lukeprog:

BTW, one of my favorite takedowns of postmodernism is this one.

Thanks for the link. I skimmed the article and it seems well written and quite informative; I'll read it in full later.

In my opinion, there are some good insights in postmodernism, but as someone (Eysenck?) said about Freud, what's true in it isn't new, and what's new isn't true. In a sense, postmodernism itself provides perhaps the most fruitful target for a postmodernist analysis (of sorts). What these people say is of little real interest when taken at face value, but some fascinating insight can be obtained by analyzing the social role of them and their intellectual output, their interactions and conflicts with other sorts of intellectuals, and the implicit (conscious or not) meanings of their claims.

Comment author: cousin_it 11 March 2011 09:17:12PM *  2 points [-]

what's true in it isn't new, and what's new isn't true

The logical redundancy in this phrase has long bothered me.

Comment author: Vladimir_M 13 March 2011 04:09:47AM 2 points [-]

If I remember correctly, you're Russian? Those Slavic double negatives must be giving you constant distress, if you're so bothered by (seeming) deficiencies of logic in natural language.

Comment author: Emile 11 March 2011 09:48:43PM 0 points [-]

It's not redundant; it's a more witty and elegant way of saying that there are some new things, some true things, but none that are both.

Comment author: Sniffnoy 11 March 2011 10:24:34PM *  8 points [-]

It technically is redundant, though, because it has the form (A=>~B)&(B=>~A), while A=>~B and B=>~A are equivalent to each other. It doesn't need to be symmetrized because the statement was symmetric in the first place, even if it wasn't stated in an obviously symmetric form such as ~(A&B). (Going to have to say I like the redundant version for emphasis, though.)

Comment author: Eliezer_Yudkowsky 11 March 2011 05:33:12AM 4 points [-]

If we're talking about CEV, I agree. It needs rewriting. So does the Intuitive Explanation of Bayesian Reasoning, or any number of other documents produced by earlier Eliezers.

It was the linked Sobel paper which called forth that particular comment by me, if you're wondering. I looked at it in hopes of finding useful details about how to construe an extrapolated volition, and got a couple of pages in before I decided that I wasn't willing to read this paper unless someone had produced a human-readable version of it. (Scanning the rest did not change my mind.)

I'm not sure I want to import FAI ethics into philosophical academia as a field where people can garner prestige by writing papers that are hard to read. Maybe it makes sense to put up a fence around it and declare that if you can't write plain-English papers you shouldn't be writing about FAI.

Comment author: lukeprog 11 March 2011 06:17:33AM *  8 points [-]

Yes, the Sobel paper is definitely not an example of how I would write a philosophy paper, and your not reading it was a wise choice. Unfortunately, it is one of the major pieces in the literature on the subject of informed preference. But have you really never read any journal-published philosophy you thought was clear, such that you think one cannot write clearly if writing philosophy for journals? That would be shocking if true.

You will not stop people from garnering prestige by writing papers that are hard to read. You will also not stop people from writing hard-to-read papers on Friendly AI. That subject is already becoming a major field, whether you call it "machine ethics" or "machine morality" or "artificial morality" or "computational ethics" or "friendly AI." (As it turns out, "machine ethics" looks like it will win.)

But one can write clear and easy-to-read papers on Friendly AI. Who knows? Maybe it will even make your work stand out among all the other people writing on the subject, for example those proposing Kantian machines. (The horror!)

Comment author: Eliezer_Yudkowsky 11 March 2011 06:48:16AM 5 points [-]

Bostrom writes clearly.

But I will suggest for the record that we can probably get away with just ignoring anything that was written for other philosophers rather than for the general public or competent AGI researchers, since those are the only two constituencies we care about. If anyone in philosophy has something to contribute to the real FAI discussion, let them rewrite it in English. I should also note that anything which does not assume analytic naturalism as a matter of course is going to be rejected out of hand because it cannot be conjugate to a computer program composed of ones and zeroes.

Philosophers are not the actual audience. The general public and competent AGI researchers are the actual audience. Now there's some case to be made for trying to communicate with the real audience via a complicated indirect method that involves rewriting things in philosophical jargon to get published in philosophy journals, but we shouldn't overlook that this is not, in fact, the end goal.

Relevance. It's what's for dinner.

Comment author: lukeprog 11 March 2011 07:03:25AM *  18 points [-]

The AGI researchers you're talking about are the people who read IEEE Intelligent Systems and Minds and Machines. That's where this kind of work is being published, except for that tiny portion of stuff produced by SIAI and by Ben Goertzel, who publishes in his own online "journal", Dynamical Psychology.

So if you want to communicate with AGI researchers and others working on Friendly AI, then you should write in the language of IEEE Intelligent Systems and Minds and Machines, which is the language I described above.

The papers in Journal of Artificial General Intelligence follow the recommendations given above, too - though as a brand new online journal with little current prestige, it's far less picky about those things than more established journals.

Moreover, if you want to communicate with others about new developments in deontic logic or decision theory for use in FAI, then those audiences are all over the philosophical terrain, in mainstream philosophy journals not focused on AI. (Deontic logic and decision theory discussions are particular prevalent in journals focused on formal philosophy.)

Also, it's not just a matter of rewriting things in philosophical jargon for the sake of talking to others. Often, the philosophical jargon has settled on a certain vocabulary because it has certain advantages.

Above, I gave the example of making a distinction between "extrapolating" from means to ends, and "extrapolating" current ends to new ends given a process of reflective equilibrium and other mental changes. That's a useful distinction that philosophers make because there are many properties of the first thing not shared by the second, and vice versa. Conflating the two doesn't carve reality at its joints terribly well.

And of course I agree that anything not assuming reductionism must be dismissed.

But then, it seems you are interested in publishing for mainstream academia anyway, right? I know SIAI is pushing pretty hard on that Singularity Hypothesis volume from Springer, for example. And of course publishing in mainstream academia will bring in funds and credibility and so on, as I stated. It's just that, as you said, you don't have many people who can do that kind of thing, and those people are tied up with other things. Yes?

Comment author: lukeprog 11 March 2011 09:07:14AM *  14 points [-]

I just realized that maybe I'm confusing things by talking about philosophy journals, when really I mean to include cognitive science journals in general.

But what I said in my original post applies to cognitive science journals as well, it's just that when you're talking about philosophy (e.g. idealized preference theories of value), you place what you're saying in the context of the relevant philosophy, and when you're talking about neuroscience (e.g. the complexity of human values) then you place what you're saying in the context of the relevant neuroscience, and when you're talking about AI (e.g. approaches to AGI) then you place what you're saying in the context of relevant AI research. You can do all three in the same paper.

The kind of philosophy I spend most of my time reading these days is just like that, actually. Epistemology and the Psychology of Human Judgment spends just as much time discussing work done by psychologists like Dawes and Kahneman as it does discussing epistemologists like Goldman and Stich. Philosophy and Neuroscience: A Ruthlessly Reductive Account spends much more time discussing neuroscience than it does philosophy. Three Faces of Desire is split about 60/40 between philosophy and neuroscience. Many of the papers on machine ethics aka Friendly AI are split about 50/50 between philosophy and AI programming. Cognitive science is like this, after all.

In fact, I've been going through the Pennachin & Goertzel volume, reading it as a philosophy of mind book when most people, I guess, are probably considering it a computer science book. Whatever. Cognitive science is probably what I should have said. This is all cognitive science, whether it's slightly more heavy on philosophy or computer science or neuroscience or experimental psychology or whatever. The problem is that philosophy almost just is cognitive science, to me. Cognitive science + logics/maths.

Anyway, sorry if the 'philosophy' word caused any confusion.

Comment author: komponisto 12 March 2011 02:02:48AM 6 points [-]

You probably should have just titled it "How SIAI could publish in mainstream academic journals".

Comment author: lukeprog 12 March 2011 02:37:29AM *  2 points [-]

Maybe. But while I'm pretty familiar with philosophy journals and cognitive science journals, I'm not familiar with some other types of journals, and so I'm not sure whether my advice applies to, for example, math journals.

Comment author: komponisto 12 March 2011 02:51:08AM 4 points [-]

I'm not sure whether my advice applies to, for example, math journals.

It definitely does.

Comment author: Vladimir_Nesov 21 March 2011 01:06:05PM *  2 points [-]

Above, I gave the example of making a distinction between "extrapolating" from means to ends, and "extrapolating" current ends to new ends given a process of reflective equilibrium and other mental changes. That's a useful distinction that philosophers make because there are many properties of the first thing not shared by the second, and vice versa. Conflating the two doesn't carve reality at its joints terribly well.

Could you write up the relevant distinction, as applied to CEV, perhaps as a discussion post? I don't know the terminology, but expect that given the CEV ambition to get a long way towards the normative stuff, the distinction becomes far less relevant than when you discuss human decision-making.

(Prompted by the reference you made in this comment.)

Comment author: lukeprog 21 March 2011 05:00:35PM 1 point [-]

Did you read the original discussion post to which the linked comment is attached? I go into more detail there.

Comment author: Vladimir_Nesov 21 March 2011 05:47:14PM 1 point [-]

Yes, I read it, and it's still not clear. Recent discussion made a connection with terminal/instrumental values, but it's not clear in what context they play a role.

I expect I could research this discussion in more detail and figure out what you meant, but that could be avoided and open the issue to a bigger audience if you make, say, a two-paragraph self-contained summary. I wouldn't mention this issue if you didn't attach some significance to it by giving it as an example in a recent comment.

Comment author: lukeprog 21 March 2011 05:59:15PM 2 points [-]

I'm not sure what to say beyond what I said in the post. Which part is unclear?

In any case, it's kind of a moot point because Eliezer said that it is a useful distinction to make, he just chose not to include it in his CEV paper because his CEV paper doesn't deep enough into the detailed problems of implementing CEV where the distinction I made becomes particularly useful.

Comment author: JoshuaZ 11 March 2011 05:53:53AM 7 points [-]

Disagree. The area of philosophy I'm most familiar with (phil sci) is generally very easily understandable. I'm also not even sure this is a substantial objection. In many areas of learning, there are specialized vocabulary that take a lot of effort to understand. That's due to the deepness of the areas. Math is one example of this. I actually have more trouble reading papers in math, which is my own field, than I often do in biology (although this may be connected to the fact that I don't read highly technical papers in bio.) So even if your claim were true, it isn't at all clear to me why it would be relevant.

As a simple status issue, if you can get the philosophers to take you seriously, it will cause the people who aren't philosophers but who respect philosophy to take you more seriously.

Comment author: grouchymusicologist 10 March 2011 09:12:18PM 23 points [-]

Disagree. Pointlessly difficult and jargon-laden writing is not an inevitable feature of academic philosophical writing, just a common one. The best philosophical writing is as technical as it needs to be but also is clear, vivid, and even fun, and surely this should be the standard to aspire to for any SIAI-sponsored effort to produce papers pitched at the academic philosophy community.

Comment author: lukeprog 11 March 2011 02:03:48AM *  1 point [-]

Yes × 100.

Comment author: AlexMennen 10 March 2011 10:56:39PM 5 points [-]

I don't know. But even if you are correct, writing papers for the sole purpose of getting SIAI to be taken seriously, rather than actually explaining anything, sounds like a good idea anyway (assuming that it would work and that actually explaining things would not, that is).

Comment author: komponisto 10 March 2011 10:11:41PM 10 points [-]

How much academic philosophy have you personally read?

I've read a fair amount, and I don't find it particularly abstruse. This includes not only quasi-popular books by the highest-status practitioners like Dennett and Chalmers but also ordinary journal papers by professors at my undergraduate institution.

It might be worth taking a look at Chalmers' philosophy of mind anthology if you haven't already.

Comment author: lukeprog 11 March 2011 02:18:59AM 5 points [-]

Agree with your point, though I wouldn't say the extremely diverse set of essays in Chalmers' compilation are a shining example of philosophical clarity. I would recommend something like Epistemology and the Psychology of Human Judgment or Philosophy and Neuroscience: A Ruthlessly Reductive Account.

Comment author: komponisto 11 March 2011 06:12:39AM 2 points [-]

I wouldn't say the extremely diverse set of essays in Chalmers' compilation are a shining example of philosophical clarity

Oh, certainly not -- it's a sampler, and all levels of clarity and confusion present in the field are represented. I cited it to show the typical writing style of papers in philosophy (over the years, since as I recall it starts with Descartes!).

Comment author: David_Gerard 10 March 2011 11:38:13PM *  3 points [-]

We were discussing future fame and press coverage at the London meetup on Sunday (because a Fast Company journalist was present, no less - and participating in discussion very productively, I might add). I noted from Wikipedia's experience that the tech press are best treated with gunfire - do not talk to them under any circumstances. (There are individuals who are worth talking to, but they're very rare.) In retrospect, Wikipedia should really have gone headlong for the academic-interest press then the mainstream, bypassing the tech press entirely, from the beginning. An important place to apply the rule "taking someone seriously just because they pay you attention may not be a good idea."

What you do is philosophical engineering. Hit the philosophy journals and the tech press may find something more interesting to troll about.

(This ties into this thread.)