There's lots of different libertarian theories of ethics. Can you be more specific?
"Universally Preferable Behavior" by Stefan Molyneux, "Argumentation Ethics" by Hans Hermann Hoppe, and of course Objectivism, to name the most famous ones. Generally the ones I'm referring to all try to deduce some sort of Objective Ethics and (surprise) it turns out that property rights are an inherent property of the universe and capitalism is a moral imperative.
Forgive me if you're thinking of some other libertarians who don't have crazy ethical theories. I didn't mean to make gross generalizations. I've just observed that libertarian philosophers who consciously promote their theories of ethics tend to be of this flavor.
Why is the discrimination problem "unfair"? It seems like in any situation where decision theories are actually put into practice, that type of reasoning is likely to be popular. In fact I thought the whole point of advanced decision theories was to deal with that sort of self-referencing reasoning. Am I misunderstanding something?
Maybe "progress" doesn't refer to equality, but autonomy. It does seem like the progression of social organization generally leads to individual autonomy and equality of opportunity. Egalitarianism is a nice talking point for politicians, but when we say "progress" we really mean individual autonomy.
The kinds of writing produced based on 'praxeology', by people like Ludwig von Mises and Murray Rothbard, I have changed my mind to think as being not very reliable. I don't think that the arguments put forth by these Austrian Economists trying to justify when they reject empiricism and make weird assumptions are very good. This actually hurts a sizable amount of writing by these people, most of which I haven't read but any that I did end up having believed I now place very little trust in. The wikipedia synopsis of one of Mises' works should give you a fairly good idea of the kinds of assumptions he makes..
The links are fairly broad because I was admittedly trying to get a better grip on this stuff after having accepted much of it for a while, having given the whole school much benefit of the doubt early because the kinds of ideas it concluded with (stuff like lassiez-faire econ, not to regulate) fitted in with my preconcieved, often political, beliefs. Politics is the mind-killer, don'tcha know. Sorry for the lack of specific propositions.
I've read some Hayek too, he seems better, but I don't really know.
edit: changed a word to make a sentence make sense
Austrian-minded people definitely have some pretty crazy methods, but their economic conclusions seem pretty sound to me. The problem arises when they apply their crazy methods to areas other than economics (see any libertarian theory of ethics. Crazy stuff)
I've changed my mind on the persuasiveness of a specific argument. I used to hold a high degree of confidence in the line of reasoning that "since nobody can agree on just about anything about god, it is likely that god doesn't exist". But then, in an unrelated conversation, someone pointed out that it would be foolish to say that "since nobody can agree on the shape of the earth, earth has no shape." I must be question-begging!
I think the correct comparison would be, "since no one can agree on the nature of Earth/Earth's existence, Earth must not exist" but this is ridiculous since everyone agrees on at least one fact about Earth: we live on it. The original argument still stands. Denying the existence of god(s) doesn't lead to any ridiculous contradictions of universally experienced observations. Denying Earth's geometry does.
You know, I've always thought that graph was pretty inaccurate. A village idiot is not overwhelmingly smarter than a chimp. There was a time when the term "idiot" designated people who suffered from profound mental retardation, and anyone that low on the scale is probably significantly dumber than the average chimp. It's not like humans are so much smarter than other animals that our intelligence curves are totally non-overlapping with any other species. Koko the gorilla would probably be somewhere from moderately to severely retarded by human standards, although some of her mental faculties might be significantly better than those of a mentally retarded human. Even highly mentally disabled humans might be graced with neurological anatomy that's only appeared in the last few million years of evolutionary history, but that doesn't mean it's all functioning properly.
A modern car might have tremendous technological advantages over old ones, but drain enough of its lubricant and it'll still run worse than a Model T.
You are merely objecting to Eliezer's choice of scale. The distances between "intelligences" are pretty arbitrary. Plus he's using a linear scale, so there's no room for intelligence curves.
I think the DRH quote is pretty out of context, and Eliezer's commentary on it is pretty unfair. DRH has a deeply personal respect for human intelligence. He doesn't look forward to the singularity because he (correctly) points out that it will be the end of humanity. Most SI/LessWrong people accept that and look forward to it, but for Hofstadter the current view of the singularity is an extremely pessimistic view of the future. Note that this is simply a result of his personal beliefs. He never claims that people are wrong to look forward to superintelligence, brain emulation and things like that, just that he doesn't. See this interview for his thoughts on the subject.
You should collect data on time spent using the app and success. Do Science and stuff.
By the way, I spent a good amount of time using it yesterday and I just finished an entire Hershey's bar. Apparently it's not working for me.
In any decision involving an Omega like entity that can run perfect simulations of you, there wouldn't be a way to tell if you were inside the simulation or in the real universe. Therefore, in situations where the outcome depends on the results of the simulation, you should act as though you are in the simulation. For example, in counterfactual mugging, you should take the lesser amount because if you're in Omega's simulation you guarantee your "real life" counterpart the larger sum.
Of course this only applies if the entity you're dealing with happens to be able to run perfect simulations of reality.
View more: Next
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)
I read this book seven years ago, and I have vague memory that Adams was purposefully controversial rather than truth-seeking and, while maybe not out-and-out wrong, relies on misleading connotations to make his point. Based on my memory alone, I would downvote this. I'm curious well this memory holds up, so I'm holding off judgment for the moment while I skim it.
ETA: Wow, that was absolutely horrible. Meant as "serious philosophy" or not, that really failed. Perhaps for someone with no exposure to philosophy, it might useful to nudge them towards getting them to "think." Seems about as likely to be actively harmful, though. Adams is on the edge of clearing up some confusions, but then falls into worse ones himself.
Some gems:
It's not meant to be "serious philosophy". He's not presenting the ideas in the book as being literally true, he's just provoking the reader to look at the issues in the book in a different light. Forcing the reader to consider alternative hypotheses, if you will.