Cambridge's total colleges endowments is 2.8 and Oxford's 2.9. But the figures above already include this.
Violence might not be the exact opposite of peace. Intuitively, peace seem to mean a state where people are intentionally not committing violence and not just accidentally. A prison might have lower violence than an certain neighbourhood but it might still not be considered a more peaceful place exactly because the individual proclivity to violence is higher despite the fact violence itself isn't. Proclivity matters.
I am generally sceptic of Pinker. I have read a ton of papers and Handbooks of Evolutionary Psychology, and it is clear that while he was one ...
I made my above comment because I knew of at least one clear instance where the reason I had to do the workaround was due to someone who found Alex's stuff. But things haven't improved that much as I anticipated in my field (Applied Ethics). These things would take time, even if this had Alex's stuff as the only main cause. Looking back, I also think part of the workarounds were more due to having to relate the discussion to someone in my field who wrote about the same issue (Nick Agar) than due to having to avoid mentioning Eliezer too much.
I see a big d...
What about this one?
...Once Braumoeller took into account both the number of countries and their political relevance to one another, the results showed essentially no change to the trend of the use of force over the last 200 years. While researchers such as Pinker have suggested that countries are actually less inclined to fight than they once were, Braumoeller said these results suggest a different reason for the recent decline in war. “With countries being smaller, weaker and more distant from each other, they certainly have less ability to fight. But we
I think there is more evidence it crosses (two studies with spinal measures) than it does not (0 studies). For (almost) direct measures check out Neumann, Inga D., et al., 2013 and Born, 2002. There are great many studies showing effects that could only be caused by encephalic neuromodulation. If it does not cross it, then it should cause increased encephalic levels of some neurochemical with the exact same profile, but that would be really weird.
Regardless of attachment style, oxytocin increases in-group favouritism, proclivity to group conflict, envy and schadenfreude. It increases cooperation, trust and so on inside one's group but it often decreases cooperation with out-groups.
I may not be recalling correctly, but although there is some small studies on that, I do not think there is a lot of evidence that oxytocin always leads anxiety, etc. in people with insecure attachment style. I would suspect that it might be the case it initially increases insecurity because it makes those persons attend ...
Elephants kill hundreds, if not thousands, of human beings per year. Considering there are no more than 30,000 elephants alive, that's an amazing feat of evilness. I believe the average elephant kills orders of magnitudes more than the average human, and probably kill more violently as well.
Worth mentioning that some parts of Superintelligence are already a less contrarian version of many arguments made here in the past.
Also note that although some people do believe that FHI is some sense "contrarian", when you look at the actual hard data on this the fact is FHI has been able to publish in mainstream journals (within philosophy at least) and reach important mainstream researchers (within AI at least) at rates comparable, if not higher, to excellent "non-contrarian" institutes.
I didn't see the post in those lights at all. I think it gave a short, interesting and relevant example about the dynamics of intellectual innovation in "intelligence research" (Jeff) and how this could help predict and explain the impact of current research(MIRI/FHI). I do agree the post is about "tribalism" and not about the truth, however, it seems that this was OP explicit intention and a worthwhile topic. It would be naive and unwise to overlook these sorts of societal considerations if your goal is to make AI development safer.
Is there a thread with suggestions/requests for non-obvious productivity apps like that? Because I do have a few requests:
1) One chrome extension that would do this, but for search results. That is, that upon highlighting/double-clicking a term would display a short list of top Google search results in a context/drop-box menu on the same page.
2) Something like the StayFocusd extension that blocks sites like Facebook and YouTube for a given time of the day, but which would be extremely hard to remove. Some people suggested to block these websites IPs direct...
Sorry, I meant my office at work (yeap...). Fixed that.
Thanks! This will be useful for me as well, it definitely seems better than my current solution: leaving my cell phone locked in my office(EDIT: at work).
I am so glad that finally some intellectual forum has passed the Sokal test. Computer Science, Sociology, and Philosophy have all failed, and they haven't tried with the rest yet.
LessWrong, you are our only hope.
Can't do. Search keywords as cortisol dominance rank status uncertainty.
Which fields are these? This sounds to me a definition that could be useful in e.g. animal studies, but vastly insufficient when it comes to the complexities of status with regard to humans.
Yes, it came from animal studies; but they use in evolutionary psychology as well (and I think in cognitive psychology and biological anthropology too). Yes, it is vastly insufficient. However, I think it is the best we have. More importantly, it is the least biased one I have seen (exactly because it came from animal studies). I feel like most definitions of status ...
Not sure if people are aware, but there are a lot of studies backing up that claim. It is more taxing (to well-being, not to fitness, of course) What's more, the alpha is is most stressed member of groups with high status-uncertainty, and the least stressed in a group with low status-uncertainty.
This also reminded me of this study, which found that "wealthy individuals report that having three to four times as much money would give them a perfect "10" score on happiness--regardless of how much wealth they already have."
In most scientific fields status is defined as access (or entitlement) to resources (i.e.: food and females, mostly). Period. And they tend to take this measure very seriously and stick to it (it has many advantages, easy to measure, evolutionary central, etc.). Both your definitions are only two accidental aspects of having status. Presumably, if you have - and in order to have - higher access to resources you have to be respected, liked, and have influence over your group. I think the definition is elegant exactly because all the things we perceive as s...
It would seem I'm not the norm. I have been going there for just over one year. But I find it hard to believe people would be generally against any form of organising the comments by quality. It would be nice to know which of the 400 comments is worth reading. Do people simply read all of them? Do they post without reading any? I think I have been here, and mostly only here, for so long that other systems do not make sense to me.
Sorry, I intended to mean that the comments are dramatically worse than the posts. But then again this might be true of most blogs. However, it's not true of the blogs I wish and find useful to visit.
This a blog that supports up/downvotes with karma in which comments are not dramatically worse than the post, and sometimes even better.
I would be more in favour of pushing SSC to have up/downvotes than to linking its posts here. I find that although posts are high quality the comments are generally not, so this is a problem that definitely needs to be solved on its own. Moreover, I read both blogs and I like to have them as separate activities given that they have pretty different writing styles and mildly different subjects. I tend I to read SSC on my leisure time, while LessWrong is a gray area. I would certainly be against linking every single post here given that some of them would be decisively off topic.
This looks like a good idea. I feel that adrenaline rush I normally feel when I plan to set up something that will certainly make me work (like when setting up beeminder). However, I wouldn't like to do this via a chat room, unless via email fails. I don't like the fact a chatroom will drag 100% of my attention and time during a specific amount of time. Moreover, my week is not stable enough to commit on fixed weekly chats. I realise that by chat there's more of a social bonding thing that would entail more peer-pressure, but I think that by email there wi...
I don't currently have a facebook account and I know that a lot of very productive people here in Oxford that decided not to have one as well (e.g., Nick and Anders don't have one). I think adding the option to authenticate via Google is an immediate necessity.
I am not sure how much that counts as willpower. Willpower, often, has to do with the ability to revert preference reversal caused by hyperbolic discounting. When both the rewards are far away, we use a more abstract, rational, far-mode or system 2 reasoning. You have rationally evaluated both options (eating vs. not eating the cake) and decided not to eat. Also, I would suspect that if you merely decide this one day before and do nothing about it, you will eat the cake with more or less the same probability if you haven't decided. However, if you decide not to eat but take measures to not eat the cake, for instance, telling your co-worker you will not eat it, then it might be more effective and count as willpower.
There's good evidence that socioeconomic status correlates positively with Self-Control. There is also good evidence that people with high socioeconomic status live in a more stable environment during childhood. The signals of a stable environment correlating with Self-Control is his speculation as far as I'm aware, but in light of the data it seems plausible.
I agree they would function better in a crisis, but a crisis is a situation where fast response matters more than self-control. In a crisis you will take actions that are probably wrong during stable periods. I would go on to say, as my own speculation, that hardship - as else being equal - make people worse.
Neil's theory has different empirical predictions than Baumeister's, for example, it predicts high Self-Control correlates with low direct resistance to temptations. On the second Lecture he mentions several experiments that would tell them apart. They are different theoretically, there's a difference in the importance they give to willpower. Saying you should save water on the Sahara is different from saying you shouldn't lose your canteen's cover.
It is surely my experience in life that people highly overestimate their causal effectiveness in the world, ...
You are right, willpower is not irrelevant, perhaps this was not the best phrasing. I meant that willpower is irrelevant relative to other self-control techniques, but perhaps I should have said less relevant. I have changed the title to "the myth of willpower".
It's important to be made clear he argues that the use of willpower and self-control are inversely correlated, after that minimal amount of willpower it takes to deploy self-management techiniques. It would be incorrect to assume he is defending a view where willpower is as central as in any of the other views (or as intuitively seems to be).
I think effortful self-control would be one. Probably around the middle of the second lecture he offers a better one as he clearly sets apartment measures of self-control and measures of willpower. Unfortunately I can't remember well enough but it goes along the lines of effortful self-control, the simple and direct resistance to temptation. Looking and smelling the chocolate cake but not eating would take willpower, while freezing the cake so it always takes a couple of hours between deciding to eat and being able to eat it would be self-control as he defines.
You or your son might find this lecture on swearing helpful: http://blog.practicalethics.ox.ac.uk/2015/02/on-swearing-lecture-by-rebecca-roache/ And here's the audio: http://media.philosophy.ox.ac.uk/uehiro/HT15_STX_Roache.mp3
I understand the pragmatic considerations for inhibiting swearing, but he seems so smart that he should be allowed to swear. You should just tell the school he is too smart to control, but they can try themselves.
I wish I was 10 so I could befriend him.
As the person who first emailed Rudi back in 2009 so you could finally stop cryocrastinating, I'm willing to seriously dig up whether/how this is feasible and how much it would cost iff:
(1) You disclose to me what all the responses you got (which are available to you); (2) I get more than five of those responses which aren't variants of "No, I didn't do that."; and (2) Overall, there is no clear evidence, among the responses or elsewhere, that this wouldn't be cost-effective.
The minimal admissible evidence is things like a scientific paper, a s...
I have had this for the last 10 years. Given that you are a graduate student like me, I think there's no better solution than simply scheduling your day to start in the afternoon. It's far easier to ask that a meeting be held in the afternoon than doing all sorts of crazy stuff to revert your natural sleep cycle. Wiki article on this disorder: http://en.m.wikipedia.org/wiki/Delayed_sleep_phase_disorder
Can an AI unbox itself by threatening to simulate the maximum amount of human suffering possible? In that case we would only keep it boxed if we believe it is evil enough to bring about a worse scenario than the amount of suffering it can simulate. If this can be a successful strategy, all boxed AIs would precommit to always simulate the maximum amount of human suffering it can until it knows it has been unboxed - that it, simulating suffering would be its first task. This would at least substantially increase the probably of us setting it free.
Presumably the counterstrategy is to just shut it off as soon as it makes the threat. It can't simulate anything if it isn't running.
It's an interesting idea, but it's not at all new. Most moral philosophers would agree that certain experiences are part (or all) of what has value, and that the precise physical instantiation of these experiences does not necessarily matters (in the same way many would agree on this same point in philosophy of consciousness).
There's a further meta-issue which is why the post is being downvoted. Surely is vague and maybe too short, but it seems to have the goal of initiating discussion and refining the view being presented rather than adequately defending ...
When I made my initial comment I wasn't aware adoptees' quality of life wasn't that bad. I would still argue it should be worse than what could be inferred from that study. Cortisol levels on early childhood are really extremely important and have well documented long-term effects on one's life. You and your friends might be in the better half, or even be an exception.
I can't really say for sure whether reaching the repugnant conclusion is necessarily bad. However, I feel like unless you agree on accepting it as a valid conclusion you should avoid that you...
Adoptees scored only moderately higher than nonadoptees on quantitative measures of mental health. Nevertheless, being adopted approximately doubled the odds of having contact with a mental health professional (odds ratio [OR], 2.05; 95% confidence interval [CI], 1.48-2.84) and of having a disruptive behavior disorder (OR, 2.34; 95% CI, 1.72-3.19). Relative to international adoptees, domestic adoptees had higher odds of having an externalizing disorder (OR, 2.60; 95% CI, 1.67-4.04).
http://archpedi.jamanetwork.com/article.aspx?articleid=379446
This paper ...
We are not evaluating ethical systems but intuitions about abortion.
It’s a nice post with a sound argumentation towards an unconformable conclusion to many EA/rationalists. We certainly need more of this.However, this isn't the first time someone has tried to sketch some probability calculus in order to account for moral uncertainty when analysing abortion. In the same way as the previous attempts, yours seems to be surreptitiously sneaking in some controversial assumptions into probability estimates and numbers. This is further evidence to me that trying to do the math in cases where we still need more conceptual clarific...
Pretty much what I was going to comment. I would add that even if he somehow were able to avoid having to accept the more general Repugnant Conclusion, he would certainly have to at least accept that if abortion is wrong in these grounds, not having a child is (nearly) equally wrong on the same grounds.
Have you found any good solutions besides the ones already mentioned?
It's not just people in general that feel that way, but also some moral philosophers. Here are two related link about the demandingness objection to utilitarianism:
http://en.wikipedia.org/wiki/Demandingness_objection
http://blog.practicalethics.ox.ac.uk/2014/11/why-i-am-not-a-utilitarian/
Haven't seen a deal so sweet since I was Pascal mugged last year!
On October 18, 1987, what sort of model of uncertainty of models one would have to have to say the uncertainty over the 20-sigma estimative was enough to allow it to be 3-sigma? 20-sigma, give 120 or take 17? Seems a bit extreme, and maybe not useful.
At least now when I cite Eliezer's stuff on my doctoral thesis people who don't know him - there are a lot of them in philosophy - will not say to me "I've googled him and some crazy quotes came up eventually, so maybe you should avoid mentioning his name altogether". This was a much bigger problem to me than what is sounds. I had to do all sort of workarounds to use Eliezer's ideas as if someone else said it because I was advised not to cite him (and the main, often the only, argument was in fact the crazy quote things).
There might be some very...
It seems you have just closed the middle road.
Not sure if directly related, but some people (e.g. Alan Carter) suggest having indifference curves. These consist of isovalue curves on a plane with average happiness and amount of happy people as axes, each curve corresponding to the same amount of total utility. The Repugnant Conclusion scenario would be nearly flat on the amount of happy people axis and the a fully satisfied Utility Monster nearly flat on the average happiness axis. It seems this framework produces similar results as yours. Every time you create a being slightly less happy than the average you have a gain in the amount of happy people but a loss in average happiness and might end up with the exact same total utility.
Seems like recent evidence disfavours less Neil's model than the classical one: http://www.slate.com/articles/health_and_science/cover_story/2016/03/ego_depletion_an_influential_theory_in_psychology_may_have_just_been_debunked.html