You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Open thread, September 15-21, 2014

6 Post author: gjm 15 September 2014 12:24PM

If it's worth saying, but not worth its own post (even in Discussion), then it goes here.


Notes for future OT posters:

1. Please add the 'open_thread' tag.

2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)

3. Open Threads should be posted in Discussion, and not Main.

4. Open Threads should start on Monday, and end on Sunday.

Comments (339)

Comment author: FiftyTwo 21 September 2014 10:55:03AM 2 points [-]

What supplements do people take?

I currently take Vitamin D, fish oil, creatine, lithium, iron, multivitamin and melatonin (at bedtime).

Comment author: RichardKennaway 23 September 2014 10:25:32AM 1 point [-]

What supplements do people take?

It would be interesting to also know their reasons, and if they notice positive effects.

I take calcium and vitamin D, prescribed for medical reasons. Nothing else. No real way to tell what the effect is short of DEXA scans, but those are x-rays, so you can't do many of them. I'm not breaking any bones now, but I never did.

Comment author: ChristianKl 23 September 2014 09:11:48AM 1 point [-]

On what basis do you take Iron?

Comment author: FiftyTwo 23 September 2014 10:29:41PM 0 points [-]

There is some decent evidence of correlation between iron deficiency and depression and anxiety, which is an issue I have.

Comment author: ChristianKl 24 September 2014 09:57:08AM *  0 points [-]

I think it would make more sense to get yourself tested for the deficiency. I think the current view is that most people overconsume iron.

Comment author: Viliam_Bur 22 September 2014 07:30:16PM 1 point [-]

Iron.

Comment author: Torello 20 September 2014 10:20:57PM *  1 point [-]

From an article I'm reading:

"For example, the life-time risk for an individual in the United States to develop Crohn’s disease is about 1/1000. How helpful is it for clinicians and patients if that risk shifts to 1/500 or 1/2000?"

It may be hard to tell without the context, but they are suggesting that these revised risk assessments would not be useful. My initial thought is: "If having an estimate is helpful, having a more accurate estimate would be better, and there seems to be a big difference between 1/500 and 1/1000.

Any thoughts?

Full article: https://d396qusza40orc.cloudfront.net/ethicalsocialgenomic/DeflatingTheGenomicBubble.pdf

Comment author: Douglas_Knight 22 September 2014 06:41:36PM *  1 point [-]

There are common diseases you should worry about and rare diseases you shouldn't worry about. A factor of 2 does not move Crohn's from rare to common. The difference between a 70% chance of dying of heart disease and a 30% chance sounds pretty big, but what would you do differently? Either way, it is a big chunk of likely mortality. A factor of 2 is unlikely to change the cost-benefit analysis of actions that might protect you from heart disease. If such an action is useful, it is useful for most people.

Some rare genes do move diseases from rare to common. A broken BRCA (1 in 10k) moves a woman from a 10% chance of dying of breast cancer to an 80% chance of dying of breast cancer, and dying at a young age. Mammograms are valuable for the second woman and not for the first. Some women have prophylactic mastectomies. But if you ask Myriad to test your BRCA, in addition to this useful information, it will also talk about minor variations with useless effects on the risk.

Comment author: Torello 24 September 2014 10:57:22PM 0 points [-]

Thanks for your reply.

Do you think it would be fair to say that for rare diseases (that are not determined by single loci mutations, like Huntington's or BRCA, as you described) it's silly to get a test because a small movement in your risk profile is meaningless in that it wouldn't impact your treatment or behavior in a meaningful way?

Could you explain what you mean by:

Either way, it is a big chunk of likely mortality

Do you work in a related field? You explained this rather concisely, thanks.

Comment author: Douglas_Knight 25 September 2014 12:34:58AM 0 points [-]

Whether they are about rare diseases or common diseases, almost all results that you get out of 23andMe are silly because they don't have rational effects on potential behavior. (They may have irrational effects - if you can use it motivate actions that you ought to be doing anyway, that's great. But there are also bad irrational reactions.)

Depending on your genetics, your chance of dying of heart disease might be as low as 30% or as high as 70%. (I made up those numbers; I suspect the real range evaluable with current genetics is much narrower.) Even the low number, 30% is very high. If you have a 30% of dying of something, you should think about it and react to it. Even in the best case, you still have to think about heart disease.

Comment author: beoShaffer 20 September 2014 04:42:22PM 4 points [-]

I just did a tried to do a Fermi calculation on the value of getting a fire-proof, theft resistant document safe, but can't find a good number for the cost of identity theft. Does anyone have one on hand?

Comment author: RichardKennaway 23 September 2014 10:10:52AM 1 point [-]

I don't, but the cases of identity theft I hear about in the news aren't done by entering someone's home to acquire their papers. What scenarios are you intending to defend against with the safe?

Comment author: beoShaffer 23 September 2014 11:44:39PM 0 points [-]

I'm not sure, basically I hear a lot of vague references to it being good to have such a safe, but can't figure out what if anything it is actually important for.

Comment author: RicardoFonseca 20 September 2014 01:37:46AM 6 points [-]

Hi. I'm Portuguese and live near Lisbon. Are there any LWers out there that live nearby?

The latest survey (2013) shows zero people living in Portugal, and so I feel a bit lonely out here, especially when I read the locations for the LW meetups. They seem so close, only not really...

I guess I could make an effort to start my own meetup in Lisbon or something, maybe, I don't know. I am a little shy and I don't think I am capable of starting something like that on my own.

I work in academia, in the field of computer science, and thus am surrounded by people that would find this website appealing. I have in fact introduced this site (sometimes subtly) to some people I know, but haven't seen anyone taking the time to read the Sequences and get in sync with this community. I want to try harder, though.

What would you say is the most effective way of capturing the interest on this site? My tools are Facebook, and the chance to make a presentation about anything I want at the University and getting an audience of at most 30 people.

Comment author: ChristianKl 23 September 2014 09:26:14AM 3 points [-]

Getting people to read HPMOR is easier than getting them to read the sequences.

Comment author: RichardKennaway 19 September 2014 06:16:25PM 2 points [-]

I've had several unexplained jumps in karma over the last few days, amounting to around 80-100 points. Someone else mentioned the same, and I believe it's happened to quite a few people. If that's a side effect of reverting the votes of systematic downvoters, fine, but if we now have a systematic upvoter, I really don't want to see this. It doesn't have the same emotional overtones as downvotes, but it obscures the signal in the same way.

Comment author: RichardKennaway 20 September 2014 07:11:44AM 2 points [-]

Another possibility is that a new reader, or more than one, is reading through the archives and voting on whatever they feel voteworthy. That's fine as well.

Comment author: shminux 19 September 2014 11:34:11PM 1 point [-]

I recall this being the norm before the dark days of Euginiering.

Comment author: Jiro 19 September 2014 10:37:46PM *  1 point [-]

I got this too, but I was probably the worst recipient of downvotes percentage-wise and the upvotes didn't even make up for the downvotes yet in terms of absolute karma value (let alone in ratio, which would require getting many times the upvotes).

I also noticed that my recent upvotes included a fair number of 2s and 3's and higher numbers, and there were some posts that didn't get voted up at all--in other words, they were distributed in a way I would expect if the upvotes came from multiple people. The downvotes from Eugine were not distributed that way, making it a dead giveaway that they all came from one person.

Comment author: polymathwannabe 19 September 2014 10:17:33PM *  0 points [-]

Now I'm listed among the 15 "Top contributors," which is absolutely not possible.

Comment author: gjm 19 September 2014 07:39:02PM 1 point [-]

I too have had some unexpected karma-jumps lately, and I feel the same way.

Comment author: gjm 20 September 2014 03:28:12AM 1 point [-]

... Aaaand now I just lost about 40 within an hour or two, including downvotes on some obviously unobjectionable comments. Looks like someone's taken a dislike to me. Anyone else had the same?

Comment author: ChristianKl 19 September 2014 06:44:30PM 1 point [-]

If that's a side effect of reverting the votes of systematic downvoters, fine, but if we now have a systematic upvoter, I really don't want to see this.

Given that we just got a new moderator, it might very well be that someone wants to test out the response about what happens when he goes and votes up systematically.

I personally also got similar jumps in my karma.

Comment author: ChristianKl 18 September 2014 09:43:49AM 2 points [-]

When speaking about battling ISIS, the alternatives for the West seems to be either air strikes or boots on the ground. Boots on the ground means actual personal. Why isn't there a version of boots on the ground that's completely robot based? Why are human bodies still needed for waging intercity warfare?

Comment author: TylerJay 18 September 2014 07:33:48PM 6 points [-]

Considering that this is the state-of-the-art in animal-like robot movement, I can see why we still use meat-soldiers.

Comment author: NancyLebovitz 18 September 2014 10:17:24AM 1 point [-]

Because warfare is complicated? Are you talking about drone robots?

Comment author: ChristianKl 18 September 2014 10:30:12AM 2 points [-]

The word drone refer to something that flies. You could miss flying and non-flying robots.

What's the bottleneck, where robots don't perform?

Comment author: hyporational 19 September 2014 02:02:45AM 1 point [-]
Comment author: bramflakes 18 September 2014 03:53:15PM 9 points [-]

What's the bottleneck, where robots don't perform?

  • Rough terrain
  • Adverse weather conditions
  • Dealing with civilians
  • Going up and down flights of stairs
  • Taking prisoners
  • Medical care
  • Being underground

probably a lot more

Comment author: Lumifer 18 September 2014 04:19:59PM 9 points [-]

Prolonged functioning at high energy levels far from usable energy sources.

Comment author: ChristianKl 18 September 2014 04:11:50PM 1 point [-]

To what extend are those issue likely to be resolved in 10 to 20 years to an extend that they change the geopolitical situation?

Comment author: Lumifer 18 September 2014 04:31:38PM 8 points [-]

Not very likely. In 10-20 years we might get a self-driving car which is a MUCH easier problem than a battlefield robot.

Comment author: ChristianKl 18 September 2014 09:54:33PM 2 points [-]

Google already has self-driving cars. The issue is more about making them safe enough that they don't get sued to the ground when the cars get into accidents. Additionally you need to pass laws that make them legal.

Military technology doesn't suffer from the same hurdle.

Comment author: Azathoth123 19 September 2014 03:29:32AM 6 points [-]

On the other hand, they have to drive through terrain that has been intentionally modified to be difficult for their algorithms.

Comment author: Lumifer 19 September 2014 02:07:08AM 12 points [-]

Google already has self-driving cars

Kinda sorta maybe not really.

Comment author: NancyLebovitz 22 September 2014 12:25:37PM 5 points [-]

Dammit, I've got to pay more attention to those feelings of "really?" Driverless cars at current levels of tech seemed faintly implausible, but I ignored that in favor of "I keep hearing it in the news" and "google=magic".

On the other hand, self-driving cars might make sense for slow-moving traffic jams.

Comment author: Furcas 22 September 2014 05:27:27AM 2 points [-]

Huh, looks like I've been fooled by journalists again. Thanks!

Comment author: CronoDAS 18 September 2014 11:03:33AM *  2 points [-]

I'd guess that communications are a problem - you'd need more bandwidth to send enough video back to drive a car remotely than to fly a plane, and it's probably easier to lose contact, too. Not to mention the difficulties of fighting inside a city you don't want to simply destroy: can your robot open a door and go up a flight of stairs?

This is the kind of thing that's being researched by the dreaded Military-Industrial Complex, though.

Comment author: RichardKennaway 18 September 2014 12:08:56PM 2 points [-]

This is the kind of thing that's being researched by the dreaded Military-Industrial Complex, though.

This is where they've got to (scroll down to the archive link). It isn't yet anywhere near good enough for the task.

For remote rather than automonous operation, there would be major humanitarian applications as well, but the technical problems are still huge. There's latency and reliability of communications, terrain that would be challenging even for people on the spot, dexterity in confined spaces, and the problem of refuelling. None of this is a Simple Matter Of Engineering.

Comment author: ahbwramc 18 September 2014 05:15:18AM 3 points [-]

I've never been entirely sure about the whole "it should all add up to normality" thing in regards to MWI. Like, in particular, I worry about the notion of intrusive thoughts. A good 30% of the time I ride the subway I have some sort of weak intrusive thought about jumping in front of the train (I hope it goes without saying that I am very much not suicidal). And since accepting MWI as being reasonably likely to be true, I've worried that just having these intrusive thoughts might increase the measure of those worlds where the intrusive thoughts become reality. And then I worry that having that thought will even further increase the measure of such worlds. And then I worry...well, then it usually tapers off, because I'm pretty good at controlling runaway thought processes. But my point is...I didn't have these kinds of thoughts before I learned about MWI, and that sort of seems like a real difference. How does it all add up to normality, exactly?

Comment author: CellBioGuy 20 September 2014 10:40:39PM 2 points [-]

You don't see other people doing so, and I can assure you many more people than jump have such thoughts. Any MWI weirdness would only affect what you recall of your OWN actions in this case.

Comment author: gjm 18 September 2014 07:54:01AM 9 points [-]

Whatever argument you have in mind about "the measure of those worlds" will go through just the same if you replace it with "the probability of the world being that way". You should be exactly equally concerned with or without MWI.

The question that actually matters to you should be something like: Are people with such intrusive thoughts who aren't generally suicidal more likely to jump in front of trains? I think I remember reading that the answer is no; if it turns out to be yes (or if you find those thoughts disturbing) then you might want to look into CBT or something; but MWI doesn't have anything to do with it except that maybe something about it bothers you psychologically.

Comment author: ahbwramc 21 September 2014 03:35:23PM 0 points [-]

Okay, fair enough, forget the whole increasing of measure thing for now. There's still the fact that every time I go to the subway, there's a world where I jump in front of it. That for sure happens. I'm obviously not suggesting anything dumb like avoiding subways, that's not my point at all. It's just...that doesn't seem very "normal" to me, somehow. MWI gives this weird new weight to all counterfactuals that seems like it makes an actual difference (not in terms of any actual predictions, but psychologically - and psychology is all we're talking about when assessing "normality"). Probably though this is all still betraying my lack of understanding of measure - worlds where I jump in front of the train are incredibly low measure, and so they get way less magical reality fluid, I should care about them less, etc. I still can't really grok that though - to me and my naive branch-counting brain, the salient fact is that the world exists at all, not that it has low probability.

Comment author: gjm 21 September 2014 07:01:53PM 3 points [-]

and psychology is all we're talking about when assessing "normality"

I have always taken "it all adds up to normality" to mean not "you should expect everything to feel normal" but "actually, when you work out the physics, all this counterintuitive weird-feeling stuff produces the world you're already used to, and if it feels weird then you should try to adjust your intuitions if possible".

I'm not sure there's much I can say to help -- it's clear from your comments that you understand in theory what's going on, and it's just that your "naive branch-counting brain" is naive and cares about the wrong things :-).

Maybe this will help: Suppose you're visiting a big city. Consider the following two propositions. (1) There is one person in this city who would cheerfully knock you on the head and steal your wallet. (2) Half the people in this city would cheerfully knock you on the head and steal your wallet. I don't know about you, but I would be really scared to learn #2 and totally unsurprised and unmoved by #1. Similarly: "there are branches in which you jump in front of the train" -- well, sure there are, and there are branches where I abruptly decide to declare myself Emperor of the World and get taken off to a mental hospital, and branches where the earth is about to get hit by an asteroid that miraculously got missed by everyone's observations and we all die. But there aren't "a lot" of any of these sorts of branch (i.e., the measure is very small). What would worry me is to find that a substantial fraction of branches (reckoned by measure) have me jumping in front of the train. But what it takes to make that true is exactly the same thing as it takes to make it true that "with high probability, you will jump in front of the train".

Comment author: DataPacRat 17 September 2014 06:32:07PM *  4 points [-]

A little help communicating some ideas?

Anyone up for beta reading a 2,000 word section of my attempt at an aspiring-rationalist story, S.I.?

I've just finished putting together an initial draft of Bunny pontificating about the ideas discussed in https://www.reddit.com/r/rational/comments/2g09xh/bstqrsthsf_factchecking_some_quantum_math/ . I could really use some feedback to make sure I'm having her explain them in a way that's actually comprehensible to the reader. Anyone who'd like to help me with this, I've pasted the initial draft to a GoogleDoc at https://docs.google.com/document/d/1lOQAAM3fdnF2ew7CgBQqSLtk21_4Foa8n7IB3Ay0ze8/edit?usp=sharing , which is set to allow comments.

Comment author: zedzed 19 September 2014 09:01:53AM 2 points [-]

Are you operating under Crocker's rules?

Also, if you want strictly writing help, I was recently made aware of the existence of /r/destructivereaders (h/t Punoxysm).

Comment author: DataPacRat 19 September 2014 09:08:54AM 2 points [-]

I've claimed to operate under Crocker's rules for some time now - though this might be the first time anyone has invoked that.

I'll take a look at that subreddit; but at the moment, I'm mainly concerned with figuring out how to best communicate the new ideas presented in the draft (the Lottery Oracle, etc) to the reader (keeping in mind that said reader is going to have gone through roughly 120,000 words of my attempt at rationalist fiction to get this far), rather than any particular grammar or style details that don't affect that goal.

Comment author: DataPacRat 19 September 2014 09:48:48AM 4 points [-]

I'm afraid that /r/destructivereaders doesn't look like a good place for me. They're set up so they (just about) require submitters to have previously critiqued multiple other submissions, and I'm already trying to come up with clever ways to make sure I spend so much time each day working on my story, instead of being distracted by all the shiny things on the internet.

From what I've looked at so far, it looks like they tend to focus on the basics. I already know that I over-use semicolons, and use sentences that are too long and complicated (and contain multiply nested subclauses (like these)) for many readers' comfort, and that I've been skimping on descriptions which aren't directly plot-relevant. I don't anticipate that being told these facts yet again would be worth the time I'd spend critiquing other posts for my entry fee.

Comment author: ChristianKl 20 September 2014 05:46:50PM 0 points [-]

I already know that I over-use semicolons, and use sentences that are too long and complicated (and contain multiply nested subclauses (like these)) for many readers' comfort

If you know, why don't you solve the issue? Rereading a post and asking yourself for every sentence: "Can I break this down into multiple sentences?" and "Does this sentence really need that many words?" isn't complicated.

I don't anticipate that being told these facts yet again would be worth the time I'd spend critiquing other posts for my entry fee.

Identifying common errors in the writing of other people is good training to then identify the same errors in your own writing.

Comment author: DataPacRat 20 September 2014 07:12:19PM 0 points [-]

If you know, why don't you solve the issue?

I also know that I should arrange my diet to have more fruits and vegetables than it currently does. However, after a few decades of doing things one way which isn't optimal, but is good enough to get the job done, simply "deciding" to do things a different way isn't enough to alter all the various unconscious mental sub-units whose interactions lead to the behaviour in question.

I bought a pumpkin pie yesterday instead of a box of cookies. I've started using beta-readers. Neither one is a perfect solution - but each is a single-step improvement over the previous situation, a step that is within the range of behaviours I can get my unconscious processes to actually accomplish, and hopefully, isn't going to be the only step.

Comment author: ChristianKl 20 September 2014 07:23:34PM 0 points [-]

While I won't claim that my writing is very good I think I learned to handle the issue of writing sentences that are too long. It just takes a decision to allocate some time to reviewing your own writing and then looking at every sentence.

I don't do that for everyone of my LW posts, but if I would write a blog or fiction I would.

Comment author: polymathwannabe 19 September 2014 01:55:09AM 0 points [-]

Done. See comments by name Carlos.

Comment author: DataPacRat 19 September 2014 09:50:13AM 1 point [-]

Thank you kindly. :)

I've accepted most of your suggestions, and responded to the ones that I haven't.

Comment author: [deleted] 17 September 2014 05:40:53PM *  -1 points [-]

I should probably update this prediction. Considering Yudkowsky's recent pwnedness and pieces like this becoming common it is at least 10%.

Moldbug in 2008.

There is a very easy resolution to this problem: adopt the principle that no person is illegal. This rule is perfectly consistent with "applied Christianity." It is taught at all our great universities. It is implied every time a journalist deploys the euphemism "undocumented." And I'm sure there are dozens of ways in which it could be incorporated into our great Living Constitution. There is only one problem: the people are not quite ready for it.

But perhaps in thirty years they will be. Perhaps? I would bet money on it. And I would also bet that, by the time this principle is established, denying it will be the equivalent of racism. Us old fogeys who were born in the 1970s will be convulsed with guilt and shame at the thought that the US actually considered it ethically acceptable to turn away, deport, and otherwise penalize our fellow human beings, on the ridiculous and irrelevant grounds that they were born somewhere else.

So the Cathedral wins coming and going. Today, it does not suffer the political backlash that would be sure to ensue if the Inner Party endorsed opening the borders to... everyone. Still less if it actually did so. (Unless it let the new Americans vote as soon as they set foot on our sacred soil, which of course would be the most Christian approach.) And in 2038, having increased North America's population to approximately two billion persons, none of them illegal, and all living in the same Third World conditions which it has already inflicted on most of the planet, our blessed Cathedral will have the privilege of berating the past with its guilt for not having recognized the obvious truth that no person is illegal. Ain't it beautiful?

Comment author: [deleted] 22 September 2014 08:16:37AM 5 points [-]

Yudkowsky's recent pwnedness

I'm not under the impression he was much less pwned in 2007. Are you thinking of something in particular?

Comment author: Sarunas 21 September 2014 04:17:29PM 2 points [-]

Curiously, the anti-immigration movement in the US would be very different than those in Western Europe, and, I would guess, significantly weaker. While economic arguments are somewhat similar in both countries (although e.g. studies differ whether (and what level of) immigration actually increases unemployment levels as in the short run immigrants seem to go to the countries where the unemployment is decreasing), in Western Europe immigrants are vastly overrepresented in crime statistics compared to local population, which is not the case in the US. Nor (it was my impression, I am not from the US) immigrants are often thought as a demographic group whose individuals are the most prone to commit crimes (it must be noted, however, that immigrants aren't a homogeneous group (both in Western Europe and the US) and their effect (and/or the perception thereof) on a country of destination might differ), and as they are not at the top of crime statistics, they are less likely to be a target of blame. Furthermore, it was my impression that due to the long history of immigration, the national identity of the US is not based on ethnicity, while in most European countries it definitely is. As you can see, it is not surprising that Americans are somewhat more likely than Europeans to think that immigration should be increased, and about two thirds of them think that on the whole immigration is a good thing, only in Sweden we find similar support for it.

Nevertheless, even given this unusually positive attitude towards immigration, I would guess that the US population reaching even 600 million (let alone 2 billion, which I believe must have been a hyperbole) by 2038 (U.S. Census Bureau projects approximately 400 millions) has a probability less than 1-2 percent. The reason is that while for a pundit it is often a good strategy to make bold claims about e.g. opening all borders, as it gains him attention (therefore I can believe that we will hear a lot of such claims from people who compete for attention) whether or not the claims about doubling World's GDP are correct (in the long run it may be correct, I do not know), it might not be such a good thing for a politician or a civil servant to do bold actions as it is a risk of losing control of the situation and/or getting fired, and, as a general rule, politicians and civil servants want neither of those, therefore , I would guess, they have to more cautious when they become actual decision makers. Therefore, even if one wants to promote the idea of open borders and free mobility, one could probably try to encourage more Schengen/EU style regional agreements and then gradually "merging" them with new bilateral agreements between those unions, as it seems less risky than simply welcoming all immigrants. And in the far future it may happen that large parts of the world is covered by a Schengen-like agreement making "nobody illegal" in a similar sense that it is relatively easy for a person from one EU country to work in another. But that would probably take much more than 25-30 years. However, I would guess that even then it seems highly unlikely that US population would exceed even 1 billion, let alone 2 billion, since as the economies of the developing countries improve, there will be less incentives for people to leave them for the US. One would expect a huge influx of immigrants only if US government loosens immigration restrictions "faster" than developing countries manage to improve.

Comment author: [deleted] 22 September 2014 07:06:26AM *  3 points [-]

Good counter-argument, updated

Though I will point out civil servants in the position to decide such things are practically unfierable and that politicians' public persona are down-stream from public opinion, if the media and academia that are mostly upstream decide open borders really is a moral crime akin to segregation (not hard since it fundamentally is segregation - not that I think this in itself makes it immoral), then public opinion would try to resist by a few populist politicians but would eventually succumb like it has on all other issues where its interests or opinions were pitted against the former.

Comment author: shminux 17 September 2014 09:04:04PM 8 points [-]

Is this post just an excuse for an NRxic quote? And what does the "Yudkowsky's recent pwnedness" swipe refer to?

Comment author: jaime2000 20 September 2014 08:22:59PM *  1 point [-]

And what does the "Yudkowsky's recent pwnedness" swipe refer to?

For what it's worth, when I asked Nyan Sandwich and Nydwracu whether they thought Eliezer Yudkowsky was pwned, they cited his support for open borders and his evangelical polyamory as evidence that he was, indeed, pwned.

Comment author: [deleted] 19 September 2014 08:34:37PM *  2 points [-]

Is this post just an excuse for an NRxic quote?

No. I honestly think the probability of this prediction coming true has increased.

Comment author: ChristianKl 20 September 2014 03:10:24PM 3 points [-]

If that's the case why don't you simple update your predictionbook entry?

Comment author: [deleted] 22 September 2014 07:05:32AM *  4 points [-]

Trivial inconvenience of forgetting my password, also I wanted to talk about it with other people.

Edit: Predictionbook entry updated.

Comment author: ChristianKl 22 September 2014 10:45:28AM 2 points [-]

Trivial inconvenience of forgetting my password

Software like KeePass really helps for that purpose.

Comment author: ChristianKl 18 September 2014 11:27:38PM 3 points [-]

And what does the "Yudkowsky's recent pwnedness" swipe refer to?

Probably the idea of relocating to a place outside of the US where it's easier to get visas.

Comment author: Nornagest 17 September 2014 06:57:12PM *  6 points [-]

I'll bet a thousand 2014 dollars at even odds that it's less than a quarter of that.

(Clarification: I'm talking about the population of the US, as per the PredictionBook entry, not of North America as per Moldbug's quote; the population of the latter is already over 500 million. Should probably stipulate present borders too, just in case. 2 billion isn't credible in any case, though.)

Comment author: knb 17 September 2014 06:56:47PM 7 points [-]

Two billion is a crazy population prediction even if open borders was enacted. Relative quality of life would decline very quickly with open borders, and the immigration level would slow down dramatically.

I also read some estimate that "only" 500 Million people world wide want to immigrate to the US. Overall I expect the quality-of-life gap between USA and the 3rd world to continue declining over the next couple decades.

Comment author: Lumifer 17 September 2014 06:42:50PM 5 points [-]

As a counterpoint look at the rise of anti-immigration movements in Europe and e.g. the success of Marine Le Pen in France.

Comment author: NancyLebovitz 17 September 2014 06:26:05PM 3 points [-]

I'm dubious about borders getting opened that soon, considering how long it's taking to make moderate moves towards drug legalization.

Comment author: bbleeker 17 September 2014 04:58:49PM *  4 points [-]

Donation sent. !@#% those !@#&!.
EDIT: Oops, wrong place, this was supposed to go under ITakeBets' post.

Comment author: Kaj_Sotala 17 September 2014 11:51:35AM *  7 points [-]

In the interest of trying out stuff outside the usual sphere-of-things-that-I'm-doing, I now have a fashion/lifestyle blog.

It's in Finnish, but it has a bunch of pictures of me, which ought to be language-neutral. Also my stuffed animals. (And yes, I know that I need a better camera.)

Comment author: Evan_Gaensbauer 17 September 2014 08:10:24AM *  1 point [-]

A friend and I hope to host a MIRIxVancouver workshop in Vancouver, Canada sometime in October. We haven't filed an application with MIRI yet, and we haven't set a date, so there's no schedule yet. So, this is just a shout-out to anyone who might want to get involved in it over a weekend, including if you want to visit from Seattle, or Oregon, or anywhere nearby. Comment below, or send me a PM, if you're interested in attending.

Comment author: Evan_Gaensbauer 17 September 2014 08:13:22AM 1 point [-]

Vancouver has enough of a diversity of people interested in the MIRI who can host their own friends that I believe it will make sense to host multiple different MIRIx workshops. Like, I find the MIRI very interesting, but I want to grasp technically what it's about. Another one of my friends is interested in the philosophy of A.I., and yet another friend is a former MIRI intern who will invite a bunch of his friends from the university math department over. So, I'll likely host workshops at different levels of depth, or with different topics.

Comment author: ike 17 September 2014 04:53:51AM *  1 point [-]

I was thinking about anthropics after seeing some posts here about it. I read the series of posts on ADT including http://lesswrong.com/r/discussion/lw/8aw/anthropic_decision_theory_iv_solving_selfish_and/, and EY's posts http://lesswrong.com/lw/17c/outlawing_anthropics_an_updateless_dilemma/, http://lesswrong.com/lw/19d/the_anthropic_trilemma/, and http://lesswrong.com/lw/17d/forcing_anthropics_boltzmann_brains/. I had a few questions about those posts.

First, how is average utilitarian defined in a non-circular way? I'm trying to wrap my head around why I don't agree with the conclusions of the first post I linked, and it seems to come down to not understanding average utilitarians.

More specifically, do they define two levels of utility? Or do they exclude themselves from the calculation? I thought it was just a different way of allocating your own utility, but how do you calculate which way will give you the most utility by giving the world a greater average utility, without knowing the answer of your own utility to plug in?

Second, in http://lesswrong.com/lw/19d/the_anthropic_trilemma/ EY ended off with

I will be extremely impressed if Less Wrong solves this one.

Has he been officially "impressed" yet? Should I read any specific attempts to solve the trilemma? What reading can I do on anthropics to get an idea of the major ideas in the field?

It seems to me that SIA, in the way it's been applied, is obviously correct, and in general I feel like I have very clear intuitions on these kind of problems. I plan on writing up something eventually, after I understand the argument against my point-of-view to argue coherently.

Comment author: KnaveOfAllTrades 17 September 2014 05:39:40AM *  4 points [-]

First, how is average utilitarian defined in a non-circular way?

If you can quantify a proto-utility across some set of moral patients (i.e. some thing that is measurable for each thing/person we care about), then you can then call your utility the average of proto-utility over moral patients. For example, you could define your set of moral patients to be the set of humans, and each human's proto-utility to be the amount of money they have, then average by summing the money and dividing by the number of humans.

I don't necessarily endorse that approach, of course.

Has he been officially "impressed" yet?

I think Eliezer says he's still confused about anthropics.

What reading can I do on anthropics to get an idea of the major ideas in the field?

So far as I know, Nick Bostrom's book is the orthodox foremost work in the field. You can read it immediately for free here. Personally, I would guess that absorbing UDT and updateless thinking is the best marginal thing you can do to make progress on anthropics, but that's probably not even a majority opinion on LW, let alone among anthropics scholars.

Comment author: Metus 17 September 2014 01:58:05AM 11 points [-]

According to the efficient market hypothesis index funds should be the best way for the average person to gain a return from investment. Now there is a plethora of indices to invest in. How should one find the 'best' one?

Further, only a relatively small part of return generating assets are captured in publically tradeable assets. What about private equity and real estate, huge parts of the economy?

Comment author: hyporational 17 September 2014 12:57:26PM *  2 points [-]

IIRC real estate prices in the US rise about 1% per year inflation adjusted while stock markets rise about 7 % on average. An average person needs a huge loan to invest in real estate and go all in which means zero spread of risk. Real estate is also relatively illiquid not only because of practical reasons but because the return of investment depends on timing of the transaction. You're shit out of luck if you need money while the price of your house is plummeting.

How should one find the 'best' one?

Depends on your risk tolerance. The bigger the index, the lower the risk and the lower the possible returns, generally. Also bigger index funds are usually more liquid. Transaction costs matter quite a lot unless you have a big lump sum to invest, and even then you should consider dollar cost averaging.

Comment author: niceguyanon 19 September 2014 06:59:52PM 1 point [-]

IIRC real estate prices in the US rise about 1% per year inflation adjusted...

There is also real estate taxes just for holding the asset and upkeep expenses too! But to be fair, asset appreciation isn't the only return on real estate, many investment properties are income producing assets. But then again you can just get that exposure from REITS anyway.

Comment author: ESRogs 17 September 2014 10:33:46PM 8 points [-]

An average person needs a huge loan to invest in real estate

That's not true. It's easy to get exposure to real estate through REITs. For example, through my wealthfront.com portfolio, I'm invested in Vanguard's US REIT ETF, VNQ.

Comment author: hyporational 18 September 2014 03:53:13AM 5 points [-]

I stand corrected.

Comment author: roystgnr 17 September 2014 03:50:13PM 6 points [-]

IIRC real estate prices in the US rise about 1% per year inflation adjusted while stock markets rise about 7 % on average.

YRC. I thought you were forgetting to adjust the stock market returns for inflation, so I went to hunt for more accurate numbers, but apparently 1950-2009 S&P500 inflation-adjusted returns (counting not just price rise, but dividends) averaged to 7% per year.

Comment author: hyporational 17 September 2014 06:08:45PM 2 points [-]

Thanks. If you care about transaction costs you should probably invest in funds that reinvest dividends automatically.

Comment author: RowanE 17 September 2014 12:01:20PM 7 points [-]

Funds take a fraction of the earnings out, as management fees, and you want the fund that charges the lowest such fees. The early retirement blogs I read seem to agree on Vanguard being the best choice, at least in the US.

Comment author: Ixiel 17 September 2014 11:25:57AM 1 point [-]

I an efficient market the expected value wouldn't be all that different between options, so base it on your risk management preferences.

Comment author: ITakeBets 17 September 2014 12:18:28AM 35 points [-]

I'm posting here on behalf of Brent Dill, known here and elsewhere as ialdabaoth-- you may have enjoyed some of his posts. If you read the comments at SSC, you'll recognize him as a contributor of rare honesty and insight. If you'd had the chance to talk with him as much as I have, you'd know he's an awesome guy: clever, resourceful, incisive and deeply moral. Many of you see him as admirable, most as relatable, some as a friend, and more, I hope, as a member of our community.

He could use some help.

Until last Thursday he was gainfully employed as a web developer for a community college in Idaho. Recently, he voluntarily mentioned to his boss that he was concerned that seasonal affective disorder was harming his job performance, who mentioned it to his boss, who suggested in all good faith that Brent should talk to HR to see if they might help through their Employee Assistance Program. In Brent's words: "Instead, HR asked me a lot of pointed questions about when my performance could turn around and whether I wanted to work there, demanded that I come up with all the solutions (after I admitted that I was already out of brainpower and feeling intimidated), and then directed me to turn in my keys and go home, and that HR would call me on Monday to tell me the status of my employment." Now, at the end of the day Tuesday, they still haven't let him know what's happening, but it doesn't look good.

I think we can agree that this is some of the worst horseshit.

On the other hand, he's been wanting to get out of Idaho and into a city with an active rationalist community for a while, so in a sense this is an opportunity. Ways to help: Brent needs, in order of priority: a job, a place to stay, and funds to cover living and moving expenses-- details below. Signal boosts and messages of support are also helpful and appreciated. Ways NOT to help: Patronizing advice/other-optimizing (useful information is of course welcome), variations on 'cool story bro' (the facts here have been corroborated to my satisfaction with hard-to-fake evidence), disrespect in general.

1. Job: Leads and connections would help more than anything else. He's looking to end up, again, in a good-sized city with an active rationalist community. Candidates include the Bay Area, New York, Boston, Columbus, San Diego, maybe DC or Ann Arbor. He has an excessively complete resume here, but, in short: C#/.NET and SQL developer, also computer game development experience, tabletop board/card game design experience, graphic art and user interface experience, and some team leadership / management experience.

2. Crash space: If you are in one of the above cities, do you have/know of a place for a guy and his cat? How much will it cost, and when will it be available? Probably he'll ultimately want a roommate situation, but if you're willing to put him up for a short time that's also useful information.

3. Funds: Brent is not now in immediate danger of going hungry or homeless, but a couple of months will exhaust his savings, and (although it is hard to know in the current state of things) he has been told that the circumstances constitute "cause" sufficient to keep him from drawing unemployment. Moving will almost certainly cost more than he has on hand. There is a possible future in which he runs out of money stranded in Idaho, which would be not good.

If you feel moved to help, he has set up a gofundme account here. (The goal amount is set at his calculated maximum expenses, but any amount at all would help and be greatly appreciated-- he would have preferred not to set a funding goal at all.) Though Brent has pledged to eventually donate double the amount he raises to Effective Altruist causes, we wouldn't like you to confuse contributing here with charitable giving. Rather, you might want to give in order to show your appreciation for his writing, or to express your solidarity in the struggles and stigma around mental illness, or as a gesture of friendship and community, or just to purchase fuzzies. Also, you can make him do stuff on Youtube, you know, if you want.

Thank you so much for your time and kindness. -Elissa Fleming

Comment author: shminux 19 September 2014 09:34:10PM *  2 points [-]

I also hope someone can help out with writing a better resume, this one is seriously subpar. A single page of achievements based on http://www.kalzumeus.com/2011/10/28/dont-call-yourself-a-programmer/ might be a start: "describe yourself by what you have accomplished for previously employers vis-a-vis increasing revenues or reducing costs".

Comment author: ITakeBets 19 September 2014 11:20:41PM 1 point [-]

Yes, thanks, this has been discussed elsewhere. (That said I'll repeat the request to avoid disrespect or patronizingly phrased advice.)

Comment author: ialdabaoth 17 September 2014 09:37:25PM 14 points [-]

Official update: HR "explored every possible option" but "ultimately we have to move forward with your termination process" after "making certain there was unanimous consensus".

Apparently several people in my now ex-office are upset about this.

Comment author: Decius 17 September 2014 03:49:30AM *  7 points [-]

That narrative is unambiguously a case of illegal discrimination. Idaho law Defines:

"Disability" means a physical or mental condition of a person, whether congenital or acquired, which constitutes a substantial limitation to that person and is demonstrable by medically accepted clinical or laboratory diagnostic techniques. A person with a disability is one who (a) has such a disability, or (b) has a record of such a disability, or (c) is regarded as having such a disability;

and

It shall be a prohibited act to discriminate against a person because of, or on the basis of, disability in [employment]... provided that the prohibition against discrimination because of disability shall not apply if the particular disability, even with a reasonable accommodation, prevents the performance of the work required in that job.

I am also very confused as to how actual HR drones in an actual HR department wouldn't be familiar with the law and able to create a suitable enough pretext for termination.

Comment author: Prismattic 17 September 2014 05:23:18AM 9 points [-]

I already mentioned the A.D.A. to Ialdabaoth, but fighting a discrimination case probably takes more money than he's looking to raise to move, as well as being psychologically exhausting.

Comment author: CronoDAS 18 September 2014 01:08:14AM 1 point [-]

but fighting a discrimination case probably takes more money than he's looking to raise to move

You might be able to get a lawyer to work on a contingency basis - they only get paid if you win.

Comment author: Decius 17 September 2014 05:35:04AM 2 points [-]

Either of those reasons is probably enough to convince a rational person. The spirit of Immanuel Genovese still sits on my shoulder screaming "Passive complicity!" at /me/ every time I contemplate accepting an outcome in which it is normal that this kind of treatment happens.

Comment author: ialdabaoth 17 September 2014 05:40:22PM 8 points [-]

Me too.

The problem is... this is a complex and delicate situation, as all real-life situations are.

There are co-workers who have gone the extra mile to help me and protect me. They didn't do everything they could, because they have families, and they know that if they rock the boat too hard it will be them, not HR, that get thrown overboard.

They aren't rationalists themselves (although I was slowly working on one of them), but they are caring and intelligent people who are themselves struggling to find meaning and stability in a harsh world.

If I could find a way to laser-lance out the demons of stupidity from my workplace, I would do so in an instant. If I could do so in a way that could add net funds to my own cause, I would already be doing so.

But as it is, I know exactly who would suffer for it.

(That doesn't mean that I have committed to a decision yet; I am still weighing necessary evils.)

Comment author: othermaciej 19 September 2014 11:43:01PM 6 points [-]

I hope this is not patronizing advice but rather useful info. To be clear, I am not pressuring you to do anything, I know there are many reasons not to pursue discrimination claims, but I wanted to make sure you are aware of all your options.

The Equal Employment Opportunity Commission is a possibly less costly and less adversarial way of pursuing a discrimination claim. They will investigate independently and try to arrange a settlement if they find discrimination. If settlement is impossible, they may even sue on your behalf. They have won a lot of ADA-related claims. I'm pretty sure they will consult with you for free, so the only initial costs are time and emotional energy.

Comment author: Decius 18 September 2014 01:15:38AM 1 point [-]

I'm letting you know about what my shoulder angel/demon is shouting, because if I follow his advice I am not optimizing for giving you good advice.

Comment author: tadrinth 17 September 2014 03:38:51AM *  14 points [-]

Is Austin on the list? I work at a not-evil tech startup called SchoolAdmin that does school admissions software for a mix of public/private/charter schools. We're not hiring devs right now, but that might possibly change since we have a product manager coming in October. The company is REALLY not evil; we've had three different people come down with mental or physical health issues, and the president's mantra has been 'your job is to get better' in every case.

I could possibly also offer a place to crash, I've got a futon, a study it could be moved to, and already have cats.

Comment author: btrettel 21 September 2014 02:23:26AM 2 points [-]

I would recommend Austin as well. There are loads of developer jobs here, though I don't know any particular place that is hiring right now. We have an active, close-knit rationalist community that I think is pretty fantastic. Worth consideration.

Comment author: juliawise 24 September 2014 11:52:31AM 1 point [-]

I was going to make a plug for Boston, but with SAD, someplace with a sunny winter like Austin sounds like it might be nicer.

Comment author: KnaveOfAllTrades 17 September 2014 01:04:00AM 3 points [-]

Woah, well done everyone who donated so far. I made a small contribution. Moreover, to encourage others and increase the chance the pooled donations reach critical mass, I will top up my donation to 1% of whatever's been donated by others, up to at least $100 total from me. I encourage others to pledge similarly if you're also worrying about making a small donation or worrying the campaign won't reach critical mass.

Comment author: gjm 17 September 2014 08:37:39AM 5 points [-]

If 102 people all pledge to donate 1% of everyone else's total, the consequences could be interesting. (Of course it's vanishingly unlikely. But pedantic donors might choose to word their pledges carefully.)

Comment author: VAuroch 16 September 2014 11:33:04PM 5 points [-]

I remember seeing this organization on LW but cannot find it again or remember the name: it was a for-profit school-like entity that does a short training program (might have been 6 six week, maybe 3 months, that range), which is free upfront and takes their payment entirely as a percentage of the salary from the job they place you in afterward. If I remember correctly, it is run in the Bay Area and took a small pool each session, with a school-like application process.

Can anyone point me to this?

Comment author: Nornagest 16 September 2014 11:37:42PM 6 points [-]

That sounds like App Academy or one of its competitors.

Comment author: VAuroch 16 September 2014 11:51:46PM 2 points [-]

App Academy was the one I was thinking of specifically, thanks.

Comment author: Punoxysm 16 September 2014 08:58:00PM *  0 points [-]

It seems like there are a lot of fan-fiction fans here. Fan-fiction fans, I am curious as to what draws you to the fan-fiction of which you are fans. Is part of it that you're fans of other fan-fiction fans? I guess that depending on the cosplay you could even be a fan of fan-fiction fans' fans.

Comment author: jaime2000 21 September 2014 01:39:18AM *  1 point [-]

I cannot parse your question. Can you rephrase?

Comment author: Nornagest 16 September 2014 09:19:19PM *  8 points [-]

This is a different question, but I've occasionally wondered why some franchises (like Star Trek, Buffy, or Harry Potter) generate a lot of fanfic and others generate much less. Part of this is raw popularity, of course, but quite a bit isn't; the film Avatar (the one with blue aliens, not the one with kung fu) was far more popular than, say, Pirates of the Caribbean, but the latter spawned a thriving fic community and the former has a smattering of stories mostly intended to illustrate critical points.

I don't think there's any single answer, but a franchise's chances seem to be improved if: it's suitable for episodic storytelling (Pirates is a self-contained story, but it's framed like an entry in a serial); it's got strong and ideally archetypal characters (Kirk, McCoy, Spock: action/emotion/reason, easy to write but easy to give depth to); and it's got an open setting with a lot of depth and unexplored bits (few settings outside spec-fic generate a lot of fanfic, and most of those that do are period pieces or procedurals). We're looking at works as toolkits for storytelling, in other words; tight plotting might actually be detrimental.

Comment author: Jiro 19 September 2014 10:48:54PM 2 points [-]

Star Trek is a special case because back when it was created, there weren't a lot of geekish series one could write fanfic about unless you resorted to books. There was no anime fandom, comic books were aimed at much younger people, and non-anthology TV genre fiction with enough merit to gain a fanbase was rare.

Comment author: Jodika 16 September 2014 06:25:48PM 5 points [-]

I have yet to find any thoughts on Effective Altruism that do not assume vast amounts of disposable income on the part of the reader. What I am currently trying to determine are things like 'at what point does it make sense to give away some of your income versus the utility of having decent quality of life yourself and insuring against the risk that you end up consuming charitable resources because something happened and you didn't have an emergency fund'. Does anyone know of any posts or similar that tackle the effective utilitarian use of resources when you don't have a lot of resources to begin with?

Comment author: ChristianKl 18 September 2014 07:46:06AM 2 points [-]

Putting money into an emergency fund here it can gather interest doesn't mean that you can't donate the same money 10 years from now.

Comment author: dspeyer 18 September 2014 04:36:03AM 2 points [-]

I don't have a link, but I suspect cutting this fine is not very valuable. That last $10k would be a lot to you, but that wouldn't make it more than any other $10k to a charity. Instead, ask how you could come to have a vast amount of disposable income. Including whether it makes sense to spend some money toward that end. You may be able to get a very high rate of return investing in yourself.

Comment author: drethelin 16 September 2014 11:38:34PM 2 points [-]

to me EA is more about how to answer the question "how should I be charitable?" than "Should I be charitable and to what extent?"

Comment author: gjm 16 September 2014 10:18:11PM 0 points [-]

Most of the EA stuff I've seen doesn't appear to me to assume vast amounts of disposable income; merely enough to be willing to give some away. Then EA is about what to do with your charity budget, whether it's large or small.

How you prioritize helping others versus helping yourself (and your family, if any) is a more or less orthogonal question.

(I might suggest, snarkily, that for someone who requires "vast amounts of disposable income" before being willing to give any away no term with "altruism" in it is very appropriate. But that wouldn't be fair because e.g. your intention might be to secure yourself a reasonably comfortable life and then give away every penny you can earn beyond that, or something.)

Comment author: Jodika 17 September 2014 01:48:20PM 1 point [-]

That's it, basically; it's about how much of a buffer I'm 'allowed' to give myself on 'reasonably comfortable'; I'm supporting myself and full-time student partner and not in permanent full-time employment so my instinct whenever I have a sniff of an excess is to hoard it against a bad month for getting work rather than do anything charitable with it (or it all goes on things we've put off replacing for monetary reasons, like shoes that are still wearable but worn out enough to no longer be waterproof).

I think Lumifer articulated better than I could what I really wanted to know the answer to, and while there may not be a general answer it does mean that I can at least go looking for things to read now my real question is clearer to me. So thanks!

Comment author: John_Maxwell_IV 19 September 2014 08:31:26AM *  6 points [-]

There are two questions here. The first is how you trade off the value you place on your own welfare vs the value you place on the welfare of distant others. And the second is how having extra cash will benefit your mental health, energy levels, free time, etc. and whether by improving those attributes of yours you'll increase the odds of doing more good for the world in the future.

I consider myself a pretty hardcore EA; I gave $20K to charity last year. But this year I'm saving all my money so my earning-to-give startup will have a bigger cash buffer. And I spend about $100/month on random stuff from Amazon that I think will make my life better (a weighted jump rope for exercising with, an acupressure mat for relaxing more effectively, nootropics, Larry Gonick's cartoon guides to the history of the universe so I can relax & educate myself away from my computer, etc.)

So I guess the point I'm trying to make is you don't even have to deal with the first values question if you decide that investing in yourself is a good investment from a long-run EA perspective. Don't be penny-wise and pound-foolish... your mental energy is limited and if you find wet feet at all stressful, it's worth considering replacing your shoes even before personal welfare gets added in to the equation.

In other words, I personally am more optimistic about you spending all of your money on yourself and spending some of your time and energy on a credible plan for significantly increasing your future EA impact than I am about you donating spare cash to charity and not spending any time and energy and such a credible plan. (In general, I suspect that the potential EA impact of time and energy is underrated; this article gives a good explanation.)

Comment author: Jodika 19 September 2014 09:21:27AM 2 points [-]

Thanks for the link; very helpful and interesting.

Comment author: gjm 17 September 2014 02:54:53PM 1 point [-]

I don't think "allowed" is the right way to think about it (and your quotation marks suggest that probably you don't either). If you mean something like "what position do other reasonable people take?" or "what is the range of options that won't make other people who think of themselves as EAs disapprove of me?", I have no information on anyone else's positions but my own is something like this:

  • If you are having difficulty feeding yourself healthily, paying for somewhere to live that isn't falling down, etc., then feel free not to give anything.
  • Otherwise, I think it's psychologically valuable to keep up the habit of giving something, even if it's very little.
  • If you expect to be substantially better off in the future than you are now, there's a lot to be said for optimizing lifetime income and aiming to give some fraction of that rather than feeling guilty about not giving a lot now. If hoarding some for now gives you more stability later, that's probably better for everyone.
  • Once you're reasonably comfortable financially, I think the traditional figure of 10% is a reasonable benchmark; you can answer the question "10% of what?" in various different ways, all somewhat defensible and leading to substantially varying levels of giving.
  • There are people who give quite a lot more. No one will think ill of you for not being one of them.
Comment author: Jodika 18 September 2014 12:22:38AM 1 point [-]

Thanks gjm, that's a really helpful comment. (And yes, quotation marks indicate 'this is the word I can think of but it is not necessarily the right word.)

I think points number 1 and 3 are especially relevant for me right now, and I have found talking it through on here to be very helpful in defeating an entirely non-useful lingering sense of guilt for not giving more when I really can't afford to, yet.

Comment author: Lumifer 17 September 2014 03:21:05PM 3 points [-]

How do you think saving (in the standard financial sense) and giving should be balanced?

Comment author: gjm 17 September 2014 10:35:36PM 1 point [-]

Um. With careful consideration?

Seriously: I don't have any very strong opinions on this, nor any reason to think that anyone should care what my opinions are. In my own family's case, we save substantially more than we give (very crudely, about 50% of income versus about 10%) but I'm not at all sure that if I thought about it longer and harder I wouldn't conclude that we should be weighting global welfare higher relative to our own financial security.

Comment author: Lumifer 18 September 2014 01:08:04AM 2 points [-]

:-) Don't mean to pick on you, but the impression I get from EAs on LW is that your free cash flow is supposed to go save the world and I got curious about the apparent/potential disconnect from the general meme that people are supposed to save more so as not to be a drag on the society if something happens to them or when they retire...

Comment author: gjm 18 September 2014 01:27:18AM 0 points [-]

The idea that your free cash flow should all go to save the world is generally based on a pretty straightforward utilitarian calculation, and it seems pretty clear that the same calculation would put saving lives in poor countries ahead of the small adverse consequences of drawing more on one's own country's social safety net. So I don't think there's much "disconnect" there.

In practice, very few people are quite so heroically altruistic as to reduce themselves to (what locally passes for) poverty so as to give everything to help the global poor. I bet the few who are are already largely neglecting saving; the rest of us, I think, first decide how much we want to give away and then how we want to balance saving and consumption. So a tradeoff between saving and giving, as such, doesn't arise.

For the avoidance of doubt, i very much don't think of myself as any sort of heroic or expert altruist (effective or otherwise). My only role here is Some Guy Who Got Into A Thread About Effective Altruism :-).

Comment author: Jodika 18 September 2014 12:31:53PM 1 point [-]

Do you have a link? I'm just not sure that it's that obvious that pumping my (hypothetical) money overseas is a utilitarian good if I end up costing my own society more than I give away (which is pretty likely - to use a US example, hypothetical-me might end up costing orders of magnitude more to treat in an emergency room when I get sick because I didn't spend my own money on preventative healthcare).

Obviously the money hypothetical-I save the government isn't automatically going to go to good causes, but by doing my bit to make the society poorer, am I reducing people's overall tendency to have extra money to give away?

I dunno, probably need an economist and a lot of time to properly answer that question...

Comment author: gjm 18 September 2014 01:14:06PM 3 points [-]

Do you have a link?

Nope. Just a lot of handwaving. Sorry.

But, e.g., if you get old and sick and it costs $100k to cure you in the USA, then the utilitarian optimum is probably to let you die and send the money to save 20 or more lives in sub-Saharan Africa. For the avoidance of doubt, I am not suggesting that you should endorse that policy; but if you bite the bullet that says you should send all your spare money to Oxfam or GiveDirectly or whoever, then I think you should probably also be biting the bullet that says you should be prepared to give up and die if you get sick and curing you would be too expensive.

On the other hand, if you're young and get similarly sick, it might (on the same assumptions) be worth curing you so that you can carry on earning money and pumping it to the desperately needy. In which case it might indeed be worth spending some money first to stop that happening. But I'll hazard a guess that the amount you need to spend to make it rather unlikely that you lose a lot of income because of ill health isn't terribly large.

I suggest that unless you're seriously inclined to really heroic charitable giving, you would do better not to worry about such things, and take decent care of yourself as I'm sure you would rather do, and give generously without impoverishing yourself. Especially at present -- if you don't have a lot of money, the difference between heroism and ordinary giving is going to be pretty small. Once you're in a better situation financially, you can reconsider how much of a hero you want to be.

Comment author: Lumifer 18 September 2014 03:14:47AM 2 points [-]

the rest of us, I think, first decide how much we want to give away and then how we want to balance saving and consumption.

I am not sure this is the case. Saving and giving are kinda fungible without any immediate impact on you -- that's got to be tempting...

Comment author: Lumifer 16 September 2014 06:33:02PM 6 points [-]

I don't think there is a general answer to the question "How much should I consume?"

Comment author: Jodika 18 September 2014 12:34:05PM 1 point [-]

Is this a thing we should be asking if someone who is an expert on Effective Altruism and economics and similar could have a go at answering?

Comment author: Lumifer 18 September 2014 03:08:40PM 2 points [-]

You can ask, but why the answer would be anything else other than someone's personal opinion?

It's a straightforward question about personal values. Do you think it's a good idea to have experts in EA or economics tell you what your values should be?

Comment author: Jodika 18 September 2014 05:13:23PM 3 points [-]

It's a straightforward question about personal values. Do you think it's a good idea to have experts in EA or economics tell you what your values should be?

No, but they might know things like the scale of diminishing returns in terms of spending money on yourself, or at what minimum level of wealth do an acceptable majority of people (in x culture or x country) report being satisfied with their lives?

They might have a personal anecdote about how they earn a million dollars a year and live in a ditch and have never been happier, and they might know the psychological reasoning why some people are happy to do that and some people aren't.

I mean, yes, it's true that their answer is not going to be everybody's. But an attempt to answer the question seems very likely to turn up useful information that could help people make their own decisions.

Comment author: NancyLebovitz 16 September 2014 03:36:15PM 8 points [-]

I have a notion that an FAI will be able to create better friends and lovers for you than actual humans could be. Family would be a more complex case if you value the history as well as the current experience.

I'm not talking about catgirls-- if some difficulties in relationships are part of making relationships better in the long haul, then the FAI will supply difficulties.

If people eventually have relationships with FAI-created humans rather than humans generated by other means, is this a problem?

Comment author: gattsuru 19 September 2014 12:39:39AM *  0 points [-]

If people eventually have relationships with FAI-created humans rather than humans generated by other means, is this a problem?

Depends on what the machine has optimized for. I'm not convinced that many definitions of better friends or lovers are vital optimization goals, or even good ones, in themselves. It's quite easy to imagine a set of relationships that trigger every desirable stimuli trigger an individual enjoys, complete with short-term difficulties if necessary, but leaves the victim trapped in a situation where his or her preferences remain at a local optima or are otherwise Not Correct by some grander standard.

Interaction with external minds and external situations not built toward you seem like very important parts of jostling folk out from such environments. Better optimization goals might do that, but it's not an assumption you can easily take.

I'd argue that non-catgirl created beings are people (tautologically), and while relationships with artificially-produced people is fine itself, there are also some possible ethical issues with creating minds optimized for better relationships for certain people, as well, though they're likely outside the scope of this thread (energy efficiency compared to sorting existing minds, harmful desires, House Elves).

Comment author: hyporational 17 September 2014 03:08:00PM 2 points [-]

What's a catgirl?

Comment author: Lumifer 17 September 2014 03:19:28PM 5 points [-]

An indistinguishable-from-live sex toy.

Comment author: Jayson_Virissimo 17 September 2014 07:32:22PM 7 points [-]

With cat-ears.

Comment author: Gunnar_Zarncke 17 September 2014 09:55:25AM *  5 points [-]

See also EYs Failed Utopia #4-2

FAI-created X ... is this a problem?

I'm not sure we can extrapolate this currently. If we knew more, thought faster... maybe.

For me this means that one contraint on FAI is that it may not perform changes arbitrarily fast. Too fast for humans to react and adapt. There must be a 'smooth' trajectory. Surely not the abrupt change suggested in Failed Utopia.

Comment author: blacktrance 16 September 2014 09:42:11PM 1 point [-]

I say it's not a problem, but my views are outside the LW mainstream on this.

Comment author: lmm 16 September 2014 07:08:34PM 3 points [-]

I'm not talking about catgirls-- if some difficulties in relationships are part of making relationships better in the long haul, then the FAI will supply difficulties.

I thought that was already part of catgirls?

Comment author: shminux 16 September 2014 06:22:22PM 4 points [-]

Let's first separate sexual aspects from the need for other companionship. Suppose everyone gets their sexual needs, if any satisfied by catgirls+ (+ for the upgrade which includes relationship problems if necessary). If you have a crush on your coworker (or your sibling, ew!), just add a catgirl copy of them to your harem.

Further suppose that the reproduction aspect is also taken care of.

Now you have a race of essentially asexual humans, as far as human-to-human interactions go.

The question is, does it make sense to have friendbots? What, if anything, is lost when you switch from socializing with meat humans to socializing with simulated ones?

Comment author: Azathoth123 17 September 2014 04:11:19AM 5 points [-]

Suppose everyone gets their sexual needs, if any satisfied by catgirls+ (+ for the upgrade which includes relationship problems if necessary). If you have a crush on your coworker (or your sibling, ew!), just add a catgirl copy of them to your harem.

This strikes me as superstimulating. In particular, the more cat girls you have, the more and kinkier cat girls you want.

Comment author: shminux 17 September 2014 04:22:04AM *  1 point [-]

Not necessarily, Plenty of people are happy with vanilla sex (or without). I suspect that even the kinkiest ones out there also have their limit. If not, let's talk about those who do.

Comment author: Azathoth123 17 September 2014 05:19:42AM 3 points [-]

That's because vanilla sex isn't as stimulating. The more superstimulating something is, the more experiencing it causes you to want more of it.

Comment author: gjm 19 September 2014 02:52:44PM 1 point [-]

vanilla sex isn't as stimulating

For people who are into one or another variety of kink, or would be if only they knew about it / were prepared to try it. I don't think it's obvious that that's everyone.

Comment author: Lumifer 17 September 2014 03:50:48PM *  2 points [-]

The more superstimulating something is, the more experiencing it causes you to want more of it.

That doesn't seem to be the case, see e.g. yummy food.

I think you're confusing "stimulating" and "addictive".

Comment author: shminux 17 September 2014 03:37:53PM 0 points [-]

That "explanation" is easily falsified. There are plenty of people who tried kinkier sex, enjoyed it, but reverted back to vanilla. There are plenty of people who tried roller-coasters once or twice but decided it's too much "stimulation".

Comment author: Azathoth123 19 September 2014 03:35:38AM 4 points [-]

There are plenty of people who tried kinkier sex, enjoyed it, but reverted back to vanilla.

Different people have different thresholds. If I remember the study correctly, none of the rats that tried directly stimulating their pleasure center ever went back.

Comment author: shminux 19 September 2014 05:51:53AM 1 point [-]

Rats != people...

Comment author: Azathoth123 19 September 2014 05:59:41AM 2 points [-]

Yes, well it would be unethical to repeat that experiment with people.

Comment author: Tripitaka 29 September 2014 10:05:01AM -1 points [-]

Well, there have been experiments on humans. http://en.wikipedia.org/wiki/Pleasure_center#Human_experiments

At its most frequent, the patient self-stimulated throughout the day, neglecting personal hygiene and family commitments. >A chronic ulceration developed at the tip of the finger used to adjust the amplitude dial and she frequently tampered with >the device in an effort to increase the stimulation amplitude. At times, she implored her to limit her access to the >stimulator, each time demanding its return after a short hiatus. During the past two years, compulsive use has become >associated with frequent attacks of anxiety, depersonalization, periods of psychogenic polydipsia and virtually complete >inactivity.

Comment author: Jodika 19 September 2014 09:29:30AM 2 points [-]

People, however, (as shminux said) do try kink all the time. It would not be unethical to do a study on people who are already kinky and see if they get kinkier over time.

Anecdotally, they start doing kink, they either decide it isn't for them and stop, or they do get kinkier for a while - because they're exploring what they like and it makes sense to start at the less extreme end of things.

Then they figure out what they like, which is often a range of things at differing levels of 'kinkiness/extremeness', and do that.

I mean, it's almost trivially obvious that compared to the size of the kink community, there is an almost negligible amount of people doing the human equivalent of directly stimulating their pleasure centres to the exclusion of everything else. They tend to make the news. The moderately kinky majority do not.

Comment author: Lumifer 16 September 2014 06:34:30PM 8 points [-]

Let's first separate sexual aspects from the need for other companionship

It's not self-evident to me that they are separable.

Comment author: hyporational 17 September 2014 03:52:37PM 2 points [-]

When my heterosexual male friends tell me companionship isn't about sex I ask them how many male companions they've had. Not many, I've gathered from the silence.

Comment author: EStokes 20 September 2014 02:45:55PM 0 points [-]

Ah, but it's quite likely that they're heteroromantic as well as heterosexual.

Comment author: hyporational 20 September 2014 04:14:29PM 1 point [-]

Perhaps, but why haven't I come across any homoromantic heterosexuals or heteroromantic homosexuals?

Comment author: EStokes 20 September 2014 04:21:00PM *  1 point [-]

AFAIK people with mismatched romantic and sexual orientations, though very much existent, are quite rare and the -romantic terms are most often used by asexual spectrum people to describe their romantic preferences.

Comment author: hyporational 20 September 2014 04:31:17PM 2 points [-]

Asexuals with romantic orientations came across my mind too. I can't imagine romantic and sexual orientations as separate, but the stakes aren't high enough for me to commit the typical mind fallacy so I'll keep my mind open to the possibility :)

Comment author: Lumifer 17 September 2014 04:00:06PM 5 points [-]

how many male companions they've had.

For hetero males the usual term for male companions is "close friends". I bet the great majority have some.

But go ask some hetero women whether they think sex and companionship are well-separable :-/

Comment author: Azathoth123 19 September 2014 03:37:26AM 6 points [-]

Also I get the feeling 21th century Americans have fewer close friends than the historical human norm.

Comment author: Lumifer 19 September 2014 05:33:37AM 2 points [-]

I don't know what the "historical human norm" is and I suspect there is a lot of variation there.

Comment author: Azathoth123 20 September 2014 08:02:41PM 3 points [-]

Try reading literature written before the past 50 years and preferably before the 20th century. That will give you an idea.

Comment author: Lumifer 21 September 2014 12:43:32AM 3 points [-]

Try reading literature written before the past 50 years and preferably before the 20th century.

I am afraid Victorian England is not all that representative of the historical human norm.

Comment author: Azathoth123 23 September 2014 03:22:47AM 4 points [-]

I wasn't primarily thinking of Victorian England. Also "before the 20th century" isn't just the 19th century.

Comment author: hyporational 17 September 2014 04:07:33PM *  2 points [-]

In Finnish the connotations of "companion" are more obviously sexual I see, at least in my circles.

Comment author: Lumifer 17 September 2014 04:16:24PM 3 points [-]

It's probably a language issue, in standard English the word "companion" has no sexual overtones.

More to the point, this subthread is explicitly about separating sex from companionship.

Comment author: cousin_it 16 September 2014 06:21:05PM *  4 points [-]

You've asked that before.

I don't have any new thoughts on this question, so I'll just quote my answer from there:

Yeah. People need to be needed, but if FAI can satisfy all other needs, then it fails to satisfy that one. Maybe FAI will uplift people and disappear, or do something more creative.

Comment author: Lumifer 16 September 2014 03:38:20PM 2 points [-]

If people eventually have relationships with FAI-created humans rather than humans generated by other means, is this a problem?

This looks to be wireheading lite and if you got there I don't see why you wouldn't make the next step as well -- the FAI will create the entire world for you to enjoy inside your head.

Comment author: NancyLebovitz 16 September 2014 03:41:35PM 2 points [-]

I thought wireheading meant stable high pleasure without content rather than an enjoyable simulated world. What do other people think wireheading means?

Comment author: Lumifer 16 September 2014 03:49:45PM 1 point [-]

Well, technically the term "wireheading" comes from experiments which involved inserting an electrode (a "wire") into a rat's pleasure center and giving the rat a pedal to apply electric current to this wire. So yes, in the narrow sense wireheading is just the direct stimulation of the pleasure center.

However I use "wireheading" in the wide sense as well and there it means, essentially, the focus on deriving pleasure from externally caused but internal experiences and the lack of interest in or concern with the outside world. Wireheading in the wide sense is, basically, purified addiction.

Comment author: NancyLebovitz 16 September 2014 04:28:35PM 2 points [-]

If we're living inside an FAI, "outside world" might be getting a little vague. This might even be true if we're still living in our DNA-based bodies.

Do you think an FAI would let people have access to anything it isn't at least monitoring, and more likely controlling?

Comment author: Lumifer 16 September 2014 04:32:49PM *  1 point [-]

If we're living inside an FAI

Uploads/ems are a bit of a different case.

Do you think an FAI would let people have access to anything it isn't at least monitoring, and more likely controlling?

I don't know, but in such a case I probably would not consider it a FAI.

Comment author: hyporational 17 September 2014 03:16:09PM *  1 point [-]

Uploads/ems are a bit of a different case.

How? Why does it matter in what substrate the information pattern called you resides in this case? I doubt the meat brain will have any connectibility issues once we have uploads.

Comment author: Lumifer 17 September 2014 03:24:11PM 1 point [-]

Why does it matter in what substrate the information pattern called you resides in this case?

I am not an information pattern having, for example, a considerable somatic component :-D

Comment author: hyporational 17 September 2014 03:30:26PM 1 point [-]

Depends. You could have a robotic somatic component, or a human body grown in a vat.

Comment author: Lumifer 17 September 2014 03:45:20PM *  0 points [-]

I don't see much difference between a human body grown in a vat and one grown in a womb.

But, generally speaking, in the context of wireheading the somatic component matters.

Comment author: NancyLebovitz 16 September 2014 03:17:29PM *  8 points [-]

It Ain't Necessarily So: Why Much of the Medical Literature Is Wrong

Some of the material will be familiar, but there are examples I hadn't seen before of how really hard it is to be sure you've asked the right question and squeezed out the sources of error in the answer.

What follows is what I consider to be a good parts summary-- if you want more theory, you should read the article.

Consider a study published in the NEJM that showed an association between diabetes and pancreatic cancer.[3] The casual reader might conclude that diabetes causes pancreatic cancer. However, further analysis showed that much of the diabetes was of recent onset. The pancreatic cancer preceded the diabetes, and the cancer subsequently destroyed the insulin-producing islet cells of the pancreas. Therefore, this was not a case of diabetes causing pancreatic cancer but of pancreatic cancer causing the diabetes.

....

To illustrate the point, consider the ISIS-2 trial,[8] which showed reduced mortality in patients given aspirin after myocardial infarction. However, subgroup analyses identified some patients who did not benefit: those born under the astrological signs of Gemini and Libra; patients born under other zodiac signs derived a clear benefit with a P value < .00001.

I guessed at a seasonal effect, but Gemini and Libra aren't adjacent signs.

The frequency of these false-positive studies in the published literature can be estimated to some degree.[2] Consider a situation in which 10% of all hypotheses are actually true. Now consider that most studies have a type 1 error rate (the probability of claiming an association when none exists [ie, a false positive]) of 5% and a type 2 error rate (the probability of claiming there is no association when one actually exists [ie, a false negative)] of 20%, which are the standard error rates presumed by most clinical trials. This allows us to create the following 2x2 table.

I didn't realize that the false negative effect (not seeing a relationship when there actually is one) is higher than the false positive rate. This might mean that a lot of useful medical tools get eliminated before they'can be explored.

Also (credit given to Seth Roberts), if a minority of people respond very well to a treatment being tested, this is very unlikely to be explored because the experiment is structured to see whether the treatment is good for people in general (actually, people in general in the group being tested). This wasn't in the NEJM piece.

One classic example of selection bias occurred in 1981 with a NEJM study showing an association between coffee consumption and pancreatic cancer.[15] The selection bias occurred when the controls were recruited for the study. The control group had a high incidence of peptic ulcer disease, and so as not to worsen their symptoms, they drank little coffee. Thus, the association between coffee and cancer was artificially created because the control group was fundamentally different from the general population in terms of their coffee consumption. When the study was repeated with proper controls, no effect was seen.[16]

....

Information bias, as opposed to selection bias, occurs when there is a systematic error in how the data are collected or measured. Misclassification bias occurs when the measurement of an exposure or outcome is imperfect; for example, smokers who identify themselves as nonsmokers to investigators or individuals who systematically underreport their weight or overreport their height.[17] A special situation, known as recall bias, occurs when subjects with a disease are more likely to remember the exposure under investigation than controls. In the INTERPHONE study, which was designed to investigate the association between cell phones and brain tumors, a spot-check of mobile phone records for cases and controls showed that random recall errors were large for both groups with an overestimation among cases for more distant time periods.[18] Such differential recall could induce an association between cell phones and brain tumors even if none actually exists.

::::

An interesting type of information bias is the ecological fallacy. The ecological fallacy is the mistaken belief that population-level exposures can be used to draw conclusions about individual patient risks.[4] A recent example of the ecological fallacy, was a tongue-in-cheek NEJM study by Messerli[19} showing that countries with high chocolate consumption won more Nobel prizes. The problem with country-level data is that countries don't eat chocolate, and countries don't win Nobel prizes. People eat chocolate, and people win Nobel prizes. This study, while amusing to read, did not establish the fundamental point that the individuals who won the Nobel prizes were the ones actually eating the chocolate.[20]

On the other hand, if you want to improve the odds of your children winning a Nobel, maybe you should move to a chocolate-eating country.

.....

A 1996 study sought to compare laparoscopic vs open appendectomy for appendicitis.[29] The study worked well during the day, but at night the presence of the attending surgeon was required for the laparoscopic cases but not the open cases. Consequently, the on-call residents, who didn't like calling in their attendings, adopted a practice of holding the translucent study envelopes up to the light to see if the person was randomly assigned to open or laparoscopic surgery. When they found an envelope that allocated a patient to the open procedure (which would not require calling in the attending and would therefore save time), they opened that envelope and left the remaining laparoscopic envelopes for the following morning. Because cases operated on at night were presumably sicker than those that could wait until morning, the actions of the on-call team biased the results. Sicker cases preferentially got open surgery, making the outcomes of the open procedure look worse than they actually were.[30] So, though randomized trials are often thought of as the solution to confounding, if randomization is not handled properly, confounding can still occur. In this case, an opaque envelope would have solved the problem.

Remembering that humans aren't especially compliant is hard.

From reading Guinea Pig Zero: The Journal for Human Research Subjects-- human beings are not necessarily going to comply with onerous food regimes. I expect that most who don't simply don't want to, but the magazine had the argument of not wanting to comply because the someone who's a human research subject is never going to be able to afford treatment based on the results of the research.

Comment author: Lumifer 16 September 2014 03:05:10PM 10 points [-]

An interesting paper. The abstract says:

Rationality leads people to imitate those with similar tastes but different information. But people who imitate common sources develop correlated beliefs, and rationality demands that later social learners take this correlation into account. This implies severe limits to rational imitation. We show that (i) in most natural observation structures besides the canonical single-file case, full rationality dictates that people must “anti-imitate” some of those they observe; and (ii) in every observation structure full rationality dictates that people imitate, on net, at most one person and are imitated by, on net, at most one person, over any set of interconnected players. We also show that in a very broad class of settings, any learning rule in which people regularly do imitate more than one person without anti-imitating others will lead to a positive probability of people converging to confident and wrong long-run beliefs.

Comment author: RichardKennaway 17 September 2014 11:23:18AM *  2 points [-]

I was this moment moved to search for the origin of a certain quote, and the process described in that paper seems to apply quite well to the promulgation of wrong citations. Here's a history of the idea of "three stages of truth". Actually, the situation for citations is even worse. The doctors in the example of the paper are observing their own outcomes as well as copying their predecessors' decisions, but someone copying a citation may make no observation of its accuracy.

More generally, memetic propagation.

Comment author: Metus 17 September 2014 02:03:34AM 3 points [-]

Ungated version?

Comment author: Douglas_Knight 17 September 2014 06:03:45AM 1 point [-]

Learn to use google scholar

Comment author: Lumifer 17 September 2014 04:04:43AM 2 points [-]

I don't know of one.

Comment author: Metus 16 September 2014 01:34:02PM *  2 points [-]

We managed to reduce performance on any number of tests to essentially a single number, g, together with a couple more for domain-specific skill. We managed to reduce the huge variation in personalities to five numbers, the OCEAN dimensions. I even recall reading that there is quite some correlation between those five numbers and that they might be reduced to a single one but I can't find the source any more.

Can we construct a whole host of other, similar numbers, like "math skills" and thus measure the impact of education and aging?

Another number I have in mind is, can we construct three numbers general health gh, mental health mh and physical health ph, and measure their correlations? I have the vague observation that medical issues tend to cluster, that is people with mental issues tend to not only exhibit any one of ADHD, depression, OCD and so on, but more than one of them. Similarly I have the impression that people tend to complain of many physical symptoms at once.

I seem to recall that BMI and/or WHR tend to be excellent predictors of physical health. Together with a couple of more measures these predictions can further be improved. The advantage of having a single number would be for research purposes on population health and it is easier to have a single mumber for personal assesment.

Comment author: Lumifer 16 September 2014 03:11:00PM 5 points [-]

We managed to reduce...

Not quite reduce. We managed to develop certain approximations which, albeit crude, work sufficiently well for some purposes. Of course, not all purposes.

I seem to recall that BMI and/or WHR tend to be excellent predictors of physical health.

I seem to recall they tend not. In particular, BMI is a flawed indicator as it has a pronounced bias for short and tall people.

these predictions can further be improved

Which "these predictions" -- what are you forecasting?

Comment author: hyporational 18 September 2014 04:25:51AM 1 point [-]

In particular, BMI is a flawed indicator as it has a pronounced bias for short and tall people.

And muscular people. What's wrong with WHR?

Comment author: ChristianKl 16 September 2014 03:10:27PM 1 point [-]

A high pressure is a good predictor for someone being unhealthy. On the other hand statins that reduce blood pressure don't provide the returns that people hoped for.

Goodhard's law applies very much.

Before dying with a heart attack Seth Roberts had a year where he improvement on the score that's the best predictor for heart attacks, while most people don't improve on the score as they age.

Using metrics like BMI and WHR seems to me very primitive. We should have no problem running a 3D scan of the whole body. I would estimate that obesitey[3D scan + complex algorithm] is a much better metric than obesity[BMI], obseity[WHR] or obesitey[BMI/WHR].

That's to be further improved by not only going for the visible light spectrum but adding infrared to get information about temperature. And you can follow it up by giving the person a west with hundreds of electrodes and measuring the conductance.

The tricoder xprice is also interesting.

As quantified self devices get cheaper it will also be possible to use their data to develop new metrics. A nursing home could decide to give every member a device that tracks heart rate 24/7. After a few years time the can give the data to some university bioinformatics folks who try to get good prediction algorithms.

Can we construct a whole host of other, similar numbers, like "math skills" and thus measure the impact of education and aging?

Math skills can mean multiple things to different people. Some people take it to mean the ability to calculate 34*61 in a short amount of time and without mistakes. Other people take it to mean doing mathematical proofs.

Comment author: NancyLebovitz 16 September 2014 03:38:09PM 2 points [-]

We might even find something more sophisticated than fat percentage. Not all fat people are ill/heading towards illness. Not all thin people are healthy.

Comment author: hyporational 18 September 2014 04:34:30AM 3 points [-]

Accumulation of fat to vital organs like the liver could be a better predictor. Fatty liver can be diagnosed via ultrasound, which is cheap.

Being fat is a risk even if you get sick for other reasons. Rehabilitation suffers.

Comment author: NancyLebovitz 18 September 2014 10:11:56AM 2 points [-]

Cite?

Comment author: hyporational 18 September 2014 12:53:56PM 2 points [-]
Comment author: ChristianKl 16 September 2014 03:57:14PM 3 points [-]

Yes, we have to try many different metrics and see which ones work best and for what purposes.

Comment author: Metus 16 September 2014 01:26:05PM *  5 points [-]

A medical issue is a problem if the patient recognises it as one. If a patient suffers from something that is not recognised as medical problem we call it hypochondria. Is there the concept of something we see as medical problem but the patient does not realise as one e.g. because they don't know that their condition is not normal?

Comment author: hyporational 17 September 2014 12:23:25PM *  3 points [-]

Terminology regarding missing symptom awareness depends on what is thought to be the cause. Anosognosia and other agnosias would be used for neurological disorders where self-monitoring is specifically impaired while denial, delusions and hallucinations would be used for psychiatric disorders. Denial could also be a psychiatric symptom concerning a somatic disorder. I'm not sure if other somatic fields than neurology have special terminology.

If a patient suffers from something that is not recognised as medical problem we call it hypochondria.

Not really. For example grief is not recognized as a medical problem, people suffer from it and we don't call it hypochondria.

Hypochondria is excessive worry about having a serious illness.

ETA: I think that whatever we choose to call a medical problem largely depends on our values and mere diversion from the biological norm does not a medical problem make. So the hypothetical patient could also simply disagree with others about what constitutes a medical problem.

Comment author: polymathwannabe 16 September 2014 03:30:51PM 5 points [-]
Comment author: ChristianKl 16 September 2014 03:11:07PM 2 points [-]

I think the standard terminology is "undiagnosed illness".

Comment author: chaosmage 16 September 2014 01:12:50PM *  4 points [-]

Since 23andme has been prohibited from giving health-related genetic reports, is there anyone else (outside the FDA's jurisdiction) who provides similar services?

Edit: I have found Promethease, which works with 23andme's raw data. I'm still interested in additional options.

Edit2: This page lists various 23andme competitors, although it was last updated in early 2013. More recent information is appreciated.

Comment author: CellBioGuy 16 September 2014 01:25:41PM *  5 points [-]

You can download your raw SNP-call data from them and run it through a plethora of third-party programs. Won't have the slick interface or the collation of multiple SNPs that affect the same trait but definitely tells you what you want to know about Mendelian diseases and you can sift through the rest.

See http://www.23andyou.com/3rdparty for some of the tools.

Comment author: chaosmage 16 September 2014 02:20:11PM 2 points [-]

Thanks a lot!

Comment author: michaelkeenan 16 September 2014 06:47:44AM *  8 points [-]

If you liked Scott Alexander's essay, Meditations on Moloch, you might like this typographic poster-meme I made. It was a minor success on Facebook.

(If you haven't read Scott Alexander's essay, Meditations on Moloch, then you might want to check it out. As Stuart Armstrong said, it's a beautiful, disturbing, poetical look at the future.)

Comment author: shminux 16 September 2014 06:32:44PM *  0 points [-]

I don't understand... The point of the essay is that one should not anthropomorhize Moloch, and your meme does exactly that.

Comment author: michaelkeenan 16 September 2014 07:06:50PM 7 points [-]

There is the line "thinking of the system as an agent throws into relief the degree to which the system isn’t an agent" so I see what you mean. But I think that just means that there's no sane agent to deal with, no law of the universe that says we can appease Moloch in exchange for something.

But anthropomorphizing Moloch, perhaps poetically, is different, and there's plenty of anthropomorphizing Moloch in the essay:

"But if we have bound Moloch as our servant, the bonds are not very strong, and we sometimes find that the tasks he has done for us move to his advantage rather than ours."

"We will break our back lifting Moloch to Heaven, but unless something changes it will be his victory and not ours."

"In the very near future, we are going to lift something to Heaven. It might be Moloch. But it might be something on our side. If it is on our side, it can kill Moloch dead."

"Moloch is exactly what the history books say he is. He is the god of Carthage. He is the god of child sacrifice, the fiery furnace into which you can toss your babies in exchange for victory in war. He always and everywhere offers the same deal: throw what you love most into the flames, and I will grant you power. As long as the offer is open, it will be irresistable. So we need to close the offer. Only another god can kill Moloch. We have one on our side, but he needs our help. We should give it to him."

My frail human mind is more motivated by war on a hated enemy than by abstractly maximizing utility, so I like the idea of frustrating a raging Moloch.

Comment author: MaximumLiberty 16 September 2014 04:17:27AM 5 points [-]

I like the notion of the Superintelligence reading group: http://lesswrong.com/lw/kw4/superintelligence_reading_group/. But the topic of AI doesn't really interest me much.

A reading group on some other topic that is more along CFAR's lines than MIRI's would. For example, reading recent studies of cognitive bias would be interesting to me. Discussion on how practically to combat them might evolve from discussing the studies.

Max L.

Comment author: ChristianKl 16 September 2014 11:41:15AM 2 points [-]

I would be up for it.

Comment author: TylerJay 16 September 2014 01:13:13AM *  4 points [-]

Here are two bookmarklets that have really helped my article-reading workflow. I named the bookmark for #1 "Clean" and #2 "Squirt":

1. javascript:window.location.replace("<http://justread.mpgarate.com/read?url=>" + escape(document.URL))
2. javascript:(function()%7Bsq%3Dwindow.sq%3Dwindow.sq%7C%7C%7B%7D%3Bif(sq.script)%7Bsq.again()%3B%7Delse%7Bsq.bookmarkletVersion%3D%270.3.0%27%3Bsq.iframeQueryParams%3D%7Bhost:%27//squirt.io%27,userId:%27f98bd133-7ae4-45da-be5d-16a4919b902c%27,%7D%3Bsq.script%3Ddocument.createElement(%27script%27)%3Bsq.script.src%3Dsq.iframeQueryParams.host%2B%27/bookmarklet/frame.outer.js%27%3Bdocument.body.appendChild(sq.script)%3B%7D%7D)()%3B

#1 is a nice, simple frontend for the Readability API. Just enter the URL of a page with something you want to read on it, and it extracts the content without any sidebars, ads, or other junk and gives it to you in an easy-to-read format.

#2 is Squirt, a speed-reading application that takes the text of any webpage and displays it to you one word at an adjustable speed. The default is 450wpm I think, but after you make an adjustment, it remembers what speed you want for next time. If you need to read a part more carefully or go back because you missed something, that's easy. Hit the spacebar to pause and it will show the context, then use the left and right arrows to move around. Hit the spacebar again to resume speed reading. Another awesome feature is that it tells you exactly how long it will take to finish if you don't stop, so you can decide if it's worth your time or not.

The two work really well together, as squirt alone will sometimes grab text you don't want. What I do is "Clean" a page by clicking on the bookmarklet, and then sometimes hit "Squirt" to speed read it.

Try it out and let me know what you think!