Open thread, September 15-21, 2014
If it's worth saying, but not worth its own post (even in Discussion), then it goes here.
Notes for future OT posters:
1. Please add the 'open_thread' tag.
2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)
3. Open Threads should be posted in Discussion, and not Main.
4. Open Threads should start on Monday, and end on Sunday.
Loading…
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)
Comments (339)
I seem to have high karma, but don't know why. Looking through my contribution history, I seem to only have a total of 47 net upvotes on anything I've ever posted, but have 74 karma points, including 10 in the last 30 days. Looking at the LW wiki FAQ, it says that you can get 10 karma per upvote if you post in main, but I haven't done that. Does anyone know why this might be happening?
Recently the discussion on modeling a better version of yourself was moved to Main. Had you made any comments there?
No, but I just realised that everything adds up if I assume that meetup posts also get 10 karma for every upvote. Given that this sort of makes sense but that I can't find it mentioned anywhere, I'm not sure whether it's a feature or a bug.
It's an undocumented feature.
Mystery solved!
I seem to have picked up 30-40 karma I can't account for over the past week. I wondered if it was some effort to undo the efforts of identified mass-downvoters.
That doesn't seem likely in my case, since the only non-meetup things I've posted before today have been about the MWI and Scott Aaronson's take on integrated information theory.
So you're saying there's a mass up-voter lurking out there? :) (If so I may have been the object of their efforts too.)
I was thinking more mod/admin efforts. I've never apparently been the victim of mass-downvotes, but if there was an intervention to remove all downvotes made by a troublesome user, the sudden removal of all their cumulative downvotes over time might look like an unexpected karma boost.
See the last open thread - several people (including me) have been victims of a mysterious mass upvoter who has gone through everything we've posted in the last month or so adding +1s indiscriminately up to about +30 to +40.
Personally I suspect it is a psychological experiment, examining the difference in reaction between this mysterious mass upvoter and the recent mysterious mass downvoter. Or, yknow, a troll.
Or someone who read one of your comments, like it, and then read others and liked them as well.
That's probably not it, given that I was one of Eugine's identified victims, and my karma has not changed in > 30 days.
Also, here is the discussion from the previous OT.
I have also noticed a bit of a spike, but if it were a de-euginiering effort, the change would be 1000+ points for me, not 30-40. So probably something else is going on.
I've had a similar thing. Prior to that I had a bunch of -1s that arrived all at once and which I considered unjustified.
I don't have any sensible way of learning about current affairs. I don't consume broadcast or print news. Most news stories reach me through social media, blogs, word of mouth or personal research, and I will independently follow up on the ones I think are worthy of interest. This is nowhere near optimal. It means I will probably find out about innovations in robotic bees before I find out about natural disasters or significant events in world politics.
Regular news outlets seem to be messy, noisy attention traps, rather than the austere factual repositories I wish them to be. Quite importantly, there seems to be a lot of stuff in the news that isn't actually news. I'm pretty sure smart people with different values will converge on what a lot of this stuff is.
Has this problem been solved already? I'm willing to put in time/effort/money for minimalist, noise-free, sensibly-prioritised news digest that I care about.
ETA: Although I haven't replied to all these responses individually, they seem very useful and I will be following them up. Thanks!
What sort of current events do you want to find out about how quickly, and why?
You should consider, if you haven't already, the possibility that the value of learning about such things quickly is almost always almost exactly zero. Suppose e.g. there's an enormous earthquake half-way around the world from you, and many thousands of people die. That's a big deal, it's very important -- but what immediate difference should it make to your life?
One possibility: you might send a lot of money to a charity working in the affected place. But it seems unlikely to me that there's much real difference in practice between doing so on the day of the disaster and doing it a week later.
Another possibility (albeit a kinda callous one): it may come up in conversation and you may not want to sound bad. But I bet that in practice "social media, blogs, word of mouth or personal research" do just fine at keeping you sufficiently up to date that you don't sound stupid or ignorant. In any case, what you need to know about in order to sound up to date is probably roughly what you get from existing news sources, rather than from a hypothetical new source of genuinely important, sensibly prioritized news.
I appreciate the distinction you make between urgent and non-urgent news.
Finding out about things quickly isn't necessarily my priority. In fact, one of my problems with "regular" news outlets is that they have poor sense of time sensitivity, and promote news that's stopped being useful. The value of knowing about Icelandic volcanoes grounding all northern European air traffic is actually very useful to me when it's just happened, but in a week's time I may as well read about it on Wikipedia.
I'm more concerned about finding out about things at all. My ad hoc news accretion drops the ball more often than I'd like. My ideal wish-upon-a-star would be a daily digest saying "here are a list of things that have happened today in two sentences or less". I can then decide whether to follow it up or not.
(I have a secondary motive of wanting to associate events in my memory to improve the granularity of my recall. I know, for example, that Eyjafjallajökull erupting was concurrent with the run-up to the 2010 UK General Election, which helps me position it in time quite accurately, as well as position personal events that I remember happening around the same time.)
Do you have some examples in mind of things you never found out about but would have been better off for knowing?
(Of course if you literally never found out about something you can't know. But I'm guessing there are things you did find out about but not until much too late.)
A couple of semi-recent examples would be the referendum on Scottish independence and the Islamic State business in the middle east. I obviously found out about them, but it felt like I found out about them a lot later than I would have liked. It's not so much that these have an immediate impact on my life (Scottish independence does, but it's not like I'd be able to remain ignorant by the time it's resolved), but they're massive news events that I basically didn't notice until everyone else was talking about them. This suggests I'm probably missing other events that people aren't talking about, and that makes me want to up my game.
What about the recent Swedish election results?
Incidentally, it was disturbingly hard to find an article about them that didn't put a misleading spin on the results.
Hilariously, a good option for you may be an actual newspaper. Made out of paper.
It comes once a day, it summarizes a few dozen major events in a reasonably succinct way, and many of them try to minimize reporting bias. You could consider specific papers based on size and editorial style (most offer free or cheap trials), and then sign up for a short subscription to see how you like it.
And the greatest advantage is that it has no hyperlinks to click. Thus, you only spend limited time reading it.
But it has a lot of the same stuff you'd have found beyond the hyperlinks -- right underneath the headlines, without even needing to click. I'm not sure that's a win.
No additional clicks from there, though, so still bounded. You can read through all the interesting stories in a paper (I used to do this) and then you're done; with the web there's no obvious stopping place.
I find daily takes up too much time, and the reporting doesn't have enough distance. So I'd recommend reading a Sunday paper instead - or, better still, a weekly or monthly magazine. If you're in the UK then Prospect is fantastic; I also read TIME (I've heard allegations that the US edition is dumbed down, so try to get a European or Asian edition).
Interesting, so the European edition of TIME is not a complete insult to their readers' intelligence?
That hasn't been my experience with newspapers.
I get The New York Times, and I find it pretty good in those regards (depending on your definition of "reasonably succinct"). And as a bonus, its science reporting is not hair-rippingly terrible at all generally.
Get an RSS reader and read only the headlines. That way you can process hundreds of news in a few minutes and only open the ones that seem seriously important.
A trivial inconvenience which could make a huge difference -- if there was a software which would put all those headlines in plain-text format, to reduce the temptation of clicking. (There is still google if something is irresistible.)
Wikipedia's current events portal is relatively minimalist and low-noise. It's not prioritized very impressively.
I scan google news headlines, top stories section and click on the items of interest. Yes, there are still attention grabs and non-news, but this is usually fairly clear from the article names.
I don't think it's a problem. Social media is good enough to tell you about significant events in world politics.
When a new topic bobbles up were I want to have an informed opinion I found vox.com or the Wikipedia summary to be good.
The Economist has a Politics This Week and Business This Week section. Both are only a page each and are international in scope.
In a similar vein: How do I find out what to read and what to learn more generally? I don't care about reading the latest Piketty but I want to read the best summary and interpretation of a philosopher from the last 10 years instead of the original from 500 years ago. Same goes for Physics text books and so on, and literature.
My foreign news comes almost exclusively from the CFR Daily News Brief, which sounds like exactly what you're looking for. The daily briefs also link to their Backgrounders, which are excellent and relatively short summaries of the backgrounds to many hot-topic issues.
Did you mean to link here?
Yes, thanks. Apparently didn't copy/paste correctly. Fixed now.
It was quite an entertaining copy/paste error.
:). Was the CFR stuff what you were looking for?
Oh, yes, thanks. I'm making a collection of news-digesty bookmarks, and it's in there.
I can't find it, but I once read an article from a guy a trust about how he just stopped following news, assuming that if anything sufficiently important happened, he'd find out about it anyway. His quality of life immediately rose. Having followed this approach for a few years now, I would suggest consuming zero news (is minimalist, completely devoid of noise, and exceptionally well-organized).
"Remember, if it’s in the news don’t worry about it. The very definition of news is “something that almost never happens.” When something is so common that it’s no longer news — car crashes, domestic violence — that’s when you should worry about it." - Bruce Schneier
I get my news from instapundit.
I don't wish to get into a mindkilling debate about this here, but for sixes-and-sevens benefit, I'll note that Instapundit is a highly ideological libertarian (alternatively, in the view of many progressives, a partisan Republican pretending to be a libertarian). If you use him as a news source, you should balance with a progressive source.
ETA: This advice holds even if you are skipping narrowly political articles and reading about crises/disasters, etc., since ideology informs what kinds of crises people consider salient.
Are you using "progressive" to mean left-leaning, or in the usual way? Just for clarity; if you meant the latter disregard.
I thought "left-leaning" was the usual way? What else, in the political sphere, does "progressive" mean?
I've heard it it as synonymous with "good," "new" and anti rich tax policy. Can you make a recommendation? Either just left or, since libertarian is socially liberal fiscally conservative, a good source that is fiscally liberal and socially conservative? I asked the DNC for the former and just got on their mailing list. Not impressed.
Of course a progressive will think that progressivism is good, and part of progressivism is that it is good becuase it is new (the clue is in the name). Those who are not progressives will hardly agree. And anti rich tax policy is a straightforward left-leaning policy.
It is tempting for progressives to define the word to mean "good" and "new", as it saves them the trouble of defending the ideology. The ideology can then be treated not as any set of beliefs about the reality, but as reality itself.
Amen. Just saying I've heard that use from other moderates as well who don't think too hard about it.
Anyway, the other question is the more interesting to me. Any good left-leaning or socially-conservative-fiscally-liberal (short name?) news source?
Short name = Christian.
I honestly had not considered a Christian news option.
"Christian" covers a lot of ground. That's a fair description of the mainline Catholic viewpoint, but looking up a random Christian news source in the US could get you fiscal viewpoints ranging from lukewarm left to hardline right to more or less apolitical.
(It's reliably socially conservative, though, generally speaking.)
No, that's not it. It doesn't mean you can't have new things happen that are bad. It does refer to a time derivative, but it's more of a goal than a statement of fact: government and society are not as good as they could be, and we can engineer the government to improve both. That's 'progress'. (Note: this summary is not an endorsement)
Progressive tax structures are not named so due to this time derivative. They are named so due to the derivative in income. Regressive tax structures exist, but they aren't named so due to being more like the past.
That is progress, but that is not what is meant by "progressive" in the political sense. The belief that government can be engineered to improve things is shared by everyone except those in despair of it ever happening. Moldbug has proposals to do that -- is he a "progressive"?
No, "progressive" means certain specific views about what is valued as an improvement, and specific beliefs about what policies will make those improvements. These values and views are accurately summarised as "left-leaning".
The US "left" is considerably to the right of the European left, and LW has a broad international readership, so I think just saying "left" would be more confusing ("liberal" would even more confusing, given the dispute between libertarians and progressives over who is the legitimate heir of 19th century liberalism). But yes, in this case, I meant progressive in the sense of "mainstream center-left."
Some of the US "left" (notably, the mainstream Democrats) are considerably to the right of the European left. "Left" encompasses a rather large landscape.
Right right, thanks. Any source you'd recommend?
This looks like the classic grey fallacy.
Looks like, but isn't. The goal isn't that you take one viewpoint and take another viewpoint and find "something in the middle"; the point is that having multiple independent viewpoints makes it easier to spot mistakes in each.
It feels natural for us to think critically when our preconceptions are contradicted and to accept information uncritically when our preconceptions are supported. If you want to improve the odds that you're reading critical thought about any given topic, you need sources with a wide range of different preconceptions.
I agree and wouldn't have objected if Prismattic advised to read multiple sources from a variety of viewpoints. As it is, he just said "you need to read progressives as well" and that's a different claim.
I'm not arguing that the views should be averaged, but that the combined sample of news stories will be less likely to suffer from politically motivated selection bias. A libertarian/fusionist source is likely to devote more coverage to, say, stories of government corruption and less to stories of corporate wage theft or environmental degradation; a progressive source to do the opposite. All of those stories might be important (in general or to sixes-and-sevens in particular), so the combined news feed is in that sense better.
So why did you recommend progressives and not, say, news coming from the Roman Catholic Church, from marxists, from PETA, from infowars, from Al-Jazira, etc. etc.?
Well, taking those specific examples as non-rhetorical: PETA, the Catholic Church, and Infowars are various kinds of insane in ways that extend beyond ordinary political mindkilling, so I'd be unlikely to recommend them. Al-Jazeera English is actually pretty good as a news source, but its website is an adjunct of being a broadcast news source, which is less helpful from a time-investment perspective. I predict that a center-left news source will provide coverage on a broader range of issues than a far-left news source, but your mileage may vary.
The center-left source is also most likely to compensate specifically for the coverage holes in a center-right source. That still isn't averaging their factual claims.
You're not averaging factual claims, you're averaging exposure to viewpoints.
I would argue that this summing, not averaging exposure. There's a difference between saying "You should read both GreenNetNews and BlueCast" and saying "To save time, read GreenNetNews on odd-numbered days and BlueCast on even-numbered days".
I think it's averaging because your capacity to absorb news/viewpoints is limited.
Any particular reason you didn't make a similar reply to Christian's suggestion of the ideologically progressive vox dot com?
Because I hadn't seen it.
I find the implied accusation of bias amusing. I've actually tweeted at Matt Yglesias once to complain about the quality of an article on Vox.
Instapundit is highly ideological libertarian, so you should balance it out with a reactionary news source like Theden.tv or Steve Sailer.
As it happens I also read Steve Sailer, although he isn't so much news as editorial cometary whereas instupundit is more "list of headlines" of the kind sixes-and-sevens was asking about.
In what way? Do you wish you spent more time following current affairs? I don't follow them, but don't see any problem with it - if anything, I occasionally have to resist the urge of looking up what's going on in the world, which I put in the same mental bucket as the urge to look at the top entries of /r/funny.
I don't think in ten years time having read one more news item on the Gaza Strip will change my life more than having seen one more picture of a cat stuck in a bowl.
(I do however sometimes go more into a binge of "reading up on something and trying to understand it", but I rely more on Wikipedia than on news for that; "breaking news" tends to repeat the same points over and over again, and doesn't put much focus on the big picture)
I used to read the wikipedia current events page, which I found a nice summary of what's going on without going into too many details.
I trust my brain to collect facts and raise them to my attention when they're important. "Current affairs" describes a class of fact that I don't think is being adequately collected.
The Wikipedia current events page is a very good example of what I'm looking for.
Okay, I know next to nothing about Haskell, and next to nothing about provability logic, so maybe what I'm about to ask doesn't make any sense, but here's something that's making me very curious right now. How do I implement a function like this:
using some typeclass like this:
The idea is that the implementation of loeb should follow the steps of this proof, and the methods of Prov should correspond to the assumptions of that proof. The question was inspired by this post by sigfpe, but he thought that the class should be Functor, which seems wrong to me.
Apologies for the formatting, it turns out LW collapses whitespace even in preformatted blocks.
Functor is a weak typeclass; the only thing it implies about □ is that □(a→b)→□a→□b (which we already know to be true). So the idea of using Functor was just to make the minimum possible assumption.
I don't know enough Haskell to answer your question, but if you can write it in Scala I'll give it a go.
The problem is that Functor's operation isn't quite what you wrote, but rather (a→b)→□a→□b, which is way too strong. I don't think it holds in provability logic. That's why I want to define an even weaker typeclass Prov, but I'm not sure what methods it should have, apart from dup and mp.
If you give me some Scala code, I think I'll be able to make sense of it :-)
You're right, I was taking the linked thing at face value.
The signature given is almost exactly Comonad. If I'm reading this right, Loeb's theorem gives you something vaguely interesting: it's a function from C[A] => A to A. So it tells you that any function that "flattens" a comonad must have some kind of "zero" value - e.g. any Stream[A] => A must give rise to a distinguished value of type A - which you can extract without ever having an instance of Stream[A].
I've replied with Scala code upthread.
I'm guessing that we get soundness :: a -> p a for free in your notation?
I think you wanted loeb :: Prov p => (p a -> a) -> p a.
Scala implementation, I think:
Thanks a lot! But your soundness condition isn't one of the assumptions of the theorem, and it's still too strong for my taste. Maybe I should clarify more.
I want the implementation of loeb to actually be a proof, by Curry-Howard. That means the implementation needs to be translatable into a total language, because non-totality turns the type system into an unsound logic where you can prove anything. An extreme case is this silly implementation of loeb, which will happily typecheck in Haskell:
Sadly, your version of loeb can't be made total. To see why, consider the identity wrapper type:
You immediately see that loeb specialized for Id is just the Y combinator with type (a -> a) -> a, which shouldn't be implementable in a total language, because (a -> a) -> a isn't true in any logic. To avoid that, the typeclass must have some methods that aren't implementable for Id. But all methods in your typeclass are easily implementable for Id, therefore your loeb can't be made total.
The same argument also rules out some other implementations I've tried. Maybe the root problem is the use of types like Fix, because they don't have quite the same properties as the diagonal lemma which is used in the original proof. But I don't really understand what's going on yet, it's a learning exercise for me.
I think the problem is hidden beneath the fixed point operator (I've edited the code to be more correct). For Id, is there actually a type Psi such that Psi <~> (Id[Psi] => A)? Isn't that where the Y combinator comes in - the only reason loeb is able to act as a Y combinator is that it gets one passed in via the fixed point lemma? Isn't a fixed point lemma always going to be the same as a Y combinator?
(I wish we were working with a proof that didn't invoke this external lemma)
Hmm, you're right, in a total language you can't define such a Psi for Id. Maybe the type class should go like this:
I don't know if that's enough. One problem is that fix and unfix are really inconvenient to use without type lambdas, which Haskell doesn't have. Maybe I should try Scala, which has them.
Another problem is in step 13. To do that without soundness, I probably need versions of dup and mp that work inside p, as well as outside. But that makes the type class too complicated.
How do I build the habit of writing down a fleeting thought that seems interesting? Way too often I notice that I just wanted to do something or write something down. Or should I just accept the thought as gone?
Do you have a smartphone? Just hit the voice command button and say "Make a note: [whatever you need to make a note of]"
Alternatively, I find it's easier to do things when you've already started. Maybe make a note on your phone and add a couple ideas. Then, the next time you have an idea, you already have a dedicated place to put it, so you just put it there instead of wondering what to do with it and ending up doing nothing because you can't get past the inertia of starting. Then award yourself a mental point as a reward to train your mind to keep coming up with ideas and writing them down.
And if you do forget something, don't worry about it too much. The whole "I just had a really good idea but now I forgot, oh no!" pattern is really common. Much more common than the "I just had a great idea and I wrote it down and I still think it's great" pattern. I've always taken this as evidence that the idea you forgot wasn't actually that great and it only feels that way because you forgot about it and you're suffering from a kind of forbidden-fruit/grass-is-greener bias. People tend to remember really good ideas because they're contextual or actionable.
get a twitter
I carry at all times a tiny notebook (smaller than my hand) and a pen so small I can barely use it comfortably. That's low tech and not very efficient (because I'll need to type it up later), but very quick, easily survives the sometimes inhospitable pocket environment, doesn't need electricity and works for non-textual thoughts.
I do this too, and it holds my former wallet contents in a pocket, so not even an extra thing to track.
I have Simplenote installed on my phone, and I pull things out and note them there very frequently.
(Later, they become blog posts/to-dos/etc)
Here are two bookmarklets that have really helped my article-reading workflow. I named the bookmark for #1 "Clean" and #2 "Squirt":
#1 is a nice, simple frontend for the Readability API. Just enter the URL of a page with something you want to read on it, and it extracts the content without any sidebars, ads, or other junk and gives it to you in an easy-to-read format.
#2 is Squirt, a speed-reading application that takes the text of any webpage and displays it to you one word at an adjustable speed. The default is 450wpm I think, but after you make an adjustment, it remembers what speed you want for next time. If you need to read a part more carefully or go back because you missed something, that's easy. Hit the spacebar to pause and it will show the context, then use the left and right arrows to move around. Hit the spacebar again to resume speed reading. Another awesome feature is that it tells you exactly how long it will take to finish if you don't stop, so you can decide if it's worth your time or not.
The two work really well together, as squirt alone will sometimes grab text you don't want. What I do is "Clean" a page by clicking on the bookmarklet, and then sometimes hit "Squirt" to speed read it.
Try it out and let me know what you think!
Which speed are you using at the moment and how long did it take you to come to that speed?
I started at 350 and that's still what I use most of the time. For light, non-technical articles, I can do 450, but it's a bit uncomfortable to focus that hard and I do miss things occasionally. I can usually tell if it was important or not though, so I know if I need to go pause and rewind. After playing with speedreeding off and on for a few years, I've come to the conclusion that it's definitely possible to read faster than I normally do with equal comprehension, but that there really is a limit and the claims you see from speedreading courses are hyperbolic. The thing I like about Squirt is that it eliminates the need to use a pacer.
1 didn't work for me.
For example, trying it on tvtropes.org got me http://tvtropes.org/pmwiki/pmwiki.php/%3Chttp://justread.mpgarate.com/read?url=%3Ehttp%3A//tvtropes.org/pmwiki/pmwiki.php/Main/HomePage
Hmm... Yeah, that's not right. Maybe there was a problem when I pasted it? Here it is again.
Only other thing I can think of is you may have a browser extension interfering.
It still doesn't work. It could be an extension, but I was guessing it was just the browser. I'm using Chrome. javascript:alert("test") seems to work if I type it direction or use a bookmark. It doesn't work if I copy and paste.
Did you try copy/pasting my link into a bookmark? That's what I was recommending, sorry if that wasn't clear. When I copy/paste anything starting with javascript: directly into the URL bar, the javascript: part gets dropped.
Looks like this is a bug with the way LW parses markdown. You need to remove the angle brackets just inside the quotes.
That fixed it. Thanks.
I like the notion of the Superintelligence reading group: http://lesswrong.com/lw/kw4/superintelligence_reading_group/. But the topic of AI doesn't really interest me much.
A reading group on some other topic that is more along CFAR's lines than MIRI's would. For example, reading recent studies of cognitive bias would be interesting to me. Discussion on how practically to combat them might evolve from discussing the studies.
Max L.
I would be up for it.
If you liked Scott Alexander's essay, Meditations on Moloch, you might like this typographic poster-meme I made. It was a minor success on Facebook.
(If you haven't read Scott Alexander's essay, Meditations on Moloch, then you might want to check it out. As Stuart Armstrong said, it's a beautiful, disturbing, poetical look at the future.)
I don't understand... The point of the essay is that one should not anthropomorhize Moloch, and your meme does exactly that.
There is the line "thinking of the system as an agent throws into relief the degree to which the system isn’t an agent" so I see what you mean. But I think that just means that there's no sane agent to deal with, no law of the universe that says we can appease Moloch in exchange for something.
But anthropomorphizing Moloch, perhaps poetically, is different, and there's plenty of anthropomorphizing Moloch in the essay:
"But if we have bound Moloch as our servant, the bonds are not very strong, and we sometimes find that the tasks he has done for us move to his advantage rather than ours."
"We will break our back lifting Moloch to Heaven, but unless something changes it will be his victory and not ours."
"In the very near future, we are going to lift something to Heaven. It might be Moloch. But it might be something on our side. If it is on our side, it can kill Moloch dead."
"Moloch is exactly what the history books say he is. He is the god of Carthage. He is the god of child sacrifice, the fiery furnace into which you can toss your babies in exchange for victory in war. He always and everywhere offers the same deal: throw what you love most into the flames, and I will grant you power. As long as the offer is open, it will be irresistable. So we need to close the offer. Only another god can kill Moloch. We have one on our side, but he needs our help. We should give it to him."
My frail human mind is more motivated by war on a hated enemy than by abstractly maximizing utility, so I like the idea of frustrating a raging Moloch.
Since 23andme has been prohibited from giving health-related genetic reports, is there anyone else (outside the FDA's jurisdiction) who provides similar services?
Edit: I have found Promethease, which works with 23andme's raw data. I'm still interested in additional options.
Edit2: This page lists various 23andme competitors, although it was last updated in early 2013. More recent information is appreciated.
You can download your raw SNP-call data from them and run it through a plethora of third-party programs. Won't have the slick interface or the collation of multiple SNPs that affect the same trait but definitely tells you what you want to know about Mendelian diseases and you can sift through the rest.
See http://www.23andyou.com/3rdparty for some of the tools.
Thanks a lot!
A medical issue is a problem if the patient recognises it as one. If a patient suffers from something that is not recognised as medical problem we call it hypochondria. Is there the concept of something we see as medical problem but the patient does not realise as one e.g. because they don't know that their condition is not normal?
I think the standard terminology is "undiagnosed illness".
https://en.wikipedia.org/wiki/Anton%E2%80%93Babinski_syndrome
https://en.wikipedia.org/wiki/Anosognosia
https://en.wikipedia.org/wiki/Anosodiaphoria
https://en.wikipedia.org/wiki/Somatoparaphrenia
https://en.wikipedia.org/wiki/Confabulation
Terminology regarding missing symptom awareness depends on what is thought to be the cause. Anosognosia and other agnosias would be used for neurological disorders where self-monitoring is specifically impaired while denial, delusions and hallucinations would be used for psychiatric disorders. Denial could also be a psychiatric symptom concerning a somatic disorder. I'm not sure if other somatic fields than neurology have special terminology.
Not really. For example grief is not recognized as a medical problem, people suffer from it and we don't call it hypochondria.
Hypochondria is excessive worry about having a serious illness.
ETA: I think that whatever we choose to call a medical problem largely depends on our values and mere diversion from the biological norm does not a medical problem make. So the hypothetical patient could also simply disagree with others about what constitutes a medical problem.
We managed to reduce performance on any number of tests to essentially a single number, g, together with a couple more for domain-specific skill. We managed to reduce the huge variation in personalities to five numbers, the OCEAN dimensions. I even recall reading that there is quite some correlation between those five numbers and that they might be reduced to a single one but I can't find the source any more.
Can we construct a whole host of other, similar numbers, like "math skills" and thus measure the impact of education and aging?
Another number I have in mind is, can we construct three numbers general health gh, mental health mh and physical health ph, and measure their correlations? I have the vague observation that medical issues tend to cluster, that is people with mental issues tend to not only exhibit any one of ADHD, depression, OCD and so on, but more than one of them. Similarly I have the impression that people tend to complain of many physical symptoms at once.
I seem to recall that BMI and/or WHR tend to be excellent predictors of physical health. Together with a couple of more measures these predictions can further be improved. The advantage of having a single number would be for research purposes on population health and it is easier to have a single mumber for personal assesment.
A high pressure is a good predictor for someone being unhealthy. On the other hand statins that reduce blood pressure don't provide the returns that people hoped for.
Goodhard's law applies very much.
Before dying with a heart attack Seth Roberts had a year where he improvement on the score that's the best predictor for heart attacks, while most people don't improve on the score as they age.
Using metrics like BMI and WHR seems to me very primitive. We should have no problem running a 3D scan of the whole body. I would estimate that obesitey[3D scan + complex algorithm] is a much better metric than obesity[BMI], obseity[WHR] or obesitey[BMI/WHR].
That's to be further improved by not only going for the visible light spectrum but adding infrared to get information about temperature. And you can follow it up by giving the person a west with hundreds of electrodes and measuring the conductance.
The tricoder xprice is also interesting.
As quantified self devices get cheaper it will also be possible to use their data to develop new metrics. A nursing home could decide to give every member a device that tracks heart rate 24/7. After a few years time the can give the data to some university bioinformatics folks who try to get good prediction algorithms.
Math skills can mean multiple things to different people. Some people take it to mean the ability to calculate 34*61 in a short amount of time and without mistakes. Other people take it to mean doing mathematical proofs.
We might even find something more sophisticated than fat percentage. Not all fat people are ill/heading towards illness. Not all thin people are healthy.
Yes, we have to try many different metrics and see which ones work best and for what purposes.
Accumulation of fat to vital organs like the liver could be a better predictor. Fatty liver can be diagnosed via ultrasound, which is cheap.
Being fat is a risk even if you get sick for other reasons. Rehabilitation suffers.
Not quite reduce. We managed to develop certain approximations which, albeit crude, work sufficiently well for some purposes. Of course, not all purposes.
I seem to recall they tend not. In particular, BMI is a flawed indicator as it has a pronounced bias for short and tall people.
Which "these predictions" -- what are you forecasting?
And muscular people. What's wrong with WHR?
An interesting paper. The abstract says:
Ungated version?
I don't know of one.
Learn to use google scholar
I was this moment moved to search for the origin of a certain quote, and the process described in that paper seems to apply quite well to the promulgation of wrong citations. Here's a history of the idea of "three stages of truth". Actually, the situation for citations is even worse. The doctors in the example of the paper are observing their own outcomes as well as copying their predecessors' decisions, but someone copying a citation may make no observation of its accuracy.
More generally, memetic propagation.
It Ain't Necessarily So: Why Much of the Medical Literature Is Wrong
Some of the material will be familiar, but there are examples I hadn't seen before of how really hard it is to be sure you've asked the right question and squeezed out the sources of error in the answer.
What follows is what I consider to be a good parts summary-- if you want more theory, you should read the article.
....
I guessed at a seasonal effect, but Gemini and Libra aren't adjacent signs.
I didn't realize that the false negative effect (not seeing a relationship when there actually is one) is higher than the false positive rate. This might mean that a lot of useful medical tools get eliminated before they'can be explored.
Also (credit given to Seth Roberts), if a minority of people respond very well to a treatment being tested, this is very unlikely to be explored because the experiment is structured to see whether the treatment is good for people in general (actually, people in general in the group being tested). This wasn't in the NEJM piece.
....
::::
An interesting type of information bias is the ecological fallacy. The ecological fallacy is the mistaken belief that population-level exposures can be used to draw conclusions about individual patient risks.[4] A recent example of the ecological fallacy, was a tongue-in-cheek NEJM study by Messerli[19} showing that countries with high chocolate consumption won more Nobel prizes. The problem with country-level data is that countries don't eat chocolate, and countries don't win Nobel prizes. People eat chocolate, and people win Nobel prizes. This study, while amusing to read, did not establish the fundamental point that the individuals who won the Nobel prizes were the ones actually eating the chocolate.[20]
On the other hand, if you want to improve the odds of your children winning a Nobel, maybe you should move to a chocolate-eating country.
.....
Remembering that humans aren't especially compliant is hard.
From reading Guinea Pig Zero: The Journal for Human Research Subjects-- human beings are not necessarily going to comply with onerous food regimes. I expect that most who don't simply don't want to, but the magazine had the argument of not wanting to comply because the someone who's a human research subject is never going to be able to afford treatment based on the results of the research.
I have a notion that an FAI will be able to create better friends and lovers for you than actual humans could be. Family would be a more complex case if you value the history as well as the current experience.
I'm not talking about catgirls-- if some difficulties in relationships are part of making relationships better in the long haul, then the FAI will supply difficulties.
If people eventually have relationships with FAI-created humans rather than humans generated by other means, is this a problem?
This looks to be wireheading lite and if you got there I don't see why you wouldn't make the next step as well -- the FAI will create the entire world for you to enjoy inside your head.
I thought wireheading meant stable high pleasure without content rather than an enjoyable simulated world. What do other people think wireheading means?
Well, technically the term "wireheading" comes from experiments which involved inserting an electrode (a "wire") into a rat's pleasure center and giving the rat a pedal to apply electric current to this wire. So yes, in the narrow sense wireheading is just the direct stimulation of the pleasure center.
However I use "wireheading" in the wide sense as well and there it means, essentially, the focus on deriving pleasure from externally caused but internal experiences and the lack of interest in or concern with the outside world. Wireheading in the wide sense is, basically, purified addiction.
If we're living inside an FAI, "outside world" might be getting a little vague. This might even be true if we're still living in our DNA-based bodies.
Do you think an FAI would let people have access to anything it isn't at least monitoring, and more likely controlling?
Uploads/ems are a bit of a different case.
I don't know, but in such a case I probably would not consider it a FAI.
How? Why does it matter in what substrate the information pattern called you resides in this case? I doubt the meat brain will have any connectibility issues once we have uploads.
I am not an information pattern having, for example, a considerable somatic component :-D
Depends. You could have a robotic somatic component, or a human body grown in a vat.
I don't see much difference between a human body grown in a vat and one grown in a womb.
But, generally speaking, in the context of wireheading the somatic component matters.
You've asked that before.
I don't have any new thoughts on this question, so I'll just quote my answer from there:
Let's first separate sexual aspects from the need for other companionship. Suppose everyone gets their sexual needs, if any satisfied by catgirls+ (+ for the upgrade which includes relationship problems if necessary). If you have a crush on your coworker (or your sibling, ew!), just add a catgirl copy of them to your harem.
Further suppose that the reproduction aspect is also taken care of.
Now you have a race of essentially asexual humans, as far as human-to-human interactions go.
The question is, does it make sense to have friendbots? What, if anything, is lost when you switch from socializing with meat humans to socializing with simulated ones?
It's not self-evident to me that they are separable.
When my heterosexual male friends tell me companionship isn't about sex I ask them how many male companions they've had. Not many, I've gathered from the silence.
For hetero males the usual term for male companions is "close friends". I bet the great majority have some.
But go ask some hetero women whether they think sex and companionship are well-separable :-/
In Finnish the connotations of "companion" are more obviously sexual I see, at least in my circles.
It's probably a language issue, in standard English the word "companion" has no sexual overtones.
More to the point, this subthread is explicitly about separating sex from companionship.
This strikes me as superstimulating. In particular, the more cat girls you have, the more and kinkier cat girls you want.
Not necessarily, Plenty of people are happy with vanilla sex (or without). I suspect that even the kinkiest ones out there also have their limit. If not, let's talk about those who do.
That's because vanilla sex isn't as stimulating. The more superstimulating something is, the more experiencing it causes you to want more of it.
That "explanation" is easily falsified. There are plenty of people who tried kinkier sex, enjoyed it, but reverted back to vanilla. There are plenty of people who tried roller-coasters once or twice but decided it's too much "stimulation".
That doesn't seem to be the case, see e.g. yummy food.
I think you're confusing "stimulating" and "addictive".
I thought that was already part of catgirls?
I say it's not a problem, but my views are outside the LW mainstream on this.
See also EYs Failed Utopia #4-2
I'm not sure we can extrapolate this currently. If we knew more, thought faster... maybe.
For me this means that one contraint on FAI is that it may not perform changes arbitrarily fast. Too fast for humans to react and adapt. There must be a 'smooth' trajectory. Surely not the abrupt change suggested in Failed Utopia.
What's a catgirl?
An indistinguishable-from-live sex toy.
With cat-ears.
I have yet to find any thoughts on Effective Altruism that do not assume vast amounts of disposable income on the part of the reader. What I am currently trying to determine are things like 'at what point does it make sense to give away some of your income versus the utility of having decent quality of life yourself and insuring against the risk that you end up consuming charitable resources because something happened and you didn't have an emergency fund'. Does anyone know of any posts or similar that tackle the effective utilitarian use of resources when you don't have a lot of resources to begin with?
I don't think there is a general answer to the question "How much should I consume?"
Most of the EA stuff I've seen doesn't appear to me to assume vast amounts of disposable income; merely enough to be willing to give some away. Then EA is about what to do with your charity budget, whether it's large or small.
How you prioritize helping others versus helping yourself (and your family, if any) is a more or less orthogonal question.
(I might suggest, snarkily, that for someone who requires "vast amounts of disposable income" before being willing to give any away no term with "altruism" in it is very appropriate. But that wouldn't be fair because e.g. your intention might be to secure yourself a reasonably comfortable life and then give away every penny you can earn beyond that, or something.)
That's it, basically; it's about how much of a buffer I'm 'allowed' to give myself on 'reasonably comfortable'; I'm supporting myself and full-time student partner and not in permanent full-time employment so my instinct whenever I have a sniff of an excess is to hoard it against a bad month for getting work rather than do anything charitable with it (or it all goes on things we've put off replacing for monetary reasons, like shoes that are still wearable but worn out enough to no longer be waterproof).
I think Lumifer articulated better than I could what I really wanted to know the answer to, and while there may not be a general answer it does mean that I can at least go looking for things to read now my real question is clearer to me. So thanks!
I don't think "allowed" is the right way to think about it (and your quotation marks suggest that probably you don't either). If you mean something like "what position do other reasonable people take?" or "what is the range of options that won't make other people who think of themselves as EAs disapprove of me?", I have no information on anyone else's positions but my own is something like this:
How do you think saving (in the standard financial sense) and giving should be balanced?
Um. With careful consideration?
Seriously: I don't have any very strong opinions on this, nor any reason to think that anyone should care what my opinions are. In my own family's case, we save substantially more than we give (very crudely, about 50% of income versus about 10%) but I'm not at all sure that if I thought about it longer and harder I wouldn't conclude that we should be weighting global welfare higher relative to our own financial security.
:-) Don't mean to pick on you, but the impression I get from EAs on LW is that your free cash flow is supposed to go save the world and I got curious about the apparent/potential disconnect from the general meme that people are supposed to save more so as not to be a drag on the society if something happens to them or when they retire...
The idea that your free cash flow should all go to save the world is generally based on a pretty straightforward utilitarian calculation, and it seems pretty clear that the same calculation would put saving lives in poor countries ahead of the small adverse consequences of drawing more on one's own country's social safety net. So I don't think there's much "disconnect" there.
In practice, very few people are quite so heroically altruistic as to reduce themselves to (what locally passes for) poverty so as to give everything to help the global poor. I bet the few who are are already largely neglecting saving; the rest of us, I think, first decide how much we want to give away and then how we want to balance saving and consumption. So a tradeoff between saving and giving, as such, doesn't arise.
For the avoidance of doubt, i very much don't think of myself as any sort of heroic or expert altruist (effective or otherwise). My only role here is Some Guy Who Got Into A Thread About Effective Altruism :-).
I am not sure this is the case. Saving and giving are kinda fungible without any immediate impact on you -- that's got to be tempting...
Thanks gjm, that's a really helpful comment. (And yes, quotation marks indicate 'this is the word I can think of but it is not necessarily the right word.)
I think points number 1 and 3 are especially relevant for me right now, and I have found talking it through on here to be very helpful in defeating an entirely non-useful lingering sense of guilt for not giving more when I really can't afford to, yet.
to me EA is more about how to answer the question "how should I be charitable?" than "Should I be charitable and to what extent?"
I don't have a link, but I suspect cutting this fine is not very valuable. That last $10k would be a lot to you, but that wouldn't make it more than any other $10k to a charity. Instead, ask how you could come to have a vast amount of disposable income. Including whether it makes sense to spend some money toward that end. You may be able to get a very high rate of return investing in yourself.
It seems like there are a lot of fan-fiction fans here. Fan-fiction fans, I am curious as to what draws you to the fan-fiction of which you are fans. Is part of it that you're fans of other fan-fiction fans? I guess that depending on the cosplay you could even be a fan of fan-fiction fans' fans.
This is a different question, but I've occasionally wondered why some franchises (like Star Trek, Buffy, or Harry Potter) generate a lot of fanfic and others generate much less. Part of this is raw popularity, of course, but quite a bit isn't; the film Avatar (the one with blue aliens, not the one with kung fu) was far more popular than, say, Pirates of the Caribbean, but the latter spawned a thriving fic community and the former has a smattering of stories mostly intended to illustrate critical points.
I don't think there's any single answer, but a franchise's chances seem to be improved if: it's suitable for episodic storytelling (Pirates is a self-contained story, but it's framed like an entry in a serial); it's got strong and ideally archetypal characters (Kirk, McCoy, Spock: action/emotion/reason, easy to write but easy to give depth to); and it's got an open setting with a lot of depth and unexplored bits (few settings outside spec-fic generate a lot of fanfic, and most of those that do are period pieces or procedurals). We're looking at works as toolkits for storytelling, in other words; tight plotting might actually be detrimental.
I remember seeing this organization on LW but cannot find it again or remember the name: it was a for-profit school-like entity that does a short training program (might have been 6 six week, maybe 3 months, that range), which is free upfront and takes their payment entirely as a percentage of the salary from the job they place you in afterward. If I remember correctly, it is run in the Bay Area and took a small pool each session, with a school-like application process.
Can anyone point me to this?
That sounds like App Academy or one of its competitors.
App Academy was the one I was thinking of specifically, thanks.
I'm posting here on behalf of Brent Dill, known here and elsewhere as ialdabaoth-- you may have enjoyed some of his posts. If you read the comments at SSC, you'll recognize him as a contributor of rare honesty and insight. If you'd had the chance to talk with him as much as I have, you'd know he's an awesome guy: clever, resourceful, incisive and deeply moral. Many of you see him as admirable, most as relatable, some as a friend, and more, I hope, as a member of our community.
He could use some help.
Until last Thursday he was gainfully employed as a web developer for a community college in Idaho. Recently, he voluntarily mentioned to his boss that he was concerned that seasonal affective disorder was harming his job performance, who mentioned it to his boss, who suggested in all good faith that Brent should talk to HR to see if they might help through their Employee Assistance Program. In Brent's words: "Instead, HR asked me a lot of pointed questions about when my performance could turn around and whether I wanted to work there, demanded that I come up with all the solutions (after I admitted that I was already out of brainpower and feeling intimidated), and then directed me to turn in my keys and go home, and that HR would call me on Monday to tell me the status of my employment." Now, at the end of the day Tuesday, they still haven't let him know what's happening, but it doesn't look good.
I think we can agree that this is some of the worst horseshit.
On the other hand, he's been wanting to get out of Idaho and into a city with an active rationalist community for a while, so in a sense this is an opportunity. Ways to help: Brent needs, in order of priority: a job, a place to stay, and funds to cover living and moving expenses-- details below. Signal boosts and messages of support are also helpful and appreciated. Ways NOT to help: Patronizing advice/other-optimizing (useful information is of course welcome), variations on 'cool story bro' (the facts here have been corroborated to my satisfaction with hard-to-fake evidence), disrespect in general.
1. Job: Leads and connections would help more than anything else. He's looking to end up, again, in a good-sized city with an active rationalist community. Candidates include the Bay Area, New York, Boston, Columbus, San Diego, maybe DC or Ann Arbor. He has an excessively complete resume here, but, in short: C#/.NET and SQL developer, also computer game development experience, tabletop board/card game design experience, graphic art and user interface experience, and some team leadership / management experience.
2. Crash space: If you are in one of the above cities, do you have/know of a place for a guy and his cat? How much will it cost, and when will it be available? Probably he'll ultimately want a roommate situation, but if you're willing to put him up for a short time that's also useful information.
3. Funds: Brent is not now in immediate danger of going hungry or homeless, but a couple of months will exhaust his savings, and (although it is hard to know in the current state of things) he has been told that the circumstances constitute "cause" sufficient to keep him from drawing unemployment. Moving will almost certainly cost more than he has on hand. There is a possible future in which he runs out of money stranded in Idaho, which would be not good.
If you feel moved to help, he has set up a gofundme account here. (The goal amount is set at his calculated maximum expenses, but any amount at all would help and be greatly appreciated-- he would have preferred not to set a funding goal at all.) Though Brent has pledged to eventually donate double the amount he raises to Effective Altruist causes, we wouldn't like you to confuse contributing here with charitable giving. Rather, you might want to give in order to show your appreciation for his writing, or to express your solidarity in the struggles and stigma around mental illness, or as a gesture of friendship and community, or just to purchase fuzzies. Also, you can make him do stuff on Youtube, you know, if you want.
Thank you so much for your time and kindness. -Elissa Fleming
Woah, well done everyone who donated so far. I made a small contribution. Moreover, to encourage others and increase the chance the pooled donations reach critical mass, I will top up my donation to 1% of whatever's been donated by others, up to at least $100 total from me. I encourage others to pledge similarly if you're also worrying about making a small donation or worrying the campaign won't reach critical mass.
If 102 people all pledge to donate 1% of everyone else's total, the consequences could be interesting. (Of course it's vanishingly unlikely. But pedantic donors might choose to word their pledges carefully.)
Is Austin on the list? I work at a not-evil tech startup called SchoolAdmin that does school admissions software for a mix of public/private/charter schools. We're not hiring devs right now, but that might possibly change since we have a product manager coming in October. The company is REALLY not evil; we've had three different people come down with mental or physical health issues, and the president's mantra has been 'your job is to get better' in every case.
I could possibly also offer a place to crash, I've got a futon, a study it could be moved to, and already have cats.
That narrative is unambiguously a case of illegal discrimination. Idaho law Defines:
and
I am also very confused as to how actual HR drones in an actual HR department wouldn't be familiar with the law and able to create a suitable enough pretext for termination.
I already mentioned the A.D.A. to Ialdabaoth, but fighting a discrimination case probably takes more money than he's looking to raise to move, as well as being psychologically exhausting.
Either of those reasons is probably enough to convince a rational person. The spirit of Immanuel Genovese still sits on my shoulder screaming "Passive complicity!" at /me/ every time I contemplate accepting an outcome in which it is normal that this kind of treatment happens.
Me too.
The problem is... this is a complex and delicate situation, as all real-life situations are.
There are co-workers who have gone the extra mile to help me and protect me. They didn't do everything they could, because they have families, and they know that if they rock the boat too hard it will be them, not HR, that get thrown overboard.
They aren't rationalists themselves (although I was slowly working on one of them), but they are caring and intelligent people who are themselves struggling to find meaning and stability in a harsh world.
If I could find a way to laser-lance out the demons of stupidity from my workplace, I would do so in an instant. If I could do so in a way that could add net funds to my own cause, I would already be doing so.
But as it is, I know exactly who would suffer for it.
(That doesn't mean that I have committed to a decision yet; I am still weighing necessary evils.)
I'm letting you know about what my shoulder angel/demon is shouting, because if I follow his advice I am not optimizing for giving you good advice.
You might be able to get a lawyer to work on a contingency basis - they only get paid if you win.
Official update: HR "explored every possible option" but "ultimately we have to move forward with your termination process" after "making certain there was unanimous consensus".
Apparently several people in my now ex-office are upset about this.
According to the efficient market hypothesis index funds should be the best way for the average person to gain a return from investment. Now there is a plethora of indices to invest in. How should one find the 'best' one?
Further, only a relatively small part of return generating assets are captured in publically tradeable assets. What about private equity and real estate, huge parts of the economy?
I an efficient market the expected value wouldn't be all that different between options, so base it on your risk management preferences.
Funds take a fraction of the earnings out, as management fees, and you want the fund that charges the lowest such fees. The early retirement blogs I read seem to agree on Vanguard being the best choice, at least in the US.
IIRC real estate prices in the US rise about 1% per year inflation adjusted while stock markets rise about 7 % on average. An average person needs a huge loan to invest in real estate and go all in which means zero spread of risk. Real estate is also relatively illiquid not only because of practical reasons but because the return of investment depends on timing of the transaction. You're shit out of luck if you need money while the price of your house is plummeting.
Depends on your risk tolerance. The bigger the index, the lower the risk and the lower the possible returns, generally. Also bigger index funds are usually more liquid. Transaction costs matter quite a lot unless you have a big lump sum to invest, and even then you should consider dollar cost averaging.
YRC. I thought you were forgetting to adjust the stock market returns for inflation, so I went to hunt for more accurate numbers, but apparently 1950-2009 S&P500 inflation-adjusted returns (counting not just price rise, but dividends) averaged to 7% per year.
Thanks. If you care about transaction costs you should probably invest in funds that reinvest dividends automatically.
That's not true. It's easy to get exposure to real estate through REITs. For example, through my wealthfront.com portfolio, I'm invested in Vanguard's US REIT ETF, VNQ.
I stand corrected.
I was thinking about anthropics after seeing some posts here about it. I read the series of posts on ADT including http://lesswrong.com/r/discussion/lw/8aw/anthropic_decision_theory_iv_solving_selfish_and/, and EY's posts http://lesswrong.com/lw/17c/outlawing_anthropics_an_updateless_dilemma/, http://lesswrong.com/lw/19d/the_anthropic_trilemma/, and http://lesswrong.com/lw/17d/forcing_anthropics_boltzmann_brains/. I had a few questions about those posts.
First, how is average utilitarian defined in a non-circular way? I'm trying to wrap my head around why I don't agree with the conclusions of the first post I linked, and it seems to come down to not understanding average utilitarians.
More specifically, do they define two levels of utility? Or do they exclude themselves from the calculation? I thought it was just a different way of allocating your own utility, but how do you calculate which way will give you the most utility by giving the world a greater average utility, without knowing the answer of your own utility to plug in?
Second, in http://lesswrong.com/lw/19d/the_anthropic_trilemma/ EY ended off with
Has he been officially "impressed" yet? Should I read any specific attempts to solve the trilemma? What reading can I do on anthropics to get an idea of the major ideas in the field?
It seems to me that SIA, in the way it's been applied, is obviously correct, and in general I feel like I have very clear intuitions on these kind of problems. I plan on writing up something eventually, after I understand the argument against my point-of-view to argue coherently.
If you can quantify a proto-utility across some set of moral patients (i.e. some thing that is measurable for each thing/person we care about), then you can then call your utility the average of proto-utility over moral patients. For example, you could define your set of moral patients to be the set of humans, and each human's proto-utility to be the amount of money they have, then average by summing the money and dividing by the number of humans.
I don't necessarily endorse that approach, of course.
I think Eliezer says he's still confused about anthropics.
So far as I know, Nick Bostrom's book is the orthodox foremost work in the field. You can read it immediately for free here. Personally, I would guess that absorbing UDT and updateless thinking is the best marginal thing you can do to make progress on anthropics, but that's probably not even a majority opinion on LW, let alone among anthropics scholars.
A friend and I hope to host a MIRIxVancouver workshop in Vancouver, Canada sometime in October. We haven't filed an application with MIRI yet, and we haven't set a date, so there's no schedule yet. So, this is just a shout-out to anyone who might want to get involved in it over a weekend, including if you want to visit from Seattle, or Oregon, or anywhere nearby. Comment below, or send me a PM, if you're interested in attending.
Vancouver has enough of a diversity of people interested in the MIRI who can host their own friends that I believe it will make sense to host multiple different MIRIx workshops. Like, I find the MIRI very interesting, but I want to grasp technically what it's about. Another one of my friends is interested in the philosophy of A.I., and yet another friend is a former MIRI intern who will invite a bunch of his friends from the university math department over. So, I'll likely host workshops at different levels of depth, or with different topics.
In the interest of trying out stuff outside the usual sphere-of-things-that-I'm-doing, I now have a fashion/lifestyle blog.
It's in Finnish, but it has a bunch of pictures of me, which ought to be language-neutral. Also my stuffed animals. (And yes, I know that I need a better camera.)
Donation sent. !@#% those !@#&!.
EDIT: Oops, wrong place, this was supposed to go under ITakeBets' post.
I should probably update this prediction. Considering Yudkowsky's recent pwnedness and pieces like this becoming common it is at least 10%.
Moldbug in 2008.
I'm dubious about borders getting opened that soon, considering how long it's taking to make moderate moves towards drug legalization.
As a counterpoint look at the rise of anti-immigration movements in Europe and e.g. the success of Marine Le Pen in France.
Two billion is a crazy population prediction even if open borders was enacted. Relative quality of life would decline very quickly with open borders, and the immigration level would slow down dramatically.
I also read some estimate that "only" 500 Million people world wide want to immigrate to the US. Overall I expect the quality-of-life gap between USA and the 3rd world to continue declining over the next couple decades.
I'll bet a thousand 2014 dollars at even odds that it's less than a quarter of that.
(Clarification: I'm talking about the population of the US, as per the PredictionBook entry, not of North America as per Moldbug's quote; the population of the latter is already over 500 million. Should probably stipulate present borders too, just in case. 2 billion isn't credible in any case, though.)
Is this post just an excuse for an NRxic quote? And what does the "Yudkowsky's recent pwnedness" swipe refer to?
A little help communicating some ideas?
Anyone up for beta reading a 2,000 word section of my attempt at an aspiring-rationalist story, S.I.?
I've just finished putting together an initial draft of Bunny pontificating about the ideas discussed in https://www.reddit.com/r/rational/comments/2g09xh/bstqrsthsf_factchecking_some_quantum_math/ . I could really use some feedback to make sure I'm having her explain them in a way that's actually comprehensible to the reader. Anyone who'd like to help me with this, I've pasted the initial draft to a GoogleDoc at https://docs.google.com/document/d/1lOQAAM3fdnF2ew7CgBQqSLtk21_4Foa8n7IB3Ay0ze8/edit?usp=sharing , which is set to allow comments.
I've never been entirely sure about the whole "it should all add up to normality" thing in regards to MWI. Like, in particular, I worry about the notion of intrusive thoughts. A good 30% of the time I ride the subway I have some sort of weak intrusive thought about jumping in front of the train (I hope it goes without saying that I am very much not suicidal). And since accepting MWI as being reasonably likely to be true, I've worried that just having these intrusive thoughts might increase the measure of those worlds where the intrusive thoughts become reality. And then I worry that having that thought will even further increase the measure of such worlds. And then I worry...well, then it usually tapers off, because I'm pretty good at controlling runaway thought processes. But my point is...I didn't have these kinds of thoughts before I learned about MWI, and that sort of seems like a real difference. How does it all add up to normality, exactly?