Open thread, Dec. 15 - Dec. 21, 2014
If it's worth saying, but not worth its own post (even in Discussion), then it goes here.
Notes for future OT posters:
1. Please add the 'open_thread' tag.
2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)
3. Open Threads should be posted in Discussion, and not Main.
4. Open Threads should start on Monday, and end on Sunday.
Loading…
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)
Comments (309)
I am looking to set a morning routine for myself and wanted to hear if you have some unusual component in your morning routine other people might benefit from.
One thing I might start experimenting with is a version of morning contemplation. Ancient stoicism seems to suggest to reflect on one's principles in the morning, christian tradition has morning prayers and Benjamin Franklin reviewed his virtues every morning, so why not do a little personalised version of it? Things like the serenity prayer or Tarski's litany.
I have a terrible problem where I wake up from my alarm, turn off the alarm, then go back to sleep (I've missed several morning lectures this way). The solution I've been trialing is to put a glass of water and some caffeine pills on my bedside table when I go to sleep. That way, when I wake up I can turn off the alarm, take the pill and give in to the urge to put my head back on the pillow, confident that the caffeine will wake me up again a few minutes later. This has worked every time I've remembered to put out the pills.
I got this idea from someone else on LW but I've forgotten who, so credit to whomever it was.
My preferred solution for this problem is to have the alarm on the other side of the room so I have to actively get out of bed and walk over to it.
I tried that before, I'd just turn it off and get back into bed.
Hmm, here's another solution which a friend of mine tried which seems to require careful timing- set multiple alarms and have each alarm set to go off a few seconds after the other, having them lead you all the way around the room.
In general, I dislike posting comments that just express agreement without any content, but this is such an great idea that I just had to pop in and say, "That's awesome." So, without further ado:
That's awesome.
It also happened to me. I used to wake up at 7:15, snoozing up until 7:30, which was very bad for my morning routine. Putting the alarm clock on the other side of the room helped, but when I tried to set the waking time at 7:00, I would just get up, snooze and go back to bed. Now I'm trying to set it to 7:10 and it seems to work well.
There probably is a subconscious set point which needs to be adjusted gradually. You also need stuff to do when you wake up: if you wake up just for the sake of it, your subconscious will know and put you back to sleep.
I would too, so I replaced my alarm with a cron job on my computer to play music at a certain time -- usually an Insane Clown Posse song (specifically Vultures, which starts loudly and distinctively enough to work well structurally as an alarm), to make sure I wouldn't stay in bed and listen.
The hack is due to Anders Sandberg, with modafinil tablets though [ETA: this last part is false, see Kaj's reply]. Works wonderfully (whether with modafinil or caffeine).
Historical note: his original blog post on it specifically used caffeine pills.
I set two alarms, about 1-2 minutes apart. The first one is soft and beside my bed, the second is loud and on the other side of the room. I get a couple minutes to achieve consciousness in a civilized fashion, and then the second alarm actually gets me out of bed. I've previously tried just having the second, but I got very good very fast at turning it off without really gaining consciousness.
I've cultivated a habit of following to the letter any deliberate precommitments I make to myself, because it seemed like a useful sort of habit, and I haven't failed yet. I might even choose to make that part of my identity in time...
With that already in place for a few months, over the last few weeks I've noticed that I have woken (and stayed) up in time iff I went to sleep in time (about half the time) or used the precommitment of, freely translated, "When I next wake up, if I notice that my ringing mobile phone has the piece of duct tape stuck to it that I put there to trigger the memory of this precommitment, and remember this precommitment, I will not touch my mattress after turning it off for the following 10 minutes." (about half of the remaining times)
You see, my morning self doesn't actually want to sleep in, it just apparently hasn't internalized induction on enough of a gut level to draw the connection between going back to bed and falling back asleep. What it does understand, is the connection between touching that mattress and tarnishing that record.
Edit: It now occurs to me that it might be a nonobvious warning that you carefully start with lesser precommitments, lest you ruin this forever for yourself by having an akrasia'd self not care enough about the habit. Your reach exceeds your grasp by default on this one.
Edit: Shit, I actually just went back to bed after those 10 minutes. Gonna have to modify that...
I think the best thing is to start implementing it as soon as you get up, to not give yourself time/opportunity to sabotage it. Also, get up as soon as you wake up.
This may be the biggest causative factor of how my day goes (though it is itself subject to causes such as when I went to bed).
Don't neglect this.
I exercise for 5 minutes within 5 minutes of getting up.
Before I did this, I sometimes had the habit of checking e-mail for like an hour before getting out of bed. After adding this to my routine, I never did that again.
Yes, when I was in high school, I would do about 20 push ups as soon as I got out of bed. After that, my heartrate is a bit too high to go back to bed. Of course, after a while the 20 push ups would get easy so I would increase the number until I was doing about 50-70.
I have gotten myself a manual coffee grinder. Using this device it takes me about 4 minutes to grind enough beans for a large Melitta pot. Attending to this repetitive task gives me a small but regular block of time during which I find it easy to practice mindfullness. And when I am done, I make a pot of coffee.
I forget who suggested it, but I actually needed 2 alarm clocks to get my sleep schedule under control.
One wakes me up. It is across the room, on my computer chair. When I goes off I move it up to the shelf. Something about doing a task makes me loathe to go back to bed.
One makes me go to sleep. When it goes off, I stand up and put it on the computer chair. That way I can't keep browsing the net, because I'm stuck standing up.
This was only necessary for like a week. Once I was going to bed at the same time every day I didn't need alarms to wake up then. Nowadays I wake up almost automatically.
The key for me was going to bed at a set hour every day.
I get my shower the night to keep my bed clean and reduce morning prep time, allowing me to wake up later. Then just wet my hair with water in the morning for combing. I also have an Ensure as soon as I wake up, even if I'm going to have a more substantial breakfast shortly thereafter, because otherwise I'm a zombie.
Shower first. Then make breakfast/coffee.
One part is writing down whatever dreams I can remember right upon awaking. This has led to me occasionally experiencing lucid dreams without really trying.
Also since I am writing dreams anyway, this makes it easy to do the other writing which I find beneficial. Namely, writing the major plan of the day and gratitude stuff.
I don't know how unusual these are, but some of the components of my routine are:
- putting the alarm on the other side of the room, so I get sure I don't snooze;
- workout with a 10 min torcher;
- cold shower after workout;
- a brief five minutes meditation before leaving.
You might be interested in this, if you haven't already seen it.
[Meta] I often see threads like this where people recommend things that require a very high level of conscientiousness or planning ability to start with, (e.g. if you are tired in the mornings get out of bed immediately and do x, requirs you to be capable of forcing yourself to do x when you are tired.)
I think you are mistaken about what is easy and difficult. Most of these are about dealing with lack of willpower, suggesting that the authors found something that was easier than it looked. Most of them don't compare before and after, but jsteinhardt's does.
In the newest episode of Person of Interest there is explicit mention of Friendly Artificial Intelligence, and the main character seems have an attittude towards AI similar to the one found on LessWrong. The exact quote is: "Even if I had succeeded in creating a benevolent machine, as if any such thing could exist, never forget that even a so-called 'Friendly Artificial Super Intelligence' would be every bit as dangerous as an unfriendly one."
Since the LessWrongWiki is the second google hit (at least for me) under both "Friendly Artificial Intelligence" and "Friendly Artificial Super Intelligence" (the first being this wikipedia page, which explicitly mentions Eliezer in the second sentence) this might lead to an increased amount of visitors? Also it should be noted that in this series almost everything said about artificial intelligence is compatible with the ideas found on LessWrong (sometimes tweaked a little bit to make the series more interesting) - this series might well promote some beliefs on AI from LessWrong to the attention of a broader audience.
Yeah, I watch this series and this season in particular makes me feel like the writers read LessWrong.
Like any technical subject I'm familiar with, I get a little weirded out by the treatment it receives in the mass media as there's always little (or big) bits that are just dumb for the stories sake, but Person of Interest seems to do a better job talking about AI than most shows do talking about whatever other technical subject.
What magazines would a LW member probably want to read/subscribe to?
My subscriptions: GQ Vogue Popular Science Time New York Times Vanity Fair Rolling Stone
For a family gift exchange game, I am giving a gift of the form: "25 dollars will be donated to a charity of you choice from this list"
Please help me form that list. My goal is to make the list feel as diverse as possible, so it feels like there is a meaningful decision, without sacrificing very much effectiveness.
My current plan is to take the 8 charities on givewell's top charities list, and remove a couple that have the same missions as other charities on the list. Maybe adding MIRI or other Xrisk charities (give me ideas) that are very difficult to compare effectiveness with the givewell charities.
What would you put on the list?
I find it annoying that there is no way to view all the posts at once -- you can see either main or discussion posts, but not both at once. I think it would be good to add "All" next to the tabs for Main and Discussion.
There is an undocumented way: all posts, all comments. But those URLs are not linked from anywhere.
Thanks! I'll just add those to the browser as bookmarks.
Is there a way to get my LW notifications to be forwarded to my email?
I use Feedly + IFTTT for things like this. Try here or here.
How do you access the inbox RSS feed? As far as I can tell there's no way for the feed reader to authenticate to access your LW account.
You may be right. I don't use Feedly for that particular RSS feed, just similar feeds.
For those who are intending to have children A recent study shows a substantial drop in IQ of children who were exposed to phthalates in the womb, and this pattern lasts at least until age 7. This is the first such study, so it may not turn out this way, but it seems worth noting, and it underscores how much low-hanging fruit there may still be in terms of environmental impacts on intelligence (similar to what happened with iodine in the 1920s and 30s).
Do we have a catalog of Not Less Wrong rationality guides?
I know we have the list of rationality blogs, but I'm asking about a collection of material that educates at an entry level of formalized rationality but sits at lower inferential distances that the sequences.
I'm not aware of such a catalog but there are definitely resources that fit into this category. One good one is You Are Not So Smart.
I haven't looked into it much myself, but a couple of people have mentioned RibbonFarm as being something like that.
When you go to GiveWell's Donate page, one of the questions is,
And you can choose the options:
Grants to recommended charities
Unrestricted donation
I notice I'm reluctant to pick "Unrestricted," fearing my donation might be "wasted" on GiveWell's operations, instead of going right to the charity. But that seems kind of strange. Choosing "Unrestricted" gives GiveWell strictly more options than choosing "Grants to recommended charities" because "Unrestricted" allows them to use the money either for their own operations, or just send it to the charities anyway. So as long as I trust GiveWell's decision-making process, "Unrestricted" is the best choice. And I presumably do trust GiveWell's decision-making, since I'm giving away some money based on their say-so. But I'm nevertheless inclined to hit "Grants to recommended charities," despite, like, mathematical proof that that's not the best option.
Can we talk about this a little? How can I get less confused?
I wouldn't make a restricted donation to a charity unless there was a cause I really cared about but I didn't think the charity behind it was well-run and I didn't know a better way of helping that cause.
I do not consider money to keep a good charity running as "wasted"--if anything I am deeply dubious of any charity which claims to have minimal to no administration costs, because it's either untrue (the resources to manage it effectively must come from somewhere, maybe from the founders' own personal resources) or a likely sign of bad management (they think that skimping on the funds needed to manage it effectively in the name of maximizing the basket of "program expenses" is a good organizational strategy). An organization that I think is well-run wants to spend on its cause as much as possible, but is mindful of needing to spend on itself also. If it cannot spend on itself--to hire good staff, to have good training, to use resources that cost money and save time, to plan its strategy and maintain regulatory compliance, to do whatever else an efficient organization needs to do--how can it possibly have the capacity to spend well on its programs? The money to sustain that charity is providing for its cause to be effectively addressed now and into the future.
"Unrestricted" says that you believe GiveWell is competent to make these allocations correctly between itself and its recommended charities. For GiveWell in particular, if you do not believe they can do this, why do you think they can evaluate other charities' effectiveness? Presumably you want to give to the other charities because GiveWell has told you they are worth it, because you think GiveWell is competent at assessing organizational effectiveness. (For other charities, I would have lower expectations for assessment ability--but still I expect that I want to give to one in particular because it is effective at spending for its cause. There are few causes where you do not have much choice of how to direct your money to affect it. An effective one will be competent at running itself--not perfect surely, but competent enough that I don't think I will do a better job at allocating its funds than it will by giving a restricted donation.)
Also, many people's gut feelings direct them to give restricted donations to avoid "wasting" their money; it's a feel-good option but one that does not help the charity stay around in the long term. People who are more considered should compensate for that by allowing the charity to use their funds unrestricted. I have no idea if GiveWell gets grants or not, but grant support from foundations is often restricted as well; it's much harder to get grants for general operating support. But I won't start that rant here.
(For background, I've been heavily involved in nonprofits for the past 10 years, as volunteer, staff, and board.)
Yeah, I think that's right. I'm the same as people who don't want to give to charities who have too much "overhead," leading to perverse incentives, as you say. GiveWell itself can be looked at as overhead for the charities it recommends, even though technically it's a different organization. As such they deserve to be supported too.
Will click "Unrestricted" in the future.
There's a big difference between trusting someone about a third party and trusting someone about themselves.
Maybe this would be a coherent position:
I'm not sure I'd buy it myself... it seems like it really only makes sense if you don't think anybody else is going to be giving money to GiveWell in the immediate future either (or perhaps ever?).
You could also just think that GiveWell doesn't currently have as much room for more funding as the recommended charities do, even though GiveWell may disagree with that assessment.
Holden has written about donation restrictions on the GiveWell blog back in 2009 (bold and italics in original):
See also the following on the GiveWell blog:
I think the nasty part of the Hard Problem of Consciousness is probably in finding a naturalistic explanation for how things come seem subjectively objective: for why the wavelength of red feels from the inside like a built-in quality of the world rather than a perception generated by a mind in response to a stimulus. I think the "social processing theory of consciousness" doesn't quite explain this, at least not to my satisfaction.
Of course, the random thoughts I record in Open Thread are not liable to be high-quality.
For any agent, self-reflection has to bottom out somewhere, since working memory and cognitive capacity are finite.
That said, some meditation practitioners report being able to notice "the arising and passing away" of individual sensations, so it may be that this is just a matter of training rather than an essential feature of consciousness.
That insight would be an additional layer of processing requiring energy that could be used for other things. How would such an insight increase the brain's reproductive success? We don't have insight to our outputs either, you don't intuitively understand how your muscles work in response to stimuli, which is part of the reason why it feels like we have free will and are agents in the world instead of parts of the world.
The true quality of stimuli is already lost and grossly simplified at receptor level, in the case of light the photoreceptive rod and cone cells of the retina.
Is there some resource (for instance, in the LW sequences) where I can redirect people to get a quick, clean view into the Bayesian worldview of doing things, especially in science? When I read people say things like "Consensus doesn't matter in science!" I want to respond with "Well, consensus isn't everything, but being informed about the agreement in opinion of a large number of authorities in a subject should make you update your beliefs" but I find it hard to do that without then having to explain what "update your beliefs" actually means.
People don't change their views about fundamental issues such as how they deal with knowledge quickly.
As far as articles go, Yvain's article http://slatestarcodex.com/2013/08/06/on-first-looking-into-chapmans-pop-bayesianism/ is good.
"Change your mind"? "Take that on board"?
I think "change your mind" is too specific; in common parlance it's used only for major changes, such as going from thinking "probably X" to "probably not-X".
I note that in my profile I can see posts that I have up or downvoted under "Liked" and "Disliked." Is there a way to get a similar list of comments that I've up and downvoted?
There's lots of material here in comment threads that I think is worth remembering but is tricky to re-find.
Thought: I think Pascal's Mugging can't harm boundedly rational agents. If an agent is bounded in its computing power, then what it ought to do is draw some bounded number of samples from its mixture model of possible worlds, and then evaluate the expected value of its actions in the sample rather than across the entire mixture. As the available computing power approaches infinity, the sample size approaches infinity, and the sample more closely resembles the true distribution, thus causing the expected utility calculation to approach the true expected utility across the infinite ensemble of possible worlds. But, as long as we employ a finite sample, the more-probable worlds are so overwhelmingly more likely to be sampled that the boundedly rational agent will never waste its finite computing power on Pascal's Muggings: it will spend more computing power examining the possibility that it has spontaneously come into existence as a consequence of an Infinite Improbability Drive being ignited in its near vicinity than on true Muggings.
There are other ways of taking Pascal's mugging into account. You shouldn't do that based on lack of computing power. And if you aren't doing it based on lack of computing power, why involve randomness at all? Why not work out what an agent would probably do after N samples, or something like that?
When turning in their final papers by email, a few of my students added positive comments about my class such as "Thanks for a great semester, I really enjoyed this class". All of my students will do an anonymous (to me) evaluation of my class. I imagine that both of these occurrences are commonplace among college students. It would be interesting to see if students who complement the class also give high anonymous evaluations, and this might tell us something about the honesty of praise. This would be an easy study for a college to conduct where a large group of professors tells a researcher which students complemented them, and then the researcher accesses the student evaluations, and then calculates correlations.
Norms of politeness will be a confounding variable.
This is an interesting idea, but it may be very difficult to do in practice. At many major universities the students' evaluations are fully anonymous-their names are never on them at all. And I suspect that the universities may be uncomfortable changing that even for a few classes just to help answer this question.
Thinking about a quote from HPMOR (the podcast is quite good, if anyone was interested):
...
Besides the quoted "Chimpanzee Politics" are there any other references to this hypothesis? I've tried Googling around for 5 minutes and I couldn't find anything.
Edit: seems like I was looking using the wrong keywords: Wikipedia seems to have a small paragraph on evolution of human brain due to competitive social behavior, but I'd still like to see if anyone else had any articles on the matter.
I believe one phrase for it is the Machiavellian Intelligence Hypothesis
Yet many of the really smart humans seem to have trouble with social interactions. Dark Triad people who excel at getting others to do their bidding might have higher than average IQ's, but they generally don't go into STEM fields where having high IQ's can pay off.
If you're smart and "excel at getting others to do [your] bidding", you don't want to go into STEM, you want to go into management.
STEM pays you far less for being smart and manipulative than say, being a good lawyer or salesman.
Not exactly a match, but Others in Mind: Social Origins of Self-Consciousness makes a variant of this claim, putting an arms race concerning prediction of third-party reactions to hypothetical actions as the driving force of human self-awareness.
I'm considering trying out a paleo diet. I'm not totally convinced by all the arguments for it, but are there arguments against paleo that say it is actually bad for you? By "bad" I mean are there worthwhile arguments that switching to paleo is worse for your weight and longevity than your typical American diet?
As far as long-term consequences go, my ability to digest carbs without getting gassy or stomach achey seems to have been seriously damaged by my diet.
I'm afraid that it largely depends on your initial diet. For me, paleo was a net disaster (3 months with quality of life substantially reduced, and 3 kgs gained), but probably because I came from an already healthier diet.
If you factor in that in a high proteic diet you have higher than normal insuline spikes and you may lose the ability to process gluten, then you might actually gain some benefit.
In my admittedly small social circle, I've seen though more people benifit from a vegetarian diet than from paleo.
Agreed. Paleo is vastly superior to SAD (standard American diet), but it's an open scientific question whether paleo is better than other common diets. (I'm paleo.)
If you do low carbohydrate paleo there can be a difficult transition period (see Atkins Flu).
Is this because there's no reason to think that it should be better or because there are contrasting data?
Because of the limits of nutritional science where they can't run long randomized experiments on people. The theoretical case for paleo is excellent.
Could you provide some solid evidence? I've never found something that didn't crumble at the first investigation.
I can't back up any of this with solid citations but: If our ancestors have been eating a food for a very long time that's Bayesian evidence that the food is safe. We have been eating meat for so long that it seems likely parts of us are dependent on stuff we can get only from meat. Cancer, heart disease, and strokes seem to be mostly diseases of civilization that were relatively rare among hunter-gatherers who ate their traditional diets. Things go really badly for hunter-gatherers who switch from their traditional diets to modern diets. Wheat is cheap to grow so even if it is unhealthy it's understandable that it would be widely consumed. It's also understandable that sugar, being a superstimulus would be widely consumed even if it is unhealthy. Lots of people who try paleo succeed in loosing weight. The modern obesity epidemic shows something is very wrong with SAD (Standard American Diet) and paleo offers a tried and true safe harbor.
Cancer and heart disease are diseases of longevity. Why expect paleo to help with them when there's every reason to believe longevity wasn't a part of that environment?
I don't have data at hand, but I think that's true only partially. Yes, the prevalence of cancer and CVD is a function of the age of the population, but as far as I remember, even after you control for age, they still show up as diseases of civilization with the "primitive" societies having considerably lesser age-adjusted rates.
At least one causal pathway for that is visible: diabetes and the metabolic syndrome in general are clearly diseases of civilization and they are strong risk factors for CVD (I don't know about cancer).
While I don't have the stats, I think that 50,000 years ago if you lived to 30, you had a reasonable chance of living to 70, and cancer and heart disease kill lots of people under 70.
"Longevity Among Hunter-Gatherers: A Cross-Cultural Examination", Gurven & Kaplan 2007; might be helpful.
When I did my paleo experiment, I wasn't really eating any more protein than before. My carbs were way down though, and my fat intake way up. I wonder if this may explain some of the difference in our outcomes.
Huh? Gluten sensitivity is an autoimmune problem, you don't acquire it by not eating wheat.
People who have done paleo can sometimes have issues processes carbs because you lose the stomach bacteria necessary to do so. It was poorly phrased. The transition back can be hard.
Even then, it's not like you "lose the ability" -- the gut microbiota changes fairly rapidly so you should be fine in a few days. But I agree that transitions between very different restrictive diets can be hard on the body.
This runs contrary to what I've ready, but not specifically on gluten-related microbiota. Apparently change in the overall ecology of gut microbes can be very difficult to recover from.
Here is some data on the ease of changing one's microbiota. However you have a valid point in that the persistence of particular kinds of microbiota is not well understood and evidently it's possible to fall into, um, local minima that are hard to get out of (thus the whole fecal transplant business).
I suspect the reality here is much more complicated than the simple "easy to change"/"hard to change" approach.
AI researcher with good academic credentials is wanted. If you are interested, send me a private message, please.
MIRI was mentioned in today's econtalk podcast on AI. Just in case anyone is interested.
Link to the podcast, with transcript.
The mention of MIRI, about (bad) AI forecasts :
I was unsure where to post this question; let me know whether it merits its own thread. I also apologize if this post is a bit messy.
If I had to title this post, I might name it, "Optimizing College Activities for a Future Programmer".
I'm a college student at an American school. It's quite a lot of work -- more than I can do in the time given, and I have a study routine that's more efficient than a lot of people I know. I was handling it relatively well last year and still getting enough sleep, exercise, socializing, etc. -- basically all the things I would consider essential for keeping me sane.
I do not do drugs. I do not watch television or movies. I am vegetarian. My room is not decorated and I do not buy expensive items. My socializing thus far has consisted of talking with people over meals while walking around campus. I am not in a relationship. I spend most of my time studying and doing school assignments. I have a relatively good GPA and have worked hard to maintain it. But the work is getting harder and I'm thinking I'll need to start putting less work into my classes and accept a lower GPA, because I cannot compromise the essentials (meditation, sleep, etc.). It's been too stressful to do as much coursework as I've been doing and to skip the essentials.
I plan to pursue a career in software engineering / outside academia. I'm double-majoring in math and CS. I do not plan to get a master's degree or a PhD (at least, not any time soon). I understand that CS students' grades don't matter much, though I do think I can benefit from doing as well as I can in my classes. (But I'm also willing to work less in college to be happy.) I also have a great coding job that I've been neglecting because of my studies, but I don't want to neglect it any longer.
Some questions:
Should I let my grades drop a bit and instead work my coding job and ensure I'm doing the "life essentials" on a daily basis? I will be replacing some of my academic work with programming, which is in my estimation pretty valuable. I'm on a scholarship and it requires that I maintain at least a 2.0 GPA, but I'm quite confident my GPA isn't in danger of dropping that low.
Should I put more emphasis on socializing and forming a network? I don't use Facebook much, and my interests don't intersect with most of my peers' (see the partying / drinking / buying expensive food stuff above). I'd rather spend time with people who are doing interesting things and who I can relate more with (is this bad?), but I'm having trouble finding such people on my campus. How do other "rationalists" form social networks in these kinds of environments, or do they?.. I don't want to miss out on something essential (like developing social skills and / or a network) if it is actually essential. (To be fair, I am a bit awkward and often find I don't have anything to say to my peers, but I think this is again because my interests differ from others'... But maybe I'm wrong.)
How should one pursue the development of social skills? How much time should one put into it, vs. into coding, studying, etc.? Based on what I've read, being friendly and someone people can get along with and want to spend time with can work wonders in all sorts of circumstances. Relative to my peers, I've put less time into meeting and hanging out with people, and I think I'm less socially adept. I'd love to improve. Any suggestions?
On a related note, how can I find people with whom I'm compatible? I frequently run into people I don't want to spend time with, but rarely do I meet people whose presence electrifies me.
Any other general advice? e.g., I haven't read anything outside of class since the summer, and I'm thinking it would be good to read during the semester.
I have two pieces of advice for you. Please take them with a grain of salt - this is merely my opinion and I am by no means an expert in the matter. Note that I can't really recommend that you do things one way or another, but I thought I would bring up some points that could be salient.
1) When thinking about the coding job, don't put a lot of emphasis on the monetary component unless you seriously need the money. You are probably earning less than you would be in a full time job, and your time is really valuable at the moment. On the other hand, if you need the money immediately or are interested in the job primarily because of networking opportunities or career advancement, then it is a different matter.
2) Keeping up a good GPA is not equivalent to learning the material well. There are certainly corners you could cut which would reduce the amount of work you need to do without losing much of the educational benefit. As the saying goes, 20% of the effort gives 80% of the results. If you are pressed for time, you may need to accept that some of your work will have to be "good enough" and not your personal best. Having said that, be very careful here, cause this is also an easy way to undermine yourself. "This isn't really the important stuff" is a fully general excuse.
Thanks for the response.
Re: 1) I'm not as focused on the money as on the programming opportunities it might later lead to.
Re: 2) I agree with everything here. What do you mean in your last sentence?
I'm having trouble finding the original sequence post that mentions it, but a "fully general excuse" refers to an excuse that can be applied to anything, independently of the truth value of the thing. In this case, what I mean is that "this isn't really the important stuff" can sound reasonable even when applied to the stuff that actually is important (especially if you don't think about it too long). It follows that if you accept that as a valid excuse but don't keep an eye on your behavior, you may find yourself labeling whatever you don't want to do at the moment as "not really important" - which leads to important work not getting done.
The post is "Knowing About Biases Can Hurt People". See also the wiki page on fully general counterarguments.
Thank you! That is exactly what I was looking for.
Before you do anything else, reconsider your class schedule. A higher GPA will probably mean more to your future career prospects than keeping the math major. Also, balance your schedule so you have a mix of lower-work/more-gently-graded classes and harder classes every semester.
Consider dropping the job depending on the criteria ChaosMote noted.
Your lifestyle sounds pretty sterile. You should make an effort to socialize more. College can be a place you make friends who last a lifetime, if you put some effort in.
What leads you to think this?
What kind of effort?
One major is enough; many companies look at GPA and a low GPA can rule you out but no second major won't. The GPA/more classes pareto curve is also usually more favorable towards one major. But if it's a very small commitment for OP my advice doesn't stand.
Talk to people, be friendly, stay in touch, initiate social activities, be a good friend.
I think you're missing,
All your suggestion are good, but it's not worthwhile to put lots of energy in building friendships with people that you don't like.
I wound up not liking anyone I met at the University of Tulsa. But then I had to transfer there after three semesters at a real university - Washington University in St. Louis. Imagine going from just below Ivy League to a place which, according to a Wikipedia page, has only one notable alum with a STEM degree from there, compared to dozens of alumni in areas like entertainment, business and sports:
http://en.wikipedia.org/wiki/List_of_University_of_Tulsa_people
Versus:
http://en.wikipedia.org/wiki/List_of_Washington_University_alumni
A high GPA is somewhat helpful when you're looking for your first full-time programming job, but probably not as helpful as how prestigious a school you went to, and definitely not as helpful as industry experience or actual nuts-and-bolts engineering skill. This latter is not to be underestimated; a lot of new CS grads can't quickly write accurate pseudocode, for example, and you will be asked to do that at some point during any half-decent interview. If you can't do it, that's going to be a deal-breaker, but having a 3.2 instead of a 3.5 GPA almost certainly won't be.
After a couple years of full-time experience, almost no one will ask about your GPA (they will, however, ask where you went to school and what degree you got), meaning that GPA is important to your long-term career prospects almost entirely insofar as it affects your standing right out of college. It'll also affect your prospects for grad school, if you're interested in going that direction at some point.
Could you join a fraternity? Best decision I made while in college.
I'm unsure this is the right decision for me, given that I don't particularly enjoy partying or drinking. Why did you join a frat? What did you get out of it?
Also the German name suggests European location which means that fraternities are pretty much dead around here.
If he thinks that his workload will ramp up, now isn't a great time to join, because he'll have to spend a semester going through the pledge process.
I do highly recommend fraternities otherwise though - they're a great network and place to meet friends, a good place to practice and develop social skills when interacting with the rest of the Greek system, have a lot of people focused on self-improvement, and they often take care of a lot of the overhead in college (by providing meals, laundry machines, a gym).
This being a good decision is of course contingent upon you actually liking the people in the fraternity as well as the particular attitude of the house in general. And if you're not into the party scene or nights out on the town, it may not be a good cultural fit.
In the same way that we have other periodic threads, how much of a good or a bad idea would it be to have a periodically posted thread where we could post our recent rationality failures?
I recently posted a discussion post on a similar topic, albeit for archival reasons, rather than regular community support. Either way, I'd encourage you try it out. There seems to be a cautious attitude on LessWrong that if one asks about starting a regular or periodic type of thread, a consensus will come forth giving great reasons either for or against the thread's institution. This doesn't really seem to happen. If you start this thread, include for the first one a note like 'If this is used lots, or well-received, then we can make it a monthly thread". If it gets lots of comments, or gets lots of upvotes, or even gets a few upvotes with a decent upvote:downvote ratio, I figure that's grounds for posting the thread a second time. The group rationality diary encourages people to report mistakes and failure modes, but trying a thread explicitly and only about reporting failures and mistakes might work.
I wouldn't mind a new monthly thread for that topic, but I think it's (in theory) already covered by the Group Rationality Diary series, at least in the sense of trying to do something rationally and failing. There's also the Mistakes repository for major life mistakes.
Predictionbook has been linked to and discussed here before. I'm one of the (few) active users, and I'm curious why more people who are regulars here don't use it or don't use it frequently. People who don't use Predictionbook, why don't you? Part of why I am curious is that if there are interface or similar issues then now might be a good time to speak up since Jayson Virissimo is working on a similar service here.
I'm not really sure how to answer this. Predictionbook is just one of thousands of websites I don't use.
I use predition book from time to time.
A huge issue of why I'm not using it more is that it's often hard to find interesting predictions to vote on. The Recent prediction page for example doesn't allow you to show a second page.
It doesn't seem to be possible to simply look at all the predictions on Tag X. I think the way stack-exchange works with tags functions quite well.
I think commenting should be separated from individual prediction making. There's no reason why the website shouldn't allow full threaded comments.
Disclaimer: This application was designed for the sole purpose of helping me obtain employment in software development and is not yet production ready. Use at your own risk (of data loss)!
Still, having said that: let me know if you have any ideas.
To judge a prediction on your site at the moment I have to:
1) Click on the prediction in the list.
2) Click on the text field which contains the number
3) Switch to keyboard to type in the number and enter to affirm the prediction
4) Use the mouse again to get back to the prediction list
The second point could be saved by autofocusing on that field. You might also think of a way to provide buttons to allow people to enter data via mouse clicks.
It would also be nice to have a "next prediction"-button on prediction page to make it easier to enter a bunch of predictions.
I tried to create a prediction and the free text field was confusing. A calender view would be a lot better. When I typed in "355 days" and it gave me a prediction that supposed to be juged in 2 hours and no way to edit it.
I used to use prediction book, but stopped because the interface was awful and the site was full of bad predictions that should have been private but weren't (e.g., I'm going to pass my exam tomorrow -- just like that, without any context).
If only publicly verifiable/falsifiable predictions were permitted, would you have stuck around?
I think the issue is not about allowing/disallowing such predictions but having good tagging that tags all those predictions with "private life" and then the option of people to block that tag the way you can block tags on stackexchange.
Yes, one of the most common issues people have brought up with PredictionBook is the lack of tagging in general which also makes it harder to search for predictions. I tried at one point to add comments to my predictions of the form [tag] but it didn't work very well and no one else was doing it. If there were a formal way to do it with searchable categories that would be really helpful.
Rather than simply forbidding them, I'd suggest having various sharing options ranging from "public" over "people I invite" to "just me."
If anyone is going to implement such a thing, I think it'd be important to have separate calibration curves for each class of publicity.
I use it when linked to it on interesting questions, usually HPMOR. But the site is too thinly used and too poorly designed to have a really robust use for me.
I recently wrote a small (python, command-line) program for prediction tracking. Above all, I wanted something that allowed very quick entry, that stored everything in a single plain-text human-readable file on my own computer, and that I could easily customize.
As far as features go, I also wanted a tagging system and to be able to do more sophisticated analysis. For example, I wanted to classify predictions as e.g."work" or "politics" when looking at accuracy, or to see how my calibration changed over different times or with how far in advance I had made a prediction.
Since the predictions that I care most about aren't very interesting to other people, I don't miss the social aspect. Nor do I want to obscure a prediction that contains personal details or store it separately.
An example:
If you wanted to predict that it will rain today, with 50% confidence, and tag the prediction with the "weather" and "external" tags, and be reminded to judge the prediction 2 days from now, you'd type this on the command line:
The created entry in the log will look something like this:
Running "predict --due" will either ask you about any predictions that are ready to be judged and update the corresponding log entry, or just change "state" to "due" so you can ctrl-f and edit the log directly. (Since the log is just text, you can manually edit any entry, like changing a prediction from "open" to "true" or "false").
Running "predict --stats" will dump statistics.
I'm very happy with this work flow, and I'm hoping to clean up the code and share it when I get a moment. It's all of ~250 lines, but I couldn't find what I wanted out there already.
I used to use it, but as I shifted to doing more on my mobile device I found it was a pain to use and so stopped. If it had a good app then I would use it again.
I don't like design and usability of the site. I think topics should be somehow divided, because not everyone is interested in soccer or currencies which are the most trivial things to predict
I've occasionally used it, but I often want to make a prediction while I'm not near a computer and doing it on a smartphone is a bit of a pain. I also don't have enough friends using it, so using it to prove I was right isn't all that possible.
Anyone here know about the International High IQ Society? I'm wondering if it would be worthwhile for me to try to get into it. Several of the free online IQ tests I've taken have put me at a few points below 120, so I probably wouldn't be able to get into Mensa, but I might be able to squeeze out ~6-7 more points from practice in order to qualify for the IHIQS, which only requires a score of 124 (95th percentile).
I am not familiar with that organization. However, psychologist Carol Dweck, the author of the book "Mindset", claims that it is better to praise children for their hard work instead of their innate intelligence, because this way one encourages them to develop "growth" mindset instead of "fixed" mindset, since the latter is less useful. On a similar note, I guess that organizations that are based on the "fixed" mindset are probably less useful than organizations or clubs that are based on some goal that could help you to develop "growth" mindset. Therefore it is probably better to join an organization that is working to achieve something and it is the result of your efforts that is measured, instead of unchanging abilities. Of course, being unfamiliar with International High IQ Society, I might be wrong about them. Perhaps instead of trying to find an organization that would help your personal growth you are simply trying to find a new place to meet new people and/or network. In that case I have very little idea whether this organization is useful for that goal, therefore I will not say anything about that.
That dichotomy looks funny to me. What about praising for results?
Relevant TED talk.
Free online IQ tests mean nothing. Don't take them more seriously than you would take a horoscope.
I don't have an experience with the organization you linked, but if it is similar to Mensa, my advice would be: don't worry about anything, just go there and take the test. If they are a serious organization, the test will probably be very different from the online tests you took. If you pass the test, there is a high probability you will be disappointed with the organization. Also, try the Mensa test.
My experience with Mensa is that it collects smart people (on the other hand, not that smart -- if you had a decent university, most of your classmates would be on that level; also most of the LW meetup participants) but then it does... pretty much nothing. You will find there also people who technically have high IQ, but otherwise are boring and/or irrational. My advice would be to pick the few individuals who seem better than the rest of the crowd, and focus on those individuals. Mensa can give you contact to these few cool people, but that's pretty much all it can do.
The IHIQS admission test actually is taken online for a fee of $5, in contrast to Mensa's which is taken in a controlled environment for a fee of $40, so I don't know how serious the IHIQS is. That's one of the reasons I was hoping someone here could share their experiences with them.
Yeah, that's kinda what I figured, but I thought I might as well see if the LW community knew anything about it.
I'm sure I'd fit right in then. :)
If networking with smart people is pretty much all Mensa's good for, could I just as well stick to LW?
I'm thinking about trying that. I wonder, though, if I should wait and try to practice for it, as the test batteries can only be taken once each. I would be very interested to hear more about your experience taking Mensa tests, even if you can only share general information about the tests themselves.
How do Bostrom type simulation arguments normally handle nested simulations? If our world spins off simulation A and B, and B spins off C and D, then how do we assign the probabilities of finding ourselves in each of those? Also troubling to me is what happens if you have a world that simulates itself, or simulations A and B that simulate each other. Is there a good way to think about this?
A world simulating itself would be highly unlikely. It's theoretically possible to have a universe simulate itself, since it can do things like just read from records instead of simulating the computer itself, but it's not really feasible. You have to simulate a smaller or more coarse-grained universe than the one you're in.
Trends at 2050?
How many linear or logarithmic trends can be forecast out as far as the year 2050? I once found some graphs of CPU speed and storage per dollar per year, but seem to have lost them; and now I'm curious what other trend-lines might be worth thinking about.
(I've been writing a story, and would be happy to make whatever such details I include as plausible as possible, while also acknowledging that relying on such trends is a mug's game.)
https://intelligence.org/2014/05/12/exponential-and-non-exponential/ ?
I'm not sure if it'll keep going, but the Flynn effect(rising IQ over time) is mysterious enough that continuation is possible, and might be interesting.
The Anti-Drug
I've seen that a lot of drugs seem to act like "gratification borrowers": they take gratification/happiness from the future and spend it all on the present, sometimes extremely quickly, then leave you feeling miserable for a certain duration, the "low" or "hangover".
I was wondering whether there was any drug that did the opposite, that functioned like delayed gratification: a drug that makes you feel utterly miserable at first, then eventually leaves you with a long-lasting feeling of satisfaction, accomplishment, and joy.
Does anyone here know of such a thing?
Reminds me of a Gregory Bateson quote:
Hm, a drug where the negative effects are first... Well, almost all medicines fit this category: unpleasant taste is common.
Illegal drugs that people want to take? Ayahuasca comes to mind:
Exercise.
But more seriously, try asking this again in the next open thread, this one seems flooded.
Actually exercise has been suggested to me as the alternative to drugs. "Spinning", specifically. Addictive, very pleasurable, and makes you healthier (unless you overdo it, but sports are much more difficult to overdo than drugs, for some reason).
There are drugs for alcoholics that make you sick if you drink, so it makes you feel miserable short-run but may help you to be stable/functional/productive long run.
In my experience Salvia divinorum works very much like this.
It seems to me that Islamist terrorists are trying to maximize defection from the larger society, and they're even able to recruit Kurds. Admittedly, they're only getting a tiny proportion of people, but why are they getting anyone at all?
Would anyone care to take a crack at whether there are conditions under which this makes sense in terms of game theory?
I think emr, above, makes some very good points, but I think you guys are all missing some crucial aspects of the situation.
The places where a distinctively Islamic terrorism has taken off (Algeria, Chechnya, Somalia, Afghanistan, Iraq, Syria) are all areas that have been ravaged by civil war or foreign occupation, leading to the breakdown of co-operative mechanisms in the wider society. In other words, the only move is defect. Yet at the same time, these societies (or sections of them) retain a distinctively Muslim identity and aspiration, so the natural way of forming new, co-operative institutions is to base them on that shared Muslim identity. Those participating in these movements no doubt see themselves as conducting a new Abbasid Revolution. This also draws in sympathisers from outside the country. Yet the outcome often becomes terrorism, because:
In other words, suppose you're a Sunni in Eastern Syria or Central/Northern Iraq, and you want to co-operate to protect your friends and family from, say, being kidnapped and tortured by the police. You tried voting for al-Iraqiyya, and indeed they won the election, but half their candidates got thrown out of Parliament, so they can't legally stop the state machinery from persecuting you. And those guys aren't gangsters, so they can't use extra-legal means to protect you. So maybe the co-operative move, at least in a lesser-of-two-evils sense, is to join ISIS. And hey, maybe that worked, because ISIS's success caused Maliki's government to collapse, and maybe the new government in Baghdad will govern in a non-sectarian way, at least for a while, in the same way that al-Qaeda-in-Iraq's initial successes gained concessions and allowed the 'Sunni Awakening.'
As for why these organisations are able to recruit worldwide - I don't know why you consider it so surprising that they should get "anyone at all." It's like medieval Christians signing up for the Teutonic Knights, or Georgian Hellenophiles going to fight for Greek independence. These people are ignorant and idiotic, with a notion of what they're doing that's utterly divorced from reality, but there's never been a shortage of romantic fools.
That may be enough to explain it, but I do think groups that compete in committing the worst public atrocities are somewhat unusual-- Hitler and Stalin made some effort to conceal the worst of what they were doing.
Thanks for the Hayek link. Do you recommend the savageleft site?
If you look at pre-20th century (and especially pre-Enlightenment) history, you see that it is Hitler and Stalin that are unusual.
I don't know anything about that site. They just happened to have a freely available copy of that chapter.
I think the linked article hits a few common themes about why this might happen:
I rather clearly shouldn't have included the link about the Kurds-- it's a distraction from the question I'm more interested in, which is why organizations like Al Qaeda and ISIS exist and are able to recruit worldwide.
Unless I've missed something, most eras don't have anything comparable.
Well, to be clear, ISIS-actually-in-Iraq-and-Sryia seems quite different from modern Al-Qaeda and self-declared-ISIS-affliate-abroad.
The Spanish Civil War might be the closest match for the foreign support for ISIS-actually-in-Iraq-and-Syria. In addition to direct foreign involvement as a proxy war, it attracted significant foreign volunteers from foreign countries whose government officially opposed their citizens going to fight (like Ireland), and whose identity partially overlapped with a particular faction, (Catholics, socialists, separatists broadly defined, etc).
Although the bulk of the recruits for ISIS have come from neighboring regions, that so many Europeans have joined ISIS-actually-in-Iraq-and-Syria might just be the result of a larger population of recent immigrants who maintain some shared identity with those involved, combined with easier travel and communication.
Modern Al-Qaeda and ISIS-affliate-abroad franchise schemes do seem more unique. You certainly could not have had comparable groups without modern communication and travel. In general, you have things like authentic branding, access to funding (Gulf States), access to expert advice, possibly better scaling, risk pooling, and a sort of meritocratic weeding-out of worse strategies that incentivize geographically separate groups to affiliate rather than work independently, in fairly direct analogy to commercial franchises, along with more local interests that favor independence. I don't know why this particular set of terminal values has geographically widespread support in the first place though.
It's all about motivation. One of the biggest driver's to join a terrorist group is a sense of powerlessness over a situation or being gullible to wider teachings. I recommend reading, "Learning to Eat Soup With A Knife." Not only does it have great bits about counter-insurgency strategy it gives you a picture of the people involved.
Because people are different and in a sufficiently large population you can find outliers (don't forget people are not rational).
I want to open up the debate again whether to split donations or to concentrate them in one place.
One camp insists on donating all your money to a single charity with the highest current marginal effectiveness. The other camp claims that you should split donations for various reasons ranging from concerns like "if everyone thought like this" to "don't put all your eggs in one basket." My position is firmly in the second camp as it seems to me obvious that you should split your donations just as you split your investments, because of risk.
But it is not obvious at all. If a utility function is concave risk aversion arises completely naturally and with it all the associated theory of how to avoid unnecessary risk. Utilitarians however seem to consider it natural that the moral utility function is completely linear in the number of people or QALYs or any other measure of human well-being. Is there any theoretical reason risk-aversion can arise if a utility function is completely linear in the way described before?
In the same vein, there seems to be no theoretical reason for having time preference in a certain world. So if we agree that we should invest our donations and donate them later it seems like there is no reason to actually donate them at any time since at any such time we could follow the same reasoning and push the donation even further. Is the conlcusion then to either donate now or not at all? Or should the answer be way more complicated involving average and local economic growth and thus the impact of money donated now or later?
Let the perfect not be the enemy of the good, but this rabbit hole seems to go deeper and deeper.
Your utility function need not be completely linear, just locally linear. If your utility function measures against the total good done in the world, your effect on the world will be small enough to be locally linear
Most people don't want to optimize the total good done, but instead care about the amount of good they do. People donate to charity until the marginal utility they derive from purchasing moral satisfaction falls below the marginal utility they derive from purchasing other things. In this case, diversification makes sense, because utility you assign to good you're responsible for is very non-linear.
If you're giving to charity at all, that's awesome. Do what motivates you.
Interesting answer. Seeing as my personal giving is completely out of pleasure not some kind of moral obligation, the argument for diversification is very strong.
Ah. Well, then there doesn't seem to be anything to debate here. If you want to do what makes you happy, then do what makes you happy.
The theoretical question still stands.
I believe donating to the best charity is essentially correct, for the reason you state. You won't find much disagreement on Less Wrong or from GiveWell. Whether that's obvious or not is a matter of opinion, I suppose. Note that in GiveWell's latest top charity recommendations, they suggest splitting one's donation among the very best charities for contingent reasons not having to do with risk aversion.
If you had some kind of donor advised fund that could grow to produce an arbitrarily large amount of good given enough time, that would present a conundrum. It would be exactly the same conundrum as the following puzzle: Suppose you can say a number, and get that much money; which number do you say? In practice, however, our choices are limited. The rule against perpetuities prevents you from donating long after your lifetime; and opportunities to do good with your money may dry up faster than your money grows. Holden Karnofsky has some somewhat more practical considerations.
...then, what? Every dollar of charity would be most efficiently invested and we'd solve the highest impact problems before moving on to others?
What happens if everyone thinks like this?
Diminishing returns. Giving $100M to a $10M problem because of information lag doesn't actually allocate money nearly as efficiently as you might hope.
Do we have any evidence about the severity of that information lag? I can't imagine it would be much more than a year. I don't know if this is a legitimatize concern. It may very well be worth a little waste value from info-lag for the benefit of overall efficiency.
You get similar results by each individual giving to a different charity. As-is, there's a few charities that are far above the rest, but if everyone thought like an effective altruist, that wouldn't be the case. Consider the idea: if everyone bought the best TV then the factory that makes that model of TV would be inundated with orders.
That's true but for now most of the problems are so large that even if almost everyone started thinking this way it would take a while before any high priority problem hit the point that it was no longer the best to donate to. Enough time for people to switch over to the next most important.
This depends on what someone believes the worthiest target for their donations is. If they're trying to optimize for goals that require the continuance of Earth-originating intelligence, or humanity, than they'll probably want to prevent human extinction. If they believe humanity faces a Great Filter or an existential risk in the coming decades, they'll want to donate eventually. Supposedly they'd drive money to organizations will decrease the chances of humanity being destroyed. Of course, this still leaves the consideration that if they invest very well, or invest at the right time, (and know how to identify those things), such that money invested now for period t, increased to value n(x), will be worth more to the target organization at the end of period t, than the original value of money, x, is worth to the organization now.
However, its been argued from within effective altruism that the longer you wait to do good, the less the good you do will be worth. A donation is worth more now than later; effort applied at a later time will result in less value than the same level of effort applied now. This is called the haste consideration. Over time, the marginal increase in value of an investment over time comes up against the (presumed) marginal decrease of donation at any given time. I figure one could try measuring or calculating the rates of change here, and then figuring out some point at which the curves of them cross which would provide the optimum time to withdraw money from investment and donate it for maximum impact. However, that seems difficult. I'm not aware of something like this previously being done, and I follow effective altruism closely. So, if this has all been quantified, that report hasn't been widely circulated.
This seems all the more important on LessWrong, because users on this site are more likely to care about existential risk reduction, whether they identify with effective altruism or not. Also, one could estimate a time past which investing rather than donating would be pointless, because that is beyond the point at which one expects humanity to preserve anything worth either investing in or donating to.
To coin a phrase, "What has government done to our money?"
https://www.google.com/webhp?sourceid=chrome-instant&rlz=1C1TSNP_enUS504US504&ion=1&espv=2&ie=UTF-8#q=oil%20price%20collapse
Because this seems to conflict with the "Austrian" doomsaying that the U.S. dollar would "collapse." Now it looks like the U.S. dollar has gone into the opposite of a "collapse" because a dollar can buy a lot more oil today than it could a few months back.
Like chaosmage said, oil is getting cheaper in all currencies (at least the ones not experiencing hyperinflation). Thus it isn't related to the country's economic policy.
Also if you update against "Austrianism" every time the price of oil drops, do you update in it's favor every time it rises?
Saudi Arabia old oil and the US frackers are currently in a price war. Oil is cheaper for every importer, not just for those who happen to use the U.S. dollar as their domestic currency.
Austrian economics has consistently made bad predictions. It doesn't give you the right answers for rates, fx, gdp or inflation.
As opposed to what kind of economics? :-/
http://krugman.blogs.nytimes.com/2011/10/09/is-lmentary/
Yes, Krugman correctly predicted that the post-2008 flood of money will not lead to quick inflation. That's the example that's I've seen literally dozens of times as the "proof" that Krugman is right and everyone else is wrong.
Can I see any other pieces of evidence?
There was that time when he predicted that fiscal tightening in 2013 would be refute his ideological opponents, and then ... totally failed to admit he was wrong when the evidence came out against him.
The effects of fiscal austerity vs deficit spending:
http://www.nytimes.com/2014/02/21/opinion/krugman-the-stimulus-tragedy.html?_r=0
That's not evidence, that's Krugman patting himself on the back. For evidence I would like to see a testable prediction made before the fact.
In this particular case, Krugman's original position was that the stimulus could be useful but was not sufficient. There is enough wiggle room for two elephants there -- if the stimulus failed, Krugman would have pointed at himself saying it was insufficient and if it obviously worked, he would have pointed at himself saying it would be useful.
In general, I find Krugman to be an interesting example of a very smart guy who either became mindkilled or deliberately decided to do propaganda "for the greater good". His columns are full of classic motivated reasoning.
As to Keynesianism, see this keeping in mind Yvain's recent Beware the Man of One Study.
Horrible news!!! Organic molecules have just been found on Mars. It appears that the great filter is ahead of us.
We've known for some time that Titan has plenty of organic molecules.
The article says:
There's also matter exchange from earth to mars that could have brought life that originated on earth to mars.
This only means that the great filter is not due to the difficulty of creating organic compounds. (In fact, creating organic compounds was already very low on a list of things that might be the cause of an early filter, or maybe even already effectively ruled out.).
It still could be a step between this and where we are now. For example, it could be the creation of Eukaryotes. Or it could be intelligence. Or other things.
Anything other than methane which is the simplest organic molecule there is (CH4) and, as far as I remember, has been detected in interstellar gas..?
Detecting methane per se isn't what's interesting about this. You don't need life to produce methane; there's plenty of it in the outer solar system, Titan for example is covered in the stuff, and it's been detected on Mars before. But it's a small enough molecule that Mars daylight temperatures can give it enough velocity to escape from the planet's gravity well, which means that if you detect nontrivial quantities in Mars' atmosphere it's probably being actively replenished somehow. Fluctuating levels of methane, which is what's actually new about the Curiosity measurements, are strong evidence for active replenishment.
This doesn't necessarily mean life -- there are other possibilities involving deep geology of various kinds -- but life is one of the candidate explanations. Though if it is life, it's probably simple, boring microbial life similar to what appeared quite early in Earth's history, so I think artemium is overselling the Great Filter implications on a couple of levels.
Sure, but there are tons of geological processes which produce methane without any involvement of life.
The discovery of methane spikes is certainly interesting, but to go from there to changing the Great Filter estimates is a looooooong jump.
Request for help: I'm looking for a study on scope insensitivity to use in one of my college entrance essays. If I recall it correctly, the study showed something like, when asked how much people would pay save one girl of a group of eight in a tribe from a disease, it was over double what people would pay to save all eight.
I checked the standard scope insensitivity post, and tried my google-fu, but can't remember where I originally heard it.
The one you refer to was mentioned in http://lesswrong.com/lw/n9/the_intuitions_behind_utilitarianism/:
Link is dead, but the article is here http://foreignpolicy.com/2007/03/13/numbed-by-numbers/.
The original study seems to be Kogut and Ritov, 2005. I found a pdf through Google scholar search (always a good way to find studies) over here.
Thank you very much :-)
Every night as I'm lying in bed trying to fall asleep, I think of five of six things I want to remember habitually or in the short-run, so I get up and write them down. This costs me at least 25 minutes of sleep. I'm sure I'm not the only one with this problem; does anyone know good ways to store or record these ideas?
Keep lot of slips of paper on a clipboard next to the bed. Give each idea its own slip (so they can easily be sorted later). Lay down on the bed 25 minutes early, which will give you time for the thoughts to arise, then you'll get your full sleep time.
Would it be possible for me to buy an actual printed version of HPMoR in some way? (Or at least, of the first X chapters of HPMoR, where X is at least a significant part of the story).
Thanks!
Fanfiction writers are generally prohibited from making money without the original author's permission. This includes charging for prints.
If you just want a printed copy, you can order one printed from a PDF yourself (search terms: PDF bound book). It is likely to cost you upwards from $100 for the currently released chapters and come in 4-5 volumes.
If you want to thank Eliezer for writing the story, donate to MIRI the amount you find appropriate.
Is there a significant difference between the mathematical universe hypothesis and Hegelian absolute idealism. Both seem to claim the primacy of ideas over matter (mind in case of Hegel, and math in case of MUH), and conclude that matter should follow the law of ideas. MUH just makes one step forward, and says that if there are different kind of maths, there should be different kinds of universes, while Hegel haven't claimed the same about different minds.
I've always thought of the MU hypothesis as a derivative of Plato's theory of forms, expressed in a modern way.
Reduced consumption of animal products, more specifically meat should help my health and both my purse and the global poor through reduced food prices. For reducing meat consumption in general it seems easy to just replace meat in a lot of dishes with cheese or substitute meaty dishes with some scrambled eggs. What can I do for variety? I am especially looking for cheap, fast and/or convenient methods to put together a meal. I am very willing to trade off fast for the other two as I can listen to audiobooks or similar while preparing food.
Chickens produce more meat than eggs in a given time-period. I originally encountered this when reading about animal cruelty, but I would expect the same would apply to saving money and lowering the demand for food. If anything, you should be replacing eggs with meat.
The self-indication assumption seems to violate some pretty basic probability. Suppose that there are two possible universes: universe A has one person, and universe B has 99 people. Suppose that the prior probability for each is 50%. Under SIA, you are 1% likely to be each one of these 100 possible people. But that means that universe B has 99% probability, even though we just assigned it 50% probability. It can't change without updating on evidence, which we never did. What happened?
We updated on the fact that we exist. SSA does this a little too: specifically, the fact that you exist means that there is at least one observer. One way to look at it is that there is initially a constant number of souls that get used to fill in the observers of a universe. In this formulation, SIA is the result of the normal Bayesian update on the fact that soul-you woke up in a body.
Suppose there is a 2^(-n) chance of universe U_n with n people for n > 0. Initially, there's nothing paradoxical about this. SIA converges. But look at the evidence you get from existing. Call that E.
P(U_n|E) = knP(U_n) for some k
P(U_n|E) = P(U_n&E)/P(E)
P(U_n|E) < P(U_n)/P(E)
P(E) < P(U_n)/P(U_n|E)
P(E) < P(U_n)/(knP(U_n))
P(E) < 1/kn
Since k is constant and this is true for all n, P(E) = 0
So, existence is infinitely unlikely? Or we must assume a priori that the universe definitely doesn't have more than n people for some n? Or P(U_n&E) is somehow higher than P(U_n)?