Open thread, Oct. 27 - Nov. 2, 2014

5 Post author: MrMind 27 October 2014 08:58AM

If it's worth saying, but not worth its own post (even in Discussion), then it goes here.

Notes for future OT posters:

1. Please add the 'open_thread' tag.

2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)

3. Open Threads should be posted in Discussion, and not Main.

4. Open Threads should start on Monday, and end on Sunday.

Comments (400)

Comment author: CAE_Jones 03 November 2014 04:57:35AM 2 points [-]

Two years ago, I wrote this cringe-worthy thing.

I can't tell if things have gotten worse, or if they've stayed the same. I lean toward worse.

4 years ago, I asked a psychiatrist about my soul-crushing Akrasia issues. He prescribed Focalin, at 5mg/day for the first week, then 10mg/day for the second. The first week saw improvements--I didn't feel like I had much choice over what I wound up focusing on, but I actually finished things--the second week did not work at all, and a pile of unpleasant things all hit at once on one of those nights. So we switched to Prozac, nothing came of it, and here we are.

For reasons, skills that are basic and trainable for most people are down-right mutant to me. (It's almost as though any problems that aren't sheltered-nerd problems amplified by blindness are blindness problems amplified by sheltered nerdiness. There exist blind nerds without this suite of doom-problems, but on investigation nothing seems to generalize.)

I didn't mention psychiatric problems in the 2012 post because all the psych professionals I've spoken to don't seem to believe I have them. But I'm pretty sure I have symptoms of ADHD-PI. And Schizoid Personality. And Avoidant Personality. And Social Anxiety. And Atypical Depression but I'm pretty sure that's a response to everything else. Whether any of these is actually the case is unknown to me, and all of the above mean that finding a professional to ask (then actually telling them everything) is stupidly difficult. Then they need to be competent.

I really have no idea what the best starting place would be. I'm trying to find another psychiatrist (though I dunno if I can actually communicate the problems), I've exhausted the less dramatic training facility and the return-to-college option, and am considering one of the National Federation of the Blind training centers (as a general rule, everyone who is not a member finds the NFB offputting, but it's pretty much their programs or nothing, if the internet is to be believed).

TL;DR: My life sucks and if I can't fix it soon, I will start complaining that decent wireheading is not yet available.

Comment author: Strangeattractor 05 November 2014 03:31:05PM 0 points [-]

Have you considered the idea of learning echolocation? Here is the beginning of a series of blog posts from blind programmer Austin Seraphim about how he learned to use echolocation to navigate the environment and get a spatial sense of things without touching them. He learned it from a teacher from World Access for the Blind.

It came to mind because you mentioned a National Federation of the Blind training center, and I'm not sure what you would learn there, but I'm pretty sure they don't offer echolocation training.

Comment author: ChristianKl 05 November 2014 01:03:23PM 1 point [-]

Instead of a psychiatrist maybe a psychologist might be the better option?

Comment author: FiftyTwo 02 November 2014 07:07:29PM *  2 points [-]

Can anyone recommend any good books/resources on dyspraxia?

Ideally suitable for adults with a reasonable background understanding of psychology. Most of the stuff I've been able to find has been aimed for teachers/parents.

Comment author: fubarobfusco 01 November 2014 05:53:58PM 3 points [-]

I've seen a few discussions recently where people seem to argue past one another because they're using different senses of the terms "subjective" and "objective".

Some things are called "subjective" because they are parametrized by subject. For instance, everyone who can see has a field of vision, but no two people have the same field of vision (because two people can't stand in the same spot at the same time). However, we can reason and calculate accurately about someone else's field of vision.

Other things are called "subjective" because they are internal to a subject. For instance, since humans are not telepathic we don't have access to the thoughts or mood of another person. The only way we can discover them is by being told about them — or, theoretically, brain-scans — and even this doesn't convey how it feels to be that person.

Comment author: ChristianKl 02 November 2014 06:38:41PM 0 points [-]

For instance, since humans are not telepathic we don't have access to the thoughts or mood of another person. The only way we can discover them is by being told about them — or, theoretically, brain-scans

I think various people are better at mood detection via reading body language than brain-scans. Both brain scans and reading body language are cases where you have partial information and use that to do pattern matching. I have multiple experiences where I meet people who can perceive my own mood better than I can myself.

There are many times where I get a better idea about someone mood by hugging that person then by asking them verbally and them telling me about how they feel.

Comment author: Curiouskid 31 October 2014 06:26:24AM 6 points [-]

Bayesianism and Causality, or, Why I am only a Half-Bayesian (Judea Pearl)

“The bulk of human knowledge is organized around causal, not probabilistic relationships, and the grammar of probability calculus is insufficient for capturing those relationships.”

Comment author: NancyLebovitz 30 October 2014 10:07:48PM *  7 points [-]

How communities Work, and What Wrecks Them

One of the first things I learned when I began researching discussion platforms two years ago is the importance of empathy as the fundamental basis of all stable long term communities. The goal of discussion software shouldn't be to teach you how to click the reply button, and how to make bold text, but how to engage in civilized online discussion with other human beings without that discussion inevitably breaking down into the collective howling of wolves.

Behavior patterns that grind communities down: endless contrarianism, axe-grinding, persistent negativity, ranting, and grudges.

Comment author: Nornagest 30 October 2014 10:21:22PM *  3 points [-]

I agree about all of that except for contrarianism (and yes, I'm aware of the irony). You want to have some amount of contrarianism in your ecosystem, because people sometimes aren't satisfied with the hivemind and they need a place to go when that happens. Sometimes they need solutions that work where the mainstream answers wouldn't, because they fall into a weird corner case or because they're invisible to the mainstream for some other reason. Sometimes they just want emotional support. And sometimes they want an argument, and there's a place for that too.

What you don't want is for the community's default response to be "find the soft bits of this statement, and then go after them like a pack of starving hyenas tearing into a pinata made entirely of ham". There need to be safe topics and safe stances, or people will just stop engaging -- no one's always in the mood for an argument.

On the other hand, too much agreeableness leads to another kind of failure mode -- and IMO a more sinister one.

Comment author: Adele_L 31 October 2014 12:31:57AM 2 points [-]

The article talked about endless contrarianism, where people disagree as a default reaction, instead of because of a pre-existing difference in models. I think that is a problem in the LW community.

Comment author: TrE 31 October 2014 03:47:08PM *  3 points [-]

On the contrary, from my experience it isn't.

Sorry, I could not resist the opportunity. But seriously, I don't often see people disagreeing for the sake of disagreeing. More often, they'll point out different aspects, or their own perspective on a topic. To be honest, support and affirmation are perhaps a bit rarer than they should be, but I've rarely perceived disagreement to be hostile, as opposed to misunderstanding, or legitimate and resolvable via further discussion.

More datapoints, anyone?

Comment author: ChristianKl 01 November 2014 08:22:51PM *  2 points [-]

If other people disagree with what I write they usually do it for the sake of disagreeing. However if I disagree... ;)

Comment author: [deleted] 30 October 2014 07:13:47PM 0 points [-]

This article discusses a paper that seems interesting from the perspective of effective altruism and how peoples behavior changes based on where they think their money might be going:

http://www.vox.com/2014/10/30/7131345/overhead-free-donations-charity-fundraising-seed-matching-gneezy

If you want a link directly to the paper, that link is both in the article and reposted here:

http://www.sciencemag.org/content/346/6209/632

Short summary: When considering donations, people in the study donated more when they know their donation is not going to overhead.

Comment author: khafra 30 October 2014 06:29:33PM 2 points [-]

Are there lists of effective charities for specific target domains? For social reasons, I sometimes want to donate to a charity focused on some particular cause; but given that constraint, I'd still like to make my donation as effective as possible.

Comment author: NancyLebovitz 30 October 2014 03:33:58AM 7 points [-]
Comment author: Vulture 31 October 2014 10:48:27PM 1 point [-]

This could be a big deal for the bestiality debate (although conducting the necessary training without falling afoul of the original ethical concerns would probably be a trick).

Comment author: NancyLebovitz 01 November 2014 03:09:29PM 3 points [-]

A general training in do want, don't want for ordinary things like blankets and types of food could go a long way to solving the problem.

Comment author: Vulture 01 November 2014 09:33:10PM 1 point [-]

That seems like a huge leap in terms of capability, though, to add the free parameter of "condition to be started/stopped" somehow.

Comment author: fubarobfusco 01 November 2014 07:41:03PM 2 points [-]

Warning: this comment is a ramble without a conclusion. Horses participating in tell culture? Cool. Preferences and consent are complicated.

This line of thinking seems to lead to some interesting places about the idea of consent.

I'm increasingly of the opinion that the whole notion of "consent" is socially constructed (that is, learned) — that it is desirable but cannot be assumed to be natural or inherent. People have to learn, not only to ask others' consent, but to recognize when their consent is being asked: not only to ask "Do you want this?" but to know when someone wants them to have and express a preference.

Indeed, the idea of developing preferences of one's own has to be learned. (Possibly the whole notion of having an identity, too.)

People raised in very controlling households seem to have trouble with this — with formulating and communicating preferences and seeking consent, rather than just ① going ahead and doing things that affect others and then seeing how those others react, or ② expecting others to do the reciprocal. They expect interactions to be, not necessarily forced, but certainly not negotiated. "Better to ask forgiveness than seek permission" is one thing as a maxim for decision-making in a bureaucratic office, but quite another thing in personal relationships!

This leads to communications problems between these folks and people who have been taught to exchange consent. For instance, "Would you like to do thus-and-so with me?" for one person can mean "I expect you to do thus-and-so with me and will be disappointed or angry if you don't" whereas for another it can mean "I actually don't know if thus-and-so would be worth doing for us; what do you think?"

Previously I thought that this difference was that (to put it overly strongly) people from controlling households had had their free will beaten out of them — that they had been abused or neglected in a way that made them alieve that people would not respect their preferences or dissent, and so did not bother to express any. But now I think the opposite: "just do stuff and see how others react" is the state of nature, whereas "formulate and express preferences and negotiate with others" is socially constructed.

And as a society, it seems we are demanding more and more of it. That sounds like a pretty good thing to me, especially for people whose preferences would otherwise be denied or disregarded. But it isn't free or obvious; it's a big structure of socially-constructed-stuff that people have to learn.

Computationally speaking, preferences aren't free. Even if we model people as agents with utility functions (which I'm not sure we should!), having a utility function doesn't mean having explicit knowledge of what your utility function is! In order to express preferences, an agent has to notice facts about itself, notice regularities about those facts, figure out what it might want another agent to do ... and so on. All that requires brain power.

Teaching a horse to express preferences — that it can communicate something that will influence its handler's actions, to get something done that it can't do for itself — seems like a pretty big deal. Affirmatively communicating about a specific action is more "consent-like" than, say, merely expressing an emotional state of dissatisfaction or contentment.

I get the sense that people who live with animals generally do have a notion of what the animals like or dislike. But that isn't the same as communicating preferences or consent.

On zoophilia/bestiality, I at one point thought something like: "A dog or horse can obviously express dissatisfaction with physical acts it doesn't like — by pulling away, kicking, biting, etc. Some animals can clearly 'propose' sexual acts with humans, such as a dog humping a person's leg. And we don't expect people to seek animals' consent to a hell of a lot of things that we do expect them to seek consent from humans — such as medical treatment or being put in a cage. So what's the big deal?"

But a dog humping someone's leg isn't proposing a sexual act or consenting to one; it's initiating one. If a human did the equivalent, to random people they didn't have an existing relationship with, well, we wouldn't want to put up with that sort of thing.

People (and, I suspect, horses) have different degrees of insight into their own preferences. It is perfectly possible to be wrong about your preferences: to believe that you would be happier if you ate a bag of candy, when in fact you would give yourself a stomachache and be less happy.

Comment author: ChristianKl 01 November 2014 08:43:02PM 1 point [-]

Consent is really tricky.

Imagine a woman sitting at the bar. The woman knows what she's doing and knows that when she smiles in a certain way at a man there a 90% chance that the man will approach her, however only in 10% of the cases the man has an idea that the woman did something to make the woman approach.

If the woman initiates an interaction like that does she have informed consent? Is there some ethical imperative for her to inform the man that she initiated the interaction?

To frame the question in another way, if all you are doing is trigger the system 1 of the other person do let the person engage in certain actions, but you never ask a question to give system 2 the opportunity to reflect, do you have consent?

Comment author: fubarobfusco 01 November 2014 08:58:48PM *  0 points [-]

Guess cultures are really tricky!

If it is indeed the case that everyone knows for certain what the signals mean, then they can be very specific communications of intent and consent: there is not actually any guessing going on! But if the point of using facial expressions and gestures rather than words is that the former are deniable, then it probably can't be the case that everyone knows for certain: deniability relies on ambiguity.

If two people have slightly different interpretations of what the signals mean, then they can end up with extremely divergent interpretations of what happened in a particular exchange.

For that matter, if everyone in the bar grew up in the same town and went to the same schools, that's a pretty different situation from if the bar is an assemblage of people from wildly different backgrounds who happen to have landed in the same location.

(I may be computing from stereotypes in saying this ... but I expect that guess cultures prize uniformity, and fear diversity as a source of confusion; whereas tell cultures may consider uniformity boring, and prize diversity as a source of novelty.)

Sexually, it seems to me that if all you are doing is triggering the System 1 of the other person and neither person is waiting around for System 2 to engage and reflect, that may be very hot indeed — Erica Jong's "zipless fuck" — but the failure modes are correspondingly huge.

Comment author: ChristianKl 01 November 2014 09:33:35PM 1 point [-]

If it is indeed the case that everyone knows for certain what the signals mean, then they can be very specific communications of intent and consent: there is not actually any guessing going on!

It's possible to send signal A and the other person not understanding what the signal means and doing nothing.

But it's also possible that they don't understand the signal but the signal causes them to feel a certain emotion and that emotion lets them engage in an action without them having any idea of the casual chain.

The more I learn about how humans work the more I get those practical ethical dilemmas. Even worse, to really know what I'm doing I have to experiment and I'm curious ;)

Comment author: Dias 30 October 2014 02:54:47AM 6 points [-]

Suppose I was an unusual moral, unusually insightful used car saleswoman. I have studied the dishonest sales techniques my colleagues use, and because I am unusually wise, worked out the general principles behind them. I think it is plausible that this analysis is new, though I guess it could already exist in an obscure journal.

Is it moral of me to publish this research, or should I practice the virtue of silence?

  • It might help people resist such techniques.
  • It might help salesmen employ these immoral techniques better.
  • Salesmen are more likely to already understand much of the content - vulnerable outsiders would have more to learn
  • Salesmen are more incentivized to learn from my analysis.
  • It is quite interesting to read as a purely abstract matter.
  • I like producing and sharing interesting research.

Obviously the dishonest car salesman is just an example so don't get too tied up on the efficiency of the second hand car market.

Comment author: gjm 30 October 2014 01:32:37PM 4 points [-]

Robert Cialdini did something a bit like this in researching his book "Influence", and so far as I can tell pretty much everyone agrees it's a good thing he wrote it.

I suspect attitudes to your doing this would depend on what your publication looked like. You could write

  • a book called "Secrets of Successful Second-hand Sales", aimed at used car salespeople, advising them on how to manipulate their customers;
  • a book called "Secrets of the Sinister Second-hand Sellers", aimed at used car buyers, advising them on what sort of things they should expect to be done to them and how to see through the bullshit and resist the manipulation;
  • a book called "A Scientific Study of Second-hand Sales Strategies", aimed at psychologists and other interested parties, presenting the information neutrally for whatever use anyone wants to make.

(As an unusually moral person you probably wouldn't actually want to write the first of those books. But some others in a similar situation might.)

My gut reaction to the first would be "ewww", to the second would be "oh, someone trying to drum up sales by attention-grabbing hype",and to the third would be "hey, that's interesting". Other people's guts may well differ from mine. Cialdini's book is mostly the third, with a little touch of the second.

Comment author: ChristianKl 30 October 2014 01:37:57PM 5 points [-]

Cialdini's book is mostly the third, with a little touch of the second.

And read by people who want to read the first ;)

Comment author: gjm 30 October 2014 07:02:37PM 1 point [-]

And also who want to read the second or the third. But yes, of course, writing for one audience won't stop others taking advantage.

Comment author: Douglas_Knight 31 October 2014 12:51:14AM 3 points [-]

I estimate that 95% of readers of Cialdini read it for business.

Comment author: ChristianKl 30 October 2014 11:03:18AM 2 points [-]

I think it depends very much on the case.

There are things in the social skill space that I discovered via experimentation that I don't openly share.

Sales man aren't the only people who care about getting people to make decisions. In medicine compliance is pretty important and choice engineering as a field isn't completely evil.

Understanding our decision making can also give us insight into issues like akrasia.

Comment author: mare-of-night 29 October 2014 05:56:37PM 5 points [-]

There have been discussions here in the past about whether "extreme", lesswrong-style rationality is actually useful, and why we don't have many extremely successful people as members of the community.

I've noticed that Ramit Sethi often uses concepts we talk about here, but under different names. I'm not sure if he's as high a level as we're looking for as evidence, but he appears to be extremely successful as a businessman. I think he started out in life/career coaching, and then switched to selling online courses when he got popular. His stuff is generally around the theme of "how to win at life", but focused on his own definition of that, which is mainly having a profitable and interesting career. (He has a lot of free content which is only inconvenience-walled by being part of a mailing list - this video is one of those things.)

I'm curious if anyone else here knows of him, and what you think of him.

Comment author: hyporational 30 October 2014 06:01:47AM *  2 points [-]

why we don't have many extremely successful people as members of the community.

I'm not sure if the community has been around long enough for this to be a useful kind of a measurement. Success doesn't happen in an instant and there's a lot of turnover. People who are already successful don't have much pressure to join in.

Comment author: RowanE 30 October 2014 01:46:56PM 2 points [-]

Additionally, "extreme success" is usually defined in zero sum terms that make it definitionally extremely rare, in addition to the strong influence of chance in whether one achieves success in most fields. So a community as small as ours with "not many extremely successful people" may still be completely worthwhile and have a high rate of extreme success per capita compared to most groups.

Comment author: wadavis 29 October 2014 10:35:01PM 6 points [-]

Side point: I've found material like his, "concepts we talk about here, but under different names", extremely useful when I want to explain the idea of rationality to someone without having to work around the lesswrong lingo and trying to have a conversion while tabooing all the lesswrong phases and cached thoughts.

Comment author: IlyaShpitser 30 October 2014 10:54:59AM *  4 points [-]

Yes! In my opinion, it's a great habit to be on the lookout for things under a different name. This is the "academic coordination problem:" things are often rediscovered again and again, because people have incentives to write but not to read.

Comment author: wadavis 29 October 2014 10:30:51PM 2 points [-]

Fully agree that he uses concepts used with less wrong, under different names. And I've seen him referenced frequently on less wrong as somewhere to look for rational financial / career advice.

I follow his free material, it has provided me with inspiration/direction/confidence to aggressively pursue increased compensation, successfully. I've been tempted to purchase his material before, but am always discouraged last second by the smell of snake oil.

Comment author: mare-of-night 30 October 2014 03:05:37AM 3 points [-]

I've been doing the same thing, for a while. I also get turned off a bit by the snake oil, and I've been following some of the mailing lists long enough that the content starts to feel repetitive. I might still buy, if he ever put out anything inexpensive (doesn't seem likely, but Jeff Walker did a while ago even though his business has a similar strategy, so it might happen..).

I wonder if everyone gets that slight snake oil feeling from him? And in particular, whether the kinds of marketing he's using still work when the reader recognizes what tactic is being used.

Comment author: wadavis 30 October 2014 02:39:44PM 0 points [-]

The question kept coming up; If I can smell snake oil, am I the target audience?

Even if it is legit and honest (I think it is), it kept on reminding me of nigerian phishers using poor language to discourage all but the most gullible from wasting their time.

Comment author: Wes_W 29 October 2014 04:47:22PM 1 point [-]

I stumbled across an article about Amelia, a program that can supposedly perform low-level human jobs like call center operator. A brief search hasn't turned up anything particularly illuminating. Has this been discussed on LW before?

On the one hand, everything I read about her sounds sufficiently vague that I suspect it's hype (and possibly native advertising). Still, I'm curious about the underlying tech - is it some kind of substantial improvement over past attempts, or is she just Siri++ in the way that Eugene Goostman was a slightly better chatbot?

Comment author: Douglas_Knight 30 October 2014 05:11:49AM 1 point [-]

Probably Siri-- in the way that Eugene Goostman was a slightly worse chatbot.

Comment author: polymathwannabe 29 October 2014 05:41:55PM -1 points [-]

The manufacturer's website is only merely illustrative.

Comment author: NancyLebovitz 29 October 2014 04:27:41PM 3 points [-]
Comment author: DataPacRat 29 October 2014 04:07:21PM 6 points [-]

Seeking LWist Caricatures

I've written the existence of a cult-like "Bayesian Conspiracy" of mostly rebellious post-apocalypse teens - and now I'm looking for individuals to populate it with. What I /want/ to do is come up with as many ways that someone who's part of the LW/HPMOR/Sequences/Yudkowsky-ite/etc memeplex could go wrong, that tend not to happen to members of the regular skeptical community. Someone who's focused on a Basilisk, someone on Pascal's Mugging, someone focused on dividing up an infinity of timelines into unequal groups...

Put another way, I've been trying to think of the various ways that people outside the memeplex see those inside it as weirdos.

(My narrative goal: For my protagonist to experience trying to be a teacher. I'd be ecstatic if I could have at least one of the cultists be able to teach her a thing or two in return, but since I've based her knowledge of the memeplex on mine, that's kind of tricky to arrange.)

I can't guarantee that I'll end up spending more than a couple of sentences on any of this - but I figure that the more ideas I have to try building with, the more likely I will.

(Also asked on Reddit at https://www.reddit.com/r/rational/comments/2kopgx/qbst_seeking_lwist_caricatures/ .)

Comment author: CellBioGuy 30 October 2014 03:30:22PM *  4 points [-]

Someone who applies useful effective behaviors towards the achievement of a ridiculous or reprehensible end goal.

Comment author: DataPacRat 30 October 2014 04:16:16PM 2 points [-]

I think I have this one covered; my character entry is simply "I wanna be a pony!".

(And, now that I think about it, my protagonist has said that if they don't have any other end goals they can think of, they're going to act as if their end goal is to "read comics".)

Comment author: IlyaShpitser 30 October 2014 11:03:57AM *  3 points [-]

people outside the memeplex see those inside it as weirdos.

When judging how weird a community is, people often approximate a kind of "weirdness Pagerank" by looking at people the community holds in high esteem. I think Yudkowsky can come across as weird and offputting to some folks (not in person, but online. This is a bit of a tangent, but I think it is very interesting to think about the systematic ways our online and offline personas differ and why they do so). If people perceive that, their alarms immediately go off and they conclude folks are brainwashed since they are not seeing the weirdness themselves.

Comment author: DataPacRat 30 October 2014 04:18:47PM 0 points [-]

This can add some useful background detail. My protagonist is acting as a pseudo-Yudkowsky to the group, and has already been called the "Mad Queen" at least once.

Comment author: fubarobfusco 30 October 2014 07:49:10AM 3 points [-]

The person who airs fringe supremacist (or even eliminationist) views ... then is surprised and offended when members of the targeted groups shun him or her instead of arguing the points as if they were a matter of abstract intellectual interest.

No, wait, that's probably not LW-specific enough.

Comment author: CellBioGuy 03 November 2014 07:44:00AM *  0 points [-]

I dunno, it seems to be happening here a disturbingly large amount lately.

Comment author: Sjcs 30 October 2014 12:41:14AM 2 points [-]

Put another way, I've been trying to think of the various ways that people outside the memeplex see those inside it as weirdos.

The lurker, who may not be gaining as much utility as they would if they participated. However, they still receive the same (or a degree of) connotations from those outside the memeplex, due to their association with the group. These percepts from the outside may be either good or bad.

Comment author: philh 29 October 2014 05:45:09PM 4 points [-]

The person who uses ev psych to justify their romantic preferences to potential and current partners. (There's a generalisation of this that I'm not sure how to describe, but I've fallen into it when talking with friends about the game-theoretical value of friendship.)

Comment author: skeptical_lurker 31 October 2014 01:24:47PM 1 point [-]

If the problem is that you shouldn't have to justify your romantic preferences then I can see where you are coming from, but if you do want a justification, what is wrong with evo psych?

Comment author: NancyLebovitz 01 November 2014 03:28:05PM 0 points [-]

Evo psych tends to be too general and too unproven.

Comment author: skeptical_lurker 04 November 2014 12:11:53PM 0 points [-]

I dunno if that's true, but regardless its a general argument against evo psych, rather than an aguement against using ev psych to justify romantic preferences.

Comment author: fubarobfusco 30 October 2014 07:40:05AM 1 point [-]

One possible generalization: Being insecure about personal preferences, and so seeking to show that one's personal likes are rooted directly in something universal — something outside one's own personal history, culture, subculture, upbringing, etc.

Comment author: ChristianKl 29 October 2014 04:48:50PM 5 points [-]

Calculating Bayes rule for everything can be quite weird for a lot of people. I remember a case where someone found it weird that another person asked on LW how to do a Bayesian calculation for the likelihood that a specific girl likes him.

Calculating probabilities for many everyday issues is hugely weird for many people. You might even have to take care to make it sound believable even if you do describe a real world character.

I remember an anecdote of a person doing an utility calculation that suggest having sex without a condom and being exposed to the chance of getting AIDS is quite okay.

Another of those things that CFAR preaches that can be seen as pretty weird is purposeful comfort zone extension. It's the kind of topic where you also have to worry about believability if you just tell real world stories.

Comment author: Lumifer 30 October 2014 03:45:16PM 2 points [-]

Calculating probabilities for many everyday issues is hugely weird for many people.

And rightly so. The great majority of people are badly calibrated, can't estimate priors properly, etc. If they tried to calculate probabilities for "many everyday issues" I would bet most of them would land straight in the valley of bad rationality.

Comment author: Azathoth123 02 November 2014 05:48:21AM 0 points [-]

Heck, many people here can't do it right. I'm in particular thinking of the recent thread about computing probability of UFOs or aliens.

Comment author: NancyLebovitz 29 October 2014 04:02:13PM *  9 points [-]

The outside view.... (The whole link is quoted.)

Yesterday, before I got here, my dad was trying to fix an invisible machine. By all accounts, he began working on the phantom device quite intently, but as his repairs began to involve the hospice bed and the tubes attached to his body, he was gently sedated, and he had to leave it, unresolved.

This was out-of-character for my father, who I presumed had never encountered a machine he couldn’t fix. He built model aeroplanes in rural New Zealand, won a scholarship to go to university, and ended up as an aeronautical engineer for Air New Zealand, fixing engines twice his size. More scholarships followed and I first remember him completing his PhD in thermodynamics, or ‘what heat does’, as he used to describe it, to his six-year-old son.

When he was first admitted to the hospice, more than a week go, he was quite lucid – chatting, talking, bemoaning the slow pace of dying. “Takes too long,” he said, “who designed this?” But now he is mostly unconscious.

Occasionally though, moments of lucidity dodge between the sleep and the confusion. “When did you arrive?” he asked me in the early hours of this morning, having woken up wanting water. Once the water was resolved he was preoccupied about illusory teaspoons lost among the bedclothes, but then chatted in feint short sentences to me and my step-mum before drifting off once more.

Drifting is a recent tendency, but in the lucidity he has remained a proud engineer. It’s more of a vocation, he always told his students, than a career.

Last week, when the doctors asked if he would speak to medical trainees, he was only too happy to have a final opportunity to teach. Even the consultants find his pragmatic approach to death somewhat out of the ordinary and they funnelled eager learners his way where he was happy to answer questions and demonstrate any malfunctioning components.

“When I got here”, he explained to them, “I was thermodynamically unstable but now I think I’m in a state of quasi-stability. It looks like I have achieved thermal equilibrium but actually I’m steadily losing energy.”

“I’m not sure”, I said afterwards, “that explaining your health in terms of thermodynamics is exactly what they’re after.”

“They’ll have to learn,” he said. “You can’t beat entropy.”

Comment author: dthunt 29 October 2014 06:02:05AM 6 points [-]

Hey, does anyone else struggle with feelings of loneliness?

What strategies have you found for either dealing with the negative feelings, or addressing the cause of loneliness, and have they worked?

Comment author: MrMind 03 November 2014 10:31:37AM 0 points [-]

What strategies have you found for either dealing with the negative feelings, or addressing the cause of loneliness, and have they worked?

On one side, a feeling of loneliness is a signal that in my life I should socialize and connect more.
Other times though, decisions and actions taken under that emotion turned up to be pretty bad: it would have been better to just be and feel alone.
I have thus filled up my week but have left slots of time to be alone, and I know that any feeling of loneliness that I get is just a withdrawal symptom.
I've filled my social life with dancing classes, founding a local go club, joined a teaching class and time to go out with my generic friends. On the other side, when I still feel alone I just take some minutes to sit quietly and imagine being in a pleasant social or sexual situation, trying to focusing on every detail. This is usually more than enough to clean me from any negative state of mind.

Comment author: IlyaShpitser 30 October 2014 11:08:19AM 2 points [-]

Sometimes negative emotions are just bad weather -- you have to get stuff done anyways. I also agree with and second sensible advice below on dealing with causes.

Comment author: Manfred 29 October 2014 06:40:29PM 4 points [-]

Joining clubs is good - especially if you're willing to put in enough work for it to be implicitly joining a social scene (unfortuanately, this bit has plenty of caveats, but trial and error sometimes works fine). Do you make music? There are scenes for that. Dance, ditto. Playing card games, ditto.

LW is almost big enough to work for this, actually - certainly if one lives in a big city.

Comment author: cousin_it 29 October 2014 01:16:12PM 5 points [-]

In my experience, "dealing with the negative feelings" is useless, because if you deal with them today and you're still lonely tomorrow, the feelings will just come back. It's better to find people who are interested in the same things as you, and hang out with them.

Comment author: ChristianKl 29 October 2014 11:06:41AM 11 points [-]

Do you feel lonely because you spent your time alone or because you will you don't connect with the people with whom you spend your time?

Two separate problems.

Comment author: dthunt 30 October 2014 07:44:39PM *  3 points [-]

Not feeling connected with people, or, increasingly feeling less connection with people.

I actively socialize myself, and this helps, but the other thing maybe suggests to me I'm doing something wrong.

(Edit: to clarify, my empathy thingy works as well (maybe better) than it ever has, I just feel like the things I crave from social interactions are getting harder to acquire. Like, people "getting" you, or having enough things in common that you can effectively talk about the stuff that interests you. So, like, obviously, one of the solutions there is to hang out with more bright-and-happy CFAR-ish/LW-ish/EA-ish people.)

Comment author: Ben_LandauTaylor 01 November 2014 05:24:55PM 1 point [-]

I found the Nonviolent Communication method extremely helpful for feeling more connected to my friends.

Comment author: ChristianKl 31 October 2014 01:08:51PM 1 point [-]

www.meetup.com can be a good place to find groups of likeminded people.

Comment author: polymathwannabe 29 October 2014 12:38:09AM 3 points [-]

NY Times on the wrongness of political party-related discrimination.

Comment author: Lumifer 30 October 2014 03:39:53PM *  1 point [-]

David Brooks is more or less correct about the US where the two mainstream parties are not very distinguishable. He is entirely wrong about many other places of the world. There are enough countries where someone's political views are "a marker for basic decency".

P.S. I am amused by a piece of incidental research he cites:

For example, political scientists Shanto Iyengar and Sean Westwood gave 1,000 people student résumés and asked them which students should get scholarships. The résumés had some racial cues (membership in African-American Students Association) and some political cues (member of Young Republicans). Race influenced decisions. Blacks favored black students 73 percent to 27 percent, and whites favored black students slightly.

That is called blatant racism and in case of s/black/white/ would be cause for much hand-wringing, soul-searching, and probably obligatory "diversity training" for everyone.

Comment author: fubarobfusco 30 October 2014 08:05:16AM 1 point [-]

I doubt this generalizes very well.

There have clearly been cases in the history of the world where one party made it clear that they really did intend to hurt or kill their perceived opponents. And then, after acceding to power, went on to do just that.

I've seen remarks here on LW from at least one person in a central European country that he or she felt increasingly personally unsafe due to particular political factions in that country producing increasingly violent rhetoric. I would not tell that person that he or she would be wrong to shun people who advocated political violence against him or her.

Here in the U.S., it sure seems that political eliminationist rhetoric (of the "All the Other Party should be killed as traitors" sort) is produced largely as a form of commercial entertainment, not serious political advocacy. But I say that from a position of relative security and privilege ....

Comment author: falenas108 28 October 2014 03:03:09PM 2 points [-]

I keep finding the statistic that "one pint of donated blood can save up to 3 lives!" But I can't find the average number of lives saved from donating blood. Does anyone know/is able to find?

Comment author: ChristianKl 28 October 2014 04:53:11PM 1 point [-]

What do you mean with "lives saved by donating blood" in the first place?

Quantity people who would die without any blood donations


Liters of blood donated

That's not a pretty useful number if you want to make personal decisions based on it. If our Western system would need more blood, raising the incentives for donations isn't that hard.

Comment author: polymathwannabe 28 October 2014 05:04:30PM 1 point [-]

WHO prefers all blood donations to be unpaid:

"Regular, unpaid voluntary donors are the mainstay of a safe and sustainable blood supply because they are less likely to lie about their health status. Evidence indicates that they are also more likely to keep themselves healthy."

Comment author: ChristianKl 28 October 2014 05:39:48PM 1 point [-]

Interesting. So the core question seems to be: "How much value is produced by healthy blood donors making decisions to donate without incentives, compared to blood that's "brought"".

Comment author: Lumifer 28 October 2014 03:11:08PM 1 point [-]

I keep finding the statistic that "one pint of donated blood can save up to 3 lives!"

The expression "can save up to" should immediately trigger your bullshit detector. It's a reliable signal that the following number is meaningless.

Comment author: NancyLebovitz 28 October 2014 01:45:31PM *  3 points [-]

I did a little research to find out whether there are free survey sites that offer "check all answers that apply" questions.

Super Simple Survey probably does, but goddamned if I'll deal with their website to make sure.

On the almost free side, Live Journal enables fairly flexible polls (including checkboxes) for paid accounts, and you can get a paid account for a month for $3. Live Journal is a social media site.

Comment author: Manfred 29 October 2014 06:43:56PM 1 point [-]
Comment author: Salemicus 28 October 2014 10:47:30AM 4 points [-]

It has been experimentally shown that certain primings and situations increase utilitarian reasoning; for instance, people are more willing to give the "utilitarian" answer to the trolley problem when dealing with strangers, rather than friends. Utilitarians like to claim that this is because people are able to put their biases aside and think more clearly in those situations. But my explanation has always been that it's because these setups are designed to maximise the psychological distance between the subject and the harm they're going to inflict - the more people are confronted with the potential consequences of their actions, the less likely they are to make the utilitarian mistake. And now, a new paper suggests that I was right all along! Abstract:

The hypothetical moral dilemma known as the trolley problem has become a methodological cornerstone in the psychological study of moral reasoning and yet, there remains considerable debate as to the meaning of utilitarian responding in these scenarios. It is unclear whether utilitarian responding results primarily from increased deliberative reasoning capacity or from decreased aversion to harming others. In order to clarify this question, we conducted two field studies to examine the effects of alcohol intoxication on utilitarian responding. Alcohol holds promise in clarifying the above debate because it impairs both social cognition (i.e., empathy) and higher-order executive functioning. Hence, the direction of the association between alcohol and utilitarian vs. non-utilitarian responding should inform the relative importance of both deliberative and social processing systems in influencing utilitarian preference. In two field studies with a combined sample of 103 men and women recruited at two bars in Grenoble, France, participants were presented with a moral dilemma assessing their willingness to sacrifice one life to save five others. Participants’ blood alcohol concentrations were found to positively correlate with utilitarian preferences (r = .31, p < .001) suggesting a stronger role for impaired social cognition than intact deliberative reasoning in predicting utilitarian responses in the trolley dilemma. Implications for Greene’s dual-process model of moral reasoning are discussed.

However, given my low opinion of such experiments, perhaps I should be very careful about uncritically accepting evidence that supports my priors.

Comment author: lmm 28 October 2014 07:24:39PM 3 points [-]

I highly doubt the subjects were drunk enough to have trouble figuring out that 5 > 1. So one could equally offer an interpretation that e.g. drunk people answered honestly, while sober people wanted to signal that they were too caring to kill someone under any circumstances.

It's a fascinating result, but I don't think the interpretation is a slam dunk.

Comment author: Coscott 30 October 2014 01:11:12AM -1 points [-]

I doubt this. I conjecture that more people lie and say they would be utilitarian than lie and say they would not be utilitarian. I hope that I would do the utilitarian thing, but I am not sure that I actually would be able to get myself to do it. (Maybe I would be more likely to actually do it if I were drunk)

Comment author: lmm 30 October 2014 01:07:24PM 2 points [-]

On LW sure, being utilitarian is the thing you want to signal here. Ordinary people in a bar? I highly doubt it. Being unwilling to kill is far, far more socially acceptable than the utilitarian answer.

Comment author: Lumifer 28 October 2014 03:27:36PM 0 points [-]

In two field studies with a combined sample of 103 men and women recruited at two bars in Grenoble, France

Field studies are hard work :-D

Comment author: ChristianKl 28 October 2014 04:44:35PM 2 points [-]

They needed the native habitat for the alcohol consumption.

Comment author: NancyLebovitz 28 October 2014 01:31:40PM 2 points [-]

I've been wondering whether utilitarianism undervalues people's loyalty to their own relationships and social networks.

Comment author: Artaxerxes 28 October 2014 10:43:30AM *  8 points [-]

Someone has created a fake Singularity Summit website.

(Link is to MIRI blog post claiming they are not responsible for the site.)

MIRI is collaborating with Singularity University to have the website taken down. If you have information about who is responsible for this, please contact luke@intelligence.org.

Comment author: Gunnar_Zarncke 28 October 2014 08:32:52AM 12 points [-]

Today I had an aha moment when discussing coalition politics (I didn't call it that, but it was) with elementary schoolers, 3rd grade.

As a context: I offer an interdisciplinary course in school (voluntary, one hour per week). It gives a small group of pupils a glimpse of how things really work. Call it rationality training if you want.

Today the topic was pairs and triple. I used analogies from relationships: Couples, parents, friendships. What changes in a relationship when a new element appears. Why do relationships form in the first place? And this revealed differences in how friendships work among boys and among girls. And that in this class at this moment at least the girl friendships were largely coalition politics: "If you do this your are my best friend," or "No we can't be best friends if she it your best friend." For the boys it appears to be at least wquantitatively different. But maybe just the surface differs.

I the end I represented this as graphs (kind of) on the board. And the children were delighted to draw their own coalition diagrams, even abbreviating names by single letters. You wouldn't have bet that these diagrams were from 3rd grade.

Comment author: ChristianKl 29 October 2014 11:26:42AM 2 points [-]

How did you deal with the prospect of one of the kids being emotional hurt by the whole process of being explicit about relationships?

Comment author: Gunnar_Zarncke 29 October 2014 11:53:05AM *  1 point [-]

I of course have an eye on the emotional wellbeing of the children. But I'm not really clear what kind of emotional hurt you mean. Being exposed to e.g. be the loner possibly? I probably wouldn't try it in this relatively direct way if the group weren't that small (4 children) when I can keep the discourse inspirational and playful at all time.

Comment author: ChristianKl 29 October 2014 11:57:06AM 2 points [-]

Being exposed to e.g. be the loner possibly?

Yes. Getting children to openly state: "We can't be best friend because you are best friends with X" seems to ask for trouble but if you have enough presence in the room to keep the discourse inspirational and playful it might be fine.

Comment author: Gunnar_Zarncke 29 October 2014 12:14:52PM 3 points [-]

Ah yes. "We can't be best friend because you are best friends with X" wasn't literally said with respect to someone in the room. Something like that was quoted by a girl as an example thus it wasn't personal in that moment but I assume that it is a real statement too.

Comment author: MrMind 28 October 2014 02:59:05PM 4 points [-]

I wonder what would happen if we trained monkeys to reveal this kind of detalis with us.

Comment author: Emile 28 October 2014 09:46:19PM 9 points [-]

You may be interested in "Chimpanzee Politics", by Frans de Waals (something like that), which is about exactly that (observing a group of Chimps in a zoo, and how their politics and alliances evolves, with a couple coups).

Comment author: MrMind 29 October 2014 08:14:17AM 1 point [-]

Great! Added to my Amazon whislist ;)

Comment author: Gunnar_Zarncke 28 October 2014 05:44:14PM 3 points [-]

But maybe we could. Considering the tricky setups scientists use to compare the intelligence of mice and rats I'd think that it should be possible to devise an experiment which teaches monkeys to reveal their clan structure. I'm thinking along the line of first training association of buttons with clan members (photos) and the allowing to select groups which should get or not get a treat.

Comment author: Evan_Gaensbauer 28 October 2014 07:00:43AM 9 points [-]

I posted a link to the 2014 survey in the 'Less Wrong' Facebook group, and some people commented they filled it out. Another friend of mine started a Less Wrong account to comment that she did the survey, and got her first karma. Now I'm curious how many lurkers become survey participants, and are then incenitivized to start accounts to get the promised karma by commenting they completed it. If it's a lot, that's cool, because having one's first comment upvoted after just registering an account on Less Wrong seems like a way of overcoming the psychological barrier of 'oh, I wouldn't fit in as an active participant on Less Wrong...'

If you, or someone you know, got active on Less Wrong for the first time because of the survey, please reply as a data point. If you're a regular user who has a hypothesis about this, please share. Either way, I'm curious to discover how strong an effect this is, or is not.

Comment author: [deleted] 29 October 2014 05:33:49AM *  3 points [-]

My first comment was after I completed the 2014 survey. I've only been lurking for about a month, and this was the first survey I've participated in.

Comment author: Sjcs 29 October 2014 04:02:39AM 3 points [-]

I have been an on-and-off lurker for ~15 months, and only recently created an account (not because of the survey though). I have participated in both 2013 and 2014's surveys.

Comment author: cursed 28 October 2014 06:51:27AM *  4 points [-]

Those who are currently using Anki on a mostly daily or weekly basis: what are you studying/ankifying?

To start: I'm working on memorizing programming languages and frameworks because I have trouble remembering parameters and method names.

Comment author: philh 28 October 2014 07:09:47PM 3 points [-]

Geography: "what direction [relative to central london] is this tube stop in?", English counties (locations), U.S. states (locations, capitals), Canadian territories and provinces (locations and capitals), countries (locations, capitals, and at some point I'll add flags). (Most of these came from ankiweb originally, but I had to add reverse cards.)

Bayes: conversions between odds, probabilities and decibels (specific numbers and more recently the general formulas)

Miscellaneous: the NATO phonetic alphabet, logs (base 2 of 1.25, 1.5, 1.75, and base 10 of 2 through 9), some words I can never remember how to spell (this turns out not to help), some computer stuff (the order of the arguments in python's datetime.strptime, and the difference between a left join and a right join), some definitions in machine learning, some historical dates (e.g. wars, first moon landing, introduction of the model T), some historical inflation rates, some astronomical facts.

Also a deck based on the twelve virtues of rationality essay. (This one and most of the bayes one I found through LW.)

I'm not sure most of this is useful, but most of it hasn't cost me significant effort either.

Comment author: Coscott 30 October 2014 01:17:08AM 3 points [-]

if you memorize logs, I recommend memorizing natural logs of primes. This is all you need to quickly calculate natural log, log2, and log10 of any integer.

You get ln of any number by adding together the natural logs of the prime factors, and you get log_m of n by the formula

log_m(n)=ln(n)/ln(m)

(maybe memorize ln(10) too to make the calculation a little easier)

Comment author: philh 30 October 2014 10:31:38AM 2 points [-]

I can't do real division in my head, but if I wanted to maximise my logarithm-ability while minimizing my number of cards, I would go for logs base (probably 10) of primes, and 1/log(e) and 1/log(2).

But I'm not too fussed about minimizing cards, or about natural logs. Learning more primes might be helpful, but I can get them approximately. E.g. I don't have log_10(11) memorized, but I know it's between log_10(10) and log_10(2*6) which are 1 and 1.08, and it would be closer to the latter (my calculator says 1.041, which is slightly lower than I would have guessed, but if I put it in Anki I'd only go to 1.04 anyway).

Comment author: Emile 28 October 2014 11:02:01AM *  4 points [-]

These days, most of my time on Anki is on Japanese (which I'm learning for fun) and Chinese (which I already know, but I'm brushing up on tones and characters).

Looking through my decks, I also have decks on:

  • Algorithms and data structures (from a couple books I read on that)
  • Communication (misc. tips on storytelling, giving talks, etc.)
  • Game Design (insights and concepts that seemed valuable)
  • German
  • Git and Unix Command Line commands
  • Haskell
  • Insight (misc. stuff that seemed interesting/important)
  • Mnemonics
  • Productivity (notes from Lukeprog's posts and vairous other sources)
  • Psychology and neuroscience
  • Rationality Habits (one of the few decks I have that come all made, from Anna Salmon I think, though I also added some stuff and delted others)
  • Statistics
  • Web Technologies (some stuff on Angular JS and CSS that I got tired of looking up all the time)

(also a few minor decks with very few cards)

I review those pretty much every day (I sometimes leave a few unfinished, depending on how much idle time I have in queues, transport, etc.)

Comment author: cursed 28 October 2014 09:23:39PM 2 points [-]

That's fantastic. How many cards total do you have, and how many minutes a day do you study?

Comment author: Emile 28 October 2014 10:01:47PM 2 points [-]

Apparently I have 6887 cards (though that includes those I suspended because they're boring, useless, too difficult, duplicated, or possibly wrong; I tend to often suspend cards instead of deleting them); of those around 3000 are Chinese pinyin cards I automatically created with a Python script (I set them up to get between 1 and 5 new ones per day, depending on how busy I tend to be), 1000 are Japanese (the biggest deck of manually-entered cards), and the remaining decks rarely go over 300 cards.

I study probably between 20 and 40 minutes per day, usually in public transit or during "downtime" (waiting in line, carrying the baby around the house hoping for him to sleep, in the restroom, the elevator...). The time depends of how many new cards I entered recently.

Comment author: Evan_Gaensbauer 28 October 2014 12:37:28AM *  3 points [-]

The following model is my new hypothesis for generating better OKCupid profiles for myself while remaining honest.

  • I brainstorm what I want to include in my profile in a positive way without lying. This may include goal-factoring on what honest signals I'm trying to send. Then, I see how what I brainstormed fits into the different prompts on OKCupid profiles.

  • I generate multiple clause-like chunks for each item/object/quality of myself I'm trying to express in my profile. I then A/B test the options for each item across a cross-section of individuals similar to the ones I would want to attract on OKCupid. This may include random assignment to conditions to participants to some extent. I would still need to think of metrics or ratings for this to best suit my goals.

  • Construct complete paragraphs for the various sections of my profile using whichever were the most successful Caveats: I would want enough experimental control to ensure the test participants were people I could trust to respond honestly, and without trolling me. However, this would decrease random selection. How much should I care about random selection, and thus external validity, in this case?

Otherwise, what do you think of the model? What's wrong with it? If it's not completely awful, I'll play-test it with an OKCupid profile just for the value of information, and see if we can't learn something.

Comment author: MrMind 28 October 2014 03:02:03PM 1 point [-]

Just test it and report back the result :) That will teach you and us many things we can't see right now.

Comment author: Capla 28 October 2014 12:02:10AM 2 points [-]

It had never occurred to me that the term "applause light" could be taken so literally.

Comment author: RomeoStevens 29 October 2014 07:10:44PM 2 points [-]

Politician, noun: a person who cheers in-group values professionally.

Comment author: Evan_Gaensbauer 28 October 2014 07:58:30AM 2 points [-]

My friend recently attended an event at which Ray Kurzweil and an urban planner named Richard Florida were speaking. He didn't like Richard Florida as a speaker, citing how Richard Florida 'sounded just like a politician', and was speaking 'only in applause lights'. I noted it was funny to use 'applause light' in that context, as an auditorium where the speaker looks over a crowd while bathed in light, saying things specifically to garner applause, is just about the most literal interpretation of 'applause light' I could think of.

Comment author: Douglas_Knight 28 October 2014 11:29:43PM 3 points [-]

"Applause lights" is a metaphor based on a concrete thing that really exists

Comment author: MathiasZaman 27 October 2014 09:28:54PM *  7 points [-]

I've recently started a tumblr dedicated to teaching people what amounts to Rationality 101. This post isn't about advertising that blog, since the sort of people that actually read Less Wrong are unlikely to be the target audience. Rather, I'd like to ask the community for input on what are the most important concepts I could put on that blog.

(For those that would like to follow this endeavor, but don't like tumblr, I've got a parallel blog on wordpress)

Comment author: dthunt 30 October 2014 08:37:42PM 2 points [-]

Noticing confusion is the first skill I tried to train up last year, and is definitely a big one, because knowing what your models predict and noticing when they fail is a very valuable feedback loop that prevents you from learning if you can't even notice it.

Picturing what sort of evidence would unconvince you of something you actively believe is a good exercise to pair with the exercise of picturing what sort of evidence would convince you of something that seems super unlikely. Noticing unfairness there is a big one.

Realizing when you are trying to "win" at truthfinding, which is... ugh.

Comment author: ruelian 28 October 2014 07:56:44PM 1 point [-]

Map and territory - why is rationality important in the first place?

Comment author: jkadlubo 28 October 2014 12:11:16PM 4 points [-]

Excercises in small rational behaviours. E.g. people genrally are very reluctant to apologize about anything, even if the case means little to them and a lot to the other person. Maybe it's "if I apologize, that will mean I was a bad person in the first place" thinking, maybe something else.

It's a nice excercise: if somebody seems to want something from you or apparently is angry with you when you did nothing wrong, stop for a moment and think: how much will it cost me to just say "I'm sorry, I didn't mean to offend you". After all, those are just words. You don't have to "win" every confrontation and convince the other person you are right and their requirements are ridiculus. And if you apologize, in fact you both will have a better day - the other person will feel appreciated and you will be proud you did something right.

(A common situation from my experience is that somebody pushes me in a queue, I say "excuse me, but please don't stand so close to me/don't look over my arm when I'm writing the PIN code etc." and then the pusher often starts arguing how my behaviour is out of line - making both of us and the cashier upset)

Come to think of it, it's a lot like Quirrell's second lesson in HPMoR...

Comment author: Manfred 27 October 2014 10:49:27PM 2 points [-]

Taking stock of what information you have, and what might be good sources for information, well in advance of making a decision.

Comment author: wadavis 27 October 2014 10:18:57PM 11 points [-]

Admitting you are wrong.

Comment author: Manfred 27 October 2014 10:48:28PM 12 points [-]

Highly related: When you even might be wrong, get curious about that possibility rather than scared of it.

Comment author: Jackercrack 27 October 2014 07:14:21PM *  7 points [-]

I'd like to ask LessWrong's advice. I want to benefit from CFAR's knowledge on improving ones instrumental rationality, but being a poor graduate I do not have several thousand in disposable income nor a quick way to acquire it. I've read >90% of the sequences but despite having read lukeprog's and Alicorn's sequences I am aware that I do not know what I do not know about motivation and akrasia. How can I best improve my instrumental rationality on the cheap?

Edit: I should clarify, I am asking for information sources: blogs, book recommendations, particularly practice exercises and other areas of high quality content. I also have a good deal of interest in the science behind motivation, cognitive rewiring and reinforcement. I've searched myself and I have a number of things on my reading list, but I wanted to ask the advice of people who have already done, read or vetted said techniques so I can find and focus on the good stuff and ignore the pseudoscience.

Comment author: gjm 29 October 2014 12:12:48AM 5 points [-]

(Apologies for the slight thread hijack here.)

It occurs to me that CFAR's model of expensive workshops and generous grants to the impoverished (note: I am guessing about the generosity) is likely to produce rather odd demographics: there's probably a really big gap between (1) the level of wealth/income at which you could afford to go, and (2) the level of wealth/income at which you would feel comfortable going, especially as -- see e.g. cursed's comments in this thread -- it's reasonable to have a lot of doubt about whether they're worth the cost. (The offer of a refund mitigates that a bit.)

Super-handwavy quantification of the above: I would be really surprised if a typical person whose annual income is $30k or more were eligible for CFAR financial aid. I would be really surprised if a typical person whose income is $150k or less were willing to blow $4k on a CFAR workshop. (NB: "typical". It's easy to imagine exceptions.) Accordingly, I would guess that a typical CFAR workshop is attended mostly by people in three categories: impoverished grad students, etc., who are getting big discounts; people on six-figure salaries, many of them quite substantial six-figure salaries; and True Believers who are exceptionally convinced of the value of CFAR-style rationality, and willing to make a hefty sacrifice to attend.

I'm not suggesting that there's anything wrong with that. In fact, it strikes me as a pretty good recipe for getting an interesting mix of people. But it does mean there's something of a demographic "hole".

Comment author: Jackercrack 29 October 2014 06:11:52AM 2 points [-]

I rather think there may be demand for a cheaper, less time dependent method of attending. It may be several seasons before they end up back in my country for example. Streaming/recording the whole thing and selling the video package seems like it could still get a lot of the benefits across. Their current strategy only really makes sense to me if they're still in the testing and refining stage.

Comment author: dthunt 30 October 2014 08:51:26PM 0 points [-]

You can always shoot someone an email and ask about the financial aid thing, and plan a trip stateside around a workshop if, with financial aid, it looks doable, and if after talking to someone, it looks like the workshop would predictably have enough value that you should do it now rather than when you have more time and money.

Comment author: ChristianKl 29 October 2014 11:54:14AM 1 point [-]

Their current strategy only really makes sense to me if they're still in the testing and refining stage.

I think they are. If everything goes well they will have published papers that proves that their stuff works by the time they move out of the testing and refining stage.

Comment author: Jackercrack 29 October 2014 11:51:31PM 1 point [-]

Any idea how long that will be (months, years, decades)?

Comment author: cursed 28 October 2014 06:50:06AM 9 points [-]

I've been to several of CFAR's classes throughout the last 2 years (some test classes and some more 'official' ones) and I feel like it wasn't a good use of my time. Spend your money elsewhere.

Comment author: hyporational 28 October 2014 11:12:04AM 4 points [-]

What made it poor use of your time?

Comment author: cursed 28 October 2014 09:15:02PM *  15 points [-]

I didn't learn anything useful. They taught, among other things, "here's what you should do to gain better habits". Tried it and didn't work on me. YMMV.

One thing that really irked me was the use of cognitive 'science' to justify their lessons 'scientifically'. They did this by using big scientific words that felt like they were trying to attempt to impress us with their knowledge. (I'm not sure what the correct phrase is - the words weren't constraining beliefs? don't pay rent? they could have made up scientific sounding words and it would have had the same effect.)

Also, they had a giant 1-2 page listing of citations that they used to back up their lessons. I asked some extremely basic questions about papers and articles I've previously read on the list and they had absolutely no idea what I was talking about.

ETA: I might go to another class in a year or two to see if they've improved. Not convinced that they're worth donating money towards at this moment.

Comment author: Unnamed 30 October 2014 01:24:29AM 7 points [-]

(This is Dan from CFAR again)

We have a fair amount of data on the experiences of people who have been to CFAR workshops.

First, systematic quantitative data. We send out a feedback survey a few days after the workshop which includes the question "0 to 10, are you glad you came?" The average response to that question is 9.3. We also sent out a survey earlier this year to 20 randomly selected alumni who had attended workshops in the previous 3-18 months, and asked them the same question. 18 of the 20 filled out the survey, and their average response to that question was 9.6.

Less systematically but in more fleshed out detail, there are several reviews that people who have attended a CFAR workshop have posted to their blogs (A, B+pt2, C +pt2) or to LW (1, 2, 3). Ben Kuhn's (also linked above under "C") seems particularly relevant here, becaue he went into the workshop assigning a 50% probability to the hypothesis that "The workshop is a standard derpy self-improvement technique: really good at making people feel like they’re getting better at things, but has no actual effect."

In-person conversations that I've had with alumni (including some interviews that I've done with alumni about the impact that the workshop had on their life) have tended to paint a similar picture to these reviews, from a broader set of people, but it's harder for me to share those data.

We don't have as much data on the experiences of people who have been to test sessions or shorter events. I suspect that most people who come to shorter events have a positive experience, and that there's a modest benefit on average, but that it's less uniformly positive. Partly that's because there's a bunch of stuff that happens with a full workshop that doesn't fit in a briefer event - more time for conversations between participants to digest the material, more time for one-on-one conversations with CFAR staff to sort through things, followups after the workshop to work with someone on implementing things in your daily life, etc. The full workshop is also more practiced and polished (it has been through many more iterations) - much moreso than a test session; one-day events are in between (the ones advertised as alpha tests of a new thing are closer to the test session end of the spectrum).

Comment author: cursed 02 November 2014 01:35:27PM *  0 points [-]

Hey Dan, thanks for responding. I wanted to ask a few questions:

You noted the non-response rate for the 20 randomly selected alumni. What about the non-response rate for the feedback survey?

"0 to 10, are you glad you came?" This is a biased question, because you frame that the person is glad. A similar negative question may say "0 to 10, are you dissatisfied that you came?" Would it be possible to anonymize and post the survey questions and data?

We also sent out a survey earlier this year to 20 randomly selected alumni who had attended workshops in the previous 3-18 months, and asked them the same question. 18 of the 20 filled out the survey, and their average response to that question was 9.6.

It's great that you're following up with people long after the workshops end. Why not survey all alumni? You have their emails.

I've read most of the blog posts about CFAR workshops that you linked to - they were one of my main motivations for attending a workshop. I notice that all reviews are from people who have already participated in LessWrong and related communities. (all refer to some prior CFAR, EA and rationality related topics before they attended camp). Also, it seems like in person conversations are majorly subjected to the availability bias, as the people who attended workshops || know people who work at MIRI/CFAR || are involved in LW meetups in Berkeley and surrounding areas would contribute to the positivity of these conversations.. Also, the evaporative cooling effect may also play a role, in that people who weren't satisfied with the workshop would leave the group. Are there reviews from people who are not already familiar with LW/CFAR staff?

Also, I agree with MTGandP. It would be nice if CFAR could write a blog post or paper on how effective their teachings are, compared to a control group. Perhaps two one-day events, with subjects randomized across both days, should work well as a starting point.

Comment author: MTGandP 30 October 2014 02:41:47AM *  3 points [-]

We send out a feedback survey a few days after the workshop which includes the question "0 to 10, are you glad you came?" The average response to that question is 9.3.

I've seen CFAR talk about this before, and I don't view it as strong evidence that CFAR is valuable.

  • If people pay a lot of money for something that's not worth it, we'd expect them to rate it as valuable by the principle of cognitive dissonance.
  • If people rate something as valuable, is it because it improved their lives, or because it made them feel good?

For these ratings to be meaningful, I'd like to see something like a control workshop where CFAR asks people to pay $3900 and then teaches them a bunch of techniques that are known to be useless but still sound cool, and then ask them to rate their experience. Obviously this is both unethical and impractical, so I don't suggest actually doing this. Perhaps "derpy self-improvement" workshops can serve as a control?

Comment author: Unnamed 30 October 2014 01:17:06AM 6 points [-]

(Dan from CFAR here)

Hi cursed - glad to hear your feedback, though I'm obviously not glad that you didn't have a good experience at the CFAR events you went to.

I want to share a bit of information from my point of view (as a researcher at CFAR) on 1) the role of the cognitive science literature in CFAR's curriculum and 2) the typical experience of the people who come to a CFAR workshop. This comment is about the science; I'll leave a separate comment about thing 2.

Some of the techniques that CFAR teaches are based pretty directly on things from the academic literature (e.g., implementation intentions come straight from Peter Gollwitzer's research). Some of our techniques are not from the academic literature (e.g., the technique that we call "propagating urges" started out in 2011 as something that CFAR co-founder Andrew Critch did).

The not-from-the-literature techniques have been through a process of iteration, where we theorize about how we think the technique works, then (with the aid of our best current model) we try to teach people to use the technique, and then we get feedback on how it goes for them. Then repeat. The "theorizing" step of this process includes digging into the academic literature to get a better understanding of how the relevant parts of the mind work, and that often plays a role in shaping the class. With "propagating urges," at first none of the people that Critch taught it to were able to get it to work for them, but then Critch made a connection to some neuroscience he'd been reading, we updated our model of how the technique was supposed to work, and then more people were able to make use of the technique. (I'm tempted to go into more specifics here, but that feels like a tangent and this comment is going to be long enough without it.)

Classes based on from-the-academic-literature techniques also go through a similar process of iteration. For example, there are a lot of studies that have shown that people who are instructed to come up with implementation intentions for a particular goal make more progress towards that goal. But I don't know of any academic research on attempts to teach people the skill of being able to create implementation intentions, and the habit of actually using them in day-to-day life. And that's what we're trying to do at CFAR workshops, so that class has gone through a similar process of iteration as we get feedback on whether people are making use of implementation intentions and how it goes for them. (One simple change that helped get more people to use implementation intentions: giving the technique a different name. We now call it "trigger action planning").

So the cognitive science literature plays both of these roles for us: it's a source of evidence about particular techniques that have been tested and found to work (or to not work), and it's a source of models of how the mind works so that we can develop better techniques. We mention both of these types of scientific references in class (and in the further resources), and we try to be careful to distinguish them. Sharing our models in class (e.g., saying a few sentences in the propagating urges class about what we think the orbitofrontal cortex might be doing in this process) seems to be helpful for getting people to use the technique as we understand it (rather than getting confused about the steps, or rounding the technique off to the nearest cached thought). It also seems to help with getting people to take ownership of the technique and treat it as something that they can tinker with, rather than as a rote series of steps for them to follow (cf. learned blankness).

Finally, a brief comment on this:

Also, they had a giant 1-2 page listing of citations that they used to back up their lessons. I asked some extremely basic questions about papers and articles I've previously read on the list and they had absolutely no idea what I was talking about.

Each CFAR class has one staff member who takes the lead in developing the class, and I'm the research specialist who does a lot of digging into the literature and sharing/discussing research with whoever is developing the class. The aim is for the two of us to be conversant in the relevant academic literature. For the rest of the CFAR team, the priority is to be able to use the techniques and help other people use them, not to know all the studies. (Often there will be more than just us two puzzling things over together, but it typically isn't the whole team.) The instructor who teaches a class at a CFAR event isn't always the person who has been developing it, especially at one-day events which are just being run by 2 instructors instead of the full CFAR staff. If I'd been at the event you came to, the instructor who you asked about the articles probably would've referred you to me and we could've had an interesting conversation.

Comment author: Jackercrack 29 October 2014 04:44:07AM 3 points [-]

Do you think it was unhelpful because you already had a high level of knowledge on the topics they were teaching and thus didn't have much to learn or because the actual techniques were not effective? Do you think your experience was typical? How useful do you think it would be to an average person? An average rationalist?

Comment author: cursed 29 October 2014 05:12:00AM 7 points [-]

Do you think it was unhelpful because you already had a high level of knowledge on the topics they were teaching and thus didn't have much to learn or because the actual techniques were not effective?

I don't believe I had a high level of knowledge on the specific topics they were teaching (behavior change, and the like). I did study some cognitive science in my undergraduate years, and I take issue with the 'science'.

Do you think your experience was typical?

I believe that the majority of people don't get much, if anything, from CFAR's rationality lessons. However, after the lesson, people may be slightly more motivated to accomplish whatever they want to, in the short term just because they've paid money towards a course to increase their motivation.

How useful do you think it would be to an average person?

There was one average person at one of the workshops I attended. e.g. never read LessWrong/other rationality material. He fell asleep a few hours into the lesson, I don't think he gained much from attending. I'm hesitant to extrapolate, because I'm not exactly sure what an average person entails.

An average rationalist?

I haven't met many rationalists, but would believe they wouldn't benefit much/at all.

Comment author: Jackercrack 29 October 2014 06:05:28AM 2 points [-]

Well that's a bit dispiriting, though I suppose looking back my view of CFAR was a bit unrealistic. Downregulating chance that CFAR is some kind of panacea.

Comment author: RomeoStevens 27 October 2014 08:34:19PM 4 points [-]

CFAR has financial aid.

Also, attending LW meetups and asking about organizing meetups based on instrumental rationality material is cheap and fun.

Comment author: Jackercrack 27 October 2014 09:56:20PM 2 points [-]

Somehow I doubt the financial aid will stretch to the full amount, and my student debt is already somewhat fearsome.

I'm on the LW meetups already as it happens. I'm currently attempting to have my local one include more instrumental rationality but I lack a decent guide of what methods work, what techniques to try or what games are fun and useful. For that matter I don't know what games there are at all beyond a post or two I stumbled upon.

Comment author: Vaniver 27 October 2014 10:28:23PM 5 points [-]

Somehow I doubt the financial aid will stretch to the full amount, and my student debt is already somewhat fearsome.

You could ask Metus how much they covered for them, or someone at CFAR how much they'd be willing to cover. The costs for asking are small, and you won't get anything you don't ask for.

Comment author: Jackercrack 27 October 2014 10:57:55PM 3 points [-]

Fair point, done. On a related note, I wonder how I can practice convincing my brain that failure does not mean death like it did in the old ancestral environment.

Comment author: ChristianKl 29 October 2014 10:52:18AM 3 points [-]

CFAR suggests doing exercises to extend your comfort zone for that purpose.

Comment author: NancyLebovitz 28 October 2014 02:27:29AM 3 points [-]

Even in the ancestral environment, not all failures (I suspect a fairly small proportion of them) meant death.

Comment author: Metus 27 October 2014 11:16:54PM 7 points [-]

Exposure therapy: Fail on small things, then larger ones, where it is obvious that failiure doesn't mean death. First remember past experiences where you failed and did not die, then go into new situations.

Comment author: Strangeattractor 27 October 2014 06:39:38PM 1 point [-]

After reading through the Quantum Physics sequence, I would like to know more about the assumptions and theories behind the idea that an amplitude distribution factorizes, or approximately factorizes. Where would be a good place to learn more about this? I would appreciate some recommendations for journal articles to read, or specific sections of specific books, or if there's another better way to learn this stuff, please let me know.

In the blog posts in the sequence, an analogy comes up a few times, saying that it doesn't make sense to distinguish between the two factors of 3 when multiplying 3 x 3 x 2 to get 18, and that similarly, the amplitude blobs in configuration space that can sometimes appear to be like particles are factors of...I'm not sure what. The wavefunction? A probability density function (but we're calling it amplitude instead of probabilities)? Something else? I didn't entirely follow that section, so I'm not sure how to look it up.

When I searched on Google Scholar for "quantum factorization" I got journal articles about how to use quantum computers to factor prime numbers. When I looked up "particle indistinguishability" I got papers about very small numbers of particles in a state of quantum entanglement. When I searched for "amplitude distribution factorizes" I got articles about tomography and mesons and keys for quantum cryptography.

I'm also confused about: what precisely is an amplitude distribution? Amplitude of what? Distributed over what? I can make some guesses, but how do I look it up?

I would also like to know: what needs to be true in order for this concept to be true? Does it depend on the many worlds interpretation of quantum mechanics, or would it hold true in the other interpretations? Does it require the wavefunction? Just how good is the analogy about the factors of 18, and where would the analogy break down? What do the equations look like that lead to these conclusions, and what are they called, so I can look them up? What assumptions are used to formulate the equations? What is the difference between factorizing exactly and approximately? Why does the idea of roughly factorizing come up at all, why isn't it all exact? How accurate would it be to describe a person as a factor of the wave function, and what does that mean? Is there a technical term for "blob of amplitude"?

The Quantum Physics sequence is the best introduction to quantum mechanics that I've read, but it is rather incomplete about explicitly stating the assumptions it is using, or giving references to where to learn more about each topic.

Help?

Comment author: Manfred 27 October 2014 10:52:40PM *  3 points [-]

Relevant wikipedia link. The keyword is something like "many-body wavefunction."

But seriously, if you're curious, try to find a textbook online, or a series of video lectures for an introductory course (you might either watch the whole course, or skip to what you want to learn and then try and figure out what the prerequisites are, then do the same thing for the prerequisites).

Comment author: DanielLC 27 October 2014 08:47:12PM 0 points [-]

I think the factorization is a reference to from quantum field theory. I haven't learned quantum field theory though, so I can't comment much. From what I can gather, multiplying something by the creation operator gets you the same state but with an extra particle.

I can tell you that at the very minimum, assuming Copenhagen and the minimal amount of physics to allow entanglement to happen at all, whenever two of the same kind of particle are entangled, they have a 50% chance of swapping. If you use MWI, it's that I can find a universe with the same probability density in which those particles are swapped.

Comment author: James_Miller 27 October 2014 06:08:05PM 10 points [-]

Assume that Jar S contains just silver balls, whereas Jar R contains ninety percent silver balls and ten percent red balls.

Someone secretly and randomly picks a jar, with an equal chance of choosing either. This picker then takes N randomly selected balls from his chosen jar with replacement. If a ball is silver he keeps silent, whereas if a ball is red he says “red.”

You hear nothing. You make the straightforward calculation using Bayes’ rule to determine the new probability that the picker was drawing from Jar S.

But then you learn something. The red balls are bombs and if one had been picked it would have instantly exploded and killed you. Should learning that red balls are bombs influence your estimate of the probability that the picker was drawing from Jar S?

I’m currently writing a paper on how the Fermi paradox should cause us to update our beliefs about optimal existential risk strategies. This hypothetical is attempting to get at whether it matters if we assume that aliens would spread at the speed of light killing everything in their path.

Comment author: jkaufman 30 October 2014 04:34:35PM *  3 points [-]

This is related to the Sleeping Beauty Problem, and in general the answer depends what you're trying to do with "probability". For lots and lots more, Bostrom's PhD thesis is very detailed: Anthropic Bias: Observation Selection Effects in Science and Philosophy.

Bostrom's Observation Selection Effects and Human Extinction Risks paper is less philosophical and sounds like it's more relvant to the paper you're working on.

Comment author: Manfred 27 October 2014 11:06:35PM *  0 points [-]

If the two jar scenarios start with equal anthropic measure (i.e. looking in from the outside), then you really are less likely to have jar R if you're not dead.

Comment author: Lumifer 27 October 2014 06:45:09PM 3 points [-]

A side note: under the cherry bomb scenario the probability of you hearing the word "red" is zero.

Comment author: private_messaging 27 October 2014 06:31:17PM 7 points [-]

I had a conversation with another person regarding this Leslie's firing squad type stuff. Basically, I came up with a cavemen analogy with the cavemen facing lethal threats. It's pretty clear - from the outside - that the cavemen which do probability correctly and don't do anthropic reasoning with regards to tigers in the field, will do better at mapping lethal dangers in their environment.

Comment author: James_Miller 28 October 2014 01:57:13PM 2 points [-]

Thanks for letting me know about "Leslie's firing squad[s]"

Comment author: private_messaging 28 October 2014 03:17:10PM *  3 points [-]

You're welcome. So what's your actual take on the issue? I never seen a coherent explanation why bombs must make a difference. I seen appeals to "but you wouldn't be thinking anything if it was red", which ought to perfectly cancel out if you apply that to the urn choice as well.

edit: i.e. this anthropics, to me, is sort of like how you could calculate the forces in a mechanical system, but make an error somewhere, and that yields an apparent perpetuum mobile, as forces on your wheel with water and magnets fail to cancel out. Likewise, you evaluate impacts of some irrelevant information, and you make an error somewhere, and irrelevant information makes a difference.

Comment author: James_Miller 29 October 2014 06:26:25PM 1 point [-]

To a first approximation I don't think it makes a difference, but it does add some logical uncertainty. Also, intuitively I want to be able to use anthropic reasoning to say "there is only a tiny chance that the universe would have condition X, but I'm not surprised by X because without X observers such as us won't exist", but I think doing this implies I have to give a different estimate if red = bomb.

Comment author: private_messaging 13 December 2014 04:50:50AM 3 points [-]

Also, intuitively I want to be able to use anthropic reasoning to say "there is only a tiny chance that the universe would have condition X, but I'm not surprised by X because without X observers such as us won't exist"

Hmm, that's an interesting angle on the issue, I didn't quite realize that was the motivation here.

I would be surprised by our existence if that was the case, and not further surprised by observation of X (because I already observed X by the way of perceiving my existence).

Let's say I remember that there was an strange, surprising sign painted on the wall, and I go by the wall, and I see that sign, and I am surprised that there's that sign on the wall at all, but I am not surprised that I am seeing it (because I can perform an operation in my head that implies existence of the sign - my memory tells me I seen it before). Same with the existence, I am surprised we exist at all but I am not surprised when I observe something necessary for my existence because I could've derived it from prior observations.

Comment author: jnarx 30 October 2014 08:26:37PM 1 point [-]

I think this particular example doesn't really exemplify what I think you're trying to demonstrate here.

A simpler example would be:

You draw one ball our of a jar containing 99% red balls and 1% silver balls (randomly mixed).

The ball is silver. Is this surprising? Yes.

What if you instead draw a ball in a dark room so you can't see the color of the ball (same probability distribution). After drawing the ball, you are informed that the red balls contain a high explosive, and if you draw a red ball from the jar it would instantly explode, killing you.

The lights go on. You see that you're holding a silver ball. Does this surprise you?

Comment author: private_messaging 13 December 2014 04:41:59AM 2 points [-]

Well, being alive would surprise me, but not the colour of the ball. Essentially what happens is that the internal senses (e.g. perceiving own internal monologue) end up sensing the ball colour (by the way of the high explosive).

Comment author: polymathwannabe 27 October 2014 06:13:18PM 2 points [-]

Before I actually do the math, "you hear nothing" appears to affect my estimate exactly in the same way as "you're still alive."

Comment author: Kindly 13 December 2014 04:34:42PM 0 points [-]

This seems like the obvious answer to me as well. What am I missing?

Comment author: polymathwannabe 13 December 2014 08:49:51PM *  -1 points [-]

Now that I see this problem again, my thoughts on it are slightly different.

In the version with no bombs, there's a possible scenario where the picker draws a red ball but lies to you by keeping silent. So, there's a viable way for "you hear nothing" AND "Jar R" to happen.

But in the version with bombs, the scenario with "you are alive" AND "Jar R" can never happen. So, being alive in the with-bomb version is stronger evidence for Jar S than hearing nothing in the no-bomb version.

Comment author: Kindly 14 December 2014 04:05:39AM 0 points [-]

Okay, sure. The picker could be lying or speaking quietly; the bomb could be malfunctioning or have a timer that hasn't gone off yet. (Note to self: put down the ball as soon as you find out that it could be a bomb.) These things don't seem like they should be the point of a thought experiment.

Comment author: ruelian 27 October 2014 05:13:24PM 8 points [-]

I have a question for anyone who spends a fair amount of their time thinking about math: how exactly do you do it, and why?

To specify, I've tried thinking about math in two rather distinct ways. One is verbal and involves stating terms, definitions, and the logical steps of inference I'm making in my head or out loud, as I frequently talk to myself during this process. This type of thinking is slow, but it tends to work better for actually writing proofs and when I don't yet have an intuitive understanding of the concepts involved.

The other is nonverbal and based on understanding terms, definitions, theorems, and the ways they connect to each other on an intuitive level (note: this takes a while to achieve, and I haven't always managed it) and letting my mind think it out, making logical steps of inference in my head, somewhat less consciously. This type of thinking is much faster, though it has a tendency to get derailed or stuck and produces good results less reliably.

Which of those, if any, sounds closer to the way you think about math? (Note: most of the people I've talked to about this don't polarize it quite so much and tend to do a bit of both, i.e. thinking through a proof consciously but solving potential problems that come up while writing it more intuitively. Do you also divide different types of thinking into separate processes, or use them together?)

The reason I'm asking is that I'm trying to transition to spending more of my time thinking about math not in a classroom setting and I need to figure out how I should go about it. The fast kind of thinking would be much more convenient, but it appears to have downsides that I haven't been able to study properly due to insufficient data.

Comment author: RichardKennaway 28 October 2014 09:51:03AM 1 point [-]

Which of those, if any, sounds closer to the way you think about math?

Each serves its own purpose. It is like the technical and artistic sides of musical performance: the technique serves the artistry. In a sense the former is subordinate to the latter, but only in the sense that the foundation of a building is subordinate to its superstructure. To perform well enough that someone else would want to listen, you need both.

This may be useful reading, and the essays here (from which the former is linked).

Comment author: ruelian 28 October 2014 04:53:12PM 1 point [-]

reads the first essay and bookmarks the page with the rest

Thanks for that, it made for enjoyable and thought-provoking reading.

Comment author: Bundle_Gerbe 28 October 2014 07:36:22AM 1 point [-]

As someone with a Ph.D. in math, I tend to think verbally in as much as I have words attached to the concepts I'm thinking about, but I never go so far as to internally vocalize the steps of the logic I'm following until I'm at the point of actually writing something down.

I think there is another much stronger distinction in mathematical thinking, which is formal vs. informal. This isn't the same distinction as verbal vs. nonverbal, for instance, formal thinking can involve manipulation of symbols and equations in addition to definitions and theorems, and I often do informal thinking by coming up with pretty explicitly verbal stories for what a theorem or definition means (though pictures are helpful too).

I personally lean heavily towards informal thinking, and I'd say that trying to come up with a story or picture for what each theorem or definition means as you are reading will help you a lot. This can be very hard sometimes. If you open a book or paper and aren't able to get anywhere when you try do this to the first chapter, it's a good sign that you are reading something too difficult for your current understanding of that particular field. At a high level of mastery of a particular subject, you can turn informal thinking into proofs and theorems, but the first step is to be able to create stories and pictures out of the theorems, proofs, and definitions you are reading.

Comment author: Fhyve 28 October 2014 07:18:47AM 1 point [-]

I'm a math undergrad, and I definitely spend more time in the second sort of style. I find that my intuition is rather reliable, so maybe that's why I'm so successful at math. This might be hitting into the "two cultures of mathematics", where I am definitely on the theory builder/algebraist side. I study category theory and other abstract nonsense, and I am rather bad (relative to my peers) at Putnam style problems.

Comment author: Luke_A_Somers 27 October 2014 10:40:48PM 3 points [-]

I don't tend to do a lot of proofs anymore. When I think of math, I find it most important to be able to flip back and forth between symbol and referent freely - look at an equation and visualize the solutions, or (to take one example of the reverse) see a curve and think of ways of representing it as an equation. Since when visualizing numbers will often not be available, I tend to think of properties of a Taylor or Fourier series for that graph. I do a visual derivative and integral.

That way, the visual part tells me where to go with the symbolic part. Things grind to a halt when I have trouble piecing that visualization together.

Comment author: ruelian 28 October 2014 04:55:11PM 1 point [-]

This appears to be a useful skill that I haven't practiced enough, especially for non-proof-related thinking. I'll get right on that.

Comment author: Gunnar_Zarncke 27 October 2014 09:52:21PM 1 point [-]

I don't see a clear verbal vs. non-verbal dichotomy - or at least the non-verbal side has lots of variants. To gain an intuitive non-verbal understanding can involve

  • visual aids (from precise to vague): graphs, diagrams, patterns (esp. repetitions), pictures, vivid imagination (esp. for memorizing)

  • acoustic aids: rhythms (works with muscle memory too), patterns in the spoken form, creating sounds for elements

  • abstract thinking (from precise to vague): logical inference, semantic relationships (is-a, exists, always), vague relationships (discovering that the more of this seems to imply the more of that)

Note: Logical inference seems to be the verbal part you mean, but I don't think symbolic thinking is always verbal. Its conscious derivation may be though.

And I hear that the verbal side despite lending itself to more symbolic thinking can nonetheless work its grammar magic on an intuitive level too (though not for me).

Personally if I really want to solve a mathematical problem I immerse myself in it. I try lots of attack angles from the list above (not systematically but as it seems fit). I'm an abstract thinker and don't rely on verbal, acoustic or motor cues a lot. Even visual aids don't play a large role though I do a lot of sketching, listing/enumerating combinations, drawing relations/trees, tabulating values/items. If I suspect a repeating pattern I may tap to it to sound it out. If there is lengthy logical inference involved that I haven't internalized I speak the rule repeatedly to use the acoustic loop as memory aid. I play around with it during the day visualizing relationships or following steps, sometimes until in the evening everyting blurs and I fall asleep.

Comment author: TsviBT 27 October 2014 09:46:38PM 1 point [-]

Personally, the nonverbal thing is the proper content of math---drawing (possibly mental) pictures to represent objects and their interactions. If I get stuck, I try doing simpler examples. If I'm still stuck, then I start writing things down verbally, mainly as a way to track down where I'm confused or where exactly I need to figure something out.

Comment author: lmm 27 October 2014 07:36:52PM 1 point [-]

I don't really draw that distinction. I'd say that my thinking about mathematics is just as verbal as any other thinking. In fact, a good indication that I'm picking up a field is when I start thinking in the language of the field (i.e. I will actually think "homology group" and that will be a term that means something, rather than "the group formed by these actions...")

Comment author: ruelian 27 October 2014 08:13:33PM 1 point [-]

I'd say that my thinking about mathematics is just as verbal as any other thinking.

Just to clarify, because this will help me categorize information: do you not do the nonverbal kind of thinking at all, or is it all just mixed together?

Comment author: lmm 27 October 2014 10:26:07PM 1 point [-]

I'm not really conscious of the distinction, unless you're talking about outright auditory things like rehearsing a speech in my head. The overwhelming majority of my thinking is in a format where I'm thinking in terms of concepts that I have a word for, but probably not consciously using the word until I start thinking about what I'm thinking about. Do you have a precise definition of "verbal"? But whether you call it verbal or not, it feels like it's all the same thing.

Comment author: ruelian 28 October 2014 04:50:20PM 1 point [-]

I don't really have good definitions at this point, but in my head the distinction between verbal and nonverbal thinking is a matter of order. When I'm thinking nonverbally, my brain addresses the concepts I'm thinking about and the way they relate to each other, then puts them to words. When I'm thinking verbally, my brain comes up with the relevant word first, then pulls up the concept. It's not binary; I tend to put it on a spectrum, but one that has a definite tipping point. Kinda like a number line: it's ordered and continuous, but at some point you cross zero and switch from positive to negative. Does that even make sense?

Comment author: lmm 28 October 2014 07:00:46PM 1 point [-]

It makes sense but it doesn't match my subjective experience.

Comment author: ruelian 28 October 2014 07:33:15PM 1 point [-]

Alright, that works too. We're allowed to think differently. Now I'm curious, could you define your way of thinking more precisely? I'm not quite sure I grok it.

Comment author: lmm 29 October 2014 11:24:33PM 1 point [-]

So, I'd say there are three modes of thinking I can identify:

  • Normal thinking, what I'm doing the vast majority of the time. I'm thinking by manipulating concepts, which are just, well, things.
  • Introspective thinking, where I'm doing the first kind of thinking, and thinking about it. Because the map can't be the territory, when I'm thinking about thinking the concepts I'm thinking about are represented by something simpler than themselves - if you're thinking about thinking about sheep then the sheep you're thinking about thinking about can't be as complex as the sheep you're thinking about. In fact they're represented either by words, or by something isomorphic to words - labels for concepts. So when I'm thinking about thinking, the thinking-about-thinking is verbal - but the thinking isn't (although there's a light-in-the-fridge effect that might make one think it was).
  • Auditory thinking, where I'm thinking in words in my head, planning a speech (or more likely a piece of writing - and most of the time I never actually write or say it). This is the only kind of thinking I'm conscious of doing that really feels verbal, but it feels sensory rather than thinking in words; I'm hearing a voice in my cartesian theater.
Comment author: Strangeattractor 27 October 2014 06:52:59PM 3 points [-]

I usually think about math nonverbally. I am not usually doing such thinking to come up with proofs. My background is in engineering, so I got a different sort of approach to math in my education about math than the people who were in the math faculty at the university I attended.

Sometimes I do go through a problem step by step, but usually not verbally. I sometimes make notes to help me remember things as I go along. Constraints, assumptions, design goals, etc. Explicitly stating these, which I usually do by writing them on paper, not speaking them aloud, if I'm working by myself on a problem, can help. But sometimes I am not working by myself and would say them out loud to discuss them with other people.

Also, there is often more than one way to visualize or approach a problem, and I will do all of them that come to mind.

I would suggest, to spend more time thinking about math, find something that you find really beautiful about math and start there, and learn more about it. Appreciate it, and be playful with it. Also, find a community where you can bounce ideas around and get other people's thoughts and ideas about the math you are thinking about. Some of this stuff can be tough to learn alone. I'm not sure how well this advice might work, your mileage may vary.

When I am really understanding the math, it seems like it goes directly from equations on the paper right into my brain as images and feelings and relations between concepts. No verbal part of it. I dream about math that way too.

Comment author: ruelian 27 October 2014 07:26:20PM 2 points [-]

I only got to a nonverbal level of understanding of advanced math fairly recently, and the first time I experienced it I think it might have permanently changed my life. But if you dream about math...well, that means I still have a long way to go and deeper levels of understanding to discover. Yay!

Follow-up question (just because I'm curious): how do you approach math problems differently when working on them from the angle of engineering, as opposed to pure math?

Comment author: Strangeattractor 28 October 2014 07:57:17AM 3 points [-]

It seemed to me that the people I knew who were studying pure math spent a lot of time on proofs, and that math was taught to them with very little context for how the math might be used in the real world, and without a view as to which parts were more important than others.

In engineering classes we proved things too, but that was usually only a first step to using the concepts to work on some other problem. There was more time spent on some types of math than on others. Some things were considered to be more useful and important than others. Usually some sort of approximations or assumptions would be used, in order to make a problem simpler and able to be solved, and techniques from different branches of math were combined together whenever useful, often making for some overlap in the notation that had to be dealt with.

There was also the idea that any kind of math is only an approximate model of the true situation. Any model is going to fail at some point. Every bridge that has been built has been built using approximations and assumptions, and yet most bridges stay up. Learning when one can trust the approximations and assumptions is vital. People can die if you get it wrong. Learning the habit of writing down explicitly what the assumptions and approximations are, and to have a sense for where they are valid and where they are not, is a skill that I value, and have carried over into other aspects of my life.

Another thing is that math is usually in service of some other goal. There are design constraints and criteria, and whatever math you can bring in to get it done is welcome, and other math is extraneous. The beauty of math can be admired, but a kludgy theory that is accurate to real world conditions gets more respect than a pretty theory that is less accurate. In fact, sometimes engineers end up making kludgy theory that solves engineering problems into some sophisticated mathematics that looks more formal and has some interesting properties, and then it has a beauty of its own, although some of the beauty comes from knowing how it fits into a real world phenomenon.

Also, engineers tend to work in teams, not alone. So communicating with each other, and making sure that all the people on the team have a similar understanding of a situation, is a non-trivial part of the work. You don't want a situation where one person has one type of abstraction in their head, and another person has a different one, and they don't realize it, and when they go off to do their separate work, it doesn't match up. This can lead to all sorts of problems, not limited to cost overruns, design flaws, delays, and even deaths. So, if you hear engineers discussing nitpicky details and going over technical concepts more than once, that is one major reason why. You really need people to be on the same page.

Teamwork is so important to engineering that when taking classes, we were encouraged to talk to each other and work together on problems, before submitting answers. Whereas the people over in math were forbidden to talk to each other about their work before handing it in. That policy might be different at different schools. But I think it shows an important difference in culture.

Math is certainly something that can be enjoyed and practiced solo. But especially on some of the most tricky concepts of math that I have learned, I benefitted a lot from being able to discuss it with people, and get new insights and understanding from their perspectives. Sometimes I didn't even realize that I didn't properly understand a concept until I attempted to use it, and got a completely different answer from someone else who was attempting to use it.

I said it can get kludgy, and that the focus is on real world problems, but there are times when it does feel clean and pure, especially when people make real world objects that correspond pretty well to ideal mathematical objects. For example, using 4th-order differential equations to calculate the bending moments for I-beams felt peaceful and pretty, once I got the hang of it, and I think not it is not unlike something you might find in a pure math course.

I'm pretty enthusiastic about math, it's one of my favourite things to think about and do.

Comment author: wadavis 27 October 2014 06:50:48PM 2 points [-]

As someone employed doing mid-level math (structural design), I'm much like most others you've talked to. The entirely non-verbal intuitive method is fast, and it tends to be highly correct if not accurate. The verbal method is a lot slower, but it lends itself nicely to being put to paper and great for getting highly accurate if not correct answers. So everything that matters gets done twice, for accurate correct results. Of course, because it is fast the intuitive method is prefered for brainstorming, then the verbal method verifies any promising brainstorms.

Comment author: ruelian 27 October 2014 07:28:04PM 2 points [-]

Could you please explain what you mean by "correct" and "accurate" in this case? I have a general idea, but I'm not quite sure I get it.

Comment author: wadavis 27 October 2014 08:14:41PM 1 point [-]

Correct and Precise may have been better terms. By correct I mean a result that I have very high confidence in, but that is not precise enough to be useable. By accurate I mean a result that is very precise but with far less confidence that it is correct.

As an example, consider a damped oscillation word problem from first year. You are very confident that as time approaches infinity that the displacement will approach a value just by looking at it, but you don't know that value. Now when you crunch the numbers (the verbal process in the extreme) you get a very specific value that the function approaches, but have less confidence that that value is correct, you could have made any of a number of mistakes. In this example the classic wrong result is the displacement is in the opposite direction as the applied force.

This is a very simple example so it may be hard to separate the non-verbal process from the verbal, but there are many cases where you know what the result should look like but deriving the equations and relations can turn into a black box.

Comment author: ruelian 27 October 2014 08:40:03PM *  2 points [-]

Right, that makes much more sense now, thanks.

One of my current problems is that I don't understand my brain well enough for nonverbal thinking not to turn into a black box. I think this might be a matter of inexperience, as I only recently managed intuitive, nonverbal understanding of math concepts, so I'm not always entirely sure what my brain is doing. (Anecdotally, my intuitive understanding of a problem produces good results more often than not, but any time my evidence is anecdotal there's this voice in my head that yells "don't update on that, it's not statistically relevant!")

Does experience in nonverbal reasoning on math lend actually itself to better understanding of said reasoning, or is that just a cached thought of mine?

Comment author: wadavis 27 October 2014 10:02:24PM 1 point [-]

Doing everything both ways, nonverbal and verbal, has lent itself to better understanding of the reasoning. Which touches on the anecdote problem, if you test every nonverbal result; you get something statistically relevant. If your odds are more often than not with nonverbal, testing every result and digging for the mistakes will increase your understanding (disclaimer: this is hard work).

Comment author: ruelian 28 October 2014 05:23:58PM 2 points [-]

So, essentially, there isn't actually any way of getting around the hard work. (I think I already knew that and just decided to go on not acting on it for a while longer.) Oh well, the hard work part is also fun.

Comment author: RowanE 27 October 2014 06:38:09PM 2 points [-]

I'm only a not-very-studious undergraduate (in physics), and don't spend an awful lot of time thinking about maths ourside of that, but I pretty much only think about maths in the nonverbal way - I can understand an idea when verbally explained to me, but I have to "translate it" into nonverbal maths to get use out of it.

Comment author: Artaxerxes 27 October 2014 04:11:35PM 12 points [-]

Luke's IAMA on reddit's r/futurology in 2012 was pretty great. I think it would be cool if he did another, a lot has changed in 2+ years. Maybe to coincide with the December fundraising drive?

Comment author: [deleted] 28 October 2014 08:20:26AM 3 points [-]

If he could not repeat the claim that UFAI is so easily compressible it could "spread across the world in seconds" through the internet, that would be quite helpful, actually. Even in the rich world, with broadband, transferring an intelligent agent all across the world will take whole hours, especially given the time necessary for the bugger to crack into and take control of the relevant systems (packaging itself as a trojan horse and uploading itself to 4chan in a "self-extracting zip" of pornography will take even longer).

Comment author: Evan_Gaensbauer 28 October 2014 06:53:20AM 3 points [-]

I just sent a message to Luke. Hopefully he will notice it.

Comment author: Lumifer 27 October 2014 04:11:03PM 5 points [-]

An good semi-rant by Ken White of popehat on GamerGate. I recommend it as an excellent example of applied rationality and sorting out through the hysterics.

Comment author: Azathoth123 30 October 2014 06:08:18AM 0 points [-]

Meh, he seems to be trying to hard to pretend to be wise. One of the more egregious examples:

There's no excuse for threats to anyone, whatever "side" they are on. Posting someone's home address or private phone number or financial details will almost never be relevant to a good-faith dispute — it's clearly intended to terrorize, and it risks empowering disturbed people to do real harm. These things are wrong no matter who does them, no matter the motive, and no mater the target.

That's like saying "Violence is wrong no matter who does it, therefore if an armed gang invades your neighborhood you should passively comply".

Comment author: Lumifer 30 October 2014 02:36:52PM 2 points [-]

he seems to be trying to hard to pretend to be wise

He does a better job of it than most people I know :-) It's not that I completely agree with him, but he writes well and makes a lot of valid points.

The biggest objection to his position is, essentially, Yvain's post on weak men.

Comment author: Azathoth123 04 November 2014 05:53:52AM 0 points [-]

Sarah Hoyt has a good description of the problems with the post here.

Comment author: Lumifer 04 November 2014 06:18:15AM 2 points [-]

Meh. It's a rant and not a particularly well-thought-out one.

Comment author: Omid 27 October 2014 03:56:13PM 8 points [-]

What chores do I need to learn how to do in order to keep a clean house?

Comment author: Risto_Saarelma 01 November 2014 03:43:59PM 2 points [-]

Learn to notice things that need cleaning.

Know a good way to get rid of everything you possess when you no longer need it (bookcrossing, electronic waste recycling or just a trash bag). Learn to notice when you have things cluttering up the place that you no longer need.

Comment author: hyporational 28 October 2014 10:57:30AM *  0 points [-]

If you've got the money and a simple enough apartment layout, I recommend a vacuum cleaning robot. My crawling saucer collects a ridiculous amount of dust from the floor every day, and this seems to keep other surfaces and the air dustless too. There's no way I could clean up that much dust myself, and I'd do the cleaning so rarely that the dust would get all over the place.

Comment author: RichardKennaway 28 October 2014 09:56:37AM 0 points [-]

Avoid these and you'll be off to a good start. :)

Comment author: Manfred 27 October 2014 11:37:30PM *  3 points [-]

Adding on to Emily:

Having a particular hamper or even corner of your room where you put dirty laundry, so that it isn't all over your floor. When this hamper / corner is full, do your laundry.
Analogous organized or occasionally-organized places for paperwork or whatever else is being clutter-y.
If you have ancient carpet and it's dirty and stinky, learn how to rent a Rug Doctor-type steam cleaner from a nearby supermarket.
If you have a bunch of broken or dirty / stinky stuff in your house, learn how to get the trash people to haul it away, and learn where to buy cheap used furniture / cheap online kitchen supplies / whatever to replace your old junk.
Having tools handy to tidy up nails / tighten loose screws etc. when you notice them.
Keeping a bush and plunger near your toilet.
If your sink has clogged any time in the past 6 months, also consider having chemical unclogger / a long skinny "snake" (that's what it's actually called) that you shove down the drain and wiggle around to bust clogs.
Figure out where all the places that are hard to clean are. These are the places that will have 50 years of accumulated nasty dirt that will make the whole house smell better when you get rid of it.

Comment author: Emily 27 October 2014 04:03:06PM 12 points [-]

Laundry (plus ironing, if you have clothes that require that - I try not to), washing up (I think this is called doing the dishes in America), mopping, hoovering (vacuuming), dusting, cleaning bathroom and kitchen surfaces, cleaning toilets, cleaning windows and mirrors. That might cover the obvious ones? Seems like most of them don't involve much learning but do take a bit of getting round to, if you're anything like me.

Comment author: RichardKennaway 27 October 2014 04:13:52PM 9 points [-]

I'd add, not leaving clutter lying around. It both collects dust, and makes cleaning more of an effort. Keep it packed away in boxes and cupboards. (Getting rid of clutter entirely is a whole separate subject.)

Comment author: Omid 27 October 2014 04:11:18PM *  1 point [-]

Thank you, how many hours a week do you spend doing these things?