Open Thread, May 11 - May 17, 2015

3 Post author: Gondolinian 11 May 2015 12:16AM

If it's worth saying, but not worth its own post (even in Discussion), then it goes here.


Notes for future OT posters:

1. Please add the 'open_thread' tag.

2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)

3. Open Threads should be posted in Discussion, and not Main.

4. Open Threads should start on Monday, and end on Sunday.

Comments (247)

Comment author: ChristianKl 13 May 2015 01:48:43PM 14 points [-]

Recent news suggest that having measles weakens the immune system for 2 to 3 years afterwards and therefore the Measles vaccine manages to reduce a lot of childhood deaths that weren't thought to be measles related.

NPR article Academic paper

Comment author: Toggle 11 May 2015 02:35:54AM 9 points [-]

It looks like AI is overtaking Arimaa, which is notable because Arimaa was created specifically as a challenge to AI. Congratulations to the programmer, David Wu.

Comment author: Kindly 12 May 2015 03:21:36PM 1 point [-]

On the subject of Arimaa, I've noted a general feeling of "This game is hard for computers to play -- and that makes it a much better game!"

Progress of AI research aside, why should I care if I choose a game in which the top computer beats the top human, or one in which the top human beats the top computer? (Presumably both the top human and the top computer can beat me, in either case.)

Is it that in go, you can aspire (unrealistically, perhaps) to be the top player in the world, while in chess, the highest you can ever go is a top human that will still be defeated by computers?

Or is it that chess, which computers are good at, feels like a solved problem, while go still feels mysterious and exciting? Not that we've solved either game in the sense of having solved tic-tac-toe or checkers. And I don't think we should care too much about having solved checkers either, for the purposes of actually playing the game.

Comment author: jacob_cannell 11 May 2015 05:49:13AM 1 point [-]

I hadn't heard of Armaa before, but based on about 5 minutes worth of reading about the game, I don't understand how it is significantly more suited to natural reasoning that chess. It inherits many of chess's general features that make serial planning more effective than value function knowledge - thus favoring fast thinkers over slow deep thinkers. Go is much more of a natural reasoning game.

Comment author: ShardPhoenix 12 May 2015 09:24:38AM 0 points [-]

From the sound of it, the AI works more or less like chess AI - search with a hand-tuned evaluation function.

Comment author: NancyLebovitz 11 May 2015 04:27:29PM *  7 points [-]

Transhumanism in the real world

Rugby players who get a bottle opener to replace a missing tooth.

Comment author: Lumifer 11 May 2015 04:39:11PM -1 points [-]

That's no more transhumanism than this.

Comment author: drethelin 11 May 2015 06:59:09PM 5 points [-]

False! It's adding functionality rather than just a cosmetic change.

Comment author: Lumifer 11 May 2015 07:03:42PM *  1 point [-]

Cosmetic changes can be highly functional. Ask any girl :-)

On a slightly more serious note, I tend to think of tranhumanist modifications as ones which confer abilities that unenhanced humans do not have. Opening beer bottles isn't one of them.

Comment author: Epictetus 12 May 2015 01:27:20AM *  4 points [-]

Having been in a group of drunk people who found that they had no bottle opener, and having seen what bizarre ideas they concoct to get the bottles open, I'd say a bottle opener in one's tooth merits the status of transhumanist modification.

Comment author: [deleted] 12 May 2015 07:42:50AM 1 point [-]

There was a saying in my youth: "There is no item that is not a beer opener." There was a bit of a competition for creative moves (drinking beer was considered a high status adult move for teenagers, opening them in creative ways even more). Keys. Lighters. Doors, the part where the "tongue" goes in on the frame, not sure the English term. Edges of tables or edges of anything. Using two bottles, locking the caps to each other and pulling apart. I still consider it the coolest manly way to open a beer when you sit at a fairly invulnerable e.g. stone table to just put the cap against the edge and hit it. Another 101 ways.

Comment author: Nornagest 13 May 2015 11:30:53PM *  0 points [-]

Doors, the part where the "tongue" goes in on the frame, not sure the English term.

Strike plate?

Comment author: TezlaKoil 12 May 2015 12:23:59AM *  4 points [-]

Would you consider a Wikipedia brain implant to be a transhumanist modification? After all, ordinary humans can query Wikipedia too!

Comment author: Lumifer 12 May 2015 02:48:26PM 1 point [-]

Would you consider a Wikipedia brain implant to be a transhumanist modification?

That's a weird way of putting it. Would I consider an implant which consists of a large chunk of memory with some processing and an efficient neural interface to be transhumanist? Yes, of course. It will give a lot of useful abilities and just filling it with Wikipedia looks like a waste of potential.

I don't think trivializing transhumanism to minor cosmetics is a useful approach. Artificial nails make better screwdrivers than natural nails, so is that also a transhumanist modification?

Comment author: [deleted] 13 May 2015 07:57:57PM 1 point [-]

But nigh-effortless verbatim memorization is, so if you carry around a pen and a pad of paper...

Comment author: Lumifer 13 May 2015 08:14:53PM 0 points [-]

"nigh-effortless"

"a pen and a pad of paper"

Ahem.

Comment author: RowanE 11 May 2015 06:49:43PM 4 points [-]

Well, you know what they say about "one man's modus ponens".

Comment author: Lumifer 11 May 2015 06:56:27PM *  2 points [-]

Here is your first transhumanist then, from pre-Columbian Maya...

Comment author: RowanE 12 May 2015 04:41:27PM 1 point [-]

I think the attitude toward the modifications is a relevant factor as well - wanting to be "more than human" in some respect, even if only a trivial respect such as "more awesome-looking than a regular human" or "more able to open beer bottles than a regular human" - but given that, yeah I'd be totally on board with considering some pre-Columbian Maya or other stone-age person "the first transhumanist".

Comment author: Lumifer 12 May 2015 04:53:30PM *  1 point [-]

I see a smooth transition into tool-using, then. Picking up a stick certainly gives you more capabilities compared to a stick-less hominid, and probably makes you much more awesome as well.

Comment author: Lumifer 14 May 2015 02:54:37PM *  6 points [-]

Pretty awesome set of trolley problems

Sample:

There’s an out of control trolley speeding towards Immanuel Kant. You have the ability to pull a lever and change the trolley’s path so it hits Jeremy Bentham instead. Jeremy Bentham clutches the only existing copy of Kant’s Groundwork of the Metaphysic of Morals. Kant holds the only existing copy of Bentham’s The Principles of Morals and Legislation. Both of them are shouting at you that they have recently started to reconsider their ethical stances.

Comment author: Romashka 11 May 2015 06:34:15PM 5 points [-]

(Random) by analogy with what Quirrellmort led people to believe about his imminent death, it would be cool to read a fic, 6th year AU, in which Dumbledore teaches defence...

Comment author: knb 14 May 2015 08:51:50PM 2 points [-]

What does 6th year AU mean?

Comment author: zedzed 15 May 2015 04:37:46AM 2 points [-]

6th year = book 6 (6th year at Hogwarts) = Half-Blood Prince.

AU = Alternate Universe.

Comment author: Dahlen 12 May 2015 12:33:59AM 4 points [-]

Is utilitarianism foundational to LessWrong? Asking because for a while I've been toying with the idea of writing a few posts with morality as a theme, from the standpoint of, broadly, virtue ethics -- with some pragmatic and descriptive ethics thrown in. (The themes are quite generous and interlocking, and to be honest I don't know where to start or whether I'll finish it.) This perspective treats stable character traits, with their associated emotions, drives, and motives as the most reasonably likely determiner of moral behaviour, and means to encourage people to "build character" so as to become more moral beings or improve their behaviour. It doesn't concern itself with quantitative approaches to welfare. Frankly, I find it hard to take seriously the numerical applications of utilitarianism, and my brain just shuts down upon some ethical problems usually enjoyed around here (torture vs. dust specks, repugnant conclusion, contrived deals with strange gods and so on).

I know that Eliezer's virtues-of-rationality post is widely appreciated by many people around here, but it's a declaration of (commitment to) values more than anything. It never seemed to be the dominant paradigm. I guess I just want to know whether a virtue-ethical approach would be well-received here, and the extent to which a utilitarian and a virtue ethicist can usefully discuss morality without jumping a meta level into theories of normative ethics.

Comment author: ChristianKl 12 May 2015 01:37:48AM 5 points [-]

If it helps you, the 2014 census gave for moral beliefs:

Moral Views Accept/lean towards consequentialism: 901, 60.0% Accept/lean towards deontology: 50, 3.3% Accept/lean towards natural law: 48, 3.2% Accept/lean towards virtue ethics: 150, 10.0% Accept/lean towards contractualism: 79, 5.3% Other/no answer: 239, 15.9%

Meta-ethics Constructivism: 474, 31.5% Error theory: 60, 4.0% Non-cognitivism: 129, 8.6% Subjectivism: 324, 21.6% Substantive realism: 209, 13.9%

In general I don't think there are foundational ideas on LW that shouldn't be questioned. Any idea is up for investigation provided the case is well argued.

Comment author: falenas108 12 May 2015 04:13:29AM 6 points [-]

In general I don't think there are foundational ideas on LW that shouldn't be questioned. Any idea is up for investigation provided the case is well argued.

But there are certain ideas that will be downvoted and dismissed because people feel like they aren't useful to be talking about, like if God exists. I think OP was asking if it was a topic that fell under this category.

Comment author: ChristianKl 12 May 2015 11:32:30AM 4 points [-]

But there are certain ideas that will be downvoted and dismissed because people feel like they aren't useful to be talking about, like if God exists.

The problem with "does God exist" isn't about the fact that LW is atheist. It's that it's hard to say interesting things about the subject and provide a well argued case.

I don't expect to learn something new when I read another post about whether or not God exists. If someone knows the subject well enough to tell me something new, then there no problem with them writing a post to communicate that insight.

Comment author: ilzolende 13 May 2015 03:20:10AM 4 points [-]

I endorse discussion of virtue ethics on LW mostly because I haven't seen many arguments for why I should use it or discussions of how using it works. I've seen a lot of pro-utilitarianism and "how to do things with utilitarianism" pieces and a lot of discussion of deontology in the form of credible precommitments and also as heuristics and rule utilitarianism, but I haven't really seen a virtue ethics piece that remotely approaches Yvain's Consequentialism FAQ in terms of readability and usability.

Comment author: pianoforte611 12 May 2015 01:43:48AM 4 points [-]

When you say virtue ethics, it sounds like you are describing consequentialism implemented on human software.

If we're talking about the philosopher's virtue ethics, this question should clarify: Are virtues virtuous because they lead to moral behavior? Or is behavior moral because it cultivates virtue?

The first is just applied consequentialism. The second is the philosopher's virtue ethics.

Comment author: Dahlen 12 May 2015 03:02:11AM 1 point [-]

The thing is... that's really beyond the scope of what I care to argue about. I understand the difference, but it's so small as to not be worth the typing time. It's precisely the kind of splitting hairs I don't want to go into.

The theme that would get treated is morality, not ethics. It kind of starts off assuming that it is self-evident why good is good, and that human beings do not hold wildly divergent morals or have wildly different internal states in the same situation. Mostly. Sample topics that I'm likely to touch on are: rationality as wisdom; the self-perception of a humble person and how that may be an improvement from the baseline; the intent with which one enters an interaction; a call towards being more understanding to others; respect and disrespect; how to deflect (and why to avoid making) arguments in bad faith; malicious dispositions, and more. Lots of things relevant to community maintenance.

These essays aren't yet written, so perhaps that's why it all sounds (and is) so chaotic. There may be more topics which conflict more obviously with utilitarianism, especially if there's a large number of individuals concerned. As for conflicts with consequentialism, they're less likely, but still probable.

Comment author: pianoforte611 12 May 2015 07:26:34PM 1 point [-]

If you don't want to talk about the difference then I respect that, and I wasn't suggesting that you do. If anything I would suggest avoiding the term "virtue ethics" entirely and instead talking about virtue which is more general and a component of most moral systems.

I disagree that it is splitting hairs though or a small difference. It makes a large whether or not you wish to cultivate virtue for its own sake (regardless or independent of consequence), or because it helps you achieve other goals. The latter makes fewer assumptions about the goals of your reader.

Comment author: BrassLion 13 May 2015 02:03:08AM 3 points [-]

Consequentialism, where morality is viewed through a lens of what happens due to human actions, is a major part of LessWrong. Utilitarianism specifically, where you judge an act by the results, is a subset of consequentialism and not nearly as widely accepted. Virtue Ethics are generally well liked and it's often said around here that "Consequentialism is what's right, Virtue Ethics are what works." I think that practical guide to virtue ethics would be well received.

Comment author: Vaniver 12 May 2015 01:41:25PM 3 points [-]

Is utilitarianism foundational to LessWrong?

No. Individual utility calculations are, as a component of decision theory, but decision-theoretic utility and interpersonal-comparison utility are different things with different assumptions.

encourage people to "build character" so as to become more moral beings or improve their behaviour.

This is a solid view, and one of the main ones I take--but I observe that listing out goals and developing training regimens have different purposes and uses.

Comment author: [deleted] 12 May 2015 08:30:24AM 2 points [-]

I think virtue ethics is sufficiently edgy, new, different these days to be interesting. Go on.

Comment author: IlyaShpitser 12 May 2015 09:41:06AM 8 points [-]

new

I agree, scholarship is a problem.

Comment author: [deleted] 12 May 2015 11:33:42AM 3 points [-]

Okay, ancient enough, but fell into disuse around the Enlightenment, was hardly considered 100-120 years ago, returned among academic philosophers like Philippa Foot, Catholics like MacIntyre tryed to keep it alive, and it is only roughly about now that it is something slowly considered again by the hip young atheist literati classes for whom karma is merely a metaphor and do not literally believe in bad deeds putting a stain on the soul, so in this sense it is only a newly fashionable thing again.

Comment author: Gunnar_Zarncke 14 May 2015 06:24:50AM 1 point [-]

Again I recommend a poll:

Is utilitarianism foundational to LessWrong? (use the middle option to see results only)

Agree Disagree

Submitting...

Comment author: Vaniver 12 May 2015 07:02:42PM 0 points [-]

Also, have you read this post? The virtue tag only points at it and one other, but searching will likely find more.

Comment author: taygetea 17 May 2015 10:57:13PM 3 points [-]

Hi. I don't post much, but if anyone who knows me can vouch for me here, I would appreciate it.

I have a bit of a Situation, and I would like some help. I'm fairly sure it will be positive utility, not just positive fuzzies. Doesn't stop me feeling ridiculous for needing it. But if any of you can, I would appreciate donations, feedback, or anything else over here: http://www.gofundme.com/usc9j4

Comment author: John_Maxwell_IV 11 May 2015 02:08:07PM *  3 points [-]

So there was recently an advance related to chips for running neural networks. I'm having a hard time figuring out if we should be happy or sad. I'm not sure if this qualifies as a "computing power" advance or a "cell modeling" one.

Comment author: Houshalter 14 May 2015 05:56:16AM *  0 points [-]

I doubt it will help scientists reverse engineer the function of the brain any faster. However it could potentially be very useful for helping AI researchers develop artificial neural networks.

ANNs aren't really tied to neuroscience research, and they probably won't help with emulations. But they are the current leading approach to AI, and increased computing power would significantly help AI research, as it has in the past.

Comment author: Error 11 May 2015 01:29:47AM 3 points [-]

Just posted this in the previous open thread; reposting here: Has anyone here used fancyhands.com or a similar personal-assistant service? If so, what was your experience like?

(context: I have anxiety issues making phone calls to strangers and certain other ugh fields, and am thinking I may be better off paying someone else to take care of such things rather than trying to bull through the ugh fields.)

Comment author: [deleted] 16 May 2015 08:25:13PM 1 point [-]

Even with my own business, I found it incredibly hard to find lots of tasks which I could hand off to an assistant from whom I didn't know what their strengths were, and couldn't train.

Comment author: dxu 12 May 2015 02:58:56AM *  5 points [-]

To any physicists out there:

This idea came to me while I was replaying the game Portal. Basically, suppose humanity one day developed the ability to create wormholes. Would one be able to generate an infinite amount of energy by placing one end of a wormhole directly below the other before dropping an object into the lower portal (thus periodically resetting said object's gravitational potential energy while leaving its kinetic energy unaffected)? This seems like a blatant violation of the first law of thermodynamics, so I'm guessing it would fail due to some reason or other (my guess goes to weird behavior of the gravitational field near the wormhole, which interferes with the larger field of the Earth), but since I'm nowhere close to being a physicist, I thought I'd ask about it on LessWrong.

So? Any ideas as to what goes wrong in the above example?

Comment author: CellBioGuy 12 May 2015 03:20:37AM 4 points [-]

Gravity is a conserved vector field. Any closed path through a gravitational potential leaves you with the same energy you started with. And if it doesn't you've stolen energy that was creating the gravity in the first place leaving less for the next circle to take and are thus just transforming energy from one to another.

Comment author: DanielLC 12 May 2015 05:37:41AM 0 points [-]

Does that apply when the space isn't simply connected? It could be conservative at every neighborhood, but not conservative over all if you allow portals.

Comment author: CellBioGuy 12 May 2015 01:32:54PM *  2 points [-]

Gravity would propagate through the connected space. The potential would probably be very VERY weird shaped but i see no reason it wouldnt be conservative or otherwise consistent with GR (the math of which is far beyond me). Though keep in mind in GR space curvature IS gravity and can change over time, I doubt you could maintain a knife edge thin aperture, it would all smooth out.

What's really fun though are gravetomagnetic effects.

These are to gravity what magnetism is to the electric field. Both the electric field and gravitational field are conservative. But changing or accelerating charges generate magnetic fields which are NOT conservative hence how an electron spinning around a coil in a field in a generator gains energy even though it returns to its starting point. However in doing so it accelerates up to velocity as well generating a counteracting field that cancels out some of the field accelerating it, the motion moving it, or both. Thus the nonconservative fields have a potential energy associated with them that can be extracted from them or used to couple two phenomena that are both coupled to them.

To get gravetomagnetic effects you need huuuuuge mass flows and accelerations. But you can similarly steal the energy that drives them. Think frame dragging and extraction of black hole rotation.

Comment author: DanielLC 12 May 2015 06:39:01PM 0 points [-]

but i see no reason it wouldnt be conservative or otherwise consistent with GR

I'm assuming GR holds. Does it actually prove the field is conservative, or just irrotational? If it's irrotational and simply connected, then it's conservative, but if you stick a portal in it, it might not be.

I doubt you could maintain a knife edge thin aperture, it would all smooth out.

Portals don't need a knife edge. For example: Flight Through a Wormhole. You do need negative energy density or it will collapse, but that on its own shouldn't break conservation of energy.

Comment author: shminux 14 May 2015 07:08:51PM *  2 points [-]

I have a PhD in Physics and my thesis was, in part, related to wormholes, so here it goes. (Squark covered most of your question already, though.)

If something falls into a black hole, it increases the black hole mass. If something escapes a black hole (such as Hawking radiation), it decreases the black hole mass. Same with white holes. A wormhole is basically two black/white holes connected by a throat. One pass through the portal would increase the mass of the entrance and decrease the mass of the exit by the mass of the passing object.

A portal with two ends having opposite masses would behave rather strangely: they sort of repel (the equivalent of Newton's law of gravity), but the gravitational force acting on the negative-mass end propels it toward the positive-mass end. As a result, the portal as a whole will tend to accelerate toward the positive end (entrance) and fly away, albeit rather slowly.

In addition, due to momentum and angular momentum conservation, the portal will start spinning to counteract the motion of the passing object.

Comment author: Squark 14 May 2015 06:19:13PM *  2 points [-]

Wormholes don't quite behave like portals in the game.

When something drops into a wormhole with zero velocity, the apparent mass of the entry end increases by the mass of the object and the apparent mass of the exit end decreases by the mass of the object. At some point one of the ends should acquire negative mass. I'm not sure what that means: either it literally behaves as a negative mass object or this is an indication of the wormhole becoming unstable and collapsing.

Similarly, when something with momentum drops into a wormhole, the momentum is added to the apparent momentum of the entry end and subtracted from the apparent momentum of the exit end. The apparent masses change in a way that ensures energy conservation. This means that the gain in energy of the "cycling" object comes from wormhole mass loss and transfer of mass from the high end to the low end. Again, if it's true that the wormhole becomes unstable when its mass is supposed to go negative, that would be the end of the process.

Comment author: shminux 14 May 2015 07:14:11PM 1 point [-]

If you already postulate having enough negative energy to create a wormhole, there is no extra issues due to one of the throats having negative mass, except the weird acceleration effect, as I mentioned in my other reply.

Comment author: Squark 07 June 2015 07:13:15PM 0 points [-]

Maybe. However, how will the geometry look like when the sign flip occurs? Will it be non-singular?

Comment author: shminux 07 June 2015 09:23:57PM 0 points [-]

There isn't as much difference between negative- and positive-mass wormholes as between negative- and positive mass black holes. Negative-mass black holes have no horizons and a naked repulsive timelike singularity. A negative- (at infinity) mass wormhole would look basically like a regular wormhole. The local spacetime curvature would, of course, be different, but the topology would remain the same, S^2xRxR or similar.

Comment author: ZankerH 14 May 2015 11:28:46AM 1 point [-]

At a glance, it seems like you're asking for extrapolation from a "suppose X - therefore X" - type statement, where X is the invalidation of conservation laws.

Comment author: dxu 15 May 2015 04:12:38AM *  1 point [-]

I don't quite understand this statement. The only real premise I can see in my original comment is

suppose humanity one day developed the ability to create wormholes.

(Please feel free to correct me if you were in fact referring to some other premise.)

Wormholes are generally agreed to be a possible solution to Einstein's equations--they don't, in and of themselves, violate conservation of energy. The scenario I proposed above is a method for generating infinite energy if physics actually worked that way, but since I'm confident that it doesn't, the proposed scenario is almost certainly flawed in some way. I asked my question because I wasn't sure how it was flawed. Whatever the flaw is, however, I doubt it lies in the wormhole premise.

EDIT: Also see the replies from shminux and Squark.

Comment author: Slider 18 May 2015 03:22:20PM 0 points [-]

I am just taking wormholes to mean "altered connectivity of space" and leave out the "massive concentrations of mass" aspect.

The curious thing about portals portals is that they somehow magically know to flip gravity when a object travels thourht. If the portal is just ordinary space there shouldn't be a sudden gradient to the gravity field but it should go smoothly from one direction to the other. And in additon gravity ought to work throught portals. that would mean that if you have a portal in a ceiling it ought to pull stuff throught it towards the ceiling (towards the center of mass beyond the portal). That is a standard "infinite fall" portal setup should feel equal gravity up and down midway between the portals. That kind of setup could be used to store kinetic energy but it doesn't generate it per se.

However if portals aftected the gravity fields it could be that the non-standard gravity environment could be a major problem and would work even when you didn't want it to. That is since the net 0 gravity point of a infinite fall setup needs to transition smoothly to the "standard gravity environment" that likely means that quite a ways "outside" the portal pair there would be a reduced gravity environment.

Comment author: falenas108 12 May 2015 04:11:26AM 0 points [-]

You can probably think about it as the lines of a gravity field also going through the wormhole, and I believe the gravitational force would be 0 around the wormhole.

The actual answer involves thinking about gravity and spacetime as a geometry, which I don't think you want to answer your question.

Comment author: OrphanWilde 13 May 2015 07:13:30PM 2 points [-]

Anybody care to weigh in on adding a flag to newbies, and make it part of the LessWrong culture to explain downvotes to flagged newbies?

Identifying what you've done incorrectly to provoke downvotes is a skill that requires training. (Especially since voting behavior in Discussion is much less consistent to voting behavior in Main.)

Comment author: Gunnar_Zarncke 14 May 2015 06:29:57AM 2 points [-]

You can detect newbies by their low karma and moderate positive ratio. The registration age doesn't mean much really.

Comment author: philh 14 May 2015 01:39:04PM *  0 points [-]

You have to click through to discover that though, and there are exceptions who have a low ratio but don't need downvotes explained to them. (I don't know if there are such users with a low ratio and low total, though.)

Comment author: Gunnar_Zarncke 14 May 2015 03:46:37PM 0 points [-]

You can see the ration in the tool tip over the karma.

Interestingly there are users with positive ratios arbitrarily close to 50%.

Comment author: philh 14 May 2015 05:31:34PM 2 points [-]

You can see a comment's karma ratio by hovering on the comment's karma, but to see the user's ratio you need to click through to the user's page and hover over their karma there.

Comment author: Epictetus 11 May 2015 06:56:34PM 2 points [-]

For some time I've been thinking about just how much of our understanding of the world is tied up in stories and narratives.

Let's take gravity. Even children playing with balls have a good idea of where a ball is going to land after they throw it. They don't know anything about spacetime curvature or Newton's laws. Instead, they amass a lot of data about the behavior of previously-thrown balls and from this they can predict where a newly-thrown ball will land. With experience, this does not even require conscious thought--a skilled ball-player is already moving into position by the time he's consciously aware of what's happening.

You can do the same thing with computers. Once you have enough raw data, you can tabulate it and use various methods to make predictions. These can range from simple interpolation to more complicated statistical modeling. The point is that you don't need any deeper understanding of the underlying phenomena to make it work. You can get good results as long as the phenomena are nice enough and the initial conditions aren't far removed from the data you used to construct the model. Going back to balls, you'll do fine predicting how a ball thrown by a human will behave, but the methods will probably fail if you shot a ball out of a cannon.

So far, so good. You can treat actual phenomena as a black box and make predictions based only on initial conditions, and this works for everyday life. Yet, we feel a drive to explain things. We like to come up with stories. Most of these are silly, but largely harmless. Sometimes, though, we happen upon a useful story or analogy. These stories transcend the role of explanations and enable us to make predictions outside of our accumulated data. Aristotle's gravity didn't have a detrimental effect on the engineering of the day due to the aforementioned use of experience, but Newton's gravity let us push things so much further.

Modern physics is full of these stories which are wrong, but make for good enough analogies to be useful. Take continuum mechanics. We know matter is made of atoms and molecules, but we sometimes assume it's continuous. Then we take this continuous matter and assume it's made up of tiny boxes (finite elements), each subjected to a constant force. We look at how the force acts on these tiny boxes and add up all the contributions to get an idea of what happens to a large object. Take limits, neglect higher-order terms, and you've got yourself a nice set of equations. In this case, a good story can be more useful than the truth.

The hard part is coming up with a good narrative framework. Working out the details is a lot easier once you have a mental picture of where you are and where you're going. It's easy to come up with a story that doesn't add anything--some ad-hoc tale to satisfy your desire for an explanation and let you go on doing what you were doing with your black-box model.

Sorry if this is a bit disjointed. I'm still trying to straighten it out in my own mind.

Comment author: Lumifer 11 May 2015 07:01:38PM 6 points [-]

I think what would be useful is to distinguish a story (a typically linear narrative) and a model (a known-to-be-simplified map of some piece of reality). They are sufficiently different and often serve different goals. In particular, stories are rarely quantitative and models usually are.

Comment author: Epictetus 11 May 2015 07:27:24PM 0 points [-]

I like to think about how the two complement each other. You can build a model out of a mass of data, but extrapolation outside the data is tricky business. You can also start with a qualitative description of the phenomena involved and work out the details. A lot of models start off by making some assumptions and figuring out the consequences.

Example: you can figure out gas laws by taking lots of measurements, or you can start with the assumption that gases are made of molecules that bounce around and go from there.

Comment author: Lumifer 11 May 2015 08:33:12PM 4 points [-]

We might be understanding the word "story" differently.

To me a "story" is a narrative (a linear sequence of words/sentences/paragraphs/etc.) with the general aim of convincing your System 1. It must be simple enough for the System 1 and must be able to be internalized to become effective. There are no calculations in stories and they generally latch onto some basic hardwired human instincts.

For example, a simple and successful story is "There are tiny organisms called germs which cause disease. Wash your hands and generally keep clean to avoid disease". No numbers, plugs into the purity/disgust template, mostly works.

The three laws of Newton are not a story to me, to pick a counter-example. Nor is the premise that gas consists of identical independent molecules in chaotic motion -- that's an assumption which underlies a particular class of models.

Models, as opposed to stories, are usually "boxes" in the sense that you can throw some inputs into the hopper, turn the crank, and get some outputs from the chute. They don't have to be intuitive or even understandable (in which case the box is black), they just have to output correct predictions. The Newton's laws, for example, make correct predictions (within their sphere of applicability and to a limited degree of precision), but we still have no idea how gravity really works.

Comment author: Epictetus 12 May 2015 01:16:25AM 1 point [-]

I was using "story" in a much more general sense. Perhaps I should have chosen a different word. I saw a story as some bit of exposition devised to explain a process. In that sense, I would view the kinetic theory of gases as a story. A gas has pressure because all these tiny particles are bumping into the walls of its container. Temperature is related to the average kinetic energy of the particles. The point here is that we can't see these particles, nor can we directly measure their state.

Consider, in contrast, the presentation in Fermi's introductory Thermodynamics book. He eschewed an explanation of what exactly was happening internally and derived his main results from macroscopic behavior. Temperature was defined initially as that which a gas thermometer measures, and later on he developed a thermodynamic definition based on the behavior of reversible heat engines. This sort of approach treats the inner workings of a gas as unknown and only uses that which we can directly observe through instrumental readings.

I guess what I really want to distinguish are black boxes from our attempts to guess what's in the box. The latter is what I tried to encapsulate by "story".

Comment author: Lumifer 12 May 2015 02:49:16PM 1 point [-]

Isn't that what science usually calls a "theory"?

Comment author: IlyaShpitser 11 May 2015 09:34:08PM 1 point [-]

You are talking about prediction vs causality. I agree, we understand via causality, and causality lets us take data beyond what is actually observed into the realm of the hypothetical. Good post.

Comment author: chaosmage 11 May 2015 10:28:03AM 2 points [-]

I have nearly finished the second of the Seven Secular Sermons, which is going to premiere at the European Less Wrong Community Weekend in Berlin in June. For final polishing, I'm looking for constuctive feedback especially from native speakers of English. If you'd like to help out, PM me for the current draft.

Comment author: [deleted] 11 May 2015 12:29:01PM 1 point [-]

Offtopic, but I like your theory of depression have you ever written about it elsewhere longer? Or any recommended online readings?

Comment author: chaosmage 11 May 2015 02:01:52PM *  1 point [-]

Glad you like it. No I haven't written about it at more length than in that post, and it is entirely my own speculation, based only on the phenomenology of clinical depression and the rank theory I referenced.

I don't have any reading on depression to recommend that is anywhere near as good as SSC. And that's despite my working as a research associate at a depression-focused nonprofit.

Comment author: NancyLebovitz 11 May 2015 02:45:35PM 2 points [-]

The status possibility doesn't explain post-partum depression.

Comment author: chaosmage 11 May 2015 03:27:32PM 0 points [-]

Why not?

Comment author: NancyLebovitz 11 May 2015 03:32:49PM 0 points [-]

The baby is very vulnerable to the mother.

Comment author: ChristianKl 11 May 2015 08:42:58PM 1 point [-]

But the baby is also able to often dictate when the mother sleeps and has power over the mother. At least if the mother lets it.

Comment author: NancyLebovitz 11 May 2015 09:14:11PM 1 point [-]

Here's a quote from the link above: "What triggers the depression response is a lack of obviously relatively weaker (dependent or safe to bully) group members".

Comment author: ChristianKl 11 May 2015 10:25:36PM 1 point [-]

"Weakness" isn't a straightforward word. It's not a precise word. In general this theory hasn't had the amount of work needed to be precise.

It can be that the thing that matters for weakness is having power over other people. The power relationship between a mother and her child is complex.

Comment author: chaosmage 12 May 2015 07:54:56AM 0 points [-]

Good point. While this theory does predict that nobody who has recently won a physical fight or successfully bullied someone (in a non-virtual setting) should have acute depression symptoms, I'd rather be cautious about less obviously one-sided imbalances. After all, kids are quite dependent for several more years after postpartum subsides, and they evidently don't confer depression symptoms immunity for that entire period.

Comment author: NancyLebovitz 12 May 2015 01:58:40PM 0 points [-]

It wouldn't surprise me if some bosses are seriously depressed even though they have a complex relationship with employees.

Comment author: Elo 13 May 2015 12:57:28AM 0 points [-]

I would also like to remain updated about this theory and subsequent writings. (repeat to ping you too)

Comment author: Elo 13 May 2015 12:57:13AM 0 points [-]

I would also like to remain updated about this theory and subsequent writings.

Comment author: John_Maxwell_IV 11 May 2015 07:42:42AM *  3 points [-]

I recently found this blog post by Ben Kuhn where he briefly summarizes ~5 classic LW posts in the space of one blog post. A couple points:

  • I don't think that much of the content of the original posts is lost in Ben's summary, and it's a lot faster to read. Do others agree? Do we think producing a condensed summary of the LW archives at some point might be valuable? (It's possible that, for instance, the longer treatment of these concepts in the original posts pushes them deeper in to your brain, or that since people are so used to skimming, conceptually dense content can actually be bad.)

  • Might it be worthwhile to link to this from the LW about page or FAQ for newcomers?

Submitting...

Comment author: Kindly 11 May 2015 11:37:06PM 5 points [-]

This ought to be verified by someone to whom the ideas are genuinely unfamiliar.

Comment author: [deleted] 11 May 2015 08:02:51AM *  2 points [-]

In what conceiveable (which does not imply logicality) universes would Rationalism not work in the sense of unearthing only some truths, not all truths? Some realms of truth would be hidden to Rationalists? To simplify it, I mean largely the aspect that of empiricism, of tying ideas to observations via prediction. What conceivable universes have non-observational truths, for example, Platonic/Kantian "pure apriori deduction" type of mental-only truths? Imagine for convenience's sake a Matrix type simulated universe, not necessarily a natural one, so it does not really need to be lawful nor unfold from basic laws.

Reason for asking: if you head over to a site like The Orthosphere, they will tell you Rationalism can only find some but not all truths. And one good answer would be: "This could happen in universes of the type X, Y, Z. What are your reasons for thinking ours could be one of them?"

Comment author: OrphanWilde 11 May 2015 07:26:15PM 4 points [-]

Depends on how you decide what truth is, and what qualifies it to be "unearthed."

But for one universe in which some truth, for some value of truth, can be unearthed, for some value of unearthed, while other truth can't be:

Imagine a universe in which 12.879% (exactly) of all matter is a unique kind of matter that shares no qualities in common with any other matter, and is almost entirely nonreactive with all other kinds of matter, and was created by a process not shared in common with any other matter, which had no effect whatsoever on any other matter. Any truths about this matter, including its existence and the percentage of the universe composed of it, would be completely non-observational. The only reaction this matter has with any other matter is when it is in a specific configuration which requires extremely high levels of the local equivalent of negative entropy, at which point it emits a single electromagnetic pulse. This was used once by an intelligence species composed of this unique matter who then went on to die in massive wars, to encode in a series of flashes of light every detail they knew about physics, and was observed by one human-equivalent monk ascetic, who used a language similar to morse code to write down the sequence of pulses, which he described as a holy vision. Centuries later, these pulses were translated into mathematical equations which described the unique physics of this concurrent universe of exotic matter, but no mechanism of proving the existence or nonexistence of this exotic matter, save that the equations are far beyond the mathematics of anyone alive at the time the signal was encoded, and it has become a controversial matter whether or not it was an elaborate hoax by a genius.

Comment author: ChristianKl 11 May 2015 12:33:49PM 4 points [-]

What do you mean with "Rationalism"?

The LW standard definition is that it's about systematized winning. If the Matrix overlords punish everybody who tries to do systematized winning than it's bad to engage in it. Especially when the Matrix overlords do it via mind reading. The Christian God might see it as a sin.

If you don't use the LW definition of rationalism, then rationalism and empiricism are not the same thing. Rationalism generally refers to gathering knowledge by reasoning as opposed to gathering it by other ways such as experiments or divine revelation.

they will tell you Rationalism can only find some but not all truths

Gödel did prove that it's impossible to find all truths. This website is called Lesswrong because it's not about learning all truths but just about becoming less wrong.

Comment author: ike 11 May 2015 02:54:31PM 1 point [-]

Gödel did prove that it's impossible to find all truths.

That's misleading. With a finite amount of processing power/storage/etc, you can't find all proofs in any infinite system. We need to show that short truths can't be found, which is a bit harder.

Comment author: Houshalter 14 May 2015 06:15:41AM 0 points [-]

I don't think that's correct. My best understanding of Godel's theorem is that if your system of logic is powerful enough to express itself, then you can create a statement like "this sentence is unprovable". That's pretty short and doesn't rely on infiniteness.

Comment author: ike 14 May 2015 01:32:23PM 0 points [-]

The statement "this sentence is unprovable" necessarily includes all information on how to prove things, so it's always larger than your logical system. It's usually much larger, because "this sentence" requires some tricks to encode.

To see this another way, the halting problem can be seen as equivalent to Godel's theorem. But it's trivially possible to have a program of length X+C that solves the halting problem for all programs of length X, where C is a rather low constant; see https://en.wikipedia.org/wiki/Chaitin'sconstant#Relationshiptothehalting_problem for how.

Comment author: Houshalter 15 May 2015 09:25:38AM 0 points [-]

I'm not sure how much space it would take to write down formally, and I'm not sure it matters. At worst it's a few pages, but not entire books, let alone some exponentially huge thing you'd never encounter in reality.

It's also not totally arbitrary axioms that would never be encountered in reality. There are reasons why someone might want to define the rules of logic within logic, and then 99% of the hard work is done.

But regardless, the interesting thing is that such an unprovable sentence exists at all. That its not possible to prove all true statements with any system of logic. It's possible that the problem is limited to this single edge case, but for all I know these unprovable sentences could be everywhere. Or worse, that it is possible to prove them, and therefore possible to prove false statements.

I think the halting problem is related, but I don't see how it's exactly equivalent. In any case the halting problem work around is totally impractical, since it would take multiple ages of the universe to prove the haltingness of a simple loop. If you are referring to the limited memory version, otherwise I'm extremely skeptical.

Comment author: ike 15 May 2015 04:53:50PM 0 points [-]

At worst it's a few pages, but not entire books, let alone some exponentially huge thing you'd never encounter in reality.

That's only if your logical system is simple. If you're a human, then the system you're using is probably not a real logical system, and is anyway going to be rather large.

I think the halting problem is related, but I don't see how it's exactly equivalent.

See http://www.solipsistslog.com/halting-consequences-godel/

Comment author: ChristianKl 11 May 2015 04:11:36PM 0 points [-]

We need to show that short truths can't be found, which is a bit harder.

DeVliegendeHollander post didn't speak about short truths but about all truths.

Comment author: ike 11 May 2015 06:13:24PM *  1 point [-]

If we're talking about all truths, then a finiteness argument shows we can never get all truths, no need for Godel. Godel shows that given infinite computing power, we still can't generate all truths, which seems irrelevant to the question.

If we can prove all truths smaller than the size of the universe, that would be pretty good, and it isn't ruled out by Godel.

Comment author: Douglas_Knight 11 May 2015 05:32:54PM 0 points [-]

Gödel did prove that it's impossible to find all truths.

While Gödel killed HIlbert's program as a matter of historical fact, it was later Tarski who proved the theorem that truth is undecidable.

Comment author: SolveIt 11 May 2015 08:50:25AM 3 points [-]

There's no guarantee we should be able to find any truths using any method. It's a miracle that the universe is at all comprehensible. The question isn't "when can't we learn everything?", it's "why can we learn anything at all?".

Comment author: Lumifer 11 May 2015 04:08:21PM -2 points [-]

"why can we learn anything at all?"

Because entities which can't do not survive.

Comment author: CronoDAS 16 May 2015 01:16:05AM 2 points [-]

Counterexample: Plants. Do they learn?

Comment author: Lumifer 16 May 2015 01:30:41AM 1 point [-]

Of course. Leaves turn to follow the sun, roots grow in the direction of more moist soil...

Comment author: CronoDAS 16 May 2015 08:17:07PM 2 points [-]

Is that really learning, or just reacting to stimuli in a fixed, predetermined pattern?

Comment author: Romashka 17 May 2015 09:53:17AM 1 point [-]

Does vaccination imply memory?.. Does being warned by another's volatile metabolites that a herbivore is attacking the population?

(Higher) plants are organized by very different principles than animals; it is a never-ending debate on what constitutes 'identity' in them. Without first deciding upon that, can one speak about learning? I don't think they have it, but their patterns of predetermined answers can be very specific.

Comment author: Romashka 16 May 2015 07:40:55AM 0 points [-]

Also, there is an interesting study, 'Kin recognition, not competitive interactions, predicts root allocation in young Cakile edentula seedling pairs'. This seems to be more difficult to do than following the sun!

Comment author: SolveIt 11 May 2015 09:27:34PM 1 point [-]

That just pushes the question back a step. Why can any entity learn?

Comment author: tim 12 May 2015 03:32:59AM 1 point [-]

In the spirit of Lumifer's comment, anything we would consider an entity would have to be able to learn or we wouldn't be considering it at all.

Comment author: RichardKennaway 16 May 2015 06:02:43AM 0 points [-]

Why can any entity learn?

A more useful question to ask would be "how do entities, in fact, learn?" This avoids the trite answer, "because if they didn't, we wouldn't be asking the question".

Comment author: Lumifer 12 May 2015 02:32:01PM 0 points [-]

I think if we follows this chain of questions, what we'll find at the end (except for turtles, of course) is the question "Why is the universe stable/regular instead of utterly chaotic?" A similar question is "Why does the universe even have negentropy?"

I don't know any answer to these questions except for "That's what our universe is".

Comment author: SolveIt 12 May 2015 02:38:40PM 1 point [-]

I suppose what I want to know is the answer to "What features of our universe make it possible for entities to learn?".

Which sounds remarkably similar to DeVliegendeHollander's question, perhaps with an implicit assumption that learning won't be present in many (most?) universes.

Comment author: Lumifer 12 May 2015 04:24:14PM 1 point [-]

The fact that the universe is stable/regular enough to be predictable. Subject predictability is a necessary requirement for learning.

Comment author: fubarobfusco 11 May 2015 08:55:00PM 0 points [-]

For that matter, a world in which it is impossible for an organism to become better at surviving by modeling its environment (i.e. learning) is one in which intelligence can't evolve.

(And a world in which it is impossible for one organism to be better at surviving than another organism, is one in which evolution doesn't happen at all; indeed, life wouldn't happen.)

Comment author: IlyaShpitser 11 May 2015 08:18:59AM *  3 points [-]

Don't need to posit crazy things, just think about selection bias -- are the sorts of people that tend to become rationalist randomly sampled from the population? If not, why wouldn't there be blind spots in such people just based on that?

Comment author: [deleted] 11 May 2015 08:40:52AM 1 point [-]

Yes, but if I get the idea right, it is to learn to think in a self-correcting, self-improving way. For example, maybe Kanazawa is right in intelligence suppressing instincts / common sense, but a consistent application of rationality sooner or later would lead to discovering it and forming strategies to correct it.

For this reason, it is more of the rules (of self-correction, self-improvement, self-updating sets of beliefs) than the people. What kinds of truths would be potentially invisible to a self-correcting observationalist ruleset even if this was practiced by all kinds of people?

Comment author: IlyaShpitser 11 May 2015 08:49:46AM *  3 points [-]

Just pick any of a large set of things the LW-sphere gets consistently wrong. You can't separate the "ism" from the people (the "ists"), in my opinion. The proof of the effectiveness of the "ism" lies in the "ists".

Comment author: NancyLebovitz 11 May 2015 11:58:23AM 2 points [-]

Which things are you thinking of?

Comment author: IlyaShpitser 11 May 2015 02:38:57PM *  2 points [-]

A lot of opinions much of LW inherited uncritically from EY, for example. That isn't to say that EY doesn't have many correct opinions, he certainly does, but a lot of his opinions are also idiosyncratic, weird, and technically incorrect.

As is true for most of us. The recipe here is to be widely read (LW has a poor scholarship problem too). Not moving away from EY's more idiosynchratic opinions is sort of a bad sign for the "ism."

Comment author: NancyLebovitz 11 May 2015 02:44:49PM 1 point [-]

Could you mention some of the specific beliefs you think are wrong?

Comment author: IlyaShpitser 11 May 2015 02:58:42PM *  8 points [-]

Having strong opinions on QM interpretations is "not even wrong."

LW's attitude on B is, at best, "arguable."

Donating to MIRI as an effective use of money is, at best, "arguable."

LW consequentialism is, at best, "arguable."

Shitting on philosophy.

Ratonalism as part of identity (aspiring rationalist) is kind of dangerous.

etc.


What I personally find valuable is "adapting the rationalist kung fu stance" for certain purposes.

Comment author: NancyLebovitz 11 May 2015 03:30:51PM 2 points [-]

Thank you.

LW's attitude on B is, at best, "arguable."

B?

Comment author: IlyaShpitser 11 May 2015 03:32:27PM 0 points [-]

Bayesian.

Comment author: OrphanWilde 11 May 2015 03:17:40PM *  2 points [-]

Ratonalism as part of identity (aspiring rationalist) is kind of dangerous.

[Edited formatting] Strongly agree. http://lesswrong.com/lw/huk/emotional_basilisks/ is an experiment I ran which demonstrates the issue. Eliezer was unable to -consider- the hypothetical; it "had" to be fought.

The reason being, the hypothetical implies a contradiction in rationality as Eliezer defines it; if rationalism requires atheism, and atheism doesn't "win" as well as religion, then the "rationality is winning" definition Eliezer uses breaks; suddenly rationality, via winning, can require irrational behavior. Less Wrong has a -massive- blind spot where rationality is concerned; for a web site which spends a significant amount of time discussing how to update "correctness" algorithms, actually posing challenges to "correctness" algorithms is one of the quickest ways to shut somebody's brain down and put them in a reactionary mode.

Comment author: ChristianKl 11 May 2015 04:25:40PM 0 points [-]

if rationalism requires atheism

I don't think that's argued. It's also worth noting that the majority of MIRI's funding over it's history comes from a theist.

Comment author: RichardKennaway 12 May 2015 12:05:43PM 1 point [-]

Eliezer was unable to -consider- the hypothetical; it "had" to be fought.

It seems to me that he did consider your hypothetical, and argued that it should be fought. I agree: your hypothetical is just another in the tedious series of hypotheticals on LessWrong of the form, "Suppose P were true? Then P would be true!"

BTW, you never answered his answer. Should I conclude that you are unable to consider his answer?

Eliezer also has Harry Potter in MoR withholding knowledge of the True Patronus from Dumbledore, because he realises that Dumbledore would not be able to cast it, and would no longer be able to cast the ordinary Patronus.

Now, he has a war against the Dark Lord to fight, and cannot take the time and risk of trying to persuade Dumbledore to an inner conviction that death is a great evil in order to enable him to cast the True Patronus. It might be worth pursuing after winning that war, if they both survive.

All this has a parallel with your hypothetical.

Comment author: TheAncientGeek 12 May 2015 05:00:22PM *  0 points [-]

I 've notice that problem, but I think it is a bit dramatic to call it rationality breaking. I think it's more of a problem of calling two things, the winning thing amd the truth seeking thing, by one name.

Comment author: Luke_A_Somers 12 May 2015 12:19:38AM 0 points [-]

Well...

QM: Having strong positive beliefs on the subject would be not-even-wrong. Ruling out some is much less so. And that's what he did. Note, I came to the same conclusion long before.

MIRI: It's not uncritically accepted on LW more than you'd expect given who runs the joint.

Identity: If you're not letting it trap you by thinking it makes you right, if you're not letting it trap you by thinking it makes others wrong, then what dangers are you thinking of? People will get identities. This particular one seems well-suited to mitigating the dangers of identities.

Others: more clarification required

Comment author: ChristianKl 11 May 2015 04:21:16PM 0 points [-]

Ratonalism as part of identity (aspiring rationalist) is kind of dangerous.

I think there's plenty of criticism voiced about that concept on LW and there are articles advocating to keep one's identity small.

Comment author: IlyaShpitser 11 May 2015 04:37:07PM 2 points [-]

And yet...

Comment author: drethelin 11 May 2015 07:00:17PM 2 points [-]

A universe where humans are running on brains with certain glitches that prevent them from coming to correct conclusions through reasoning about specific topics.

Comment author: CronoDAS 16 May 2015 01:05:48AM 1 point [-]

I can't understand what my girlfriend is saying when she uses her cell phone to call my cell phone. Often entire words are simply dropped from the audio stream. Is there anything I can do to improve the voice quality?

Comment author: moreati 16 May 2015 05:55:37PM 0 points [-]

A few thoughts based on eliminating/ruling out possible causes:

  • Can you avoid making cell -> cell calls? If you're both on a smartphone with wifi could you use e.g. Skype or Messenger?
  • Can you both use a hands free kit? This should eliminate poor positioning of the microphone/earpiece.
  • Are you or your girlfriend in a poor signal area? Does going outside to make the call reduce the problem? There are options such as GSM signal boosters and femtocells that might be worth exploring.
  • Some carriers have deployed HD Voice service. See if your phone and carrier(s) support this.
Comment author: CronoDAS 16 May 2015 08:28:35PM *  0 points [-]

I have a Nokia Lumia 822 (a Windows Phone). My girlfriend has an iPhone 5s. Our carrier is Verizon.

Comment author: moreati 17 May 2015 10:10:20AM 0 points [-]

Ah, sorry. I wasn't seeking answers to those questions. They were meant to suggest lines of enquiry you should follow, and workarounds you could try.

Comment author: taygetea 15 May 2015 10:40:41PM 1 point [-]

I've begun to notice discussion of AI risk in more and more places in the last year. Many of them reference Superintelligence. It doesn't seem like a confirmation bias/Baader-Meinhoff effect, not really. It's quite an unexpected change. Have others encountered a similar broadening in the sorts of people you encounter talking about this?

Comment author: Manfred 15 May 2015 11:41:29PM *  3 points [-]

Yup. Nick Bostrom is basically the man. Above and beyond being the man, he's a respectable focal point for a sea change that has been happening for broader reasons.

Comment author: TsviBT 12 May 2015 02:03:29AM 1 point [-]

PSA: If you wear glasses, you might want to take a look behind the little nosepads. Some... stuff... can build up there. According to this unverified source it is oxidized copper from glasses frame + your sweat, and can be cleaned with an old toothbrush + toothpaste.

Comment author: Dorikka 12 May 2015 02:17:57AM 6 points [-]

Sounds like the only disutility of the stuff is that it annoys some people, but it can't annoy you if you dont notice it...so why bring it up?

Comment author: philh 12 May 2015 08:45:52AM 6 points [-]

It mildly annoys me, but I hadn't thought to use toothpaste on it, so there's that.

Comment author: advancedatheist 11 May 2015 12:48:52AM *  -1 points [-]

Transhumanism-related blog posts:

In Praise of Life (Let’s Ditch the Cult of Longevity)

http://www.patheos.com/blogs/friendlyatheist/2015/05/08/in-praise-of-life-lets-ditch-the-cult-of-longevity/

Overcoming Bias: Why Not?

http://futurisms.thenewatlantis.com/2015/05/overcoming-bias-why-not.html

Also noteworthy:

Prepping for cataclysms, neglecting ordinary emergencies

http://akinokure.blogspot.com/2015/05/prepping-for-cataclysms-neglecting.html

Interesting books:

A cryonics novel:

The New World: A Novel Hardcover – May 5, 2015 by Chris Adrian (Author), Eli Horowitz (Author)

http://www.amazon.com/New-World-Novel-Chris-Adrian/dp/0374221812

Futurology, from the looks of it:

Tomorrowland: Our Journey from Science Fiction to Science Fact Paperback – May 12, 2015 by Steven Kotler (Author)

http://www.amazon.com/Tomorrowland-Journey-Science-Fiction-Fact/dp/0544456211/

Cryonics news:

Another of cryonics' founding generation goes into cryo, though under really bad circumstances.

Dr. Laurence Pilgeram becomes Alcor’s 135th patient on April 15, 2015

http://www.amazon.com/Tomorrowland-Journey-Science-Fiction-Fact/dp/0544456211/

Comment author: ZankerH 11 May 2015 06:31:50AM 14 points [-]

Despite medical and police personnel aware of his Alcor bracelet, he was taken to the medical examiner’s office in Santa Barbara, as they did not understand Alcor’s process and assumed that the circumstances surrounding his death would pre-empt any possible donation directives. Since this all transpired late on a Friday evening, Alcor was not notified of the incident until the following Monday morning.

How the hell are they treating this as a successful preservation? The body spent two days "warm and dead".

Looking at their past case reports, this seems to be fairly normal. Unless you're dying of a known terminal condition and go die in their hospice in Arizona, odds are the only thing getting froze is a mindless, decaying corpse.

Comment author: advancedatheist 11 May 2015 02:32:07PM 10 points [-]

Cryonicist Ben Best has put a lot of effort into studying and testing personal alarm gadgets you can wear which signal cardiac arrest to try to reduce the incidence of these unattended deanimations and long delays before cryopreservation. I plan to look into those myself.

Ironically, I've noticed that cryonicists talk a lot about how much they believe in scientific, medical and technological progress, but then they don't seem to want to act on it when you present them with evidence of the correctable deficiencies of real, existing cryonics.

Reference:

Personal Alarm Systems for Cryonicists

http://www.benbest.com/cryonics/alarms.html

Comment author: RichardKennaway 11 May 2015 07:43:13AM *  9 points [-]

In Praise of Life (Let’s Ditch the Cult of Longevity)

That article would be better titled "In Praise of Death", and is a string of the usual platitudes and circularities.

Overcoming Bias: Why Not?

Why not? Because (the article says) rationalists are cold, emotionless Vulcans, and valuing reason is a mere prejudice.

Prepping for cataclysms, neglecting ordinary emergencies

Maybe there are people who do that, but the article is pure story-telling, without a single claim of fact. File this one under "fiction".

A cryonics novel:

The New World: A Novel Hardcover – May 5, 2015 by Chris Adrian (Author), Eli Horowitz (Author)

The previous links scored 0 out of 3 for rational content, so coming to this one, I thought, what am I likely to find? Clearly, the way to bet is that it's against cryonics. There's only about a blogpost's worth of story in the idea of corpsicles just being unrevivable, so the novel will have to have revival working, but either it works horribly badly, or the revived people find themselves in a bad situation.

Click through...and I am, I think, pleasantly surprised to find that it might, in the end, be favourable to the idea. Or maybe not, there are no reviews and it's difficult to tell from the blurb:

Furious and grieving, Jane fights to reclaim Jim from Polaris [the "shadowy" cryonics company]. Revived in the future, Jim learns that he must sacrifice every memory of Jane if he wants to stay alive in the new world.

Spoiler request! How does it play out in the end?

Tomorrowland: Our Journey from Science Fiction to Science Fact Paperback – May 12, 2015 by Steven Kotler (Author)

Yep, futurological journalism. Pass.

Another of cryonics' founding generation goes into cryo, though under really bad circumstances.

Dr. Laurence Pilgeram becomes Alcor’s 135th patient on April 15, 2015

Shit happens.

Comment author: CAE_Jones 11 May 2015 10:29:11AM 2 points [-]

In Praise of Life (Let’s Ditch the Cult of Longevity)

That article would be better titled "In Praise of Death", and is a string of the usual platitudes and circularities.

I'm now curious: where are the essays that make actual arguments in favor of death? The linked article doesn't make any; it just asserts that death is OK and we're being silly for fighting it, without actually providing a reason (they cite Borges's distopias at the end, but this paragraph has practically nothing in common with the rest of the article, which seems to assume immortality is impossible anyway).

Preference goes to arguments against Elven-style immortality (resistant but not completely immune to murder or disaster, suicide is an option, age-related disabilities are not a thing).

Comment author: jkaufman 11 May 2015 03:42:52PM 1 point [-]

Here's my argument for why death isn't the supreme enemy: http://www.jefftk.com/p/not-very-anti-death

Comment author: Lumifer 11 May 2015 04:23:52PM *  5 points [-]

I have a feeling a lot of discussions of life extension suffer from being conditioned on the implicit set point of what's normal now.

Let's imagine that humans are actually replicants and their lifespan runs out in their 40s. That lifespan has a "control dial" and you can turn it to extend the human average life expectancy into the 80s. Would all your arguments apply and construct a case against meddling with that control dial?

Comment author: Kawoomba 11 May 2015 04:39:15PM *  3 points [-]

That's a good argument if you were to construct the world from first principles. You wouldn't get the current world order, certainly. But just as arguments against, say, nation-states, or multi-national corporations, or what have you, do little do dissuade believers, the same applies to let-the-natural-order-of-things-proceed advocates. Inertia is what it's all about. The normative power of the present state, if you will. Never mind that "natural" includes antibiotics, but not gene modification.

This may seem self-evident, but what I'm pointing out is that by saying "consider this world: would you still think the same way in that world?" you'd be skipping the actual step of difficulty: overcoming said inertia, leaving the cozy home of our local minimum.

Comment author: Lumifer 11 May 2015 04:56:00PM *  5 points [-]

Inertia is what it's all about. The normative power of the present state, if you will.

That's fine as long as you understand it and are not deluding yourself with a collection of reasons why this cozy local minimum is actually the best ever.

The considerable power wielded by inertia should be explicit.

Comment author: jkaufman 16 May 2015 02:23:27AM 0 points [-]

Huh? It feels like you're responding to a common thing people say, but not to anything I've said (or believe).

Comment author: Lumifer 16 May 2015 02:40:59AM 0 points [-]

I meant this as a response specifically to

But dramatically fewer children? Much less of the total human experience spent in early learning stages? Would we become less able to make progress in the world because people have trouble moving on from what they first learned?

Comment author: jkaufman 18 May 2015 12:00:59PM *  1 point [-]

More context:

A world in which we have ended death ... may be better than the world now, but I could also see it being worse. On one hand, not having to see your friends and family die, increased institutional memory, more time to get deeply into subjects and achieve mastery, and time to really build up old strong friendships sound good. But dramatically fewer children? Much less of the total human experience spent in early learning stages? Would we become less able to make progress in the world because people have trouble moving on from what they first learned?

I don't think our current lifespan is the perfect length, but there's a lot of room between "longer is probably better" and "effectively unlimited is ideal".

Comment author: Lumifer 18 May 2015 03:46:27PM 0 points [-]

there's a lot of room between "longer is probably better" and "effectively unlimited is ideal".

Yes, but are you saying there's going to a maximum somewhere in that space -- some metric will flip over and start going down? What might that metric be?

Comment author: jkaufman 20 May 2015 02:40:18PM 0 points [-]

As I wrote in that post, there are some factors that lead to us thinking longer lives would be better, and others that shorter would be better.

Maybe this is easier to think about with a related question: what is the ideal length of tenure at a company? Do companies do best when they have entirely employees-for-life, or is it helpful to have some churn? (Ignoring that people can come in with useful relevant knowledge they got working elsewhere.) Clearly too much churn is very bad for the company, but introducing new people to your practices and teaching them help you adapt and modernize, while if everyone has been there forever it can be hard to make adjustments to changing situations.

The main issue is that people tend to fixate some on what they learn when they're younger, so if people get much older on average then it would be harder to make progress.

Comment author: CellBioGuy 12 May 2015 03:15:58AM 0 points [-]

My take: there's a big difference between calling something good and dealing with a fact.

Comment author: advancedatheist 11 May 2015 02:42:14PM *  2 points [-]

I know "preppers" in Arizona who don't have any savings because they have spent all their money on this survivalist nonsense. They would do better to have put that money in the bank and applied for subsidized health insurance.

The blogger agnostic does have a point about how the prepper mentality shows an abandonment of wanting to produce for and sustain the existing society, so that instead you can position yourself to become a scavenger and a parasite on the wealth produced by others if some apocalyptic collapse happens. That ridiculous Walking Dead series, which amounts to nonstop prepper porn, feeds some very damaging fantasies that I don't think we should encourage.

Comment author: passive_fist 11 May 2015 10:20:54PM *  -1 points [-]

Just a PSA: advancedatheist has a fixation on dehumanizing rationalists with an especial focus on rationalists 'not being able to get laid'. Here's some of his posts on this matter:

http://lesswrong.com/lw/lzb/open_thread_apr_01_apr_05_2015/c7gr

http://lesswrong.com/lw/m4h/when_does_technological_enhancement_feel_natural/cc09

http://lesswrong.com/lw/m1p/open_thread_apr_13_apr_19_2015/cams

http://lesswrong.com/lw/dqz/a_marriage_ceremony_for_aspiring_rationalists/72wr

It's best not to 'feed the trolls', so to speak.

Comment author: knb 12 May 2015 03:13:30AM 10 points [-]

So why lash out at him for this now when he isn't currently doing that? In any case I don't think he was trolling (deliberately trying to cause anger) so much as he was just morbidly fixated on a topic, and couldn't stop bringing it up,

Comment author: passive_fist 12 May 2015 03:20:39AM 0 points [-]

I'm pointing it out for the benefit of others who may not understand where AA is coming from.

Comment author: NancyLebovitz 12 May 2015 04:14:55AM 9 points [-]

I recommend responding to whatever specific problematic things he might say rather than issuing a general warning.

Comment author: passive_fist 12 May 2015 05:01:23AM -1 points [-]

I am responding to quite specific problematic things he's saying. My comment is in response to AAs and is in reply to a reply to his comment. If I were to directly reply to him saying the same thing, my intentions would probably be misunderstood.

Comment author: philh 12 May 2015 09:03:32AM 2 points [-]

Another thing AA seems to do quite a lot is link to pro-death blog posts and articles that he doesn't endorse. I get the impression that's what he was doing with some of the above links. IIRC he's signed up for cryonics, so it seems unlikely that he's trying to push a pro-death agenda.

Comment author: RichardKennaway 12 May 2015 12:20:12PM 1 point [-]

IIRC he's signed up for cryonics, so it seems unlikely that he's trying to push a pro-death agenda.

So, AA, if you're reading down here, why are you signed up for cryonics while posting pro-death links and complaining at length about never getting laid? Optimism for a hereafter, despair for the present, and bitterness for the past. This is not a good conjunction.

Comment author: OrphanWilde 14 May 2015 02:31:32PM 1 point [-]

Maybe he just sees value on challenging the status quo?

Comment author: philh 14 May 2015 05:33:23PM 1 point [-]

I interpret it more as "look at these awful things people are saying about us".

Comment author: Luke_A_Somers 12 May 2015 12:07:14AM 4 points [-]

That's a weirdly weak collection of posts to complain about. It seems more like AA is noting his OWN lack of ability to get laid and has a degree of curiosity on the subject that would naturally result from such a situation. He also (correctly, I expect) anticipates that a noticeable number of people who are or have been in the same boat as him are on LW.

I have seen some really obnoxious posts by AA, but these don't strike me as great examples. I am not about to go digging for them.

Comment author: RichardKennaway 12 May 2015 12:10:50PM 3 points [-]

Just a PSA: advancedatheist has a fixation on dehumanizing rationalists with an especial focus on rationalists 'not being able to get laid'.

I've noticed. While it certainly informs my attitude to everything he posts, he is mostly still at the level of worth responding to.

Comment author: Error 11 May 2015 01:31:17AM 1 point [-]

Just FYI, it looks like you goofed that last link.

Comment author: [deleted] 11 May 2015 07:23:43AM 3 points [-]
Comment author: Dorikka 11 May 2015 02:32:59AM 0 points [-]

Works now for me

Comment author: Error 11 May 2015 02:36:16AM 0 points [-]

That's odd, because it still doesn't for me. For me the last link is a duplicate of the second-to-last. Um?

Comment author: Dorikka 11 May 2015 04:14:50AM 1 point [-]

Um, sorry. I reflexively checked the last link and it went to a valid page, didn't notice it was the same as the ones above. User error creates the weirdest problems, eh?

Comment author: SolveIt 12 May 2015 09:55:05AM 1 point [-]

The rationalist Tumblr community looks interesting. Any tips on how to start?

Comment author: Gondolinian 12 May 2015 11:45:38PM 5 points [-]

Well, for a start there's the Rationalist Masterlist currently hosted by Yxoque (MathiasZaman here on LW). You could announce your presence there and ask to be added to the list, or just lurk around some of the blogs for a while and send anonymous asks to people to get a feel for the community before you set up an account.

Comment author: SolveIt 13 May 2015 01:49:12AM 1 point [-]

Thanks!

Comment author: Silver_Swift 12 May 2015 08:51:39AM 1 point [-]

Is this the place to ask technical questions about how the site works? If so, then I'm wondering why I can't find any of the rationality quote threads on the main discussion page anymore (I thought we'd just stopped doing those, until I saw it pop up in the side bar just now). If not, then I think I just asked anyway. :P

Comment author: NancyLebovitz 12 May 2015 02:04:33PM 6 points [-]

This is a good place to ask about how the site works.

Comment author: gjm 12 May 2015 09:05:36AM 5 points [-]

Here -- it's in Main rather than Discussion.

Comment author: Silver_Swift 12 May 2015 01:16:23PM 2 points [-]

Thanks!

Comment author: Archelon 17 May 2015 11:50:44PM 0 points [-]

According to this article, a traumatic brain injury turned a furniture salesman into a mathematician. (Not without side effects, but still.)

There is a bit of conventional wisdom in evolutionary biology that drastic improvements in efficacy are not available through trivial modifications (and that nontrivial modifications which are random are not improvements). This is an example of the principle that evolution is supposed to have already 'harvested' any 'low-hanging fruit'. Although I don't think much of this type of website (note the lack of external links), the story seems to be based in reality; it is thus one of the most surprising things I have ever heard. And, oddly, heartening as well---insofar as it suggests both a potential shortening of the timescale for human intelligence augmentation and the possibility that such augmentation may be relatively more accessible (than I previously thought) by comparison to computer-based artificial intelligence developments.

Comment author: the-citizen 17 May 2015 06:52:33AM *  0 points [-]

Suffering and AIs

Disclaimer - Under utilitarianism suffering is an intrinsically bad thing. While I am not a utilitarian, many people are and I will treat it as true for this post because it is the easiest approach for this issue. Also, apologies if others have already discussed this idea, which seems quite possible

One future moral issue is that AIs may be created for the purpose of doing things that are unpleasant for humans to do. Let's say an AI is designed with the ability to have pain, fear, hope and pleasure of some kind. It might be reasonable to expect in such cases the unpleasant tasks might result in some form of suffering. Added to this problem is the fact that a finite lifespan and an approaching termination/shutdown might cause fear, another form of suffering. Taking steps to shut down an AI would also then become morally unacceptable, even though they perform an activity that might be useless or harmful. Because of this, we might face a situation where we cannot shutdown AIs even when there is good reason to.

Basically, if suffering AIs were some day extremely common, we would be introducing a massive amount of suffering into the world, which under utilitarianism is unacceptable. Even assuming some pleasure is created, we might search for ways to create that pleasure without creating the pain.

If so, would it make sense to adopt a principle of AI design that says AIs should be designed so it (1) does not suffer or feel pain (2) should not fear death/shutdown (eg. views own finite life as acceptable). This would minimise suffering (potentially you could also attempt to maximise happiness).

Potential issues with this: (1) Suffering might be in some way relative, so that a neutral lack of pleasure/happiness might become "suffering". (2) Pain/suffering might be useful to create a robot with high utility, and thus some people may reject this principle. (3) I am troubled by this utilitarian approach I have used here as it seems to justify tiliing the universe with machines whose only purpose and activity is to be permanently happy for no reason. (4) Also... killer robots with no pain or fear of death :-P

Comment author: hairyfigment 17 May 2015 03:46:48PM 0 points [-]
Comment author: the-citizen 19 May 2015 07:40:49AM 0 points [-]

That seems like an interesting article, though I think it is focused on the issue of free-will and morality which is not my focus.

Comment author: Fatcat 14 May 2015 08:36:31PM 0 points [-]

Article on transhumanism (intro bit perfunctory) - but has interviews with Anders Sandberg and Steve Fuller on implications of transhumanist thought. Quite interesting in parts - http://www.theworldweekly.com/reader/i/humanity-20/3757

Same journalist did reasonable job of introducing AI dangers last month - http://www.theworldweekly.com/reader/i/irresistible-rise-ai/3379

Comment author: [deleted] 13 May 2015 08:02:59AM 0 points [-]

Asking for article recommendations: difference between intelligence vs. intellectualism, how a superintelligence is not the same as a superintellectual.

Comment author: sixes_and_sevens 11 May 2015 10:16:37AM *  0 points [-]

An idea: auto-generated anki-style flashcards for mathematical notation.

Let's say you struggle reading set builder notation. This system would prompt you with progressively more complicated set builder expressions to parse, keeping track of what you find easy or difficult, and providing tooltips/highlighting for each individual term in the expression. If it were an anki card, the B-side would be how you'd read the expression out in natural language. This wouldn't be a substitute for learning how to use set builder notation, but it would give you a lot of practice in reading it.

There's an easy version of this you could cobble together in an afternoon which has a bunch of randomly-populated templates it renders with MathJax or something. There's a more sophisticated extended project which uses generative grammars, gamified progress visibility and spaced-repetition algorithms.

I've been thinking about putting something like this together, but realistically I don't have the time or the complete skill-set to do it justice, and it would never get finished. Having read this thread about having difficulty in reading mathematical notation, I'm convinced a lot of other people might benefit from it.

ETA: it was probably misguided of me to liken this to Anki decks. I'm not talking about generating a bunch of static flashcards to be used with an existing system like Anki, but something separate that generates dynamic examples of what you're trying to learn, against which you'd record your success at parsing each example in a way similar to Anki. There are, of course, all sorts of problems with memorising specific examples of mathematical notation with an Anki deck, which respondents have prudently picked up on.

Comment author: RichardKennaway 11 May 2015 12:40:59PM 1 point [-]

An idea: auto-generated anki-style flashcards for mathematical notation.

Auto-generated exercises might be better. Compared with e.g. learning a language, there aren't many elementary components to mathematical notation to be memorised.

The exercises might be auto-rated for complexity, and a generalised Anki for this sort of material would generate random examples of various degrees of complexity, and make the distribution of complexity depend in some way on the distribution of your errors with respect to complexity.

Language learning materials might be similarly generalised from the simple vocabulary lists that flashcards are usually used for.

Comment author: sixes_and_sevens 11 May 2015 02:21:06PM 0 points [-]

I agree that auto-generated exercises would be a superior utility, but that seems like a much trickier proposition.

Also, for clarification, this wouldn't be used for memorising notation, but for training fluency in it. My use of Anki as a comparison might have been misguided.

Comment author: Strangeattractor 11 May 2015 09:38:52PM 0 points [-]

I like the idea of making it easier to understand mathematical notation, and get more practice at it. However, using flash cards to implement it could be problematic.

As I learned more and more mathematical notation while studying engineering, it became clear that a lot of the interpretation of the notation depends upon context. For example, if you see vertical lines to either side of an expression, does that mean absolute value or the determinant of a matrix? Is i representing the imaginary number, or current, or the vectors in the same direction as the x-axis? (As an example, electrical engineers use j for the imaginary number, since I represents current.)

For a sufficiently narrow topic, the flashcards might be useful, but it might set up false expectations that the meaning of the symbols will apply outside that narrow topic. There is not a one-to-one correspondence between symbols and meaning.

Comment author: sixes_and_sevens 11 May 2015 11:42:45PM 0 points [-]

I was envisioning some sort of context-system, in part for the reason you describe and in part because people probably have specific learning needs, and at any given time they'd probably be focusing on a specific context.

Also I reiterate what I've said to other commenters: likening it to Anki flashcards was probably misguided on my part. I'm not talking about generating a bunch of static flashcards, but about presenting a user with a dynamically-generated statement for them to parse. The interface would be reminiscent of something like Anki, but it would probably never show you the same statement twice.

Comment author: ChristianKl 11 May 2015 12:21:38PM 0 points [-]

It's important to understand the notation before you put it into Anki. Automatically generated cards with mathematical notation that the person doesn't yet understand is asking for trouble.

Comment author: sixes_and_sevens 11 May 2015 02:13:08PM *  1 point [-]

I may not have presented this well in the original comment. This wouldn't be generating random static cards to put into an Anki deck, but a separate system which dynamically presents expressions made up of known components, and tracks those components instead of specific cards. It seems plausible to restrict these expressions to those composed of notation you've already encountered.

In fact, this could work to its advantage. It also seems plausible to determine which components are bottlenecks, and therefore which concepts are the most effective point of intervention for the person studying. If the user hasn't learned, say, hat-and-tilde notation for estimators, and introducing that notation would result in a greater order of available expressions than the next most bottleneck-y piece of notation, it could prompt the user with "hey, this is hat-and-tilde notation for estimators, and it's stopping you from reading a bunch of stuff". It could then direct them to some appropriate material on the subject.

Comment author: [deleted] 15 May 2015 11:45:02AM *  -1 points [-]

Would this work at least as an early crude hypothesis of how neurotypicals function?

Neurotypicals like social mingling primarily because they play a constant game of social status points, both in the eyes of others (that is real status) and just feeling like getting status (this is more like self-esteem). This should not be understood as a harsh machiavellian cruel game. Usually not. Often it is very warm and friendly. For example, we on the spectrum often finding things like greeting each other superfluous. Needless custom. You notice when people arrive or when not they will talk to you when they want something. But for neurotypicals exchanging mutual warm greetings makes sense, it is a mutual reassurance or reinforcement of each others status. Sometimes people will ignore someone's greeting, even not shake an extended hand, that will be seen as a rude move to reduce the status of the other, as it is embarrassing. This is rudely dominant move they only pull if they are angry at each other. However, it will more often happen that there is some community, group, like a Toastmasters meeting or the local Linux user club or something and someone arrives and hands out a warmish greeting to the whole group but gets only nods in return, or a very short hi, this usually means "you are an outsider, newcomer, not fully accepted in this community yet, so while we don't reduce your status we are cautious about affirming it either, we stay noncommittant until you prove yourself". But then again someone will often be very warm to the newcomer, because acting as a mentor of noobs raises one status inside the group. It a friendly, helpful but clearly dominant move over the newcomer (let me show you the ropes here is the subconscious message), it also reaffirms one as a central member of the group as people who are themselves newbies would not do it, and as new and new people get into the group the people who mentored them cal slowly drift into leadership.

But greetings are just a simple example of the many ways neurotypicals enjoy playing status games. Again it is not a harsh thing, often very warm. But their constant social mingling, constant "purring" with each other is nothing but a set of microtransactions in status. All this small talk thing, plus the body language etc.

To enjoy gaining a score is a general human trait, even we on the spectrum do enjoy getting on the high score table in arcade games or levelling up in an RPG videgame or becoming a thane in Skyrim, we just don't notice how neurotypicals keep doing this all the time. All the social niceties boil down on handing out a micropayment of positive status, any tiny notion of kindness acts as so, the basic small talk of two random people meeting at a garden party will be feeling out each others status as a first step, getting liked by micropaying status ("Wow, that sounds like you have an interesting job!") in exchange for liking or raising your own status by a bit of boasting etc. And that is the nicer part, the uglier part is when people try to reduce each others scores, that is where bad blood comes from.

  1. Am I on a remotely right track here?
  2. Does the statement "nerds / neckbeards often have poor social skills" unpack into "people on the spectrum not even noticing that neurotypicals don't just mindlessly follow social customs, but they are involved in a status micropayment exchange" ?
  3. Any article or book that helps me go on in this line of thought?
Comment author: Zubon 15 May 2015 03:45:45PM *  3 points [-]
  1. Am I on a remotely right track here?

That sounds like a subset of neurotypical behavior. I'm neurotypical and from the very first sentence ("Neurotypicals like social mingling primarily because they play a constant game of social status points, both in the eyes of others (that is real status) and just feeling like getting status (this is more like self-esteem).") I found it contrary to my experience. Which is not to say it is wrong, and it certainly looks like behavior I have seen, but it kind of suggests that there is One Neurotypical Experience as opposed to a spectrum.

That is reading the initial "neurotypicals" as "all/most neurotypicals" as opposed to "some neurotypicals" or "some subset of neurotypicals." I think you are trying to describe typical neurotypical behavior, so I would read that "neurotypicals" as trying to describe how most neurotypicals behave.

But I am not the most central example of a neurotypical, so others may find it a more accurate description of their social experiences. I don't like social mingling, and I avoid most games of social status points. My extroversion score is 7 out of 100, which is likely a factor in not seeing myself in your description of neurotypicals.

2. Does the statement "nerds / neckbeards often have poor social skills" unpack into "people on the spectrum not even noticing that neurotypicals don't just mindlessly follow social customs, but they are involved in a status micropayment exchange" ?

There seem to be several assumptions built into that unpacking. For example, it suggests that all/most nerds are on the spectrum. My characterization of neurotypical socialization would include mindlessly following social customs as well as enjoying the social game. I don't think highly social neurotypicals would describe their behavior as a "status micropayment exchange"; that seems like the wrong metaphor and suggests the dominant model as a fixed-sum status game, whereas many (most?) social interactions have no need for an exchange of status points.

Even when a social status point game is in play, I would expect more interactions to involve the recognition of point totals rather than an exchange. "Mutual reassurance or reinforcement of each others status" seems on point.

If the above is the start of a hypothesis, it seems to me that it links greetings and status point exchange too strongly. Greetings are rarely an occasion to gain or lose points, although they may be occasions to discover the current score.

"Pinging" is a metaphor I have seen used productively in these attempts to explain neurotypical social behavior. The greeting is a ping, a mutual recognition that someone is there and potentially responsive to interaction, potentially also exchanging some basic status information.

Comment author: ChristianKl 16 May 2015 04:01:09PM *  1 point [-]

If you say Bob likes X because of Y, what do you mean with it? Do you mean that if Y wouldn't be there Bob wouldn't like X?

I don't think that there a good reason to believe that if you take status away no neurotically would engage in social mingling or like engaging in it.

Apart from that "status" is a word that's quite abstract. It's much more something "map" than "territory". That produces danger to get into too vague to be wrong territory.

Comment author: [deleted] 18 May 2015 08:29:22AM 1 point [-]

Apart from that "status" is a word that's quite abstract. It's much more something "map" than "territory".

Let's get more meta here. Usually the map-terrain distinction is used to describe how human minds interpret the chunks of reality that are not man-made. When we are talking about something that arises from the behavior of humans, how can we draw that distinction. Plato's classic "What is justice?" is map or terrain? Here the terrain is in human minds too, as justice exists only inside minds and nowhere else, so the distinction seems to be more like is it the grand shared map or a more private map of maps? And the same with status. It does not exist outside the human perception of it. Similar to money, esp. paper/computer number money.

I don't think that there a good reason to believe that if you take status away no neurotically would engage in social mingling or like engaging in it.

I will consider it a typo, assuming you meant neurotypicals like I did i.e. people outside the autism spectrum, or in other words non-geeks. I got the idea from here. If status microtransactions are so important...

(to be continued gotta go now)

Comment author: ChristianKl 18 May 2015 12:46:45PM 0 points [-]

If you look at the link you posted it argues:

Status is a confusing term, unless it’s understood as something one does. You may be low in status, but play high, and vice versa. … We always like it when a tramp is mistaken for the boss, or the boss for a tramp. … I should really talk about dominance and submission, but I’d create a resistance.

That means that dominance and submission map more directly to the territory than status does.

The author doesn't argue that people care about mutually reinforcement of each other status as being high but that people also consciously make moves to submit and place themselves at a low status position.

The text invalidates your idea that people engage primarily in social interaction to maximize the amount of status.

You don't pick that up if you make the error of not treating status as a model but as reality. Reality is complex. Models simplify reality. Sometimes the simplification keeps the essential elements of what you want to describe. Other times it doesn't.

I will consider it a typo, assuming you meant neurotypicals like I did i.e. people outside the autism spectrum, or in other words non-geeks.

Yes, it's a typo likely because my spellchecker didn't know "neurotypicals".

Comment author: DataPacRat 14 May 2015 04:37:57AM *  0 points [-]

Acting on A Gut Feeling

I've been planning an overnight camping trip for sometime this week; but something about the idea is making me feel... disquiet. Uneasy. I can't figure out why; I've got a nice set of equipment, I have people who know where I'm going, and so on. But I can't shake something resembling an "ugh field" that eases when I think of /not/ taking the trip.

And so, I'm concluding that the rational thing to do is to pay attention to my gut, on the chance that one part of my mind is aware of some detail that the rest of my mind hasn't figured out, and postpone my camping trip until I'm feeling more self-assured about the whole thing.

Comment author: DataPacRat 14 May 2015 06:34:11AM 0 points [-]

After some further mental gymnastics, the plan I've come up with which seems to most greatly reduce the disquiet is to buy a backup cellphone, small enough to turn off, stick in a pocket and forget about until I drop my smartphone in a stream. Something along the lines of taking one of the watchphones from http://www.dx.com/s/850%2b1900?category=529&PriceSort=up and snipping off the wristband, or one of the smaller entries in http://www.dx.com/s/850%2b1900?PriceSort=up&category=531 ; along with the $25/year plan from http://www.speakout7eleven.ca/ . Something on the order of $65 to $85 seems a moderate price for peace of mind.

I am, however, going to take at least a day before placing any such order, to find out if such a plan still seems like it /will/ offer increased peace of mind. Not to mention, whether I can come up with (or get suggestions for) any plans which reveal that my actual disquiet arises from some other cause.

Comment author: Miguelatron 15 May 2015 02:16:32PM 1 point [-]

Have you gone camping like this before? If you have, were you by yourself when you did? I'm just trying to eliminate the source of your unease being something simple like stepping out of your comfort zone.