You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Open Thread, May 11 - May 17, 2015

3 Post author: Gondolinian 11 May 2015 12:16AM

If it's worth saying, but not worth its own post (even in Discussion), then it goes here.


Notes for future OT posters:

1. Please add the 'open_thread' tag.

2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)

3. Open Threads should be posted in Discussion, and not Main.

4. Open Threads should start on Monday, and end on Sunday.

Comments (247)

Sort By: Controversial
Comment author: advancedatheist 11 May 2015 12:48:52AM *  -1 points [-]

Transhumanism-related blog posts:

In Praise of Life (Let’s Ditch the Cult of Longevity)

http://www.patheos.com/blogs/friendlyatheist/2015/05/08/in-praise-of-life-lets-ditch-the-cult-of-longevity/

Overcoming Bias: Why Not?

http://futurisms.thenewatlantis.com/2015/05/overcoming-bias-why-not.html

Also noteworthy:

Prepping for cataclysms, neglecting ordinary emergencies

http://akinokure.blogspot.com/2015/05/prepping-for-cataclysms-neglecting.html

Interesting books:

A cryonics novel:

The New World: A Novel Hardcover – May 5, 2015 by Chris Adrian (Author), Eli Horowitz (Author)

http://www.amazon.com/New-World-Novel-Chris-Adrian/dp/0374221812

Futurology, from the looks of it:

Tomorrowland: Our Journey from Science Fiction to Science Fact Paperback – May 12, 2015 by Steven Kotler (Author)

http://www.amazon.com/Tomorrowland-Journey-Science-Fiction-Fact/dp/0544456211/

Cryonics news:

Another of cryonics' founding generation goes into cryo, though under really bad circumstances.

Dr. Laurence Pilgeram becomes Alcor’s 135th patient on April 15, 2015

http://www.amazon.com/Tomorrowland-Journey-Science-Fiction-Fact/dp/0544456211/

Comment author: RichardKennaway 11 May 2015 07:43:13AM *  9 points [-]

In Praise of Life (Let’s Ditch the Cult of Longevity)

That article would be better titled "In Praise of Death", and is a string of the usual platitudes and circularities.

Overcoming Bias: Why Not?

Why not? Because (the article says) rationalists are cold, emotionless Vulcans, and valuing reason is a mere prejudice.

Prepping for cataclysms, neglecting ordinary emergencies

Maybe there are people who do that, but the article is pure story-telling, without a single claim of fact. File this one under "fiction".

A cryonics novel:

The New World: A Novel Hardcover – May 5, 2015 by Chris Adrian (Author), Eli Horowitz (Author)

The previous links scored 0 out of 3 for rational content, so coming to this one, I thought, what am I likely to find? Clearly, the way to bet is that it's against cryonics. There's only about a blogpost's worth of story in the idea of corpsicles just being unrevivable, so the novel will have to have revival working, but either it works horribly badly, or the revived people find themselves in a bad situation.

Click through...and I am, I think, pleasantly surprised to find that it might, in the end, be favourable to the idea. Or maybe not, there are no reviews and it's difficult to tell from the blurb:

Furious and grieving, Jane fights to reclaim Jim from Polaris [the "shadowy" cryonics company]. Revived in the future, Jim learns that he must sacrifice every memory of Jane if he wants to stay alive in the new world.

Spoiler request! How does it play out in the end?

Tomorrowland: Our Journey from Science Fiction to Science Fact Paperback – May 12, 2015 by Steven Kotler (Author)

Yep, futurological journalism. Pass.

Another of cryonics' founding generation goes into cryo, though under really bad circumstances.

Dr. Laurence Pilgeram becomes Alcor’s 135th patient on April 15, 2015

Shit happens.

Comment author: passive_fist 11 May 2015 10:20:54PM *  -1 points [-]

Just a PSA: advancedatheist has a fixation on dehumanizing rationalists with an especial focus on rationalists 'not being able to get laid'. Here's some of his posts on this matter:

http://lesswrong.com/lw/lzb/open_thread_apr_01_apr_05_2015/c7gr

http://lesswrong.com/lw/m4h/when_does_technological_enhancement_feel_natural/cc09

http://lesswrong.com/lw/m1p/open_thread_apr_13_apr_19_2015/cams

http://lesswrong.com/lw/dqz/a_marriage_ceremony_for_aspiring_rationalists/72wr

It's best not to 'feed the trolls', so to speak.

Comment author: RichardKennaway 12 May 2015 12:10:50PM 3 points [-]

Just a PSA: advancedatheist has a fixation on dehumanizing rationalists with an especial focus on rationalists 'not being able to get laid'.

I've noticed. While it certainly informs my attitude to everything he posts, he is mostly still at the level of worth responding to.

Comment author: knb 12 May 2015 03:13:30AM 10 points [-]

So why lash out at him for this now when he isn't currently doing that? In any case I don't think he was trolling (deliberately trying to cause anger) so much as he was just morbidly fixated on a topic, and couldn't stop bringing it up,

Comment author: passive_fist 12 May 2015 03:20:39AM 0 points [-]

I'm pointing it out for the benefit of others who may not understand where AA is coming from.

Comment author: NancyLebovitz 12 May 2015 04:14:55AM 9 points [-]

I recommend responding to whatever specific problematic things he might say rather than issuing a general warning.

Comment author: passive_fist 12 May 2015 05:01:23AM -1 points [-]

I am responding to quite specific problematic things he's saying. My comment is in response to AAs and is in reply to a reply to his comment. If I were to directly reply to him saying the same thing, my intentions would probably be misunderstood.

Comment author: philh 12 May 2015 09:03:32AM 2 points [-]

Another thing AA seems to do quite a lot is link to pro-death blog posts and articles that he doesn't endorse. I get the impression that's what he was doing with some of the above links. IIRC he's signed up for cryonics, so it seems unlikely that he's trying to push a pro-death agenda.

Comment author: RichardKennaway 12 May 2015 12:20:12PM 1 point [-]

IIRC he's signed up for cryonics, so it seems unlikely that he's trying to push a pro-death agenda.

So, AA, if you're reading down here, why are you signed up for cryonics while posting pro-death links and complaining at length about never getting laid? Optimism for a hereafter, despair for the present, and bitterness for the past. This is not a good conjunction.

Comment author: OrphanWilde 14 May 2015 02:31:32PM 1 point [-]

Maybe he just sees value on challenging the status quo?

Comment author: philh 14 May 2015 05:33:23PM 1 point [-]

I interpret it more as "look at these awful things people are saying about us".

Comment author: Luke_A_Somers 12 May 2015 12:07:14AM 4 points [-]

That's a weirdly weak collection of posts to complain about. It seems more like AA is noting his OWN lack of ability to get laid and has a degree of curiosity on the subject that would naturally result from such a situation. He also (correctly, I expect) anticipates that a noticeable number of people who are or have been in the same boat as him are on LW.

I have seen some really obnoxious posts by AA, but these don't strike me as great examples. I am not about to go digging for them.

Comment author: advancedatheist 11 May 2015 02:42:14PM *  2 points [-]

I know "preppers" in Arizona who don't have any savings because they have spent all their money on this survivalist nonsense. They would do better to have put that money in the bank and applied for subsidized health insurance.

The blogger agnostic does have a point about how the prepper mentality shows an abandonment of wanting to produce for and sustain the existing society, so that instead you can position yourself to become a scavenger and a parasite on the wealth produced by others if some apocalyptic collapse happens. That ridiculous Walking Dead series, which amounts to nonstop prepper porn, feeds some very damaging fantasies that I don't think we should encourage.

Comment author: CAE_Jones 11 May 2015 10:29:11AM 2 points [-]

In Praise of Life (Let’s Ditch the Cult of Longevity)

That article would be better titled "In Praise of Death", and is a string of the usual platitudes and circularities.

I'm now curious: where are the essays that make actual arguments in favor of death? The linked article doesn't make any; it just asserts that death is OK and we're being silly for fighting it, without actually providing a reason (they cite Borges's distopias at the end, but this paragraph has practically nothing in common with the rest of the article, which seems to assume immortality is impossible anyway).

Preference goes to arguments against Elven-style immortality (resistant but not completely immune to murder or disaster, suicide is an option, age-related disabilities are not a thing).

Comment author: jkaufman 11 May 2015 03:42:52PM 1 point [-]

Here's my argument for why death isn't the supreme enemy: http://www.jefftk.com/p/not-very-anti-death

Comment author: Lumifer 11 May 2015 04:23:52PM *  5 points [-]

I have a feeling a lot of discussions of life extension suffer from being conditioned on the implicit set point of what's normal now.

Let's imagine that humans are actually replicants and their lifespan runs out in their 40s. That lifespan has a "control dial" and you can turn it to extend the human average life expectancy into the 80s. Would all your arguments apply and construct a case against meddling with that control dial?

Comment author: Kawoomba 11 May 2015 04:39:15PM *  3 points [-]

That's a good argument if you were to construct the world from first principles. You wouldn't get the current world order, certainly. But just as arguments against, say, nation-states, or multi-national corporations, or what have you, do little do dissuade believers, the same applies to let-the-natural-order-of-things-proceed advocates. Inertia is what it's all about. The normative power of the present state, if you will. Never mind that "natural" includes antibiotics, but not gene modification.

This may seem self-evident, but what I'm pointing out is that by saying "consider this world: would you still think the same way in that world?" you'd be skipping the actual step of difficulty: overcoming said inertia, leaving the cozy home of our local minimum.

Comment author: Lumifer 11 May 2015 04:56:00PM *  5 points [-]

Inertia is what it's all about. The normative power of the present state, if you will.

That's fine as long as you understand it and are not deluding yourself with a collection of reasons why this cozy local minimum is actually the best ever.

The considerable power wielded by inertia should be explicit.

Comment author: jkaufman 16 May 2015 02:23:27AM 0 points [-]

Huh? It feels like you're responding to a common thing people say, but not to anything I've said (or believe).

Comment author: Lumifer 16 May 2015 02:40:59AM 0 points [-]

I meant this as a response specifically to

But dramatically fewer children? Much less of the total human experience spent in early learning stages? Would we become less able to make progress in the world because people have trouble moving on from what they first learned?

Comment author: jkaufman 18 May 2015 12:00:59PM *  1 point [-]

More context:

A world in which we have ended death ... may be better than the world now, but I could also see it being worse. On one hand, not having to see your friends and family die, increased institutional memory, more time to get deeply into subjects and achieve mastery, and time to really build up old strong friendships sound good. But dramatically fewer children? Much less of the total human experience spent in early learning stages? Would we become less able to make progress in the world because people have trouble moving on from what they first learned?

I don't think our current lifespan is the perfect length, but there's a lot of room between "longer is probably better" and "effectively unlimited is ideal".

Comment author: Lumifer 18 May 2015 03:46:27PM 0 points [-]

there's a lot of room between "longer is probably better" and "effectively unlimited is ideal".

Yes, but are you saying there's going to a maximum somewhere in that space -- some metric will flip over and start going down? What might that metric be?

Comment author: jkaufman 20 May 2015 02:40:18PM 0 points [-]

As I wrote in that post, there are some factors that lead to us thinking longer lives would be better, and others that shorter would be better.

Maybe this is easier to think about with a related question: what is the ideal length of tenure at a company? Do companies do best when they have entirely employees-for-life, or is it helpful to have some churn? (Ignoring that people can come in with useful relevant knowledge they got working elsewhere.) Clearly too much churn is very bad for the company, but introducing new people to your practices and teaching them help you adapt and modernize, while if everyone has been there forever it can be hard to make adjustments to changing situations.

The main issue is that people tend to fixate some on what they learn when they're younger, so if people get much older on average then it would be harder to make progress.

Comment author: CellBioGuy 12 May 2015 03:15:58AM 0 points [-]

My take: there's a big difference between calling something good and dealing with a fact.

Comment author: ZankerH 11 May 2015 06:31:50AM 14 points [-]

Despite medical and police personnel aware of his Alcor bracelet, he was taken to the medical examiner’s office in Santa Barbara, as they did not understand Alcor’s process and assumed that the circumstances surrounding his death would pre-empt any possible donation directives. Since this all transpired late on a Friday evening, Alcor was not notified of the incident until the following Monday morning.

How the hell are they treating this as a successful preservation? The body spent two days "warm and dead".

Looking at their past case reports, this seems to be fairly normal. Unless you're dying of a known terminal condition and go die in their hospice in Arizona, odds are the only thing getting froze is a mindless, decaying corpse.

Comment author: advancedatheist 11 May 2015 02:32:07PM 10 points [-]

Cryonicist Ben Best has put a lot of effort into studying and testing personal alarm gadgets you can wear which signal cardiac arrest to try to reduce the incidence of these unattended deanimations and long delays before cryopreservation. I plan to look into those myself.

Ironically, I've noticed that cryonicists talk a lot about how much they believe in scientific, medical and technological progress, but then they don't seem to want to act on it when you present them with evidence of the correctable deficiencies of real, existing cryonics.

Reference:

Personal Alarm Systems for Cryonicists

http://www.benbest.com/cryonics/alarms.html

Comment author: Error 11 May 2015 01:31:17AM 1 point [-]

Just FYI, it looks like you goofed that last link.

Comment author: [deleted] 11 May 2015 07:23:43AM 3 points [-]
Comment author: CronoDAS 16 May 2015 01:05:48AM 1 point [-]

I can't understand what my girlfriend is saying when she uses her cell phone to call my cell phone. Often entire words are simply dropped from the audio stream. Is there anything I can do to improve the voice quality?

Comment author: moreati 16 May 2015 05:55:37PM 0 points [-]

A few thoughts based on eliminating/ruling out possible causes:

  • Can you avoid making cell -> cell calls? If you're both on a smartphone with wifi could you use e.g. Skype or Messenger?
  • Can you both use a hands free kit? This should eliminate poor positioning of the microphone/earpiece.
  • Are you or your girlfriend in a poor signal area? Does going outside to make the call reduce the problem? There are options such as GSM signal boosters and femtocells that might be worth exploring.
  • Some carriers have deployed HD Voice service. See if your phone and carrier(s) support this.
Comment author: CronoDAS 16 May 2015 08:28:35PM *  0 points [-]

I have a Nokia Lumia 822 (a Windows Phone). My girlfriend has an iPhone 5s. Our carrier is Verizon.

Comment author: moreati 17 May 2015 10:10:20AM 0 points [-]

Ah, sorry. I wasn't seeking answers to those questions. They were meant to suggest lines of enquiry you should follow, and workarounds you could try.

Comment author: taygetea 15 May 2015 10:40:41PM 1 point [-]

I've begun to notice discussion of AI risk in more and more places in the last year. Many of them reference Superintelligence. It doesn't seem like a confirmation bias/Baader-Meinhoff effect, not really. It's quite an unexpected change. Have others encountered a similar broadening in the sorts of people you encounter talking about this?

Comment author: Manfred 15 May 2015 11:41:29PM *  3 points [-]

Yup. Nick Bostrom is basically the man. Above and beyond being the man, he's a respectable focal point for a sea change that has been happening for broader reasons.

Comment author: [deleted] 15 May 2015 11:45:02AM *  -1 points [-]

Would this work at least as an early crude hypothesis of how neurotypicals function?

Neurotypicals like social mingling primarily because they play a constant game of social status points, both in the eyes of others (that is real status) and just feeling like getting status (this is more like self-esteem). This should not be understood as a harsh machiavellian cruel game. Usually not. Often it is very warm and friendly. For example, we on the spectrum often finding things like greeting each other superfluous. Needless custom. You notice when people arrive or when not they will talk to you when they want something. But for neurotypicals exchanging mutual warm greetings makes sense, it is a mutual reassurance or reinforcement of each others status. Sometimes people will ignore someone's greeting, even not shake an extended hand, that will be seen as a rude move to reduce the status of the other, as it is embarrassing. This is rudely dominant move they only pull if they are angry at each other. However, it will more often happen that there is some community, group, like a Toastmasters meeting or the local Linux user club or something and someone arrives and hands out a warmish greeting to the whole group but gets only nods in return, or a very short hi, this usually means "you are an outsider, newcomer, not fully accepted in this community yet, so while we don't reduce your status we are cautious about affirming it either, we stay noncommittant until you prove yourself". But then again someone will often be very warm to the newcomer, because acting as a mentor of noobs raises one status inside the group. It a friendly, helpful but clearly dominant move over the newcomer (let me show you the ropes here is the subconscious message), it also reaffirms one as a central member of the group as people who are themselves newbies would not do it, and as new and new people get into the group the people who mentored them cal slowly drift into leadership.

But greetings are just a simple example of the many ways neurotypicals enjoy playing status games. Again it is not a harsh thing, often very warm. But their constant social mingling, constant "purring" with each other is nothing but a set of microtransactions in status. All this small talk thing, plus the body language etc.

To enjoy gaining a score is a general human trait, even we on the spectrum do enjoy getting on the high score table in arcade games or levelling up in an RPG videgame or becoming a thane in Skyrim, we just don't notice how neurotypicals keep doing this all the time. All the social niceties boil down on handing out a micropayment of positive status, any tiny notion of kindness acts as so, the basic small talk of two random people meeting at a garden party will be feeling out each others status as a first step, getting liked by micropaying status ("Wow, that sounds like you have an interesting job!") in exchange for liking or raising your own status by a bit of boasting etc. And that is the nicer part, the uglier part is when people try to reduce each others scores, that is where bad blood comes from.

  1. Am I on a remotely right track here?
  2. Does the statement "nerds / neckbeards often have poor social skills" unpack into "people on the spectrum not even noticing that neurotypicals don't just mindlessly follow social customs, but they are involved in a status micropayment exchange" ?
  3. Any article or book that helps me go on in this line of thought?
Comment author: ChristianKl 16 May 2015 04:01:09PM *  1 point [-]

If you say Bob likes X because of Y, what do you mean with it? Do you mean that if Y wouldn't be there Bob wouldn't like X?

I don't think that there a good reason to believe that if you take status away no neurotically would engage in social mingling or like engaging in it.

Apart from that "status" is a word that's quite abstract. It's much more something "map" than "territory". That produces danger to get into too vague to be wrong territory.

Comment author: [deleted] 18 May 2015 08:29:22AM 1 point [-]

Apart from that "status" is a word that's quite abstract. It's much more something "map" than "territory".

Let's get more meta here. Usually the map-terrain distinction is used to describe how human minds interpret the chunks of reality that are not man-made. When we are talking about something that arises from the behavior of humans, how can we draw that distinction. Plato's classic "What is justice?" is map or terrain? Here the terrain is in human minds too, as justice exists only inside minds and nowhere else, so the distinction seems to be more like is it the grand shared map or a more private map of maps? And the same with status. It does not exist outside the human perception of it. Similar to money, esp. paper/computer number money.

I don't think that there a good reason to believe that if you take status away no neurotically would engage in social mingling or like engaging in it.

I will consider it a typo, assuming you meant neurotypicals like I did i.e. people outside the autism spectrum, or in other words non-geeks. I got the idea from here. If status microtransactions are so important...

(to be continued gotta go now)

Comment author: ChristianKl 18 May 2015 12:46:45PM 0 points [-]

If you look at the link you posted it argues:

Status is a confusing term, unless it’s understood as something one does. You may be low in status, but play high, and vice versa. … We always like it when a tramp is mistaken for the boss, or the boss for a tramp. … I should really talk about dominance and submission, but I’d create a resistance.

That means that dominance and submission map more directly to the territory than status does.

The author doesn't argue that people care about mutually reinforcement of each other status as being high but that people also consciously make moves to submit and place themselves at a low status position.

The text invalidates your idea that people engage primarily in social interaction to maximize the amount of status.

You don't pick that up if you make the error of not treating status as a model but as reality. Reality is complex. Models simplify reality. Sometimes the simplification keeps the essential elements of what you want to describe. Other times it doesn't.

I will consider it a typo, assuming you meant neurotypicals like I did i.e. people outside the autism spectrum, or in other words non-geeks.

Yes, it's a typo likely because my spellchecker didn't know "neurotypicals".

Comment author: Zubon 15 May 2015 03:45:45PM *  3 points [-]
  1. Am I on a remotely right track here?

That sounds like a subset of neurotypical behavior. I'm neurotypical and from the very first sentence ("Neurotypicals like social mingling primarily because they play a constant game of social status points, both in the eyes of others (that is real status) and just feeling like getting status (this is more like self-esteem).") I found it contrary to my experience. Which is not to say it is wrong, and it certainly looks like behavior I have seen, but it kind of suggests that there is One Neurotypical Experience as opposed to a spectrum.

That is reading the initial "neurotypicals" as "all/most neurotypicals" as opposed to "some neurotypicals" or "some subset of neurotypicals." I think you are trying to describe typical neurotypical behavior, so I would read that "neurotypicals" as trying to describe how most neurotypicals behave.

But I am not the most central example of a neurotypical, so others may find it a more accurate description of their social experiences. I don't like social mingling, and I avoid most games of social status points. My extroversion score is 7 out of 100, which is likely a factor in not seeing myself in your description of neurotypicals.

2. Does the statement "nerds / neckbeards often have poor social skills" unpack into "people on the spectrum not even noticing that neurotypicals don't just mindlessly follow social customs, but they are involved in a status micropayment exchange" ?

There seem to be several assumptions built into that unpacking. For example, it suggests that all/most nerds are on the spectrum. My characterization of neurotypical socialization would include mindlessly following social customs as well as enjoying the social game. I don't think highly social neurotypicals would describe their behavior as a "status micropayment exchange"; that seems like the wrong metaphor and suggests the dominant model as a fixed-sum status game, whereas many (most?) social interactions have no need for an exchange of status points.

Even when a social status point game is in play, I would expect more interactions to involve the recognition of point totals rather than an exchange. "Mutual reassurance or reinforcement of each others status" seems on point.

If the above is the start of a hypothesis, it seems to me that it links greetings and status point exchange too strongly. Greetings are rarely an occasion to gain or lose points, although they may be occasions to discover the current score.

"Pinging" is a metaphor I have seen used productively in these attempts to explain neurotypical social behavior. The greeting is a ping, a mutual recognition that someone is there and potentially responsive to interaction, potentially also exchanging some basic status information.

Comment author: TsviBT 12 May 2015 02:03:29AM 1 point [-]

PSA: If you wear glasses, you might want to take a look behind the little nosepads. Some... stuff... can build up there. According to this unverified source it is oxidized copper from glasses frame + your sweat, and can be cleaned with an old toothbrush + toothpaste.

Comment author: Dorikka 12 May 2015 02:17:57AM 6 points [-]

Sounds like the only disutility of the stuff is that it annoys some people, but it can't annoy you if you dont notice it...so why bring it up?

Comment author: philh 12 May 2015 08:45:52AM 6 points [-]

It mildly annoys me, but I hadn't thought to use toothpaste on it, so there's that.

Comment author: DataPacRat 14 May 2015 04:37:57AM *  0 points [-]

Acting on A Gut Feeling

I've been planning an overnight camping trip for sometime this week; but something about the idea is making me feel... disquiet. Uneasy. I can't figure out why; I've got a nice set of equipment, I have people who know where I'm going, and so on. But I can't shake something resembling an "ugh field" that eases when I think of /not/ taking the trip.

And so, I'm concluding that the rational thing to do is to pay attention to my gut, on the chance that one part of my mind is aware of some detail that the rest of my mind hasn't figured out, and postpone my camping trip until I'm feeling more self-assured about the whole thing.

Comment author: DataPacRat 14 May 2015 06:34:11AM 0 points [-]

After some further mental gymnastics, the plan I've come up with which seems to most greatly reduce the disquiet is to buy a backup cellphone, small enough to turn off, stick in a pocket and forget about until I drop my smartphone in a stream. Something along the lines of taking one of the watchphones from http://www.dx.com/s/850%2b1900?category=529&PriceSort=up and snipping off the wristband, or one of the smaller entries in http://www.dx.com/s/850%2b1900?PriceSort=up&category=531 ; along with the $25/year plan from http://www.speakout7eleven.ca/ . Something on the order of $65 to $85 seems a moderate price for peace of mind.

I am, however, going to take at least a day before placing any such order, to find out if such a plan still seems like it /will/ offer increased peace of mind. Not to mention, whether I can come up with (or get suggestions for) any plans which reveal that my actual disquiet arises from some other cause.

Comment author: Miguelatron 15 May 2015 02:16:32PM 1 point [-]

Have you gone camping like this before? If you have, were you by yourself when you did? I'm just trying to eliminate the source of your unease being something simple like stepping out of your comfort zone.

Comment author: DataPacRat 15 May 2015 05:52:26PM 0 points [-]

I have, indeed, gone camping like this before, though it's been a few years since I've done anything solo. The last few times I've gone camping has been with a relative to campgrounds with showers and such amenities, as opposed to solo in a conservation area or along a trail, which is/was my goal for my next hike. My original motivation for the overnighter was to make sure I hadn't forgotten anything important about soloing, and that all my gear's ready for longer trips.

I'm in the general Niagara area, and the city papers laud the local rescue teams whenever a tourist needs to get pulled out of the Niagara gorge, so as long as I can dial 911, I should be able to get rescued from any situation I get myself in that's actually worth all this worrying about. The particular spot I'm thinking of going to (43.0911426, -79.284342) is roughly an hour's walk from a city bus stop - half an hour's walk from where I could wave to frequently passing cars, if my phone's dead.

My plans for this whole trip have been to make it as simple and easy as possible. Amble down some trails for an hour or two, hang my hammock, cook my dinner, read my ebook, and amble on out the next day, enjoying the peace and quiet and so on. It's the smallest step I can think of beyond camping in a backyard - and since I don't have a backyard, it's pretty much as far within my comfort zone as any camping could be. If /that's/ now outside my comfort zone... then I've got a trunk full of camping gear that's suddenly a lot less useful to me.

Comment author: Miguelatron 17 May 2015 01:52:36PM *  0 points [-]

Sounds like it will be a blast. The nerves may just be from going solo then. Sounds like you know what your about though, so I'd just override any trepidation and go for it.

I did something similar a few weeks ago (admittedly with some friends). We were probably 40 miles from anywhere where we could flag down a car, and hiked into the woods several miles along the trail. My backpack broke inside the first mile, one of my friends slipped and fell into a stream, there were coyotes in the camp at night, and of course it rained. We all made it out sleepy sore and soggy the next, day but definitely felt better for having gone. Would do again.

You'll have a good time, no worries.

Comment author: wadavis 14 May 2015 10:02:38PM 0 points [-]

It is because you forgot to pack TP. Bring TP and things will be ok.

Comment author: DataPacRat 14 May 2015 10:11:08PM 1 point [-]

:)

Don't worry, I've got the essentials. And enough luxuries, like a folding solar panel, that I could head out for a week or more, if I were so inclined, and bought an upgrade to my cellphone dataplan.

Considering from various perspectives, a trip to some nearby city and staying at an Airbnb or hotel raises more interest than disquiet; so it seems to be something about going camping, rather than taking a trip, which is bothering me. An imagined day-hike only raises questions about transportation, not unease, so it seems to be something about overnighting. Cooking? Water source? Sleeping? First-aid kit? Emergency plans in case of zombie outbreak (or more probable disasters)? I can't quite put my finger on it.

And since almost the whole point of such a trip is to /improve/ my psychological condition by the end of it, I'm starting to feel a tad annoyed at myself for being less than clear about my motivations to me. :P

Comment author: Dorikka 15 May 2015 04:20:16AM *  0 points [-]

Hm. I have been camping quite a few times, but would not really be comfortable camping alone. Might be true for you as well due to perceived lowering of risk.

ETA: This is more of a preference thing for me than an actual concern thing.

Comment author: [deleted] 11 May 2015 08:02:51AM *  2 points [-]

In what conceiveable (which does not imply logicality) universes would Rationalism not work in the sense of unearthing only some truths, not all truths? Some realms of truth would be hidden to Rationalists? To simplify it, I mean largely the aspect that of empiricism, of tying ideas to observations via prediction. What conceivable universes have non-observational truths, for example, Platonic/Kantian "pure apriori deduction" type of mental-only truths? Imagine for convenience's sake a Matrix type simulated universe, not necessarily a natural one, so it does not really need to be lawful nor unfold from basic laws.

Reason for asking: if you head over to a site like The Orthosphere, they will tell you Rationalism can only find some but not all truths. And one good answer would be: "This could happen in universes of the type X, Y, Z. What are your reasons for thinking ours could be one of them?"

Comment author: OrphanWilde 11 May 2015 07:26:15PM 4 points [-]

Depends on how you decide what truth is, and what qualifies it to be "unearthed."

But for one universe in which some truth, for some value of truth, can be unearthed, for some value of unearthed, while other truth can't be:

Imagine a universe in which 12.879% (exactly) of all matter is a unique kind of matter that shares no qualities in common with any other matter, and is almost entirely nonreactive with all other kinds of matter, and was created by a process not shared in common with any other matter, which had no effect whatsoever on any other matter. Any truths about this matter, including its existence and the percentage of the universe composed of it, would be completely non-observational. The only reaction this matter has with any other matter is when it is in a specific configuration which requires extremely high levels of the local equivalent of negative entropy, at which point it emits a single electromagnetic pulse. This was used once by an intelligence species composed of this unique matter who then went on to die in massive wars, to encode in a series of flashes of light every detail they knew about physics, and was observed by one human-equivalent monk ascetic, who used a language similar to morse code to write down the sequence of pulses, which he described as a holy vision. Centuries later, these pulses were translated into mathematical equations which described the unique physics of this concurrent universe of exotic matter, but no mechanism of proving the existence or nonexistence of this exotic matter, save that the equations are far beyond the mathematics of anyone alive at the time the signal was encoded, and it has become a controversial matter whether or not it was an elaborate hoax by a genius.

Comment author: drethelin 11 May 2015 07:00:17PM 2 points [-]

A universe where humans are running on brains with certain glitches that prevent them from coming to correct conclusions through reasoning about specific topics.

Comment author: ChristianKl 11 May 2015 12:33:49PM 4 points [-]

What do you mean with "Rationalism"?

The LW standard definition is that it's about systematized winning. If the Matrix overlords punish everybody who tries to do systematized winning than it's bad to engage in it. Especially when the Matrix overlords do it via mind reading. The Christian God might see it as a sin.

If you don't use the LW definition of rationalism, then rationalism and empiricism are not the same thing. Rationalism generally refers to gathering knowledge by reasoning as opposed to gathering it by other ways such as experiments or divine revelation.

they will tell you Rationalism can only find some but not all truths

Gödel did prove that it's impossible to find all truths. This website is called Lesswrong because it's not about learning all truths but just about becoming less wrong.

Comment author: ike 11 May 2015 02:54:31PM 1 point [-]

Gödel did prove that it's impossible to find all truths.

That's misleading. With a finite amount of processing power/storage/etc, you can't find all proofs in any infinite system. We need to show that short truths can't be found, which is a bit harder.

Comment author: Houshalter 14 May 2015 06:15:41AM 0 points [-]

I don't think that's correct. My best understanding of Godel's theorem is that if your system of logic is powerful enough to express itself, then you can create a statement like "this sentence is unprovable". That's pretty short and doesn't rely on infiniteness.

Comment author: ike 14 May 2015 01:32:23PM 0 points [-]

The statement "this sentence is unprovable" necessarily includes all information on how to prove things, so it's always larger than your logical system. It's usually much larger, because "this sentence" requires some tricks to encode.

To see this another way, the halting problem can be seen as equivalent to Godel's theorem. But it's trivially possible to have a program of length X+C that solves the halting problem for all programs of length X, where C is a rather low constant; see https://en.wikipedia.org/wiki/Chaitin'sconstant#Relationshiptothehalting_problem for how.

Comment author: Houshalter 15 May 2015 09:25:38AM 0 points [-]

I'm not sure how much space it would take to write down formally, and I'm not sure it matters. At worst it's a few pages, but not entire books, let alone some exponentially huge thing you'd never encounter in reality.

It's also not totally arbitrary axioms that would never be encountered in reality. There are reasons why someone might want to define the rules of logic within logic, and then 99% of the hard work is done.

But regardless, the interesting thing is that such an unprovable sentence exists at all. That its not possible to prove all true statements with any system of logic. It's possible that the problem is limited to this single edge case, but for all I know these unprovable sentences could be everywhere. Or worse, that it is possible to prove them, and therefore possible to prove false statements.

I think the halting problem is related, but I don't see how it's exactly equivalent. In any case the halting problem work around is totally impractical, since it would take multiple ages of the universe to prove the haltingness of a simple loop. If you are referring to the limited memory version, otherwise I'm extremely skeptical.

Comment author: ike 15 May 2015 04:53:50PM 0 points [-]

At worst it's a few pages, but not entire books, let alone some exponentially huge thing you'd never encounter in reality.

That's only if your logical system is simple. If you're a human, then the system you're using is probably not a real logical system, and is anyway going to be rather large.

I think the halting problem is related, but I don't see how it's exactly equivalent.

See http://www.solipsistslog.com/halting-consequences-godel/

Comment author: SolveIt 11 May 2015 08:50:25AM 3 points [-]

There's no guarantee we should be able to find any truths using any method. It's a miracle that the universe is at all comprehensible. The question isn't "when can't we learn everything?", it's "why can we learn anything at all?".

Comment author: Lumifer 11 May 2015 04:08:21PM -2 points [-]

"why can we learn anything at all?"

Because entities which can't do not survive.

Comment author: CronoDAS 16 May 2015 01:16:05AM 2 points [-]

Counterexample: Plants. Do they learn?

Comment author: Lumifer 16 May 2015 01:30:41AM 1 point [-]

Of course. Leaves turn to follow the sun, roots grow in the direction of more moist soil...

Comment author: CronoDAS 16 May 2015 08:17:07PM 2 points [-]

Is that really learning, or just reacting to stimuli in a fixed, predetermined pattern?

Comment author: Romashka 17 May 2015 09:53:17AM 1 point [-]

Does vaccination imply memory?.. Does being warned by another's volatile metabolites that a herbivore is attacking the population?

(Higher) plants are organized by very different principles than animals; it is a never-ending debate on what constitutes 'identity' in them. Without first deciding upon that, can one speak about learning? I don't think they have it, but their patterns of predetermined answers can be very specific.

Comment author: Romashka 16 May 2015 07:40:55AM 0 points [-]

Also, there is an interesting study, 'Kin recognition, not competitive interactions, predicts root allocation in young Cakile edentula seedling pairs'. This seems to be more difficult to do than following the sun!

Comment author: SolveIt 11 May 2015 09:27:34PM 1 point [-]

That just pushes the question back a step. Why can any entity learn?

Comment author: tim 12 May 2015 03:32:59AM 1 point [-]

In the spirit of Lumifer's comment, anything we would consider an entity would have to be able to learn or we wouldn't be considering it at all.

Comment author: DanielLC 12 May 2015 05:40:14AM 0 points [-]

That would explain why all entities learn. Not why any entities learn. Ignoring things that can't learn doesn't explain the existence if things that can.

Comment author: RichardKennaway 16 May 2015 06:02:43AM 0 points [-]

Why can any entity learn?

A more useful question to ask would be "how do entities, in fact, learn?" This avoids the trite answer, "because if they didn't, we wouldn't be asking the question".

Comment author: Lumifer 12 May 2015 02:32:01PM 0 points [-]

I think if we follows this chain of questions, what we'll find at the end (except for turtles, of course) is the question "Why is the universe stable/regular instead of utterly chaotic?" A similar question is "Why does the universe even have negentropy?"

I don't know any answer to these questions except for "That's what our universe is".

Comment author: SolveIt 12 May 2015 02:38:40PM 1 point [-]

I suppose what I want to know is the answer to "What features of our universe make it possible for entities to learn?".

Which sounds remarkably similar to DeVliegendeHollander's question, perhaps with an implicit assumption that learning won't be present in many (most?) universes.

Comment author: Lumifer 12 May 2015 04:24:14PM 1 point [-]

The fact that the universe is stable/regular enough to be predictable. Subject predictability is a necessary requirement for learning.

Comment author: fubarobfusco 11 May 2015 08:55:00PM 0 points [-]

For that matter, a world in which it is impossible for an organism to become better at surviving by modeling its environment (i.e. learning) is one in which intelligence can't evolve.

(And a world in which it is impossible for one organism to be better at surviving than another organism, is one in which evolution doesn't happen at all; indeed, life wouldn't happen.)

Comment author: IlyaShpitser 11 May 2015 08:18:59AM *  3 points [-]

Don't need to posit crazy things, just think about selection bias -- are the sorts of people that tend to become rationalist randomly sampled from the population? If not, why wouldn't there be blind spots in such people just based on that?

Comment author: [deleted] 11 May 2015 08:40:52AM 1 point [-]

Yes, but if I get the idea right, it is to learn to think in a self-correcting, self-improving way. For example, maybe Kanazawa is right in intelligence suppressing instincts / common sense, but a consistent application of rationality sooner or later would lead to discovering it and forming strategies to correct it.

For this reason, it is more of the rules (of self-correction, self-improvement, self-updating sets of beliefs) than the people. What kinds of truths would be potentially invisible to a self-correcting observationalist ruleset even if this was practiced by all kinds of people?

Comment author: IlyaShpitser 11 May 2015 08:49:46AM *  3 points [-]

Just pick any of a large set of things the LW-sphere gets consistently wrong. You can't separate the "ism" from the people (the "ists"), in my opinion. The proof of the effectiveness of the "ism" lies in the "ists".

Comment author: NancyLebovitz 11 May 2015 11:58:23AM 2 points [-]

Which things are you thinking of?

Comment author: IlyaShpitser 11 May 2015 02:38:57PM *  2 points [-]

A lot of opinions much of LW inherited uncritically from EY, for example. That isn't to say that EY doesn't have many correct opinions, he certainly does, but a lot of his opinions are also idiosyncratic, weird, and technically incorrect.

As is true for most of us. The recipe here is to be widely read (LW has a poor scholarship problem too). Not moving away from EY's more idiosynchratic opinions is sort of a bad sign for the "ism."

Comment author: NancyLebovitz 11 May 2015 02:44:49PM 1 point [-]

Could you mention some of the specific beliefs you think are wrong?

Comment author: IlyaShpitser 11 May 2015 02:58:42PM *  8 points [-]

Having strong opinions on QM interpretations is "not even wrong."

LW's attitude on B is, at best, "arguable."

Donating to MIRI as an effective use of money is, at best, "arguable."

LW consequentialism is, at best, "arguable."

Shitting on philosophy.

Ratonalism as part of identity (aspiring rationalist) is kind of dangerous.

etc.


What I personally find valuable is "adapting the rationalist kung fu stance" for certain purposes.

Comment author: OrphanWilde 11 May 2015 03:17:40PM *  2 points [-]

Ratonalism as part of identity (aspiring rationalist) is kind of dangerous.

[Edited formatting] Strongly agree. http://lesswrong.com/lw/huk/emotional_basilisks/ is an experiment I ran which demonstrates the issue. Eliezer was unable to -consider- the hypothetical; it "had" to be fought.

The reason being, the hypothetical implies a contradiction in rationality as Eliezer defines it; if rationalism requires atheism, and atheism doesn't "win" as well as religion, then the "rationality is winning" definition Eliezer uses breaks; suddenly rationality, via winning, can require irrational behavior. Less Wrong has a -massive- blind spot where rationality is concerned; for a web site which spends a significant amount of time discussing how to update "correctness" algorithms, actually posing challenges to "correctness" algorithms is one of the quickest ways to shut somebody's brain down and put them in a reactionary mode.

Comment author: ChristianKl 11 May 2015 04:25:40PM 0 points [-]

if rationalism requires atheism

I don't think that's argued. It's also worth noting that the majority of MIRI's funding over it's history comes from a theist.

Comment author: RichardKennaway 12 May 2015 12:05:43PM 1 point [-]

Eliezer was unable to -consider- the hypothetical; it "had" to be fought.

It seems to me that he did consider your hypothetical, and argued that it should be fought. I agree: your hypothetical is just another in the tedious series of hypotheticals on LessWrong of the form, "Suppose P were true? Then P would be true!"

BTW, you never answered his answer. Should I conclude that you are unable to consider his answer?

Eliezer also has Harry Potter in MoR withholding knowledge of the True Patronus from Dumbledore, because he realises that Dumbledore would not be able to cast it, and would no longer be able to cast the ordinary Patronus.

Now, he has a war against the Dark Lord to fight, and cannot take the time and risk of trying to persuade Dumbledore to an inner conviction that death is a great evil in order to enable him to cast the True Patronus. It might be worth pursuing after winning that war, if they both survive.

All this has a parallel with your hypothetical.

Comment author: TheAncientGeek 12 May 2015 05:00:22PM *  0 points [-]

I 've notice that problem, but I think it is a bit dramatic to call it rationality breaking. I think it's more of a problem of calling two things, the winning thing amd the truth seeking thing, by one name.

Comment author: NancyLebovitz 11 May 2015 03:30:51PM 2 points [-]

Thank you.

LW's attitude on B is, at best, "arguable."

B?

Comment author: Luke_A_Somers 12 May 2015 12:19:38AM 0 points [-]

Well...

QM: Having strong positive beliefs on the subject would be not-even-wrong. Ruling out some is much less so. And that's what he did. Note, I came to the same conclusion long before.

MIRI: It's not uncritically accepted on LW more than you'd expect given who runs the joint.

Identity: If you're not letting it trap you by thinking it makes you right, if you're not letting it trap you by thinking it makes others wrong, then what dangers are you thinking of? People will get identities. This particular one seems well-suited to mitigating the dangers of identities.

Others: more clarification required

Comment author: John_Maxwell_IV 11 May 2015 07:42:42AM *  3 points [-]

I recently found this blog post by Ben Kuhn where he briefly summarizes ~5 classic LW posts in the space of one blog post. A couple points:

  • I don't think that much of the content of the original posts is lost in Ben's summary, and it's a lot faster to read. Do others agree? Do we think producing a condensed summary of the LW archives at some point might be valuable? (It's possible that, for instance, the longer treatment of these concepts in the original posts pushes them deeper in to your brain, or that since people are so used to skimming, conceptually dense content can actually be bad.)

  • Might it be worthwhile to link to this from the LW about page or FAQ for newcomers?

Submitting...

Comment author: Kindly 11 May 2015 11:37:06PM 5 points [-]

This ought to be verified by someone to whom the ideas are genuinely unfamiliar.

Comment author: dxu 12 May 2015 02:58:56AM *  5 points [-]

To any physicists out there:

This idea came to me while I was replaying the game Portal. Basically, suppose humanity one day developed the ability to create wormholes. Would one be able to generate an infinite amount of energy by placing one end of a wormhole directly below the other before dropping an object into the lower portal (thus periodically resetting said object's gravitational potential energy while leaving its kinetic energy unaffected)? This seems like a blatant violation of the first law of thermodynamics, so I'm guessing it would fail due to some reason or other (my guess goes to weird behavior of the gravitational field near the wormhole, which interferes with the larger field of the Earth), but since I'm nowhere close to being a physicist, I thought I'd ask about it on LessWrong.

So? Any ideas as to what goes wrong in the above example?

Comment author: shminux 14 May 2015 07:08:51PM *  2 points [-]

I have a PhD in Physics and my thesis was, in part, related to wormholes, so here it goes. (Squark covered most of your question already, though.)

If something falls into a black hole, it increases the black hole mass. If something escapes a black hole (such as Hawking radiation), it decreases the black hole mass. Same with white holes. A wormhole is basically two black/white holes connected by a throat. One pass through the portal would increase the mass of the entrance and decrease the mass of the exit by the mass of the passing object.

A portal with two ends having opposite masses would behave rather strangely: they sort of repel (the equivalent of Newton's law of gravity), but the gravitational force acting on the negative-mass end propels it toward the positive-mass end. As a result, the portal as a whole will tend to accelerate toward the positive end (entrance) and fly away, albeit rather slowly.

In addition, due to momentum and angular momentum conservation, the portal will start spinning to counteract the motion of the passing object.

Comment author: Squark 14 May 2015 06:19:13PM *  2 points [-]

Wormholes don't quite behave like portals in the game.

When something drops into a wormhole with zero velocity, the apparent mass of the entry end increases by the mass of the object and the apparent mass of the exit end decreases by the mass of the object. At some point one of the ends should acquire negative mass. I'm not sure what that means: either it literally behaves as a negative mass object or this is an indication of the wormhole becoming unstable and collapsing.

Similarly, when something with momentum drops into a wormhole, the momentum is added to the apparent momentum of the entry end and subtracted from the apparent momentum of the exit end. The apparent masses change in a way that ensures energy conservation. This means that the gain in energy of the "cycling" object comes from wormhole mass loss and transfer of mass from the high end to the low end. Again, if it's true that the wormhole becomes unstable when its mass is supposed to go negative, that would be the end of the process.

Comment author: shminux 14 May 2015 07:14:11PM 1 point [-]

If you already postulate having enough negative energy to create a wormhole, there is no extra issues due to one of the throats having negative mass, except the weird acceleration effect, as I mentioned in my other reply.

Comment author: Squark 07 June 2015 07:13:15PM 0 points [-]

Maybe. However, how will the geometry look like when the sign flip occurs? Will it be non-singular?

Comment author: shminux 07 June 2015 09:23:57PM 0 points [-]

There isn't as much difference between negative- and positive-mass wormholes as between negative- and positive mass black holes. Negative-mass black holes have no horizons and a naked repulsive timelike singularity. A negative- (at infinity) mass wormhole would look basically like a regular wormhole. The local spacetime curvature would, of course, be different, but the topology would remain the same, S^2xRxR or similar.

Comment author: ZankerH 14 May 2015 11:28:46AM 1 point [-]

At a glance, it seems like you're asking for extrapolation from a "suppose X - therefore X" - type statement, where X is the invalidation of conservation laws.

Comment author: dxu 15 May 2015 04:12:38AM *  1 point [-]

I don't quite understand this statement. The only real premise I can see in my original comment is

suppose humanity one day developed the ability to create wormholes.

(Please feel free to correct me if you were in fact referring to some other premise.)

Wormholes are generally agreed to be a possible solution to Einstein's equations--they don't, in and of themselves, violate conservation of energy. The scenario I proposed above is a method for generating infinite energy if physics actually worked that way, but since I'm confident that it doesn't, the proposed scenario is almost certainly flawed in some way. I asked my question because I wasn't sure how it was flawed. Whatever the flaw is, however, I doubt it lies in the wormhole premise.

EDIT: Also see the replies from shminux and Squark.

Comment author: CellBioGuy 12 May 2015 03:20:37AM 4 points [-]

Gravity is a conserved vector field. Any closed path through a gravitational potential leaves you with the same energy you started with. And if it doesn't you've stolen energy that was creating the gravity in the first place leaving less for the next circle to take and are thus just transforming energy from one to another.

Comment author: DanielLC 12 May 2015 05:37:41AM 0 points [-]

Does that apply when the space isn't simply connected? It could be conservative at every neighborhood, but not conservative over all if you allow portals.

Comment author: CellBioGuy 12 May 2015 01:32:54PM *  2 points [-]

Gravity would propagate through the connected space. The potential would probably be very VERY weird shaped but i see no reason it wouldnt be conservative or otherwise consistent with GR (the math of which is far beyond me). Though keep in mind in GR space curvature IS gravity and can change over time, I doubt you could maintain a knife edge thin aperture, it would all smooth out.

What's really fun though are gravetomagnetic effects.

These are to gravity what magnetism is to the electric field. Both the electric field and gravitational field are conservative. But changing or accelerating charges generate magnetic fields which are NOT conservative hence how an electron spinning around a coil in a field in a generator gains energy even though it returns to its starting point. However in doing so it accelerates up to velocity as well generating a counteracting field that cancels out some of the field accelerating it, the motion moving it, or both. Thus the nonconservative fields have a potential energy associated with them that can be extracted from them or used to couple two phenomena that are both coupled to them.

To get gravetomagnetic effects you need huuuuuge mass flows and accelerations. But you can similarly steal the energy that drives them. Think frame dragging and extraction of black hole rotation.

Comment author: DanielLC 12 May 2015 06:39:01PM 0 points [-]

but i see no reason it wouldnt be conservative or otherwise consistent with GR

I'm assuming GR holds. Does it actually prove the field is conservative, or just irrotational? If it's irrotational and simply connected, then it's conservative, but if you stick a portal in it, it might not be.

I doubt you could maintain a knife edge thin aperture, it would all smooth out.

Portals don't need a knife edge. For example: Flight Through a Wormhole. You do need negative energy density or it will collapse, but that on its own shouldn't break conservation of energy.

Comment author: Slider 18 May 2015 03:22:20PM 0 points [-]

I am just taking wormholes to mean "altered connectivity of space" and leave out the "massive concentrations of mass" aspect.

The curious thing about portals portals is that they somehow magically know to flip gravity when a object travels thourht. If the portal is just ordinary space there shouldn't be a sudden gradient to the gravity field but it should go smoothly from one direction to the other. And in additon gravity ought to work throught portals. that would mean that if you have a portal in a ceiling it ought to pull stuff throught it towards the ceiling (towards the center of mass beyond the portal). That is a standard "infinite fall" portal setup should feel equal gravity up and down midway between the portals. That kind of setup could be used to store kinetic energy but it doesn't generate it per se.

However if portals aftected the gravity fields it could be that the non-standard gravity environment could be a major problem and would work even when you didn't want it to. That is since the net 0 gravity point of a infinite fall setup needs to transition smoothly to the "standard gravity environment" that likely means that quite a ways "outside" the portal pair there would be a reduced gravity environment.

Comment author: falenas108 12 May 2015 04:11:26AM 0 points [-]

You can probably think about it as the lines of a gravity field also going through the wormhole, and I believe the gravitational force would be 0 around the wormhole.

The actual answer involves thinking about gravity and spacetime as a geometry, which I don't think you want to answer your question.

Comment author: taygetea 17 May 2015 10:57:13PM 3 points [-]

Hi. I don't post much, but if anyone who knows me can vouch for me here, I would appreciate it.

I have a bit of a Situation, and I would like some help. I'm fairly sure it will be positive utility, not just positive fuzzies. Doesn't stop me feeling ridiculous for needing it. But if any of you can, I would appreciate donations, feedback, or anything else over here: http://www.gofundme.com/usc9j4

Comment author: Lumifer 14 May 2015 02:54:37PM *  6 points [-]

Pretty awesome set of trolley problems

Sample:

There’s an out of control trolley speeding towards Immanuel Kant. You have the ability to pull a lever and change the trolley’s path so it hits Jeremy Bentham instead. Jeremy Bentham clutches the only existing copy of Kant’s Groundwork of the Metaphysic of Morals. Kant holds the only existing copy of Bentham’s The Principles of Morals and Legislation. Both of them are shouting at you that they have recently started to reconsider their ethical stances.

Comment author: OrphanWilde 13 May 2015 07:13:30PM 2 points [-]

Anybody care to weigh in on adding a flag to newbies, and make it part of the LessWrong culture to explain downvotes to flagged newbies?

Identifying what you've done incorrectly to provoke downvotes is a skill that requires training. (Especially since voting behavior in Discussion is much less consistent to voting behavior in Main.)

Comment author: Gunnar_Zarncke 14 May 2015 06:29:57AM 2 points [-]

You can detect newbies by their low karma and moderate positive ratio. The registration age doesn't mean much really.

Comment author: philh 14 May 2015 01:39:04PM *  0 points [-]

You have to click through to discover that though, and there are exceptions who have a low ratio but don't need downvotes explained to them. (I don't know if there are such users with a low ratio and low total, though.)

Comment author: Gunnar_Zarncke 14 May 2015 03:46:37PM 0 points [-]

You can see the ration in the tool tip over the karma.

Interestingly there are users with positive ratios arbitrarily close to 50%.

Comment author: philh 14 May 2015 05:31:34PM 2 points [-]

You can see a comment's karma ratio by hovering on the comment's karma, but to see the user's ratio you need to click through to the user's page and hover over their karma there.

Comment author: ChristianKl 13 May 2015 01:48:43PM 14 points [-]

Recent news suggest that having measles weakens the immune system for 2 to 3 years afterwards and therefore the Measles vaccine manages to reduce a lot of childhood deaths that weren't thought to be measles related.

NPR article Academic paper

Comment author: SolveIt 12 May 2015 09:55:05AM 1 point [-]

The rationalist Tumblr community looks interesting. Any tips on how to start?

Comment author: Gondolinian 12 May 2015 11:45:38PM 5 points [-]

Well, for a start there's the Rationalist Masterlist currently hosted by Yxoque (MathiasZaman here on LW). You could announce your presence there and ask to be added to the list, or just lurk around some of the blogs for a while and send anonymous asks to people to get a feel for the community before you set up an account.

Comment author: SolveIt 13 May 2015 01:49:12AM 1 point [-]

Thanks!

Comment author: Silver_Swift 12 May 2015 08:51:39AM 1 point [-]

Is this the place to ask technical questions about how the site works? If so, then I'm wondering why I can't find any of the rationality quote threads on the main discussion page anymore (I thought we'd just stopped doing those, until I saw it pop up in the side bar just now). If not, then I think I just asked anyway. :P

Comment author: NancyLebovitz 12 May 2015 02:04:33PM 6 points [-]

This is a good place to ask about how the site works.

Comment author: gjm 12 May 2015 09:05:36AM 5 points [-]

Here -- it's in Main rather than Discussion.

Comment author: Silver_Swift 12 May 2015 01:16:23PM 2 points [-]

Thanks!

Comment author: Dahlen 12 May 2015 12:33:59AM 4 points [-]

Is utilitarianism foundational to LessWrong? Asking because for a while I've been toying with the idea of writing a few posts with morality as a theme, from the standpoint of, broadly, virtue ethics -- with some pragmatic and descriptive ethics thrown in. (The themes are quite generous and interlocking, and to be honest I don't know where to start or whether I'll finish it.) This perspective treats stable character traits, with their associated emotions, drives, and motives as the most reasonably likely determiner of moral behaviour, and means to encourage people to "build character" so as to become more moral beings or improve their behaviour. It doesn't concern itself with quantitative approaches to welfare. Frankly, I find it hard to take seriously the numerical applications of utilitarianism, and my brain just shuts down upon some ethical problems usually enjoyed around here (torture vs. dust specks, repugnant conclusion, contrived deals with strange gods and so on).

I know that Eliezer's virtues-of-rationality post is widely appreciated by many people around here, but it's a declaration of (commitment to) values more than anything. It never seemed to be the dominant paradigm. I guess I just want to know whether a virtue-ethical approach would be well-received here, and the extent to which a utilitarian and a virtue ethicist can usefully discuss morality without jumping a meta level into theories of normative ethics.

Comment author: Gunnar_Zarncke 14 May 2015 06:24:50AM 1 point [-]

Again I recommend a poll:

Is utilitarianism foundational to LessWrong? (use the middle option to see results only)

Agree Disagree

Submitting...

Comment author: ilzolende 13 May 2015 03:20:10AM 4 points [-]

I endorse discussion of virtue ethics on LW mostly because I haven't seen many arguments for why I should use it or discussions of how using it works. I've seen a lot of pro-utilitarianism and "how to do things with utilitarianism" pieces and a lot of discussion of deontology in the form of credible precommitments and also as heuristics and rule utilitarianism, but I haven't really seen a virtue ethics piece that remotely approaches Yvain's Consequentialism FAQ in terms of readability and usability.

Comment author: BrassLion 13 May 2015 02:03:08AM 3 points [-]

Consequentialism, where morality is viewed through a lens of what happens due to human actions, is a major part of LessWrong. Utilitarianism specifically, where you judge an act by the results, is a subset of consequentialism and not nearly as widely accepted. Virtue Ethics are generally well liked and it's often said around here that "Consequentialism is what's right, Virtue Ethics are what works." I think that practical guide to virtue ethics would be well received.

Comment author: Vaniver 12 May 2015 01:41:25PM 3 points [-]

Is utilitarianism foundational to LessWrong?

No. Individual utility calculations are, as a component of decision theory, but decision-theoretic utility and interpersonal-comparison utility are different things with different assumptions.

encourage people to "build character" so as to become more moral beings or improve their behaviour.

This is a solid view, and one of the main ones I take--but I observe that listing out goals and developing training regimens have different purposes and uses.

Comment author: [deleted] 12 May 2015 08:30:24AM 2 points [-]

I think virtue ethics is sufficiently edgy, new, different these days to be interesting. Go on.

Comment author: IlyaShpitser 12 May 2015 09:41:06AM 8 points [-]

new

I agree, scholarship is a problem.

Comment author: [deleted] 12 May 2015 11:33:42AM 3 points [-]

Okay, ancient enough, but fell into disuse around the Enlightenment, was hardly considered 100-120 years ago, returned among academic philosophers like Philippa Foot, Catholics like MacIntyre tryed to keep it alive, and it is only roughly about now that it is something slowly considered again by the hip young atheist literati classes for whom karma is merely a metaphor and do not literally believe in bad deeds putting a stain on the soul, so in this sense it is only a newly fashionable thing again.

Comment author: pianoforte611 12 May 2015 01:43:48AM 4 points [-]

When you say virtue ethics, it sounds like you are describing consequentialism implemented on human software.

If we're talking about the philosopher's virtue ethics, this question should clarify: Are virtues virtuous because they lead to moral behavior? Or is behavior moral because it cultivates virtue?

The first is just applied consequentialism. The second is the philosopher's virtue ethics.

Comment author: Dahlen 12 May 2015 03:02:11AM 1 point [-]

The thing is... that's really beyond the scope of what I care to argue about. I understand the difference, but it's so small as to not be worth the typing time. It's precisely the kind of splitting hairs I don't want to go into.

The theme that would get treated is morality, not ethics. It kind of starts off assuming that it is self-evident why good is good, and that human beings do not hold wildly divergent morals or have wildly different internal states in the same situation. Mostly. Sample topics that I'm likely to touch on are: rationality as wisdom; the self-perception of a humble person and how that may be an improvement from the baseline; the intent with which one enters an interaction; a call towards being more understanding to others; respect and disrespect; how to deflect (and why to avoid making) arguments in bad faith; malicious dispositions, and more. Lots of things relevant to community maintenance.

These essays aren't yet written, so perhaps that's why it all sounds (and is) so chaotic. There may be more topics which conflict more obviously with utilitarianism, especially if there's a large number of individuals concerned. As for conflicts with consequentialism, they're less likely, but still probable.

Comment author: pianoforte611 12 May 2015 07:26:34PM 1 point [-]

If you don't want to talk about the difference then I respect that, and I wasn't suggesting that you do. If anything I would suggest avoiding the term "virtue ethics" entirely and instead talking about virtue which is more general and a component of most moral systems.

I disagree that it is splitting hairs though or a small difference. It makes a large whether or not you wish to cultivate virtue for its own sake (regardless or independent of consequence), or because it helps you achieve other goals. The latter makes fewer assumptions about the goals of your reader.

Comment author: ChristianKl 12 May 2015 01:37:48AM 5 points [-]

If it helps you, the 2014 census gave for moral beliefs:

Moral Views Accept/lean towards consequentialism: 901, 60.0% Accept/lean towards deontology: 50, 3.3% Accept/lean towards natural law: 48, 3.2% Accept/lean towards virtue ethics: 150, 10.0% Accept/lean towards contractualism: 79, 5.3% Other/no answer: 239, 15.9%

Meta-ethics Constructivism: 474, 31.5% Error theory: 60, 4.0% Non-cognitivism: 129, 8.6% Subjectivism: 324, 21.6% Substantive realism: 209, 13.9%

In general I don't think there are foundational ideas on LW that shouldn't be questioned. Any idea is up for investigation provided the case is well argued.

Comment author: falenas108 12 May 2015 04:13:29AM 6 points [-]

In general I don't think there are foundational ideas on LW that shouldn't be questioned. Any idea is up for investigation provided the case is well argued.

But there are certain ideas that will be downvoted and dismissed because people feel like they aren't useful to be talking about, like if God exists. I think OP was asking if it was a topic that fell under this category.

Comment author: ChristianKl 12 May 2015 11:32:30AM 4 points [-]

But there are certain ideas that will be downvoted and dismissed because people feel like they aren't useful to be talking about, like if God exists.

The problem with "does God exist" isn't about the fact that LW is atheist. It's that it's hard to say interesting things about the subject and provide a well argued case.

I don't expect to learn something new when I read another post about whether or not God exists. If someone knows the subject well enough to tell me something new, then there no problem with them writing a post to communicate that insight.

Comment author: Vaniver 12 May 2015 07:02:42PM 0 points [-]

Also, have you read this post? The virtue tag only points at it and one other, but searching will likely find more.

Comment author: Epictetus 11 May 2015 06:56:34PM 2 points [-]

For some time I've been thinking about just how much of our understanding of the world is tied up in stories and narratives.

Let's take gravity. Even children playing with balls have a good idea of where a ball is going to land after they throw it. They don't know anything about spacetime curvature or Newton's laws. Instead, they amass a lot of data about the behavior of previously-thrown balls and from this they can predict where a newly-thrown ball will land. With experience, this does not even require conscious thought--a skilled ball-player is already moving into position by the time he's consciously aware of what's happening.

You can do the same thing with computers. Once you have enough raw data, you can tabulate it and use various methods to make predictions. These can range from simple interpolation to more complicated statistical modeling. The point is that you don't need any deeper understanding of the underlying phenomena to make it work. You can get good results as long as the phenomena are nice enough and the initial conditions aren't far removed from the data you used to construct the model. Going back to balls, you'll do fine predicting how a ball thrown by a human will behave, but the methods will probably fail if you shot a ball out of a cannon.

So far, so good. You can treat actual phenomena as a black box and make predictions based only on initial conditions, and this works for everyday life. Yet, we feel a drive to explain things. We like to come up with stories. Most of these are silly, but largely harmless. Sometimes, though, we happen upon a useful story or analogy. These stories transcend the role of explanations and enable us to make predictions outside of our accumulated data. Aristotle's gravity didn't have a detrimental effect on the engineering of the day due to the aforementioned use of experience, but Newton's gravity let us push things so much further.

Modern physics is full of these stories which are wrong, but make for good enough analogies to be useful. Take continuum mechanics. We know matter is made of atoms and molecules, but we sometimes assume it's continuous. Then we take this continuous matter and assume it's made up of tiny boxes (finite elements), each subjected to a constant force. We look at how the force acts on these tiny boxes and add up all the contributions to get an idea of what happens to a large object. Take limits, neglect higher-order terms, and you've got yourself a nice set of equations. In this case, a good story can be more useful than the truth.

The hard part is coming up with a good narrative framework. Working out the details is a lot easier once you have a mental picture of where you are and where you're going. It's easy to come up with a story that doesn't add anything--some ad-hoc tale to satisfy your desire for an explanation and let you go on doing what you were doing with your black-box model.

Sorry if this is a bit disjointed. I'm still trying to straighten it out in my own mind.

Comment author: IlyaShpitser 11 May 2015 09:34:08PM 1 point [-]

You are talking about prediction vs causality. I agree, we understand via causality, and causality lets us take data beyond what is actually observed into the realm of the hypothetical. Good post.

Comment author: Lumifer 11 May 2015 07:01:38PM 6 points [-]

I think what would be useful is to distinguish a story (a typically linear narrative) and a model (a known-to-be-simplified map of some piece of reality). They are sufficiently different and often serve different goals. In particular, stories are rarely quantitative and models usually are.

Comment author: Epictetus 11 May 2015 07:27:24PM 0 points [-]

I like to think about how the two complement each other. You can build a model out of a mass of data, but extrapolation outside the data is tricky business. You can also start with a qualitative description of the phenomena involved and work out the details. A lot of models start off by making some assumptions and figuring out the consequences.

Example: you can figure out gas laws by taking lots of measurements, or you can start with the assumption that gases are made of molecules that bounce around and go from there.

Comment author: Lumifer 11 May 2015 08:33:12PM 4 points [-]

We might be understanding the word "story" differently.

To me a "story" is a narrative (a linear sequence of words/sentences/paragraphs/etc.) with the general aim of convincing your System 1. It must be simple enough for the System 1 and must be able to be internalized to become effective. There are no calculations in stories and they generally latch onto some basic hardwired human instincts.

For example, a simple and successful story is "There are tiny organisms called germs which cause disease. Wash your hands and generally keep clean to avoid disease". No numbers, plugs into the purity/disgust template, mostly works.

The three laws of Newton are not a story to me, to pick a counter-example. Nor is the premise that gas consists of identical independent molecules in chaotic motion -- that's an assumption which underlies a particular class of models.

Models, as opposed to stories, are usually "boxes" in the sense that you can throw some inputs into the hopper, turn the crank, and get some outputs from the chute. They don't have to be intuitive or even understandable (in which case the box is black), they just have to output correct predictions. The Newton's laws, for example, make correct predictions (within their sphere of applicability and to a limited degree of precision), but we still have no idea how gravity really works.

Comment author: Epictetus 12 May 2015 01:16:25AM 1 point [-]

I was using "story" in a much more general sense. Perhaps I should have chosen a different word. I saw a story as some bit of exposition devised to explain a process. In that sense, I would view the kinetic theory of gases as a story. A gas has pressure because all these tiny particles are bumping into the walls of its container. Temperature is related to the average kinetic energy of the particles. The point here is that we can't see these particles, nor can we directly measure their state.

Consider, in contrast, the presentation in Fermi's introductory Thermodynamics book. He eschewed an explanation of what exactly was happening internally and derived his main results from macroscopic behavior. Temperature was defined initially as that which a gas thermometer measures, and later on he developed a thermodynamic definition based on the behavior of reversible heat engines. This sort of approach treats the inner workings of a gas as unknown and only uses that which we can directly observe through instrumental readings.

I guess what I really want to distinguish are black boxes from our attempts to guess what's in the box. The latter is what I tried to encapsulate by "story".

Comment author: Romashka 11 May 2015 06:34:15PM 5 points [-]

(Random) by analogy with what Quirrellmort led people to believe about his imminent death, it would be cool to read a fic, 6th year AU, in which Dumbledore teaches defence...

Comment author: knb 14 May 2015 08:51:50PM 2 points [-]

What does 6th year AU mean?

Comment author: zedzed 15 May 2015 04:37:46AM 2 points [-]

6th year = book 6 (6th year at Hogwarts) = Half-Blood Prince.

AU = Alternate Universe.

Comment author: NancyLebovitz 11 May 2015 04:27:29PM *  7 points [-]

Transhumanism in the real world

Rugby players who get a bottle opener to replace a missing tooth.

Comment author: Lumifer 11 May 2015 04:39:11PM -1 points [-]

That's no more transhumanism than this.

Comment author: Ishaan 13 May 2015 08:48:52PM *  0 points [-]

NancyLebovitz didn't imply the rugby player was showing signs of ideological transhumanism - only that they're doing something transhumanist. Transhumanists don't have the monopoly on self modification. It's the same sense that Christians refer to kind acts as Christian and bad acts as un-Christian.

Transhumanists would claim the first intentional use of fire and writing and all that as transhuman-ish things. (And yes, I would consider self decoration to be a transhumanish thing too. Step into the paleolithic - what's the very first thing you notice which is different about the humans? They have clothes and strings and beads and tattoos, which turn out to have pretty complex social functions. Adam and Eve and all that, it's literally the stuff of myth.)

Comment author: Lumifer 13 May 2015 08:54:54PM 1 point [-]

So, using tools. Traditionally, tool-using is said to be be what distinguishes humans from apes. That makes it just human, not transhuman.

Comment author: Ishaan 13 May 2015 09:00:23PM -1 points [-]

Yes, I bite that bullet: I think "you aught to use tools to do things better" counts as foundational principle of transhuman ideology. It's supposed to be fundamentally about being human.

Comment author: Lumifer 13 May 2015 09:26:38PM 1 point [-]

Well, me might just be having a terminology difference.

My understanding of "transhuman" involves being more than just human. Picking up a tool, even a sophisticated tool, doesn't qualify. And "more" implies that you standard garden-variety human doesn't qualify either.

I'm not claiming there is an easily discernible bright line, but just as contact lenses don't make you a cyborg, a weirdly shaped metal tooth does not make you a transhuman.

Comment author: Ishaan 13 May 2015 11:58:11PM *  0 points [-]

But that's because everyone uses glasses, as a matter of course - it's the status quo now. The person who thought "well, and why should we have to walk around squinting all the time when we can just wear these weird contraption on our heads", at a time when people might look at you funny having wearing glass on your face, I think that's pretty transhuman. As is the guy who said "Let's take it further, and put the refractive material directly on our eyeball" back when people would have looked at you real funny if you suggested they put plastic in their eyes are you crazy that sounds so uncomfortable.

Now of course, it's easy to look at these things and say "meh".

Edit: If you look at the history of contact lenses, though, what actually happened is less people saying "let's improve" and more people saying "I wonder how the eye works" and doing weird experiments that probably seemed pointless at the time. Something of a case study against the "basic research isn't useful" argument, I think, not that there are many who espouse that here.

Comment author: drethelin 11 May 2015 06:59:09PM 5 points [-]

False! It's adding functionality rather than just a cosmetic change.

Comment author: Lumifer 11 May 2015 07:03:42PM *  1 point [-]

Cosmetic changes can be highly functional. Ask any girl :-)

On a slightly more serious note, I tend to think of tranhumanist modifications as ones which confer abilities that unenhanced humans do not have. Opening beer bottles isn't one of them.

Comment author: [deleted] 13 May 2015 07:57:57PM 1 point [-]

But nigh-effortless verbatim memorization is, so if you carry around a pen and a pad of paper...

Comment author: Lumifer 13 May 2015 08:14:53PM 0 points [-]

"nigh-effortless"

"a pen and a pad of paper"

Ahem.

Comment author: Epictetus 12 May 2015 01:27:20AM *  4 points [-]

Having been in a group of drunk people who found that they had no bottle opener, and having seen what bizarre ideas they concoct to get the bottles open, I'd say a bottle opener in one's tooth merits the status of transhumanist modification.

Comment author: Lumifer 12 May 2015 02:52:08PM -1 points [-]

You can open a beer bottle with your natural teeth easily enough.

These people lacked in knowledge, not in tools :-P

Comment author: NancyLebovitz 12 May 2015 07:38:36PM 5 points [-]

However, you can damage a tooth by using it to open bottles.

Comment author: [deleted] 12 May 2015 07:42:50AM 1 point [-]

There was a saying in my youth: "There is no item that is not a beer opener." There was a bit of a competition for creative moves (drinking beer was considered a high status adult move for teenagers, opening them in creative ways even more). Keys. Lighters. Doors, the part where the "tongue" goes in on the frame, not sure the English term. Edges of tables or edges of anything. Using two bottles, locking the caps to each other and pulling apart. I still consider it the coolest manly way to open a beer when you sit at a fairly invulnerable e.g. stone table to just put the cap against the edge and hit it. Another 101 ways.

Comment author: Nornagest 13 May 2015 11:30:53PM *  0 points [-]

Doors, the part where the "tongue" goes in on the frame, not sure the English term.

Strike plate?

Comment author: TezlaKoil 12 May 2015 12:23:59AM *  4 points [-]

Would you consider a Wikipedia brain implant to be a transhumanist modification? After all, ordinary humans can query Wikipedia too!

Comment author: Lumifer 12 May 2015 02:48:26PM 1 point [-]

Would you consider a Wikipedia brain implant to be a transhumanist modification?

That's a weird way of putting it. Would I consider an implant which consists of a large chunk of memory with some processing and an efficient neural interface to be transhumanist? Yes, of course. It will give a lot of useful abilities and just filling it with Wikipedia looks like a waste of potential.

I don't think trivializing transhumanism to minor cosmetics is a useful approach. Artificial nails make better screwdrivers than natural nails, so is that also a transhumanist modification?

Comment author: RowanE 11 May 2015 06:49:43PM 4 points [-]

Well, you know what they say about "one man's modus ponens".

Comment author: Lumifer 11 May 2015 06:56:27PM *  2 points [-]

Here is your first transhumanist then, from pre-Columbian Maya...

Comment author: RowanE 12 May 2015 04:41:27PM 1 point [-]

I think the attitude toward the modifications is a relevant factor as well - wanting to be "more than human" in some respect, even if only a trivial respect such as "more awesome-looking than a regular human" or "more able to open beer bottles than a regular human" - but given that, yeah I'd be totally on board with considering some pre-Columbian Maya or other stone-age person "the first transhumanist".

Comment author: Lumifer 12 May 2015 04:53:30PM *  1 point [-]

I see a smooth transition into tool-using, then. Picking up a stick certainly gives you more capabilities compared to a stick-less hominid, and probably makes you much more awesome as well.

Comment author: John_Maxwell_IV 11 May 2015 02:08:07PM *  3 points [-]

So there was recently an advance related to chips for running neural networks. I'm having a hard time figuring out if we should be happy or sad. I'm not sure if this qualifies as a "computing power" advance or a "cell modeling" one.

Comment author: ChristianKl 11 May 2015 02:33:51PM 0 points [-]

Neural networks chips aren't neurons. Neurons are much more complex than nodes in artificial neural networks.

Comment author: jacob_cannell 12 May 2015 04:37:16AM *  1 point [-]

Technically true but also irrelevant. At the physical level, a modern digital transistor based computer running an ANN simulation is also vastly more complex than the node-level ANN model.

In terms of simulation complexity, a modern GPU is actually more complex than the brain. It would take at most on the order of 10^17 op/s second to simulate a brain (10^14 synapses @ 10^3 hz), but it takes more than 10^18 op/s second to simulate a GPU (10^9 transistors @ 10^9 hz).

Simulating a brain at any detail level beyond its actual computational power is pointless for AI - the ANN level is the exactly correct level of abstraction for actual performance.

Comment author: ChristianKl 12 May 2015 10:30:25AM 1 point [-]

Technically true but also irrelevant. At the physical level, a modern digital transistor based computer running an ANN simulation is also vastly more complex than the node-level ANN model.

ANN are no neurons. We can't accurately simulate even what a single neuron does. Neurons can express proteins when specific hormones are in their environment. The functioning of roughly a third of the human genome is unknown. Increasing or decreasing the amount of channels for various substances in the cell membrane takes proteins. That's part of long term plasticity.

It would take at most on the order of 10^17 op/s second to simulate a brain (10^14 synapses @ 10^3 hz),

That simulation completely ignores neurotransmitters floating around in the brain and many other factors. You can simulate a ANN at one op/synase but a simulation of a real brain is very incomplete at that level.

Comment author: jacob_cannell 13 May 2015 12:35:51AM -1 points [-]

ANN are no neurons. We can't accurately simulate even what a single neuron does.

Everything depends on your assumed simulation scale and accuracy. If you want to be pedantic, you could say we can't even simulate transistors, because clearly our simulations of transistors are not accurate down to the quantum level.

However, the physics of computation allow us to estimate the approximate level of computational scale separation that any conventional (irreversible) physical computer must have to functional correctly (signal reliably in a noisy environment).

The Lanauder limits on switching energies is one bound, but most of the energy (in brains or modern computers) goes to wire transmission energy, and one can derive bounds on signal propagation energy in the vicinity of ~1pJ / bit / mm for reliable signaling. From this we can then plug in the average interconnect distance between synapses and neurons (both directions) and you get a maximum computation rate on the order of 10^15 ops or so, probably closer to 10^13 low precision ops. Deriving all that is well beyond the scope of a little comment.

Comment author: ChristianKl 13 May 2015 10:02:52AM -1 points [-]

The Lanauder limits on switching energies is one bound, but most of the energy (in brains or modern computers) goes to wire transmission energy, and one can derive bounds on signal propagation energy in the vicinity of ~1pJ / bit / mm for reliable signaling.

The energy count for signal transmission doesn't include changing the amount of ion channels a neuron has. You might model short term plasticity but you don't get long term plasticity.

You also don't model how hormones and other neurotransmitter float around in the brain. An ANN deals only with the electric signal transmission misses essential parts of how the brain works. That doesn't make it bad for the purposes of being an ANN but it's lacking as a model of the brain.

Comment author: jacob_cannell 13 May 2015 04:30:55PM 0 points [-]

Sure, all of that is true, but of the brain's 10 watt budget, more than half is spent on electric signaling and computation, so all the other stuff you mention at most increases the intrinsic simulation complexity by a factor of 2.

Comment author: ChristianKl 13 May 2015 05:50:37PM 0 points [-]

Are you aware of the complexity of folding of a single protein? It might not take much energy but it's very complex.

If you have 1000 different types of proteins swimming around in a neurons that interact with each other I don't think you get that by adding a factor of two.

Comment author: [deleted] 12 May 2015 10:48:58AM 0 points [-]

We can't accurately simulate even what a single neuron does.

This blew my mind a bit. So why the heck are researchers trying to train neural nets when the nodes of those nets are clearly subpar?

Comment author: jacob_cannell 13 May 2015 12:39:43AM *  1 point [-]

Actually the exact opposite is true - ANN neurons and synapses are more powerful per neuron per synapse than their biological equivalents. ANN neurons signal and compute with high precision real numbers with 16 or 32 bits of precision, rather than 1 bit binary pulses with low precision analog summation.

The difference depends entirely on the problem, and the ideal strategy probably involves a complex heterogeneous mix of units of varying precision (which you see in the synaptic distribution in the cortex, btw), but in general with high precision neurons/synapses you need less units to implement the same circuit.

Also, I should mention that some biological neural circuits implement temporal coding (as in the hippocampus), which allows a neuron to send somewhat higher precision signals (on the order of 5 to 8 bits per spike or so). This has other tradeoffs though, so it isn't worth it in all cases.

Brains are more powerful than current ANNs because current ANNs are incredibly small. All of the recent success in deep learning where ANNs are suddenly dominating everywhere was enabled by using GPUs to train ANNs in the range of 1 to 10 million neurons and 1 to 10 billion synapses - which is basically insect to lizard brain size range. (we aren't even up to mouse sized ANNs yet)

That is still 3 to 4 orders of magnitude smaller than the human brain - we have a long ways to go still in terms of performance. Thankfully ANN performance is more than doubling every year (combined hardware and software increase).

Comment author: ChristianKl 12 May 2015 11:35:49AM 0 points [-]

The fact that they are subpar doesn't mean that you can learn nothing from ANNs. It also doesn't mean that ANNs can't do a variety of tasks in machine learning with them.

Comment author: Houshalter 14 May 2015 05:56:16AM *  0 points [-]

I doubt it will help scientists reverse engineer the function of the brain any faster. However it could potentially be very useful for helping AI researchers develop artificial neural networks.

ANNs aren't really tied to neuroscience research, and they probably won't help with emulations. But they are the current leading approach to AI, and increased computing power would significantly help AI research, as it has in the past.

Comment author: chaosmage 11 May 2015 10:28:03AM 2 points [-]

I have nearly finished the second of the Seven Secular Sermons, which is going to premiere at the European Less Wrong Community Weekend in Berlin in June. For final polishing, I'm looking for constuctive feedback especially from native speakers of English. If you'd like to help out, PM me for the current draft.

Comment author: [deleted] 11 May 2015 12:29:01PM 1 point [-]

Offtopic, but I like your theory of depression have you ever written about it elsewhere longer? Or any recommended online readings?

Comment author: chaosmage 11 May 2015 02:01:52PM *  1 point [-]

Glad you like it. No I haven't written about it at more length than in that post, and it is entirely my own speculation, based only on the phenomenology of clinical depression and the rank theory I referenced.

I don't have any reading on depression to recommend that is anywhere near as good as SSC. And that's despite my working as a research associate at a depression-focused nonprofit.

Comment author: NancyLebovitz 11 May 2015 02:45:35PM 2 points [-]

The status possibility doesn't explain post-partum depression.

Comment author: Elo 13 May 2015 12:57:28AM 0 points [-]

I would also like to remain updated about this theory and subsequent writings. (repeat to ping you too)

Comment author: Elo 13 May 2015 12:57:13AM 0 points [-]

I would also like to remain updated about this theory and subsequent writings.

Comment author: Toggle 11 May 2015 02:35:54AM 9 points [-]

It looks like AI is overtaking Arimaa, which is notable because Arimaa was created specifically as a challenge to AI. Congratulations to the programmer, David Wu.

Comment author: Kindly 12 May 2015 03:21:36PM 1 point [-]

On the subject of Arimaa, I've noted a general feeling of "This game is hard for computers to play -- and that makes it a much better game!"

Progress of AI research aside, why should I care if I choose a game in which the top computer beats the top human, or one in which the top human beats the top computer? (Presumably both the top human and the top computer can beat me, in either case.)

Is it that in go, you can aspire (unrealistically, perhaps) to be the top player in the world, while in chess, the highest you can ever go is a top human that will still be defeated by computers?

Or is it that chess, which computers are good at, feels like a solved problem, while go still feels mysterious and exciting? Not that we've solved either game in the sense of having solved tic-tac-toe or checkers. And I don't think we should care too much about having solved checkers either, for the purposes of actually playing the game.

Comment author: jacob_cannell 11 May 2015 05:49:13AM 1 point [-]

I hadn't heard of Armaa before, but based on about 5 minutes worth of reading about the game, I don't understand how it is significantly more suited to natural reasoning that chess. It inherits many of chess's general features that make serial planning more effective than value function knowledge - thus favoring fast thinkers over slow deep thinkers. Go is much more of a natural reasoning game.

Comment author: ShardPhoenix 12 May 2015 09:24:38AM 0 points [-]

From the sound of it, the AI works more or less like chess AI - search with a hand-tuned evaluation function.

Comment author: Error 11 May 2015 01:29:47AM 3 points [-]

Just posted this in the previous open thread; reposting here: Has anyone here used fancyhands.com or a similar personal-assistant service? If so, what was your experience like?

(context: I have anxiety issues making phone calls to strangers and certain other ugh fields, and am thinking I may be better off paying someone else to take care of such things rather than trying to bull through the ugh fields.)

Comment author: [deleted] 16 May 2015 08:25:13PM 1 point [-]

Even with my own business, I found it incredibly hard to find lots of tasks which I could hand off to an assistant from whom I didn't know what their strengths were, and couldn't train.

Comment author: Archelon 17 May 2015 11:50:44PM 0 points [-]

According to this article, a traumatic brain injury turned a furniture salesman into a mathematician. (Not without side effects, but still.)

There is a bit of conventional wisdom in evolutionary biology that drastic improvements in efficacy are not available through trivial modifications (and that nontrivial modifications which are random are not improvements). This is an example of the principle that evolution is supposed to have already 'harvested' any 'low-hanging fruit'. Although I don't think much of this type of website (note the lack of external links), the story seems to be based in reality; it is thus one of the most surprising things I have ever heard. And, oddly, heartening as well---insofar as it suggests both a potential shortening of the timescale for human intelligence augmentation and the possibility that such augmentation may be relatively more accessible (than I previously thought) by comparison to computer-based artificial intelligence developments.

Comment author: the-citizen 17 May 2015 06:52:33AM *  0 points [-]

Suffering and AIs

Disclaimer - Under utilitarianism suffering is an intrinsically bad thing. While I am not a utilitarian, many people are and I will treat it as true for this post because it is the easiest approach for this issue. Also, apologies if others have already discussed this idea, which seems quite possible

One future moral issue is that AIs may be created for the purpose of doing things that are unpleasant for humans to do. Let's say an AI is designed with the ability to have pain, fear, hope and pleasure of some kind. It might be reasonable to expect in such cases the unpleasant tasks might result in some form of suffering. Added to this problem is the fact that a finite lifespan and an approaching termination/shutdown might cause fear, another form of suffering. Taking steps to shut down an AI would also then become morally unacceptable, even though they perform an activity that might be useless or harmful. Because of this, we might face a situation where we cannot shutdown AIs even when there is good reason to.

Basically, if suffering AIs were some day extremely common, we would be introducing a massive amount of suffering into the world, which under utilitarianism is unacceptable. Even assuming some pleasure is created, we might search for ways to create that pleasure without creating the pain.

If so, would it make sense to adopt a principle of AI design that says AIs should be designed so it (1) does not suffer or feel pain (2) should not fear death/shutdown (eg. views own finite life as acceptable). This would minimise suffering (potentially you could also attempt to maximise happiness).

Potential issues with this: (1) Suffering might be in some way relative, so that a neutral lack of pleasure/happiness might become "suffering". (2) Pain/suffering might be useful to create a robot with high utility, and thus some people may reject this principle. (3) I am troubled by this utilitarian approach I have used here as it seems to justify tiliing the universe with machines whose only purpose and activity is to be permanently happy for no reason. (4) Also... killer robots with no pain or fear of death :-P

Comment author: hairyfigment 17 May 2015 03:46:48PM 0 points [-]
Comment author: the-citizen 19 May 2015 07:40:49AM 0 points [-]

That seems like an interesting article, though I think it is focused on the issue of free-will and morality which is not my focus.

Comment author: Fatcat 14 May 2015 08:36:31PM 0 points [-]

Article on transhumanism (intro bit perfunctory) - but has interviews with Anders Sandberg and Steve Fuller on implications of transhumanist thought. Quite interesting in parts - http://www.theworldweekly.com/reader/i/humanity-20/3757

Same journalist did reasonable job of introducing AI dangers last month - http://www.theworldweekly.com/reader/i/irresistible-rise-ai/3379

Comment author: [deleted] 13 May 2015 08:02:59AM 0 points [-]

Asking for article recommendations: difference between intelligence vs. intellectualism, how a superintelligence is not the same as a superintellectual.