Welcome to Less Wrong! (5th thread, March 2013)

27 Post author: orthonormal 01 April 2013 04:19PM
If you've recently joined the Less Wrong community, please leave a comment here and introduce yourself. We'd love to know who you are, what you're doing, what you value, how you came to identify as a rationalist or how you found us. You can skip right to that if you like; the rest of this post consists of a few things you might find helpful. More can be found at the FAQ.

(This is the fifth incarnation of the welcome thread; once a post gets over 500 comments, it stops showing them all by default, so we make a new one. Besides, a new post is a good perennial way to encourage newcomers and lurkers to introduce themselves.)

A few notes about the site mechanics

Less Wrong comments are threaded for easy following of multiple conversations. To respond to any comment, click the "Reply" link at the bottom of that comment's box. Within the comment box, links and formatting are achieved via Markdown syntax (you can click the "Help" link below the text box to bring up a primer).

You may have noticed that all the posts and comments on this site have buttons to vote them up or down, and all the users have "karma" scores which come from the sum of all their comments and posts. This immediate easy feedback mechanism helps keep arguments from turning into flamewars and helps make the best posts more visible; it's part of what makes discussions on Less Wrong look different from those anywhere else on the Internet.

However, it can feel really irritating to get downvoted, especially if one doesn't know why. It happens to all of us sometimes, and it's perfectly acceptable to ask for an explanation. (Sometimes it's the unwritten LW etiquette; we have different norms than other forums.) Take note when you're downvoted a lot on one topic, as it often means that several members of the community think you're missing an important point or making a mistake in reasoning— not just that they disagree with you! If you have any questions about karma or voting, please feel free to ask here.

Replies to your comments across the site, plus private messages from other users, will show up in your inbox. You can reach it via the little mail icon beneath your karma score on the upper right of most pages. When you have a new reply or message, it glows red. You can also click on any user's name to view all of their comments and posts.

It's definitely worth your time commenting on old posts; veteran users look through the recent comments thread quite often (there's a separate recent comments thread for the Discussion section, for whatever reason), and a conversation begun anywhere will pick up contributors that way.  There's also a succession of open comment threads for discussion of anything remotely related to rationality.

Discussions on Less Wrong tend to end differently than in most other forums; a surprising number end when one participant changes their mind, or when multiple people clarify their views enough and reach agreement. More commonly, though, people will just stop when they've better identified their deeper disagreements, or simply "tap out" of a discussion that's stopped being productive. (Seriously, you can just write "I'm tapping out of this thread.") This is absolutely OK, and it's one good way to avoid the flamewars that plague many sites.

EXTRA FEATURES:
There's actually more than meets the eye here: look near the top of the page for the "WIKI", "DISCUSSION" and "SEQUENCES" links.
LW WIKI: This is our attempt to make searching by topic feasible, as well as to store information like common abbreviations and idioms. It's a good place to look if someone's speaking Greek to you.
LW DISCUSSION: This is a forum just like the top-level one, with two key differences: in the top-level forum, posts require the author to have 20 karma in order to publish, and any upvotes or downvotes on the post are multiplied by 10. Thus there's a lot more informal dialogue in the Discussion section, including some of the more fun conversations here.
SEQUENCES: A huge corpus of material mostly written by Eliezer Yudkowsky in his days of blogging at Overcoming Bias, before Less Wrong was started. Much of the discussion here will casually depend on or refer to ideas brought up in those posts, so reading them can really help with present discussions. Besides which, they're pretty engrossing in my opinion.

A few notes about the community

If you've come to Less Wrong to  discuss a particular topic, this thread would be a great place to start the conversation. By commenting here, and checking the responses, you'll probably get a good read on what, if anything, has already been said here on that topic, what's widely understood and what you might still need to take some time explaining.

If your welcome comment starts a huge discussion, then please move to the next step and create a LW Discussion post to continue the conversation; we can fit many more welcomes onto each thread if fewer of them sprout 400+ comments. (To do this: click "Create new article" in the upper right corner next to your username, then write the article, then at the bottom take the menu "Post to" and change it from "Drafts" to "Less Wrong Discussion". Then click "Submit". When you edit a published post, clicking "Save and continue" does correctly update the post.)

If you want to write a post about a LW-relevant topic, awesome! I highly recommend you submit your first post to Less Wrong Discussion; don't worry, you can later promote it from there to the main page if it's well-received. (It's much better to get some feedback before every vote counts for 10 karma- honestly, you don't know what you don't know about the community norms here.)

If you'd like to connect with other LWers in real life, we have  meetups  in various parts of the world. Check the wiki page for places with regular meetups, or the upcoming (irregular) meetups page. There's also a Facebook group. If you have your own blog or other online presence, please feel free to link it.

If English is not your first language, don't let that make you afraid to post or comment. You can get English help on Discussion- or Main-level posts by sending a PM to one of the following users (use the "send message" link on the upper right of their user page). Either put the text of the post in the PM, or just say that you'd like English help and you'll get a response with an email address.
* Normal_Anomaly
* Randaly
* shokwave
* Barry Cotter

A note for theists: you will find the Less Wrong community to be predominantly atheist, though not completely so, and most of us are genuinely respectful of religious people who keep the usual community norms. It's worth saying that we might think religion is off-topic in some places where you think it's on-topic, so be thoughtful about where and how you start explicitly talking about it; some of us are happy to talk about religion, some of us aren't interested. Bear in mind that many of us really, truly have given full consideration to theistic claims and found them to be false, so starting with the most common arguments is pretty likely just to annoy people. Anyhow, it's absolutely OK to mention that you're religious in your welcome post and to invite a discussion there.

A list of some posts that are pretty awesome

I recommend the major sequences to everybody, but I realize how daunting they look at first. So for purposes of immediate gratification, the following posts are particularly interesting/illuminating/provocative and don't require any previous reading:

More suggestions are welcome! Or just check out the top-rated posts from the history of Less Wrong. Most posts at +50 or more are well worth your time.

Welcome to Less Wrong, and we look forward to hearing from you throughout the site!

Note from orthonormal: MBlume and other contributors wrote the original version of this welcome post, and I've edited it a fair bit. If there's anything I should add or update on this post (especially broken links), please send me a private message—I may not notice a comment on the post. Finally, once this gets past 500 comments, anyone is welcome to copy and edit this intro to start the next welcome thread.

Comments (1750)

Comment author: AABoyles 26 September 2014 03:14:57PM *  4 points [-]

Hi Everyone! I'm AABoyles (that's true most places on the internet besides LW).

I first found LW when a colleague mentioned That Alien Message over lunch. I said something to the effect of "That sounds like an Arthur C. Clarke short story. Who is the author?" "Eliezer Yudkowsky," He said, and sent me the link. I read it, and promptly forgot about it. Fast forward a year, and another friend posts the link to HPMOR on Facebook. The author's name sounded very familiar. I read it voraciously. I subscribed to the Main RSS feed and lurked for a year.

I joined the community last month because I wanted to respond to a specific discussion, but I've been having a lot of fun since I got here. I'm interested in finding ways to achieve the greatest good (read: reducing the number of lost Disability Adjusted Life Years), including Effective Altruism and Global Catastrophic Risk Reduction.

Comment author: lirene 10 August 2014 01:48:53PM 3 points [-]

Hello community.

I've been aware of LW for a while, reading individual posts linked in programmer/engineering hangouts now and then, and I independently came across HPMOR in search of good fanfiction. But the decision to un-lurk myself came after I attended a CFAR workshop (a major positive life change) and realized that I want to keep being engaged with the community.

I'm very interested in anti-aging research (both from the effective altruism point of view, and because I find the topic really exciting and fascinating) and want to learn about it in as much depth as time permits. So far I would come across science articles about single related discoveries in specialized fields (molecular biology, brain science, ... ) but I haven't found a good resource (book, coursera course, whatever) where I can learn the necessary medicine/biology background and how it all comes together in the current state of the art (I'm thinking of something similar to all the remarkable physics books we have on the market). Any pointers are appreciated.

Comment author: shelikavoid 25 July 2014 05:57:44PM 3 points [-]

Hi I'm N. Currently a systems engineer. Lurked for sometime and finally decided to create an account. I am interested in mathematics and computer science and typography. Fonts can give me happiness or drive me crazy.

I am currently in SoCal.

Comment author: MelbourneLW 23 July 2014 03:59:10AM 4 points [-]

This account is used by a VA to post events for the Melbourne Meetup group. Comment is to accrue 2 karma to allow posting.

Comment author: more_wrong 26 May 2014 04:59:23PM 3 points [-]

I chose more_wrong as a name because I'm in disagreement with a lot of the lesswrong posters about what constitutes a reasonable model of the world. Presumably my opinions are more wrong than opinions that are lesswrong, hence the name :)

My rationalist origin story would have a series of watershed events but as far as I can tell, I never had any core beliefs to discard to become rational, because I never had any core beliefs at all. Do not have a use for them, never picked them up.

As far as identifying myself as an aspiring rationalist, the main events that come to mind would be: 1. Devouring as a child anything by Isaac Asimov that I could get my hands on. In case you are not familiar with the bulk of his work, most of it is scientific and historical exposition, not his more famous science fiction; see especially his essays for rationalist material.

  1. Working on questions in physics like "Why do we call two regions of spacetime close to each other?", that is, delving into foundational physics.

  2. Learning about epistemology and historiography from my parents, a mathematician and a historian.

  3. Thinking about the thinking process itself. Note: Being afflicted with neurological and psychological conditions that shut down various parts of my mentality, notably severe intermittent aphasia, has given me a different perspective on the thinking process.

  4. Making some effort to learn about historical perspectives on what constitutes reason or rationality, and not assuming that the latest perspectives are necessarily the best.

    I could go on but that might be enough for an intro.

    My hope is to both learn how to reason more effectively and, if fortunate, make a contribution to the discussion group that helps us to learn the same as a community. mw

Comment author: Algernoq 23 May 2014 03:55:56AM 1 point [-]

Hello there! I really enjoyed HPMOR, because it expanded on some of my thoughts and made me feel less alone. I joined now to post a realization about Harry's (and my) personality. See my 1st post.

Comment author: tsarani 13 February 2014 09:46:00AM *  1 point [-]

Hi! I read some articles here a few years ago, decided they were good, and moved on. I think I am a pretty practical person, and I have some ways of deciding things that are utility based (and some that are not).

I would like to ask the community for some help with a couple of reading recommendations:

  1. how to leave for places (like work) on time and not get distracted;
  2. general time- and direction-sense improvements, should any reliably exist; and
  3. how to maintain composure when your body is making moods happen at you (dealing with becoming hangry).

Thanks very much, and I hope to be optimizing more things soon. It's nice to meet you!

Comment author: zoltanistvan 08 February 2014 08:13:49PM 1 point [-]

Hi, I'm reposting my introduction here from 2 days ago, as it was moved for some reason, perhaps accidentally. Anyway, hello, my name is Zoltan Istvan. I'm a transhumanist, futurist, journalist, and the author of the philosophical novel "The Transhumanist Wager." I've been checking out this site for some time, but decided to create an account a few days ago to become closer to the community. I thought I'd start by posting an essay I recently wrote, which sums up some of my ideas. Feel free to share it if you like, and I look forward to interacting here. Cheers.

"When Does Hindering Life Extension Science Become a Crime—or even Genocide?"

Every human being has both a minimum and a maximum amount of life hours left to live. If you add together the possible maximum life hours of every living person on the planet, you arrive at a special number: the optimum amount of time for our species to evolve, find happiness, and become the most that it can be. Many reasonable people feel we should attempt to achieve this maximum number of life hours for humankind. After all, very few people actually wish to prematurely die or wish for their fellow humans' premature deaths.

In a free and functioning democratic society, it's the duty of our leaders and government to implement laws and social strategies to maximize these life hours that we want to safeguard. Regardless of ideological, political, religious, or cultural beliefs, we expect our leaders and government to protect our lives and ensure the maximum length of our lifespans. Any other behavior cuts short the time human beings have left to live. Anything else becomes a crime of prematurely ending human lives. Anything else fits the common legal term we have for that type of reprehensible behavior: criminal manslaughter.

In 2001, former President George W. Bush restricted federal funding for stem cell research, one of the most promising fields of medicine in the 21st Century. Stem cells can be used to help fight disease and, therefore, can lengthen lives. Bush restricted the funding because his conservative religious beliefs—some stem cells came from aborted fetuses—conflicted with his fiduciary duty of helping millions of ailing, disease-stricken human beings. Much medical research in the United States relies heavily on government funding and the legal right to do the research. Ultimately, when a disapproving President limits public resources for a specific field of science, the research in that field slows down dramatically—even if that research would obviously lengthen and improve the lives of millions.

It's not just politicians that are prematurely ending our lives with what can be called "pro-death" policies and ideologies. In 2009, on a trip to Africa, Pope Benedict XVI told journalists that the epidemic of AIDS would be worsened by encouraging people to use condoms. More than 25 million people have died from AIDS since the first cases began being reported in the news in the early 1980s. In numerous studies, condoms have been shown to help stop the spread of HIV, the virus that causes AIDS. This makes condoms one of the simplest and most affordable life extension tools on the planet. Unfathomably, the billion-person strong Catholic Church actively supports the idea that condom usage is sinful, despite the fact that such a malicious policy has helped sicken and kill a staggering amount of innocent people.

Regrettably, in 2014, America continues to be permeated with an anti-life extension culture. Genetic engineering experiments in humans often have to pass numerous red-tape-laden government regulatory bodies in order to conduct any tests at all, especially at publically funded universities and research centers. Additionally, many states still ban human reproductive cloning, which could one day play a critical part in extending human life. The current US administration is also culpable. The White House is simply not doing enough to extend American lifespans. The US Government spends just 2% of the national budget on science and medical research, while their defense budget is over 20%, according to a 2011 US Office of Management Budget chart. Does President Obama not care about this fact, or is he unaware that not actively funding and supporting life extension research indeed shortens lives?

In my philosophical novel The Transhumanist Wager, there is a scene which takes place outside of a California courthouse where transhumanist activists are holding up a banner. The words inscribed on the banner sum up some eye-opening data: "By not actively funding life extension research, the amount of life hours the United States Government is stealing from its citizens is thousands of times more than all the American life hours lost in the Twin Towers tragedy, the AIDS epidemic, and the Vietnam War combined. Demand that your government federally fund transhuman research, nullify anti-science laws, and promote a life extension culture. The average human body can be made to live healthily and productively beyond age 150."

Some longevity experts think that with a small amount of funding—$50 billion dollars—targeted specifically towards life extension research and ending human mortality, average human lifespans could be increased by 25-50 years in about a decade's time. The world's net worth is over $200 trillion dollars, so the species can easily spare a fraction of its wealth to gain some of the most valuable commodities humans have: health and time.

Unfortunately, our species has already lost a massive amount of life hours; billions of lives have been unnecessarily cut short in the last 50 years because of widespread anti-science attitudes and policies. Even in the modern 21st Century, our evolutionary development continues to be significantly hampered by world leaders and governments who believe in non-empirical, faith-driven religious doctrines—most of which require the worship of deities whose teachings totally negate the need for radical life extension science. Virtually every major leader on the planet believes their "God" will give them an afterlife in a heavenly paradise, so living longer on planet Earth is just not that important.

Back in the real world, 150,000 people died yesterday. Another 150,000 will cease to exist today, and the same amount will disappear tomorrow. A good way to reverse this widespread deathist attitude should start with investigative government and non-government commissions examining whether public fiduciary duty requires acting in the best interest of people's health and longevity. Furthermore, investigative commissions should be set up to examine whether former and current top politicians and religious leaders are guilty of shortening people's lives for their own selfish beliefs and ideologies. Organizations and other global leaders that have done the same should be scrutinized and investigated too. And if fault or crimes against humanity are found, justice should be administered. After all, it's possible that the Catholic Church's stance on condoms will be responsible for more deaths in Africa than the Holocaust was responsible for in Europe. Over one million AIDS victims died in Africa last year alone. Catholicism is growing quickly in Africa, and there will soon be nearly 200 million Catholics on the continent. Obviously, the definition of genocide needs to be reconsidered by the public.

As a civilization of advanced beings who desire to live longer, better, and more successfully, it is our responsibility to put government, religious institutions, big business, and other entities that endorse pro-death policies on notice. Society should stand ready to prosecute anyone that deliberately promotes agendas and actions that prematurely end people's useful lives. Stifling or hindering life extension science, education, and practices needs to be recognized as a legitimate crime.

Comment author: hg00 18 February 2014 01:57:21AM *  0 points [-]

I think I'm in favor of life extension research, but I wonder if your position is a little extreme for people who are not already transhumanists. See this blog post. Edit: though, moralizing sometimes works.

Comment author: Kawoomba 08 February 2014 08:23:31PM 0 points [-]

Welcome "onboard", good to have you! You've led an interesting life, though it did take you long enough to sail to our shores.

Comment author: zoltanistvan 09 February 2014 11:20:16PM 0 points [-]

Thanks Kawoomba! I appreciate that. Cheers.

Comment author: [deleted] 07 February 2014 11:04:41AM *  2 points [-]

Comment author: Stefan_Schubert 07 February 2014 11:57:40AM 0 points [-]

Hi Elias, nice to see that you've found your way here. What are your academic interests? Philosophy, it seems, but what kind? And what else are you interested in?

Comment author: zoltanistvan 07 February 2014 02:03:10AM *  2 points [-]

Hi, My name is Zoltan Istvan. I'm a transhumanist, futurist, journalist, and the author of the philosophical novel "The Transhumanist Wager." I've been checking out this site for some time, but decided to create an account today to become closer to the community. I thought I'd start by posting an essay I recently wrote, which sums up some of my ideas. Feel free to share it if you like, and I hope you find it moving. Cheers.

"When Does Hindering Life Extension Science Become a Crime—or even Genocide?"

Every human being has both a minimum and a maximum amount of life hours left to live. If you add together the possible maximum life hours of every living person on the planet, you arrive at a special number: the optimum amount of time for our species to evolve, find happiness, and become the most that it can be. Many reasonable people feel we should attempt to achieve this maximum number of life hours for humankind. After all, very few people actually wish to prematurely die or wish for their fellow humans' premature deaths.

In a free and functioning democratic society, it's the duty of our leaders and government to implement laws and social strategies to maximize these life hours that we want to safeguard. Regardless of ideological, political, religious, or cultural beliefs, we expect our leaders and government to protect our lives and ensure the maximum length of our lifespans. Any other behavior cuts short the time human beings have left to live. Anything else becomes a crime of prematurely ending human lives. Anything else fits the common legal term we have for that type of reprehensible behavior: criminal manslaughter.

In 2001, former President George W. Bush restricted federal funding for stem cell research, one of the most promising fields of medicine in the 21st Century. Stem cells can be used to help fight disease and, therefore, can lengthen lives. Bush restricted the funding because his conservative religious beliefs—some stem cells came from aborted fetuses—conflicted with his fiduciary duty of helping millions of ailing, disease-stricken human beings. Much medical research in the United States relies heavily on government funding and the legal right to do the research. Ultimately, when a disapproving President limits public resources for a specific field of science, the research in that field slows down dramatically—even if that research would obviously lengthen and improve the lives of millions.

It's not just politicians that are prematurely ending our lives with what can be called "pro-death" policies and ideologies. In 2009, on a trip to Africa, Pope Benedict XVI told journalists that the epidemic of AIDS would be worsened by encouraging people to use condoms. More than 25 million people have died from AIDS since the first cases began being reported in the news in the early 1980s. In numerous studies, condoms have been shown to help stop the spread of HIV, the virus that causes AIDS. This makes condoms one of the simplest and most affordable life extension tools on the planet. Unfathomably, the billion-person strong Catholic Church actively supports the idea that condom usage is sinful, despite the fact that such a malicious policy has helped sicken and kill a staggering amount of innocent people.

Regrettably, in 2014, America continues to be permeated with an anti-life extension culture. Genetic engineering experiments in humans often have to pass numerous red-tape-laden government regulatory bodies in order to conduct any tests at all, especially at publically funded universities and research centers. Additionally, many states still ban human reproductive cloning, which could one day play a critical part in extending human life. The current US administration is also culpable. The White House is simply not doing enough to extend American lifespans. The US Government spends just 2% of the national budget on science and medical research, while their defense budget is over 20%, according to a 2011 US Office of Management Budget chart. Does President Obama not care about this fact, or is he unaware that not actively funding and supporting life extension research indeed shortens lives?

In my philosophical novel The Transhumanist Wager, there is a scene which takes place outside of a California courthouse where transhumanist activists are holding up a banner. The words inscribed on the banner sum up some eye-opening data: "By not actively funding life extension research, the amount of life hours the United States Government is stealing from its citizens is thousands of times more than all the American life hours lost in the Twin Towers tragedy, the AIDS epidemic, and the Vietnam War combined. Demand that your government federally fund transhuman research, nullify anti-science laws, and promote a life extension culture. The average human body can be made to live healthily and productively beyond age 150."

Some longevity experts think that with a small amount of funding—$50 billion dollars—targeted specifically towards life extension research and ending human mortality, average human lifespans could be increased by 25-50 years in about a decade's time. The world's net worth is over $200 trillion dollars, so the species can easily spare a fraction of its wealth to gain some of the most valuable commodities humans have: health and time.

Unfortunately, our species has already lost a massive amount of life hours; billions of lives have been unnecessarily cut short in the last 50 years because of widespread anti-science attitudes and policies. Even in the modern 21st Century, our evolutionary development continues to be significantly hampered by world leaders and governments who believe in non-empirical, faith-driven religious doctrines—most of which require the worship of deities whose teachings totally negate the need for radical life extension science. Virtually every major leader on the planet believes their "God" will give them an afterlife in a heavenly paradise, so living longer on planet Earth is just not that important.

Back in the real world, 150,000 people died yesterday. Another 150,000 will cease to exist today, and the same amount will disappear tomorrow. A good way to reverse this widespread deathist attitude should start with investigative government and non-government commissions examining whether public fiduciary duty requires acting in the best interest of people's health and longevity. Furthermore, investigative commissions should be set up to examine whether former and current top politicians and religious leaders are guilty of shortening people's lives for their own selfish beliefs and ideologies. Organizations and other global leaders that have done the same should be scrutinized and investigated too. And if fault or crimes against humanity are found, justice should be administered. After all, it's possible that the Catholic Church's stance on condoms will be responsible for more deaths in Africa than the Holocaust was responsible for in Europe. Over one million AIDS victims died in Africa last year alone. Catholicism is growing quickly in Africa, and there will soon be nearly 200 million Catholics on the continent. Obviously, the definition of genocide needs to be reconsidered by the public.

As a civilization of advanced beings who desire to live longer, better, and more successfully, it is our responsibility to put government, religious institutions, big business, and other entities that endorse pro-death policies on notice. Society should stand ready to prosecute anyone that deliberately promotes agendas and actions that prematurely end people's useful lives. Stifling or hindering life extension science, education, and practices needs to be recognized as a legitimate crime.

Comment author: RichardKennaway 07 February 2014 12:16:46PM 2 points [-]

Society should stand ready to prosecute anyone that deliberately promotes agendas and actions that prematurely end people's useful lives.

Do you apply this stirring declaration to the beginning of a life as well as to the end of one?

Comment author: zoltanistvan 07 February 2014 09:19:50PM 1 point [-]

First, let me just say that the essay is designed to provoke and challenge, while also aiming to move the idea forward in hopes life extension can be taken more seriously. I realize the incredible difficulties and violations of freedom, as the ideas in the essay would require. But to answer your question, I tend to concentrate on "useful" lives, so the declaration would not apply to the beginning of life, but rather to those lives that are already well under way.

Comment author: James_Miller 07 February 2014 04:48:46AM 3 points [-]

it is our responsibility to put government, religious institutions, big business, and other entities that endorse pro-death policies on notice. Society should stand ready to prosecute anyone that deliberately promotes agendas and actions that prematurely end people's useful lives.

A tiny minority group such as transhumanists should not make threats against the powers that be.

Comment author: thirdfloornorth 07 February 2014 06:37:45AM *  2 points [-]

He's making it himself, not as a spokesperson of the movement. However, as a transhumanist myself, I can't say I disagree with him. Morally speaking, when does not only actively hindering, but choosing to not vehemently pursue, life extension research constitute a threat on our lives?

Maybe it is time (or if not, it will be very soon) for transhumanism and transhumanists to enter the public sphere, to become more visible and vocal.

We have the capacity, for the first time in human history, to potentially end death, and not for our progeny but for ourselves, now. Yet we are disorganized, spread thin, essentially invisible in terms of public consciousness. People are having freakouts about something as mundane as Google Glass: We are talking about the cyberization or gross genetic manipulation of our bodies, increasing life spans to quickly approach "indefinite", etc., and not in some distant future, but in the next twenty or thirty years.

We are being held back by lack of funding, poor cohesion, and a general failure of imagination, and that is largely our own fault for being content to be quiet, to remain a fringe element, optimistically debating and self-congratulating in nooks and niches of various online communities, bothering and being bothered by few if any.

I believe it is our moral imperative to, now that is is possible, pursue life extension with every cent and scrap of resources we have available to us. To do otherwise is reprehensible.

http://www.nickbostrom.com/fable/dragon.html

Let Mr. Istvan make his threats, as long as it gets people talking about us.

Comment author: James_Miller 07 February 2014 04:10:27PM 1 point [-]

I believe it is our moral imperative to, now that is is possible, pursue life extension with every cent and scrap of resources we have available to us. To do otherwise is reprehensible.

This means taking a consequentialist public relations strategy. Imagine that group X advocates Y, and you know little about X and based on superficial analysis Y seems somewhat silly. How would your opinion of group X change if you find members of this group want to "prosecute anyone" who stands in the way of Y?

Comment author: zoltanistvan 07 February 2014 06:17:12AM 2 points [-]

Hi, Thanks for the response. I should be clear; transhumanists are not making the threat. I'm making it myself. And I'm doing it as publicly and openly as possible so there can be no misunderstanding:

http://www.psychologytoday.com/blog/the-transhumanist-philosopher/201401/when-does-hindering-life-extension-science-become-crime

http://ieet.org/index.php/IEET/more/istvan20140131

The problem is that lives are on the line. So I feel someone needs to openly state what seems to be quite obvious. Thanks for considering my thoughts.

Comment author: DanielFilan 31 January 2014 02:31:03AM *  0 points [-]

Found the newest welcome thread, posted there instead.

Comment author: higurashimerlin 23 January 2014 06:35:55PM 2 points [-]

My name is Morgan. I was brought here by my brother and have been lurking for awhile. I've have read most of the sequences which have cleared up some of my confused thinking. There were things that I didn't think about because I didn't have an answer for them. Free will and morality used to confuse me and so I never thought much about them since I didn't have a guarantee that they were answerable.

Lesswrong has helped me get back into programming. It has helped me learn to think about things with precision. And to understand how an Cognitive algorithm feels from the inside to dissolve questions.

I am going to join this community and improve my skills. Tsuyoku Naritai.

Comment author: briancrisan 21 January 2014 11:30:25AM 3 points [-]

Greetings!

I'm Brian. I'm a full-time police dispatcher and part-time graduate student in the marriage and family therapy/counseling master's degree program at the University of Akron (in northeast Ohio). Before I began studies in my master's program, I earned a bachelor's degree in emergency management. I am an atheist and skeptic. I think I can trace my earliest interest in rationality back to my high school days, when I began critically examining theism (generally) and Catholicism (in particular) while taking an elective religion class called "Questions About God." It turned out the class raised more questions than answers, for me.

I found LessWrong by way of browsing CFAR's website and wishing that I had the money to attend one of their workshops. With that being said, I haven't been lurking around LW proper for very long. Thus, I anticipate it will take some time for me to become acquainted with norms of this platform. However, after briefly browsing around, I get the sense that this is a thoughtful community of people that value rationality. That's exciting to me! I hope to get more involved, as time permits, and to eventually become a valuable contributor.

Comment author: radu_floricica 12 January 2014 09:14:36AM 3 points [-]

Hello,

I'm a 34 yo programmer/entrepreneur in Romania, with a long time interest in rationality - long before I called it by that name. I think the earliest name I had for it was "wisdom", and a desire to find a consistent, repeatable way to obtain it. Must admit at that time I didn't imagine it was going to be so complicated.

Spent some of my 20s believing I already know everything, and then I made a decision that in retrospect was the best I ever made: never to look at the price when I buy a book, but only at the likelihood of finishing it. Which is something I strongly recommend even (or especially) to cash-starved students. The first one happened to be Nassim Taleb's Black Swan, which was another huge stroke of luck. Not only it exposed me to some pretty revolutionary concepts and destroyed my illusions of omniscience, but he's a frequent name dropper and provided a lot of leads for future reading material. And the rest, as they say, is history.

Introduction aside, I'm a long time lurker and I actually came here with a request for comments. There is an often mentioned thought experiment in the sequences that compares a lot of harm done to a person (like torture) with minimum harm done to a lot of people, like a mote in the eye of a billion people. I've always found it a bit disturbing, but couldn't escape the conclusion that harm is additive and comparable. Except I now think it's not.

I've recently read Anti-fragile and found the concept of "hormesis", i.e. small harm done to a complex system generates an over-compensatory response, resulting in overall improvement. Simple examples: cold showers or lifting weights. So small harm done to a lot of people is possible to overall have net positive effects.

Two holes I see in this argument: some harms like going to the gym create hormesis, while motes in the eye don't. Also, you could just up the harm: use a big enough mote that the overall effect is a net negative, like maybe cause some permanent damage. But both holes are plugged by the fact that complex systems will always find ways to compensate. Small cornea damage gets compensated at processing level, muscle damage turns into new muscle, neuron damage means rerouting etc. There are tipping points and limits, but they're still counter-intuitive. Killing the n-th neuron will put somebody in a wheelchair, but their happiness level still bounces back. There is harm, but it's very non-linear in respect to the original damage. So I can't help but conclude that harm is simply non-additive and non-comparable, at least not easily.

Comment author: kotrfa 09 January 2014 08:54:28PM *  2 points [-]

Hello,

I'd like to get some opinions about my future goals.

I'm 21 and I'm a second-year student of engineering in Prague, Czech Republic, focusing mainly on math and then physics.

My background is not stunning - I was born in 93, visiting sporting primary school and then general high school. Until I was in second year of high school, I behaved as an idiot with below-average results in almost everything, paradoxically except extraordinary "general study presupposes" (whatever it means). My not so bad IQ - according to IQ test I took when I was 15 - is about 130 points. When I was 17, I realized that there is something about the world that needs to be done with. I started to study, mainly math and physics. I was horrible at it - I had very big disadvantage because I missed basics and wasn't able to recognize it. Anyway, I tried (but, unfortunately, not as much as I had to) and reached so-so level and I got on the technical university. Here I tried really hard and I achieved relatively good results and got into the best maths-focused student group. I'm below-average in this group (about 30 students) and my results are satisfactory. I'm quite popular thanks to collaborating on some non-study events for my schoolmates. I also created a presentation for high school students about engineering and I distributed it among faculty workers and students, who are connected to propagation.

About 10 years I obtained ECDL and it started my curiosity about informatics. But nothing special - I was autodidact in HTML and "computer administration" for regular usage. I was also very interested in economy, as my father is working in this area. I actively did cross-country skiing and play on piano and trombone.

I have high charisma, authority and ability to organize people and some bigger events, which I was usually asked to prepare (the graduate prom, matriculation etc.). I have good reasoning skills and ability to negotiate even under heavy pressure and stress. People usually enjoy time with me and appreciates me for my honesty, empathy and "cold-think" reasoning solutions, which in most time shows there were the best possible. I'm in healthy relationship for two years. My family is good background for my activities and support me. They also support me financially. My expense per month is not more than 300 USD including accommodation with in an apartment (university students, two of them from my university and domain), food and social activities.

Currently, apart from my school activities, I'm also attending some kind of philosophy group every week, where we usually discuss some topic about epistemology, relationships, culture, religions etc., we read some philosophic works (Platon), deal with art (classical music or paintings) or we write some kind of voluntary essays. I'm really interested in discussions about these topics and I try to develop my reasoning skills as often I can. For example, now I contacted a priest from local temple with whom I want to discuss some religion based questions. I autodidact psychology (last book I've read was Kahneman: Think fast and slow), rationality (started to read LW sequences), and programming. I enjoy using open source software on my Archlinux laptop and now I dived into Python as a scripting language. I also develop some web for my mother using Django and I also signed for a statistical research task about datamining in Python (pandas, numpy, scikit-learn...) or R. In school I have courses of C++ also. I'm not the most talented or generally best mathematician or programmer, but I have quite good learning (and also teaching) skills.

I've chosen my "path" - I'd like to do what's right and true and seek for the truth whenever it is possible. I feel that I'm not getting everything (e.g. from my school) I need for changing the world to a better place. I could do more. I can't decide where to focus and how to divide my attention and possibilities. Should I do aggressive autodidact of sequences? Should I focus on maths and algorithms or biases? Should I try to develop my social skills?

And the second question is simple:" Are there any Czechs who are interested in meetups in PRAGUE?"

Thank you

Comment author: Stefan_Schubert 07 January 2014 10:57:51PM 4 points [-]

Hi,

I'm a philosopher (postdoc) at the London School of Economics who recently discovered Less Wrong. I am now reading through lots of old posts, especially Yudkowsky's and lukeprog's philosophy-related material, which I find very interesting.

I think lukeprog is right when he points out that the general thrust of Yudkowsky's philosophy belongs to a naturalistic tradition often associated with Quine's name. In general, I think it would be useful to situate Yudkowsky's ideas visavi the philosophical tradition. I hope to be able to contributre something here at some point (though I should point out that I'm not an expert in the history of philosophy).

lukeprog argues for these ideas in two excellent articles:

http://lesswrong.com/lw/4vr/less_wrong_rationality_and_mainstream_philosophy/ http://lesswrong.com/lw/4zs/philosophy_a_diseased_discipline/

I agree with most of what is said there, and am myself very critical of mainstream analytical philosophy. It also seems to me that the overall program advocated here - to let psychological knowledge permeate all philosophical arguments in a very radical way - is very promising. Though there are philosophers who make use of psychology, do experiments, etc., few let it influence their thinking as radically as it is done here.

The site seems very interesting in other respects as well. I am presently reading up on cognitive science (I found this site after googling on Stanovichs Rationality and the Reflective Mind, which I now have read) and am grateful for the info on this subject gathered on Less Wrong.

Comment author: tanagrabeast 02 January 2014 02:56:50AM 10 points [-]

Greetings.

I'm a long-time singularitarian and (intermediate) rationalist looking be a part of the conversation again. By day I am an English teacher in a suburban American high school. My students have been known to Google me. Rather than self-censor I am using a pseudonym so that I will feel free to share my (anonymized) experiences as a rationalist high school teacher.

I internet-know a number of you in this community from early years of the Singularity Institute. I fleetingly met at a few in person once, perhaps. I used to write on singularity-related issues, and was a proud "sniper" of the SL4 mailing list for a time. For the last 6-7 years I've mostly dropped off the radar by letting "life" issues consume me, though I have continued to follow the work of the key actors from afar with interest. I allow myself some pride for any small positive impact I might have once had during a time of great leverage for donors and activists, while recognizing that far too much remains undone. (If you would like to confirm your suspicions of my identity, I would love to hear from you with a PM. I just don't want Google searches of my real name pulling up my LW activity.)

High school teaching has been a taxing path, along with parenting, and it has been all too easy to use these as excuses to neglect my more challenging (yet rewarding) interests. I let my inaction and guilt reinforce each other until I woke up one day, read HPMoR, and realized I had long-ago regressed into an NPC.

Screw that.

Other background tidbits: I'm one of those atheist ex-mormons that seem so plentiful on this page (since 2000ish). I'm a self-taught "hedge coder" who has successfully used inelegant-but-effective programming in the service of my career. I feel effective in public education, which is not without its rewards. But on some important levels teaching at an American public high school is also a bit like working security at Azkaban, and I'm not sure how many more years I'll be able to keep my patronus going.

I've been using GTD methodologies for the last eight years or so, which has been great for letting me keep my mind clear to work on important tasks at hand; however, my dearest personal goals (which involve writing, both fiction and non) live among some powerful Ugh Fields. If I had been reading LW more closely, I probably would've discovered the Pomodoro method a lot sooner. This is helping.

My thanks to all who share their insights and experiences on this forum.

Comment author: SaidAchmiz 02 January 2014 03:51:52AM 0 points [-]

Welcome to Less Wrong!

Is your user name a reference to "Darmok"?

Comment author: tanagrabeast 02 January 2014 04:25:34AM 0 points [-]

Yes. It's amazing how memorable people find that one episode. Props to the writers.

Comment author: Faustus2 14 December 2013 10:36:36PM 0 points [-]

Hello to you all, I am Chris.

I live in England and attend my local High school (well, in England we call the senior years/curriculums a sixth form). I take Mathematics, Further mathematics, physics and philosophy. I actually happened upon Lesswrong two years ago, when I was 16, whilst searching for interesting discussions on worldviews. Although I had never really been interested in rationality (up until that point I hasten to add!), I had a seething urge to sort out my worldview as quickly as I could. I just got so sick of the people who went to sunday school coming out with claims about the universe that didn't jive with the understanding of modern physics. So I read the reductionism sequence and realised I was a reductionist. The way Eliezer 'spelled it out' just really struck me as a great way to say what I had started to feel. Shortly afterwards naturalism , or rather metaphysical naturalism, became my first great love. I have a good collection of friends, but none of them have really cared for 'waxing on worldviews' like me. I guess I'm just really happy that I get to speak with a community that has stuff in common with me (not just worldviews, but other cool topics as well). I guess camaraderie is eagerly sought. I would love to talk with people of my age group (I suppose 15-24) but of course I should love to meet with anyone with a similar mindset to me. I live near Reading. If anyone would like to speak with me, whether it be through Lesswrong, Facebook or just meeting up for a chat, just message me and I shall do my utmost to entertain/be friendly with you. :)

Comment author: ericgarr 09 December 2013 12:47:24AM 2 points [-]

Hi everybody,

My name is Eric, and I'm currently finishing up my last semester of undergraduate study and applying to Ph.D. programs in cognitive psychology/cognitive neuroscience. I recently became interested in the predictive power offered by formal rational models of behavior after working in Paul Glimcher's lab this past summer at NYU, where I conducted research on matching behavior is rhesus monkeys. I stumbled upon Less Wrong while browsing the internet for behavioral economics blogs. After reading a couple of posts, I decided to join.

Some sample topics that I like reading about and discussing include intertemporal choice, risk preferences, strategic behavior in the context of games, reinforcement learning, and the evolution of cooperation. I look forward to chatting with some of you!

Comment author: [deleted] 03 December 2013 12:05:33PM 3 points [-]

Hello, LW,

One of my names is holist. I am 45. Self-employed family man, 6 kids, 2 dogs, 1 cat. Originally a philosopher (BA+MA from Sussex, UK), but I've been a translator for 19 years now... it is wearing thin. Music and art are also important parts of my life (have sold music, musically directed a small circus, have exhibited pictures), and recently, with dictatorship being established here in Hungary, politics seems increasingly urgent, too. I dabble in psychotherapy and call myself a Discordian. Recently, I started thinking about doing a PhD somewhere. My topic is very general: what has caused the increasingly hostile relationship between individual and culture, and what are the remedies available? My window of opportunity is a few years away: I am intermittently thinking about possible supervisors. I have a blog at holist.hu, some of it is in English and it has a lot of pictures. The discordians at PD kicked me out, I found the Secular Café to be indescribably boring, I am a keen MeFite, but I'd like something a little more discussioney. A friend pointed me at LW. I hope it works out.

For pointers, here's a slightly random list of some books that are very important to me: Fiction: Mason and Dixon by Thomas Pynchon, Ulysses by James Joyce, Karnevál by Béla Hamvas, Moby Dick by Herman Melville, The Diamong Age by Neal Stephenson. Non-fiction: The Continuum Concept by Jean Liedloff, The Facts of Life by R. D. Laing, Children of the Future by Wilhelm Reich, The Drama of the Gifted Child by Alice Miller, The Story of B by Daniel Quinn, Tools for Conviviality by Ivan Illich

I guess I'll start by lurking, but you never know :)

Comment author: aquaticko 29 November 2013 05:07:02PM 2 points [-]

Hello, my name is Luke. I'm an urban planning graduate student at Cleveland State University, having completed an undergrad in philosophy at the University of New Hampshire a year ago. It was the coursework I did at that school which lead me to be interested in the nebulous and translucent topic of rationality, and I'm happy to see so many people involved and interested in the same conversations I'd spend hours having with classmates. Heck, the very question I was asking myself in something of an ontological sense--am I missing the trees for the forest--is what led me here, specifically to Eliezer's article on the fallacies of compression, which was somewhat helpful. Suffice to say, I tend to think I'm not missing the trees for the forest, and that in fact the original form of the idiom remains true for most other people, though thankfully, not many here.

I'm deeply interested in epistemology, metaphysics, aesthetics, and metaethics, all of which I attempt to approach in systemic ways. As for what led me to consider myself a rationalist in these endeavors...I'm not sure I do. In fact, I'm not sure anyone can or should think of themselves a rationalist, considering that basic beliefs, other than solipsism, are inductive and inferential, and thus fallible. We could argue in circles forever (as others have) what constitutes knowledge, but any definition seems, in my view, to be arbitrary and thus non-universal and therefore, again, fallible--even mathematical knowledge and formal logic.

Granted, I don't sit in a corner rocking back and forth sucking my thumb, driven mad by the uncertainty of it all, but I also operate with the knowledge that whatever I deem rational behavior and thought processes only seem rational because I've pre-decided what constitutes rational behavior (i.e., circularity, or coherentism at best...feeling like I'm writing a duplicate of a different post). Of course, all that seems like too easy an exit from a number of hard problems, so I keep reading to make sure that, in fact, I oughtn't be rocking back and forth in a corner sucking my thumb for the utility of it, turning into a kind of utility monster. An absurdist I remain, but one with a pretty strong intuitive consequentialist metaethical framework which allows me to find great joy in the topics covered on LW.

Comment author: David_Chapman 23 November 2013 10:57:56PM 2 points [-]

Hi, I have a site tech question. (Sorry if this is the wrong place to post that!—I couldn't find any other.)

I can't find a way to get email notifications of comment replies (i.e. when my inbox icon goes red). If there is one, how do I turn it on?

If there isn't one, is that a deliberate design feature, or a limitation of the software, or...?

Thanks (and thanks especially to whoever does the system maintenance here—it must be a big job.)

Comment author: TheOtherDave 23 November 2013 11:42:00PM 1 point [-]

There's no way I know of to get email notifications, and I've looked enough that I'm pretty confident one doesn't exist.
No idea if it's a deliberate choice or a software limitation.

Comment author: alexg 13 November 2013 12:33:03PM 2 points [-]

G'day

As you can probably guess, I'm Alex. I'm a high school student from Australia and have been disappointed with the education system here from quite some time.

I came to LW via HPMoR which was linked to me by a fellow member of the Aus IMO team. (I seriously doubt I'm the only (ex-)Olympian around here - seems just the sort of place that would attract them). I've spent the past few weeks reading the sequences by EY, as well as miscellaneous other stuff. Made a few (inconsequential) posts too.

I have very little in the way of controversial opinions to offer (relative to the demographics of this site) as just about all the unusual positions it takes I already agreed with (e.g. athiesm) or seemed pretty obvious to me after some thought (e.g. transhumanism). Maybe it's just hindsight bias.

I'm slightly disappointed with the ban on political discussion. I do agree that it should not be mentioned when not relevant but it seems a shame to waste this much rationality in one place by forbidding them to use it where it's most needed. A possible compromise would be to create a politics dicussion page to discuss pros and cons to particular ideologies. (If one already exists point me to it). A reason cited is that there are other sites to discuss politics - if any do so rationally I'd like to see them.

It is a relief to be somewhere where I don't have to constantly take into account inferential distance, and I shall try to make the most of this. I only resolve to write just that which has not been written.

Comment author: Vaniver 13 November 2013 05:02:43PM *  3 points [-]

Welcome!

There have been previous political threads, like here, here, or here. If you search "politics," you'll find quite a bit. Here was my response to the proposal that we have political discussion threads; basically, I think politics is a suboptimal way to spend your time. It might feel useful, but that doesn't mean it is useful. Here's Raemon's comment on the norm against discussing politics. Explicitly political discussion can be found on MoreRight, founded by posters active on LessWrong, as well as on other blogs. (MoreRight is part of 'neoreaction', which Yvain has recently criticized here, for example.)

I don't see what you mean by the 'pros and cons' of holding a particular ideology. Ideologies are, generally, value systems- they define what is a pro and what is a con.

Comment author: Lumifer 13 November 2013 05:20:36PM 2 points [-]

I must add that not all political discussion is a mud-flinging match between the Cyans and the Magentas.

For example, the Public Choice theory is a bona fide intellectual topic, but it's also clearly political.

I would also argue that knowing things like the scope of NSA surveillance is actually useful.

Comment author: Vaniver 13 November 2013 05:29:16PM *  1 point [-]

Cyans and the Magentas

I'm curious why you'd divert from the historically compelling example of the Blues and the Greens.

For example, the Public Choice theory is a bona fide intellectual topic, but it's also clearly political.

It's about politics, but the methodology is not political. The part of politics that's generally fun for people is putting forth an impassioned defense of some idea or policy. That's generally not useful on LessWrong unless it's about a site policy- and even then, the passion probably doesn't help.

I would also argue that knowing things like the scope of NSA surveillance is actually useful.

Sure.

Comment author: Lumifer 13 November 2013 05:47:22PM 3 points [-]

I'm curious why you'd divert from the historically compelling example of the Blues and the Greens.

I strongly associate the Greens with, well, the Greens -- a set of political parties in Europe and the whole environmentalist movement.

Blue is a politically-associated color in the US as well.

The part of politics that's generally fun for people is putting forth an impassioned defense of some idea or policy.

True, but LW is VERY unrepresentative sample :-) and maybe we could do a bit better. You're right in that discussing the "pros and cons" of ideological positions is not a good idea, but putting "Warning: mindkill" signs around a huge area of reality and saying "we just don't go there" doesn't look appealing either.

Comment author: MathiasZaman 31 October 2013 03:46:38PM 1 point [-]

Hello everyone,

My name is Mathias. I've been thinking about coming here for quite some while and I finally made the jump. I've been introduced to this website/community through Harry Potter and the Methods of Rationality and I've been quite active on it's subreddit (under the alias of Yxoque).

For a while now, I've been self-identifying as "aspiring rationalist" and I want to level up even further. One way I learn quickly is through conversation, so that's why I finally decided to come here. Also because I wanted to attend a meet-up, but if felt wrong to join one of a community I'm not a member of.

As for info on myself, I'm not sure how interesting that is. I've recently graduated as a Bachelor in Criminology at the University of Brussels and I'm currently looking for a suitable job. I still need to figure out my comparative advantage.

I'm also reading through a pdf of all the blog posts, currently I'm on the Meta-ethics sequence. In my free time I'm (slowly) working on annotating HPMOR and convincing people to write Rational!Animorphs fanfiction.

Comment author: Andrew_93 26 October 2013 06:25:13PM *  1 point [-]

Hi there, I am Andrew, living in Hungary and studying to be an IT physicist some day in an utmost lazy way. I've just recently discovered this site and so far I can't really believe what I'm seeing. I have been thinking myself before about a website whose main purpose is basically making its users wiser and/or more rational. - about which my main question later will be put that if u could answer would be great; also, excuse my English, it's not my native language.

I believe rationality can be expressed as the set of "right" algorithms in a given context. The rightfulness of algorithms in this case is dependent on the goal which the context defines.

My question is that are the majority of people here generally conscious of the significance of finding the "fancy word for global, important and unique" goal, or as I shall put "pure wisdom" OR do they - or you - here just feel the necessity to lay down the healthy plain soil for building the "tower of wisdom" and care less about the actual adventure of building it.

My main point is that even tough our best tool to gain wisdom is through rational thinking - and a little extra - , how we react and how far we go on this road as rational beings is the function of each one of our unique perspective of life. Are there individuals here whose (main) point in life is to possess the right perspective of life?

For example if we accepted all sciences and set aside our big hopes like the one for afterlife, we may come to the conclusion that any kind of necessity other than our primal ones that have evolved in our ancestors over the millions of years are flawed or delusional. Not taking the alternative of nihilism -its reflection to humans- or physicalism as its sort of parallel philosophy into account is neither irrational nor rational, but unwise I believe even though the philosophy itself doesn't bring or promise much materialistic profit.

Here, my goals in the first place would - I mean: will - be to observe, learn and then adjust my knowledge and beliefs since inevitably I'm gonna bump into other people's belief systems which are built on a rational, in its broadest sense, and healthy ground. Looking forward to it. And looking forward getting to know other people's struggles achieving similar goals and their personalities, have a nice day :)

Comment author: Mitchell_Porter 29 October 2013 09:14:09AM 3 points [-]

Are there individuals here whose (main) point in life is to possess the right perspective of life?

There are individuals here whose main point in life is to ensure that the first superhuman artificial intelligence possesses the right perspective of life. Is that close enough? :-)

One thing that distinguishes LW rationalism from other historic rationalist movements is a strong interest in transhumanism and the singularity. Historically, wisdom has usually been about accepting the limited and disappointing nature of life, whether your attitude is stoic, epicurean, or bodhisattvic. But the cultural DNA of LW includes nanotechnology, space travel, physical immortality, mind uploading, and computer-brains the size of whole solar systems. There is a strong tendency to think that rationality consists of remaking nature in the image of your goals, rather than vice versa, and that the struggle is to determine which values will shape the universe. This is a level of Promethean ambition more common in apocalyptic movements than in rationalist movements.

This aspect of LW comes and goes in prominence.

Comment author: Jennifer_H 23 October 2013 04:19:48AM 13 points [-]

Hello!

I'm Jennifer; I'm currently a graduate student in medieval literature and a working actor. Thanks to homeschooling, though, I do have a solid background and abiding interest in quantum physics/pure mathematics/statistics/etc., and 'aspiring rationalist' is probably the best description I can provide! I found the site through HPMoR.

Current personal projects: learning German and Mandarin, since I already have French/Latin/Spanish/Old English/Old Norse taken care of, and much as I personally enjoy studying historical linguistics and old dead languages, knowing Mandarin would be much more practical (in terms of being able to communicate with the greatest number of people when travelling, doing business, reading articles, etc.)

Comment author: komponisto 01 January 2014 12:32:57AM 0 points [-]

Your self-description is one of the best arguments for homeschooling I have ever seen or could imagine being made. (See also: Lillian Pierce.)

Welcome to LW, and please keep existing.

Comment author: Adele_L 23 October 2013 04:54:33AM 2 points [-]

Hey, another homeschooled person! There seem to be a lot of us here. How was your experience? Mine was the crazy religious type, but I still consider it to have been an overall good thing for my development relative to other feasible options.

Comment author: lavalamp 01 January 2014 07:08:18AM 1 point [-]

Me three-- I thought I was the only one, where are we all hiding? :)

Comment author: Jennifer_H 23 October 2013 05:12:29AM 2 points [-]

My experience was, overall, excellent - although my parents are definitely highly religious. (To be more precise, my father is a pastor, so biology class certainly contained some outdated ideas!) However, I'm in complete agreement - relative to any other possible options, I don't think I could have gotten a better education (or preparation for postsecondary/graduate studies) any other way.

Comment author: Adele_L 24 October 2013 03:46:37AM 2 points [-]

Yeah, I got taught young earth creationism instead of evolution. But despite this, i think I was better prepared academically than most of my peers.

Comment author: shminux 23 October 2013 04:39:21AM -1 points [-]

Impressive! How do you plan to learn Mandarin? Immersion? Rosetta Stone?

Comment author: Jennifer_H 23 October 2013 05:00:37AM *  3 points [-]

Combination of methods based on what has worked for me in the past with other languages! I've used Rosetta Stone before, for French & Spanish, and while it's definitely got advantages, I (personally - I also know people who love it!) also found it very time-consuming for very little actual learning, and it's also expensive for what it is.

Basically:

a) I have enough friends who are either native or fluent speakers of Mandarin that once I'm a little more confident with the basics, I will draft them to help me practice conversation skills :)

b) My university offers inexpensive part-time courses to current students.

c) Lots of reading, textbook exercises, watching films, listening to music, translating/reading newspapers, etc. in the language.

d) I'm planning to go to China to teach English in the not-too-distant future, so while I'd like to have basic communication skills down before I go, immersion will definitely help!

Comment author: aarongertler 10 October 2013 08:38:04PM 2 points [-]

Salutations!

My name is Aaron. I'm a college junior on the tail end of the cycle of Bar Mitzvah to New Atheist to info-omnivorous psychology geek to attempted systems thinker. Prospective Psychology/Cognitive Science major at Yale, very interested in meeting other rationalists in the New Haven area. I'm on the board of the Yale Humanist Community, I'm a research assistant in a neuroscience lab, and I do a lot of writing.

Big problems I've been thinking a lot about: Why are most people wildly irrational in the amount of time they're willing to devote to information search (that is, reducing uncertainty around uncertain decisions)? How can humanists and rationalists build a compelling community that serves adults of all ages as well as children? What sorts of media tend to encourage the "shift" from bad thinking to good thinking, and/or passive to active thinking (NPC vs. hero mindset, sort of--this one is complicated), and how can we get that media in the hands of more people?

I read HPMoR without really noticing Less Wrong, but have been linked to a few posts over the years. Last spring, I found "Privileging the Question", which rang so true that I went on to read the Sequences and much of the rest. I was never very certain in my philosophy before finding the site, but now I'm pretty sure I at least know how to think about philosophy, which is nice.

The next few years hopefully involve me getting a job out of college that will allow me to build savings while donating plenty, while aligning me to take a position in some high-upside sector of tech or in the rationalist arena, but a lot of people say that, and I'm very unsure about what will actually happen if I flunk my case interviews. Still, the future will be better than the past regardless, and that thought keeps me going (as does knowing how many people are out there working to avoid future-is-worse-than-past scenarios).

Comment author: JoshElders 10 October 2013 08:13:48PM 3 points [-]

I am a celibate pedophile. That means I feel a sexual and romantic attraction to young girls (3-12) but have never acted on that attraction and never will. In some forums, this revelation causes strong negative reactions and a movement to have me banned. I hope that's not true here.

From a brief search, I see that someone raised the topic of non-celibate pedophilia, and it was accepted for discussion. http://lesswrong.com/lw/67h/the_phobia_or_the_trauma_the_probem_of_the_chcken/ Hopefully celibate pedophilia is less controversial.

I have developed views on the subject, though I like to think that I can be persuaded to change them, and one thing I hope to get here on LessWrong is reason-based challenges. Hopefully others will find the topics informative as well. In the absence of advice on a better way to proceed, I plan to make posts in Discussion now and then on various aspects of the topic.

I'm in my 50s and am impressed with the LessWrong approach in general and have done my best to follow some of its precepts for years. I have read most of the core sequences.

Comment author: wedrifid 10 October 2013 09:13:39PM *  -1 points [-]

I have developed views on the subject, though I like to think that I can be persuaded to change them, and one thing I hope to get here on LessWrong is reason-based challenges. Hopefully others will find the topics informative as well. In the absence of advice on a better way to proceed, I plan to make posts in Discussion now and then on various aspects of the topic.

Reason based challenges? There doesn't seem to be much to say. You have some attraction, you choose not to act on it for altruistic and/or pragmatic reasons. Nothing much to challenge.

Perhaps an aspect that could provoke discussion would be the decision of whether to self-modify to not have those desires if the technological capability to do so easily were available. I believe there is a Sci. Fi. short story out there that explores this premise. My take is that given an inclusive preference to not have sex with children I'd be perfectly content to self modify to remove the urge which I would never endorse acting on. I would consider this to be a practical choice that allowed me to experience more pleasure and less frustration without sacrificing my values. I would not consider it a moral obligation for people to do so.

A variant situation would be when that same technology is available, along with the technology to detect both preferences and decision making traits in people. Consider a case where it is detected that someone with a desire to do a forbidden thing and who is committed to not doing that thing and yet who has also identifiable deficits in willpower or decision making that make it likely that they will act on the desires anyway. In that case it seems practical to enforce a choice of either self-modifying or submitting to restrictions of behaviour and access.

A further scenario would be one in which for technological or evolutionary reasons there are 12 year old girls who would not be physically or psychologically harmed by sexual liaisons with adults and who have been confirmed (by brain scan and superintelligent extrapolation) to prefer outcomes where they engage in such practice than those in which they don't. That would tend to make any residual moral objection to pedophilia to be not about altruistic consideration of consequenes and all about prudishness. (I of course declare vehemently that I would still oppose pedophilia regardless of consequences. Rah Blues!)

For some real controversy I suppose you could do an analysis of the research on just how much physical and psychological damage is done to children in loving-but-sexual relationships with adults versus how much damage is done by other adults who wish to signal their opposition to the crime by the way they treat the victims. Mind you, that is something that if it were ever to be discussed would perhaps best be discussed by people who are not celibate pedophiles. It would be offensive enough to many people to even see it considered a valid avenue of enquiry by an individual with zero sexual interest in children. Protection of victims from 'righteous' authorities is something best done by those with no interest in committing the crime.

Comment author: volya 07 October 2013 01:17:08PM 8 points [-]

Hi, I am Olga, female, 40, programmer, mother of two. Got here from HPMoR. Can not as yet define myself as a rationalist, but I am working on it. Some rationality questions, used in real life conversations, have helped me to tackle some personal and even family issues. It felt great. In my "grown-up" role I am deeply concerned to bring up my kids with their thoughts process as undamaged as I possibly can and maybe even to balance some system-taught stupidity. I am at the start of my reading list on the matter, including LW sequences.

Comment author: [deleted] 07 October 2013 07:42:25PM 1 point [-]

Welcome!

Can not as yet define myself as a rationalist, but I am working on it.

Many people here call themselves aspiring rationalists.

Comment author: [deleted] 07 October 2013 12:35:34PM *  -1 points [-]

Hi, I always like to be less wrong and try to verify and falsify my own take on philosophical Modernism, which I have developed since my student days. (FYI, I am twice that age now.) I believe we should all have opinions about everything and look for independent confirmation, rational or emotional, both for the give and for the take, when earned, to find truth. When it doesn't happen, I try to improve on my theory. I have done so online for years at http://crpa.co. I would like you to try and find any or all flaws and make a case out of it. Thanks in advance. I will be critical too of your work if you would like me to. Best regards ~Ron dW.

Comment author: DSimon 07 October 2013 03:11:10PM 1 point [-]

I've been trying very hard to read the paper at that link for a while now, but honestly I can't figure it out. I can't even find anything content-wise to criticize because I don't understand what you're trying to claim in the first place. Something about the distinction between map and territory? But what the heck does that have to do with ethics and economics? And why the (seeming?) presumption of Christianity? And what does any of that have to do with this graph-making software you're trying to sell?

It would really help me if you could do the following:

  1. Read this short story: http://yudkowsky.net/rational/the-simple-truth/
  2. Please explain, using language no more complicated or technical than that used in the story, whether the idea of "truth" that the story proposes lines up with your philosophy or not, and why or why not.
Comment author: shminux 07 October 2013 05:30:35PM 0 points [-]

His pile of CRPA reads like autogenerated entries from http://snarxiv.org/.

Comment author: [deleted] 07 October 2013 03:53:18PM 0 points [-]

Thanks for your reply DSimon. I like Yudkowsky's story. Truth as I understand it, is simply that which can be detected by independent confirmation, if we look for it. It is the same methodology as being used in science, justice, journalism and other realms.

Comment author: nasrin 02 October 2013 11:31:31PM 2 points [-]

Hi! Everyone below are superbly impressive! I'm a physicist, in my second year of teaching English and that's as much rationality as I can provide at the moment. Looking to relocate to China in an effort to be superhuman. Would really appreciate a few pointers on teaching institutions to avoid/ embrace.

Excellent reading here, thanks! Nas

Comment author: LM7805 18 September 2013 01:38:32AM 5 points [-]

Hi. I've been a distant LW lurker for a while now; I first encountered the Sequences sometime around 2009, and have been an avid HP:MOR fan since mid-2011.

I work in computer security with a fair bit of software verification as flavoring, so the AI confinement problem is of interest to me, particularly in light of recent stunts like arbitrary computation in zero CPU instructions via creative abuse of the MMU trap handler. I'm also interested in applying instrumental rationality to improve the quality and utility of my research in general. I flirt with some other topics as well, including capability security, societal iterated game theory, trust (e.g., PKI), and machine learning; a meta-goal is to figure out how to organize my time so that I can do more applied work in these areas.

Apart from that, lately I've become disillusioned with my usual social media circles, in part due to a perceived* uptick in terrible epistemology and in part due to facing the fact that I use them as procrastination tools. I struggle with akrasia, and am experiencing less of it since quitting my previous haunts cold turkey, but things could still be better and I hope to improve them by seeking out positive influences here.

*I haven't measured this. It's entirely possible I've become more sensitive to bad epistemology, or some other influence is lowering my tolerance to bad epistemology.

Comment author: Roman_Yampolskiy 16 September 2013 10:35:39PM 11 points [-]

Hey, my name is Roman. You can read my detailed bio here, as well as some research papers I published on the topics of AI and security. I decided to attend a local LW meet up and it made sense to at least register on the site. My short term goal is to find some people in my geographic area (Louisville, KY, USA) to befriend.

Comment author: Wei_Dai 17 September 2013 02:04:40PM 1 point [-]

Hi Roman. Would you mind answering a few more questions that I have after reading your interview with Luke? Carl Shulman and Nick Bostrom have a paper coming out arguing that embryo selection can eventually (or maybe even quickly) lead to IQ gains of 100 points or more. Do you think Friendly AI will still be an unsolvable problem for IQ 250 humans? More generally, do you see any viable path to a future better than technological stagnation short of autonomous AGI? What about, for example, mind uploading followed by careful recursive upgrading of intelligence?

Comment author: shminux 17 September 2013 04:42:29PM -1 points [-]

Carl Shulman and Nick Bostrom have a paper coming out arguing that embryo selection can eventually (or maybe even quickly) lead to IQ gains of 100 points or more.

I wonder how they propose to avoid the standard single-trait selective breeding issues, like accumulation of undesirable traits. For example, those geniuses might end up being sickly and psychotic.

Comment author: arundelo 17 September 2013 05:01:03PM 0 points [-]

It seems to me that this would not be a problem with iterated embryo selection, but I might be wrong.

See also Yvain's "modal human" post.

Comment author: [deleted] 17 September 2013 04:46:17PM 0 points [-]

Would it matter? C.f. goldmage.

Comment author: Roman_Yampolskiy 17 September 2013 03:19:41PM 2 points [-]

Hey Wei, great question! Agents (augmented humans) with IQ of 250 would be superintelligent with respect to our current position on the intelligence curve and would be just as dangerous to us, unaugment humans, as any sort of artificial superintelligence. They would not be guaranteed to be Friendly by design and would be as foreign to us in their desires as most of us are from severely mentally retarded persons. For most of us (sadly?) such people are something to try and fix via science not something for whom we want to fulfill their wishes. In other words, I don’t think you can rely on unverified (for safety) agent (event with higher intelligence) to make sure that other agents with higher intelligence are designed to be human-safe. All the examples you give start by replacing humanity with something not-human (uploads, augments) and proceed to ask the question of how to safe humanity. At that point you already lost humanity by definition. I am not saying that is not going to happen, it probably will. Most likely we will see something predicted by Kurzweil (merger of machines and people).

Comment author: [deleted] 17 September 2013 08:25:00PM *  0 points [-]

I'm going to throw out some more questions. You are by no means obligated to answer.

In your AI Safety Engineering paper you say, "We propose that AI research review boards are set up, similar to those employed in review of medical research proposals. A team of experts in artificial intelligence should evaluate each research proposal and decide if the proposal falls under the standard AI – limited domain system or may potentially lead to the development of a full blown AGI."

But would we really want to do this today? I mean, in the near future--say the next five years--AGI seems pretty hard to imagine. So might this be unnecessary?

Or, what if later on when AGI could happen, some random country throws the rules out? Do you think that promoting global cooperation now is a useful way to address this problem, as I assert in this shamelessly self-promoted blog post?

The general question I am after is, How do we balance the risks and benefits of AI research?

Finally you say in your interview, "Conceivable yes, desirable NO" on the question of relinquishment. But are you not essentially proposing relinquishment/prevention?

Comment author: Roman_Yampolskiy 17 September 2013 10:12:39PM 1 point [-]

Just because you can’t imaging AGI in the next 5 years, doesn’t mean that in four years someone will not propose a perfectly workable algorithm for achieving it. So yes, it is necessary. Once everyone sees how obvious AGI design is, it will be too late. Random countries don’t develop cutting edge technology; it is always done by the same Superpowers (USA, Russia, etc.). I didn’t read your blog post so can’t comment on “global cooperation”. As to the general question you are asking, you can get most conceivable benefits from domain expert AI without any need for AGI. Finally, I do think that relinquishment/delaying is a desirable thing, but I don’t think it is implementable in practice.

Comment author: TheOtherDave 18 September 2013 12:53:20AM 1 point [-]

you can get most conceivable benefits from domain expert AI without any need for AGI.

Is there a short form of where you see the line between these two types of systems? For example, what is the most "AGI-like" AI you can conceive of that is still "really a domain-expert AI" (and therefore putatively safe to develop), or vice-versa?

My usual sense is that these are fuzzy terms people toss around to point to very broad concept-clusters, which is perfectly fine for most uses, but if we're really getting to the point of trying to propose policy based on these categories, it's probably good to have a clearer shared understanding of what we mean by the terms.

That said, I haven't read your paper; if this distinction is explained further there, that's fine too.

Comment author: Roman_Yampolskiy 18 September 2013 09:06:38PM 0 points [-]

Great question. To me a system is domain specific if it can’t be switched to a different domain without re-designing it. I can’t take Deep Blue and use it to sort mail instead. I can’t take Watson and use it to drive cars. An AGI (for which I have no examples) would be capable of switching domains. If we take humans as an example of general intelligence, you can take an average person and make them work as a cook, driver, babysitter, etc, without any need for re-designing them. You might need to spend some time teaching that person a new skill, but they can learn efficiently and perhaps just by looking at how it should be done. I can’t do this with domain expert AI. Deep Blue will not learn to sort mail regardless of how many times I demonstrate that process.

Comment author: TheOtherDave 18 September 2013 10:24:54PM 0 points [-]

(nods) That's fair. Thanks for clarifying.

Comment author: shminux 17 September 2013 04:50:19PM *  -2 points [-]

with IQ of 250

Technicality: there is no such thing as IQ of 250, since IQ is a score on a test, and there is no test calibrated for 1 in 10^22 humans. What you probably mean is the mysterious hypothetical g-factor, partly responsible for the IQ scores, or maybe some other intelligence marker.

Comment author: Moss_Piglet 17 September 2013 05:05:54PM 4 points [-]

If you understand the point there's no reason to make a comment like this except as an attempt to show off. Changing "250 IQ" to "+10 sd out from the mean intelligence" only serves to make the original point less accessible to people not steeped in psychometry.

Comment author: TheOtherDave 17 September 2013 06:24:56PM 1 point [-]

You don't have to be steeped in psychometry to understand what a standard deviation is.
And if we're going to talk about intelligence at all, it is often helpful to keep in mind the difference between IQ and intelligence.

Comment author: Wei_Dai 17 September 2013 04:14:43PM 4 points [-]

I think if I became an upload (assuming it's a high fidelity emulation) I'd still want roughly the same things that I want now. Someone who is currently altruistic towards humanity should probably still be altruistic towards humanity after becoming an upload. I don't understand why you say "At that point you already lost humanity by definition".

Comment author: Dr_Manhattan 17 September 2013 09:05:45PM 3 points [-]

Someone who is currently altruistic towards humanity should

Wei, the question here is would rather than should, no? It's quite possible that the altruism that I endorse as a part of me is related to my brain's empathy module, much of which might be broken if I see cannot relate to other humans. There are of course good fictional examples of this, e.g. Ted Chiang's "Understand" - http://www.infinityplus.co.uk/stories/under.htm and, ahem, Watchmen's Dr. Manhattan.

Comment author: Eliezer_Yudkowsky 17 September 2013 10:24:39PM 2 points [-]

Logical fallacy: Generalization from fictional evidence.

A high-fidelity upload who was previously altruistic toward humanity would still be altruistic during the first minute after awakening; their environment would not cause this to change unless the same sensory experiences would have caused their previous self to change.

If you start doing code modification, of course, some but not all bets are off.

Comment author: MugaSofer 20 September 2013 06:42:03PM 0 points [-]

Perhaps the idea is that the sensory experience of no longer falling into the category of "human" would cause the brain to behave in unexpected ways?

I don't find that especially likely, mind, although I suppose long-term there might arise a self-serving "em supremacy" meme.

Comment author: Bugmaster 18 September 2013 03:46:30AM -1 points [-]

their environment would not cause this to change unless the same sensory experiences would have caused their previous self to change.

I don't see why this is necessarily true, unless you treat "altruism toward humanity" as a terminal goal.

When I was a very young child, I greatly valued my brightly colored alphabet blocks; but today, I pretty much ignore them. My mind had developed to the point where I can fully visualize all the interesting permutations of the blocks in my head, should I need to do so for some reason.

Comment author: somervta 18 September 2013 08:30:35AM 1 point [-]

I don't see why this is necessarily true, unless you treat "altruism toward humanity" as a terminal goal.

Well, yes. I think that's the point. I certainly don't only value other humans for the way that they interest me - If that were so, I probably wouldn't care about most of them at all. Humanity is a terminall value to me - or, more generally, the existence and experiences of happy, engaged, thinking sentient beings. Humans qualify, regardless of whether or not uploads exist (and, of course, also qualify.

Comment author: Bugmaster 19 September 2013 10:37:30PM 0 points [-]

How do you know that "the existence and experiences of happy, engaged, thinking sentient beings" is indeed one of your terminal values, and not an instrumental value ?

Comment author: Dr_Manhattan 18 September 2013 02:42:08AM 3 points [-]

Well, I did put a disclaimer by using the standard terminology :) Fiction is good for suggesting possibilities, you cannot derive evidence from it of course.

I agree on the first-minute point, but do not see why it's relevant, because there is the 999999th minute by which value drift will take over (if altruism is strongly related to empathy). I guess upon waking up I'd make value preservation my first order of business, but since an upload is still evolution's spaghetti code it might be a race against time.

Comment author: Bugmaster 17 September 2013 09:31:49PM 1 point [-]

+1 for linking to Understand ; I remembered reading the story long ago, but I forgot the link. Thanks for reminding me !

Comment author: Roman_Yampolskiy 17 September 2013 08:22:36PM 2 points [-]

We can talk about what high fidelity emulation includes. Will it be just your mind? Or will it be Mind + Body + Environment? In the most common case (with an absent body) most typically human feelings (hungry, thirsty, tired, etc.) will not be preserved creating a new type of an agent. People are mostly defined by their physiological needs (think of Maslow’s pyramid). An entity with no such needs (or with such needs satisfied by virtual/simulated abandoned resources) will not be human and will not want the same things as a human. Someone who is no longer subject to human weaknesses or relatively limited intelligence may lose all allegiances to humanity since they would no longer be a part of it. So I guess I define “humanity” as comprised on standard/unaltered humans. Anything superior is no longer a human to me, just like we are not first and foremost Neanderthals and only after homo sapiens.

Comment author: TheOtherDave 18 September 2013 12:43:45AM 0 points [-]

Well, yes, a lot depends on what we assume the upload includes, and how important the missing stuff is.
If Dave!upload doesn't include X1, and X2 defines Dave!original's humanity, and X1 contains X2, then Dave!upload isn't human... more or less tautologically.

We can certainly argue about whether our experiences of hunger, thirst, fatigue, etc. qualify as X1, X2, or both... or, more generally, whether anything does. I'm not nearly as confident as you sound about either of those things.

But I'm not sure that matters.

Let's posit for the sake of comity that there exists some set of experiences that qualify for X2. Maybe it's hunger, thirst, fatigue, etc. as you suggest. Maybe it's curiosity. Maybe it's boredom. Maybe human value is complex and X2 actually includes a carefully balanced brew of a thousand different things, many of which we don't have words for.

Whatever it is, if it's important to us that uploads be human, then we should design our uploads so that they have X2. Right?

But you seem to be taking it for granted that whatever X2 turns out to be, uploads won't experience X2.
Why?

Comment author: Roman_Yampolskiy 19 September 2013 04:12:03PM 0 points [-]

Just because you can experience something someone else can does not mean that you are of the same type. Belonging to a class of objects (ex. Humans) requires you to be one. A simulation of a piece of wood (visual texture, graphics, molecular structure, etc.) is not a piece of wood and so does not belong to the class of pieces of wood. A simulated piece of wood can experience simulated burning process or any other wood-suitable experience, but it is still not a piece of wood. Likewise a piece of software is by definition not a human being, it is at best a simulation of one.

Comment author: TheOtherDave 19 September 2013 04:42:41PM 0 points [-]

Ah.

So when you say "most typically human feelings (hungry, thirsty, tired, etc.) will not be preserved creating a new type of an agent" you're making a definitional claim that whatever the new agent experiences, it won't be a human feeling, because (being software) the agent definitionally won't be a human. So on your view it might experience hunger, thirst, fatigue, etc., or it might not, but if it does they won't be human hunger, thirst, fatigue, etc., merely simulated hunger, thirst, fatigue, etc.

Yes? Do I understand you now?

FWIW, I agree that there are definitions of "human being" and "software" by which a piece of software is definitionally not a human being, though I don't think those are useful definitions to be using when thinking about the behavior of software emulations of human beings. But I'm willing to use your definitions when talking to you.

You go on to say that this agent, not being human, will not want the same things as a human.
Well, OK; that follows from your definitions.

One obvious followup question is: would a reliable software simulation of a human, equipped with reliable software simulations of the attributes and experiences that define humanity (whatever those turn out to be; I labelled them X2 above), generate reliable software simulations of wanting what a human wants?

Relatedly, do we care? That is, given a choice between an upload U1 that reliably simulates wanting what a human wants, and an upload U2 that doesn't reliable simulate wanting what a human wants, do we have any grounds for preferring to create U1 over U2?

Because if it's important to us that uploads reliably simulate being human, then we should design our uploads so that they have reliable simulations of X2. Right?

Comment author: Roman_Yampolskiy 19 September 2013 06:39:44PM -1 points [-]

So uploads are typically not mortal, hungry for food, etc. You are asking if we create such exact simulations of humans that they will have all the typical limitations would they have the same wants as real humans, probably yes. The original question Wei Dai was asking me was about my statement that if we becomes uploads "At that point you already lost humanity by definition". Allow me to propose a simple thought experiment. We make simulated version of all humans and put them in cyberspace. At that point we proceed to kill all people. Does the fact that somewhere in the cyberspace there is still a piece of source code which wants the same things as I do makes a difference in this scenario? I still feel like humanity gets destroyed in this scenario, but you are free to disagree with my interpretation.

Comment author: CCC 19 September 2013 07:51:47PM 1 point [-]

We make simulated version of all humans and put them in cyberspace. At that point we proceed to kill all people.

At the very lesat, by this point we've killed a lot of people. the fact that they've been backed up doesn't make the murder less henious.

Whether or not 'humanity' gets destroyed in this scenario depends on the definition that you aply to the word 'humanity'. If you mean the flesh and blood, the meat and bone, then yes, it gets destroyed. If you mean values and opinions, thoughts and dreams, then some of them are destroyed but not all of them - the cyberspace backup still have those things (presuming that they're actually working cyberspace backups).

Comment author: TheOtherDave 19 September 2013 07:42:05PM 2 points [-]

You are asking if we create such exact simulations of humans that they will have all the typical limitations would they have the same wants as real humans, probably yes.

I'm also asking, should we care?
More generally, I'm asking what is it about real humans we should prefer to preserve, given the choice? What should we be willing to discard, given a reason?

The original question Wei Dai was asking me was about my statement that if we becomes uploads "At that point you already lost humanity by definition".

Fair enough. I've already agreed that this is true for the definitions you've chosen, so if that's really all you're talking about, then I guess there's nothing more to say. As I said before, I don't think those are useful definitions, and I don't use them myself.

Does the fact that somewhere in the cyberspace there is still a piece of source code which wants the same things as I do makes a difference in this scenario?

Source code? Maybe not; it depends on whether that code is ever compiled.
Object code? Yes, it makes a huge difference.

I still feel like humanity gets destroyed in this scenario, but you are free to disagree with my interpretation.

Some things get destroyed. Other things survive. Ultimately, the question in this scenario is how much do I value what we've lost, and how much do I value what we've gained?
My answer depends on the specifics of the simulation, and is based on what I value about humanity.

The thing is, I could ask precisely the same question about aging from 18 to 80. Some things are lost, other things are not. Does my 18-year-old self get destroyed in the process, or does it just transform into an 80-year-old? My answer depends on the specifics of the aging, and is based on what I value about my 18-year-old self.

We face these questions every day; they aren't some weird science-fiction consideration. And for the most part, we accept that as long as certain key attributes are preserved, we continue to exist.

Comment author: hairyfigment 19 September 2013 07:00:48PM 0 points [-]

Well, if nothing else happens our new computer substrate will stop working. But if we remove that problem - in what sense has this not already happened?

If you like, we can assume that Eliezer is wrong about that. In which case, I'll have to ask what you think is actually true, whether a smarter version of Aristotle could tell the difference by sitting in a dark room thinking about consciousness, and whether or not we should expect this to matter.

Comment author: shminux 19 September 2013 06:56:33PM *  1 point [-]

We make simulated version of all humans and put them in cyberspace. At that point we proceed to kill all people.

Ah, The Change in the Prime Intellect scenario. Is it possible to reconstruct meat humans if the uploads decide to do so? If not, then something has been irrecoverably lost.

Comment author: Nornagest 17 September 2013 08:44:11PM *  1 point [-]

Insofar as Maslow's pyramid accurately models human psychology (a point of which I have my doubts), I don't think the majority of people you're likely to be speaking to on the Internet are defined in terms of their low-level physiological needs. Food, shelter, physical security -- you might have fears of being deprived of these, or even might have experienced temporary deprivation of one or more (say, if you've experienced domestic violence, or fought in a war) but in the long run they're not likely to dominate your goals in the way they might for, say, a Clovis-era Alaskan hunter. We treat cases where they do as abnormal, and put a lot of money into therapy for them.

If we treat a modern, first-world, middle-class college student with no history of domestic or environmental violence as psychologically human, then, I don't see any reason why we shouldn't extend the same courtesy to an otherwise humanlike emulation whose simulated physiological needs are satisfied as a function of the emulation process.

Comment author: Roman_Yampolskiy 17 September 2013 10:21:30PM 2 points [-]

I don’t know you, but for me only a few hours a day is devoted to thinking or other non-physiological pursuits, the rest goes to sleeping, eating, drinking, Drinking, sex, physical exercise, etc. My goals are dominated by the need to acquire resources to support physiological needs of me and my family. You can extend any courtesy you want to anyone you want but you (human body) and a computer program (software) don’t have much in common as far as being from the same group is concerned. Software is not humanity; at best it is a partial simulation of one aspect of one person.

Comment author: Nornagest 17 September 2013 10:48:27PM *  2 points [-]

It seems to me that there are a couple of things going on here. I spend a reasonable amount of time (probably a couple of hours of conscious effort each day; I'm not sure how significant I want to call sleep) meeting immediate physical needs, but those don't factor much into my self-image or my long-term goals; I might spend an hour each day making and eating meals, but ensuring this isn't a matter of long-term planning nor a cherished marker of personhood for me. Looked at another way, there are people that can't eat or excrete normally because of one medical condition or another, but I don't see them as proportionally less human.

I do spend a lot of time gaining access to abstract resources that ultimately secure my physiological satisfaction, on the other hand, and that is tied closely into my self-image, but it's so far removed from its ultimate goal that I don't feel that cutting out, say, apartment rental and replacing it with a proportional bill for Amazon AWS cycles would have much effect on my thoughts or actions further up the chain, assuming my mental and emotional machinery remains otherwise constant. I simply don't think about the low-level logistics that much; it's not my job. And I'm a financially independent adult; I'd expect the college student in the grandparent to be thinking about them in the most abstract possible way, if at all.

Comment author: Bugmaster 17 September 2013 06:54:05PM 2 points [-]

Have you ever had the unfortunate experience of hanging out with really boring people; say, at a party ? The kind of people whose conversations are so vapid and repetitive that you can practically predict them verbatim in your head ? Were you ever tempted to make your excuses and duck out early ?

Now imagine that it's not a party, but the entire world; and you can't leave, because it's everywhere. Would you still "feel altruistic toward humanity" at that point ?

Comment author: TheOtherDave 17 September 2013 08:41:47PM 1 point [-]

It's easy to conflate uploads and augments, here, so let me try to be specific (though I am not Wei Dai and do not in any way speak for them).

I experience myself as preferring that people not suffer, for example, even if they are really boring people or otherwise not my cup of tea to socialize with. I can't see why that experience would change upon a substrate change, such as uploading. Basically the same thing goes for the other values/preferences I experience.

OTOH, I don't expect the values/preferences I experience to remain constant under intelligence augmentation, whatever the mechanism. But that's kind of true across the board. If you did some coherently specifiable thing that approximates the colloquial meaning of "doubled my intelligence" overnight, I suspect that within a few hours I would find myself experiencing a radically different (from my current perspective) set of values/preferences.

If instead of "doubling" you "multiplied by 10" I expect that within a few hours I would find myself experiencing an incomprehensible (from my current perspective) set of values/preferences.

Comment author: Bugmaster 17 September 2013 09:05:37PM *  -1 points [-]

It's easy to conflate uploads and augments, here...

Wait, why shouldn't they be conflated ? Granted, an upload does not necessarily have to possess augmented intelligence, but IMO most if not all of them would obtain it in practice.

I can't see why that experience would change upon a substrate change, such as uploading.

Agreed, though see above.

If you did some coherently specifiable thing that approximates the colloquial meaning of "doubled my intelligence" overnight, I suspect that within a few hours I would find myself experiencing a radically different (from my current perspective) set of values/preferences. If instead of "doubling" you "multiplied by 10" I expect that within a few hours I would find myself experiencing an incomprehensible (from my current perspective) set of values/preferences.

I agree completely; that was my point as well.

Edited to add:

I believe that, however incomprehensible one's new values might be after augmentation, I am reasonably certain that they would not include "an altruistic attitude toward humanity" (as per our current understanding of the term). By analogy, I personally neither love nor hate individual insects; they are too far beneath me.

Comment author: TheOtherDave 18 September 2013 12:24:54AM 1 point [-]

Mostly, I prefer not to conflate them because our shared understanding of upload is likely much better-specified than our shared understanding of augment.

I agree completely; that was my point as well.

Except that, as you say later, you have confidence about what those supposedly incomprehensible values would or wouldn't contain.

By analogy, I personally neither love nor hate individual insects; they are too far beneath me.

Turning that analogy around.... I suspect that if I remembered having been an insect and then later becoming a human being, and I believed that was a reliably repeatable process, both my emotional stance with respect to the intrinsic value of insect lives and my pragmatic stance with respect to their instrumental value would be radically different than they are now and far more strongly weighted in the insects' favor.

With respect to altruism and vast intelligence gulfs more generally... I dunno. Five-day-old infants are much stupider than I am, but I generally prefer that they not suffer. OTOH, it's only a mild preference; I don't really seem to care all that much about them in the abstract. OTGH, when made to think about them as specific individuals I end up caring a lot more than I can readily justify over a collection. OT4H, I see no reason to expect any of that to survive what we're calling "intelligence augmentation", as I don't actually think my cognitive design allows my values and my intelligence (ie my optimize-environment-for-my-values) to be separated cleanly. OT5H, there are things we might call "intelligence augmentation", like short-term-memory buffer-size increases, that might well be modular in this way.

Comment author: Bugmaster 18 September 2013 12:53:57AM 0 points [-]

Except that, as you say later, you have confidence about what those supposedly incomprehensible values would or wouldn't contain.

More specifically, I have confidence only about one specific thing that these values would not contain. I have no idea what the values would contain; this still renders them incomprehensible, as far as I'm concerned, since the potential search space is vast (if not infinite).

I suspect that if I remembered having been an insect and then later becoming a human being...

I am not entirely convinced that a vastly augmented mind would remember being a regular human in the same way that we humans remember what we had for lunch yesterday. The situation may be more analogous to remembering what it was like being a newborn.

Most people don't remember what being a newborn baby was like; but even if you could recall it with perfect clarity, how much of that information would you find really useful ? A newborn's senses are dull; his mind is mostly empty of anything but basic desires; his ability to affect the world is negligible. There's not much there that is even worth remembering... and, IMO, there's a good chance that a transhuman intelligence would feel the same way about its past humanity.

... and I believed that was a reliably repeatable process, both my emotional stance with respect to the intrinsic value of insect lives and my pragmatic stance with respect to their instrumental value would be radically different than they are now and far more strongly weighted in the insects' favor.

I agree with your later statement:

OT4H, I see no reason to expect any of that to survive what we're calling "intelligence augmentation", as I don't actually think my cognitive design allows my values and my intelligence (ie my optimize-environment-for-my-values) to be separated cleanly.

To expand upon it a bit:

I agree with you regarding the pragmatic stance, but disagree about the "intrinsic value" part. As an adult human, you care about babies primarily because you have a strong built-in evolutionary drive to do so. And yet, even that powerful drive is insufficient to overcome many people's minds; they choose to distance themselves from babies in general, and refuse to have any of their own, specifically. I am not convinced that an augmented human would retain such a built-in drive at all (only targeted at unaugmented humans instead/in addition to infants), and even if they did, I see no reason to believe that it would have a stronger hold over transhumans than over ordinary humans.

Comment author: TheOtherDave 18 September 2013 01:15:57AM 0 points [-]

Like you, I am unconvinced that a "sufficiently augmented" human would continue to value unaugmented humans, or infants.

Unlike you, I am also unconvinced it would cease to value unaugmented humans, or infants.

Similarly, I am unconvinced that it would continue to value its own existence, or, well, anything at all. It might turn out that all "sufficiently augmented" human minds promptly turn themselves off. It might turn out that they value unaugmented humans more than anything else in the universe. Or insects. Or protozoa. Or crystal lattices. Or the empty void of space. Or paperclips.

More generally, when I say I expect my augmented self's values to be incomprehensible to me, I actually mean it.

I am not entirely convinced that a vastly augmented mind would remember being a regular human in the same way that we humans remember what we had for lunch yesterday.

Mostly, I think that will depend on what kinds of augmentations we're talking about. But I don't think we can actually sustain this discussion with an answer to that question at any level more detailed than a handwavy notion of "vastly augmented" and analogies to insects and protozoa, so I'm content to posit either that it does, or that it doesn't, whichever suits you.

My own intuition, FWIW, is that some such minds will remember their true origins, and others won't, and others will remember entirely fictionalized accounts of their origins, and still others will combine those states in various ways.

There's not much there that is even worth remembering.

You keep talking like this, as though these kinds of value judgments were objective, or at least reliably intersubjective. It's not at all clear to me why. I am perfectly happy to take your word for it that you don't value anything about your hypothetical memories of infancy, but generalizing that to other minds seems unjustified.

For my own part... well, my mom is not a particularly valuable person, as people go. There's no reason you should choose to keep her alive, rather than someone else; she provides no pragmatic benefit relative to a randomly selected other person. Nevertheless, I would prefer that she continue to live, because she's my mom, and I value that about her.

My memories of my infancy might similarly not be particularly valuable as memories go; I agree. Nevertheless, I might prefer that I continue to remember them, because they're my memories of my infancy.

And then again, I might not. (Cf incomprehensible values of augments, above.)

Comment author: Moss_Piglet 17 September 2013 04:04:07PM *  1 point [-]

I've heard repeatedly that the correlation between IQ and achievement after about 120 (z = 1.33) is pretty weak, possibly even with diminishing returns up at the very top. Is moving to 250 (z = 10) passing a sort of threshold of intelligence at some point where this trend reverses? Or is the idea that IQ stops strongly predicting achievement above 120 wrong?

This is something I've been curious about for a while, so I would really appreciate your help clearing the issue up a bit.

Comment author: Lumifer 18 September 2013 08:33:18PM 2 points [-]

Is moving to 250 (z = 10) passing a sort of threshold of intelligence

First, IQ tests don't go to 250 :-) Generally speaking standard IQ tests have poor resolution in the tails -- they cannot reliably identify whether you have the IQ of, say, 170 or 190. At some point all you can say is something along the lines of "this person is in the top 0.1% of people we have tested" and leave it at that.

Second, "achievement" is a very fuzzy word. People mean very different things by it. And other than by money it's hard to measure.

Comment author: ESRogs 17 September 2013 06:20:04PM *  6 points [-]

In agreement with Vaniver's comment, there is evidence that differences in IQ well above 120 are predictive of success, especially in science. For example:

  • IQs of a sample of eminent scientists were much higher than the average for science PhDs (~160 vs ~130)

  • Among those who take the SAT at age 13, scorers in the top .1% end up outperforming the top 1% in terms of patents and scientific publications produced as adults

I don't think I have good information on whether these returns are diminishing, but we can at least say that they are not vanishing. There doesn't seem to be any point beyond which the correlation disappears.

Comment author: Moss_Piglet 18 September 2013 08:15:44PM 2 points [-]

I just read the "IQ's of eminent scientists" and realized I really need to get my IQ tested.

I've been relying on my younger brother's test (with the knowledge that older brothers tend to do slightly better but usually within an sd) to guesstimate my own IQ but a) it was probably a capped score like Feynman's since he took it in middle school and b) I have to know if there's a 95% chance of failure going into my field. I'd like to think I'm smart enough to be prominent, but it's irrational not to check first.

Thanks for the information; you might have just saved me a lot of trouble down the line, one way or the other.

Comment author: EHeller 19 September 2013 03:13:25PM 4 points [-]

I just read the "IQ's of eminent scientists" and realized I really need to get my IQ tested.

I'd be very careful generalizing from that study to the practice of science today. Science in the 1950s was VERY different, the length of time to the phd was shorter, postdocs were very rare, and almost everyone stepped into a research faculty position almost immediately.

In today's world, staying in science is much harder- there are lots of grad students competing for many postdocs competing for few permanent science positions. In today's world, things like conscientiousness, organization skills,etc (grant writing is now a huge part of the job) play a much larger role in eventually landing a job in the past, and luck is a much bigger driver (whether a given avenue of exploration pays off requires a lot of luck. Selecting people whose experiments ALWAYS work is just grabbing people who have been both good AND lucky). It would surprise me if the worsening science career hasn't changed the make up of an 'eminent scientist'.

Comment author: Moss_Piglet 19 September 2013 04:21:02PM 1 point [-]

At the same time, all of those points except the luck one could be presented as evidence that the IQ required to be eminent has increased rather than the converse. Grant writing and schmoozing are at least partially a function of verbal IQ, IQ in general strongly predicts academic success in grad school, and competition tends to winnow out the poor performers a lot more than the strong.

Not that I really disagree, I just don't see it as particularly persuasive.

whether a given avenue of exploration pays off requires a lot of luck. Selecting people whose experiments ALWAYS work is just grabbing people who have been both good AND lucky

That's just one of the unavoidable frustrations of human nature though; an experiment which dis-confirms it's hypothesis worked perfectly, it just isn't human nature to notice negatives.

Comment author: EHeller 19 September 2013 06:20:37PM *  2 points [-]

At the same time, all of those points except the luck one could be presented as evidence that the IQ required to be eminent has increased rather than the converse.

I disagree for several reasons. Mostly, conscientiousness, conformity,etc are personality traits that aren't strongly correlated with IQ (conscientiousness may even be slightly negatively correlated).

IQ in general strongly predicts academic success in grad school, and competition tends to winnow out the poor performers a lot more than the strong.

Would it surprise you to know that the most highly regarded grad students in my physics program all left physics? They had a great deal of success before and in grad school (I went to a top 5 program) , but left because they didn't want to deal with the administrative/grant stuff, and because they didn't want to spend years at low pay.

I'd argue that successful career in science is selecting for some threshhold IQ and then much more strongly for a personality type.

Comment author: Kawoomba 19 September 2013 06:22:31PM 0 points [-]

conscientiousness may even be slightly negatively correlated

No kidding.

Comment author: ESRogs 19 September 2013 05:55:19AM 1 point [-]

Are you American? If you've taken the SAT, you can get a pretty good estimate of your IQ here.

Comment author: Moss_Piglet 19 September 2013 03:59:43PM 1 point [-]

Mensa apparently doesn't consider the SAT to have a high-enough g loading to be useful as an intelligence test after 1994. Although the website's figure are certainly encouraging, it's probably best to take them with a bit of salt.

Comment author: ESRogs 20 September 2013 12:11:13AM *  1 point [-]

True, but note that, in contrast with Mensa, the Triple Nine Society continued to accept scores on tests taken up through 2005, though with a higher cutoff (of 1520) than on pre-1995 tests (1450).

Also, SAT scores in 2004 were found to have a correlation of about .8 with a battery of IQ tests, which I believe is on par with the correlations IQ tests have with each other. So the SAT really does seem to be an IQ test (and an extremely well-normed one at that if you consider their sample size, though perhaps not as highly g-loaded as the best, like Raven's).

But yeah, if you want to have high confidence in a score, probably taking additional tests would be the best bet. Here's a list of high-ceiling tests, though I don't know if any of them are particularly well-normed or validated.

Comment author: wedrifid 17 September 2013 05:10:28PM 1 point [-]

I've heard repeatedly that the correlation between IQ and achievement after about 120 (z = 1.33) is pretty weak, possibly even with diminishing returns up at the very top.

Is this what you intended to say? "Diminishing returns" seems to apply at the bottom the scale you mention. You've already selected the part where returns have started diminishing. Sometimes it is claimed that that at the extreme top the returns are negative. Is that what you mean?

Comment author: Moss_Piglet 18 September 2013 08:02:33PM 1 point [-]

Yeah, that's just me trying to do everything in one draft. Editing really is the better part of clear writing.

I meant something along the lines of "I've heard it has diminishing returns and potentially [, probably due to how it affects metabolic needs and rate of maturation] even negative returns at the high end."

Comment author: Vaniver 17 September 2013 04:36:00PM 3 points [-]

Or is the idea that IQ stops strongly predicting achievement above 120 wrong?

Most IQ tests are not very well calibrated above 120ish, because the number of people in the reference sample that scored much higher is rather low. It's also the case that achievement is a function of several different factors, which will probably become the limiting factor for most people at IQs higher than 120. That said, it does seem that in physics, first-tier physicists score better on cognitive tests than second-tier physicists, which suggests that additional IQ is still useful for achievement in the most cognitively demanding fields. It seems likely that augmented humans who do several times better than current humans on cognitive tests will also be able to achieve several times as much in cognitively demanding fields.

Comment author: shminux 16 September 2013 11:29:03PM *  1 point [-]

Nice to see more AI experts here.

Comment author: lukeprog 16 September 2013 11:07:04PM 1 point [-]

Note also that Roman co-authored 3 of the papers on MIRI's publications page.

Comment author: shminux 16 September 2013 11:35:09PM *  -1 points [-]

His paper http://cecs.louisville.edu/ry/LeakproofingtheSingularity.pdf seriously discusses ways to confine a potentially hostile superintelligence, a feat MIRI seems to consider hopeless. Did you guys have a good chat about it?

Comment author: CarlShulman 17 September 2013 12:10:07AM *  6 points [-]

I think most everyone at MIRI and FHI thinks boxing is a good thing, even if many would say not enough on its own. I don't think you will find many who think that open internet connections are a matter of indifference for AI developers working with powerful AGI.

Comment author: Eliezer_Yudkowsky 17 September 2013 10:28:55PM 6 points [-]

High-grade common sense (the sort you'd get by asking any specialist in computer security) says that you should design an AI which you would trust with an open Internet connection, then put it in the box you would use on an untrusted AI during development. (No, the AI will not be angered by this lack of trust and resent you. Thank you for asking.) I think it's safe to say that for basically everything in FAI strategy (I can't think of an exception right now) you can identify at least two things supporting any key point, such that either alone was designed to be sufficient independently of the other's failing, including things like "indirect normativity works" (you try to build in at least some human checks around this which would shut down any scary AI independently of your theory of indirect normativity being remotely correct, while also not trusting the humans to steer the AI because then the humans are your single point of failure).

Comment author: lukeprog 16 September 2013 11:43:08PM 3 points [-]

See my interview with Roman here.

Comment author: shminux 17 September 2013 06:39:14AM -1 points [-]

Thanks. Pretty depressing, though.

Comment author: David_Chapman 14 September 2013 06:04:59PM 13 points [-]

Hi!

I’ve been interested in how to think well since early childhood. When I was about ten, I read a book about cybernetics. (This was in the Oligocene, when “cybernetics” had only recently gone extinct.) It gave simple introductions to probability theory, game theory, information theory, boolean switching logic, control theory, and neural networks. This was definitely the coolest stuff ever.

I went on to MIT, and got an undergraduate degree in math, specializing in mathematical logic and the theory of computation—fields that grew out of philosophical investigations of rationality.

Then I did a PhD at the MIT AI Lab, continuing my interest in what thinking is. My work there seems to have been turned into a surrealistic novel by Ken Wilber, a woo-ish pop philosopher. Along the way, I studied a variety of other fields that give diverse insights into thinking, ranging from developmental psychology to ethnomethodology to existential phenomenology.

I became aware of LW gradually over the past few years, mainly through mentions by people I follow on Twitter. As a lurker, there’s a lot about the LW community I’ve loved. On the other hand, I think some fundamental, generally-accepted ideas here are limited and misleading. I began considering writing about that recently, and posted some musings about whether and how it might be useful to address these misconceptions. (This was perhaps ruder than it ought to have been.) It prompted a reply post from Yvain, and much discussion on both his site and mine.

I followed that up with a more constructive post on aspects of how to think well that LW generally overlooks. In comments on that post, several frequent LW contributors encouraged me to re-post that material here. I may yet do that!

For now, though, I’ve started a sequence of LW articles on the difference between uncertainty and probability. Missing this distinction seems to underlie many of the ways I find LW thinking limited. Currently my outline for the sequence has seven articles, covering technical explanations of this difference, with various illustrations; the consequences of overlooking the distinction; and ways of dealing with uncertainty when probability theory is unhelpful.

(Kaj Sotala has suggested that I ask for upvotes on this self-introduction, so I can accumulate enough karma to move the articles from Discussion to Main. I wouldn’t have thought to ask that myself, but he seems to know what he’s doing here! :-)

O&BTW, I also write about contemporary trends in Buddhism, on several web sites, including a serial, philosophical, tantric Buddhist vampire romance novel.

Comment author: Rapses 31 August 2013 02:57:11PM *  0 points [-]

Hi All LessWrongers, My name is Rupesh and I have PhD in mathematics. I been lurking here for a long time. The posts are of really very high quality. After visiting here for a while, I realised that rationality is not something you do just 9 to 5 at work, it must seep into your whole lifestyle.

Comment author: ShiraKarasu 19 August 2013 06:58:53PM 1 point [-]

Hello then.

I am a political science and international development undergrad student, residing mainly in Vienna, Austria. The story of how I came here is probably a rather common one - it started on TvTropes, where I am an on-off forum contributor and editor, where I first heard of Harry Potter and the Methods of Rationality. After reading it, I decided to look further into the rationalist community, partly because of my interest for philosophy, ethics, politics and debating, but also hoping to find novel, intelligent and helpful approaches to several key questions I have been struggling with for a while.

I hope to be able to contribute soon in an efficient and constructive way - even though I have a lot of catching up to do. There's probably going to be a bit of an archive panic moment, what with having the sequences to finish. Lots to learn, and eager to do so. See you around!

Comment author: Kawoomba 19 August 2013 07:07:44PM 1 point [-]

several key questions I have been struggling with for a while

Which are those?

Comment author: Serendipity 04 August 2013 10:39:45AM *  14 points [-]

Hi everyone, my name is Sara!

I am 21, live in Switzerland and study psychology. I am fascinated with the field of rationality and therefore wrote my Bachelor thesis on why and how critical thinking should be taught in schools. I started out with the plan to get my degree in clinical- and neuropsychology but will now change to developmental psychology for I was able to fascinate my supervising tutor and secure his full support. This will allow me to base my Master project on the development and enhancing of critical thinking and rationality, too. Do you have any recommendations?

After my Master's degree I still intend on getting an education as therapist (money reasons) or going into research (pushing the experimental research on rationality) and on giving a lot of money to the most effective charities around. I wonder if as therapist it would be smarter to concentrate on children or adults; both fields will be open for me after my university education (which will take me about 2.5-3 more years). I speak German, Swiss German, Italian, French and English (and understand some more languages), which will give me some freedom in the choice where to actually work in future.

...but I'm not only looking for advice here. I'm (mainly) interested in educating myself (and possibly other people around me). In fact, I am part of a Swiss group that translates less wrong articles into German, making the content available for more people in our surroundings (Switzerland, Germany, Austria).

I've learned a lot from this community and it has strongly shaped who I have become. There's no way I'd want to go back to my even more biased past self :)

Indeed, I am looking forward to learning more!

Comment author: Tenoke 04 August 2013 11:55:19AM 0 points [-]

Hello, Sara.

This will allow me to base my Master project on the development and enhancing of critical thinking and rationality,

Do you have any specific ideas for this. Are you aiming at enhancing rationality in adults or children? I don't have specific recommendations except perhaps people whose work is relevant, however, you would have encountered those around the site.

P.S. I am mainly commenting here because this is the second time I see you on the internet within the last 4 hours.

Comment author: Serendipity 04 August 2013 12:08:02PM 1 point [-]

Hello, Tenoke.

I am aiming on enhancing rationality in children but indeed had to often fall back on research with older people. Until now I've been concentrating on the work of Stanovich, Facione, van Gelder and Twardy. Whose work do you think would be relevant, too?

Thank you for your answer!

Comment author: Tenoke 04 August 2013 09:53:06PM *  0 points [-]

Well, Kahneman (and Tversky) would be the most obvious example out of those not mentioned. Otherwise Dennet, Gilovich, Slovic, Pinker, Taleb and Thaler would be some examples of people whose work has varying degrees of relevance to the subject. Those are the people who I can think of off the top of my head but the best way to systematically find researchers of interest would be to look at the reverse citations of Kahneman and Tversky's work or something of the sort.

Comment author: Serendipity 05 August 2013 04:11:53PM 0 points [-]

Ah, how could I forget them! Biases and heuristics play a big role in my interests for critical thinking of course. I'm a bit surprised: how come you included Dennett and Pinker? I know these two for work that's (very interesting but) mostly unrelated to my addressed topic. I'm curious, seems like I missed something important.

Comment author: Tenoke 06 August 2013 12:09:44PM 0 points [-]

I was writing on auto-pilot, you are right that their work is significantly less relevant to the topic than the others'.

Comment author: vollmer 02 August 2013 04:37:33AM 5 points [-]

I'm a Swiss medical student. I've read HPMoR and a large part of the core sequences. I've attended LW meetups in several US cities and met quite a few of you in the Bay Area and/or at the Effective Altruism Summit. I've interned for Leverage Research. I co-founded giordano-bruno-stiftung.ch (outreach organisation with German translations of some LessWrong blog posts, and other posts about rationality). Looking forward to participating in the comment section more often.

Comment author: raydpratt 27 July 2013 10:28:37PM 3 points [-]

I am a maximum-security ex-con who studied and used logic for pro se, civil-rights lawsuits. (The importance of being a maximum-security ex-con is that I was stubborn iconoclast who learned and used logic in all seriousness.) Logic helped me identify the weak links in my opponent's arguments and to avoid weak links in my own arguments, and logic helped my organize my writing and evidence. I also studied and learned to use “The Option Process” for eliminating my negative emotions and to understand other people's negative emotions. The core truth of “The Option Process” is that we choose to have negative emotions for reasons, not randomly, and not even necessarily. So, our rationality is very much a part of our emotions, and, as such, good reasoning can utterly remove negative emotions at the core of their raison d'être. However, some of my emotional and intellectual challenges have resisted solutions via logic and “The Option Process.” For example, I could not figure out how to stay objective and to behave objectively while trying to gamble for profit (not for fun). So, I began reading widely about self-control, discipline, integrity, neuroeconomics, etc. And, in the process, I found this LessWrong website.

I have only recently identified what may be at the root of my problem with gambling and why it resists both logic and “The Option Process.” Freud called it “childhood megalomania.” In our early years, whenever we cried and sniveled, the universe of Mom and Dad and others rushed to our needs. That inner baby rarely grows up well in any of us, and we still whine, snivel, and howl at the universe when things don't go our way, and we can get down right obstinate about doing so until the universe listens! The universe, in turn, responds favorably often enough to keep our inner babies convinced of our magic, temper-tantrum powers over reality.

I figured out that when I get frustrated, afraid, and challenged by the difficulties of gambling, I would rather feel safe, powerful and warm, and so I often lapse into an obstinate insistence on continuing to gamble because I want to believe and feel that I can successfully gamble whenever I want, even during objectively bad, fear-inducing, and frustrating conditions.

The universe has not been kind in that regard, but with my recent insight, I at least hope that my inner baby has grown one year older. The rest of the problem, the frustration and fear, will easily fall prey to the power of logic and “The Option Process.”

Comment author: pashakun 26 July 2013 05:13:31AM 2 points [-]

I'm Pasha, a financial journalist based in Tokyo.

I recently found out about this blog from this post on The View From Hell: http://goo.gl/DCNX4U

A few years in a school specialized in math and physics in the former Soviet Union have convinced me to seek my fortunes in liberal arts. (It's those kids in my class who would yell out an answer to a physics problem even before the teacher has finished reading the question.)

Covering the semiconductor industry here in Japan has sparked a renewed appreciation of the scientific method and revived my interest in rationality, math and computation. ... One thing leads to another and here I am ~

Comment author: djm 25 July 2013 11:27:06AM 5 points [-]

Hello, I am a 46 yr old software developer from Australia with a keen interest in Artificial Intelligence.

I don’t have any formal qualifications, which is a shame as my ideal life would be to do full time research in AI - without a PhD I realise this won’t happen, so I am learning as much as I can through books, practice and various online courses.

I came across this site today from a link via MIRI and feel like I have struck gold - the articles, sequences and discussions here are very well written, interesting and thoughtful.

My current goals are to build a framework that would allow a machine to manage its information (goals, tasks, raw data, external biases, weightings, and eventually its “knowledge”). As I understand it the last bit hasn’t been solved yet as it implies the machine needs a consciousness, but I am having fun playing around with it.

Comment author: dirtfruit 24 July 2013 09:25:17PM *  3 points [-]

Hey, I'm dirtfruit.

I've lurked here for quite a while now. LessWrong is one of the most interesting internet communities I've observed, and I'd like to begin involving myself more actively. I've been to one meetup, in NYC, a few months ago, which was nice. I've read most of the sequences (I think I've read all of them at least once, but I haven't looked hard enough to be super-confident saying that). HPMOR is cool, I enjoyed reading it and continue to check for updates. I've tried to read most of what Eliezer has written, but gave up early on anything extremely technical, as I don't have the background for it. EY seems like a righteous dude to me. I dig his cause, and would like to make myself available to help, in what ways I can.

I'm currently 21 years old. I was born and raised on the west coast of the united states, and am now attending a college on the east coast studying fine art, with a concentration in drawing. I've always read a lot. When I was young; analog fiction, mostly. Now I most often find myself reading nonfiction online .

I'd like to find ways for artists (specifically me, but also other interested artists(to a lesser degree)) to be useful to the general cause of rationality; raising waterlines and whatnot. I believe there exists a general feeling among lesswrong users that artists can be fun, but are not very instrumentally useful to their particular cause. If this belief is misplaced, I'd be overjoyed to adjust it properly. I'm obviously biased, but I believe this feeling to be more than a few shades off from correct. Pictorial communication can be super intuitive. It can communicate very quickly relative to the written word, can be very memorable, and is capable of transcending many written/spoken language barriers. It's main downsides include: time-expense (drawing a picture generally takes longer than describing something verbally(spoken or written)); and scarcity of expertise - drawing and painting's difficulty curves seems roughly similar to that of writing, but they are practiced far less often than writing, and (nowadays, in the fine art world at least) held to very different standards. Experts in visual communication should be very instrumentally useful, for clarifying concepts not well suited to words, and also for attracting/aiding/communicating with those beyond the reach of literacy. I'm not claiming expertise (I'm still building my skills as a student), but at the very least I have some experience in crafting understandable, detailed pictures to something of a high standard. I'm also somewhat talented with words; integrating textual communication with visual communication (and visa versa) is something I'm sensitive to and interested in.

I also just really like the spirit and conventions of debate here, and would very much like to hear any and all thoughts about what I just wrote. :D thanks!

(also I think we need a new welcome thread? either that or I failed to find the proper one. This thread has far exceeded 500 posts...)

Comment author: Heraclitus 21 July 2013 04:29:25AM *  3 points [-]

So: Here goes. I'm dipping my toe into this gigantic and somewhat scary pool/lake(/ocean?).

Here's the deal: I'm a recovering irrationalic. Not an irrationalist; I've never believed in anything but rationalism (in the sense it's used here, but that's another discussion), formally. But my behaviors and attitudes have been stuck in an irrational quagmire for years. Perhaps decades, depending on exactly how you're measuring. So I use "irrationalic" in the sense of "alcoholic"; someone who self-identifies as "alcoholic" is very unlikely to extol the virtues of alcohol, but nonetheless has a hard time staying away from the stuff.

And, like many alcoholics, I have a gut feeling that going "cold turkey" is a very bad idea. Not, in this case, in the sense that I want to continue being specifically irrational to some degree or another, but in that I am extremely wary of diving into the list of readings and immersing myself in rationalist literature and ideology (if that is the correct word) at this point. I have a feeling that I need to work some things out slowly, and I have learned from long and painful experience that my gut is always right on this particular kind of issue.

This does not mean that linking to suggested resources is in any way not okay, just that I'm going to take my time about reading them, and I suppose I'm making a weak (in a technical sense) request to be gentle at first. Yes, in principle, all of my premises are questionable; that's what rationalism means (in part). But...think about it as if you had a new, half-developed idea. If you tell it to people who tear it apart, that can kill it. That's kind of how I feel now. I'm feeling out this new(ish) way of being, and I don't feel like being pushed just yet (which people who know me might find quite rich; I'm a champion arguer).

Yes, this is personal, more personal than I am at all comfortable being in public. But if this community is anything like I imagine it to be (not that I don't have experience with foiled expectations!), I figure I'll probably end up divulging a lot more personal stuff anyway.

I honestly feel as if I'm walking into church for the first time in decades.

So why am I here then? Well, I was updating my long-dormant blog by fixing dead links &c, and in doing so, discovered to my joy that Memepool was no longer dead. There, I found a link to HPMOR. Reading this over the next several days contributed to my reawakening, along with other, more personal happenings. This is a journey of recovery I've been on for, depending on how you count, three to six years, but HPMOR certainly gave a significant boost to the process, and today (also for personal reasons) I feel that I've crossed a threshold, and feel comfortable "walking into church" again.

Alright, I'll anticipate the first question: "What are you talking about? Irrationality is an extremely broad label." Well, I'm not going to go into to too terribly much detail just now, but let's say that the revelation or step forward that occurred today was realizing that the extremely common belief that other people can make you morally wrong by their judgement is unequivocally false. This (that this premise is false) is what I strongly believed growing up, but...well, perhaps "strongly" is the wrong word. I had been raised in an environment that very much held that the opposite was true, that other people's opinion of you was crucial to your rightness, morality and worth as a human being. Nobody ever said it that way, of course, and would probably deny it if put that way, but that is nonetheless how most people believe. However, in my case it was so blatant that it was fairly easy to see how ridiculous it was. Nonetheless, as reasonable as my rational constructions seemed to me, there was really no way I could be certain that I was right and others were wrong, so I held a back-of-my-head belief, borne of the experience of being repeatedly mistaken that every inquisitive child experiences, that I would someday mature and come to realize I had been wrong all along.

Well, that happened. Sort of. Events in my life picked at that point of uncertainty, and I gave up my visceral devotion to rationality and personal responsibility, which led slowly down into an awful abyss that I'm not going to describe at just this moment, that I have (hopefully) at last managed to climb out of, and am now standing at the edge, blinking at the sunlight, trying to figure out precisely where to go from here, but wary of being blinded by the newfound brilliance and wishing to take my time to figure out the next step.

So again, then, why am I here? If I don't want to be bombarded with advice on how to think more rationally, why did I walk in here? I'm not sure. It seemed time, time to connect with people who, perhaps, could support me in this journey, and possibly shorten it somewhat.

I also notice that this thread has gone waaay beyond 500 comments; perhaps someone with more Karma than I can make a new Welcome thread?

Comment author: Heraclitus 21 July 2013 04:48:32AM *  0 points [-]

So since I wrote this five minutes ago, I've gotten some insights (through looking at one of the links on the welcome page above) into why I'm so wary of being bombarded with arguments explaining how to be rational. Hopefully commenting on my own comment won't discourage others from doing so.

I'm not wary because I'm afraid my newfound insight is going to be damaged somehow; quite the contrary. I'm wary because I strongly fear that all these rationalist arguments will be very seductive. However, I've tried very hard my whole life (with varying degrees of success) to make sure my thoughts and ideas were my own, and, having so recently stepped back into the light, I fear I might be very susceptible to rationalist arguments. "But that's a good thing," you might say, "because it's rationalism!" (or rather, some more complicated and convincing formulation). Well, sure, but that doesn't make any specific rationalist argument certain to be right, and I'm not sure I feel competent to evaluate the truth of claims that sound very good and I really want to believe right now.

Comment author: satt 21 July 2013 11:35:48AM 5 points [-]

So since I wrote this five minutes ago, I've gotten some insights (through looking at one of the links on the welcome page above) into why I'm so wary of being bombarded with arguments explaining how to be rational.

You might like a couple of pieces that take a similarly positive-but-tempered view of LW-style rationality (both written by the person — Yvain — who wrote the piece at your link, as it happens): "Extreme Rationality: It's Not That Great" and "Epistemic learned helplessness". You might also like Yvain's other LW posts, most of which work as standalone pieces and are worth reading.

Comment author: Heraclitus 24 July 2013 12:36:16AM 0 points [-]

Wow. Thank you. I just finished "Epistemic Learned Helplessness," and I feel much better now. Those two articles have successfully inoculated me against being sucked in too easily into the "x-rationalist" view.

I actually disagree with what he says in "Epistemic Learned Helplessness"; or rather, I don't believe that that helplessness is actually necessary, that I can--or if I can't, it is possible to with sufficient training--tell when a case has been reasonably proven and when I should suspend judgement. Or maybe he's more right than I like to admit; I have to concede that I was taken in by much of Graham Hancock's work until I tried to write a short story based on one of his ideas and it completely fell apart after some research and analysis. But regardless of whether the dilemma he poses is avoidable or not, he makes some excellent, indeed critical, points, and I can now proceed with a healthy dose of skepticism of rationalism, a phrase I would likely have been ashamed to utter before reading that article.

Comment author: Heraclitus 23 July 2013 10:32:15PM -1 points [-]

Okay, I've read the first article you linked, and I'm discovering that I was naive about what this site was about (this should not be surprising after all the times similar things have happened to me, but it apparently still is). I've read HPMOR, of course, but I didn't catch on that this site would be specifically geared to using specific, formal, scientifically-derived techniques to improve thinking. The article mentioned Scientology; this kind of sounds a little like Scientology (well, Dianetics) to me, though I'm sure it makes much more formal sense. This makes me still more wary than before; I like my own "organic" rationalist methods, and am skittish of adopting some formal "system" of thought. This is more grousing than complaint; I do not have enough information to intelligently critique at this point, although the thing that bothered me most about Harry was his overuse of formal techniques instead of just trying to grok the whole situation in a more organic fashion; that just seems like a good way to miss something. This does not mean that reading about common errors in thinking couldn't be useful.

I'm disappointed that my post didn't receive more response (poor me! I want attention! Well, alright, I was hoping for something analogous to a support group), but I appreciate yours. I'll definitely keep reading.

Comment author: satt 29 July 2013 01:43:04AM 0 points [-]

the thing that bothered me most about Harry was his overuse of formal techniques instead of just trying to grok the whole situation in a more organic fashion; that just seems like a good way to miss something.

I can't speak to how well this works out for Harry (I haven't read HPMoR) but I think I can guess why this bites people in real life.

The methods that work for someone tend to be the ones they're already familiar with. Why? At least two reasons. The boring one is that people are less likely to stick with methods that obviously don't work, so obviously bad methods get forgotten about and become unfamiliar again. The more interesting reason is that using a method makes it "better": practice allows you to apply it more quickly when it's relevant, you learn to recognize more quickly the situations where the method's relevant, and you get better at integrating what you learn from that method with your other thoughts.

This is why it can be safer to organically accrete a system of thinking piece by piece than to install a fully-fledged system in one go; you only have to keep one piece in your head at a time, and you can focus on that one piece for a while until you're used to it and can apply it without much conscious effort. By contrast, trying to take on a complete system in one go means you're constantly having to think hard about which parts of it are relevant to each problem you confront. It's the difference between seeing a loose screw sticking out of something and knowing you need a screwdriver to tighten it, and seeing a loose screw sticking out of something and emptying your toolbox on the floor so you can try each tool one-by-one.

The important distinction isn't so much between formal methods and organic methods, but between methods you've fully internalized and methods you haven't. A formal method that's permanently imprinted into your mind through practice is likely to be quicker to use, easier to use, and more effective than an informal method you've only just heard about. Eventually, if you practice a technique enough, formal or not, there's a good chance your brain will automatically reach out and apply it in the normal course of grokking a whole situation organically. (For example, if I need to predict or reason about some recurrent event in my life, I often automatically apply reference class forecasting without much thought, and I readily integrate that information with any other information I can glean about the event.)

So I think it makes sense to take this stuff at whatever pace feels comfortable. Certainly, when I first landed on LW, I didn't shoot off and read all of the sequences of core posts in one go. I just clicked around, read recent discussions, and when people referred to individual posts in the sequences while discussing other things, I'd click through and read the post they linked to. (And then if I felt like reading more, I'd look at the other posts linked by that post!)

I'll definitely keep reading.

Enjoy the site!

Comment author: [deleted] 21 July 2013 04:01:49AM *  0 points [-]

So: Here goes. I'm dipping my toe into this gigantic and somewhat scary pool/lake(/ocean?).

Here's the deal: I'm a recovering irrationalic. Not an irrationalist; I've never believed in anything but rationalism (in the sense it's used here, but that's another discussion), formally. But my behaviors and attitudes have been stuck in an irrational quagmire for years. Perhaps decades, depending on exactly how you're measuring. So I use "irrationalic" in the sense of "alcoholic"; someone who self-identifies as "alcoholic" is very unlikely to extol the virtues of alcohol, but nonetheless has a hard time staying away from the stuff.

And, like many alcoholics, I have a gut feeling that going "cold turkey" is a very bad idea. Not, in this case, in the sense that I want to continue being specifically irrational to some degree or another, but in that I am extremely wary of diving into the list of readings and immersing myself in rationalist literature and ideology (if that is the correct word) at this point. I have a feeling that I need to work some things out slowly, and I have learned from long and painful experience that my gut is always right on this particular kind of issue.

This does not mean that linking to suggested resources is in any way not okay, just that I'm going to take my time about reading them, and I suppose I'm making a weak (in a technical sense) request to be gentle at first. Yes, in principle, all of my premises are questionable; that's what rationalism means (in part). But...think about it as if you had a new, half-developed idea. If you tell it to people who tear it apart, that can kill it. That's kind of how I feel now. I'm feeling out this new(ish) way of being, and I don't feel like being pushed just yet (which people who know me might find quite rich; I'm a champion arguer).

Yes, this is personal, more personal than I am at all comfortable being in public. But if this community is anything like I imagine it to be (not that I don't have experience with foiled expectations!), I figure I'll probably end up divulging a lot more personal stuff anyway.

I honestly feel as if I'm walking into church for the first time in decades.

So why am I here then? Well, I was updating my long-dormant blog by fixing dead links &c, and in doing so, discovered to my joy that Memepool was no longer dead. There, I found a link to HPMOR. Reading this over the next several days contributed to my reawakening, along with other, more personal happenings. This is a journey of recovery I've been on for, depending on how you count, three to six years, but HPMOR certainly gave a significant boost to the process, and today (also for personal reasons) I feel that I've crossed a threshold, and feel comfortable "walking into church" again.

Alright, I'll anticipate the first question: "What are you talking about? Irrationality is an extremely broad label." Well, I'm not going to go into to too terribly much detail just now, but let's say that the revelation or step forward that occurred today was realizing that the extremely common belief that other people can make you morally wrong by their judgement is unequivocally false. This is what I strongly believed growing up, but...well, perhaps "strongly" is the wrong word. I had been raised in an environment that very much held that the opposite was true, that other people's opinion of you was crucial to your rightness, morality and worth as a human being. Nobody ever said it that way, of course, and would probably deny it if put that way, but that is nonetheless how most people believe. However, in my case it was so blatant that it was fairly easy to see how ridiculous it was. Nonetheless, as reasonable as my rational constructions seemed to me, there was really no way I could be certain that I was right and others were wrong, so I held a back-of-my-head belief, borne of the experience of being repeatedly mistaken that every inquisitive child experiences, that I would someday mature and come to realize I had been wrong all along.

Well, that happened. Sort of. Events in my life picked at that point of uncertainty, and I gave up my utter visceral devotion to rationality and personal responsibility, which led slowly down into an awful abyss that I'm not going to describe at just this moment, that I have (hopefully) at last managed to climb out of, and am now standing at the edge, blinking at the sunlight, trying to figure out precisely where to go from here, but wary of being blinded by the newfound brilliance and wishing to take my time to figure out the next step.

So again, then, why am I here? If I don't want to be bombarded with advice on how to think more rationally, why did I walk in here? I'm not sure. It seemed time, time to connect with people who, perhaps, could support me in this journey, and possibly shorten it somewhat.

Comment author: Kendra 19 July 2013 07:54:39PM 10 points [-]

Hi, I'm Denise from Germany, I just turned 19 and study maths at university. Right now, I spend most of my time with that and caring for my 3-year-old daughter. I know LessWrong for almost two years now, but never got around to write. However, I'm more or less involved with parts of the LessWrong and the Effective Altruism community, most of them originally found me via Okcupid (I stated I was a LessWrongian), and from there, it expanded.

I grew up in a small village in the middle of nowhere in Germany, very isolated without any people to talk to. I skipped a grade and did extremely well at school, but was mostly very unhappy during my childhood/teen years. Though I had free internet access, I had almost no access to education until I was 15 years old (and pregnant, and no, that wasn't unplanned), because I had no idea what to look for. I dropped out of school then and prepared for the exams -when I had time (I was mostly busy with my child)- I needed to do to be allowed to attend university. In Germany that's extremely unusual and most people don't even know you can do it without going to school.

When I was 15, I discovered enviromentalism (during pregnancy, via people who share my parenting values) and feminism. Since then, I seriously cared about making the world „a better place“. I was already very nerdy in my special fields of interest then, though still very uneducated and lacking basic concepts. Thankfully, I found LessWrong when I was just 17 and became very taken with it. I started to question my beliefs, became a utilitarian, adopted a somewhat transhumanist mindset and the usual, but the breakthrough only came last year after I started spending time with people from the community. Since then I am totally focused. Most people who have met me this year or at the end of 2012 are very surprised by this, I noticed that a lot of people completely overestimate my past selves (which is somewhat relieving, though I still feel like everyone from the LW/EA who is usually quite taken with me overestimates me). Until the beginning of this year, I even considered enviromentalism the most important problem (which is completely ridiculous for me now). Well, I had been a serious enviromentalist for three years, then I talked half an hour with another LessWrongian about it, who explained to me why it isn't the most important problem, so I dropped it on the same day. After thinking about it myself and talking to several LW/EAs (e.g. 80,000hours) I decided it's the best thing for me to study maths (my minor will be in computer science). People always tell me I worry too much about my future and I am already at a very good position, being so driven, etc. but I often think I have lost so many years now and there is so much to read and so much I don't know and so little time. Especially considering that I lose about 70% of my time awake to caring for my daughter (which people do never take into account at all. They just have no idea. Before last October, it was even 90%). I often felt extremely incompetent and lazy because other people get so much done in comparison to me. Well, I do feel a bit better after actually thinking about how big my disadvantages are, but it's still quite bad. Several people have asked me to consider internships, etc., but I mostly still feel too incompetent, and the even bigger problem, too socially awkward.

Rationality was very helpful in the past with personal problems (e.g., I have a very static mindset, which hasn't really been a problem so far because I always was able to do things despite of it, without having to work for them, but now, doing my maths degree, it doesn't work as well as in the past) and has heavily reduced them, though enough still remain. My productivity has increased a lot. There are a lot of things to do waiting for me, I can't afford losing time to personal inconveniences. (Though anyway most of my time and energy goes into my child and there isn't really much I can do about that.)

I'm very happy that I found LessWrong and like-minded people. If you have reading recommendations, please tell me. I am familiar with all the basic material (the Sequences, of course, the EA stuff, the self-improvement stuff, Bostrom's work, Kahneman...). If you have any other advice, I would also love to hear it.

Comment author: Gunnar_Zarncke 22 January 2014 10:49:58PM 1 point [-]

Hi Denise/Kendra,

sich um ein kleines Kind alleine zu kümmern ist schon viel. Wenn Du dann auch noch studierst und EA und LW Meetups machst ist das schon ziemlich viel. Ich bewundere Deine Leistung. Ich habe einiges Material zu rationaler Erziehung auf meiner Homepage verlinkt, das Du Dir evtl. mal ansehen möchtest: http://lesswrong.com/user/Gunnar_Zarncke

Ein Tipp (obwohl Du vermutlich weißt und nur nicht umsetzen konntest): Die Synergieeffekte bei der Kindererziehung sind beträchtlich. Es ist erheblich einfacher für zwei Eltern für zwei Kinder zu sorgen als 2x alleinerziehend mit Kind. Entsprechend in größeren Gruppen (das sieht man natürlich meist nur wenn sich mehrere Familien treffen). Hast Du keine Möglichkeit das zu nutzen?

Du darfst mir gerne jederzeit Fragen stellen.

Gruß aus Hamburg

Gunnar

Comment author: Creutzer 22 January 2014 11:04:34PM -1 points [-]

I don't think it's appropriate to write a comment in a language other than English. You could have sent a PM.

Comment author: Gunnar_Zarncke 22 January 2014 11:05:32PM *  1 point [-]

A yes. You are probably right.

How comes you did notice so quickly this being such an old thread?

Comment author: blacktrance 22 January 2014 11:15:11PM 0 points [-]

There's a Recent Comments column to the right.

Comment author: vollmer 02 August 2013 04:32:17AM 1 point [-]

Welcome Denise! :)

Comment author: [deleted] 20 July 2013 11:21:22PM 2 points [-]

„a better place“

<nitpick> It isn't customary that kind of quotation marks in English; “these ones” are usually used in typeset materials, but most people just use "the ones on the keyboard" on-line.

Comment author: Kawoomba 19 July 2013 08:10:20PM 3 points [-]

As another LW'er with kids in Germany, welcome!

Comment author: Tsende 19 July 2013 12:35:22AM 3 points [-]

Hi, I'm a second year engineering student at a university of California. I like engaging in rational discussions and find importance in knowing about what's going on in the world and gain more insight on controversial issues such as abortion, gay rights, sexuality, immigration, etc. Someone on Facebook directed me to this site but I easily get bored so I may or may not be much of a contribution.

Comment author: Viliam_Bur 18 September 2013 12:37:33PM 0 points [-]

gain more insight on controversial issues

It is probably better to practice rationality skills on less controversial issues. When speaking about politics, people instinctively become less rational, because politics is usually not about being correct, but about belonging to the winning tribe.

Comment author: Brendon_Wong 18 July 2013 06:21:56PM *  6 points [-]

Hello! I’m a 15 year old sophomore in high school, living in the San Francisco Bay Area. I was introduced to rationality and Less Wrong while interning at Leverage Research, which was about a month ago.

I was given a free copy of Chapters 1-17 of HPMOR during my stay. I was hooked. I finished the whole series in two weeks and made up my mind to try and learn what it would be like being Harry.

I decided to learn rationality by reading and implementing The Sequences in my daily life. The only problem was, I discovered the length of the Eliezer’s posts from 2006-2010 was around around 10 Harry Potter books. I was told it would take months to read, and some people got lost along the way due to all the dependencies.

Luckily I am very interested in self improvement, so I decided that I should learn speed reading to avoid spending months dedicated solely to reading The Sequences. After several hours of training, I increased my reading speed (with high comprehension) five times, from around 150 words per minute to 700 words per minute. At that speed, it will take me 33.3 hours to read The Sequences.

It seems like most people advise reading The Sequences in chronological order in ebook form. Is using this ebook a good way to read The Sequences? Also, If I could spend 5 seconds to a minute after each blog post doing anything, what should I do? I was thinking of making some quick notes for myself to remember everything I read, perhaps with a spaced repetition system, or figuring out all the dependencies to smooth the way for future readers, perhaps leading to the easier creation of a training program...

Thanks for all your help, and I look forward to contributing to Less Wrong in the future!

Comment author: Nisan 18 July 2013 06:51:06PM *  1 point [-]

Welcome! As you're interested in applying the Sequences to your daily life, I suggest checking out the Center for Applied Rationality. (Maybe you overlapped with them at Leverage?) As part of their curriculum development process, they offer free classes at their Berkeley office sometimes. If you sign up here you'll be put on a mailing list where they announce these sessions, usually a day or so in advance.

Comment author: Brendon_Wong 18 July 2013 06:55:27PM 0 points [-]

Thanks, I just signed up. Do you think taking a full CFAR workshop would be a good next step after The Sequences? I'll be done in about 4 days at current reading speed (no planning fallacy adjustments), so I should probably plan ahead now.

Comment author: Nisan 18 July 2013 09:12:31PM *  0 points [-]

It would definitely be a good next step. I don't know if they have a minimum age for workshops, but it doesn't hurt to apply.

Comment author: Brendon_Wong 18 July 2013 09:17:38PM 0 points [-]

I don't believe they have age constraints, the issue is the monetary constraints :p

Thanks for your help!

Comment author: Nisan 18 July 2013 09:21:00PM 0 points [-]

They offer financial aid, too.

Comment author: Brendon_Wong 18 July 2013 09:28:43PM *  0 points [-]

Since I have a total of $23, I must get my parents to pay and allow me to go for a week, that will be the tricky part

Comment author: Jiro 19 July 2013 06:03:09PM *  1 point [-]

People might not like my response, but I'd say that if you're in a situation where you believe something might be beneficial to you but it consumes a substantial portion of your resources, you should heavily lean towards not going. This applies as much to a rationality workshop attended by someone with a tiny budget as it applies to playing the stock market. Making large expenditures for an uncertain return is generally a bad bet even if the expected utility gain is positive, if failure has a very negative consequence. And human beings are notoriously bad at assessing the expected utility in such situations.

You also need to be very confident in your ability to evaluate arguments if you don't want to end up worse than before.

Obviously, this doesn't apply if you're absolutely certain that going gives you more benefit than you forego in money, time, and parental willingness to give in (which may, in fact, be in limited supply) so there is no risk of loss, but not too many people are really that certain.

Comment author: thomblake 19 July 2013 06:24:26PM 0 points [-]

But surely going to a rationality workshop is the best way to learn to evaluate whether to go to a rationality workshop. And whether it succeeds or not, you can be convinced it was a good idea!

Comment author: James_Miller 18 July 2013 06:43:11PM 6 points [-]

If I could spend 5 seconds to a minute after each blog post doing anything, what should I do?

Figure out how you would explain the main idea of the post to a smart friend.

Comment author: Brendon_Wong 18 July 2013 06:45:45PM *  0 points [-]

Thanks! Just curious, how come you chose that over simply taking short 10 second notes allowing me to memorize all the main ideas?

Comment author: Eliezer_Yudkowsky 18 July 2013 07:33:43PM 2 points [-]

IIRC notetaking is supposed to work less well than explaining something to others. I don't know about imagining how to explain something to others.

Comment author: Vaniver 18 July 2013 07:42:15PM *  6 points [-]

I don't know about imagining how to explain something to others.

I would imagine that actually explaining it out loud to a rubber duck is better than imagining explaining it to a friend, for the same reasons that it is a common debugging practice. Actually putting something into words makes weak spots in understanding obvious in a way that imagination can glide over.

Comment author: [deleted] 20 July 2013 11:51:30PM -1 points [-]

When I imagine speaking to someone, I generally imagine specific words. YMMV.

Comment author: ialdabaoth 21 July 2013 12:15:48AM 2 points [-]

Actually speaking the words activates different areas of Broca's and Wernicke's regions (and elsewhere) than merely imagining them. Physically vocalizing the words, and hearing yourself vocalize them, allows them to be processed by more areas of your brain.

Comment author: [deleted] 22 July 2013 08:06:49PM *  0 points [-]

If that made much of a difference, it would also matter whether I was talking to someone out loud vs in writing. I don't feel that is the case, though it's not like I did any Gwern-level statistics about that. (Also, some people have more vivid auditory imagery than others.)

Comment author: Brendon_Wong 18 July 2013 08:04:32PM *  0 points [-]

IIRC notetaking is supposed to work less well than explaining something to others.

Perhaps note taking works less well for understanding, but explaining it out loud without recording it down or even writing my explanation will do very little for long term recall. What good will it do if I forget everything I read, after spending many hours reading it?

Comment author: Brendon_Wong 18 July 2013 09:15:15PM 0 points [-]

At first, I think I will try explaining ideas out loud as I read to save time, then write ultrashort notes on main ideas for long term memory.

Thanks for everyone's help!

Comment author: James_Miller 18 July 2013 07:13:27PM 0 points [-]

Both would work but my idea is less obvious so perhaps more helpful.

Comment author: Brendon_Wong 18 July 2013 07:23:36PM *  0 points [-]

That's an interesting idea. I suppose it might help with better understanding the concept, but it might not work for long term memorization. Should I write the explanations down?

Comment author: James_Miller 18 July 2013 08:37:37PM 0 points [-]

That would probably help if you have the time.

Comment author: Anna_Zhang 18 July 2013 07:28:02AM 7 points [-]

Hello, Less Wrong, I'm Anna Zhang, a high school student. I found this site about half a month ago, after reading Harry Potter and the Methods of Rationality. On Mr. Yudkowsky's Wikipedia page, I found a link to his site, where I found a link to this site. I've been reading the sequence How to Actually Change Your Mind, as Mr. Yudkowsky recommended, and I've learned a lot from it (though I still have a lot to learn...)

Comment author: Brendon_Wong 18 July 2013 06:42:59PM 1 point [-]

Welcome!

If you want to meet other high schoolers, this looks like a good place to start.

Comment author: idea21 17 July 2013 08:11:51AM 1 point [-]

Hi, Less Wrong.

I am idea21, I am from Spain and I apologize for my defective english.

I got acquainted with the existence of this forum thanks to the kindness of mister Peter Singer, he recommended me to expose my own idea about altruistic cultural development after questioning him whether he knew something similar about. Apparently there is nothing similar being discussed anywhere, which turned to be very dissapointing to me. But I still feel that "it" makes sense, at least from a logical point of view.

I will post here some excerpts of the message I wrote to Peter Singer, I hope any suggestion or comment from yours could be enlightening.

"Cultural changes about ethics have happened very slowly across history. According to some people they are motivated by economic issues (land´s property, trading, industrial development…) or political ones, also connected to economy. But although I read what Norbert Elias wrote about, it disturbs me the idea that the real change happens first in the people´s minds, influencing then to economics and politics, and not the other way around.

Primitive men started to create arts in Paleolithic previous to start agriculture in Neolithic. They decided to create arts, probably for the same reasons they decided to try to settle down: social needs, sharing emotional and intellectual activity in bigger groups. Agriculture was the economic answer to the practical problem of how affording sedentary way of life.

Norbert Elias (and Steven Pinker then) explains that an economic and political necessity urged authorities in the Middle Age to try to promote values of cooperation and less violent human relationships: that idea of the “civilité”, gentlemanliness, new rules of behavior advancing to modern humanism. But it seems to me that Elias and Pinker forget that rules to control individual aggression were created previous to the date they give (XIII century, as the European kings courts promoted the new gentle habits). The real origin of that is in monasticism, San Benito´s Rules are from VI Century, and monasticism did not start with the fall of the Roman Empire either. It did not start with Christianity, as a matter of fact. Buddhism started it.

All this reminded me what Miss Karen Armstrong wrote about “compassionate religions” and the “Axial Age”. So, the thing could be this way: First, intellectual changes happened (arts, communitarian life, ethics), and then new economic phenomena came, developing social, cultural, ideological forms; second, as social life increased human relationships, a new adaptation of individual behavior is demanded to control aggression.

It seems that monasticism is the answer to the need of developing new ways to control human behavior for the benefit of outside society, the same way that animals are tamed to be used by men. Monasticism is, basically, a "High-Performance Center" for behavior, producing “new men” able to control better the violent behavior and teaching these new discoveries for the outside people.

According to some current psychologists, like Simon Baron-Cohen, there are many people bearing features of “super-empathy”, being the opposite equivalent to the psychopaths, but unlike psychopaths, who can enjoy their own fitted sub-cultural environments (the underworld of criminality), there is not today a particular sub-cultural environment fitted for people particularly able to develop self-control of aggression and antiaggressive, affectionate and altruistic behavior. But in monasticism, these “super-empathic” people were specially fitted to develop patterns of aggression self-control: that psychological personal feature proved to be adaptive.

My idea (I hope not only mine…) is that monasticism should be re-invented.

A new monasticism of XXI century could be attractive for many young people, as providing them with emotional, intellectual and affectionate experiences that probably they could find nowhere else. It must be kept in mind that old monasticism existed because, at some extent, it fulfilled this kind of social needs for many people, particularly the young ones. Nobody compromises on the hard search for a future better world if they are not expecting to get, in that process, some kind of psychological reward in the present.

A monasticism of the XXI century would be, of course, very different from that of XVII century. It should be rational, atheist and non-authoritarian, emphasizing in affectionate and cooperative behavior by secluding from mainstream society. That could make much more against poverty that all the current NGO and every current humanitarian trend.

Human behavior is the “raw material” of humanitarianism. Not only because they could influence the mainstream society by demonstrating that a full antiagressive way of life is possible and emotionally rewarding, but also because the economic activity of “super-empathic” people culturally organized would be totally focused on altruistic work. Using modern technology, extremely cooperative organization and concentration of work resources (like in a “economy of war”) the results should be very good.

Remember that in Spain, in XVII century, there was a 2 % of population secluded from “civil life”, as monks, nuns or priests (and remember also the very committed communist activists of the first half of XX century). Can you imagine what a 2 % of this planet population could do with our technology, if rationally concerned to dedicate their lives only to ease human sufferance only in exchange for emotional, affectionate and intellectual rewards? Psychopaths are between 2 and 4 %: how many people “super-empathic” ones could exist? It would be worth trying to get them organized, culturally evolving for the whole world´s benefit.

Don´t underestimate young people idealism. The problem today is that they have not an alternative to start to create a better world outside our cultural, social and political mainstream limitations."

This idea could be develop much deeper, but I hope you will understand that it deals with the creation of a last religion, rational and of course atheist, in order to allow a furtther enhancing of human abilities for mutual cooperation.

As I mean "religion", I mean the necessity of developing an own system of cultural symbols and understandable patterns of social behaviour, which could not be the same as those of the current mainstream society (which is competitive, non-idealist and still irrational). As highly cooperative society could be based only on extreme trust and mutual altruism, it could be a bit similar to some traditions of the old compassionate religions, but now detached from any irrationality, any tradition and based on rational knowledge about human behaviour.

Thank you very much for your attention.

Comment author: rafiss 11 July 2013 05:35:01AM 4 points [-]

Hi everyone! I've been lurking around here for a few years, but now I want to be more active in the great discussions that often occur on this site. I discovered Less Wrong about 4 years ago, but the Methods of Rationality fanfic brought me here as a more attentive reader. I've read some of the sequences, and found them generally to use clear reasoning to make great points. If nothing else, reading them has definitely made me think very carefully about the way nature operates and how we perceive it.

In fact, this site was my first exposure to cognitive biases, and since then I've had the chance to study them further in college and read about them independently. This has been tremendously useful for me to understand why I and others I know behave the way we do.

I recently graduated college with a major in computer science and a decent exposure to math, having done some small independent research projects in machine learning. I'll soon begin a job as a software engineer at a late-stage startup that brings machine learning to the field of education.

I find that my greatest weakness with online communities is my tendency to return to lurking, even if I find the content very engaging. I hope to avoid that problem here, and at least continue participating in the comment threads.

Comment author: caffemacchiavelli 11 July 2013 12:00:37AM 4 points [-]

Hello, everyone. I stumbled upon LW after listening to Eliezer make some surprisingly lucid and dissonance-free comments on Skepticon's death panel that inspired me to look up more of his work.

I've been browsing this site for a few days now, and I don't think I've ever had so many "Hey, this has always irritated me, too!" moments in such short intervals, from the rant about "applause lights" to the discussions about efficient charity work. I like how this site provides some actual depth to the topics it discusses, rather than hand the reader a bullet list of trivialities and have them figure out the application.

I am working as a direct marketing consultant, in the process of getting my MBA (a decision I've started to regret; my faith in the scientific validity of academic management begins to resemble a Shepard Tone) and with future ambitions in entrepreneurship, investing, scaling and other things that fit in the "things I've never done yet smart people are supposed to be good at" box.

I'm a member of Mensa, casual Poker (winning) and Mahjong (losing) player, enjoy lifting weights, cooking (in an utterly unscientific way that would make Heston Blumenthal weep) and martial arts. I also have an imaginary -5yo son/daughter who keeps me motivated to put in more hours at work so we won't have financial worries once they get born.

There are a bunch of things I'd like to do with my life long-term, with varying amounts of megalomania, but I'm generally content with focusing on increasing my financial and (practical) intellectual power in the short- to mid-term and let the future decide just how far off my predictions and plans turn out to be. Estimates range from very to utterly.

Here's hoping LW will help me with that, and that I'll be helpful to others.

Comment author: afterburger 09 July 2013 02:38:34AM 3 points [-]

Hello! I'm here because...well, I've read all of HPMOR, and I'm looking for people who can help me find the truth and become more powerful. I work as an engineer and read textbooks for fun, so hopefully I can offer some small insights in return.

I'm not comfortable with death. I've signed up for cryonics, but still perceive that option as risky. As a rough estimate, it appears that current medical research is about 3% of GDP and extends lifespans by about 2 years per decade. I guess that if medical research spending were increased to 30% of current GDP, then most of us would live forever while feeling increasingly healthy. Unfortunately, raising taxes to achieve this is not realistic -- doubling taxes for an uncertain return is a hard sell, and I have been unable to find research quantifying the link between public research spending and healthcare technology improvements. Another approach is inventing a technology to increase the overall economy size by 10x, by creating a practical self-replicating robot. This is possible in principle (as demonstrated by Hod Lipson in 2006 and by FANUC robot arm factories daily) but I am currently not a good enough programmer to design and build a fully automated RepRap assembly system in a reasonable amount of time. Also, there are many smart and innovative people at Willow Garage, FANUC and other similar organizations, and it seems unlikely I could exceed the slow and incremental progress of those groups. A third option, trying to create super-level AI to make self-replicating robots for me, is even more difficult and unlikely. A fourth option, not taking heroic responsibility, would make me uncomfortable because I'm not that optimistic about the future. As it is, since dropping out of a PhD program I'm not confident in my ability to complete such a large project. Any practical help would be appreciated, as I would prefer not to rely on the untestable promises of quantum immortality, or on the faith that life is a computer game.

Comment author: idea21 25 July 2013 08:25:15AM 0 points [-]

Hi, afterburger

I find correct that you are not comfortable with death, the opposite of that would be unnatural.

I don't know whether you have ever heard of this person

https://en.wikipedia.org/wiki/Nikolai_Fyodorovich_Fyodorov

"Fedorov argued that the struggle against death can become the most natural cause uniting all people of Earth, regardless of their nationality, race, citizenship or wealth (he called this the Common Cause)."

Fedorov speculations about a future resurrection of all, although seen today as a joke, at least they are able to beat the "Pascal's wager" and, if we keep in mind the possiibilities of new particle physics, it is rational hoping that an extremely altruistic future humanity could decide to ressurrect all of us, by using resources on technology that today we cannot imagiine (the same way that current technology could have never been imagined by Plato or Aristotle).

Although science and technology could maybe keep limits, the most important issue about that would have to do with motivations. Why should a future humanity would be interested in acting so?

The only thing we could do today about helping that, would be starting to build up already the moral and cultural foundation of a fully altruistic and rational society (which would be inevitably, extremely economically efficient). And that is not done yet.

Comment author: Camaragon 08 July 2013 09:43:53AM 5 points [-]

Hello, my name is Cam :]

My goals in life are: 1. To build a self sufficient farm I with renewable alternative energy and everything. 2. Acquire financial assets to support the building of my farm and other hobbies and activities I pursue. 3 .To further my fitness and health and maintain it. 4. Love and Romance.

That's pretty much it, hahaha, I want to learn the ways of a Rationalist to make the best decisions and solutions for problems I might encounter in pursuing these goals! I have a immature or childlike air around me, people tend to say, which is why I am often looked down upon me and not taken seriously. I think it's how I construct my sentences maybe? My English is only at decent quality. Maybe I just see things too simply and positively people see it as being naive? Well, Anyway, I look forward to having you as my one of my buddies! :D

Comment author: Viliam_Bur 18 September 2013 12:40:15PM 1 point [-]

To build a self sufficient farm I with renewable alternative energy and everything.

Have you already built something? Do you have specific plans?

Comment author: polutropon 06 July 2013 08:11:53PM 3 points [-]

Hello again, Less Wrong! I'm not entirely new — I've been lurking since at least 2010 and I had an account for a while, but since that I've let that one lie fallow for almost two years now I thought I'd start afresh.

I'm a college senior, studying cognitive psychology with a focus on irrationality / heuristics and biases. In a couple of months I'll be starting my year-long senior thesis, which I'm currently looking for a specific topic for. I'm also a novice Python programmer and a dabbler in nootropics.

I'll be trying to avoid spending too unproductive time on LW ("insight porn" really is a great description, and I've learned to be wary of being excessively cerebral), but here I am again.

Comment author: Zoe 06 July 2013 04:39:32AM 5 points [-]

Hello Less Wrong community members,

My name is Zoe, I'm a philosophy student, and increasingly discombobulated by the inadequacy of my field of study to teach me how to Actually Do Things. I discovered Less Wrong 18 months ago, thanks to the story Harry Potter and the Method of Rationality. I've read a number of articles and discussions since then, mostly whenever I felt like reading something both intelligent and relevant, but I have not systematically read through any sequence or topic.

I have recently formed the goal to develop the skills necessary to 'raise the waterline' of rationality in the meat space discussions in which I take part, but without appearing to put anyone down.

Working towards this goal will make me interact more with a greater proportion of the people that are around me, which is something that I need to do. Right now, apart from a few friends whose minds I love, I usually flee at the earliest politically correct time from most conversations, due to sheer boredom or annoyance and a huge lack of confidence in my ability to steer the conversation somewhere interesting. I want to change this by improving myself (since Less Wrong has well taught me that it would be foolish to wait or hope for others to change or improve when I could be changing myself.)

While so far my use of Less Wrong has been recreational, I'm creating an account now to be able to participate in discussions, not because I think I have anything really important to say, but because practicing rationality not just in my mind but while actually interacting is probably a good way to go about my newfound objective. I would really like to become able to introduce rationality into conversations with the average non-rationalist and do so tactfully, and I think Less Wrong can help me.

Do you agree with my assessment that the Less Wrong posts and discussion community have the potential to help me further my goal? If so, how do you think I should best use the resources here?

I'm looking very much forward to interact with all of you!

Zoé

PS : My first language is French. I really do welcome any and all nitpicks and corrections about my English.

Comment author: John_Maxwell_IV 11 July 2013 06:37:06AM *  0 points [-]

Welcome!

I would really like to become able to introduce rationality into conversations with the average non-rationalist and do so tactfully, and I think Less Wrong can help me.

Let me know if you figure something out. So far I haven't been able to do it without coming across as weird.

Comment author: [deleted] 04 July 2013 01:20:58PM *  4 points [-]

Hi, I'm Alex, high school student. Came here from hpmor and have been lurking for about 5 months for now.

I use my "rationalnoodles" nickname almost everywhere, however still can't decide if it's appropriate on LW. Would like to read what others think.

Thanks.

Comment author: Cthulhoo 04 July 2013 01:57:55PM 0 points [-]

I use my "rationalnoodles" nickname almost everywhere, however still can't decide if it's appropriate on LW. Would like to read what others think.

Considering that infanticide is a generally accepted discussion topic here, I don't think people will question a nickname ;)

Welcome!

Comment author: atorm 04 July 2013 01:34:40PM 0 points [-]

It's not INappropriate.

Comment author: GTLisa 04 July 2013 06:33:56AM 7 points [-]

Hello, my name is Lisa. I found this site through HPMOR.

I'm a Georgia Tech student double majoring in Industrial Engineering and Psychology. I know I want to further my education after graduation, probably through a PhD. However, I'm not entirely sure what field I would want to focus on.

I've been lurking for awhile and am slowly making my way through the sequences, though I'm currently studying abroad so I'm not reading particularly quickly. I'm particularly interested in behavioral economics, statistics, evolutionary psychology, and in education policy, especially in higher education.

Comment author: John_Maxwell_IV 11 July 2013 06:41:38AM 0 points [-]

education policy, especially in higher education.

Fun fact: my high level of interest in education policy quickly evaporated as soon as I was no longer going to school.

Comment author: Axion 03 July 2013 03:05:50AM 11 points [-]

Hi Less Wrong. I found a link to this site a year or so ago and have been lurking off and on since. However, I've self identified as a rationalist since around junior high school. My parents weren't religious and I was good at math and science, so it was natural to me to look to science and logic to solve everything. Many years later I realize that this is harder than I hoped.

Anyway, I've read many of the sequences and posts, generally agreeing and finding many interesting thoughts. It's fun reading about zombies and Newcomb's problem and the like.

I guess this sounds heretical, but I don't understand why Bayes theorem is placed on such a pedestal here. I understand Bayesian statistics, intuitively and also technically. Bayesian statistics is great for a lot of problems, but I don't see it as always superior to thinking inspired by the traditional scientific method. More specifically, I would say that coming up with a prior distribution and updating can easily be harder than the problem at hand.

I assume the point is that there is more to what is considered Bayesian thinking than Bayes theorem and Bayesian statistics, and I've reread some of the articles with the idea of trying to pin that down, but I've found that difficult. The closest I've come is that examining what your priors are helps you to keep an open mind.

Comment author: Viliam_Bur 18 September 2013 08:43:36AM *  6 points [-]

Bayesian theorem is just one of many mathematical equations, like for example Pythagorean theorem. There is inherently nothing magical about it.

It just happens to explain one problem with the current scientific publishing process: neglecting base rates. Which sometimes seems like this: "I designed an experiment that would prove a false hypothesis only with probability p = 0.05. My experiment has succeeded. Please publish my paper in your journal!"

(I guess I am exaggerating a bit here, but many people 'doing science' would not understand immediately what is wrong with this. And that would be those who even bother to calculate the p-value. Not everyone who is employed as a scientist is necessarily good at math. Many people get paid for doing bad science.)

This kind of thinking has the following problem: Even if you invent hundred completely stupid hypotheses; if you design experiments that would prove a false hypothesis only with p = 0.05, that means five of them would be proved by the experiment. If you show someone else all hundred experiments together, they may understand what is wrong. But you are more likely to send only the successful five ones to the journal, aren't you? -- But how exactly is the journal supposed to react to this? Should they ask: "Did you do many other experiments, even ones completely irrelevant to this specific hypothesis? Because, you know, that somehow undermines the credibility of this one."

The current scientific publishing process has a bias. Bayesian theorem explains it. We care about science, and we care about science being done correctly.