Notes for future OT posters:

1. Please add the 'open_thread' tag.

2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)

3. Open Threads should start on Monday, and end on Sunday.

4. Unflag the two options "Notify me of new top level comments on this article" and "

New to LessWrong?

New Comment
76 comments, sorted by Click to highlight new comments since: Today at 11:11 AM
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings
[-][anonymous]7y80

Final version of thesis going out within 4 days. Getting back into a semi-regular schedule after PhD defense, death in the family, and convergence of job-search on a likely candidate in quick succession. Astrobiology writing likely to restart soon. Possible topics include:

  • schools of thought in life origins research
  • the nature of LUCA
  • recent work on the evolution of potentiated smart animal lineages on Earth
  • WTF are eukaryotes anyway
  • the fallacies of the Fermi paradox/ 'great filter' concepts
  • the fallacies of SETI as it is currently performed

I'm thinking on writing a post on doing 'lazy altruism', meaning 'something having a somewhat lasting effect that costs the actor only a small inconvenience, and is not specifically calculated to do the most amount of good - only the most amount per this exact effort."

Not sure I'm not too lazy to expand on it, though.

4Elo7y
write it!
0Viliam7y
Potentially very useful, considering how many lazy people are out there.

Tyler Cowen and Ezra Klein discuss things. Notably:

Ezra Klein: The rationality community.

Tyler Cowen: Well, tell me a little more what you mean. You mean Eliezer Yudkowsky?

Ezra Klein: Yeah, I mean Less Wrong, Slate Star Codex. Julia Galef, Robin Hanson. Sometimes Bryan Caplan is grouped in here. The community of people who are frontloading ideas like signaling, cognitive biases, etc.

Tyler Cowen: Well, I enjoy all those sources, and I read them. That’s obviously a kind of endorsement. But I would approve of them much more if they called themselves the irr

... (read more)
89eB17y
I think no one would argue that the rationality community is at all divorced from the culture that surrounds it. People talk about culture constantly, and are looking for ways to change the culture to better address shared goals. It's sort of silly to say that that means it should be called the "irrationality community." Tyler Cowen is implicitly putting himself at the vantage point of a more objective observer with the criticism, which I find ironic. Where Tyler is wrong is that it's not JUST another kind of culture. It's a culture with a particular set of shared assumptions, and it's nihilistic to imply that all cultures are equal no matter from what shared assumptions they issue forth. Cultures are not interchangeable. Tyler would also have to admit (and I'm guessing he likely would admit if pressed directly) that his culture of mainstream academic thought is "just another kind of religion" to exactly the same extent that rationality is, it's just less self-aware about that fact. As an aside, I think Lumifer is a funny name. I always associate it with Lumiere from Beauty and the Beast, and with Lucifer. Basically I always picture your posts as coming from a cross between a cartoon candle and Satan.
3Lumifer7y
Notice the name of this website. It is not "The Correct Way To Do Everything". Don't ALL cultures have their own particular set of shared assumptions? Tyler's point is that the rationalist culture, says Tyler, sets itself above all others as it claims to possess The Truth (or at least know the True Paths leading in that general direction) -- and yet most cultures have similar claims. Lucifer is the bringer of light (Latin: lux). Latin also has another word for light: lumen (it's the same root but with the -men suffix). Just sayin' :-P But I will also admit that the idea of an all-singing all-dancing candelabra has merit, too :-)
2Brillyant7y
It did seem to be a pretty bold and frontal critique. And "irrationality community" is probably silly. But I agree LW, et al has at times a religious and dogmatic feel to it. In this way the RC becomes something like the opposite of the label it carries. That seems to be his point. Yes. Yes. If this wasn't exactly the mental image I had of Lumifer before, then it is now. Maybe a bit more Satan than cartoon
7Viliam7y
¯\_(ツ)_/¯ (posted mostly for the challenge of writing this properly in markdown syntax. actually, it was quite easy)
2Lumifer7y
Razib Khan also offers a comment.
2Elo7y
No one is more critical of us than ourselves. "LessWrong" is lesswrong for being humble about it. Hopefully that humility sticks around for a very long time.

No one is more critical of us than ourselves.

This seems untrue. For example, RationalWiki.

In the past I could also have pointed to some individuals (who AFAIK were not associated with RW, but they could have been) who I think would have counted. I can't think of any right now, but I expect they still exist.

0dxu7y
That's fair, but I also think it largely misses the point of Elo's comment. Here, have (an attempt at) a rephrase: No community is as prone to self-criticism as the rationalist community.
2philh7y
I think I would rather say, less superlatively, that we're unusually good at self-criticism. (I do note that I'm comparing my inside view of the rationalist community with my outside view of other communities, so I shouldn't put too much confidence in this.) (But yes, I agree that I was ignoring the thing that Elo was actually trying to point at.)
0Elo7y
yes. the fact that we know of RW, and this mention, and try to understand what is going on here. is a good start.
7MrMind7y
Humility is good, but calibration is better.
1Nate_Rausch7y
A bit tongue-in-cheeck, but how about taking Tyler's unfair label as a proposal? We could start the rationality religion, without the metaphysics or ideology of ordinary religion. Our God could be everything we do not know. We worship love. Our savior is the truth. We embrace forgiveness as the game-theoretical optimal modified tit-for-tat solution to a repeated game. And so on. We thoroughly investigate and aggregate the best knowledge humanity currently has on how to live. And we create Rationality Temples worldwide. There will be weekly congregations, with talks on a sequence, with following discussions, on topics such as signalling, bayesian thinking, cognitive biases. We propose a three step way to heaven on earth: identifying worthwhile causes, charting effective solutions and taking actions to achieve it. Lifetime goal is writing a sequence. Compassion meditation and visualisation prayer once per day. Haha, okay perhaps I'm overdoing it. Using the well-established concepts, rituals and memes of religion is easy to mock, but what if it is also an effective way to build our community and reach our goals?
2MrMind7y
It surely is an effective way, since by this mean all kinds of silly causes have been pursued. But creating a religion out of rationality (lowercase) would defeat its purpose: in the span of a year, rationality would become the password to learn by memory and the beginning structures will solidify as an attire. Religions are appealing exactly because they exempt their members from thinking on their own and accepting hard truths: rationality has instead more in common with martial arts, they are mostly a question of training and learning to take many hits.
0Nate_Rausch7y
Well, yes, but couldn't one just make a new religion without those attributes. For example the first of the 10 commandments could be: Question everything, including these texts. Be a student, not a follower. Finding fault in ourself is the highest virtue, free speech etc. ? :-)
0Lumifer7y
Literally God of the Gaps! :-) Which love?
4gjm7y
Tangential: I think the "four loves" thing is a bit of a cheat, like the "fifty Eskimo words for snow" meme. The Greeks had different words for describing different kinds of positive interpersonal affect -- but so do we! We have "affection" and "friendship" and "devotion" and "lust" and "loyalty" and "benevolence" and so on. That doesn't mean there's anything wrong with noticing when one of those words (in this case "love") gets used very broadly, and it doesn't change the fact that "see how other people classify things" is a useful heuristic. But I am entirely unconvinced that there's anything very special about the Ancient Greeks, or the notion of love, in this regard -- other than various historical contingencies involving C S Lewis, Christianity, and the history of the New Testament.
0Lumifer7y
Well, they were pretty special, being the cradle of the Western civilization 'n'all, but in this specific case all I intended was to give the OP a possible list of specific meanings of the word "love" to consider.
0Nate_Rausch7y
Fair point, well I don't think romantic love is worthy of sacred status in the irrationality religion. Tough those four neither seemed quite to fit the love I had in mind. Perhaps something closer the buddhist concept of bodhisattva, meaning altruistic love for all sentient beings?
1Lumifer7y
Sounds like the plain old Christian love, but with a new cool label :-/
2bogus7y
Ah, Christian love... the kind of altruistic love that makes people tell you that you're a fallen and depraved creature due to the sin of Adam and thus will be burning in Hell forever unless you "accept Jesus as your savior" by becoming a Christian ASAP. (If precedent is any guide, I can already guess that Roko's basilisk will be featured prominently in the new "rationality religion"!)
0Lumifer7y
Well, our baseline is what, Buddhist love? There is Buddhist hell) as well and surprise! it doesn't sound like a pleasant place. You get there through being enslaved by your lusts and desires -- unless, of course, you accept the teachings of Buddha ASAP :-P
09eB17y
Bryan Caplan responded to this exchange here
2tristanm7y
I would object to calling these "devastating counter-examples", they're more like unsolved problems. It seems overly dramatic. I'm not a perfect Bayesian agent, I use my intuitions a lot, but that is not grounds on which to reject Bayesianism, and I think we could say something similar about consequentialism. I may not know how to perfectly measure relative happiness, or perfectly predict the future, but it doesn't seem like that should be grounds to reject consequentialism entirely, in favor of alternatives which don't cope with those issues either.
2dxu7y
One very common error people make is to treat "utilitarianism" and "consequentialism" as if they were one and the same thing. Utilitarianism makes claims about what is moral and what is not. Consequentialism makes claims about what sort of properties a moral criterion should have. Criticisms about utilitarianism, therefore, are often taken also as criticisms of consequentialism, when in fact the two are distinct concepts!
0Good_Burning_Plastic7y
On the other hand, utilitarianism is a subset of consequentialism, so whereas a criticism of utilitarianism is not necessarily a criticism of consequentialism, the converse is true.

Our article about using nuclear submarines as refuges in case of a global catastrophe has been accepted for the Futures journal and its preprint is available online.

Abstract

Recently many methods for reducing the risk of human extinction have been suggested, including building refuges underground and in space. Here we will discuss the perspective of using military nuclear submarines or their derivatives to ensure the survival of a small portion of humanity who will be able to rebuild human civilization after a large catastrophe. We will show that it is a ve... (read more)

4morganism7y
Back when SeaSteading forum was going, one of the folks was working on building cement submarines, as a pre-design to growing them like coral.

Curious about if this is worth making into it's own weekly thread. Curious as to what's being worked on, in personal life, work life or just "cool stuff". I would like people to share, after all we happen to have similar fields of interest and similar fields we are trying to tackle.

Projects sub-thread:

  • What are you working on this week (a few words or a serious breakdown)(if you have a list feel free to dump it here)?
  • What do you want to be asked about next week? What do you expect to have done by then?
  • Have you noticed anything odd or puzz
... (read more)
3Elo7y
I am working on: * learning colemak * vulnerability, circling (relational therapy stuff) investigating the ideas around it. * trialling supplements: Creatine, Protein, Citrulline malate. Adding in Vitamins: C, D, Fish oil, Calcium, Magnesium, Iron, Distant future trials: 5HTP, SAMe. preliminary SAMe and 5HTP were that they make me feel like crap. * promoting the voluntary euthanasia party of NSW (Australia) (political reform of the issue) * emptying my schedule to afford me more "free" time in which to write up posts. * contemplating book topic ideas. * trying to get better routines going, contemplating things like "I only sleep at home", and cancelling all meeting and turning off all connected media for a week. my biggest bottlenecks are myself (sometimes focus) and being confident at what will have a return VS not have a return. (hence many small experiements)
0WalterL7y
You are working on vulnerability? I'm not sure I understand.
0Elo7y
https://www.youtube.com/watch?v=iCvmsMzlF7o yes. This talk seems to imply value from being vulnerable in social settings to connect with people. I am experimenting to see if it's worth it or better.
0username27y
Where else do you sleep?
0Elo7y
This may seem obvious to me but I sleep at the houses of partners, my parents house, a property I was doing some work for ~1.5 hrs away from home. The point is that I have less control over the bed, bedroom setup, my sleep and my morning routine if I am not in the bed that I engineered the routine around.
3MaryCh7y
Translating a novel (really, a collection of essays) about WWII massacres of Jews in Kyiv & the rise of neo-nazism in post-Soviet republics (and much in between). It will take me a few months, probably, since this is a side job. Overall impression: the past is easier to deal with, because it is too horrible. Imagine 10^5 deaths. Although I unfortunately know the place where it happened, & he includes personal stories (more like tidbits), so the suspension of disbelief takes some effort to maintain. But the 'present' part - a series of the author's open letters to mass-media and various officials about pogroms and suchlike that went unpunished - is hard: he keeps saying the same over and over. (Literally. And his style is heavy on the reader.) After a while the eye glazes over & notices only that the dates and the addresses change, but the content doesn't, except for the growing list of people who had not answered. Just had not answered. Now this is - easy to imagine. Maybe this isn't odd, but I had thought it would be the other way around.
2Lumifer7y
" A single death is a tragedy; a million deaths is a statistic" -- a meme
2MaryCh7y
Ah, but what if you have walked above their bones?
2Lumifer7y
You always walk over bones, it's just that you know of some and don't know of others.
2MaryCh7y
I don't know how you do it, but you seldom fail to cheer me up. Even a little bit. Thanks.
0MaryCh7y
From the book, on 'The Doctors' plot' of 1953: Among the listed people who provided medical help to the party and state leaders, there was an abrupt addition - V. V. Zakusov, professor of pharmacology. He didn't take part directly in the leaders' treatment - he was at first only called in for an opinion, and given to sign the conclusion about the prescriptions that the 'doctors-murderers' had issued to hasten their patients' death. Vasili Vasilyevitch Zakusov took the feather and, well aware of what lied ahead, wrote this: "The best doctors in the world will sign such prescriptions." In that moment he stopped being an expert and became a suspect. In jail, even after tortures, he didn't withdraw his conclusion.
3philh7y
I'm working on * a graphing library for python (ggplot2's conceptual model with a pythonic API) * writing cliffs notes for Order Without Law (a book about how people settle disputes without turning to the legal system) * learning ukelele with Yousician (currently making extremely slow progress on the "double stops" lesson, I find it really hard to consistently strum two strings at once) * trying to write an essay about how heuristics can sometimes be anti-inductive (they become less accurate the more they're applied), and how we don't really seem to have any cultural awareness of this problem even though it seems important I might have the last one complete by next week, but the others are fairly long-term projects.
3Viliam7y
Reminds me, we didn't have a bragging thread for some time. I think I'd like to see this as a separate topic (probably monthly, because such things take time). However, just do as you wish, and then we'll see how it works.
1moridinamael7y
I'm working on a podcast read-through of the web serial Worm. We've just put out the fifth episode of the series. It's becoming pretty popular by our standards.
1Darklight7y
I recently made an attempt to restart my Music-RNN project: https://www.youtube.com/playlist?list=PL-Ewp2FNJeNJp1K1PF_7NCjt2ZdmsoOiB Basically went and made the dataset five times bigger and got... a mediocre improvement. The next step is to figure out Connectionist Temporal Classification and attempt to implement Text-To-Speech with it. And somehow incorporate pitch recognition as well so I can create the next Vocaloid. :V Also, because why not brag while I'm here, I have an attempt at an Earthquake Predictor in the works... right now it only predicts the high frequency, low magnitude quakes, rather than the low frequency, high magnitude quakes that would actually be useful... you can see the site where I would be posting daily updates if I weren't so lazy... http://www.earthquakepredictor.net/ Other than that... I was recently also working on holographic word vectors in the same vein as Jones & Mewhort (2007), but shelved that because I could not figure out how to normalize/standardize the blasted things reliably enough to get consistent results across different random initializations. Oh, also was working on a Visual Novel game with an artist friend who was previously my girlfriend... but due to um... breaking up, I've had trouble finding the motivation to keep working on it. So many silly projects... so little time.

"Modafinil-Induced Changes in Functional Connectivity in the Cortex and Cerebellum of Healthy Elderly Subjects"

http://journal.frontiersin.org/article/10.3389/fnagi.2017.00085/full

"CEDs may also help to maintain optimal brain functioning or compensate for subtle and or subclinical deficits associated with brain aging or early-stage dementia."

"In the modafinil group, in the post-drug period, we found an increase of centrality that occurred bilaterally in the BA17, thereby suggesting an increase of the FC of the visual cortex with oth... (read more)

I'm still mulling over the whole "rationalism as a religion". I've come to the conclusion that there are indeed two axioms that are shared by the rational-sphere that we-cannot-quite-prove, and whose variations produce different cultures.
I call them underlying reality and people are perfect.
"Underlying reality" (U): refers to the existence of a stratum of reality that is independent from our senses and our thoughts, whose configurations gives the notion of truth as correspondence.
"People are perfect" (P): instead refers to the... (read more)

In "Strong AI Isn't Here Yet", Sarah Constantin writes that she believes AGI will require another major conceptual breakthrough in our understanding before it can be built, and it will not simply be scaled up or improved versions of the deep learning algorithms that already exist.

To argue this, she makes the case that current deep learning algorithms have no way to learn "concepts" and only operate on "percepts." She says:

I suspect that, similarly, we’d have to have understanding of how concepts work on an algorithmic leve

... (read more)
0bogus7y
The problem with AGI is not that AIs have no ability to learn "concepts", it's that the G in 'AGI' is very likely ill-defined. Even humans are not 'general intelligences', they're just extremely capable aggregates of narrow intelligences that collectively implement the rather complex task we call "being a human". Narrow AIs that implement 'deep learning' can learn 'concepts' that are tailored to their specific task; for instance. the DeepDream AI famously learns a variety of 'concepts' that relate to something looking like a dog. And sometimes these concepts turn out to be usable in a different task, but this is essentially a matter of luck. In the Amazon reviews case, the 'sentiment' of a review turned out to be a good predictor of what the review would say, even after controlling for the sorts of low-order correlations in the text that character-based RNNs can be expected to model most easily. I don't see this as especially surprising, or as having much implication about possible 'AGI'.
0entirelyuseless7y
Humans are general intelligences, and that is exactly about having completely general concepts. Is there something you cannot think about? Suppose there is. Then let's think about that thing. There is now nothing you cannot think about. No current computer AI can do this; when they can, they will in fact be AGIs.
3Lumifer7y
Your link is broken. ETA: Thomas fixed the problem.
0Thomas7y
I don't know. I can't reproduce the error, it just works when I click it. It is strange however, that this number 9056 has been given by the Wordpress. Usually it is the title of the post.
3Lumifer7y
Your link leads me to a Wordpress login page. To reproduce, try deleting your cookies (or try another browser). Specifically, your link resolves to this: https://wordpress.com/wp-login.php?redirect_to=https%3A%2F%2Fwordpress.com%2Fpost%2Fprotokol2020.wordpress.com%2F9056 Is that -- https://protokol2020.wordpress.com/2017/04/02/another-intermezzo-intermezzo-problem/ -- the link you mean?
2Thomas7y
Thank you very much! Now, should be fixed!

I have an idea about how to make rationality great again (bad joke, but I'm serious). The term "theoretical rationality" may be coined by me - idk, but the new meanings to "theoretical and applied rationality" are mine and include 1)optimisation 2)fast completion of goals and ideals 3)updating the list of desirable goals and ideals 4) repeat. Any comments?

2Elo7y
how does this compare to the instrumental and epistemic rationality separation?
0Jordan2386_duplicate0.0220539558213204157y
The distinctions are in the application. In my proposed system, the theoretic (epistemic) rationality should mostly be a byproduct of the applied (instrumental) rationality. My view puts a huge emphasis on the knowledge being derived mostly and predominantly derived through (many fast) interactions with the environment, avoiding the pitfalls of fixing scientific "laws" (which themselves are products of many observations). This is not the Bayesian view with priors on what one could expect from looking into some unexpected phenomenon. If it works, it works. If the theory says it can't work, it would still work.
0MrMind7y
Ehrm...
0Jordan2386_duplicate0.0220539558213204157y
The conference looks cool, but does not follow the main framework of "theoretical and applied rationality", as defined by me. More specifically, I can use the term "whatever works (most) effectively" not to cause confusion with the otherwise defined "theoretical and applied rationality".What I do with my approach is to completely avoid theorization, unless the interactions with the environment uncover a pattern that needs recording. The main focus of the approach is on the applied end. If one needs to, say, improve wealth for the poorest I can't think of a more efficient way to do so than fast experimentation and coming up with new and better way to do so. The cool bit is that it can be sustainable and empowering, blah, blah, blah (not the focus of this conversation). The idea is that one interacts with the environment to acquire new knowledge. Otherwise one can become detached from the environment as is most of modern maths theory (I study maths in Oxford btw). Collecting theory is very very backwards to the second pillar of rationality - achieving one's goals and ideals.

Is there a place in the existential risk community for a respected body/group to evaluate peoples ideas and put them on a danger scale? Or dangerous given assumptions.

If this body could give normal machine learning a stamp of safety then people might not have to worry about death threats etc?

In some situations my thinking becomes much more structured, I throw out the syntax and the remaining words come in very clear hierarchy and kind of seem to echo shortly. It lasts, perhaps, less than a second.

Examples: "snake - nonvenomous (snake, snake) - dead, where's the head, somebody struck it with something, probably a stick, curse them, ought to collect it, where's vodka"; "snake - viper (viper, viper) - back off, where's camera, damn, it's gone, ought to GPS the place"; "orchid - Epipactis (...pactis) - why not something rarer, again this weed".

Has it been like that for you?

I'm wondering what people would think about adopting the term Future Super Intelligences or FSI, rather than AGI or SAGI.

This would cover more scenarios (e.g. uploads/radical augments) where the motivational systems of super powerful actors may not be what we are used to. It would also show that we are less worried about current tech than talking about AIs does, there is that moment when you have to explain you are not worried about backprop.

0g_pepper7y
I never really thought that AGI implied any specific technology; certainly the "G" in AGI rules out the likelihood of AGI referring to current technology since current technology is not generally intelligent. AGI seems to capture what it is we are talking about quite well, IMO - Artificial (i.e. man-made) General Intelligence. Do you really find yourself explaining that AGI is not the same as backpropagation very often?
0whpearson7y
I think AGI gets compressed to AI by the mainstream media and then people working on current ML think that you people are worried about their work (which they find ridiculous and so don't want to engage). An example of the compression is here. We don't help ourselves calling it airisk.
2g_pepper7y
Actually, the term AI has traditionally encompassed both strong AI (a term which has been mostly replaced by LW and others with "AGI") and applied (or narrow) AI, which includes expert systems, current ML, game playing, natural language processing, etc. It is not clear to me that the mainstream media is compressing AGI into AI; instead I suspect that many mainstream media writers simply have not yet adopted the term AGI, which is a fairly recent addition to the AI jargon. The mainstream media's use of "AI" and the term "AIRISK" are not so much wrong as they are imprecise. I suspect that the term AGI was coined specifically to differentiate strong AI from narrow AI. If the mainstream media has been slow in adopting the term AGI, I don't see how adding yet another, newer term will help - in fact, doing so will probably just engender confusion (e.g. people will wonder what if anything is the difference between AGI and FSI).
0whpearson7y
AGI has been going back over 10 years? Longer than the term AIrisk has been around, as far as I can tell. We had strong vs weak before that. AGIrisk seems like a good compromise? Who runs comms for the AGIrisk community? Imprecision matters when you are trying to communicate and build communities.
0g_pepper7y
I certainly prefer it to FSIrisk. I doubt anyone does. Terms catch on or fail to catch on organically (or memetically, to be precise). Perhaps. But, I doubt that a significant amount of the reluctance to take the unfriendly AGI argument seriously is due to confusion over terminology. Nor is changing terminology likely to cause a lot of people who previously did not take the argument seriously to begin to take it seriously. For example, there are some regulars here on LW who do not think that unfriendly AGI is a significant risk. But I doubt that any LW regular is confused about the distinction between AGI and narrow AI.

Are there any studies that determine whether regular coffein consumption has a net benefit? Or does the body produce enough additional receptors to counteract it?

2Dagon7y
I like "coffein" as a portmanteau. I will try to introduce it into conversation and see who notices.
0PotterChrist7y
Google is your friend. Here is an example link: http://www.sciencedirect.com/science/article/pii/0278691595000933 I have smird'(proof)#regular question, did this just safe it but only for me?

Normal "introducing myself", normal I am of faith.

0PotterChrist7y
Regular, "dont think about the irony"