If it's worth saying, but not worth its own post, then it goes here.


Notes for future OT posters:

1. Please add the 'open_thread' tag.

2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)

3. Open Threads should start on Monday, and end on Sunday.

4. Unflag the two options "Notify me of new top level comments on this article" and "

New Comment
94 comments, sorted by Click to highlight new comments since: Today at 12:30 PM
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings

Hey guys,

I'm a lurker, but I'm a regular member of the Denver LW meetup crew, trying to get our scheduled meetups on the main map. There's a Karma limit for that sort of post, and the mod I talked to sent me to you here for help. Would you please give me internet points to make this possible? You'd make all of my transhumanist and EA dreams come true. You know, except the main ones.

Could I get a couple of upvotes so that I could post links? I'd like to put some of the LW-relevant content from weird.solar here now that link posts are a thing.

3Dagon7y
Upvoted, but I have to say I dislike link posts in discussion. The vast majority are wrong for LW - either oversimple or just repeating what's already in the wiki or sequences. Some are on-topic and well-written, but even then the community interaction is broken when there are comments both on the hosting site and LW. If you have something LW-appropriate to say, make a post and include the link. note: since links are now a thing, I'm likely in the minority.
0ProofOfLogic7y
Yeah, I think the links thing is pretty important. Getting bloggers in the rationalist diaspora to move back to blogging on LW is something of an uphill battle, whereas them or others linking to their stuff is a downhill one.
1MrMind7y
I have upvoted you. Make us proud.

Downvoting is temporarily disabled! I'm very excited about this change because in the last few weeks I've seen some good conversations deleted by someone exploiting a sockpuppet glitch. Besides, I have always preferred commenting to downvoting.

1niceguyanon7y
Agree. There have been an influx of posts and almost no comments to go with them, looks sad. SSC's most recent post is really interesting but almost no comments, perhaps because it has already been discussed on SSC, maybe this will help.
1Lumifer7y
SSC got a little burned out by the election madness.

Check out the Double Crux post in Main!

Double Crux is one of the recent CFAR methods that seems like it could spread easily and isn't too deeply reliant on other things that CFAR teaches. (Basically, it's about what leads to conversations where people can actually change their minds, and a recipe for doing so.)

There's a new post in Main! I missed it completely, because on login I head straight to Discussion... if you are like me, just be aware.

4ChristianKl7y
Given we recently got the change to have link posts in discussion it seems to me like the ideal solution would be to show main posts as well in discussion.

The reason I visit LW is it satisfies a need for community. I'm glad to see the recent efforts at revitalisation, as a large part of the value for me generated by a single conversational locus is the social support it provides. This site has been inactive for a long time - and yet to my puzzlement I still found myself checking it regularly, despite not learning anything. I discovered that it's because I just wanted to keep in touch with what's going on in rationalist circles, and hang out a bit. I see myself as an aspiring rationalist, and that's a hard th... (read more)

Moved Fact Posts to Main and promoted it; make sure you don't miss it.

[-][anonymous]7y40

I had to translate an article about testing the shelf life of a viral diagnosticum of the fourth generation, and it seemed rather fishy to me (but I'm never a chemist). The authors used the "accelerated aging" method, heated the diagnosticum up for some periods of time to some temperatures, and then tested the "functional parameters". The rationale is that a 10C increase in temperature results in double the rate of the reaction. They used the results to project the shelf life at 4 C.

As far as I can tell, they did not check test kits th... (read more)

0MrMind7y
I know very little about it, but there doesn't seem to be anything wrong. If anything, Arrhenius equation over-estimates reaction rates constants because it presupposes no barrier to the acquisition of the activation energy. Do you have any reason to believe that this diagnostic equipment has an internal source of energy which is not the temperature?

If you could pick one music track that, if turned into a music video, could most exemplify the emotions resulting from LW-style rationality, what would that song be?

3g_pepper7y
Aside from the music video constraint, I would say Schubert's piano sonata in B flat major. Overall it is an optimistic, contemplative piece (capturing the attitudes of many LWers) but the bass note trill occurring early in the piece and repeating several times throughout serves as a reminder of existential risks such as unfriendly AGI. The piece was used (to great effect) at a climactic scene of the recent movie Ex Machina.
0[anonymous]7y
I immediately thought of this.
0polymathwannabe7y
Steve Reich's Violin Phase.
0knb7y
I immediately thought of this.
[-][anonymous]7y40

The bottom left corner of Questionable Content number 3362 (http://questionablecontent.net/view.php?comic=3362). That is all.

Everyone's afraid that robots will steal manual labor. But the components for robots stealing enterpreneur's jobs are already floating around: DAO's, machine learning for copywriting, maximize profit.

0ChristianKl7y
A DAO doesn't replace any of the work of understanding what customers are willing to buy and making product development decisions. In most cases for legal products a central server costs a lot less to get the same computational power. The added complexity that the DAO provides makes no sense for most startups.
0MrMind7y
But an AI possibly can. The DAO would be just a way to redistribute the ownership of the whole "AI + robot factory" complex. It's not the AI that is distributed along the blockchain, it's the company modus operandi.
0ChristianKl7y
If there's GAI that might make the decisions that enterpreneur's make. Currently that's not on the horizon.
0MrMind7y
I believe that coordinating a series of specialized AIs could do the job, but until I write down a model I cannot go beyond a mere statement / hope.
0Lumifer7y
Don't you mean the crater where the DAO used to be..?
0scarcegreengrass7y
Extremely confusingly, there is an organization named 'The DAO' which is a DAO (a Distributed Autonomous Organization).
1Lumifer7y
Keeping in mind that dao and tao are different ways of spelling the same thing, let me quote from Tao Te Ching:
0MrMind7y
Exactly!
0Viliam7y
Does this mean that Tao has infinite Kolmogorov complexity?
0Lumifer7y
I don't know -- you might want to ask Gödel, Escher, and Bach about it :-)
0scarcegreengrass7y
Yes! This is delightful.

MIRI publishes a lot of research on 'neat' systems like first order logic reasoners, and not on 'scruffy' systems like neural networks. I heard Eliezer Yudkowsky allude to the idea that this is for convenience or budgetary reasons, and that they will do more research on neural networks (etc) in the future.

Does anyone have any more information about what MIRI thinks and intends to research about 'scruffy' AI systems?

4ProofOfLogic7y
Basically, this: https://intelligence.org/2016/07/27/alignment-machine-learning/ It's now MIRI's official 2nd agenda, with the previous agenda going under the name "agent foundations".
2scarcegreengrass7y
Okay, thanks.

I'm having an un-rational moment, and despite knowing that, it's still affecting my behaviour.

Earlier today, my newsfeed included the datum discussed here, of Trump having a phone call with the President of Taiwan; and the item discussed here, about Trump talking about 'shutting down' the Internet. And later, while listening to my music playlist of the Merry Wives of Windsor, one of the tunes that popped up was "Green Fields of France", one version of which can be heard here. And I started wondering whether I was prepared for politics to go in an... (read more)

6entirelyuseless7y
You sound paranoid. Even if there is significant military conflict, it won't affect you personally.
0DataPacRat7y
Maybe there's a bit of terminology confusion here; if a military conflict /doesn't/ affect me personally, it seems unlikely to be a 'significant' one. (Some historical ways a military conflict could affect me personally: Victory Gardens, the Order of the White Feather, the Fenian Raids, even less oversight and accountability for civilian police whose actions would otherwise end up in the subreddit "Bad Cop, No Donut".) I'm thinking of scenarios such as 'It turns out China put secret backdoors into all sorts of hardware chips, and in a fit of self-righteous pique (which they think will play well to their red-state base), the war-monger side of the American Congress doesn't see any downsides to making a demand that everyone in the world shut down their supposedly Chinese-controlled hardware under threat that if they don't, they'll send the American military to shut it down'. As far as I can tell, several versions of just this one particular scenario don't obviously break the sociological law of every political actor having to act in what they perceive to be their own self-interest. However, I no longer trust my sense of calibration for the odds of large-scale politics, given that I was willing to go along with the predicted odds of 88% for Hillary winning the election, and didn't update nearly as much as I should have by the time of the election itself. And said lack of calibration puts a sharp limit on how rationally I can act as I decide how much effort to put into preparing for the more unpleasant scenarios.
3sen7y
You sound insane desu. Stop browsing reddit for a while. Any board where attention is explicitly rewarded, whether in the form of (You)s or upboats, will almost by definition tend towards encouraging high volatility of beliefs and emotions. It sounds like you've been riding that wave a bit too long. Also, learn to recognize fear mongering.
0DataPacRat7y
I want to provide arguments offering further justification for increasing my priority for making personal offline backup copies of various online resources (such as "it's something I've been vaguely wanting to do for some time anyway, I've just never had any particular impetus to get more of the job done than my current mirrors"), but, from inside my head, it's hard to tell whether these are actual reasons or mere rationalizations. Do facts such as that I've had this username for 15 years, have said "it's not just a nom-de-net, it's a way of life", and already have offline copies of Wikipedia, Project Gutenberg, and several other multi-gigabyte text references, provide a reasonable amount of evidence that my possibly-irrational desired behaviour is merely a continuation of my existing trends, rather than being a step too far?
4sen7y
I said you sound insane because of your paranoia, not because of what you wanted to do as a result of that paranoia. Whether or not you would be creating backups in other circumstances is irrelevant, except as an indicator of how paranoid you are. I don't think such an indicator is necessary because your first two paragraphs already demonstrate what I see as an extreme level of paranoia, and so to me it's irrelevant whether you already have backups of various sites. It's perfectly reasonable for you to create backups given your beliefs. Those beliefs though I consider insane. The solution then is not to stop creating backups, as that would accomplish nothing. The solution is to stop browsing sites that are specifically designed to make you insane.
0DataPacRat7y
Ah, but is it really paranoia if "they really are out to get you"? :) I've previously demonstrated that I'm willing to make long-shot gambles on 5% odds, given that that's roughly my estimate of cryonics working and I've signed up for it. So let's try working with that number. Out of all the possible scenarios of a Trump presidency, if you leave out 95% of the most positive options, how unpleasant is the best-remaining one? Put another way, is there at least a 5% chance of American or international politics descending to the point where my current apparent paranoia seems reasonable? And don't forget, as you calibrate your answer, that according to FiveThirtyEight, on October 17, Hillary had been predicted to have over an 88% chance of winning, implying that many people, likely including myself, have been massively mis-calibrated about how likely unpleasant political events are.
4sen7y
You don't place bets based solely on probabilities. You place bets based on probabilities, odds, timescales, investments, and alternative options. Specifically, you place bets to optimize for growth of principle with respect to time. What you're doing is not placing a bet. If you were placing a bet and wanted feedback, it would have been appropriate to provide a lot more information, such as what you expect to gain from your bet, what you expect to lose in the negative case, what you're hoping to optimize, what your expected costs are, and what alternatives you're considering spending your time or money on. It's not appropriate for you to provide any of that information because what you're doing is not placing a bet. What you're doing is panicking and looking for an echo to tell you that your beliefs are sensible, that the world really is crashing, and that what you're doing is justified. Your beliefs are not sensible, and the world isn't really crashing. I don't know if what you're doing is justified, as that would require a lot more information, but honestly I think that's irrelevant.
4DataPacRat7y
(A quick FYI, I'm about to try for a good night's sleep, then compare how I was feeling when I first posted in this thread with however I feel when I wake.)
2MrMind7y
Only very mildly. The point is that you priority should be to get out of depression: in the case of a military conflict, how much helpful will be that? This is much more important for your long-term survival than a bunch of reddit branches.
0DataPacRat7y
If you have any clue for a method on how a person can reliably accomplish that - especially if it's one that I haven't tried yet - please share. With the whole world. I trust that you won't mind if I don't plan on holding my breath.
0MrMind7y
I was talking about the meta-level, and your meta-level question was "Have I made the appropriate preparations?" to which I answered: no, the biggest improvement is if you prioritize depression treatment over any other. That said, on the object level, if you have that goal, then you would try anything sensible-sounding and any combination of anything sensible until something works. But I cannot tell you what is sensible because I'm not an expert on depression.
0DataPacRat7y
I have had that goal for some time. I have tried the sensible-sounding things, in various combinations. They didn't work. So I've been shifting my focus from "trying to keep depressive bouts from happening" to "managing my life on the assumption I'm going to keep getting depressive bouts". I've hit enough such management tricks that even with my bout last week interrupting, I'm about 60,000 words into writing a novel, including 1600 words yesterday; I could be doing better, sure, but I could be doing a lot /worse/, too.
0DataPacRat7y
As a point of interest: as of when I woke up, the votes were: LessWrong, two votes for paranoid; /r/rational, two votes for not particularly crazy. Emotionally, I'm not feeling the particular "I'm going to hate myself in January 2018 if I haven't mailed copies of my archival Blu-Ray discs to certain members of my extended family stretching halfway across the continent by then, and the Net gets taken down" urgency that I did when I posted, but it still seems like a good idea to nudge my plans in the direction of being able to handle that particular scenario with minimal losses of what I find valuable.

Oops, solved

[This comment is no longer endorsed by its author]Reply

I think the unofficial, undercover ban on basilisks should be removed. Acausal trade is an important topic and should be openly discussed.

[pollid:1171]

8Unnamed7y
As of October 2015, "The Roko's basilisk ban isn't in effect anymore."
2Dagon7y
I've seen it referenced enough times recently that I suspect the ban is no longer enforced anyway. Our wiki has a complete writeup which includes the original post.

I think the unofficial, undercover ban on basilisks should be removed. Acausal trade is an important topic and should be openly discussed.

[pollid:1170]

[This comment is no longer endorsed by its author]Reply
0MrMind7y
Sorry, I messed up the number of options. The examples in the help are wrong, though.

Let's say that I have a belief running like this: "a DAO that controls the manifacturing output of robots to produce a UBI would be the solution to the robots-stealing-jobs problem".
What would be the best move for me to influence someone into believe / try this?
Take a degree in economics? Joining some kind of foundation? Shouting from the top of a carboard box in front of the Coliseum?
What else?

4Lumifer7y
You need a VC (venture capitalist) because you're effectively proposing a startup and the most important things you need are money and talent. In parallel you can write a paper (under peer-reviewed standards) which would explain why do you think it will work and/or a set of lay posts explaining the same in accessible language. The first question that comes to my mind is why do you need a DAO here and what is the difference from a state nationalising "robots to produce a UBI". Or an everyone-gets-a-share corporation if you're particularly distrustful of the state.
2MrMind7y
Well, a DAO is a kind of formalized, cryptographically secure everyone-gets-a-share corporation, so it's not much different. But you actually gave me very good ideas, thanks.
0ChristianKl7y
Not really given the DAO that we have seen. There no cryptographically secure mechanism to give everybody a share with common definition of what cryptographically secure means. Sockpuppeting can't be solved with cryptography.
0MrMind7y
Very well, "secure" in cryptography just means "robust". I was not implying that it is impossible to hack, and since the DAO we are speaking about doesn't even exists, it doesn't make sense to discuss about its pro's and con's.
0ChristianKl7y
No. Secure in cryptography generally means the ability to make security guarantees. Trust in identity is still required. There's also no reason why a DAO that destributes money to random people won't be outcompeted by one that reinvest all resources in furthering it's own goals.
2ChristianKl7y
The first step would be to get more specific about what you actually believe. List the individual beliefs that you have and write down your credence that you have for key assumptions.
2[anonymous]7y
If you already have relevant researching skills (e.g., as a statistician), why not just hunt yourself an economist and co-work?
2MrMind7y
That might indeed be something I'm willing to try, thanks!

Richard Wong, head of engineers at Coursera, in an interview at the site lifehacker.com has declared:

I used to be a PC-only person, back during my days at Microsoft, but now I’m pretty much Apple only. It has some of the best development tools for engineers.

It beats me, though. I knew that PCs are good for gaming and developing, but which are the conclusively superior development tools for engineer? I'm confused.

3Vaniver7y
Most of the cutting-edge projects that I know have simple installation instructions in Linux/Mac, along the lines of "oh, you sudo apt-get project and then it manages all the dependencies for you and just works," whereas getting them to work in Windows is something of a "alright, you need to get X, Y, and Z, and also want W but it doesn't really work in Windows so here's a hack to work around that kind of, which might stop working whenever they upgrade." I suspect this is the impression he's pointing at.
0Lumifer7y
Yep. Writing code on Windows works reasonably well if you stay within a pre-built environment (e.g. Visual Studio, Eclipse, R Studio, etc.) Stray out of it and you'll be forced to painfully kludge together some bastardised version of Unix inside Windows. Macs are Unix machines at their core (with a pretty GUI on top) so from the writing code point of view there isn't a great difference between a Mac and, say, an Ubuntu machine. By the way, Richard Wong's favourite text editor -- Sublime Text -- is available on all three platforms.
0Lumifer7y
I'm not sure what are you asking. IMHO for development Linux/Unix environment is superior, but in specific cases it all depends on what kind of application software is available.
0MrMind7y
Yeah I wasn't very clear. Can you think of any developement tools that is definitely superior / works only on Apple?
0Lumifer7y
By "development" do you mean "writing code"? You mentioned engineering which could imply, say, CAD/CAE software, and at this level you're basically comparing applications and OSes (Win/Mac/*nix) matter only to the extent that the program you need will run on them.
0Vaniver7y
Head of engineering at Coursera means software engineering.

Would I be able to tap the LW academic network to get a copy of this paper?

Extreme gratitude in advance.

[This comment is no longer endorsed by its author]Reply
2MrMind7y
Does sci-hub.io work for you? I'm behind a firewall at the moment.

Explosions in the Sky music It's very important as a rationalist, your job is to understand the machine that you are. If what you exists, how you choose your actions, seeing through the conditioning and extreme obstacles that are limiting your growth and the growth of humanity, is very important. So study neuroscience!

A reminder that rationality is a slave to our emotions and how in line our emotions are with rationality dictates how rational our actions are, for example, from one moment to the next you can become Vegan. The disconnect between emotions an... (read more)

4RowanE7y
The usual rule is to identify as an "aspiring rationalist"; identifying rationality as what you are can lead to believing you're less prone to bias than you really are, while identifying it as what you aspire to reminds you to maintain constant vigilance.
-1ingive7y
That is mostly true, you've discovered the fallacy of most humans, they identify not out of rationality's sake but from their own comfort in most cases. Because they are not honest, they wear the facade of Rationality to rationalize their behavior even though they are not rational at all or care. Stupidity can be categorized as writing useless posts on facebook - Yudkowsky. Not being Vegan. Smoking etc. Connecting with the Way emotionally will allow you to scrutinize and redevelop your belief system. It's an observation our species has made but not reviewed.

Okay, I finished reading the book, and then I also looked at the wiki. So...

A few years ago I suspected that the biggest danger for the rationalist movement could be it's own success. I mean, as long as no one give a fuck about rationality, the few nerds are able to meet somewhere at the corner of the internet, debate their hobby, and try to improve themselves if they desire so. But if somehow the word "rationality" becomes popular, all crackpots and scammers will notice it, and will start producing their own versions -- and if they won't care about the actual rationality, they will have more degrees of freedom, so they will probably produce more attractive versions. Well, Gleb Tsipursky is already halfway there, and this Athene guy seems to be fully there... except that instead of "rationality", his applause light is "logic". Same difference.

Instead of nitpicking hundred small details, I'll try to get right into what I perceive as the fundamental difference between LW and "logic nation":

According to LW, rationality is hard. It's hard, because our monkey brains were never designed by evolution to be rational in the first place. Just to use to... (read more)

0ingive7y
If you are aware of mathematics what do you think about this part: https://logicnation.org/wiki/A_simple_click#Did_God_create_logic.3F Is it falsifiable? There was an interesting talk how something can arise out of nothing and how it's relatable to the present moment which one can't ever grasp but I will have to condense it for you guys later. I agree. That's false, if it wasn't for evolution you wouldn't have the ability to be rational in the first place. Again, I agree. Yes, the one weird trick has been observed to 'work' in the different cases although we don't know more than that right now. But someone with an understanding in neuroscience, psychology-physiology connection and can bot search the academic literature would be able to connect the dots I think. 'Logic nation' has nothing to do with this type of rationality, though, it was a mistake to deliberately say it had or use the word. No, you don't want to do that, it's unlikely people will care or understand or immediately cry wolf cult. It's also with a large likelihood an inefficient use of time. There is people who want to click so it's probably better to push resources there. If Yudkowsky clicked he would probably not call up someone, instead write an article 'to rule them all' and all of you would finally get it. But this is speculation or me giving information which can be taken accounted of after you click. Because you think differently, you probably will take some time to restructure your beliefs as the "first thing". Not completely accurate, it's only if you cannot fix something with an adequate amount of time, if you, for example, scratched your leg, you can accept the pain, thus the suffering go away instantly. Doing logical things and figuring out God (Spinoza) by science could be seen as prayer, maybe neuro or something else too. But it's speculation after all since it differs for everyone based on everything, their current knowledge (see the example of Yudkowsky) and so on. Sure, ratio
3Viliam7y
This is exactly the part I called "his bastardized version of Tegmark Multiverse + Solomonoff Induction" in my previous comment. He intruduces a few complicated concepts, without going into details; it's all just "this could", "this would", "emerges from this". To be falsifiable, there needs to be a specific argument made in the first place. Preferably written explicitly, not just hinted at. For example: "Additionally it might also help us better understand the quantum weirdness such as entanglement and superposition." -- Okay, might. How exactly? Uhm, who cares, right? It's important that I said "quantum", "entanglement" and "superposition". It shows I am smart. Not like I said anything specific about quantum physics, other than that it might be connected with ones and zeroes in some unspecified way. Yeah, maybe. Statements like "If logic is your core value you automatically try to understand everything logically" are deeply in the motte-and-bailey territory. Yeah, people who value logic, are probably more likely to try using it. On the other hand, human brains are quite good at valueing one thing and automatically doing other thing. When I try going through individual statements in the text, too many of them contain some kind of weasel word. Statements that something "can be this", "could be this", "emerges from this", or "is one of reasons" are hard to disprove. Statements saying "I have been wondering about this", "I will define this", "this makes me look at the world differently" can be true descriptions of author's mental state; I have no way to verify it; but that's irrelevant for the topic itself. -- There are too many statements like this in the text. Probably not a coincidence. I really don't want to play this verbal game, because this is an exercise in rhetorics, not rationality. I suspect it already has a name in psychology, and that it does much less than Athene claims. In psychotherapy, people have "breakthrough insights" every week, and it feels
0ingive7y
Falsifiable mathematically, a theory of everything which includes the theory itself, I mean. But sure, it allows someone to pick up the torch who are going to write a paper anyway. They are just ramblings and no real inquiry has been made to investigate these 'bathtub theories', as Elon Musk puts it. But it is an easy way to explain certain things. Too much weight shouldn't be put on it, but it would be interesting with papers, so someone who is going to publish - do this. That's not correct as doing another thing still arises out of what you value emotionally according to this theory. It does: religious experience enlightenment "wikipedia.org/wiki/Enlightenment_(spiritual)" mystical experience - nondualism We don't know if it's permanent, so far data only goes for around 1 month - 1 month 1 week. With the exception of Athene himself (creator of the experiment). But enlightenment, religious experiences etc last for awhile. Well, he probably hasn't read anything, he did apply for an LW meet-up but was rejected as he had to stay for the full amount of days, before this clicking religion thing they did reach out regarding their group on here I think, at EA forums and elsewhere. Staying there is free. Regarding rationality.org and so forth I think he mentioned they're all just intellectually masturbating. By the way, what do you think about the website: https://www.asimpleclick.org/# ?
4Viliam7y
This is Athene: This is a Less Wrong article on a similar topic: An Intuitive Explanation of Solomonoff Induction I hope you understand why I am not impressed with the Athene's version. Having to stay somewhere for a few days doesn't sound to me like a regular LW meetup. I guess it was either a CFAR workshop, or an event like this. (Uhm, this is probably not the case, but asking anyway to make sure -- "they did reach out regarding their group on here I think" does not refer to this, right? Because that's the only recent attempt to reach out here that I remember.) Heh, sometimes I have a similar impression. On the other hand, some things take time. A few years ago, superintelligent AI was a completely fringe topic... now it's popular in media, and random people share articles about it on Facebook. So either the founders of LW caused this trend, or at least were smart enough to predict it. That requires some work. MIRI and CFAR have funding, which is also not simple to achieve. They sometimes publish scientific articles. If I remember correctly, they were also involved in creating the effective altruist movement. (Luke Muehlhauser, the former executive director of MIRI, now works for GiveWell.) There is probably more, but I think this already qualifies as more than "intellectual masturbation". Athene has an impressive personal track record. I admit that part. But the whole thing about "clicking" is a separate claim. (Steve Jobs was an impressive person; that doesn't prove his beliefs in reincarnation are correct.) Any specific part of it? I have already spent hours researching this topic. I have even read the Reddit forum where people describe how they "clicked" (most posts seem the same, and so do all replies, it's a bit creepy). Am I supposed to listen to the guided meditation, or watch yet another advertising video, or...?
1niceguyanon7y
I applaud your effort and hope your hours spent means others' saved.
1Viliam7y
Thanks! I feel weird about this whole thing. Similarly how I feel weird about Gleb. I don't want to make a full conclusion for others (that feels like too much responsibility), but at least I can point them directly towards the imporant parts, so they don't have to google and watch promotional videos. Here is the good part -- a PDF booklet with some useful advice on instrumental rationality. It would make a good LW article, if some parts were removed. magnet:?xt=urn:btih:e3ade7cdccc4aba33789686b9b9d765d7f14ae7b&dn=Real+Answers&tr=udp%3A%2F%2Ftracker.leechers-paradise.org%3A6969&tr=udp%3A%2F%2Fzer0day.ch%3A1337&tr=udp%3A%2F%2Fopen.demonii.com%3A1337&tr=udp%3A%2F%2Ftracker.coppersurfer.tk%3A6969&tr=udp%3A%2F%2Fexodus.desync.com%3A6969 Here are the bad parts -- wiki (just read the main page), and reddit forum (click on a few random articles, they all feel the same, and the responses all feel the same) The rest is just marketing, hyping the contents of the wiki and of the book over and over again. This is my conclusion after ~10 hours of looking at various materials; maybe there is something more that I missed, also I didn't listen to the podcasts. This all seems to be a one man show; a guy whose main strength is making popular youtube videos, and being a successful poker player in the past.
0ingive7y
I understand, that article looks interesting. I think it was an event like you linked. The message in the bottom does look a lot like Athene's writing, but the first one I don't really understand what it is about. They have mentioned they do sports betting. Athene doesn't care about lying, only if it's for the greater good, so I speculate they have some ways to make money at betting, for example, thus they thought about trying to reach out to some high IQ people to actually off-load work or have 'em join. But this is only speculation. I think they've reached out in general and around December explaining their "charity" organization and how they offer free food, housing etc. But maybe not. That makes sense. By the way gamingforgood has a 10X multiplier on donations to their newborn survival programs, how likely is it that is is more efficent then GiveWell's top charities? I was thinking more the design, but sure, the guided meditations might be good if you are doing a self-experiment of the 4 steps and so forth, they're nothing special, what you'd expect I suppose.
3Viliam7y
Yeah, the "i was a great poker player" part... I missed that previously. OK, now I am quite confused. Here are things said by "hans_jonsson" (Athene?) in that thread: So, to put things together... a guy who according to Wikipedia is a popular YouTube celebrity, raised millions for charity, and recently started the "Logic Nation"... was contacting individual LW members through private messages, instead of posting in an open thread, because posting in an open thread would be awkward bragging... and the best way to show that he was competent, was to pseudonymously post something that closely resembles popular scams. My brain has problem processing so much logic. (Could LessWrong really be so frightening for outsiders? So much that even starting your own cult feels less awkward and more humble than posting in LW open thread...) I admit I am out of my depth here. There seems to be evidence flying in both directions, and I am confused. (Still learning towards "scam", though.)
0ingive7y
Other than that, the poor grammar and spelling is typical of Athene and lacking of paragraphs, the further you scroll in his Reddit profile the worse his grammar becomes: https://www.reddit.com/user/Chiren :) He also wrote this in the thread: >and i may very well have some mental issues in regards to quite a few things I don't take things too seriously. When you type or talk you can push buttons and see how the community reacts to forward your agenda, the best option after making a mistake might not be to say "I am X and I should've posted in the open thread" instead for example say it's embarrassing, awkward and eventually that "may very well have mental issues" Well, yes, maybe. Observing what was done objectively it seems as it was as it was said, to keep it in private? I don't really see how it is leaning towards scam, you might simply have been wrong all along. The first message I don't fully understand.
0Viliam7y
Oh, this is pure gold! :D Two months ago, on Athene's Reddit forum: For context, this is Athene on LW, nine months ago: Karma is a bitch.
0ingive7y
I remember that, pretty funny. Maybe he learned from LW and sub- and consciously understood it, responded the same way 7 months later. :) Now now, if it is him who posted here, that's simply speculation but I think it's 60-70% probability.
4NatashaRostova7y
You can't just classify things as stupid because you think they are stupid, and what you think is stupid is true because you think you're rational. The idea that 'not being vegan' or 'smoking' are stupid is silly.
1Dagon7y
I agree that "stupid" is a bad label for the clustering which includes those kinds of behaviors, but I don't agree if you're saying that smoking and meat-eating are usually instrumentally rational choices for common human desires. Stupid implies either incorrect logic or lack of consideration of an action. For many humans, these behaviors are neither one. They're some combination of weakness (knowing the better choice, but failing to override the monkey brain) and value differences (preferring current/near experienced pleasures over later/distant pain). note: I eat lots of meat. I also play lots of video games and read lots of fiction, none of which is purely rationally motivated. I don't smoke or vape, but that's also not rationally motivated - I just find it disgusting.
0NatashaRostova7y
It doesn't really matter what either of us think. If someone eats too much meat, and wishes they could stop, but can't, then for a certain function we can claim it's irrational in their achievement of that goal. If I eat a fair amount of meat because I work out, because it helps me get my weight-lifting goals, it's rational for my objective. What's your objective? Well, my main point is really just that we can't abstract these sorts of things, they are empirical. "Is X irrational (implied: for all people under all conditions)?" is about as meaningful as "Does this chair really exist?"
0ingive7y
Exactly, it's better to look at the evidence, objective reality and see what's more likely to be efficient. You presume with the latter statement that your achievement of a goal is accurate. I hope that this example is simply that, eating meat is not necessary for a positive nitrogen balance and muscle hypertrophy OR strength. It might have a slight advantage, but at that point, you'd assume you're already doing everything efficiently and your genetics are on par. Very unlikely. You realize you are biased and not in line with the objective reality of things where your desires can be replaced and come from a certain place for a reason.
0ingive7y
If you define stupidity as a set of rules that we use to ensure a problem is solved longer than chance or never and is nevertheless pursued with alacrity and enthusiasm. Then it's stupid. I don't know what the problem to be solved can be boiled down to in the context of this definition, maybe evolving as a super organism although that is not an end. General solving seems more applicable.