While I find the Socrates analogy vivid and effective, I propose considering critics on posts under the same bucket as lawyers. Where Socrates had a certain set of so-called principles-- choosing to die for arbitrary reasons, I find that most people are not half as dogmatic as Socrates, and so the analogy/metaphor seems to slip short.
While my post is sitting at negative two, and no comments or feedback... Modeling commenters as if they were lawyers might be better? When the rules lawyers have to follow shows up, lawyers (usually) do change their behavior, ...
Part of the problem with verifying this is that the number of machine learning people who got into machine learning due to lesswrong. We need more machine learning people whom were able to come to doom conclusions of their own accord, independent of hpmor etc, as a control group.
As far as I can tell, the number worried about doom overlap 1:1 with lesswrong posters/readers, and if it was such a threat, we'd expect there to be some number of people coming to the conclusions independently/of their own accord.
I met an ex-Oracle sales guy-turned medium-bigwig at other companies once.
He justified it by calling it "selling ahead", and it started because the reality is that if you tell customers no, you don't get the deal. They told the customers they would have requested features. The devs would later get notice when the deal was signed, and no one on management ever complained, and everyone else on his "team" was doing it.
Any preferred critiques of Graham's Hierarchy of Disagreements?
https://en.m.wikipedia.org/wiki/File:Graham's_Hierarchy_of_Disagreement-en.svg
Extra layers or reorderings?
I'm avoiding terms like "epistemic" and "consequential" and such in this answer, and instead attempting to give a colloquial one, to what I think is the spiritual question.
(I'm also deliberately avoiding iterating over the harms of blind traditionalism and religious thinking. Assuming since you're atheist, and you don't reject most of the criticisms of religion)
(Also also, I am being brief. For more detail I would point you at the library, to go reading on Christianity's role for the rise of the working and uneducated classes in the 1600s-1800s, and ...
Perhaps "term" is the wrong, ahem, term.
Maybe you want "metrics"? There's lots of non-GDP metrics that could be used to track ai's impact on the world.
Instead of the failure mode of saying "well, GDP didn't track typists being replaced with computers," maybe the flipside question is "what metrics would have shown typists being replaced?"
If you find yourself leaning into conspiracy theories, one should consider whether they're stuck in a particular genre and need to artificially inject more variety into their intellectual, media, and audio diets.
Confirmation bias leads to one feeling like every song has the same melody, but there are many other modes of thought, and, imo, sliding into a slot that checks more boxes with <ingroup> is an indicator our information feeds/sensors are bad more than that we are stumbling on truth.
What kind of information would you look out for, that would make you change your mind about alignment-by-default?
What information would cause you to inverse again? What information would cause you to adjust 50% down? 25%?
I know that most probability mass is some measure of gutfeel, and I don't want to introduce nickel-and-diming, more get a feel for what information you're looking for.
Do you believe encouraging the site maintainers to implement degamification techniques on the site would help with your criticisms?
When you can predict that beliefs won't update towards convergence, you're predicting a mutual lack of respect and a mutual lack of effort to figure out whose lack of respect is misplaced.
Are you saying that the interlocutors should instead change to attempting to resolve their lack of mutual respect?
As a relatively new person to lesswrong, I agree.
The number of conversations which I've read which end in either party noticeably updating one way or the other have been relatively rare. The one point I'm not sure if I agree with is being able to predict a particular disagreement is a problem?
I suppose being able to predict the exact way in which your interlocutors will disagree is the problem? If you can foresee someone disagreeing in a particular way, and then accounting for it in your argument, and then they disagree anyway, in the exact way you tried to address, that's generally just bad faith.
(though sometimes I do skim posts, by god)
I try to ask myself whether the tenor of what I'm saying overshadows definitional specificity, and how I can provide a better mood or angle. If my argument is not atonal - if my points line up coherently, such that a willing ear will hear, definitionalist debates should slide on by.
As a descriptivist, rather than a prescriptivist, it really sucks to have to fall back on Socratic methods of pre-establishing definitions, except in highly-technical locations.
Thus, I prefer to avoid arguments which hinge on definitions altogether. This doesn't preclude example...
I don't think there's any place quite like lesswrong on the entire internet. It's a lot of fun to read, but it tends to be pretty one-note, and even if there is discord in lesswrong's song, it's far more controlled, Eru Illuvitar's hand can yet be felt, if not seen. (edit: that is to say, it's all the same song)
For the most part, people are generally-tolerant of Christians. There is even a Catholic who teaches (taught?) at the Center For Applied Rationality, and there's a few other rationalist-atheists who hopped to christianity, though I can't remember th...
I don't agree, but for a separate reason from trevor.
Highly-upvoted posts are a signal of what the community agrees with or disagrees with, and I think being able to more easily track down karma would cause reddit-style internet-points seeking. How many people are hooked on Twitter likes/view counts?
Or "ratio'd".
Making it easier to track these stats would be counterproductive, imo.
It seems pretty standard knowledge among pollsters that even the ordering of questions can change a response. It seems pretty blatantly obvious that if we know who a commenter is, that we will extend them more or less charity.
Even if the people maintaining the site don't want to hide votes + author name on comments and posts, it would be nice if user name + votes were moved to the bottom. I would like to at least be presented with the option to vote after I have read a comment, not before.
The reason why I don't want a scratch-space, is because I view scratch space and context equivalent to giving the ai a notecard that it can peek at. I'm not against having extra categories or asterisks for the different kinds of ai for the small test.
Thinking aloud and giving it scratch space would mean it's likely to be a lot more tractable for interpretability and alignment research, I'll grant you that.
I appreciate the feedback, and I will think about your points more, though I'm not sure if I will agree.
Given my priors + understanding of startups/silicon valley culture, it sounds more like Openai started to leave the startup phase and is entering into the "profit-seeking" phase of running the company. After they had the entire rug pulled under them by stable diffusion, I would expect them to get strategic whiplash, and then decide to button things up.
The addition of a non-compete clause within their terms of service and the deal with Microsoft seems to hint towards that. They'll announce GPT-3.5, and "next-gen language models", but it doesn't match my priors that they would hold back GPT-4 if they had it.
Time may tell, however!
What I'm asking with this particular test is, can an ai play blindfold chess, without using a context in order to recant every move in the game?
I'm confused. What I'm referring to here is https://en.wikipedia.org/wiki/Blindfold_chess
I'm not sure why we shouldn't expect an ai to be able to do well at it?
My proposed experiment / test is trying to avoid analogizing humans, but rather scope out places where the ai can't do very well. I'd like to avoid accidentally overly-narrow-scoping the vision of the tests. It won't work with an ai network where the weights are reset every time.
An alternative, albeit massively-larger-scale experiment might be:
Will a self-driving car ever be able to navigate from one end of a city to another, using street signs and just learning the streets by exploring it?
A test of this might be like the following:
Thanks. fyi, i tried making the post i alluded to:
I wonder how long it's going to be until you can get an LLM which can do the following with 100% accuracy.
I don't care about the ai winning or losing, in fact, I would leave that information to the side. I don't care if this test is synthetic, either. What I want is:
I think the lesswrong community is wrong about x-risk and many of the problems about ai, and I've got a draft longform with concrete claims that I'm working on...
But I'm sure it'll be downvoted because the bet has goalpost-moving baked in, and lots of goddamn swearing, so that makes me hesitant to post it.
Something possibly missing from the list, is breadth of first-hand experience amidst other cultures. Getting older and meeting people and really getting to know them in such a short lifespan is really, really hard!
And I don't just mean meeting people in the places we already live. Getting out of our towns and countries and living in their worlds? Yeah you can't really do that. Sure, you might be able to move to <Spain> or <the Phillippines> for a couple years, but then you come home.
It's not just death here, but the breadth of experiences we ca...
Feel free to delete this if it feels off-topic, but on a meta note about discussion norms, I was struck by that meme about C code. Basically, the premise that there is higher code quality when there is swearing.
I was also reading discussions in the linux mailinglists- the discussions there are clear, concise, and frank. And occasionally, people still use scathing terminology and feedback.
I wonder if people would be interested in setting up a few discussion posts where specific norms get called out to "participate in good faith but try to break these specif...
As a matter of pure category, yeah, it's more advanced than "don't make stuff up".
I usually see these kinds of guides as an implicit "The community is having problems with these norms"
If you were to ask me "what's the most painful aspect about comments on lesswrong?", it's reading comments that go on for 1k words a piece and neither commenter ever agrees, and it's probably the most spooky part for me as a lurker, and made me hesitant to participation.
So I guess I misread the intent of the post and why it was boosted? I dunno, are these not proposals ...
(Porting and translating comment here, because this post is great):
Goddamn I wish people would just tell me when the fuck they're not willing to fucking budge. It's a fucking waste of time for all parties if we just play ourselves to exhaustion. Fuck, it's okay to not update all at once, goddamn Rome wasn't built in a day.
I'm going to be frank, and apologize for taking so long to reply, but this sounds like a classic case of naivete and overconfidence.
It's routinely demonstrated that stats can be made to say whatever we want and conclude whatever the person who made them wants, and via techniques like the ones used in p-hacking etc, it should eventually become evident that economics are not exempt from similar effects.
Add in the replication crisis, and you have a recipe for disaster. As such, the barriers you need to clear: "this graph about economics- a field known for att...
It's details like these that you point out here, which make me SUPER hesitant when reading people making claims about correlating GDP/economy-based metrics with anything else.
What's the original charts base definitions, assumptions, and their error bars? What's their data sources, what assumptions are they making? To look at someone's charts over GDP and then extrapolate and finally go "tech has made no effect", feels naive and short-sighted, at least, from a rational perspective- we should know that these charts tend not to convey as much meaning as we'd like.
Having a default base of being extremely skeptical of sweeping claims based on extrapolations on GDP metrics seems like a prudent default.
This was an extremely enjoyable read.
Good fun.