https://michaelzuo.wordpress.com
michaelyzuo@gmail.com
Sometimes politics IS the core issue, or at least an important underlying cause of the core issue, so a blanket ban on discussing it is a very crude tool.
Because it’s effectively banning any substantial discussion on a wide range of topics, and instead replacing it, at best, with a huge pile of euphemisms and seemingly bizarre back and forths. And at worst, nothing at all.
So user competence as a factor is unlikely to be completely seperate.
Or to look at it from the other angle, in an ideal world with ideal forum participants, there would very likely be a different prevailing norm.
It seems contradictory. If LW users believe that the userbase is not competent enough on average to avoid tangential but divisive politics, then why do they believe the ’karma’ average decided by the same, matters?
It’s like a superposition of two extremes:
At one extreme there’s Reddit where a high karma is more of an anti-signal, and having extra karma beyond a pretty low threshold actually increase reader’s suspicions that it’s fluff or unusually deceiving…
At the other extreme, there are traditional old BBS forums with no karma or scoring system whatsoever. And any formal distinction is a huge positive signal.
Why do a lot of the “External reviews of “Alignment faking in large language models”” read like they were also written, or edited, by LLMs?
Are people expected to take “reviews” done seemingly pro forma at face value?
I would perhaps go even farther, most, maybe all, people don’t have any ‘sincere ethics’ whatsoever with a sharp boundary line, it’s nearly always quite muddled.
At least judging by actions.
Which upon reflection makes it sorta amazing any complex polity functions at all.
And in any case it’s probably safer to assume in business dealings that the counterparty is closer to the 50th percentile than to the 99.99th percentile.
For a test that can give certainty to claims there’s something more than that going on somewhere in the LLM. Since the OP only indicated ~90% confidence.
Otherwise it’s hard to see how it can be definitely considered ‘consciousness’ in the typical sense we ascribe to e.g. other LW readers.
Is there a definitive test? Or any prospects of such in the foreseeable future?
Such that a well informed reader can be completely confident it’s not just very fancy pattern recognition and word prediction underneath…
I already think that "the entire shape of the zeitgeist in America" is downstream of non-trivial efforts by more than one state actor. Those links explain documented cases of China and Russia both trying to foment race war in the US, but I could pull links for other subdimensions of culture (in science, around the second amendment, and in other areas) where this has been happening since roughly 2014.
This theory likely assigns too much intention to too large of a structure. The cleavage lines are so obvious in the U.S. that it wouldn’t take much more than a random PSYOP middle manager every week having a lark on a slow Friday afternoon, who decides to just deploy some of their resources to mess around.
Although it’s possible policy makers know this too and intentionally make it very low hanging fruit for bored personnel to mess around and get away with only a slap on the wrist.
The core issue, in any society, is that it’s thousands of times easier to destroy trust than to rebuild it.
Plenty of libertarians understand that some percentage of the population will degenerate when given limitless opportunities to do so. Though they usually don’t have an answer for what to do with the resulting millions of semi-deranged adults other than isolation/prison/etc...
It’s more the speed and extent that it has occurred within a short time is probably what’s surprising.
What exactly is the counterargument here?
There are a boundless number of reasons for or against anything, because real world things happen in an infinite dimensional space of possibilities… so just a listing some opposing points for those in the parent doesn’t add much.
Once you take into account real world factors, such as an expanding userbase leading to less average credibility per user, multiplying political positions, etc… which are all pretty much unavoidable due to regression to the mean…
It really becomes ever closer to an effective blanket ban, to at least try to maintain the same average quality. (Asssuming that is a goal.)
To extrapolate it to an extreme scenario, if the userbase suddenly 100X in size, then even many things considered prosaic might have to be prohibited because the userbase, on average, literally wouldn’t be capable of evaluating discussion beyond a mediocore subreddit otherwise.