For what it's worth, Inkhaven seems awesome—among the best things that Lightcone has done recently, I think. I regret that I'm not participating.
I think it's not uncommon for people to call things they don't like "religions", as a way to tacitly assert that the followers of some movement or idea are dogmatic without directly claiming it. The stronger version is calling an idea or an idiology "a cult".
See this nicely collected list of examples courtesy of Scott, in an essay that aderesses the topic:
On the last Links thread, Eric Raymond claims that environmentalism is a religion. It has “sins” like wasting energy and driving gas-guzzling SUVs. It has “taboos” like genetically modified foods. It has an “apocalypse” in the form of global warming. It even has “rituals” in the form of weekly recycling.
This reminds me of an article I read recently claiming that transhumanism is a religion. But also of the article claiming that social justice is a religion. Also, liberalism is a religion. And conservativism is a religion. Libertarianism is a religion. Communism is a religion. Capitalism is like a religion. Objectivism is a religion. An anthropologist “confirms” that Apple is a religion. But UNIX is also a religion (apparently Linux was the Protestant Reformation).
I claim that I am unusually Good (people who know me well would agree—many of them have said as much, unprompted). This is not how it works for me.
It's also plausible to me that I am more coming at this from a deontological feeling of "One should not kill everyone if one has a good reason" rather than "The world is net positive".
I agree that these are importantly different, and easily conflated!
Another source of prediction error arises not from the mismatch between model and reality, but from tension between internal models.
Is this a standard element of Predictive Processing, or are you generalized / analogizing the theory in a general way?
I'm familiar with the prediction error that results in diffs between sense data and generative models, but not between different generative models.
Your use of hyperlinks is very amusing to me.
You can't save the world without working with people at least as annoying as John.
This is a great quote, and one that I should keep in mind myself.
My coworkers are pretty amazing in how not annoying they are (they're low ego, interested in changing their mind when they're wrong, smart, invested in getting better over time) and I still sometimes feel an urge to quit in a huff.
I think all the stories and adventures and loves and lives that people in the world have lived are worth quite a lot of torture, and it's not naively the case that if the torturous experiences are larger than the other experiences, that this means they're more important.
For you, is this a quantitative question, or an "in principle" question? Like could there exist some amount of extreme suffering, for which you would judge them to be outweighing the meaningful and worthwhile experiences?
Or is the sentiment more like "if there exists a single moment of meaning, that redeems all the pain and suffering of all of history"?
I think I no longer buy this comment of mine from almost 3 years ago. Or rather I think it's pointing at a real thing, but I think it's slipping in some connotations that I don't buy.
This view seems to put forward that all the deontological constraints of an agent must be "dumb" static rules, because anything that isn't a dumb static rule will be dangerous maximizer-y consequentialist cognition.
I don’t buy this dichotomy, in principle. There’s space in between these two poles.
An agent can have deontology that recruits the intelligence of the agent, so that when it thinks up new strategies for accomplishing some goal that it has it intelligently evaluates whether that strategy is violating the spirit of the deontology.
I think this can be true, at least around human levels of capability, without that deontology being a maximizer-y goal in of itself. Humans can have a commitment to honesty without becoming personal-honesty maximizers that steer the world to extreme maxima of their own honesty. (Though a commitment to honesty does, for humans, in practice, does entail some amount of steering into conditions that are supportive of honesty.)
However, that’s not to say that something like this can never be an issue. I can see three problems.
Claude is relatively Helpful, Harmless, and Honest now, but mega-Claude that is trained continually on profit metrics from the 100,000 businesses it runs and sales-metrics on the billions of sales calls it does a year, etc, probably ends up a good deal more ruthless (though not necessarily ruthless-seeming, since seeming ruthless isn't selected for by that training).
This both seems like it might be resolvable with very careful and well-tested training setups, but it also seems like maybe the biggest issue, since I think there will be a lot of incentive to move fast and break things instead of being very slow and careful.