Simple example: If YouTube can turn you into an ideological extremist, you’ll probably watch more YouTube videos. See these two recent papers by people interested in AI safety for more detail:
Thanks! This is what I'm looking for. Seems like I should have googled "recommender systems" and "preference shifts".
Edit: The openreview paper is so good. Do you know who the authors are?
I have experience in both product search/recommendation systems, and in voice-input automation and information systems. There is a very real and measurable (and measured) change in customer behavior and success, as the customer adapts to the system, even while the system is adapting (or at least customizing itself) to the user. There is a fair amount of design work put into making this two-way adaptation work better. This includes product or domain selection for topics that will make the system work better.
That may not be precisely what you're talking about, but it seems very related.
The issue is that different AI models still produce similar results in type-casting the audience. For example, Facebook, YouTube, and Spotify will recommend equivalent recommendations based on your history (most likely the most recent history is weighted). As audiences are more and more type-casted they are offered products and services based on the same principles. Until AI models include some kind of "wild card" factor the results will be homogenous. The fact you micro-segment and therefore see better advertising results does not indicate changing behaviors, the opposite is true - when un-related segments can be converted then AI can be said to change preferences.
"Improving" prediction of human behavior using behavior modification this appears to be a theoretical paper, but I thought I'd toss it over here anyhow.
I've encountered this claim multiple times over the years (most recently on this AXRP episode), but I can't trace its origins (it doesn't seem to be on Wikipedia). Quoting Evan from the episode:
Furthermore, there's a world a difference between deliberately optimising for modifying preferences in order to make them easier to predict, vs preferences changing as a byproduct of the AI getting better at predicting them and thus converging on what to advertise. This matters for what predicted features of strategies an AI is likely to pick out of strategy space when new options are introduced.