https://micahcarroll.github.io/
As we have seen in the former post, the latter question is confusing (and maybe confused) because the value change itself implies a change of the evaluative framework.
I’m not sure which part of the previous post you’re referring to actually – if you could point me to the relevant section that would be great!
What is more, the change that the population undergoes is shaped in such a way that it tends towards making the values more predictable.
(...)
As a result, a firms’ steering power will specifically tend towards making the predicted behaviour easier to predict, because it is this predictability that the firm is able to exploit for profit (e.g., via increases in advertisement revenues).
A small misconception that lies at the heart of this section is that AI systems (and specifically recommenders) will try to make people more predictable. This is not necessarily the case.
For example, one could imagine incentives for modifying someone's values to be more unpredictable (changing constantly within some subset) but in an area of the value-space that leads to much higher reward for any AI action.
Moreover, most recommender systems (given that they only optimize instantaneous engagement) don't really optimize for making people more predictable, and can't reason about changing the human's long-term predictability. In fact, most recsystems today are "myopic": their objective is a one-timestep optimization that won't account for much change in the human, and can essentially be thought of as ~"let me find the single content item X that maximizes the probability that you'd engage with X right now". This often doesn't have much to do with long-term predictability: clickbait often will maximize the current chance of a click but might make you more unpredictable later.
For example, in the case of recommendation platforms, rather than finding an increased heterogeneity in viewing behaviour, studies have observed that these platforms suffer from what is called a ‘popularity bias’, which leads to a loss of diversity and a homogenisation in the content recommended (see, e.g., Chechkin et al. (2007), DiFranzo et al. (2017), & Hazrati et al. (2022)). As such, predictive optimisers impose pressures towards making behaviour more predictable, which, in reality, often imply pressures towards simplification, homogenisation, and/or polarisation of (individual and collective) values.
Related to my point above (and this quoted paragraph), a fundamental nuance here is the distinction between "accidental influence side effects" and "incentivized influence effects". I'm happy to answer more questions on this difference if it's not clear from the rest of my comment.
Popularity bias and homogenization have mostly been studied as common accidental influence side effects: even if you just optimize for instantaneous engagement, often in practice it seems like this homogenization effect will occur, but there's not a sense that the AI system is "trying to bring homogenization about" – it just happens by chance, similarly to how introducing TV will change the dynamics of how people produce and consume information.
I think most people's concern about AI influencing us (and our values) comes instead from incentivized influence: the AI "planning out" how to influence us in ways that are advantageous to its objective, and actively trying to change people's values because of manipulation incentives emerged from the optimization [3, 8]. For instance, various works [1-2] have shown that recommenders which optimize long-term engagement via RL (or other forms of ~planning) will have these kinds of incentives to manipulate users (potentially by making them more predictable, but not necessarily).
Regarding grounding the discussion of "mechanisms causing illegitimate value change": I do think that it makes sense to talk about performative power as a measure of how much a population can be steered, and why we would expect firms to have incentives to intentionally try to steer user values. However, imo performative power is more an issue of AI policy, misuse, and mechanism design (to discourage firms from trying to cause value change for profit), rather than the "core mechanism" of the VCP.
In part because of this, imo performative prediction/power seem like a potentially misleading lens to analyze the VCP. Here are some reasons why I've come to think so:
The story seems a lot cleaner (at least in my head) from the perspective of sequential decision problems and RL [1-5], which makes much less assumptions about the nature of the interaction. It goes something like this (even in the best case in which we are assuming a system designer aligned with the user):
On another note, in some of our work [1] we propose a way to ground a notion of value-change legitimacy based on counterfactual preference evolution (what we call "natural preference shifts"). While it's not perfect (in part also because it's challenging to implement computationally), I believe it could limit some of the main potential harms we are worried about, and might be of interest to you.
The idea behind natural preference shifts is to consider "what would have the person's value been without the actions of the AI system", and evaluate the AIs actions based on such counterfactual preferences rather than their current ones. This ensures that the AI won't drive the person to internal states that they would have judged negatively according to their counterfactual preferences. While this might prevent beneficial legitimate preference shifts from being induced by the AI (as they wouldn't have happened without the AI), it at least can guarantee that the effect of the system is not arbitrarily bad. For an alternate description of natural preference shifts, you can also see [3].
Sorry for the very long comment! Would love to chat more, and see the full version of the paper – feel free to reach out!
[1] Estimating and Penalizing Induced Preference Shifts in Recommender Systems
[2] User Tampering in Reinforcement Learning Recommender Systems
[3] Characterizing Manipulation from AI Systems
[4] Hidden Incentives for Auto-Induced Distributional Shift
[5] Path-Specific Objectives for Safer Agent Incentives
[6] Agent Incentives: A Causal Perspective
[7] Reinforcement learning based recommender systems: A survey
Cool work and results!
Is there a reason you didn't include GPT4 among the models that you test (apart from cost)? If the results would not be as strong for GPT4, would you find that to be evidence that this issue is less important than you originally thought?