Psychology and Social Cognition Language Understanding and Pragmatics

Can AI systems preserve moral value conflicts instead of averaging them?

Current AI systems wash out value tensions through majority aggregation. Can we instead model how values like honesty and friendship genuinely conflict in moral reasoning?

Note · 2026-02-23 · sourced from Design Frameworks
What kind of thing is an LLM really? How should researchers navigate LLM reasoning research?

Value pluralism holds that multiple correct values may be held in tension with one another — honesty may conflict with friendship, privacy may conflict with transparency, autonomy may conflict with safety. These tensions are not resolved by choosing a winner; they are irreducible features of moral reasoning.

AI systems, as statistical learners, fit to averages by default. Supervised systems aggregate opinions through majority votes, washing out the very value conflicts that make moral reasoning meaningful. This is not a bug in current systems — it is the default behavior of any system trained to minimize loss across a labeled dataset.

ValuePrism provides a dataset of 218k values, rights, and duties connected to 31k human-written situations. The values are generated by GPT-4 and deemed high-quality by human annotators 91% of the time. Four modeling tasks make pluralism tractable:

  1. Generation — what values, rights, and duties are relevant for a situation?
  2. Relevance — is a specific value relevant for this situation? (2-way classification)
  3. Valence — does the value support or oppose the action, or might it depend? (3-way classification)
  4. Explanation — how does the value relate to the action? (post-hoc rationale)

The valence task is critical. Disentangling whether a value supports, opposes, or contextually depends is necessary for understanding how plural considerations interact. A value like "respecting autonomy" might support one action and oppose another in the same situation.

Since Should AI alignment target preferences or social role norms?, the value pluralism framework provides a mechanism: rather than aggregating to a single preference or aligning to a universal standard, the system models the full field of relevant values and their interactions. Since Do large language models develop coherent value systems?, value pluralism offers a structural alternative to emergent value coherence — explicit modeling rather than implicit emergence.


Source: Design Frameworks

Related concepts in this collection

Concept map
15 direct connections · 124 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

value pluralism requires explicitly modeling multiple values in tension rather than aggregating by majority vote