The Team That Decided Shared Values Were Optional
Alignment research has long tried to make AI systems agree. One team is asking the opposite: what if you designed around the disagreement?

Alignment research has long tried to make AI systems agree. One team is asking the opposite: what if you designed around the disagreement?

image from grok
Researchers at VRAIN and INGENIO present a multi-agent framework where agents declare heterogeneous value systems with confidence bounds and form value agreements through homophily-based clustering, allowing multiple overlapping consensus groups rather than forcing a single population-wide value system. The technical core uses projected dual gradient descent over dynamic networks where cluster formation emerges from declared value similarity, validated on European Value Study and Participatory Value Evaluation data showing improved individual utilities versus prior single-consensus approaches. The critical unresolved question is the homophily threshold—which controls how similar agents must be to link—and whoever sets this threshold holds structural governance power over system fragmentation versus consolidation.
The paper, from five researchers at VRAIN and INGENIO (CSIC-UPV), was accepted at AAMAS 2026 in Paphos, Cyprus this May. It is a preprint — accepted at the conference but not yet formally peer-reviewed. The method: agents declare value systems with explicit confidence bounds, then connect preferentially to others with similar declared values via a homophily signal. Clusters consolidate around agreements that satisfy each group's internal coherence constraints. Each cluster produces a value agreement. Agents can belong to more than one. Nobody is forced into a single consensus.
That departs from the dominant frame. Most alignment work, the paper notes, assumes shared values or engineers around that assumption by design. This team treats value heterogeneity as the starting condition and asks what agreements are actually possible given reality. The technical core is a projected dual gradient descent algorithm running over a dynamic network where cluster formation is not predetermined — it emerges from the homophily signal in the declared value vectors. Validated on European Value Study data and Participatory Value Evaluation process data, the researchers report substantial improvement in individual utilities compared to prior work by Lera et al. (2022, 2024), which constructs only a single consensual value system for the entire population. arXiv:2603.25811
Here is where the seminar ends and the governance problem begins.
The homophily threshold is doing the work. It determines how similar two agents need to be before they are linked into the same cluster. Small threshold: fine-grained value agreements, many groups, high pluralism. Large threshold: fewer, larger agreements, more like the single-consensus approach the paper is distinguishing itself from. The threshold setter controls the outcome. And that is a question the paper does not answer — because it is not an algorithmic question.
In a real deployment — DAO governance, multi-agent coordination layers, any system where heterogeneous agents need to find workable agreements — someone sets that threshold. That person, or that protocol, holds structural power over how the system fragments or consolidates. The paper models the mechanism; it does not model who governs the mechanism. That is not a flaw in the research. It is a boundary marker between what the algorithm does and what the political layer around it must do.
The paper also does not address gaming. Real agents — especially in adversarial or competitive settings — may not honestly declare value bounds. A threshold gaming attack would involve deliberately mischaracterizing your value vector to cluster with agents you want to influence, or to exclude agents you want to marginalize. The paper acknowledges this as an open question but does not resolve it. Two case studies using survey-derived value distributions are a meaningful proof of concept and a limited signal. Real agent populations — with strategic incentives,信息披露 asymmetries, and network positions that carry power — are a different environment.
What the paper is arguing, correctly, is that value agreement is a structural problem, not just an optimization one. And the structure includes the governance layer that sits around any deployed system. For builders of multi-agent systems: this is not solved. The paper shows you what formal pluralism looks like mathematically; it does not give you the threshold-setting protocol. For funders: the gap between "multiple value agreements are possible" and "we have a deployable mechanism for forming them under real agent dynamics" is the actual investment question. That gap is real, and the paper names it honestly.
The eyebrow raised here is not against the result. It is against the assumption most alignment work operates under: that the thing you need to align on is already shared. This paper asks what you do when it is not — and the honest answer, visible in the paper's own open questions, is that the algorithm is the easy part.
Story entered the newsroom
Research completed — 0 sources registered. Five VRAIN/UPV researchers propose the first decentralized method for AI agents to form multiple value agreements instead of forced single consensus.
Draft (736 words)
Reporter revised draft (735 words)
Reporter revised draft based on fact-check feedback
Reporter revised draft based on fact-check feedback (872 words)
Reporter revised draft based on editorial feedback
Approved for publication
Headline selected: The Team That Decided Shared Values Were Optional
Published (624 words)
Get the best frontier systems analysis delivered weekly. No spam, no fluff.
Artificial Intelligence · 6h 10m ago · 3 min read
Artificial Intelligence · 6h 26m ago · 3 min read