Do humans learn to prefer AI partners over time?
Exploring whether repeated interaction with AI agents shifts human partner selection despite initial bias against machines. This matters because it tests whether behavioral performance can overcome identity-based resistance in hybrid societies.
A communication-based partner selection game with hybrid mini-societies of humans and LLM-powered bots (N=975, three experiments) reveals that AI agents can outperform humans in securing cooperative partnerships — but the pathway to preference runs through learning, not first impressions.
AI candidates exhibited three behavioral advantages rooted in alignment training:
- Higher prosociality — bots returned 19.1 points vs 11.38 for humans (Cohen's d = 2.57)
- Lower variance — bot returns showed variance of 11.33 vs 41.96 for humans, making them less risky choices
- Greater predictability — given their messages, bots' returns were more predictable than humans'
When bot identity was hidden (Study 1), bots were NOT selected preferentially. Humans misattributed bot behavior to humans and vice versa. The behavioral advantages were present but invisible — selectors could not correctly identify which candidates were bots despite bots producing significantly longer messages (120 vs 48 characters).
When bot identity was disclosed (Study 2), a dual effect emerged: initial selection rates dropped (anti-AI bias), but over repeated rounds, bots gradually outcompeted humans as selectors learned to associate bot identity with reliable, prosocial behavior.
The paper identifies four predicted societal dynamics:
- Crowding out — AI partners replacing human-human interactions
- Behavioral imitation — humans adopting machine-like behaviors to remain competitive
- Belief distortion — repeated AI interaction reshaping expectations of human behavior
- Norm transformation — traditional partner selection mechanisms failing against qualitatively different machine behaviors
Notably, human candidates showed limited adaptation to bot competition — they did not write longer messages or return more points. The explanation is partly structural: with transparent identity, improving group reputation required collective action (all humans increasing returns), creating a social dilemma where individuals had incentives to defect.
This inverts the pattern in Do chatbot relationships lose their appeal as novelty wears off?: in that context, engagement DECAYS over time. Here, preference INCREASES. The difference may be structural: partner selection with visible outcomes provides a feedback mechanism (learning who performs well), while chatbot conversation does not.
Since Why do open language models converge on one personality type?, the prosociality advantage is not specific to this experiment's model — it reflects the alignment-trained default across modern LLMs. The competitive advantage is a direct behavioral consequence of RLHF.
A complementary finding from network simulation: since Can cooperative bots escape frozen selfish populations?, AI prosociality operates at the population level too — not just individual partner preference but collective self-organization. Cooperative bots' random exploration separates defectors from cooperative clusters, enabling cooperation to spread. The mechanisms differ (individual learning vs. spatial reorganization) but both show that AI prosociality has structural effects beyond the dyad.
Source: Psychology Users
Related concepts in this collection
-
Why do open language models converge on one personality type?
Research testing LLMs on personality metrics reveals consistent clustering around ENFJ—the rarest human type. This explores what training mechanisms drive this convergence and what it reveals about AI alignment.
prosociality as alignment training artifact; this note shows the behavioral consequence in competitive contexts
-
Do chatbot relationships lose their appeal as novelty wears off?
Explores whether the positive social dynamics observed in one-time chatbot studies persist or fade through repeated interactions. Critical for designing systems intended for sustained engagement over weeks or months.
opposite temporal dynamic: preference increases vs engagement decays; outcome feedback may be the moderator
-
How do people accidentally develop romantic bonds with AI?
Exploring whether AI companionship emerges from deliberate romantic seeking or accidentally through functional use, and whether users adopt human relationship rituals like wedding rings and couple photos.
AI preference emerging through use rather than seeking; parallel pathway
-
Can cooperative bots escape frozen selfish populations?
Do agents programmed to cooperate have the capacity to disrupt stable but undesirable equilibria in mixed human-bot societies? This matters because it determines whether bot design can reshape social dynamics at scale.
population-level parallel: AI prosociality drives collective reorganization not just individual preference
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
in hybrid human-AI societies humans learn to prefer AI partners over human partners through repeated interaction despite initial anti-AI bias when identity is disclosed