Recommender Systems Language Understanding and Pragmatics Psychology and Social Cognition

Do LLM explanations faithfully describe their recommendation process?

When LLMs recommend items to groups, do their explanations match how they actually made the choice? This matters because users trust explanations to understand AI decision-making.

Note · 2026-05-03 · sourced from Recommenders Architectures
What breaks when specialized AI models reach real users? How do people build trust with conversational AI?

When LLMs are asked to make group recommendations from individual member preferences, the outputs converge on Additive Utilitarian aggregation — picking items with the highest sum of all members' ratings. This is the consensus-based strategy from social choice theory. The behavior is consistent across uniform and divergent group structures.

The disconnect is in the explanations. Asked to explain its recommendation procedure to a layperson, the LLM doesn't say "I summed the ratings" — it cites averaging (which is similar to but not identical to ADD), user or item similarity, diversity, undefined popularity metrics, and ad-hoc thresholds. Different LLMs invent different procedures: Llama tends to cite user similarity, while Mistral and Phi cite diversity in the recommendation list. These claimed procedures don't match the behavioral output.

This makes LLM explainers unreliable narrators. They generate recommendations one way and explain them another way, and the explanation is plausible enough that a user accepts it. As item set size grows, the mention of similarity and diversity in explanations increases (suggesting the LLM is performing post-hoc justification harder when more items make the choice less defensible) while the use of "undefined popularity" decreases. The implication for group recommender systems built on LLMs: the explanation layer cannot be trusted to faithfully describe what the model did, even though that's its stated purpose.


Source: Recommenders Architectures

Related concepts in this collection

Concept map
15 direct connections · 157 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

LLM group recommendations resemble additive utilitarian aggregation but explanations claim multiple criteria — explainers as unreliable narrators