Can language models bridge the gap between critique and preference?
When users express what they dislike rather than what they want, can LLMs reliably transform those critiques into positive preferences that retrieval systems can actually use?
In conversational recommendation people often state preferences as critiques of the current candidate rather than as positive descriptions of what they want. "It doesn't look good for a date" tells you something the user wants — a date-suitable place — but expressed as the negation of a property of the current option. Conventional retrieval systems can't directly act on critiques because their indexes match positive descriptors of items, not negations of properties.
The proposal is to use a large neural language model in few-shot mode to transform the critique into a positive preference. "It doesn't look good for a date" becomes "I prefer more romantic." The transformed preference is then used to retrieve reviews that mention the matching positive aspect — "Perfect for a romantic dinner" becomes a candidate review.
This works because LLMs can perform the common-sense inference required to convert a negation into a preference: knowing that "good for a date" implies "romantic" or "intimate," that the negation of one suggests the affirmation of an opposite, and that the relevant aspects to surface depend on the domain. Few-shot prompting with examples is enough to elicit this transformation; no fine-tuning is required.
The architectural pattern is general: when user feedback is naturally expressed in a form the indexing system can't consume, use an LLM as a translator between the feedback and the index's vocabulary. The LLM doesn't need to be the recommender — it just needs to bridge the linguistic gap between user expression and retrieval representation. This separates the conversational interface from the retrieval infrastructure cleanly, which means the retrieval can stay efficient (review embeddings) while the interface becomes natural.
Source: Recommenders Conversational
Related concepts in this collection
-
Why do queries and documents occupy different embedding spaces?
Queries and documents express the same information in fundamentally different ways—short and interrogative versus long and declarative. Understanding this mismatch is crucial for why direct embedding retrieval often fails.
complements: HyDE generates a hypothetical answer to bridge the query-document gap; critique-to-preference generates a hypothetical positive preference to bridge the negation-vocabulary gap — same architectural pattern in a different domain
-
Can implicit feedback reveal both preference and confidence?
When users take implicit actions like purchases or watches, do those signals carry two separable pieces of information: what they prefer and how certain we should be? Explicit ratings can't make that distinction.
complements: critiques are a third feedback type beyond explicit and implicit — natural-language negative signal that transforms into preference
-
Can unified policy learning improve conversational recommender systems?
This explores whether formulating attribute-asking, item-recommending, and timing decisions as a single reinforcement learning policy outperforms treating them as separate components. The question matters because joint optimization could improve conversation quality and system scalability.
complements: critique-handling is a sub-policy within the broader CRS policy space
-
Can users steer recommendations with natural language at inference?
Can recommendation systems let users specify their preferences in natural language at inference time without retraining? This matters because it would let new users and existing users dynamically adjust what they want to see.
extends: both let users steer recommendations via natural language at inference time; preference discerning starts from positive preferences while critique transformation starts from negative ones
-
Why do users drift away from their original information need?
When users know their knowledge is incomplete but cannot articulate what's missing, do they unintentionally shift topics? And can real-time systems detect this drift?
complements: critiques surface as users discover what they don't want — the negation expresses an articulation gap the LLM bridges
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
critique-to-preference transformation enables retrieving better recommendations from natural negative feedback