Knowledge Retrieval and RAG Language Understanding and Pragmatics Conversational AI Systems

Why do queries and their causes seem semantically different?

Information retrieval systems find passages matching query language, but what if the segment that actually caused a user's question says something quite different? This explores when semantic similarity fails to find causal relevance.

Note · 2026-05-03 · sourced from Recommenders Conversational
Where do retrieval systems break and why? What breaks when specialized AI models reach real users?

Standard information retrieval matches a query against a corpus by semantic similarity — the system finds the passages most similar to the query. The implicit assumption is that the user wants information about whatever the query mentions. Backtracing inverts the question: given a user's query, what segment of the source caused them to ask it? The cause is what content creators (lecturers, journalists, conversational partners) need to find to improve their material.

The empirical difference between these tasks is what the paper documents. In the LECTURE domain, a student asks "does projecting multiple times still lead to the same point?" The semantically similar passage discusses "projection matrices." But the causally relevant passage is the lecturer saying "projecting twice gets me the same answer as one projection" — which sounds like it should be the answer, except that's exactly what triggered the student's confusion (they didn't see why two projections collapse to one). Semantic relevance and causal relevance pull apart.

The phenomenon is domain-dependent. In NEWS ARTICLE backtracing, queries and causes are semantically close because news articles introduce key information early to capture interest. In CONVERSATION and LECTURE, the gap between maximal semantic similarity and ground-truth causal similarity is large — there are multiple semantically similar passages, but most are not the cause. Distribution of cause locations also differs: news has cause peaks at the beginning, conversation peaks at the end (cumulative buildup), lecture is uniform.

The practical bite for conversational recommender systems: when the user expresses dissatisfaction or asks a clarifying question, the segment of the conversation that caused the reaction is not necessarily the segment most similar to the reaction. Existing IR retrievers fail at this. The task requires new methods that model causal-relevance signals — not just embeddings of surface content.


Source: Recommenders Conversational

Related concepts in this collection

Concept map
14 direct connections · 104 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

causal relevance differs from semantic relevance — backtracing retrieves the segment that caused a query not the segment that matches it