Can retrieval enhancement fix explainable recommendations for sparse users?
When users have few historical interactions, embedded recommendation models struggle to generate personalized explanations. Can augmenting sparse histories with retrieved relevant reviews—selected by aspect—overcome this fundamental data limitation?
Three frameworks exist for explainable recommendation: post-hoc methods that generate explanations for a black-box model, embedded methods that learn explanations as part of model representation, and multi-task methods that share latent representations between recommendation and explanation. Post-hoc explanations have limited diversity. Embedded methods rely on historical interactions to learn representations, so they fail when the user or item has insufficient data. Multi-task methods often use Transformers but encounter limited review content, leading to performance decline in generation tasks.
ERRA's contribution is two enhancements added to multi-task learning. Retrieval enhancement is model-agnostic: identify and select the most relevant reviews from a corpus (the dataset's training set, used as a retrieval corpus rather than Wikipedia). This addresses data sparsity by augmenting the user's limited historical reviews with relevant external ones. Aspect enhancement is personalized: extract aspects from each user and item review using extraction tools, then select the most relevant aspects per user-item pair. The same user might care about environment when choosing a hotel and price when choosing a phone — aspect selection adapts to the recommendation context.
The combined effect: retrieval enhancement gives the model more information to work with when user history is sparse, and aspect enhancement makes the explanations user-specific rather than generic. The general pattern: when sparsity hits an embedding-based model, retrieval augmentation is a more effective remedy than larger embeddings — and personalizing the augmentation (aspect-aware retrieval) further improves quality.
Source: Recommenders Personalized
Related concepts in this collection
-
Do comparisons help users evaluate items better than isolated descriptions?
Can framing product evaluations relationally—by comparing to other items—ground assessment in user reasoning better than absolute descriptions? This matters because recommendation explanations often ask users to do comparison work mentally.
extends: aspects extracted from reviews are the same primitive — ERRA uses them for explanation, comparative-recommendations uses them for relational generation
-
Can LLMs explain recommenders by mimicking their internal states?
Can training language models to align with both a recommender's outputs and its internal embeddings produce explanations that are both faithful and human-readable? This explores whether dual-access interpretation solves the fundamental tension between behavioral accuracy and interpretability.
complements: surrogate-model interpretability and aspect-aware retrieval are alternative answers to the explainable-recommendation problem
-
Can attention mechanisms reveal which user taste explains each recommendation?
Single-vector user models collapse diverse tastes into one representation, losing expressiveness. Can weighting multiple personas by item relevance surface the right taste at the right time while making recommendations traceable?
complements: persona-attention and aspect-attention are parallel mechanisms for explanation — orthogonal axes of personalization
-
Why does collaborative filtering struggle with sparse user data?
Collaborative filtering datasets appear massive but hide a fundamental challenge: each user has rated only a tiny fraction of items. How does this per-user sparsity shape the modeling problem, and what techniques can overcome it?
grounds: user-history sparsity is exactly the problem ERRA's retrieval enhancement compensates for
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
explainable recommendation needs aspect-aware retrieval enhancement — sparse user history defeats embedded-only methods