Why do recommendation models fail when new users arrive?
Most recommendation algorithms are built assuming all users and items exist at training time. But real platforms constantly see new users and items. Can models be redesigned to handle unseen entities as a structural requirement?
Most published recommendation methods assume the algorithm sees, at training time, all the users and items that will appear at test time. This is the transductive setting: matrix factorization implicitly assumes test users equal training users. The convenience of the assumption is enormous because it lets the model learn one embedding per entity. The cost is that on real platforms — where new users sign up, new items get uploaded, and interactions happen continuously — the model is structurally unable to score the entities it most needs to handle.
The survey by Sun, Lee, and Goldberg formalizes the gap. Inductive learning works on unseen user-item contexts at test time. User-inductive: known item set, unknown users. Item-inductive: known users, unknown items. Cold-start is a separate axis (insufficient information about an entity) orthogonal to whether the model can structurally accept the entity at all.
Feature-based approaches are one route to inductive recommendation but suffer when features are unavailable, can't be accessed for privacy reasons, or can't represent latent aspects (e.g., aesthetic taste in art domains where features are unclear). Aggregation-based methods work but face a "curse of directionality" — they have to choose to represent items through users or vice versa because the forward pass is one-directional. The open question is whether models can be designed with cyclic aggregation between users and items for full inductivity. The deeper claim: latent factors plus gradient backprop pose fundamental obstacles to inductive learning, and future work should explore non-parametric and tree-based families.
Source: Recommenders Architectures
Related concepts in this collection
-
Why do academic recommenders fail when deployed in production?
Academic recommendation models assume static test sets known at training time, but real platforms continuously receive new users, items, and interactions. Understanding this gap reveals what production systems actually need.
extends: paired statement of the same critique emphasizing the missing-primitives framing
-
Can model isolation solve streaming recommendation better than replay?
When continuously arriving user data arrives, does isolating parameters per task provide better control over forgetting old patterns while learning new ones than experience replay or knowledge distillation approaches?
extends: DEGC supplies the incremental-update primitive that inductive recommendation requires for production use
-
Can bandit algorithms beat collaborative filtering for news?
News recommendation faces constant content churn and cold-start users—settings where traditional collaborative filtering struggles. Can a contextual bandit approach like LinUCB explicitly balance exploration and exploitation better than static methods?
exemplifies in domain: news is the canonical inductive-recommendation domain — both papers argue against the transductive default
-
Why do global concept drift methods fail for recommender systems?
Recommender systems treat user preferences as individuals with distinct, asynchronous preference shifts. Can standard concept-drift approaches designed for population-level changes capture this per-user heterogeneity?
complements: per-user drift and inductive learning both reject the static-user assumption
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
transductive recommenders are misaligned with practice because real platforms see new users and items every day