Recommender Systems

Can one text encoder unify all recommendation tasks?

Does framing diverse recommendation problems—from sequential prediction to review generation—as natural language tasks allow a single model to learn shared structure? Can this approach generalize to unseen items and new task phrasings?

Note · 2026-05-03 · sourced from Recommenders Personalized
What breaks when specialized AI models reach real users?

Different recommendation tasks — sequential recommendation, rating prediction, explanation generation, conversational recommendation — historically require different architectures, different objectives, and different feature engineering. Knowledge learned for one task does not transfer to another. A sequential recommender cannot be redeployed for review generation.

P5's move is unification: convert all data formats (user-item interactions, user descriptions, item metadata, user reviews) into natural language sequences, and train one encoder-decoder model with one language modeling loss across five task families. Tasks differ only in the personalized prompt that frames them. "Predict the next item user X would interact with given history H" and "Generate a review for user X about item Y" become the same kind of input-target text pair.

P5 matches or beats representative task-specific approaches across all five families and transfers zero-shot to new items, new domains, and new prompt phrasings — generalizations that task-specific architectures structurally cannot do. The conceptual contribution: recommendation tasks share a common substrate (user-item pool, contextual features), and natural language is general enough to encode the variation. Task-specific architectures fragmented research because each task chose its own encoding; language unification reverses the fragmentation. The cost is loss of efficiency relative to specialized models, but the gain is composability — new tasks can be added by writing prompts rather than designing new models. The frontier is scaling up base models (GPT-3, OPT, BLOOM) and incorporating retrieval augmentation.


Source: Recommenders Personalized

Related concepts in this collection

Concept map
12 direct connections · 66 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

recommendation as language processing unifies tasks under one text-to-text encoder-decoder — P5 enables zero-shot transfer to new prompts and items