Recommender Systems

Why does multinomial likelihood work better for ranking recommendations?

Explores whether the choice of likelihood function in VAE-based collaborative filtering matters for matching training objectives to ranking evaluation metrics. Why items should compete for probability mass.

Note · 2026-05-03 · sourced from Recommenders Architectures
What breaks when specialized AI models reach real users?

Variational autoencoders for collaborative filtering had been studied with Gaussian and logistic likelihoods, both of which let each item-prediction be independent — high probability on one item doesn't reduce probability on another. Liang et al. show that switching to a multinomial likelihood produces state-of-the-art results, and the mechanism explains why.

In a multinomial model the predicted probabilities over items must sum to 1. Items compete for limited probability mass. To put high probability on the items the user is likely to click, the model must lower probability on items the user is unlikely to click. This is structurally what top-N ranking demands: the goal is to put the right items at the top, which means pushing the wrong items down. Gaussian and logistic likelihoods don't encode this competition, so they optimize a target that is one step removed from the evaluation metric.

The second contribution is reinterpreting the standard VAE objective as over-regularized in this setting. The KL term, calibrated for image generation, suppresses the latent code too aggressively for sparse-implicit-feedback data. Adjusting the regularization recovers performance. Together these give a principled recipe for VAE-based CF that finally beats simpler baselines.

The general lesson: choice of likelihood is not a routine modeling decision. It encodes assumptions about what kind of competition exists between predictions, and matching that to the evaluation metric matters more than choice of architecture.


Source: Recommenders Architectures

Related concepts in this collection

Concept map
15 direct connections · 85 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

multinomial likelihoods outperform Gaussian and logistic for collaborative filtering because they enforce probability competition between items