Critical Trade-offs: Ranking Objectives, Latency, and Freshness
RANKING OBJECTIVE CHOICES
Ranking models optimize different objectives. Pointwise predicts absolute score per item (click probability). Simple to train but ignores relative ordering. Pairwise compares two items to decide which ranks higher. Captures preferences but needs O(n²) pairs. Listwise optimizes metrics over full lists, directly targeting NDCG. Most aligned with evaluation but hardest to train.
WHEN TO USE EACH
Pointwise: abundant click data, predicting engagement probability. Pairwise: explicit preference signals like A/B choices. Listwise: non decomposable metrics, 10M+ training examples, dedicated ML infrastructure. Most systems start pointwise and migrate later.
LATENCY VS QUALITY
Every choice involves this tradeoff. More ranking stages increase quality but add latency. Larger candidate pools improve recall but slow ranking. Richer features improve predictions but need more lookups. Quantify with offline experiments: measure NDCG at different candidate counts, plot quality versus latency curves.
FRESHNESS VS STABILITY
Fresh signals (recent activity, trending items) improve relevance but are expensive and noisy. Common pattern: 80% stable precomputed features updated daily, 20% real time features updated per request. Too much real time signal causes volatile recommendations that confuse users.