What is Learning to Rank and How Does It Differ From Standard Classification?
Why Standard Classification Fails
Classification predicts "relevant" or "not relevant" per item. Two problems: First, you lose ordering information within the relevant set (is item A more relevant than item B?). Second, the model optimizes per-item accuracy, not list quality. A model can have 90% accuracy but produce terrible rankings if it misclassifies the few items users actually want.
The Three LTR Approaches
Pointwise: Predict a relevance score per item, then sort. Simple but ignores item relationships. Pairwise: Predict which of two items is more relevant. Captures relative ordering but doesn"t optimize the full list. Listwise: Optimize the entire ranked list directly against metrics like NDCG. Best alignment with evaluation metrics but hardest to implement.
When You Need LTR
Use LTR when: you show users ordered lists (search results, recommendations, feeds); position matters (top-3 items get 80% of clicks); you have graded relevance labels (not just binary relevant/irrelevant). Simple heuristic scoring suffices for: small catalogs (<1000 items), when users don"t expect personalization, or when ranking logic is easily expressible as rules.