Hybrid Retrieval: Combining Dense and Sparse Methods
Fusion Strategies
Score fusion: Normalize scores from each retriever, combine with weights: score = α × dense_score + (1-α) × sparse_score. Typical α: 0.5-0.7. Rank fusion: Merge by rank position (RRF - Reciprocal Rank Fusion). More robust to score scale differences. Cascade: Use sparse for initial recall, dense for re-ranking. Reduces dense inference cost.
Why Hybrid Outperforms
Dense models miss queries with rare terms (product IDs, technical jargon, proper nouns not in training data). Sparse models miss semantic matches ("inexpensive" for "cheap"). Hybrid catches both. Empirically: hybrid improves recall@100 by 5-15% over either method alone. The improvement is largest on mixed query types; homogeneous query sets may see smaller gains.
Implementation Considerations
Run both retrievers in parallel to minimize latency. Normalize scores before fusion (dense and sparse scores are on different scales). Tune fusion weights on held-out data; optimal α varies by domain. For cascade, dense re-ranks sparse top-100 or top-200; larger candidate sets improve recall but increase cost.