Post Processing Threshold Optimization for Fairness
How It Works
Train your model normally, outputting probability scores. For each group, find the threshold achieving your fairness constraint. For demographic parity: thresholds such that positive rate is equal. For equalized odds: thresholds such that TPR and FPR are equal. This is an optimization problem: search threshold pairs to minimize fairness violation while maximizing accuracy. Grid search works for two groups; constrained optimization for more.
When to Use Post-Processing
Cannot retrain: Model training is expensive or you lack pipeline access. Regulatory compliance: Need quick fairness demonstration without long retraining. Interpretability: Threshold adjustments are easy to explain. However, post-processing is a patch. The underlying model still learned biased patterns. If features change, thresholds need recalibration.
The Accuracy Cost
Adjusting thresholds trades accuracy for fairness. If Group A threshold drops from 0.5 to 0.4, more lower-score members get approved, some as false positives. Typical accuracy loss: 2-5% for demographic parity, 3-8% for equalized odds. Cost depends on base rate differences: similar rates need minimal adjustment, large differences (60% vs 30%) cause larger hits.