What is Point in Time (PIT) Correctness in ML Systems?
Why It Matters
PIT correctness eliminates future leakage, where information from after the prediction timestamp contaminates your model. This is one of the most insidious bugs in ML systems because offline metrics look great while production performance mysteriously degrades.
Event Time vs Processing Time
The core requirement is strict separation of event time (when the fact actually happened) versus processing time (when your system saw it). A fraud detection feature computed at 3pm but reflecting transaction data from 2pm must be timestamped at 2pm, not 3pm. Without this distinction, late arriving data can leak future information into past training examples.
When PIT Becomes Critical
PIT correctness is essential when labels trail features (fraud confirmed days later, ad clicks happen hours after impression), when features are time window aggregates (user activity over past 7 days), or when data arrives out of order in streaming systems. Uber processes 100 million to 1 billion training examples using PIT joins to ensure temporal consistency.
Implementation Principle
The same principles as database point in time recovery apply: maintain immutable versioned histories with event timestamps, then reconstruct state at any moment using base snapshots plus an append only change log. This underpins reproducibility, letting you recreate the exact dataset that trained a deployed model despite ongoing pipeline evolution.