What is Root Mean Squared Error (RMSE) in Time Series?
Definition: RMSE (Root Mean Squared Error) measures forecast accuracy in the same units as the target: RMSE = √[(1/n) × Σ(actual - forecast)²]. Squaring penalizes large errors more heavily than small errors. A forecast with one large miss is worse than a forecast with many small misses of equal total absolute error.
RMSE vs MAE
MAE (Mean Absolute Error) treats all errors linearly: MAE = (1/n) × Σ|actual - forecast|. RMSE squares before averaging, then takes square root. Result: RMSE ≥ MAE always, with equality only when all errors are identical. The ratio RMSE/MAE indicates error distribution: ratio near 1 means uniform errors, ratio near √2 indicates varied error magnitudes.
Why Penalize Large Errors
In many applications, large errors are disproportionately costly. Understocking by 1000 units is worse than understocking by 10 units 100 times—customers experience stockouts, not aggregate shortage. RMSE aligns optimization with this business reality. If all errors are equally bad, use MAE instead.
Mathematical Advantage: RMSE is differentiable everywhere and convex, making it ideal for gradient-based optimization. Most ML models minimize squared error internally for this reason, even when MAPE is the reported metric.
RMSE Limitations
RMSE is scale-dependent: RMSE of 100 for sales in units cannot compare to RMSE of 1000 for sales in dollars. Use RMSE within a single series over time, not across series with different scales. For cross-series comparison, use percentage metrics (MAPE) or scaled metrics (MASE).
Sensitivity to Outliers
Squaring amplifies outlier impact. A single extreme error can dominate RMSE. If outliers represent data quality issues rather than true forecast failures, consider robust alternatives (median absolute error) or windsorize extreme values before computing RMSE.