Engine | Coefficient | Std Dev | z-score |
---|---|---|---|
MCMC | +0.705 | 0.070 | 10.08 |
MH | +0.352 | 0.111 | 3.16 |
Conjugate | +0.911 | 0.034 | 26.69 |
Engine | Coefficient | Std Dev | Role |
---|---|---|---|
MCMC | +0.246 | 0.015 | Moderate |
MH | +0.702 | 0.095 | Dominant |
Conjugate | +0.029 | 0.008 | Minimal |
Samples from conditional posteriors sequentially. Each parameter updated given current values of others. Results in tight posterior intervals due to efficient exploration. Low noise variance (0.085) means it attributes less to randomness, more to signal.
Proposes jumps, accepts/rejects based on probability ratio. 1.6% acceptance suggests narrow proposal distribution or conservative tuning. High noise variance (0.190) compensates—attributes more to stochastic effects. Lagged sales dominates (0.702) to stabilize predictions.
No sampling. Posterior derived analytically from likelihood × prior. Extremely tight intervals (σ² = 0.020). Heavily weights transactions (0.911) with minimal lag dependence. Deterministic, no randomness in inference—only in data.
Despite 10x difference in noise variance and 30x difference in transaction coefficients, 48-step-ahead forecasts align within 2% median error.
Coefficient | MCMC Fisher | MH Fisher | Conjugate Fisher | Interpretation |
---|---|---|---|---|
total_sales_lag | 1,023,816 | 456,943 | 4,390,515 | Extremely high precision—all agree on importance |
Intercept | 10,798 | 4,819 | 46,307 | Conjugate most confident in baseline |
log1p_tx | 10,798 | 4,819 | 46,307 | Uniform across features—standardized design |
Key Insight: Fisher information diagonal values are nearly identical across engines except for the lagged term. This suggests the X'X design matrix is well-conditioned (cond ≈ 46,007 for all). The massive difference in lagged-term Fisher info (100x–400x larger) reveals where models store "memory"—but they compensate by adjusting other coefficients to maintain forecast stability.
MCMC | total_sales: +0.70 |
MH | total_sales: +0.35 |
Conjugate | total_sales: +0.91 |
Conjugate believes transactions are primary driver. MH hedges with lag dependence. MCMC balances both.
MCMC | total_sales: -0.08 |
MH | total_sales: -0.35 |
Conjugate | total_sales: +0.03 |
Major disagreement. MCMC/MH see negative impact; Conjugate sees slight positive. Yet forecasts still converge because this feature has low overall influence.
Use Conjugate for fast inference with tight bounds. Use MCMC for interpretable coefficients. Use MH when you need conservative uncertainty estimates.
Since forecasts converge despite parameter differences, an equally-weighted ensemble provides robustness without over-complicating. The median of three engines is more stable than any single model.
When coefficients differ but forecasts align, it suggests model misspecification doesn't matter for prediction—only for causal interpretation. If forecasts diverge, your data needs more structure or longer history.