Process validation and it's role in uncertainty decomposition #158
divine7022
started this conversation in
Ideas
Replies: 1 comment 1 reply
-
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
understanding of how process validation fits into our uncertainty workflow:
conceptually, i am thinking of total forecast uncertainty as:
Var(Y) = Vθ + VX + VIC + Vint + Vprocess + Vobs,where we have already implemented parameter, driver, IC, and interaction terms via OAT/Sobol and ensemble variance.
What is process error?
process error represents the model's structural inability to perfectly represent reality --fundamentally different from parameter uncertainty (where the equation is correct but values are uncertain).
for agricultural systems specifically, process error dominates over stochasticity because:
• no wildfire disturbance
• controlled planting (no dispersal stochasticity)
• low background mortality
• managed pest control
This means
residual error ~= structural/process errorfor croplands.earlier discussion with @mdietze and @dlebauer , my understanding is that the gold-standard way to estimate process error is dynamic estimation via SDA (iteratively learns process error through Bayesian joint estimation of process error, state, and observation error through time), requires operational SDA pipeline; but that this is out of scope for the current project. Post-hoc validation residuals (RMSE, bias, etc.) can’t be treated as process error directly because they conflate process, observation, and accumulated errors (what we can do)
Given that, my proposed integration strategy is:
Phase 1 (current scope): Parallel reporting
• Uncertainty report: variance partitioning over parameters, drivers, ICs, and interactions
• Validation report: RMSE / bias / R2 by variable, site, or pft
without folding validation residuals into the variance budget.
Phase 2 (optional extension, only if useful): Residual variance attribution
As a conservative extension, we could compute an “unexplained variance” term
(
V_unexplained = V_total,observed − (Vθ + VX + VIC))and report bounds on V_process using observation-error estimates, rather than a point estimate, with explicit caveats.
My assumption is that Phase 1 is the expected deliverable, and Phase 2 would only be exploratory if there's interest and sufficient validation data.
Questions for discussion
(a) Separate report section (minimal integration), or
(b) Explicit "unexplained variance" term in decomposition (deeper integration)?
(a) Point estimate with caveats, or
(b) Confidence interval acknowledging we can't fully separate sources?
Beta Was this translation helpful? Give feedback.
All reactions