1. Typify model into PA or PAD based on the purpose and in-hand data #12
hyunjimoon
started this conversation in
perspectives; sense-making
Replies: 1 comment
-
Using |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Considering expanding model workflow would always bump into the boundary of deterministic parameter, classifying deterministic and stochastic model would have little value. On the other hand, empirical vs theoretical i.e. using observed data or not matters which can be uniformly expressed with PA and PAD modeling in this paper. [1] As ode integrator is embedded for ode modeling, I assumed PD which doesn't involve approximator cannot exist from the figure below.

Below are the list of each model's character from @tomfid's SIG talk "Data and Uncertainty in System Dynamics" and @hazhirr's class on Brining data into Dynamic model.
PA modeling
PAD modeling
Decision-based Bayesian Workflow
Building on the above, I drafted a three step model-building scenario.
Data
is arrays of vectors of observed,Draws
is arrays of vectors of sampled unobserved,Demand
is a functional form of policy. All of them can be dynamic i.e. function of time, but we restrictDemand
to time-heterogenous. This is based on Yaman's explanation, "A policy is expressed in general by a policy equation, not by a single parameter. Policy equations naturally have policy parameters. Can policy analysis be carried out by simply changing values of policy parameters? The answer is: not always, but to a great extent. In more advanced policy analysis, we need to change the forms of the policy equations, not just parameter values. But I now strongly suggest that you ignore this at this point and focus on parameters values of fixed policy equations.1. Verify Draws-Demand
We start from
PA
modeling. GivenP
(joint distribution of observed and unobserved) generated either explicitly (e.g. Bayesian learning) or implicitly (e.g. deep learning), we experiment within the following:2. Validate Data - [Draws - Demand] with testing-framed demand
Then we proceed to
PAD
modeling by addingD
i.e. observed, real-world data. GivenD
, we experiment within the following:We reverse engineer the first iteration by asking, "to make our demand noise-resilient, what form should our demand be, what data is needed, and what estimation method can do this most efficiently"? This order is common in Bayesian inference where we first build the generator then do the inference conditional on the data. Working in both directions is the benefit of generative model (e.g. Bayesian) which operates on joint space of observed and unobserved. Using this benefit, we argue the best practice of model building is to first make the rough mockup from data to demand then building up from the demand's end. For instance, the effect of new data on demand can be forwardly-simulated, then compared to inform us the best data. This spirit is similar to Bayesian optimization. However, unlike their description of "taking human out of the loop", we aim to design human-machine interface. In practice, human's form of demand, or activation function in Bayesian optimization, becomes clearer during the model building iteration.
Parsimoniously constructed graphs with appropriate level of parameter uncertainty (i.e. identifiable structure with tight prior distribution) will facilitate the inverse procedure by preventing degeneracy in posterior space.
On top of the above listed value of PAD modeling, Jeroen and I identified two additional benefits related to tipping points:
Jeroen's poster in ISDC 2022, Sustainable consumption transformations: Should we mobilize the young generation? below illustrates the two. Refer to Struben22_SustainTrans.pdf in
ref
folder (requested to Jeroen to share reference link (and with generosity his codes).First, the model qualification test outcome (pass/fail) depends on the value assumed parameter is conditioned upon. Like a water's boiling point (100 degree at 1 atm), tipping points boundary is unavoidable. The least we can do is to move our demand (mostly binary, at most finite) away from these sensitive borders. The easiest remedy is to adjust the fixed value of assumed parameter as this sometimes is modeler's rough assumption. However, changing estimated parameter's prior distribution or prior parameter can be another remedy, followed by changing model parameterization or components. Even collecting more data are possible and recommended if this can make our demand more resilient.
Jeroen's poster on aging model illustrates two attractors in the phase space; however, this analysis is based on the fixed value of assumed parameter (noted as green in his poster, e.g. active life duration). Fixing can be problematic in two ways: first in calibration then in analysis. Regarding calibration, the model that passed quality test when conditioned on at 20 can fail the test at = 30. In other words, the same model qualified in 1900s (or at the time when active life duration was 20 years) becomes outdated as society evolved to the era of longer active life duration. Considering both real data and parameter for calibration is well discussed in the upcoming paper on simulation-based calibration. Second, even if the model passed the quality test, under different conditioning, the analysis (two attractors in phase space) may completely change. There may be one or three attractors for instance, affecting the downstream policy proposal.
Assumed-value of nuisance parameter affecting the test result is problematic but unavoidable as we can't afford full-Bayesian model where uncertainty is imbued to every parameter. Like a Pirate Roulette game, we should pray each time that our slice of fixed parameter hyperplane does not cross the tipping points. Inference helps us in this respect as we can alternately focus on different parameters by switching back and forth between estimated parameter and assumed parameter. This zig-zag learning like that of Gibbs sampling is suboptimal but can help keep modelers away from borders by providing a more global view of the decision geometry.
Second, is to argue the state value of a particular attractor is not realistic. Globally it may be the bimodal but if the scale of state for the second mode is out of bounds we may as well consider it unimodal. This allows us to save considerable amount of further analysis (Analysis of tipping behavior from the poster) by precluding certain scenarios. However, both known and unknown bias should be acknowledged as data in hand may not be the just representation of population.
Alluded above in the second benefit of removing multi-modality, mapping boundary of tipping point in state space with that of multimodality in parameter space is my main research interest and I have listed here with four examples collected from @tomfid and @hazhirr's writing.
For "Testing-framed demand" refer to "Testing-framed decisions" in 2. Typify parameters in syntax-semantic with colors, units, bounds #7.
3. Prescribe policy parameter with [Data - [Draws - Demand]]
For the final iteration, we invite policy and behavior on top of PAD calibration flow we constructed in 1 and 2. Given function of
Demand
and observed dataD
, we experiment within the following (discussion needed):The last two is equivalent to above 2, the only difference being 2 is on physical parameter of our existing system (e.g. disease and community parameters) whereas 3 is what imaginary parameter (or at least implementation in process) for management (e.g. policy levers). Example of this parameter classification can is detailed in Colored diagrams section of 2. Typify parameters in syntax-semantic with colors, units, bounds #7.
@tomfid's suggested parallel workflow of calibration and decision in the diagram below, which I view as coflow of 2 and 3.

Both parameteric and nonparameteric approach exist to address these non-binary and sometimes continuous objects. We visit seminal works by John, Tom, Rogelio, Andrew, Yaman, Erling and seek ways to connect their softwares for the automated Bayesian workflow. For detail, refer to Supply of SilkRoad project.
Specific topics are to be decided but the following are relevant:
Beta Was this translation helpful? Give feedback.
All reactions