Sampling distribution? #67
hyunjimoon
started this conversation in
people; relating
Replies: 1 comment 2 replies
-
This reminds me of a problem we ran into long ago in consumer packaged goods modeling. Basically the problem is |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
What should be the distribution for measurement noise (or sampling distribution)? I've tried log normal which was too fat tailed as its variance is superlinear to mean. So negative binomial might be a good option.
Perhaps using normal-normal (prior-likelihood) might not be a bad choice, as it would create normal mixture model, which @tomfid and I both experienced would perform well (i.e. better out of sample prediction accuracy, as I think the comment of "fitting data well is not always good" from BDA below is warning against overfitting) when we have large amount of data. When we have many missing (tooth) data, poisson or negative binomial are better for two reasons: first for zero-inflated concept, and second as count's discreteness is emphasized near zero-value.
BDA ch.6 states
These two points are not criticisms of the lognormal distribution or power transformations. Rather, they provide warnings when using a model that has not been subjected to posterior predictive checks (for test variables relevant to the estimands of interest) and reality checks. In this context, the naive statement, ‘better fits to data mean better models which in turn mean better real-world answers,’ is not necessarily true. Statistical answers rely on prior assumptions as well as data, and better real-world answers generally require models that incorporate more realistic prior assumptions (such as bounds on municipality sizes) as well as provide better fits to data. This comment naturally leads to a general message encompassing the first two points.
This explains power lognormal function.
Beta Was this translation helpful? Give feedback.
All reactions