"There are only two problems in inference: how to assign probability distributions, and how to do integrals." - John Skilling
There are several things we can do when publishing inferences:
State clearly what priors we assumed, so that anyone reproducing our analysis can compare their results given their assumptions with ours.
Try to make it easy for others to carry out the same inference but with different assumptions. Making our posterior samples is a good example of this: those samples can often be re-weighted in an importance sampling analysis that involves different assumptions.
Carry out "sensitivity analyses" so that our readers don't have to do the above: if a conclusion is sensitive to the prior PDF we assumed, that means there was relatively little information in our data about that parameter.
At first sight, ${\rm Pr}(x) \propto 1/x$ seems like a bad idea, because it rises steeply with decreasing $x$, "biasing" the result. But suppose $x$ is galaxy mass, and you want to plead ignorance: you assign a a uniform prior between 0 and $10^{14} M_{\odot}$. With this assignment you are saying that a priori ${\rm Pr}(x > 10^{12}) = 0.99$ - a highly informative statement!
A computational problem with uninformative priors is that they can lead to parameter space volumes that are unmanageably large.
Consider a uniform joint prior PDF in N parameter dimensions. What fraction of the a priori allowed volume is in a hypercubic shell that has thickness f of the side length?
In [ ]:
f = 0.01
N = 2
# Compute difference between two hypervolumes:
dV = 'not yet coded'
print("Volume fraction for f =",f," is",dV)
This effect can cause real computational problems when attempting to characterize posterior PDFs - you've seen it already, just in our attempts with two-dimensional grids!
Examples of constraints, and the maximum entropy distributions that result from them, include:
No constraint, except that $\int p(x) dx = 1$, gives $p(x) \propto {\rm constant}$, the uniform distribution.
Known mean, $\int x p(x) dx = \mu$, gives $p(x) \propto \exp(-x/\mu)$, the exponential distribution.
Known mean $\mu$ and variance $\sigma^2$ gives $p(x) \propto \exp(-(x - \mu)^2/2\sigma^2)/\sigma$, the Gaussian distribution.
${\rm Pr}(x|d,H) \propto {\rm Pr}(d|x,H)\;{\rm Pr}(x|c,H)$
${\rm Pr}(x|d,H) \propto {\rm Pr}(d|x,H)\;{\rm Pr}(c|x,H)\;{\rm Pr}(x|H)$
More generally, a good prior PDF is one that accurately represents your beliefs.
In [ ]: