When should the inverse-variance method be applied to distributions?

By Vasco Grilo @ 2022-06-14T14:33 (+5)

Given 2 distributions  and  which are independent estimates of the distribution , this be estimated with the inverse-variance method from:

Under which conditions is this a good aproach? For example, for which types of distributions? These questions might be relevant for determining:

Some notes:


karthik-t @ 2022-06-14T15:33 (+2)

If you assume both X1 and X2 are normal then the only difference between them comes from their moments, so you can use the inverse variance formula. But that leans directly on the formula for the product of normal distributions. The formula for a general convolution of two distributions does not have such a clean form. So while I don't have a rigorous argument for this, I would be shocked if you could do the same for any two PDFs X1 and X2 with no change to the formula.

I do not know if this is really necessary for the uses you name, though. Bayes Rule determines the posterior distribution regardless of whether it follows an inverse variance formula or not.

Vasco Grilo @ 2022-06-15T09:23 (+1)

Thanks for the reply!

I also think the above formula does not formally apply to non-normal distributions, but I was wondering whether it was a good enough approximation.

Is there a simple way of applying the Bayes Rule to two arrays  and  of Monte Carlo samples? I believe this is analagous to considering that all elements of  and  are equiprobable.

karthik-t @ 2022-06-15T18:54 (+3)

I don't think I follow. Monte Carlo sampling is done from a distribution, which I assume you want to use as the basis of your likelihood function? In this case, you can just calculate the likelihood function from this distribution, and combine it with your prior to get a posterior distribution.

Vasco Grilo @ 2022-06-16T10:57 (+1)

I was thinking about cases in which X1 and X2 are non-linear functions of arrays of Monte Carlo samples generated from distributions of different types (e.g. loguniform and lognormal). To calculate E(X1), I can simply compute the mean of the elements of X1. I was looking for a similar simple formula to combine X1 and X2, without having to work with the original distributions used to compute X1 and X2.

A concrete simple example would be combining the following:

  • According to estimate 1, X is as likely to be 1, 3, 4, 6 or 8: X1 = [1, 2, 3, 4, 5].
  • According to estimate 2, X is as likely to be 2, 4, 6, 8 or 10: X2 = [2, 4, 6, 8, 10].
  • The generation mechanisms of estimates 1 and 2 are not known.
karthik-t @ 2022-06-16T15:27 (+2)

How are both X1 and X2 estimates of X when they are different distributions? At this point I am out of my depth so I do not have an informative answer for you.

Vasco Grilo @ 2022-06-16T18:19 (+1)

I will try to illustrate what I mean with an example:

Meanwhile, I have realised the inverse-variance method minimises the variance of a weighted mean of  and  (and have updated the question above to reflect this).