...

/

Attacking the Final Problem

Attacking the Final Problem

In this lesson, we will solve the problem of finding a good helper distribution `q`.

In the previous lesson, we finally wrote a tiny handful of lines of code to implement importance sampling correctly; if we have a distribution p that we’re sampling from, and a function f that we’re running those samples through, we can compute the expected value of f even if there are “black swan” regions in p.


Helper Distribution q

All we need is a helper distribution q that has the same support as p, but no black swans.

Great. How are we going to find that?

A variation of this problem has also occurred before: what should the initial and proposal distributions be when using Metropolis? If we’re using Metropolis to compute a posterior from a prior, then we can use the prior as the initial distribution. But it’s not at all clear in general how to choose a high-quality proposal distribution; there’s ...