Houjun Liu

approximate inference

Direct Sampling

Direct Sampling is an approximate inference method where we pull samples from the given joint probability distribution.

Example

Suppose we are interested in:

where we dare \(P(B^{1}|D^{1},C^{1})\).

Step 1: sort

We obtain a topological sort of this network:

\begin{equation} B, S, E, D, C \end{equation}

Step 2: sample from \(B,S\)

  • We sample \(B\). We sampled that \(B=1\) today.
  • We sample \(S\). We sampled that \(S=0\) today.

Step 3: sample from \(E\)

  • We sample \(E\) GIVEN what we already sampled, that \(B=1, S=0\), we sampled that that \(E = 1\)

Step 4: sample from \(D, C\)

  • We sample \(D\) given that \(E=1\) as we sampled.
  • We sample \(C\) given that \(E=1\) as we sampled.

Repeat

Repeat steps 2-4

Step n: Analyze

BSEDC
10101
01100
11110
00110
10111

We desire to know \(P(b^{1}|d^{1}, c^{1})\). Looks like, given this table, it would be \(100\%\).

Likelihood Weighted Sampling

Likelihood Weighted Sampling is a sampling approach whereby you force values that you wont, and then weight the results by the chance of it happening.

This is super useful when our envidence is unlikely.

Example

Suppose again you are interested in \(P(b^{1}|d^{1}, c^{1})\). In this case, we only sample \(B,S,E\):

BSE
010
101

Now, for each of these results, we the compute the chance of our priors happening given the samples.

  • Row 1: \(p(d^{1}|e^{0})p(c^{1}|e^{0})\)
  • Row 2: \(p(d^{1}|e^{1})p(c^{1}|e^{1})\)

Let’s say:

  • Row 1: \(p(d^{1}|e^{0})p(c^{1}|e^{0})=0.3\)
  • Row 2: \(p(d^{1}|e^{1})p(c^{1}|e^{1})=0.9\)

Finally, to compute \(p(b^{1}|d^{1}c^{1})\):

\begin{equation} \frac{0.9}{0.9+0.3} \end{equation}

because only row \(2\) fit with our expectations.