Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
klderivation [2024/12/24 12:55] – [Cost of deliberation] pedroortega | klderivation [2024/12/24 12:58] (current) – [Connecting to the free energy objective] pedroortega | ||
---|---|---|---|
Line 82: | Line 82: | ||
We've obtained two expectation terms. The second is proportional to the Kullback-Leibler divergence between of the posterior to the prior choice probabilities. What is the first expectation? | We've obtained two expectation terms. The second is proportional to the Kullback-Leibler divergence between of the posterior to the prior choice probabilities. What is the first expectation? | ||
- | The first expectation represents the expected cost of each individual choice if it were to occur deterministically. This is because each term $C(x \cap d|x \cap c)$ measures the cost of transforming the relative probability of a specific choice. | + | The first expectation represents the expected cost of each individual choice |
===== Connecting to the free energy objective ===== | ===== Connecting to the free energy objective ===== | ||
Line 88: | Line 88: | ||
We can transform the above equality into a variational principle by replacing the individual choice costs $C(x \cap d|x \cap c)$ with arbitrary numbers. The resulting expression is convex in the posterior choice probabilities $P(x|d)$, so we get a nice and clean objective function with a unique minimum. | We can transform the above equality into a variational principle by replacing the individual choice costs $C(x \cap d|x \cap c)$ with arbitrary numbers. The resulting expression is convex in the posterior choice probabilities $P(x|d)$, so we get a nice and clean objective function with a unique minimum. | ||
- | We can even go a step further: | + | We can even go a step further: |
\[ | \[ | ||
\sum_x P(x|d) U(x) - \frac{1}{\beta} \sum_x P(x|d) \log \frac{ P(x|d) }{ P(x|c) }. | \sum_x P(x|d) U(x) - \frac{1}{\beta} \sum_x P(x|d) \log \frac{ P(x|d) }{ P(x|c) }. |