Differences
This shows you the differences between two versions of the page.
| Both sides previous revision Previous revision Next revision | Previous revision | ||
| klderivation [2024/12/24 12:53] – [Cost of deliberation] pedroortega | klderivation [2024/12/24 12:58] (current) – [Connecting to the free energy objective] pedroortega | ||
|---|---|---|---|
| Line 82: | Line 82: | ||
| We've obtained two expectation terms. The second is proportional to the Kullback-Leibler divergence between of the posterior to the prior choice probabilities. What is the first expectation? | We've obtained two expectation terms. The second is proportional to the Kullback-Leibler divergence between of the posterior to the prior choice probabilities. What is the first expectation? | ||
| - | The first expectation represents the expected cost of each individual choice. This is because each term $C(x \cap d|x \cap c)$ measures the cost of transforming the relative probability of a specific choice. | + | The first expectation represents the expected cost of each individual choice |
| ===== Connecting to the free energy objective ===== | ===== Connecting to the free energy objective ===== | ||
| Line 88: | Line 88: | ||
| We can transform the above equality into a variational principle by replacing the individual choice costs $C(x \cap d|x \cap c)$ with arbitrary numbers. The resulting expression is convex in the posterior choice probabilities $P(x|d)$, so we get a nice and clean objective function with a unique minimum. | We can transform the above equality into a variational principle by replacing the individual choice costs $C(x \cap d|x \cap c)$ with arbitrary numbers. The resulting expression is convex in the posterior choice probabilities $P(x|d)$, so we get a nice and clean objective function with a unique minimum. | ||
| - | We can even go a step further: | + | We can even go a step further: |
| \[ | \[ | ||
| \sum_x P(x|d) U(x) - \frac{1}{\beta} \sum_x P(x|d) \log \frac{ P(x|d) }{ P(x|c) }. | \sum_x P(x|d) U(x) - \frac{1}{\beta} \sum_x P(x|d) \log \frac{ P(x|d) }{ P(x|c) }. | ||