Differences
This shows you the differences between two versions of the page.
| Both sides previous revision Previous revision | |||
| sba [2024/10/27 18:11] – [Memory-Constrained Agents] pedroortega | sba [2024/10/27 18:11] (current) – [Adding the context] pedroortega | ||
|---|---|---|---|
| Line 56: | Line 56: | ||
| The resulting distribution $P_t(\tau)$ is our bounded-rational policy. You will have to experiment with the choices of $\alpha$ (which controls the step size) and $N$ (which controls the representation quality of the target distribution) to obtain a satisfactory training time. | The resulting distribution $P_t(\tau)$ is our bounded-rational policy. You will have to experiment with the choices of $\alpha$ (which controls the step size) and $N$ (which controls the representation quality of the target distribution) to obtain a satisfactory training time. | ||
| - | ===== Adding | + | ===== Adding |
| The above algorithm generates a new prior $P(\tau)$ which places more weights on desirable strings. However, often we want policies to respond to a user-provided context string $c \in X^\ast$, i.e. we want to sample strings from $P(\tau|c)$, | The above algorithm generates a new prior $P(\tau)$ which places more weights on desirable strings. However, often we want policies to respond to a user-provided context string $c \in X^\ast$, i.e. we want to sample strings from $P(\tau|c)$, | ||