Where does value come from? In our new preprint (https://psyarxiv.com/rxf7e/ ) we explore some of the open challenges of Reinforcement Learning and reward-based decision-making more broadly. THREAD 1/13
-
-
Importantly, this circumvents the need to define arbitrary reward functions and simply assumes that the agent keeps track of physiological and cognitive repositories of their assets (e.g. hydration, social status, capital, etc.) 9/13
Prikaži ovu nit -
Interestingly, simply assuming that more than one value dimension contributes to the agent’s distance-to-goal estimation (eq. 2) yields diminishing marginal utility and convex indifference curves over value dimensions. 10/13pic.twitter.com/pknxUvAHlO
Prikaži ovu nit -
Although our theoretical framework is preliminary, we think that goal equilibrium theory may scaffold research into complex choices involving multiple, competing goals and inform our understanding of the underlying neural mechanisms. 11/13
Prikaži ovu nit -
This will require tasks which make explicit reference to multiple assets as if the agent collected distinct items whose relationships to value fluctuate as the agent’s goals change. 12/13
Prikaži ovu nit -
We hope you enjoy reading the paper and would be happy to hear thoughts, comments, or suggestions. 13/13
Prikaži ovu nit -
Novi razgovor -
Čini se da učitavanje traje već neko vrijeme.
Twitter je možda preopterećen ili ima kratkotrajnih poteškoća u radu. Pokušajte ponovno ili potražite dodatne informacije u odjeljku Status Twittera.