scholarly journals Joint modeling of choices and reaction times based on Bayesian contextual behavioral control

2021 ◽  
Author(s):  
Sarah Schw&oumlbel ◽  
Dimitrije Markovic ◽  
Michael N Smolka ◽  
Stefan Kiebel

In cognitive neuroscience and psychology, reaction times are an important behavioral measure. However, in instrumental learning and goal-directed decision making experiments, findings often rely only on choice probabilities from a value-based model, instead of reaction times. Recent advancements have shown that it is possible to connect value-based decision models with reaction time models, for example in a joint reinforcement learning and diffusion decision model. We propose a novel joint model of both choices and reaction times by combining a mechanistic account of Bayesian sequential decision making with a sampling procedure. Specifically, we use a recent context-specific Bayesian forward planning model which we extend by an MCMC sampler to obtain both choices and reaction times. We show that we can explain and reproduce well-known experimental findings in value based-decision making as well as classical inhibition and switching tasks. First, we use the proposed model to explain how instrumental learning and automatized behavior result in decreased reaction times and improved accuracy. Second, we reproduce classical results in the Eriksen flanker task. Third, we reproduce established findings in task switching. These findings show that the proposed joint behavioral model may describe common underlying processes in all these types of decision making paradigms.

2015 ◽  
Vol 112 (45) ◽  
pp. 13817-13822 ◽  
Author(s):  
Fiery Cushman ◽  
Adam Morris

Humans choose actions based on both habit and planning. Habitual control is computationally frugal but adapts slowly to novel circumstances, whereas planning is computationally expensive but can adapt swiftly. Current research emphasizes the competition between habits and plans for behavioral control, yet many complex tasks instead favor their integration. We consider a hierarchical architecture that exploits the computational efficiency of habitual control to select goals while preserving the flexibility of planning to achieve those goals. We formalize this mechanism in a reinforcement learning setting, illustrate its costs and benefits, and experimentally demonstrate its spontaneous application in a sequential decision-making task.


2011 ◽  
Vol 137 (5) ◽  
pp. 341-348 ◽  
Author(s):  
Samiul Hasan ◽  
Satish Ukkusuri ◽  
Hugh Gladwin ◽  
Pamela Murray-Tuite

Author(s):  
Ming-Sheng Ying ◽  
Yuan Feng ◽  
Sheng-Gang Ying

AbstractMarkov decision process (MDP) offers a general framework for modelling sequential decision making where outcomes are random. In particular, it serves as a mathematical framework for reinforcement learning. This paper introduces an extension of MDP, namely quantum MDP (qMDP), that can serve as a mathematical model of decision making about quantum systems. We develop dynamic programming algorithms for policy evaluation and finding optimal policies for qMDPs in the case of finite-horizon. The results obtained in this paper provide some useful mathematical tools for reinforcement learning techniques applied to the quantum world.


Sign in / Sign up

Export Citation Format

Share Document