Preprint
Article

This version is not peer-reviewed.

An Information-Theoretic Framework for Understanding Learning and Choice Under Uncertainty

A peer-reviewed article of this preprint also exists.

Submitted:

29 August 2025

Posted:

02 September 2025

You are already at the latest version

Abstract
Although information theory is widely used in neuroscience, its application has primarily been limited to the analysis of neural activity, with much less emphasis on behavioral data. This is despite the fact that the discrete nature of behavioral variables in many experimental settings—such as choice and reward outcomes—makes them particularly well-suited to information-theoretic analysis. In this study, we provide a framework for how behavioral metrics based on conditional entropy and mutual information can be used to infer an agent’s choice and learning mechanisms under uncertainty. Through simulations of various reinforcement learning models as ground truth, we illustrate how information-theoretic metrics can be applied to uncover the underlying choice and learning mechanisms. Specifically, we show that these metrics can reveal: (1) a positivity bias, reflected in higher learning rates for rewarded compared to unrewarded outcomes; (2) gradual, history-dependent changes in the learning rates indicative of metaplasticity; (3) adjustments in choice strategies driven by global reward rate; and (4) presence of alternative learning strategies. Overall, our study highlights how information theory can leverage the discrete, trial-by-trial structure of many cognitive tasks, offering a flexible framework for investigating neural mechanisms of learning and choice under uncertainty—with potential for further extension.
Keywords: 
;  ;  ;  ;  

1. Introduction

Information theory has been used widely across different domains of cognitive and systems neuroscience, ranging from analyzing spike trains to estimating uncertainty in sensory environments. For example, Shannon entropy and related metrics have been used to study information flow [1,2], functional and effective connectivity [3,4,5,6,7], and variability in neural response [8,9,10,11]. These quantitative approaches have provided significant insight into brain functions at the cellular and system levels.
Despite this widespread use of information-theoretic metrics in analyzing neural data, their application in investigating behavior has been surprisingly limited. Although several studies have employed entropy-based metrics to quantify uncertainty in stimuli [12,13,14,15] and outcomes [16,17,18], there are only a few studies that have utilized information theory to directly examine the underlying decision-making and learning mechanisms [19,20,21]. This is despite the fact that many behavioral measures—such as binary choices and reward feedback—are ideally structured for analysis using information-theoretic tools.
A common tool for studying learning and choice behavior is reinforcement learning (RL) models [22,23,24,25,26,27], due to their simplicity and interpretability. In this approach, various RL models are constructed and fit to the choice data. The best-fitting model is then identified through model selection, and its components form the basis for inferences about underlying cognitive and/or neural mechanisms. However, these models often need to be augmented with additional components to capture empirical data in specific tasks (e.g., [28,29,30,31]), while it remains unclear how many components are necessary—or sufficient—to define the “best” model. Such extensions include separate learning rates for positive versus negative prediction errors (differential learning rates), time-varying (dynamic) learning rates, modulations of learning by reward harvest rate, and arbitration between alternative models of the environment, among others [32,33,34,35,36,37,38,39,40]. Nevertheless, currently there is no systematic method to identify the critical components required in RL models. This lack of methodology can lead to important mechanisms being overlooked, even in models that provide the best fit among the tested models.
Interestingly, choice behavior under uncertainty is inherently stochastic, making it well-suited for analysis using information theory. Recently, a set of information-theoretic metrics was proposed as a model-agnostic approach to uncover the hidden learning mechanisms underlying behavior. For example, Trepka et al. [29] have suggested that behavioral metrics based on conditional entropy can quantify the consistency in local choice strategies and predict undermatching. Moreover, Woo et al. [41] have shown that mutual information, alongside other measures, can capture the influence of high-order reward statistics. These results highlight the potential of information-theoretic metrics to probe learning and decision-making processes beyond what traditional model-fitting techniques can reveal.
Here, we extended this approach by applying information-theoretic metrics to simulated choice data from different RL models—serving as ground truth— across four learning tasks. We show that this method captures keys aspects of learning and decision making without relying on model fitting. To that end, we used existing metrics and developed new ones to detect a range of learning and decision-making mechanisms and their dynamic adjustments. We begin by identifying higher learning rates following positive prediction errors (rewarded outcomes) than following negative prediction errors (unrewarded outcomes)—a phenomenon referred to as positivity bias [42,43,44,45]. We also examine changes in learning rates over time as a result of metaplasticity [46,47]. Next, we investigate the influence of reward harvest rate, which has been shown to modulate learning and decision making [34,35,48,49]. Finally, in naturalistic reward environments, choice options often possess multiple attributes, each potentially predictive of reward outcomes. Previous studies have demonstrated that humans and other animals tackle this challenge by simultaneously learning about alternative reward contingencies (i.e., models of the environment), arbitrating between these models, and deploying selective attention to guide differential learning and decision making [38,50,51,52,53,54,55,56,57,58]. Therefore, we test how information-theoretic metrics can be used to detect the presence of distinct learning strategies in complex reward environments. Overall, our results demonstrate that the patterns of information-theoretic metrics provide useful summary statistics of the behavioral signatures generated by different learning and choice mechanisms, thereby offering a complementary approach to model fitting (both discovery and recovery).

2. Materials and Methods

2.1. General Experimental Paradigm: Probabilistic Reversal Learning

We simulated choice data using different reinforcement learning (RL) agents performing two variants of the probabilistic reversal learning task (PRL), a widely used paradigm to assess cognitive flexibility across various species [59]. In a typical probabilistic reversal learning paradigm, subjects choose between two options based on reward feedback they receive on each trial, with selection of each option associated with a different probability of reward. Choice options can take various forms, such as distinct visual stimuli on a computer screen, identical stimuli presented at different spatial locations, or physical levers. One option yields reward with a higher probability (the better option) than the other (the worse option), but these reward contingencies switch at fixed or random times within an experimental session, creating ’reversals’. Crucially, reversals are not signaled and thus, subjects have to adjust their choice solely based on reward feedback to maximize their chance of winning a reward. Commonly used rewards include drops of juice or water (e.g., [60,61]), sucrose pellets (e.g., [62,63]), and monetary incentives for humans (e.g., [56,64]), which are delivered probabilistically at the end of each trial. For most simulations, we set the reward probabilities at 80 and 20 (corresponding to 80/20% reward schedule) with each block consisting of 80 trials, unless stated otherwise. Each block contains a reversal in the middle of the block, where reward probabilities on the two options switch. In the final set of simulations (Section 3.4), we also considered a generalized PRL task in which reward probabilities depended on one feature of each choice option (e.g., its color or its shape), while the other feature carried no information about reward.

2.2. Information-Theoretic Metrics

In this study, we utilized and extended the information-theoretic metrics introduced previously [29,41]. These measures—–grounded in conditional entropy, mutual information, and outcome-specific decompositions—–quantify how past rewards and choices shape the uncertainty surrounding an agent’s decision to stay with or switch from its previous choice (Figure 1b).
In general, for discrete random variables X and Y, the conditional entropy H ( Y | X ) represents the remaining uncertainty in Y given the information about the variable X. Formally, it is defined as:
H ( Y | X ) = x X P ( x ) · H ( Y | X = x ) = x X y Y P ( x , y ) log 2 P ( x , y ) P ( x )
Lower values of conditional entropy indicate that knowing the values of X reduces the uncertainty in Y, suggesting a strong dependence between the two variables.
Similarly, mutual information, denoted as I ( X ; Y ) , quantifies the information shared between discrete variables X and Y. Higher values of I ( X ; Y ) indicate a greater dependency between variables, such that knowledge of Y would make X more predictable (less uncertain). This relationship is expressed using the difference between Shannon entropy for Y, H ( Y ) , and the conditional entropy, H ( Y | X ) :
I ( X ; Y ) = H ( Y ) H ( Y | X ) = x X y Y P ( x , y ) log 2 P ( x , y ) P ( x ) P ( y )
Building on these general formulations, we define behavioral metrics to quantify uncertainty in the agent’s choice strategy, specifically in terms of whether the agent "stays" with or "switches" from the previous choice option, given certain outcomes (Figure 1a). Specifically, we aim to quantify how uncertainty in choice strategy is reduced by certain task-related information: the previous reward outcome (R), and the previously chosen option ( O p t ). The resulting metrics include the conditional entropy of reward-dependent strategy (ERDS), the conditional entropy of option-dependent strategy (EODS), and the conditional entropy of reward and option-dependent strategy (ERODS). These are paired with mutual information metrics including mutual information between reward outcome and choice strategy (MIRS), mutual information between previous choice and choice strategy (MIOS), and mutual information between reward outcome, choice options, and choice strategy (MIROS).
More specifically, ERDS measures the influence of previous reward outcomes—reward vs. no reward (referred to as win vs. loss for simplicity)—on the uncertainty of the subsequent choice strategy (stay vs. switch), as follows:
ERDS = H ( S t r t R t 1 ) = H ( S t r ) I ( R ; S t r ) = R { win , loss } S t r { stay , switch } P ( R , S t r ) log 2 P ( R , S t r ) P ( R ) .
where S t r t or S t r denotes the choice strategy between two subsequent trials (stay = 1, switch = 0) , R t 1 is the reward outcome on the previous trial (reward or win = 1 and no reward or loss = 0 ), P ( R t 1 , S t r t ) is the joint probability of choice strategy given a reward outcome on the previous trial, and P ( R ) is the probability of reward. In the equation above, H ( S t r ) is the Shannon entropy of strategy, measuring the randomness in choice strategy in terms of stay or switch:
H ( S t r ) = S t r { stay , switch } P ( S t r ) log 2 P ( S t r ) = P ( stay ) · log 2 P ( stay ) + P ( switch ) · log 2 P ( switch ) ,
where P ( S t r ) is the probability of stay or switch ( P ( s t a y ) = 1 P ( s w i t c h ) ). I ( R t 1 ; S t r t ) , which we refer to as MIRS, is the mutual information between reward outcome and strategy, equal to:
MIRS = I ( R t 1 ; S t r t ) = R { win , loss } S t r { stay , switch } P ( R , S t r ) log 2 P ( R , S t r ) P ( R ) P ( S t r ) .
Analogously, the conditional entropy of option-dependent strategy (EODS) represents the remaining uncertainty in the agent’s strategy after accounting for the choice made on the previous trial. It is defined as the difference between H ( S t r ) and the mutual information between the previous choice and strategy (MIOS), as follows:
EODS = H ( S t r t O p t t 1 ) = H ( S t r ) I ( O p t ; S t r ) = O p t { better , worse } S t r { stay , switch } P ( O p t , S t r ) log 2 P ( O p t , S t r ) P ( O p t ) ,
where O p t t 1 indicates the option chosen on the previous trial, with 1 indicating the better option and 0 indicating the worse option, as defined by the assigned reward probabilities, and P(Opt) is the probability of choosing the better option. I ( O p t t 1 ; S t r t ) denotes the mutual information between the chosen option and the agent’s strategy, referred to as MIOS, and is calculated as follows:
MIOS = I ( O p t t 1 ; S t r t ) = O p t { better , worse } S t r { stay , switch } P ( O p t , S t r ) log 2 P ( O p t , S t r ) P ( O p t ) P ( S t r ) .
To consider the combined effect of reward outcome and chosen option (i.e., winning or losing after choosing the better or worse option), we also considered a generalized metric that quantifies the combined effects of the two variables. Specifically, we define the conditional entropy of reward- and option-dependent strategy (ERODS), as follows:
ERODS = H ( S t r t R t 1 , O p t t 1 ) = H ( S t r ) I ( R , O p t ; S t r ) = R { win , loss } O p t { better , worse } S t r { stay , switch } P ( R , O p t , S t r ) log 2 P ( R , O p t , S t r ) P ( R , O p t ) .
Here, I ( R , O p t ; S t r ) denotes the mutual information between the combination of the previous reward and choice outcomes ( R , O p t ) and subsequent choice strategy (MIROS):
MIROS = I ( R t 1 , O p t t 1 ; S t r t ) = R { win , loss } O p t { better , worse } S t r { stay , switch } P ( R , O p t , S t r ) log 2 P ( O p t , S t r ) P ( R , O p t ) P ( S t r ) .

2.2.1. Decomposition of Information-Theoretic Metrics

For the metrics described above, we also analyzed their decompositions into component values associated with each specific outcome of the conditioning variable. In the case of conditional entropy measures, this corresponds to the conditional entropy of Y given a specific value of X = x , weighted by the probability P ( x ) . For instance, ERDS can be decomposed into two components based on previous reward outcomes: E R D S + and E R D S , which reflect the uncertainty reduction following rewarded (win) and unrewarded (loss) trials, respectively [29]:
ERDS + = H ( S t r t R t 1 = win ) · P ( win ) = P ( stay , win ) · log 2 P ( stay , win ) P ( win ) + P ( switch , win ) · log 2 P ( switch , win ) P ( win ) ,
ERDS = H ( S t r t R t 1 = loss ) · P ( loss ) = P ( stay , loss ) · log 2 P ( stay , loss ) P ( loss ) + P ( switch , loss ) · log 2 P ( switch , loss ) P ( loss ) .
This decomposition guarantees that ERDS + + ERDS = ERDS (Equation (3)).
Similarly, EODS can be decomposed into E O D S B and E O D S W based on the choice of the better or worse option on the previous trial:
EODS B = H ( S t r t O p t t 1 = better ) · P ( better ) = P ( stay , better ) · log 2 P ( stay , better ) P ( better ) + P ( switch , better ) · log 2 P ( switch , better ) P ( better ) ,
EODS W = H ( S t r t O p t t 1 = worse ) · P ( worse ) = P ( stay , worse ) · log 2 P ( stay , worse ) P ( worse ) + P ( switch , worse ) · log 2 P ( switch , worse ) P ( worse ) .
Finally, ERODS is decomposed into four components based on the combination of the previous reward and the choice option:
ERODS B + = H ( S t r t R t 1 = win , O p t t 1 = b e t t e r ) · P ( win , better ) = P ( stay , win , better ) · log 2 P ( stay , win , better ) P ( win , better ) + P ( switch , win , better ) · log 2 P ( switch , win , better ) P ( win , better ) ,
ERODS B = H ( S t r t R t 1 = loss , O p t t 1 = b e t t e r ) · P ( loss , better ) = P ( stay , loss , better ) · log 2 P ( stay , loss , better ) P ( loss , better ) + P ( switch , loss , better ) · log 2 P ( switch , loss , better ) P ( loss , better ) ,
ERODS W + = H ( S t r t R t 1 = win , O p t t 1 = w o r s e ) · P ( win , worse ) = P ( stay , win , worse ) · log 2 P ( stay , win , worse ) P ( win , worse ) + P ( switch , win , worse ) · log 2 P ( switch , win , worse ) P ( win , worse ) ,
ERODS W = H ( S t r t R t 1 = loss , O p t t 1 = worse ) · P ( loss , worse ) = P ( stay , loss , worse ) · log 2 P ( stay , loss , worse ) P ( loss , worse ) + P ( switch , loss , worse ) · log 2 P ( switch , loss , worse ) P ( loss , worse ) ,
where the subscripts refer to winning after choosing the better option ( B + ), losing after the better option ( B ), winning after the worse option ( W + ), and losing after the worse option ( W ). Note that ERODS can be alternatively decomposed based on either reward outcome or choice option alone, by adding the relevant components. For example, this metric can be decomposed based on whether the previous trial resulted in reward or no reward, as follows:
ERODS + = ERODS B + + ERODS W + , ERODS = ERODS B + ERODS W .
This metric can also be decomposed based on whether the better or worse option was selected on the previous trial:
ERODS B = ERODS B + + ERODS B , ERODS W = ERODS W + + ERODS W .
For decomposing mutual information metrics, we utilize the general formulation given by:
I ( X = x ; Y ) = H ( Y ) H ( Y | X = x ) = y Y P ( y ) · log 2 P ( y ) + y Y P ( y | x ) · log 2 P ( y | x ) ,
This quantity represents the information that a specific value X = x provides about Y, and is known as pointwise mutual information. The expected value of pointwise mutual information over all values of X is equal to the mutual information between X and Y (Equation (2)):
I ( X ; Y ) = H ( Y ) H ( Y | X ) = x X P ( x ) H ( Y ) x X P ( x ) H ( Y X = x ) = x X P ( x ) { H ( Y ) H ( Y X = x ) } = x X P ( x ) I ( X = x ; Y )
Unlike the mutual information I ( X ; Y ) , which is always nonnegative, the pointwise mutual information for a specific event x, I ( X = x ; Y ) , can take negative values when H ( Y ) < H ( Y | X = x ) [65]. Conceptually, negative values indicate that the outcome x is misleading—or misinformative—about Y [66].
Using this definition, we compute two components of MIRS, MIRS + and MIRS , corresponding to previously rewarded and unrewarded outcome, respectively:
MIRS + = I ( R t 1 = win ; S t r t ) · P ( win ) = { H ( S t r ) H ( S t r t R t 1 = win ) } · P ( win ) ,
MIRS = I ( R t 1 = loss ; S t r ) · P ( loss ) = { H ( S t r ) H ( S t r t R t 1 = loss ) } · P ( loss ) .
These include the decomposition terms for conditional entropy (Equations (10)-11), which satisfy the identity ( MIRS = MIRS + + MIRS ). Similarly, the decompositions for mutual information between the option chosen on the previous trial and strategy, MIOS, are defined as:
MIOS B = I ( O p t t 1 = better ; S t r t ) · P ( better ) = { H ( S t r ) H ( S t r t O p t t 1 = better ) } · P ( better ) ,
MIOS W = I ( O p t t 1 = worse ; S t r t ) · P ( worse ) = { H ( S t r ) H ( S t r t O p t t 1 = worse ) } · P ( worse ) .
These terms satisfy MIOS = MIOS B + MIOS W . Lastly, the decompositions for MIROS are given as:
MIROS B + = I ( R t 1 = win , O p t t 1 = better ; S t r t ) · P ( win , better ) = { H ( S t r ) H ( S t r t R t 1 = win , O p t t 1 = better ) } · P ( win , better ) ,
MIROS B = I ( R t 1 = loss , O p t t 1 = better ; S t r t ) · P ( loss , better ) = { H ( S t r ) H ( S t r t 1 R t 1 = loss , O p t t 1 = better ) } · P ( loss , better ) ,
MIROS W + = I ( R t 1 = win , O p t t 1 = worse ; S t r t ) · P ( win , worse ) = { H ( S t r ) H ( S t r t R t 1 = win , O p t t 1 = worse ) } · P ( win , worse ) ,
MIROS W = I ( R t 1 = loss , O p t t 1 = worse ; S t r t ) · P ( loss , worse ) = { H ( S t r ) H ( S t r t R t 1 = loss , O p t t 1 = worse ) } · P ( loss , worse ) .
These terms quantify the shared information between each instance of reward-option combination and the agent’s strategy. Alternative decompositions can be computed similarly to ERODS as in Equations (18) and (19). Notably, these decompositions reveal how specific combinations of trial outcomes influence subsequent strategy, providing additional information beyond the MIRS and MIOS metrics.

2.2.2. Normalization of Mutual Information Metrics

The mutual information metrics defined thus far quantify the total amount of shared information between the random variables X and Y. However, one might be interested in the fraction of uncertainty in Y that is reduced by the knowledge of X, rather than the absolute amount of shared information [67,68,69]. In our setting, if the agent’s strategy is largely deterministic (i.e., H ( S t r ) is low), then the mutual information will also be low due to small value of H ( S t r ) itself (e.g., see the area of gray circle specifying H ( S t r ) in Figure 1b). Therefore, to account for the baseline uncertainty in the strategy, we also considered normalized mutual information metrics, obtained by dividing the mutual information by H ( S t r ) .
More specifically, we define normalized mutual information metrics for MIRS, MIOS, MIROS as follows:
n-MIRS = I ( R t 1 ; S t r t ) / H ( S t r ) ,
n-MIOS = I ( O p t t 1 ; S t r t ) / H ( S t r ) ,
n-MIROS = I ( R t 1 , O p t t 1 ; S t r t ) / H ( S t r ) .
These metrics quantify the proportion of uncertainty in H ( S t r ) that is explained by each corresponding mutual information term. The normalized decompositions for each metric are similarly obtained by dividing each component by H ( S t r ) . For example, normalized mutual information between reward and strategy is computed as:
n-MIRS = n-MIRS + + n - MIRS , = P ( win ) · I ( R t 1 = win ; S t r t ) H ( S t r ) + P ( loss ) · I ( R t 1 = loss ; S t r t ) H ( S t r ) ,
and analogously, normalized mutual information between choice option and strategy is defined as:
n-MIOS = n-MIOS B + n-MIOS W , = P ( better ) · I ( O p t t 1 = better ; S t r t ) H ( S t r ) + P ( worse ) · I ( O p t t 1 = worse ; S t r t ) H ( S t r ) .
Finally, the normalized mutual information between reward-choice combination and strategy can be computed as follows (dropping subscripts for trials for simplicity):
n-MIROS = n-MIROS B + + n-MIROS B + n-MIOS W + + n-MIOS W + , = P ( better , win ) · I ( R = win , O p t = better ; S t r ) H ( S t r ) + P ( better , loss ) · I ( R = loss , O p t = better ; S t r ) H ( S t r ) + P ( worse , win ) · I ( R = win , O p t = worse ; S t r ) H ( S t r ) + P ( worse , loss ) · I ( R = loss , O p t = worse ; S t r ) H ( S t r ) .

2.3. Reinforcement Learning Models and Simulations

We used multiple reinforcement learning (RL) models to simulate two variants of the probabilistic reversal learning task in order to illustrate that information-theoretic metrics can detect: (1) the differential learning rates for rewarded versus unrewarded trials, (2) changes in learning rates due to metaplasticity, (3) the influence of reward harvest rate, and (4) the use of alternative learning strategies in multi-dimensional reward environments.
For each task, we generated synthetic choice data using a range of model parameters and computed a set of relevant behavioral metrics from the resulting data. Below, we outline the general architecture of the RL models used, which are based on the standard Q-learning algorithm [22,25,26,27]. In this algorithm, the reward values of the two choice options, Q A and Q B , are updated on a trial-by-trial basis using reward feedback, and the difference between these values is used to make a choice (Figure 2a). Specifically, the value of the chosen option on a given trial t, Q c , is updated according to the following rule:
Q c ( t + 1 ) = Q c ( t ) + α ( R ( t ) Q c ( t ) ) ,
where R ( t ) is the binary reward outcome (1 if rewarded, 0 if unrewarded), and α denotes the learning rate, which determines the magnitude of the value update by scaling the reward prediction error, R P E = R ( t ) Q c ( t ) . The value of the unchosen option remains unchanged. All value estimates are initialized at 0.5 for the first trial.
On each trial, the choice is determined stochastically based on the probability of selecting each option, P j , computed using the softmax function as follows:
P j ( t ) = e β Q j ( t ) e β Q j ( t ) + e β Q k ( t ) = 1 1 + e β ( Q j ( t ) Q k ( t ) ) ,
where the indices j and k correspond to the two choice alternatives, and β is the inverse temperature parameter that controls the sensitivity to value differences (i.e., slope of the softmax function). We selected the softmax function because it is widely adopted in empirical research and follows naturally from normative rationality principles [70,71]. In its simplest form, the RL model has only two free parameters: α and β . Below, we describe the variants of the RL models used in each set of simulations.

2.3.1. Simulations of Positivity Bias

In this set of simulations, we examined the positivity bias—also known as the optimism bias—which refers to the tendency of an agent to learn more from feedback with positive valence than from feedback with negative valence [42,43,44,45]. This bias is often quantified by estimating separate learning rates for the outcomes that either are “better” or “worse” than expected, or equivalently, for positive and negative prediction errors, in the context of error-driven learning. In the case of binary reward outcomes, these correspond to the learning rates following rewarded and unrewarded trials, respectively. Therefore, the learning rule in Equation (36) can be generalized as:
Q c ( t + 1 ) = Q c ( t ) + α + ( R ( t ) Q c ( t ) ) , if R ( t ) = 1 Q c ( t ) + α ( R ( t ) Q c ( t ) ) , if R ( t ) = 0 ,
where α + and α denote the learning rates used to update the value of the chosen option on rewarded and unrewarded trials, respectively. Positivity bias is thus formally characterized by the condition where α + > α .
We used the above RL model to simulate the choice behavior in a probabilistic reversal learning task. Each block consisted of 80 trials with an 80/20% reward schedule, and a reversal in reward contingencies occurred after trial 40. To examine how positivity bias can be detected using information-theoretic metrics, we simulated choice behavior using different combinations of ( α + , α ) values. We set the inverse temperature at β = 10 , consistent with values observed in previous experimental studies [42,72]. To explore a plausible range of learning rates while avoiding extreme values, we varied each learning rate α i [ 0.1 , 0.9 ] in increments of 0.05, resulting in a 17-by-17 grid of ( α + , α ) combinations. For each point in this parameter space, choice behavior was simulated over 10,000 blocks.
To test whether the information-theoretic measures contain sufficient information to determine the presence of positivity bias ( α + > α ), we utilized a linear decoder based on logistic regression, implemented using the fitclinear function in MATLAB). Decoders were trained on metrics computed from the choice behavior of training samples and tested on a separate held-out set, with the goal of predicting whether the underlying RL model used to generate the data had α + > α .
More specifically, in each decoding experiment, we randomly sampled RL agents, each assigned a true ( α + , α ) pair drawn independently from uniform distributions over α i [ 0.1 , 0.9 ] (Figure 3e). Twenty RL agents were assigned to each group, labeled as either "optimistic" ( α + > α ) or "pessimistic" ( α + < α ). For each agent, every simulated block (out of total of 100) used a distinct ( α + , α ) pair sampled from gaussian distributions ( σ = 0.03 ) centered on that agent’s true ( α + , α ) values. For each block, we obtained theoretical averages of the information-theoretic metrics by averaging across 10 repeated simulations. To evaluate decoding performance, we employed a leave-one-out procedure at the agent level: the decoder was trained on the information-theoretic metrics computed from the choice behavior of all agents except the one being tested. The training data were balanced to include an equal number of optimistic and pessimistic agents. Decoding accuracy was averaged across 100 independent decoding experiments, each using a unique set of RL agents.
Finally, to benchmark our decoder against behavioral features beyond the information-theoretic metrics, we trained a separate decoder on the coefficients of a logistic regression predicting choice from past choice and reward history. More specifically, we fit the following logistic regression model [30] to the same segment of choice data:
log ( P ( C ( t ) = j ) P ( C ( t ) = k ) ) = β 0 + i = 1 5 β i R ( R j ( t i ) R k ( t i ) ) + i = 1 5 β i U ( U j ( t i ) U k ( t j ) ) ,
where C ( t ) indicates the choice made on trial t, j and k index the two choice options, R i indicates a rewarded choice (1 if rewarded, 0 otherwise), and U i denotes an unrewarded choice (1 if unrewarded, 0 otherwise). We used the resulting ten regression coefficients ( β R and β U ) to decode whether a given agent is optimistic or pessimistic.

2.3.2. Simulations of Reward-Dependent Metaplasticity

In this set of simulations, we examined whether adjustments in the learning rates—predicted by reward-dependent metaplasticity [46,47]—can be detected using information-theoretic metrics. Importantly, reward-dependent metaplasticity enables dynamic adjustments of synaptic plasticity over time and can account for stimulus- or action-specific learning rates, including asymmetries such as positivity bias. Using the same probabilistic reversal learning task with an 80/20% reward schedule as in the positivity bias simulations, we compared the choice behavior of the metaplastic model to that of the ’plastic’ model, which is equivalent to the standard RL model described earlier. Below, we briefly describe the metaplastic model (see [46] for more details).
Importantly, the standard RL algorithm can be implemented through binary synapses that encode and update the value of each choice option through reward-dependent plasticity: transitioning from a "weak" to a "strong" state following reward (reward-dependent potentiation), and from a "strong" to a "weak" state following no reward (reward-dependent depression) [47,73,74] (Figure 4a). Importantly, the proportion of synapses in the "strong" state provides an estimate of the value of a given choice option ( Q j ), as this quantity increases and decreases following reward and no reward, respectively [73]. We assume that only the synapses associated with the chosen option undergo state transitions according to the reward outcome (reward-dependent plasticity), while those associated with the unchosen option remain unchanged—consistent with the assumption of standard Q-learning model.
The metaplastic model generalizes the above reward-dependent plasticity mechanism by introducing multiple meta-states for each level of synaptic efficacy, with synapses occupying deeper meta-states being more resistant—or stable—with respect to future changes (Figure 4b). More specifically, synapses in the metaplastic model undergo both plastic and metaplastic transitions. During plastic transitions, synaptic efficacy shifts between "weak" and "strong" states. During metaplastic transitions, synaptic efficacy remains unchanged, but the stability is modified as the synapse transitions to a deeper or shallower meta-state (Figure 4b).
Formally, higher stability of deeper meta-states is captured by decreasing transition probabilities q i for the i t h meta-state, governed by a power law as follows:
q i = q 1 i ( m 2 ) + 1 m 1 for 2 i m ,
where m is the number of meta-states, q 1 is the (baseline) transition probability between the most unstable weak and most unstable strong meta-states, and q i is the transition probability from weak (or strong) meta-state i + 1 to the most unstable strong (or weak) state in response to positive (or negative) reward feedback (reward vs. no reward), as indicated by the diagonal arrows in Figure 4b.
Additionally, metaplastic synapses can undergo transitions that do not alter their synaptic efficacy (vertical arrows in Figure 4b), with the probability of such transitions decreasing for deeper meta-states:
p i = p 1 i for 2 i m 1 ,
where p i indicate the transition probability between the i t h and the ( i + 1 ) t h meta-states. After positive reward feedback (potentiation events), "weak" synapses undergo transition toward less stable meta-states while "strong" synapses transition toward more stable, deeper meta-states. Conversely, after negative reward feedback (depression events) "weak" synapses undergo transition toward more stable, deeper meta-states while "strong" synapses transition toward less stable meta-states (vertical arrows in Figure 4b).
The value of each option in the metaplastic model is computed by summing over the fractions of synapses in the strong states:
Q j ( t ) = i = 1 m S j , i ( t ) ,
where j indexes the choice option, and S j , i ( t ) indicates the fraction of synapses in the i t h strong meta-state on a given trial.
Because the metaplastic model includes multiple meta-states with different transition rates, its update in response to reward feedback—and thus its rate of learning—changes over time. This dynamic can be quantified using two "effective" learning rates computed on each trial. The effective learning rate following reward ( α + e f f ) is defined as the fraction of synapses that transition from weak to strong states, while the effective learning rate following no reward ( α e f f ) reflects the fraction transitioning from strong to weak states, as follows:
α + e f f ( t ) = Q c ( t + 1 ) Q c ( t ) R ( t ) Q c ( t ) if R ( t ) = 1 , α e f f ( t ) = Q c ( t + 1 ) Q c ( t ) R ( t ) Q c ( t ) if R ( t ) = 0 ,
where the numerator indicates the change in the overall value of the chosen option, while the denominator corresponds to the prediction error.
To simulate choice behavior, we set α = 0.3 and β = 10 for the plastic model with a single learning rate. For the metaplastic model, we used the model with m = 4 meta-states and set p 1 = 0.4 for the baseline meta-transition probability. We set the transition probability q 1 = 0.516 , which yields an initial effective learning rate of 0.3 on the first trial of the block ( Figure 4c). We also tested the choice behavior of a plastic model with differential learning rates for rewarded and unrewarded outcomes ( α + α ). These learning rates were estimated by fitting the plastic model to the choice behavior of the metaplastic model on each block via maximum-likelihood estimation (optimized with fmincon function in MATLAB).
Similar to the positivity bias simulations, we used a linear decoder to test whether the emergence of positivity bias ( α + > α )—driven by metaplasticity over time—can be detected using the information-theoretic metrics. The training procedure involved randomly drawing a set of ( α + , α ) values from the grid of [0.2, 0.4] in a step size of 0.01, under the assumption that the approximate range of α value is known ( α = 0.3 in our case). Each parameter space was labeled as either "optimistic" ( α + > α ) or "pessimistic" ( α + < α ). The points with α + = α were dropped from the training set. We then computed the decoder’s posterior probability on held-out samples, which were generated from the choice behavior of the three models: (1) a plastic model with a single learning rate ( α + = α = α ), (2) the metaplastic model, and (3) a plastic model with separate learning rates ( α + α ) estimated from the choice behavior of the metaplastic model.

2.3.3. Simulations of Reward Harvest Rate Effects on Behavior

In this set of simulations, we examined whether the influence of the overall reward rate on a long time scale—referred to as reward harvest rate—can be detected using the information-theoretic metrics. We used the same probabilistic reversal learning task as in the positivity bias simulations and tested two RL models for generating choice behavior: (1) the standard RL model with a single learning rate, as described earlier; and (2) an augmented RL model that included an additional adjustment mechanism based on reward harvest rate (RL with RHR model; Figure 5a). The latter model served as the ground truth for a potential pathway through which reward harvest rate could influence choice strategy—specifically, by increasing the tendency to win-stay and lose-switch as the harvest reward rate increases. More specifically, the augmented model tracks a third variable H, representing the overall reward harvest rate independent of any specific choice option, and updates it as follows:
H ( t + 1 ) = H ( t ) + α H ( R ( t ) H ( t ) ) ,
where α H is the update rate for H. Therefore, H provides an exponentially weighted moving average of harvested rewards. For simplicity, we set α H = α = 0.3 in the simulations. In practice, although α H can be directly optimized to the data, one can equally justify choosing an arbitrary α H for behavioral analysis to avoid modeling assumptions and obtain model-independent measures. To incorporate a win-stay/lose-switch bias that is modulated by the reward harvest rate H, the choice probability in this model is computed as follows:
P j ( t ) = 1 1 + e β ( Q j ( t ) Q k ( t ) ) + β H B i a s j ( t ) ,
where β H is a free parameter controlling the overall influence of the win-stay/lose-switch bias term for option j, denoted as B i a s . This bias term is calculated as follows:
B i a s j ( t ) = C j ( t 1 ) · R ( t 1 ) H ( t 1 ) ,
where C j indicates whether option j was chosen in the previous trial (1 if chosen, -1 otherwise), and the term R H dynamically adjusts the agent’s tendency to repeat (after reward) or switch (after no reward) based on how much the recent reward outcome differs from reward harvest rate H. In our simulations, we set β = 10 , β H = 2 .
To quantify the effect of reward harvest rate on choice strategy, we introduced an additional mutual information metric, referred to as MIRHS, which captures the joint influence of previous reward outcome and the "reward harvest state" (as defined by reward rate) on the agent’s strategy, as follows:
MIRHS = I ( R t 1 , R H S t 1 ; S t r t ) = R { win , loss } R H S { high , low } S t r { stay , switch } P ( R , R H S , S t r ) log 2 P ( R H S , S t r ) P ( R , R H S ) P ( S t r ) ,
where RHS denotes the reward harvest state—a binary, discretized variable indicating whether the reward harvest rate H on a given (previous) trial was classified as "high" or "low", based on the median split within the current block (Figure 5b) [35]. We also used a normalized version of this metric, n-MIRHS = I ( R t 1 , R H S t 1 ; S t r t ) / H ( S t r t ) , which is decomposed as follows (dropping subscripts for trials for simplicity):
n-MIRHS = n-MIRHS h i g h + + n-MIRHS h i g h + n-MIRHS l o w + + n-MIRHS l o w + , = P ( high , win ) · I ( R = win , R H S = high ; S t r ) H ( S t r ) + P ( high , loss ) · I ( R = loss , R H S = high ; S t r ) H ( S t r ) + P ( low , win ) · I ( R = win , R H S = low ; S t r ) H ( S t r ) + P ( low , loss ) · I ( R = loss , R H S = low ; S t r ) H ( S t r ) ,
where the subscripts of the metric refer to winning during a high RHS state ( h i g h + ), losing during a high RHS state ( h i g h ), winning during a low RHS state ( l o w + ), and losing during a low RHS state ( l o w ). These metrics therefore separate the influence of immediate, reward outcome from that of the reward harvest state. To isolate the overall effect of R H S , we also computed an alternative decompositions as follows:
MIRHS h i g h = MIRHS h i g h + + MIRHS h i g h ,
MIRHS l o w = MIRHS l o w + + MIRHS l o w ,

2.3.4. Simulations of Learning in Multidimensional Reward Environments

To investigate the presence of alternative learning strategies in the agent’s behavior, we conducted a final set of simulations using a variant of the probabilistic reversal learning task in which different attributes of the choice options predicted reward outcomes at different time points.
In this task, at any given point within a block of trials, only one of the two choice attributes—the shape and the color of the stimuli—is predictive of reward probabilities (80% vs 20%). Initially, the reward schedule for a given block is assigned to the color of the stimuli regardless of their shape, with higher reward probability assigned to either green or orange objects. Therefore, unlike the reversal learning task used in the previous simulations, this task involves two types of reversals: (i) reversal in the reward probability, as previously considered, and (ii) reversal in the feature predictive of reward (Figure 6a). Therefore, in a Color-to-Color (or similarly Shape-to-Shape) block, the reward contingency was reversed between the better and worse values of the same attribute—for example, between two colors (Figure 6c). In contrast, in a Color-to-Shape block, the predictive attribute switched, and reward became associated with the shape of the object (e.g., triangle = 80% and square = 20%), regardless of its color (Figure 6d). The identity of the block type was unknown to the agent, and the two block types were randomly interleaved throughout the session. The Agent therefore had to adapt their learning and choice strategies solely based on the reward feedback they received.
To capture the strategies that more realistic agents might adopt, we used a generalized RL model that dynamically arbitrates between competing learning strategies. To that end, we used a variant of the model introduced by Woo et al. [38], which simultaneously tracks the values of multiple choice attributes (Figure 6b). In this model, the value estimates for each attribute are updated simultaneously as follows:
Q c , i ( t + 1 ) = Q c , i ( t ) + α ( R ( t ) Q c , i ( t ) ) ,
Q u , i ( t + 1 ) = ( 1 γ ) Q u , i ( t ) ,
where i indexes the choice attribute (Color or Shape), and c indexes the chosen option within each attribute (i.e., Color { g r e e n , o r a n g e } and Shape { t r i a n g l e , s q u a r e } ). The index u refers to the unchosen option within each attribute, and γ denotes the decay rate of value for unchosen options. Note that this decay mechanism is biologically motivated and included to account for the increased number of choice features in this model, reflecting the brain’s limited capacity to retain the memory of all value estimates. To generate a choice, the model first computes the overall value, V, of the two choice alternatives through a linear weighted combination of the attributes, as follows:
V j ( t ) = Q C o l o r , j × ω ( t ) + Q S h a p e , j × ( 1 ω ( t ) ) ,
where ω ( t ) is the arbitration weight on trial t specifying the relative contribution of color attribute to overall value, and the index j denotes the choice options to the left and right of the screen (for example, if the leftward option is a green triangle, V l e f t ( t ) = Q g r e e n ( t ) ω ( t ) + Q t r i a n g l e ( t ) ( 1 ω ( t ) ) ). The difference in overall values V is then used to compute choice probability, similar to Equation (37). It is important to note that linear combination of feature values to an overall value is used primarily for simplicity in describing the model and does not necessarily imply that such an integrated value is explicitly constructed. In practice, it is more plausible that the value of each attribute is compared directly—albeit with different weights—when making choices [75].
To dynamically control the arbitration weight between competing strategies, we considered the following update rule:
ω ( t + 1 ) = ω ( t ) + α ω | Δ R e l ( t ) | ( 1 ω ( t ) ) , if Δ R e l ( t ) > 0 ω ( t ) + α ω | Δ R e l ( t ) | ( 0 ω ( t ) ) , if Δ R e l ( t ) < 0 ,
where Δ R e l specifies the difference in the reliability between color and shape attributes in predicting rewards. Intuitively, the model assigns higher weight to the attribute which is estimated to have higher reliability for the given trial. We defined reliability based on the absolute value of the reward prediction error (RPE) for each attribute. Specifically, the reliability of an attribute is defined as inversely proportional to the magnitude of its RPE—a lower RPE (i.e., less surprise) indicates higher reliability. Based on this definition, the reliability difference Δ R e l is given by:
Δ R e l ( t ) = | R P E S h a p e | | R P E C o l o r | , = | R ( t ) Q c , S h a p e | | R ( t ) Q c , C o l o r | ,
where R P E C o l o r and R P E S h a p e refer to the reward prediction error between actual and predicted reward outcomes based on color or shape attribute, respectively. For example, if the color attribute yields a smaller RPE magnitude than the shape attribute, then Δ R e l > 0 , and the model increases ω to bias decision making toward the color attribute on the next trial.
We considered three distinct decision-making strategies, each implemented as a special case of the arbitration model described above. In the first case, the model fixed ω = 1 for all trials, representing an agent who relies exclusively on the color attribute. In the second case, ω = 0.5 was fixed, modeling an agent who assigns equal weight to both color and shape attributes. In the third case, the full arbitration mechanism was implemented, with the agent dynamically adjusting ω based on the relative reliability of the two attributes in predicting reward. For this model, ω was initialized at 0.5 on the first trial. For the full model, we used the following parameter values: α = 0.4 , β = 10 , γ = 0.2 , α ω = 0.4 .
To quantify the relative dominance of color-based versus shape-based learning strategies, we computed the conditional entropy of reward-dependent strategy (ERDS) separately for each attribute. Specifically, we defined stay/switch strategies separately for color and shape—denoted S t r C o l o r and S t r S h a p e , respectively—based on whether the agent repeats or switches their choice option with respect to each attribute after reward feedback. The corresponding ERDS measure for each attribute was then defined as follows:
ERDS C o l o r = H ( S t r C o l o r | R ) , ERDS S h a p e = H ( S t r S h a p e | R ) ,
To quantify the relative dominance of strategies in response to reward feedback, we computed the difference Δ E R D S = E R D S S h a p e E R D S C o l o r . A higher value of Δ E R D S indicates greater reliance on a color-based choice strategy.

3. Results

In the following sections, we illustrate the utility of information-theoretic metrics for identifying distinct learning and decision-making mechanisms using simulated data described above. We begin by presenting example behavior of an RL agent with a single learning rate, shown for three values: α = { 0.2 , 0.4 , 0.6 } (Figure 2). The time course of the performance—defined as the probability of choosing the more rewarding option P ( B e t t e r ) —shows that, for all three learning rates, performance peaks within approximately 10 to 20 trials but more quickly for larger learning rates. However, the overall performance, calculated across the entire block, does not significantly differ among the three learning rates (rank sum test, p > . 05 ; inset in Figure 2b).
Compared to the performance, the information-theoretic metrics exhibited more distinct trends across the different values of α (Figure 2c–f). For example, the entropy of strategy, H ( S t r ) , was highest for the RL agent simulated with α = 0.2 , followed by α = 0.6 and α = 0.4 , which did not differ significantly from each other (inset in Figure 1c). Moreover, the trajectory of H ( S t r ) , which measures the entropy of stay/switch strategy, decreased over trials within a block and peaked following the reversal—when the better and worse options were swapped. The normalized mutual information between reward and strategy (n-MIRS) revealed that information gained from reward feedback was highest for α = 0.6 , followed by 0.4 , and lowest for 0.2 (Figure 1d), consistent with the interpreting the learning rate as a measure of how strongly choices are adjusted based on reward feedback. The normalized mutual information between choice option and strategy (n-MIOS) indicated that the agents gradually learned to identify and persist with the better option over trials, as reflected by the increasing P ( B e t t e r ) and decreasing H ( S t r ) (Figure 1e). Finally, the normalized mutual information between option-reward combinations and strategy (n-MIROS) revealed that, despite differences in their temporal trajectories, the overall information shared between these variables plateaued at similar levels across all three learning rates (Figure 2f). However, the distinct temporal dynamics led to significant differences among agents as reflected in the average metrics values computed over the entire block (inset in Figure 2f).
Overall, this example illustrates that information-theoretic metrics can detect subtle variations in underlying mechanisms—such as variations in the learning rates—that simpler measures like performance may fail to capture (compare the inset in Figure 2b with those in Figure 2c–f). In the following four sets of simulations, we further show how these information-theoretic metrics can differentiate between the choice behavior of various RL models used as ground truths.

3.1. Revealing Positivity Bias

In the first set of simulations, we examined whether positivity bias—formally defined as a larger learning rate for positive outcomes compared to negative outcome ( α + > α )—can be captured using information-theoretic metrics. To that end, we used a standard RL (Figure 3a) to simulate choice behavior in a probabilistic reversal learning task under different reward scheudles: 80/20% and 40/10% reward schedules. We explored different combination of α + and α values to examine how asymmetric learning rates influence behavior across these environments.
We found that that higher positivity bias (larger Δ α = α + α ) resulted in higher performance across the two reward environments, especially for the 40/10% schedule (Pearson’s correlation between Δ α and P ( B e t t e r ) ; 40/10%: r = .904, p = 4.32×10-108; 80/20%: r = .702, p = 2.88×10-44) (Figure 3b). This is consistent with the previous literature [27,44,45,76], demonstrating that positivity bias can improve performance in terms of reward harvest.
Because learning rates control how value estimates are updated in response to reward feedback, we hypothesized that information-theoretic metrics related to reward—specifically ERDS and n-MIRS—would be sufficient to detect positivity bias in the choice behavior. Following this intuition, we first compared the choice behavior of the standard RL agents simulated with selected values of α + and α that exemplify "optimistic" ( α + > α ), "neutral" ( α + = α ), and "pessimistic" ( α + < α ) tendencies Figure 3c).
We found that the entropy of stay/switch strategy conditioned on reward (ERDS) was highest for the most pessimistic agent (cyan lines in Figure 3d), whereas the normalized mutual information between reward and strategy (n-MIRS) was lowest for the same agent (cyan lines in Figure 3e). However, metrics computed from the neutral agent (black lines) and the two optimistic agents (red and magenta lines) were less distinguished and plateaued at similar values. When simulating the full grid of α + and α values, we observed that ERDS was significantly correlated with Δ α , with higher positivity bias associated with lower ERDS (r = -.778, p = 6.25×10-60; Figure 3f). In contrast, n-MIRS was only weakly correlated with Δ α (r = -.137, p = .020), and the direction of this relationship was opposite to that observed in the example shown in Figure 3e, suggesting an overall nonlinear relationship. Therefore, while a stronger positivity bias is generally associated with lower ERDS, the absolute values of these metrics alone do not reliably indicate whether a given behavior reflects positivity bias.
To identify candidate metrics that may be more informative of positivity bias, we next examined the decomposition of the above metrics based on reward outcome (reward vs. no reward). Intuitively, there is greater uncertainty about whether the agents will stay/switch following unrewarded trials compared to rewarded ones. Over time, agents learn to maximize reward by adopting a stay-dominated strategy following rewarded trials, as they continue selecting the better option. In contrast, after unrewarded trials, agents are more likely to switch. However, if the chosen option still has a relatively high reward probability (i.e., it is the better option), the likelihood of staying after a no-reward trial can also increase over time.
Consistent with this notion, we found that the trajectory of the decomposition of mutual information metric following no reward (n-MIRS- in Figure 3g) became negative over time, suggesting that negative reward outcomes became less informative about the agents’ subsequent strategy (i.e., H ( S t r ) < H ( S t r | R = l o s s ) ). However, during the early portion of the block, unrewarded trials were initially informative about choice strategy,as indicated by the positive n-MIRS- values. Optimistic and pessimistic agents exhibited distinct temporal dynamics in how quickly this metric shifted toward negative values (shaded gray area in Figure 3g). When computed over the first 10 trials, the sign of the n-MIRS– metric was approximately predictive of whether an agent exhibited positivity bias or not (Figure 3g inset). Specifically, more optimistic agents exhibited more negative n-MIRS- values, indicating that their tendency to update more strongly after rewarded than unrewarded trials. This makes the no-reward trials less informative of their subsequent strategy. This pattern is consistent with the results shown in Figure 3b, where optimistic agents achieved better performance. To confirm this effect across the full parameter space, we computed n-MIRS- as a function of Δ α and found a strong linear relationship in the 80/20% environment (r = -.926, p = 1.27×10-123), as well as in the other two tested reward environments (60/40% and 40/10%; Figure 3h). These results suggest that the decomposed mutual information measures—specifically n-MIRS- computed from the early period of each block—can be used to predict the presence of positivity bias. To test this idea, we conducted a classification analysis in which we randomly sampled groups of optimistic and pessimistic agents and used the information-theoretic metrics to predict whether a given metric profile originated from an optimistic agent using a linear decoder (Figure 3i; see Section 2.3.1 for details).
We found that the n-MIRS- metric alone, when computed from the first 10 trials, yielded high cross-validated decoding accuracy across different environments (Figure 3j), significantly exceeding chance level (signed-rank test against 0.50; p < . 001 for all environments). Including all available metrics in the decoder led to a small but significant improvement in the decoding accuracy (signed-rank test; p = . 0011 for 50/10%, p < . 001 for all other environments). In comparison, a decoder based on logistic regression weights achieved comparable accuracies (white bars in Figure 3j), indicating that the information-theoretic metrics contain comparable information to traditional regression-based measures—while offering more interpretable connections to underlying learning and decision-making processes. Overall, these results illustrate that information-theoretic metrics contain sufficient information to determine the presence of positivity bias when the ground truths are provided.

3.2. Revealing Reward-Dependent Metaplasticity

One potential neural mechanism underlying differential learning rates is reward-dependent metaplasticity, which generates dynamic learning rates by naturally adapting to the recent history of rewards in the environment [46,47,73,77]. To test whether such dynamic changes in the learning rates can be detected using information-theoretic metrics, we implemented a variant of the metaplastic model proposed by Farashahi et al. [46]. In our simulations, we assumed that each of the two choice options is assigned a set of metaplastic synapses which undergo transitions in response to reward feedback (Figure 4b; Section 2.3.2). By computing the "effective" learning rate—defined as the overall rate of value update on each trial—we found that the metaplastic model exhibited diverging learning rates following rewarded and unrewarded trials, with α + > α (solid lines in Figure 4c), consistent with previous findings [46]. The difference between the effective α + and α increased over time and plateaued just before the reversal (red solid curve in Figure 4c), driven primarily by an increase in α + rather than a decrease in α . This happens because, as more reward is obtained, the "weak" synapses encoding the better option mostly occupy the most unstable meta-states w 1 (Figure 4b), which has the highest transition probability α 1 toward the "strong" state s 1 , resulting in a high effective α + .
Using the choice behavior of this metaplastic agent as ground truth, we fitted a ’plastic’ model (standard RL) with differential learning rates to the simulated choice data from each block to test whether the fitted parameters reflected the pattern α + > α (noting that these learning rates are constant by definition in this model). The results confirmed this pattern (red and blue dashed lines in Figure 4c), suggesting that empirically observed trends of α + > α in previous studies may, in fact, reflect underlying reward-dependent metaplasticity.
To test whether the distinct learning mechanisms of the plastic and metaplastic models can be identified from choice behavior, we simulated the choice behavior of a metaplastic model and two plastic models, and then computed information-theoretic metrics from the simulated behavior. Specifically, we compared metrics across three models: (1) a plastic model with a single α = 0.3 (plastic 1- α ), (2) a metaplastic model initialized with the same effective learning rate of 0.3 on the first trial, and (3) a plastic model with differential learning rates (plastic 2- α ) estimated from the metaplastic agent.
Comparison of the entropy of choice strategy indicated that the metaplastic agent was more consistent in its strategy compared to both plastic agents (rank sum test on H ( S t r ) : vs. 1- α : p = 4.43×10-28; vs. 2- α : p = 2.43×10-24; Figure 4d). Moreover, the n-MIOS metric revealed that the metaplastic agent had the advantage of increasing the mutual information between the choice of the better option and subsequent strategy (Figure 4e), reflected in its superior performance relative to both plastic models (rank sum test on n-MIOS: vs. 1- α : p = 5.58×10-4; vs. 2- α : p = .00280). The two plastic agents (1- α and 2- α ) were less distinguishable from each other in terms of H ( S t r ) (p = .0120), n-MIOS (p = .683), and overall performance (p = .848).
These results suggest that, although the parameters of the plastic 2- α model were directly estimated from the metaplastic agent, differences in the underlying learning mechanisms still give rise to distinct patterns in the information-theoretic metrics. This highlights the potential utility of such metrics in model recovery and validation—for example, by assessing whether a given candidate RL model can reproduce the information-theoretic metrics observed in the empirical data [29].
Given the observed differences between plastic and metaplastic models, we next quantified the extent to which the information-theoretic metrics discriminate between different models. We hypothesized that these metrics could be used to decode changes in the learning rates corresponding to positivity bias, using an approach similar to that employed in the previous section. To obtain highest accuracy, we utilized all available information-theoretic metrics and their decompositions to train a linear decoder to discriminate between optimistic and pessimistic agents (see Section 2.3.2 in Materials and Methods for more details). We then applied this decoder to the metrics generated by the three models described above and computed the posterior probability that the decoder would classify the behavior as exhibiting positivity bias.
We found that the posterior probabilities derived from the metaplastic model (magenta line in Figure 4f) exhibited temporal dynamics closely resembling those of effective learning rates (Figure 4c). Specifically, as trials progressed, the decoder increasingly indicated that the behavior originated from an optimistic agents with α + > α . In contrast, the posterior probabilities for the two plastic agents remained relatively constant across the block. The plastic 1- α agent (black line in Figure 4f) maintained performance at chance level, as it was neither optimistic nor pessimistic by design. On the other hand, choice behavior of the plastic 2- α agent (gray line in Figure 4f) was classified as being optimistic, consistent with its estimated learning rates reflecting α + > α (dashed lines in Figure 4c).

3.3. Revealing Behavioral Adjustments due to Reward Harvest Rate

Next, we investigated how information-theoretic metrics can be used to detect the influence cumulative overall reward feedback on all options (i.e., reward harvest rate) on an agent’s choice strategy. To that end, we used two types of RL model to generate choice behavior: (1) the standard RL with a single learning rate α , and (2) the augmented RL model incorporating a reward harvest rate component (Figure 5a). Briefly, the augmented model tracks an average reward rate H, which was not tied to any specific choice option, and uses it to modulate the agent’s win-stay and lose-switch tendencies (see Section 2.3.3 in Materials and Methods for more details ). To quantify the effect of reward harvest rate on choice behavior, we defined a discrete variable referred to as the reward harvest state (RHS)—a binary indicator of whether the reward rate on a given trial was above or below the block median [35].
As shown for an example block, high or low reward harvest state (RHS) did not consistently coincide with reward (or no reward) on the previous trial (Figure 5b), given the weak correlation between the reward rate and binary reward feedback (r = .107, p = .343). Exploiting this property, we used RHS as an additional conditioning variable for the strategy S t r , and measured the joint influence of previous reward and RHS on the agent’s subsequent strategy using the mutual information metric n-MIRHS (see Section 2.3.3 for more details). The trajectory of this metric revealed that overall information was significantly higher for the agent with the additional reward harvest rate component (RL with RHR agent) compared to the standard RL agent (rank sum test, p = 2.56×10-34), reflecting the added influence of RHS on choice strategy Figure 5c.
To gain insight on how the high and low RHS states influence behavior, we next examined the decomposition of n-MIRHS based on RHS, corresponding to the influence of high ( n-MIRHS h i g h , Figure 5d) and low RHS ( n-MIRHS l o w , Figure 5e). We found that the shared information between the h i g h state and choice strategy was overall comparable across models but significantly higher for the standard RL agent than for the RL with RHR agent (difference in mean n-MIRHS h i g h = 0.0445; rank sum test, p = 4.67×10-34). In contrast, n-MIRHS l o w revealed a more pronounced distinction: it was significantly positive for the RL with RHR agent (signed-rank test, p = 3.90×10-18), but not significantly different from zero for the standard RL model (signed-rank test, p = .064; mean = 0.0019; inset in Figure 5e).
These results indicate that during the high state—when the reward rate was higher than usual due to successful learning—both models showed positive mutual information between the high RHS and subsequent strategy. In contrast, during the low state—when the agents were receiving rewards at a lower-than-usual rate—RHS was not informative about whether the standard RL agent would stay or switch on the next trial. This was not the case for the RL with RHR agent, for which n-MIRHS l o w > 0 . Further decompositions based on reward feedback revealed that this effect was driven by a larger magnitude of n-MIRHS l o w + (following rewarded trials) compared to n-MIRHS l o w (following unrewarded trials; compare red and blue bars in the Figure 5e inset for the RL with RHR agent). This pattern reflects an overall positive shift in information content resulting from the architecture of the RL with RHR model, which modulates stay/switch tendency as a function of the reward harvest rate H.
Lastly, to gain further insight into how the n-MIRHS metric rely on the model parameters, we simulated choice behavior of the model with additional modulation by reward harvest rate using the full range of α and β H values. Using standardized regression, we then examined the contribution of the β H parameter, which determines the influence of reward harvest rate on choice behavior ( β H = 0 corresponds to the standard RL. We found that, consistent with the example in (Figure 5c), larger values of β H predicted overall higher values of n-MIRHS (standardized regression of n-MIRHS on β H , α , and their interaction: b 1 = 0.366, p = 3.13×10-216). Similar regression analyses on each of the decompositions yielded results consistent with Figure 5d,e. More specifically, β H did not significantly affect the n-MIRHS decomposition for the high reward state (standardized coefficient predicting n-MIRHS h i g h : b 1 = 0.0150, p = .329) nor its further reward-based decompositions ( n-MIRHS h i g h + : b 1 = 0.024, p = .109; n-MIRHS h i g h : b 1 = -0.00665, p = .674). In contrast, higher values of β H predicted significantly larger n-MIRHS l o w ( b 1 = 0.464, p = 4.37×10-308). The decompositions based on reward feedback revealed that this effect was strongest for n-MIRHS l o w , which was significantly higher for larger values of β H (standardized coefficient: b 1 = 0.523, p = 1.53×10-253; dashed lines in Figure 5f, right). In comparison, n-MIRHS l o w + was also significantly predicted by β H , but to a lesser degree ( b 1 = 0.121, p = 1.90×10-19; solid lines in Figure 5f, right). These results suggest that the influence of reward harvest rate on promoting heuristic win-stay/lose-switch is primarily mediated by increased information value of no reward in the low reward state. That is, the larger β H , corresponding to the effect of reward rate, was associated with less negative n-MIRHS l o w toward zero, thus increasing the informativeness of receiving no reward during this state. This can be used to detect modulations of choice behavior by reward harvest rate as these effects can change across blocks of trials.

3.4. Revealing the Presence of Alternative Learning Strategies in Multidimensional Reward Environments

Finally, we examined whether the information-theoretic metrics can be utilized to identify the presence of alternative learning strategies in multidimensional reward environments. In naturalistic settings, reward outcomes often depend on distinct features or attributes of the available choice options. To mimic such scenarios, we simulated a task in which only one of the two choice attributes—either the shape or the color of the stimuli—was associated with reward, thereby introducing uncertainty about which attribute was predictive of reward outcomes (Figure 6a; see Section 2.3.4 for more details). We considered three RL agents, each exhibiting distinct types of choice strategy: (1) an agent that only learns and chooses based on the color attribute (Color-only), (2) an agent that equally weighs the color and shape attributes without arbitration (No Arbitration), and (3) an agent which dynamically arbitrates between two attributes based on the reliability of each attribute in predicting reward (Dynamic Arbitration; Figure 6b).
Examining the arbitration weight ω across different block types revealed that the Dynamic Arbitration agent successfully adapted its strategy to the environment: it consistently increased ω during Color-to-Color blocks, indicating a growing reliance on color (magenta curve in Figure 6e, top), and shifted ω toward a shape-based strategy following reversals in Color-to-Shape blocks (magenta curve in Figure 6f, top). When reversals occurred within the color attribute (Color-to-Color blocks), the performance trajectories ( P ( B e t t e r ) ) were qualitatively similar across all agents, as expected. In each model, performance dropped immediately after reversals followed by gradual recovery (Figure 6e, bottom). As expected, the Color-only agent reached its peak performance more quickly (Figure 6e, bottom), as it did not consider shape values, which were irrelevant in this condition. During Color-to-Shape blocks, when the color information was no longer predictive of rewards after reversals, the Dynamic Arbitration agent performed the best (Figure 6f, bottom), as this agent appropriately identified the correct reward-predictive attribute by dynamically arbitrating between the two feature dimensions (Figure 6f, bottom).
Although the learning strategy of the Color-only agent can be readily distinguished by its P ( B e t t e r ) trajectory during the Color-to-Shape blocks, distinguishing among all three agents is more challenging in Color-to-Color blocks based on performance alone. Therefore, we examined whether information-theoretic metrics could be used to distinguish among the three types of learning strategies represented by the RL agents. Although the underlying arbitration weight ω can, in principle, be estimated through model fitting, behavioral metrics offer a more direct window into learning and decision-making strategies. Moreover, these metrics can be computed over a subset of trials and require far simpler computations compared to model fitting, which depends on continuity in the choice data.
To quantify reward-dependent strategies, we computed the conditional entropy of reward-dependent strategies (ERDS) separately for each attribute— ERDS S h a p e and ERDS C o l o r (Figure 7a,b). We found that during Color-to-Color blocks, the entropy of the shape-based strategy ( ERDS S h a p e ) was close to its maximum value of 1 for all the three agents, reflecting the fact that shape carried no information about reward in this block type (dashed lines in Figure 7a). In contrast, the entropy of the color-based strategy ( ERDS C o l o r ) exhibited modulation across trials, with distinct dynamics for each agent (solid lines in Figure 7a). Consistent with the performance trajectories, the Color-only agent showed the lowest entropy, followed by the Dynamic Arbitration agent (rank sum test on ERDS C o l o r , p = 7.98×10-34). To assess the relative dominance of the two learning strategies, we computed the difference between the two entropy values, ERDS S h a p e ERDS C o l o r . The resulting positive values indicated that all three agents correctly prioritized the color-based learning strategy during this block type, as reflected by lower entropy for the relevant dimension (i.e., color) Figure 7c). These two metrics thus provide insight beyond the trajectories of arbitration weight ω (Figure 6e top), which does not directly capture changes in the learned values that could drive choice behavior.
During the Color-to-Shape blocks, the entropy values ( ERDS S h a p e ) indicated near-random use of the shape-based strategy for all three agents prior to the reversal (dashed lines in Figure 7b). After the reversal—when shape attribute became informative of reward—the Dynamic Arbitration showed the most pronounced adjustment, characterized by a decreased reliance on the color-based strategy (reflected by as drop in ERDS C o l o r ; solid magenta line in Figure 7b), and an increased reliance on the shape-based strategy (reflected by a rise in ERDS S h a p e ; dashed magenta line). The difference in the two entropy values further revealed that only the agents capable of learning both attributes (No Arbitration and Dynamic Arbitration) were able to shift their strategies after the reversal, as indicated by ERDS C o l o r > ERDS S h a p e (Figure 7d). In contrast, the Color-only agent exhibited ERDS C o l o r < ERDS S h a p e throughout the entire block.
To quantify potential interactions between the color-based and shape-based learning strategies, we measured the correlation between ERDS S h a p e and ERDS C o l o r , computed separately from two distinct periods within each block. "Non-stationary" trials were defined as the first 30 trials at the beginning of each block and those immediately following the reversal (highlighted by red bars in Figure 7a,b). In contrast, "steady" trials corresponded to the final 10 trials before the reversal or at end of each block, during which performance had plateaued (blue bars in Figure 7a–d).
This analysis revealed distinct patterns of interaction between the two learning strategies across the three agents (Figure 7e,f). More specifically, the Color-only agent showed no significant correlation between ERDS S h a p e and ERDS C o l o r during either periods of the Color-to-Shape blocks (non-stationary: r = .137, p = .174; steady: r = .072, p = .477; Figure 7e). This result is expected, as the Color-only agent does not employ any shape-based strategy, which renders response to reward based on shape effectively random. In contrast, the No Arbitration agent exhibited competitive interaction between the two strategies, as indicated by a significant negative correlation between ERDS S h a p e and ERDS C o l o r (non-stationary: r = -.435, p = 6.08×10-6; steady: r = -.386, p = 7.43×10-5; Figure 7e). This pattern reflects the agent’s decision-making strategy , in which color and shape information are always weighted equally, regardless of which attribute is currently predictive of reward (fixed ω = 0.5 ). Interestingly, the Dynamic Arbitration agent exhibited different trend of interaction across the two periods of the block. During the non-stationary phase, this model exhibited a significant negative correlation between ERDS S h a p e and ERDS C o l o r (r = -.374, p = 1.28×10-4). However, this correlation disappeared during the steady phase of the block (r = -.0409, p = .686), when the color-based strategy became dominant (i.e., ω > 0.5 ). This shift in correlation pattern can be used to detect dynamic arbitration between alternative learning strategies.
We found overall consistent results during the Color-to-Shape blocks (Figure 7f). The Color-only agent showed no significant interaction between strategies in either period (non-stationary: r = -.0346, p = .733; steady: r = -.055, p = .587). In contrast, the No Arbitration agent exhibited significant negative correlations during both phases of the block (r = -.522, p = 2.59×10-8; steady: r = -.269, p = .00685). Notably, only the Dynamic Arbitration agent—capable of adapting its behavior based on relative reliability of the two attributes in predicting reward—exhibited a shift in the interaction pattern over time. Specifically, it showed a significant negative correlation during the non-stationary period (r = -.536, p = 9.19×10-9), which weakened and became non-significant during the steady phase (r = -.151, p = .135), consistent with the pattern observed in the Color-to-Color blocks. Together, these results demonstrate that a negative correlation between ERDS S h a p e and ERDS C o l o r during the non-stationary phase and the absence of such a correlation during the steady phase is indicative of dynamic arbitration between alternative learning strategies.

4. Discussion

In this study, we demonstrate how behavioral metrics inspired by information theory can be used to identify certain learning and decision-making mechanisms. In particular, we applied metrics based on conditional entropy, normalized mutual information, and their decompositions based on specific outcomes to choice behavior during two variants of the probabilistic reversal learning task—a widely adopted paradigm for studying cognitive flexibility across species. Using these metrics, we investigated whether specific neural or computational mechanisms—specified by the reinforcement learning (RL) models serving as ground truth—could be inferred from choice behavior. To that end, we examined positivity bias, gradual changes in the learning rates due to reward-dependent metaplasticity, the influence of reward harvest rate, and the adoption and arbitration between alternative learning strategies in multidimensional environments.
One of the key strengths of the proposed information-theoretic metrics lies in their versatility and flexibility. In contrast, fitting RL models to choice behavior in order to identify underlying mechanisms requires a continuous stream of choice and reward data, with the precision of the estimated parameters heavily dependent on the amount of available data. As our results demonstrate, information-theoretic metrics circumvent these limitations, as they can be computed from as few as several dozen trials, even when drawn from non-contiguous segments of the data. Directly comparing the information content of our metrics with regression weights for predicting choice (Figure 3j), we found similar decoding performance from both types of measures. However, information-theoretic metrics offer the distinct advantage of being simple and model-free, whereas regression-based features are inherently model-dependent—–requiring decisions about which predictors to include, how many lags to consider, how to evaluate model fit and statistical significance, and how to interpret the resulting regression weights.
Moreover, time courses of information-theoretic metrics can be computed for each trial point by concatenating trial-wise data across different blocks. This approach serves as a useful visualization tool for examining behavioral changes over time. Finally, the analysis in Figure 7 demonstrates how metrics computed from different periods within a block can provide insight into temporal changes in learning and choice strategies. Taken together, the efficiency and flexibility of these metrics make them valuable model-free tools for identifying and testing hypotheses about the mechanisms underlying learning and decision making.
Another important advantage of information-theoretic metrics is their adaptability to target specific variables of interest within a given study. As illustrated in simulations of reward harvest rate effects on behavior (Figure 5), the key variable—reward harvest state—can be incorporated into the mutual information metric n-MIRHS to assess how reward harvest rate influences choice behavior. This measure revealed the role of reward harvest rate in shaping choice strategy under varying reward contingencies by quantifying the information flow from prior reward outcomes and reward states into the agent’s choice strategy. Therefore, these metrics can be flexibly formulated to match the demands of a given research context and provide quantifiable insights into the role of specific variables.
In turn, the insights provided by these metrics can guide the development of improved models that better capture key aspects of behavioral data [29]. As numerous studies have pointed out [78,79,80], the best-fitting model among a set of candidates may still fail to reproduce important features of choice behavior. Therefore, validating candidate models through posterior predictive checks remains a critical step in modeling [26,81]. Information-theoretic metrics offer ideal tools for this process, as they provide summary statistics specific to the variables of interest. For example, different RL models could be validated by comparing information-theoretic metrics computed from simulated data with those derived from empirical data [29,38]. By quantifying the divergence between empirical observations and model-generated data, this comparison both reveals a model’s limitations and pinpoints the mechanisms that must be added for improvement [29]. Such an approach is especially valuable when standard model-comparison metrics are inconclusive—for example, with small or noisy datasets that inherently favor simpler, low-parameter models. In those cases, one can use information-theoretic measures to constrain the candidate models by selecting the one whose simulated statistics most closely match the empirical data. Lastly, although this study focuses on RL models, the same framework can be applied to other model classes—such as those based on Bayesian inference [32,82].
In addition to serving as useful tools for model discovery and validation, information-theoretic metrics may also contribute to data-driven approaches in studies involving choice behavior. Our decoding analyses in Section 3.1 and Section 3.2 serves as a proof-of-concept, showing that the information-theoretic measures contain sufficient information to detect positivity bias when the ground truths are known. This example highlights the broader potential of the metrics in detecting latent cognitive states or mechanisms other than positivity bias (which may lack ground truths in real data). For example, one could train a decoder to predict experimentally defined task states from either behavioral measures (reaction time, reward rate, pupil dilation) or neural recordings (spike trains, EEG, fMRI) by using information-theoretic features extracted from choice data. In this approach, the experimentally defined states serve as the ground-truth labels during training, and the decoder’s ability to recover those labels is then evaluated on held-out behavioral (or neural) data. Future research could apply this model-agnostic yet interpretable framework to infer latent cognitive or neural states.
There are some challenges and considerable potential for extending the information-theoretic framework presented here. For example, while our entropy measures quantify the overall consistency in the stay/switch strategy in response to reward and other relevant variables, they do not directly capture directionality—that is, they do not indicate what causes what. As a result, these metrics should be considered alongside other measures to provide a more complete picture. This limitation makes interpreting mutual information versus conditional entropy particularly challenging. Moreover, while information-theoretic metrics capture sensitivity to reward feedback, they cannot be uniquely mapped onto the learning rates. Finally, although we have focused on conditional entropy and normalized mutual information, other behavioral metrics based on other concepts from information theory can be developed including transfer entropy [83,84], mutual information between discrete and continuous variables [85,86,87], and partial information decomposition [88,89], among others. Incorporating these extensions would allow the behavioral metrics to be generalized to more complex tasks, including those involving more than two alternative options, higher-dimensional feature spaces, and task contexts.

Author Contributions

Conceptualization, J.H.W. and A.S.; methodology, J.H.W. and A.S.; software, J.H.W. and L.B.; validation, J.H.W.; formal analysis, J.H.W., L.B., and A.S.; investigation, J.H.W. and L.B.; resources, J.H.W. and L.B.; data curation, J.H.W.; writing—original draft preparation, J.H.W., L.B., and A.S.; writing—review and editing, J.H.W., L.B., and A.S.; visualization, J.H.W.; supervision, A.S.; project administration, J.H.W. and A.S.; funding acquisition, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Institutes of Health grant R01 DA047870 to A.S.

Data Availability Statement

The analysis codes used for this study is available at https://github.com/DartmouthCCNL/woo_etal_entropy.

Acknowledgments

We thank the members of the Computational Cognitive Neuroscience Lab, Peter Duggins, Michael Chong Wang, Aryan Yazdanpanah, and Atlas Shahamati, for their helpful comments on the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lungarella, M.; Sporns, O. Mapping information flow in sensorimotor networks. PLoS computational biology 2006, 2, e144. [Google Scholar] [CrossRef]
  2. Palmigiano, A.; Geisel, T.; Wolf, F.; Battaglia, D. Flexible information routing by transient synchrony. Nature Neuroscience 2017, 20, 1014–1022. [Google Scholar] [CrossRef] [PubMed]
  3. Honey, C.J.; Kötter, R.; Breakspear, M.; Sporns, O. Network structure of cerebral cortex shapes functional connectivity on multiple time scales. Proc Natl Acad Sci U S A 2007, 104, 10240–10245. [Google Scholar] [CrossRef] [PubMed]
  4. Stramaglia, S.; Wu, G.R.; Pellicoro, M.; Marinazzo, D. Expanding the transfer entropy to identify information circuits in complex systems. Phys. Rev. E 2012, 86, 066211. [Google Scholar] [CrossRef]
  5. Novelli, L.; Wollstadt, P.; Mediano, P.; Wibral, M.; Lizier, J.T. Large-scale directed network inference with multivariate transfer entropy and hierarchical statistical testing. Network Neuroscience 2019, 3, 827–847. [Google Scholar] [CrossRef]
  6. Vicente, R.; Wibral, M.; Lindner, M.; Pipa, G. Transfer entropy—a model-free measure of effective connectivity for the neurosciences. Journal of computational neuroscience 2011, 30, 45–67. [Google Scholar] [CrossRef]
  7. Ursino, M.; Ricci, G.; Magosso, E. Transfer Entropy as a Measure of Brain Connectivity: A Critical Analysis With the Help of Neural Mass Models. Front Comput Neurosci 2020, 14, 45. [Google Scholar] [CrossRef]
  8. Strong, S.P.; Koberle, R.; De Ruyter Van Steveninck, R.R.; Bialek, W. Entropy and Information in Neural Spike Trains. Phys. Rev. Lett. 1998, 80, 197–200. [Google Scholar] [CrossRef]
  9. Gourévitch, B.; Eggermont, J.J. Evaluating Information Transfer Between Auditory Cortical Neurons. Journal of Neurophysiology 2007, 97, 2533–2543. [Google Scholar] [CrossRef]
  10. Cofré, R.; Maldonado, C. Information Entropy Production of Maximum Entropy Markov Chains from Spike Trains. Entropy (Basel) 2018, 20, 34. [Google Scholar] [CrossRef]
  11. Shorten, D.P.; Spinney, R.E.; Lizier, J.T. Estimating Transfer Entropy in Continuous Time Between Neural Spike Trains or Other Event-Based Data. PLOS Computational Biology 2021, 17, e1008054. [Google Scholar] [CrossRef] [PubMed]
  12. Strange, B.A.; Duggins, A.; Penny, W.; Dolan, R.J.; Friston, K.J. Information theory, novelty and hippocampal responses: unpredicted or unpredictable? Neural Networks 2005, 18, 225–230. [Google Scholar] [CrossRef] [PubMed]
  13. Bach, D.R.; Dolan, R.J. Knowing how much you don’t know: a neural organization of uncertainty estimates. Nat Rev Neurosci 2012, 13, 572–586. [Google Scholar] [CrossRef] [PubMed]
  14. Sayood, K. Information Theory and Cognition: A Review. Entropy (Basel) 2018, 20, 706. [Google Scholar] [CrossRef]
  15. Seiler, J.P.H.; Dan, O.; Tüscher, O.; Loewenstein, Y.; Rumpel, S. Experienced entropy drives choice behavior in a boring decision-making task. Sci Rep 2022, 12, 3162. [Google Scholar] [CrossRef]
  16. Jepma, M.; Nieuwenhuis, S. Pupil diameter predicts changes in the exploration-exploitation trade-off: evidence for the adaptive gain theory. J Cogn Neurosci 2011, 23, 1587–1596. [Google Scholar] [CrossRef]
  17. Wang, M.Z.; Hayden, B.Y. Monkeys are curious about counterfactual outcomes. Cognition 2019, 189, 1–10. [Google Scholar] [CrossRef]
  18. Woo, J.H.; Azab, H.; Jahn, A.; Hayden, B.; Brown, J.W. The PRO model accounts for the anterior cingulate cortex role in risky decision-making and monitoring. Cognitive, Affective, & Behavioral Neuroscience 2022, 22, 952–968. [Google Scholar] [CrossRef]
  19. Lee, D.; Conroy, M.L.; McGreevy, B.P.; Barraclough, D.J. Reinforcement learning and decision making in monkeys during a competitive game. Cognitive Brain Research 2004, 22, 45–58. [Google Scholar] [CrossRef]
  20. Lee, D.; McGreevy, B.P.; Barraclough, D.J. Learning and decision making in monkeys during a rock–paper–scissors game. Cognitive Brain Research 2005, 25, 416–430. [Google Scholar] [CrossRef]
  21. Takahashi, H.; Izuma, K.; Matsumoto, M.; Matsumoto, K.; Omori, T. The Anterior Insula Tracks Behavioral Entropy during an Interpersonal Competitive Game. PLoS One 2015, 10, e0123329. [Google Scholar] [CrossRef] [PubMed]
  22. Sutton, R.S.; Barto, A.G. Reinforcement Learning, second edition: An Introduction; MIT Press, 2018. Google-Books-ID: uWV0DwAAQBAJ.
  23. Dayan, P.; Daw, N.D. Decision theory, reinforcement learning, and the brain. Cognitive, Affective, & Behavioral Neuroscience 2008, 8, 429–453. [Google Scholar] [CrossRef]
  24. Niv, Y. Reinforcement learning in the brain. Journal of Mathematical Psychology 2009, 53, 139–154. [Google Scholar] [CrossRef]
  25. Jensen, K.T. An introduction to reinforcement learning for neuroscience. arXiv 2024, arXiv:2311.07315. [Google Scholar] [CrossRef]
  26. Eckstein, M.K.; Wilbrecht, L.; Collins, A.G. What do reinforcement learning models measure? Interpreting model parameters in cognition and neuroscience. Current Opinion in Behavioral Sciences 2021, 41, 128–137. [Google Scholar] [CrossRef]
  27. Nussenbaum, K.; Hartley, C.A. Reinforcement learning across development: What insights can we draw from a decade of research? Dev Cogn Neurosci 2019, 40, 100733. [Google Scholar] [CrossRef]
  28. Bartolo, R.; Averbeck, B.B. Prefrontal cortex predicts state switches during reversal learning. Neuron 2020, 106, 1044–1054. [Google Scholar] [CrossRef]
  29. Trepka, E.; Spitmaan, M.; Bari, B.A.; Costa, V.D.; Cohen, J.Y.; Soltani, A. Entropy-based metrics for predicting choice behavior based on local response to reward. Nat Commun 2021, 12, 6567. [Google Scholar] [CrossRef]
  30. Grossman, C.D.; Bari, B.A.; Cohen, J.Y. Serotonin neurons modulate learning rate through uncertainty. Curr Biol 2022, 32, 586–599.e7. [Google Scholar] [CrossRef]
  31. Yang, M.A.; Jung, M.W.; Lee, S.W. Striatal arbitration between choice strategies guides few-shot adaptation. Nature Communications 2025, 16, 1811. [Google Scholar] [CrossRef]
  32. Nassar, M.R.; Wilson, R.C.; Heasly, B.; Gold, J.I. An Approximately Bayesian Delta-Rule Model Explains the Dynamics of Belief Updating in a Changing Environment. J. Neurosci. 2010, 30, 12366–12378. [Google Scholar] [CrossRef]
  33. Farashahi, S.; Xu, J.; Wu, S.W.; Soltani, A. Learning arbitrary stimulus-reward associations for naturalistic stimuli involves transition from learning about features to learning about objects. Cognition 2020, 205, 104425. [Google Scholar] [CrossRef] [PubMed]
  34. Cinotti, F.; Coutureau, E.; Khamassi, M.; Marchand, A.R.; Girard, B. Regulation of reinforcement learning parameters captures long-term changes in rat behaviour. European Journal of Neuroscience 2024, 60, 4469–4490. [Google Scholar] [CrossRef] [PubMed]
  35. Wittmann, M.K.; Fouragnan, E.; Folloni, D.; Klein-Flügge, M.C.; Chau, B.K.; Khamassi, M.; Rushworth, M.F. Global reward state affects learning and activity in raphe nucleus and anterior insula in monkeys. Nature communications 2020, 11, 3771. [Google Scholar] [CrossRef]
  36. Lee, S.W.; Shimojo, S.; O’doherty, J.P. Neural computations underlying arbitration between model-based and model-free learning. Neuron 2014, 81, 687–699. [Google Scholar] [CrossRef] [PubMed]
  37. Charpentier, C.J.; Iigaya, K.; O’Doherty, J.P. A neuro-computational account of arbitration between choice imitation and goal emulation during human observational learning. Neuron 2020, 106, 687–699. [Google Scholar] [CrossRef]
  38. Woo, J.H.; Costa, V.D.; Taswell, C.A.; Rothenhoefer, K.M.; Averbeck, B.B.; Soltani, A. Contribution of amygdala to dynamic model arbitration under uncertainty. bioRxiv 2024, 2024.09.13.612869. [Google Scholar] [CrossRef]
  39. Philippe, R.; Janet, R.; Khalvati, K.; Rao, R.P.; Lee, D.; Dreher, J.C. Neurocomputational mechanisms involved in adaptation to fluctuating intentions of others. Nature communications 2024, 15, 3189. [Google Scholar] [CrossRef]
  40. Collins, A.G.E.; Shenhav, A. Advances in modeling learning and decision-making in neuroscience. Neuropsychopharmacol. 2022, 47, 104–118. [Google Scholar] [CrossRef]
  41. Woo, J.H.; Aguirre, C.G.; Bari, B.A.; Tsutsui, K.I.; Grabenhorst, F.; Cohen, J.Y.; Schultz, W.; Izquierdo, A.; Soltani, A. Mechanisms of adjustments to different types of uncertainty in the reward environment across mice and monkeys. Cogn Affect Behav Neurosci 2023, 23, 600–619. [Google Scholar] [CrossRef]
  42. Lefebvre, G.; Lebreton, M.; Meyniel, F.; Bourgeois-Gironde, S.; Palminteri, S. Behavioural and neural characterization of optimistic reinforcement learning. Nat Hum Behav 2017, 1, 1–9. [Google Scholar] [CrossRef]
  43. Palminteri, S.; Lefebvre, G.; Kilford, E.J.; Blakemore, S.J. Confirmation bias in human reinforcement learning: Evidence from counterfactual feedback processing. PLoS computational biology 2017, 13, e1005684. [Google Scholar] [CrossRef] [PubMed]
  44. Lefebvre, G.; Summerfield, C.; Bogacz, R. A normative account of confirmation bias during reinforcement learning. Neural Comput 2022, 34, 307–337. [Google Scholar] [CrossRef] [PubMed]
  45. Palminteri, S.; Lebreton, M. The computational roots of positivity and confirmation biases in reinforcement learning. Trends in Cognitive Sciences 2022, 26, 607–621. [Google Scholar] [CrossRef]
  46. Farashahi, S.; Donahue, C.H.; Khorsand, P.; Seo, H.; Lee, D.; Soltani, A. Metaplasticity as a Neural Substrate for Adaptive Learning and Choice under Uncertainty. Neuron 2017, 94, 401–414.e6. [Google Scholar] [CrossRef]
  47. Khorsand, P.; Soltani, A. Optimal structure of metaplasticity for adaptive learning. PLoS Comput Biol 2017, 13, e1005630. [Google Scholar] [CrossRef] [PubMed]
  48. Niv, Y.; Daw, N.D.; Joel, D.; Dayan, P. Tonic dopamine: opportunity costs and the control of response vigor. Psychopharmacology 2007, 191, 507–520. [Google Scholar] [CrossRef]
  49. Sukumar, S.; Shadmehr, R.; Ahmed, A.A. Effects of reward and effort history on decision making and movement vigor during foraging. Journal of neurophysiology 2024, 131, 638–651. [Google Scholar] [CrossRef]
  50. Soltani, A.; Koechlin, E. Computational models of adaptive behavior and prefrontal cortex. Neuropsychopharmacology 2022, 47, 58–71. [Google Scholar] [CrossRef]
  51. O’Doherty, J.P.; Lee, S.W.; Tadayonnejad, R.; Cockburn, J.; Iigaya, K.; Charpentier, C.J. Why and how the brain weights contributions from a mixture of experts. Neuroscience & Biobehavioral Reviews 2021, 123, 14–23. [Google Scholar] [CrossRef]
  52. Leong, Y.C.; Radulescu, A.; Daniel, R.; DeWoskin, V.; Niv, Y. Dynamic interaction between reinforcement learning and attention in multidimensional environments. Neuron 2017, 93, 451–463. [Google Scholar] [CrossRef] [PubMed]
  53. Farashahi, S.; Rowe, K.; Aslami, Z.; Lee, D.; Soltani, A. Feature-based learning improves adaptability without compromising precision. Nature Communications 2017, 8, 1768. [Google Scholar] [CrossRef]
  54. Radulescu, A.; Niv, Y.; Ballard, I. Holistic reinforcement learning: the role of structure and attention. Trends in cognitive sciences 2019, 23, 278–292. [Google Scholar] [CrossRef] [PubMed]
  55. Farashahi, S.; Soltani, A. Computational mechanisms of distributed value representations and mixed learning strategies. Nature Communications 2021, 12, 7191. [Google Scholar] [CrossRef]
  56. Wang, M.C.; Soltani, A. Contributions of Attention to Learning in Multidimensional Reward Environments 2025. Publisher: Society for Neuroscience Section: Research Articles. [CrossRef]
  57. Wise, T.; Emery, K.; Radulescu, A. Naturalistic reinforcement learning. Trends in Cognitive Sciences 2024, 28, 144–158. [Google Scholar] [CrossRef]
  58. Yazdanpanah, A.; Wang, M.C.; Trepka, E.; Benz, M.; Soltani, A. Contributions of statistical learning to learning from reward feedback. bioRxiv 2024. Publisher: Cold Spring Harbor Laboratory _eprint: https://www.biorxiv.org/content/early/2024/07/26/2024.04.27.591445.full.pdf. [CrossRef]
  59. Izquierdo, A.; Brigman, J.L.; Radke, A.K.; Rudebeck, P.H.; Holmes, A. The neural basis of reversal learning: An updated perspective. Neuroscience 2017, 345, 12–26. [Google Scholar] [CrossRef]
  60. Costa, V.D.; Dal Monte, O.; Lucas, D.R.; Murray, E.A.; Averbeck, B.B. Amygdala and ventral striatum make distinct contributions to reinforcement learning. Neuron 2016, 92, 505–517. [Google Scholar] [CrossRef]
  61. Bari, B.A.; Grossman, C.D.; Lubin, E.E.; Rajagopalan, A.E.; Cressy, J.I.; Cohen, J.Y. Stable representations of decision variables for flexible behavior. Neuron 2019, 103, 922–933. [Google Scholar] [CrossRef]
  62. Hamid, A.A.; Pettibone, J.R.; Mabrouk, O.S.; Hetrick, V.L.; Schmidt, R.; Vander Weele, C.M.; Kennedy, R.T.; Aragona, B.J.; Berke, J.D. Mesolimbic dopamine signals the value of work. Nature Neuroscience 2016, 19, 117–126. [Google Scholar] [CrossRef]
  63. Aguirre, C.G.; Woo, J.H.; Romero-Sosa, J.L.; Rivera, Z.M.; Tejada, A.N.; Munier, J.J.; Perez, J.; Goldfarb, M.; Das, K.; Gomez, M.; et al. Dissociable Contributions of Basolateral Amygdala and Ventrolateral Orbitofrontal Cortex to Flexible Learning Under Uncertainty 2024. Publisher: Society for Neuroscience Section: Research Articles. [CrossRef]
  64. Rutledge, R.B.; Lazzaro, S.C.; Lau, B.; Myers, C.E.; Gluck, M.A.; Glimcher, P.W. Dopaminergic Drugs Modulate Learning Rates and Perseveration in Parkinson’s Patients in a Dynamic Foraging Task. J. Neurosci. 2009, 29, 15104–15114. [Google Scholar] [CrossRef]
  65. Grunwald, P.; Vitanyi, P. Shannon Information and Kolmogorov Complexity. 2004, arXiv:cs/0410002. [Google Scholar] [CrossRef]
  66. Lizier, J.T. Measuring the Dynamics of Information Processing on a Local Scale in Time and Space. In Directed Information Measures in Neuroscience; Wibral, M., Vicente, R., Lizier, J.T., Eds.; Springer Berlin Heidelberg: Berlin, Heidelberg, 2014; pp. 161–193. [Google Scholar] [CrossRef]
  67. Yao, Y. Information-theoretic measures for knowledge discovery and data mining. Entropy measures, maximum entropy principle and emerging applications 2003, pp. 115–136.
  68. Kvålseth, T.O. On normalized mutual information: measure derivations and properties. Entropy 2017, 19, 631. [Google Scholar] [CrossRef]
  69. Amelio, A.; Pizzuti, C. Correction for Closeness: Adjusting Normalized Mutual Information Measure for Clustering Comparison. Computational Intelligence 2017, 33, 579–601. [Google Scholar] [CrossRef]
  70. Todorov, E. Efficient computation of optimal actions. Proceedings of the national academy of sciences 2009, 106, 11478–11483. [Google Scholar] [CrossRef]
  71. Ramírez-Ruiz, J.; Grytskyy, D.; Mastrogiuseppe, C.; Habib, Y.; Moreno-Bote, R. Complex behavior from intrinsic motivation to occupy future action-state path space. Nature Communications 2024, 15, 6368. [Google Scholar] [CrossRef]
  72. Donahue, C.H.; Lee, D. Dynamic routing of task-relevant signals for decision making in dorsolateral prefrontal cortex. Nature neuroscience 2015, 18, 295–301. [Google Scholar] [CrossRef]
  73. Soltani, A.; Wang, X.J. A biophysically based neural model of matching law behavior: melioration by stochastic synapses. Journal of Neuroscience 2006, 26, 3731–3744. [Google Scholar] [CrossRef]
  74. Soltani, A.; Wang, X.J. Synaptic Computation Underlying Probabilistic Inference. Nat Neurosci 2010, 13, 112–119. [Google Scholar] [CrossRef]
  75. Farashahi, S.; Donahue, C.H.; Hayden, B.Y.; Lee, D.; Soltani, A. Flexible combination of reward information across primates. Nature human behaviour 2019, 3, 1215–1224. [Google Scholar] [CrossRef]
  76. Cazé, R.D.; van der Meer, M.A. Adaptive properties of differential learning rates for positive and negative outcomes. Biological cybernetics 2013, 107, 711–719. [Google Scholar] [CrossRef]
  77. Iigaya, K. Adaptive learning and decision-making under uncertainty by metaplastic synapses guided by a surprise detection system. eLife 2016, 5, e18073. [Google Scholar] [CrossRef] [PubMed]
  78. Nassar, M.R.; Frank, M.J. Taming the beast: extracting generalizable knowledge from computational models of cognition. Current Opinion in Behavioral Sciences 2016, 11, 49–54. [Google Scholar] [CrossRef] [PubMed]
  79. Palminteri, S.; Wyart, V.; Koechlin, E. The importance of falsification in computational cognitive modeling. Trends in cognitive sciences 2017, 21, 425–433. [Google Scholar] [CrossRef] [PubMed]
  80. Sifar, A.; Srivastava, N. Over-precise Predictions Cannot Identify Good Choice Models. Computational Brain & Behavior 2022, 5, 378–396. [Google Scholar] [CrossRef]
  81. Wilson, R.C.; Collins, A.G. Ten simple rules for the computational modeling of behavioral data. Elife 2019, 8, e49547. [Google Scholar] [CrossRef]
  82. Meyniel, F.; Dehaene, S. Brain networks for confidence weighting and hierarchical inference during probabilistic learning. Proceedings of the National Academy of Sciences 2017, 114, E3859–E3868. [Google Scholar] [CrossRef]
  83. Schreiber, T. Measuring Information Transfer. Phys. Rev. Lett. 2000, 85, 461–464. [Google Scholar] [CrossRef]
  84. Bossomaier, T.; Barnett, L.; Harré, M.; Lizier, J.T. Transfer Entropy. In An Introduction to Transfer Entropy. In An Introduction to Transfer Entropy: Information Flow in Complex Systems; Bossomaier, T., Barnett, L., Harré, M., Lizier, J.T., Eds.; Springer International Publishing: Cham, 2016; pp. 65–95. [Google Scholar] [CrossRef]
  85. Kraskov, A.; Stögbauer, H.; Grassberger, P. Estimating mutual information. Phys. Rev. E 2004, 69, 066138. [Google Scholar] [CrossRef]
  86. Gao, W.; Kannan, S.; Oh, S.; Viswanath, P. Estimating Mutual Information for Discrete-Continuous Mixtures. In Proceedings of the Advances in Neural Information Processing Systems. Curran Associates, Inc., 2017, Vol. 30.
  87. Ross, B.C. Mutual Information between Discrete and Continuous Data Sets. PLOS ONE 2014, 9, e87357. [Google Scholar] [CrossRef]
  88. Williams, P.L.; Beer, R.D. Nonnegative Decomposition of Multivariate Information. 2010, arXiv:1004.2515. [Google Scholar] [CrossRef]
  89. Luppi, A.I.; Rosas, F.E.; Mediano, P.A.M.; Menon, D.K.; Stamatakis, E.A. Information decomposition and the informational architecture of the brain. Trends in Cognitive Sciences 2024, 28, 352–368. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Probabilistic reversal learning task and information-theoretic framework for describing agent-environment interaction. (a) Illustration of agent-environment interaction and an example progression of task-related information used for computing behavioral information-theoretic metrics. (b) Left: illustration of conditional entropy and mutual information between previous reward ( R t 1 ) and current strategy ( S t r t ). Gray area represents conditional entropy of reward dependent strategy (ERDS = H ( S t r t R t 1 ) ), and black area represents mutual information between strategy and previous reward (MIRS). n-MIRS denotes mutual information normalized by H ( S t r ) . Right: similar metric defined on the joint combination of previous reward (R) and chosen option ( O p t ). Gray area represents conditional entropy of reward and option-dependent strategy (ERODS = H ( S t r R , O p t ) ), and black area represents mutual information between reward outcome, choice options, and choice strategy (MIROS = I ( R , O p t ; S t r ) ).
Figure 1. Probabilistic reversal learning task and information-theoretic framework for describing agent-environment interaction. (a) Illustration of agent-environment interaction and an example progression of task-related information used for computing behavioral information-theoretic metrics. (b) Left: illustration of conditional entropy and mutual information between previous reward ( R t 1 ) and current strategy ( S t r t ). Gray area represents conditional entropy of reward dependent strategy (ERDS = H ( S t r t R t 1 ) ), and black area represents mutual information between strategy and previous reward (MIRS). n-MIRS denotes mutual information normalized by H ( S t r ) . Right: similar metric defined on the joint combination of previous reward (R) and chosen option ( O p t ). Gray area represents conditional entropy of reward and option-dependent strategy (ERODS = H ( S t r R , O p t ) ), and black area represents mutual information between reward outcome, choice options, and choice strategy (MIROS = I ( R , O p t ; S t r ) ).
Preprints 174449 g001
Figure 2. Schematic of an RL agent performing a probabilistic reversal learning task, with behavioral quantification using information-theoretic metrics. (a) Schematic of a standard RL algorithm that makes choices based on the value difference between the two options ( Q A Q B ), using a softmax function to compute choice probabilities ( P ( A ) ). After each choice, the model updates the value estimate of the chosen option, Q c , based on the difference between the reward outcome (R) and Q c . This difference is known as the reward prediction error ( R P E ). (b) Plot shows average performance over time, defined as probability of choosing the better- rewarding option, shown separately for three different learning rates: α = 0.2 , 0.4 , 0.6 . Inset shows the average performance over 100 simulated blocks. ’n.s.’ denotes no significant difference. (c–f) Plots show the averaged time course of the entropy of strategy ( H ( S t r ) ; c), normalized mutual information between strategy and previous reward (n-MIRS; d), normalized mutual information between strategy and option (n-MIOS; e), and normalized mutual information between strategy and the previous reward/option combinations (n-MIROS; f). Asterisks indicate significant differences based on a two-sided rank sum test (*: p < .05; **: p < .01; ***: p < .001).
Figure 2. Schematic of an RL agent performing a probabilistic reversal learning task, with behavioral quantification using information-theoretic metrics. (a) Schematic of a standard RL algorithm that makes choices based on the value difference between the two options ( Q A Q B ), using a softmax function to compute choice probabilities ( P ( A ) ). After each choice, the model updates the value estimate of the chosen option, Q c , based on the difference between the reward outcome (R) and Q c . This difference is known as the reward prediction error ( R P E ). (b) Plot shows average performance over time, defined as probability of choosing the better- rewarding option, shown separately for three different learning rates: α = 0.2 , 0.4 , 0.6 . Inset shows the average performance over 100 simulated blocks. ’n.s.’ denotes no significant difference. (c–f) Plots show the averaged time course of the entropy of strategy ( H ( S t r ) ; c), normalized mutual information between strategy and previous reward (n-MIRS; d), normalized mutual information between strategy and option (n-MIOS; e), and normalized mutual information between strategy and the previous reward/option combinations (n-MIROS; f). Asterisks indicate significant differences based on a two-sided rank sum test (*: p < .05; **: p < .01; ***: p < .001).
Preprints 174449 g002
Figure 3. Behavioral signatures of positivity bias and its decoding using information-theoretic metrics. (a) Illustration of value updates following rewarded ( R = 1 ) and unrewarded ( R = 0 ) outcomes, with corresponding differential learning rates α + and α , respectively. (b) Behavioral signature of positivity bias. Heatmaps show the average performance for each combination of ( α + , α ) in two example reward schedules. (c) Correlations between performance and the difference in the learning rates ( Δ α = α + α ), shown separately for two reward schedules. (d, e) Example time courses of ERDS (d) and n-MIRS (e) for five example RL agents with different learning rates. Red/magenta lines indicate "optimistic" agents, blue/cyan lines indicates "pessimistic" agents, and black corresponds to the neutral agent. Insets show metric computed over the whole block. (f) Correlation between Δ α and ERDS metric in the 80/20% environment. (g) Example time courses for the decomposition of n-MIRS after negative reward feedback (n-MIRS-). The sign of n-MIRS-, computed from the first 10 trials (inset), is predictive of positivity bias. (h) Correlations between Δ α and n-MIRS- metric, shown separately for three example reward environments. (i) Illustration of sampling and decoding procedure to predict the sign of Δ α using information-theoretic metrics. For each RL agent in the two groups—“optimistic” with Δ α > 0 or “pessimistic” with Δ α < 0 )—a mean value of ( α + , α ) was randomly drawn. Learning rates for each session were then randomly sampled from Gaussian distributions centered on these means. Behavioral metrics were computed from the resulting choice behavior in each session and used to train a binary decoder. The inset on the right shows the predicted posterior probability generated by cross-validated decoders for each true value of Δ α , with a sigmoid curve fitted to the data. (j) Cross-validated decoding accuracy from decoders trained on the information-theoretic metrics—n-MIRS- metric only (gray) or all available metrics (black)—and logistic regression weights (white empty bars), separately for different reward environments.
Figure 3. Behavioral signatures of positivity bias and its decoding using information-theoretic metrics. (a) Illustration of value updates following rewarded ( R = 1 ) and unrewarded ( R = 0 ) outcomes, with corresponding differential learning rates α + and α , respectively. (b) Behavioral signature of positivity bias. Heatmaps show the average performance for each combination of ( α + , α ) in two example reward schedules. (c) Correlations between performance and the difference in the learning rates ( Δ α = α + α ), shown separately for two reward schedules. (d, e) Example time courses of ERDS (d) and n-MIRS (e) for five example RL agents with different learning rates. Red/magenta lines indicate "optimistic" agents, blue/cyan lines indicates "pessimistic" agents, and black corresponds to the neutral agent. Insets show metric computed over the whole block. (f) Correlation between Δ α and ERDS metric in the 80/20% environment. (g) Example time courses for the decomposition of n-MIRS after negative reward feedback (n-MIRS-). The sign of n-MIRS-, computed from the first 10 trials (inset), is predictive of positivity bias. (h) Correlations between Δ α and n-MIRS- metric, shown separately for three example reward environments. (i) Illustration of sampling and decoding procedure to predict the sign of Δ α using information-theoretic metrics. For each RL agent in the two groups—“optimistic” with Δ α > 0 or “pessimistic” with Δ α < 0 )—a mean value of ( α + , α ) was randomly drawn. Learning rates for each session were then randomly sampled from Gaussian distributions centered on these means. Behavioral metrics were computed from the resulting choice behavior in each session and used to train a binary decoder. The inset on the right shows the predicted posterior probability generated by cross-validated decoders for each true value of Δ α , with a sigmoid curve fitted to the data. (j) Cross-validated decoding accuracy from decoders trained on the information-theoretic metrics—n-MIRS- metric only (gray) or all available metrics (black)—and logistic regression weights (white empty bars), separately for different reward environments.
Preprints 174449 g003
Figure 4. Comparison of plastic and metaplastic models and their distinct behavior revealed by information-theoretic metrics. (a) Schematic of a biological implementation of a standard RL model with binary synapses that undergo stochastic reward-dependent plasticity. In this model, "weak" synapses transition to the "strong" synaptic state (i.e., are potentiated) with probability α + following reward, whereas "strong" synapses transition to the "weak" state (i.e., are depressed) in the absence of reward. (b) Schematic of synapses with reward-dependent metaplasticity, consisting of three meta-states with increasing levels of stability ( i = 1 , 2 , 3 ), each associated with one of the two levels of synaptic strengths ("weak" w or "strong" s). Transition probabilities q i govern changes from the i t h meta-state to the most unstable meta-state of opposite efficacy, with q 1 > q 2 > q 3 . Transition probabilities p i govern transitions between meta-states of the same efficacy, and are also higher for less stable states (i.e., p 1 > p 2 ). (c) Averaged trajectories of “effective learning rates” for the metaplastic model (solid lines), a plastic model with a single learning rate (dotted lines; α = 0.3 ), a the plastic model with differential learning rates estimated from the choice behavior of the metaplastic model (dashed lines). Red and blue correspond to the effective learning rates on rewarded and unrewarded trials, respectively. (d) Averaged trajectory of H ( S t r ) , shown separately for the three models as indicated in the legend. (e) Averaged trajectory of normalized mutual information between previously chosen option and strategy (n-MIOS, solid lines), separately for each of the three model. MIOS before normalization by H ( S t r ) are plotted for reference (dotted lines). (f) Posterior probability that a given set of metrics was classified as originating from an "optimistic" (opti) vs. a "pessimistic" (pess) agent, based on binary decoders applied to the choice behavior of the three models.
Figure 4. Comparison of plastic and metaplastic models and their distinct behavior revealed by information-theoretic metrics. (a) Schematic of a biological implementation of a standard RL model with binary synapses that undergo stochastic reward-dependent plasticity. In this model, "weak" synapses transition to the "strong" synaptic state (i.e., are potentiated) with probability α + following reward, whereas "strong" synapses transition to the "weak" state (i.e., are depressed) in the absence of reward. (b) Schematic of synapses with reward-dependent metaplasticity, consisting of three meta-states with increasing levels of stability ( i = 1 , 2 , 3 ), each associated with one of the two levels of synaptic strengths ("weak" w or "strong" s). Transition probabilities q i govern changes from the i t h meta-state to the most unstable meta-state of opposite efficacy, with q 1 > q 2 > q 3 . Transition probabilities p i govern transitions between meta-states of the same efficacy, and are also higher for less stable states (i.e., p 1 > p 2 ). (c) Averaged trajectories of “effective learning rates” for the metaplastic model (solid lines), a plastic model with a single learning rate (dotted lines; α = 0.3 ), a the plastic model with differential learning rates estimated from the choice behavior of the metaplastic model (dashed lines). Red and blue correspond to the effective learning rates on rewarded and unrewarded trials, respectively. (d) Averaged trajectory of H ( S t r ) , shown separately for the three models as indicated in the legend. (e) Averaged trajectory of normalized mutual information between previously chosen option and strategy (n-MIOS, solid lines), separately for each of the three model. MIOS before normalization by H ( S t r ) are plotted for reference (dotted lines). (f) Posterior probability that a given set of metrics was classified as originating from an "optimistic" (opti) vs. a "pessimistic" (pess) agent, based on binary decoders applied to the choice behavior of the three models.
Preprints 174449 g004
Figure 5. Detecting the modulation of choice strategy by reward harvest rate. (a) Schematic of an RL model that tracks reward harvest rate to modulate choice strategy. In this model, the difference between the immediate reward (R) and reward harvest rate (H) is used to directly influence win-stay and lose-switch behavior. (b) Averaged time course of the model estimated values ( Q i ’s in green and orange) and reward harvest rate (H in black) during the PRL task with a reversal on trial 40. The gray trace in the background depicts the time course of H from an example session. Filled (empty) circles represent trials with high (low) reward harvest states ( R H S ), determined via a median split of H values across the session. Simulation parameters: α = 0.3 , β = 10 , β H = 2 . (c) Averaged time course of normalized mutual information between strategy and reward/RHS combinations (n-MIRHS), shown for models with and without the reward harvest rate modulation. (d) Decomposition of the n-MIRHS metric for trials in the high RHS state. Inset shows further decompositions into previously rewarded (high+, red) and unrewarded (high-, blue) trials. (e) Decomposition of the n-MIRHS metric for trials in the low RHS state. Inset shows further decompositions into previously rewarded (low+, red) and unrewarded (low-, blue) trials. (f) Plotted are the mean simulated n-MIRHS decompositions based on high (left panel) or low (right panel) reward state, following rewarded (solid) or unrewarded (dashed lines) trials, as a function of α (X-axis) and β H (indicated by colors). The influence of β H was most pronounced in the information gained from no reward in the low RHS.
Figure 5. Detecting the modulation of choice strategy by reward harvest rate. (a) Schematic of an RL model that tracks reward harvest rate to modulate choice strategy. In this model, the difference between the immediate reward (R) and reward harvest rate (H) is used to directly influence win-stay and lose-switch behavior. (b) Averaged time course of the model estimated values ( Q i ’s in green and orange) and reward harvest rate (H in black) during the PRL task with a reversal on trial 40. The gray trace in the background depicts the time course of H from an example session. Filled (empty) circles represent trials with high (low) reward harvest states ( R H S ), determined via a median split of H values across the session. Simulation parameters: α = 0.3 , β = 10 , β H = 2 . (c) Averaged time course of normalized mutual information between strategy and reward/RHS combinations (n-MIRHS), shown for models with and without the reward harvest rate modulation. (d) Decomposition of the n-MIRHS metric for trials in the high RHS state. Inset shows further decompositions into previously rewarded (high+, red) and unrewarded (high-, blue) trials. (e) Decomposition of the n-MIRHS metric for trials in the low RHS state. Inset shows further decompositions into previously rewarded (low+, red) and unrewarded (low-, blue) trials. (f) Plotted are the mean simulated n-MIRHS decompositions based on high (left panel) or low (right panel) reward state, following rewarded (solid) or unrewarded (dashed lines) trials, as a function of α (X-axis) and β H (indicated by colors). The influence of β H was most pronounced in the information gained from no reward in the low RHS.
Preprints 174449 g005
Figure 6. Schematic of a multidimensional reward learning task and RL models used to perform it. (a) Illustration of example trials from the multidimensional reversal learning task. Reversals can occur either in the reward probabilities associated with two options (e.g., switching from green to orange as the better option), as in the standard probabilistic reversal learning task, or in the feature dimension predictive of rewards (e.g., from color to shape). (b) Schematic of the dynamic arbitration model that simultaneously tracks the values of individual color and shape features (Q). These feature values are then combined to compute the overall value (V) of the left and right options. The arbitration weight ( ω ) is dynamically adjusted based on the reliability of the chosen color and shape values, estimated from their respective reward prediction errors. (c) Illustration of a reversal where the better option reverses from the green to the orange object (Color-to-Color reversal), regardless of shape. (d) Illustration of a reversal where the better option reverses from the green to triangle object (Color-to-Shape, feature reversal). (e) Averaged time courses of arbitration weight ( ω , top) and task performance (bottom) during Color-to-Color reversal blocks for the three models. The dynamic arbitration model (Dynamic arb. gradually increases its weighting of color. The model with no dynamic arbitration (No arb. maintains a fixed weight of ω = 0.5 , equally combining color and shape values. The Color-only model does not track shape values and therefore, ω remains fixed at 1. (f) Same plots as in (e) but for Color-to-Shape (feature reversal) blocks.
Figure 6. Schematic of a multidimensional reward learning task and RL models used to perform it. (a) Illustration of example trials from the multidimensional reversal learning task. Reversals can occur either in the reward probabilities associated with two options (e.g., switching from green to orange as the better option), as in the standard probabilistic reversal learning task, or in the feature dimension predictive of rewards (e.g., from color to shape). (b) Schematic of the dynamic arbitration model that simultaneously tracks the values of individual color and shape features (Q). These feature values are then combined to compute the overall value (V) of the left and right options. The arbitration weight ( ω ) is dynamically adjusted based on the reliability of the chosen color and shape values, estimated from their respective reward prediction errors. (c) Illustration of a reversal where the better option reverses from the green to the orange object (Color-to-Color reversal), regardless of shape. (d) Illustration of a reversal where the better option reverses from the green to triangle object (Color-to-Shape, feature reversal). (e) Averaged time courses of arbitration weight ( ω , top) and task performance (bottom) during Color-to-Color reversal blocks for the three models. The dynamic arbitration model (Dynamic arb. gradually increases its weighting of color. The model with no dynamic arbitration (No arb. maintains a fixed weight of ω = 0.5 , equally combining color and shape values. The Color-only model does not track shape values and therefore, ω remains fixed at 1. (f) Same plots as in (e) but for Color-to-Shape (feature reversal) blocks.
Preprints 174449 g006
Figure 7. Distinct predictions of information-theoretic metrics for adopting and arbitrating between alternative strategies in a multidimensional reward environment. (a) Time courses of the conditional entropy of reward-dependent strategy based on color ( ERDS C o l o r = H ( S t r C o l o r | R ) , solid lines) and shape ( ERDS S h a p e = H ( S t r S h a p e | R ) , dashed lines) during the Color-to-Color blocks. Line colors indicate the three simulated agents (gray: Color-only; black: No Arbitration; magenta: Dynamic Arbitration). (b) Similar to (a) but for the Color-to-Shape (feature reversal) blocks. (c) Time course of the difference ERDS S h a p e ERDS C o l o r during the Color-to-Color blocks, measuring the relative dominance of two learning strategies. Higher values indicate dominance of the color-based strategy. (d) Similar to (c) but for the Color-to-Shape (feature reversal) blocks. (e–f) Correlation coefficients between ERDS C o l o r and ERDS S h a p e during the Color-to-Color (e) and the Color-to-Shape (feature reversal) blocks (f), computed separately from the “non-stationary” (first 30 trials of each block) or “steady” portion of the block (last 10 trials before the reversal or end of the block). Asterisks indicate significance level (*: p < . 05 ; **: p < . 01 ; ***: p < . 001 ; n . s . : not significant). The Color-only agent showed no interaction between the two strategies, with no significant correlations observed between ERDS C o l o r and ERDS S h a p e in any block or phase. The No Arbitration agent (with fixed ω = 0.5) exhibited strong negative interaction between the two strategies, as reflected by significant negative correlations during both steady and non-stationary portions of the blocks. The Dynamic Arbitration agent showed significant negative interaction during the non-stationary phase, but this correlation became non-significant during the steady phase as one strategy began to dominate.
Figure 7. Distinct predictions of information-theoretic metrics for adopting and arbitrating between alternative strategies in a multidimensional reward environment. (a) Time courses of the conditional entropy of reward-dependent strategy based on color ( ERDS C o l o r = H ( S t r C o l o r | R ) , solid lines) and shape ( ERDS S h a p e = H ( S t r S h a p e | R ) , dashed lines) during the Color-to-Color blocks. Line colors indicate the three simulated agents (gray: Color-only; black: No Arbitration; magenta: Dynamic Arbitration). (b) Similar to (a) but for the Color-to-Shape (feature reversal) blocks. (c) Time course of the difference ERDS S h a p e ERDS C o l o r during the Color-to-Color blocks, measuring the relative dominance of two learning strategies. Higher values indicate dominance of the color-based strategy. (d) Similar to (c) but for the Color-to-Shape (feature reversal) blocks. (e–f) Correlation coefficients between ERDS C o l o r and ERDS S h a p e during the Color-to-Color (e) and the Color-to-Shape (feature reversal) blocks (f), computed separately from the “non-stationary” (first 30 trials of each block) or “steady” portion of the block (last 10 trials before the reversal or end of the block). Asterisks indicate significance level (*: p < . 05 ; **: p < . 01 ; ***: p < . 001 ; n . s . : not significant). The Color-only agent showed no interaction between the two strategies, with no significant correlations observed between ERDS C o l o r and ERDS S h a p e in any block or phase. The No Arbitration agent (with fixed ω = 0.5) exhibited strong negative interaction between the two strategies, as reflected by significant negative correlations during both steady and non-stationary portions of the blocks. The Dynamic Arbitration agent showed significant negative interaction during the non-stationary phase, but this correlation became non-significant during the steady phase as one strategy began to dominate.
Preprints 174449 g007
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated