Abstract
Predicting an individual’s behaviour in one task based on their behaviour in a different task is a key challenge in modeling individual decision-making tendencies. We propose a novel framework that addresses this challenge by leveraging neural networks and introducing a concept we term the “individuality index.” This index, extracted from behaviour in a “source” task via an encoder network, captures an individual’s unique decision-making tendencies. A decoder network then utilizes this index to generate the weights of a task-specific neural network (a “task solver”), which predicts the individual’s behaviour in a “target” task. We demonstrate the effectiveness of our approach in two distinct decision-making tasks: a value-guided task and a perceptual task. Our framework offers a robust and generalizable approach for parameterizing individuality, providing a promising pathway toward computational modeling at the individual level—that is, replicating individuals in silico.
1 Introduction
Humans (and other animals) exhibit substantial commonalities in their decision-making processes. However, considerable variability is also frequently observed in how individuals perform perceptual and cognitive decision-making tasks [5, 1]. This variability arises from differences in underlying cognitive mechanisms. For example, individuals may vary in their ability or tendency to retain past experiences [12, 8], respond to events with both speed and accuracy [52, 45], or explore novel actions [17]. If these factors can be meaningfully disentangled, they would enable a concise characterization of individual decision-making processes, yielding a low-dimensional, parameterized representation of individuality. Such a representation could, in turn, be leveraged to predict future behaviours at an individual level. Shifting from population-level predictions to an individual-based approach would mark a significant advancement in domains where precise behaviour prediction is essential, such as social and cognitive sciences. Beyond prediction, this approach offers a framework for parametrizing and clustering individuals, thereby facilitating the visualization of behavioural heterogeneity, which has applications in psychiatric analysis [31, 10]. Furthermore, this parameterization offers a promising pathway toward computational modeling at the individual level—that is, replicating the cognitive and functional characteristics of individuals in silico [41].
Cognitive modelling is a standard approach for reproducing and predicting human behaviour [29, 3, 57], often implemented within a reinforcement learning framework (e.g., [30, 9, 54]). However, because these cognitive models are manually designed by researchers, their ability to accurately fit behavioural data may be limited [16, 44, 28, 13]. A data-driven approach using artificial neural networks (ANNs) offers an alternative [11, 33, 39]. Unlike cognitive models, which rely on predefined behavioural assumptions [37], ANNs require minimal prior assumptions and can learn complex patterns directly from data. For instance, convolutional neural networks (CNNs) have successfully replicated human choices and reaction times in various visual tasks [25, 35, 15]. Similarly, recurrent neural networks (RNNs) [42, 7] have been applied to model value-guided decision-making tasks such as the multi-armed bandit problem [56, 10]. A promising approach to capturing individual decision-making tendencies while preserving behavioural consistency is to tune ANN weights using a parameterized representation of individuality.
This idea was first proposed by Dezfouli et al. [10], who employed an RNN to solve a two-armed bandit task. Their study utilized an autoencoder framework [38, 48], in which behavioural recordings from a single session of the bandit task, performed by an individual, were fed into an encoder. The encoder produced a low-dimensional vector, interpreted as a latent representation of the individual. Similar to hypernetworks [20, 22], a decoder then took this low-dimensional vector as input and generates the weights of the RNN. This framework successfully reproduced behavioural recordings from other sessions of the same bandit task while preserving individual characteristics. However, since this individuality transfer has only been validated within the bandit task, it remains unclear whether the extracted latent representation captures an individual’s intrinsic tendencies across a variety of tasks.
To address this question, we aim to make the low-dimensional representation— referred as individuality index—robust to variations across individuals and tasks, thereby enhancing its generalizability. Specifically, we propose a framework that predicts an individual’s behaviours, not only in the same task but also in similar yet distinct tasks and environments. If the individuality index indeed serves as a low-dimensional representation of an individual’s decision-making process, then extracting it from one task could facilitate the prediction of that individual’s behaviours in another.
In this study, we define the problem of individuality transfer across tasks as follows (also illustrated in Figure 1). We assume access to a behavioural dataset from multiple individuals performing two tasks: a source task and a target task. We train an encoder that takes behavioural data from the source task as input and outputs an individuality index. This index is then fed into a decoder, which generates the weights of an ANN, referred to as a task solver, that reproduces behaviours in the target task. For testing, a new individual provides behavioural data from the source task, allowing us to infer their individuality index. Using this index, a task solver is constructed to predict how the test individual will behave in the target task. Importantly, this prediction does not require any behavioural data from the test individual performing the target task. We refer to this framework as EIDT, an acronym for encoder, individuality index, decoder and task solver.

The EIDT (encoder, individuality index, decoder, and task solver) framework for individuality transfer across tasks.
The encoder maps action(s) α, provided by an individual K performing a specific problem ϕ in the source task A, into an individuality index (represented as a point in the two-dimensional space in the center). The individual index is then fed into the decoder, which generates the weights for a task solver. The task solver predicts the behaviour of the same individual K in the target task B. During the training, a loss function evaluates the discrepancy between the predicted behaviour
We evaluated whether the proposed EIDT framework can effectively transfer individuality in both value-guided sequential decision-making tasks and perceptual decision-making tasks. To assess its generalizability across individuals, meaning its ability to predict the behaviour of previously unseen individuals, we tested the framework using a test participant pool that was not included in the dataset used for model training. To determine how well our framework captures each individual’s unique behavioural patterns, we compared the prediction performance of a task solver specifically designed for a given individual with the performance of task solvers designed for other individuals. Our results indicate that the proposed framework successfully mimics decision-making while accounting for individual differences.
2 Results
We evaluated our individuality transfer framework, which consists of an encoder, a decoder, a task solver, and an individuality index, using two transfer problems: a value-guided sequential decision-making task (an MDP task) and a perceptual decision-making task (a handwritten digit recognition task, referred to as the MNIST task). The dataset for both tasks include two or more distinct experimental settings. Since each human participant completed all settings, we tested transfer problems in which one setting served as the source task, while another served as the target task.
2.1 Markov decision process (MDP) task
The dataset consisted of behavioural data from 98 participants, as detailed in Section 4.2.2. Participants performed the MDP task under two different conditions: a 2-step task and a 3-step task. Each participant completed three sequence blocks for both the 2- and 3-step tasks, with each sequence comprising 50 episodes. Thus, the dataset contained a total of N = 6 (sequences) × 98 (participants) = 588 (action sequences). For model training, sequence data from 70% of the participants (68 participants) were used to train the encoder and decoder. An additional 10% (10 participants) were designated as validation data, used to determine the optimal number of training iterations. The remaining 20% (20 participants) were allocated for evaluating the performance of the model.
Neural network exhibits superior performance in behaviour prediction
We evaluated our model under two experimental situations, summarized in Table 1. The first situation, Situation SX (no transfer), does not involve individuality transfer across tasks. Since individuality is not explicitly considered in this situation, we designed average models as baselines, described below.

Simulated experimental situations.
The “Availability” column indicates data availability for a given participant pool and task during model training. In the “Participant pool” column, “train”, “valid”, and “test” refer to the training, validation, and test participant pools, respectively.
Cognitive model (CG)
We implemented a cognitive model based on Q-learning, as detailed in Section 4.2.3. The model parameters—including the learning rate qlr, forgetting factor qff, discount rate qdr, and inverse temperature qit—were estimated using a maximum likelihood approach based on behavioural data from the target task in the training and validation participant pools. Using these estimated parameters, we constructed a Q-learning agent to predict the behaviour of participants in the test participant pool.
Task solver (TS)
We employed a task solver TS(·), as defined in (1). Instead of using the EIDT framework, the weights of the task solver network were directly trained using data from the training participant pool. To determine the optimal stopping point during training, we used early stopping based on the loss measured on the validation participant pool—training was terminated when the validation loss reached its minimum. The trained task solver was then used to predict the behaviour of participants in the test participant pool.
For further details on the model training procedure, refer to Section 4.2.4.
Under Situation SX, we evaluated whether a neural network-based behaviour prediction model (i.e., the task solver) outperforms the cognitive model in predicting average performance across individuals. The result (Figure 2A) confirmed this. A two-way (model: CG/TS, task: 2/3) repeated-measures (RM) ANOVA with Greenhouse-Geisser correction (significant level was 0.05) revealed a significant effect of model (F1,19 = 29.944, p < 0.001,

Prediction error results for the MDP task.
A Under Situation SX. The box represents the interquartile range (IQR), with the central line indicating the median. Whiskers extend to the minimum and maximum values. Connected dots represent data from the same participant. B Under Situation SY. C Comparison between the transfer-to-original participant (Original) and transfer-to-other-participants (Others) settings.
EIDT framework excels in individuality transfer
Under Situation SY (transfer across tasks), we assumed that that behavioural data from the source task of the test participant pool could be leveraged to predict behaviour in the target task. This situation was designed to evaluate individuality transfer across tasks, which is the primary focus of this study. We tested two models:
Cognitive model (CG)
We implemented a Q-learning-based cognitive model, as described in Section 4.2.3. The parameters (qlr, qff, qdr, and qit) for a participant in the test pool were estimated using a maximum likelihood approach based on behavioural data from that participant’s source task. Using these estimated parameters, we constructed a Q-learning agent capable of solving the target task (not the source task) and used it to predict the participant’s behaviour in the target task.
EIDT
We employed our EIDT framework, which was trained using data from both and source and target tasks of the training participant pool. As with the TS model under Situation SX, data from the validation participant pool were used for early stopping during training. To predict behaviour in the target task for a participant K in the test participant pool, we first input each sequence from 𝒜K (behavioural data from participant K in task A) into the encoder. The individuality index for participant K was computed as the average:
For details on the model training procedure, refer to Section 4.2.4.
We compared the prediction error between models, as shown in Figure 2B. A two-way (model: CG/EIDT, transfer task set: 2→3/3→2) RM ANOVA revealed significant effects of model (F1,19 = 21.372, p < 0.001,
Individuality index forms clusters for each individual
We visualized the individuality indices in Figure 3 and evaluated them using the within-individual and between-individual distances, as defined in Section 4.4.1. For both transfer task sets, the within-individual distance was significantly shorter than the between-individual distance (t19 = −5.515, p < 0.001 for 3→2 and t19 = −2.991, p = 0.008 for 2→3). Since the within-individual distance is shorter than the between-individual distance, the individuality indices from the same individual are positioned closer together in the individuality index space, forming a distinct cluster for each individual.

Individuality indices in the test participant pool for the MDP tasks.
Dots represent the average individuality index for each participant, while shaded areas show confidence ellipses, enclosing 98.9% of the points for each participant.
We also evaluated the extent to which the variables of the individuality index are disentangled, using the Pearson correlation coefficient. A test of the null hypothesis of uncorrelation showed that the two variables (z0 and z1) are significantly correlated for the 3→2 transfer (r = 0.687, p < 0.001) and the 2→3 transfer (r = −0.323, p = 0.012).
Individuality is preserved across tasks
We evaluated whether the models could predict unique behavioural patterns for each individual. To do so, we compared prediction accuracy for a given participant’s behaviour using a task solver trained specifically for that participant with its accuracy using task solvers trained for other participants. That is, we examined accuracy for the original task solver (Original) versus accuracy for other task solvers (Others). If the task solver effectively captures individuality, it should perform significantly better for Original than for Others. For details on this comparison, refer to Section 4.4.2.
The cognitive model exhibited this tendency (Figure 2C, left). A two-way (transfer task set: 3→2/2→3, transfer-to participant(s): Original/Others) RM ANOVA revealed a significant effect of transfer task set (F1,19 = 5.383, p = 0.032,
The EIDT model exhibited a similar pattern (Figure 2C, right). A two-way RM ANOVA revealed a significant effect of transfer-to participant(s) (F1,19 = 26.976, p < 0.001,
2.2 Handwritten digit recognition (MNIST) task
The dataset was originally collected and published by Rafiei et al. [34]. It contains behavioural data from 60 human participants who performed a digit discrimination task. The experiment followed a 2 × 2 factorial design with two factors: task difficulty (easy versus difficult images, controlled by noise level) and speed pressure (accuracy versus speed focus, controlled by experimental instruction). This design resulted in four experimental settings; EA (easy, accuracy focus), ES (easy, speed focus), DA (difficult, accuracy focus), and DS (difficult, speed focus). Our individuality transfer experiments evaluated whether the proposed framework could transfer individuality from one setting to another. Each setting included 120 unique images, and each participant made a decision for each image twice, resulting in a total of 960 trials per participant. behavioural data from 70% of participants (42 participants) were used to train the encoder and decoder. Additional 10% of participants (6 participants) were assigned as validation data for early stopping during training. The remaining 20% of participants (12 participants) were used to evaluate model performance.
Our average model outperforms a no-fitting model
Situation SX, a no transfer situation, was used to evaluate how well our task solver fits human behaviour compared to an existing model. Ideally, this comparison would involve models that can be explicitly fitted to human behaviour. However, no readily adaptable model was available. As an alternative, we used an average model, RTNet [34], which predicts human-like decisions and reaction times.
RTNet (RN)
RTNet accumulates evidence generated by a CNN (see Section 4.3.2) to produce stochastic, human-like decisions and response times.
Task solver (TS)
The task solver TS(), as defined in Section 4.3, was trained using data from the training participant pool.
Our results confirmed that the task solver with fitting outperforms RT-Net without fitting, as shown in Figure 4. A two-way (model: RN/TS, task: EA/ES/DA/DS) RM ANOVA revealed significant effects of model (F3,33 = 104.744, p < 0.001,

The prediction error under Situation SX (no transfer) in the MNIST task.
Individuality index forms clusters for each individual
We visualized the individuality indices and evaluated them using the within-individual and between-individual distances (see Section S1, Supplementary materials). For most transfer task sets, the within-individual distance was significantly shorter than the between-individual distance, indicating that individuality indices from the same participant tended to cluster together. A test of the null hypothesis of no correlation revealed that the two variables of the individuality index (z0 and z1) were significantly correlated for all transfer task sets.
Individuality preserves across tasks
Situation SY (transfer across tasks) was used to evaluate whether our EIDT model could predict the unique behaviour of each individual in an unexperienced (target) task. As in Section 2.1, we compared the prediction accuracy in the target task for a given participant using a task solver trained for that participant (Original) with its accuracy using task solvers trained for other participants (Others).
As shown in Figure 5A, the percentage of correct responses was approximately 78% for the easy difficulty level and 60% for the difficult level, aligning with the accuracy levels of human participants and RTNets [34] A two-way (transfer task set: 12 sets (see x-axis in Figure 5), transfer-to participant(s): Original/Others) RM ANOVA revealed a significant effect of transfer task set (F1,121 = 402.851, p < 0.001,

Results for the transfer-to-original-participant (Original) and transfer-to-other-participants (Others) settings in the MNIST task.
A Percentage of correct responses. B Prediction error (likelihood). C Percentage of matches to actual behaviour.
A comparison using individual behavioural data revealed differences between Original and Others. Figure 5B shows the prediction accuracy for each task transfer set. A two-way RM ANOVA revealed significant effects of transfer task set (F1,121 = 84.899, p < 0.001,
Figure 6 visualizes the percentage of correct responses for each digit in the MNIST tasks, highlighting behavioural tendencies reproduced by the model. For example, Participant #33 exhibited a lower percentage for digit 3 compared to other digits, which was captured by the model. Participant #47 showed a drastically lower percentage for digit 1, a pattern also reflected in the model’s predictions.

Percentage of correct responses for each stimulus digit in human behaviour and model predictions. Model predictions were conducted under the transfer experiment from Setting ES to Setting EA.
3 Discussion
We proposed an EIDT framework for modeling the unique decision-making process of each individual. This framework enables the transfer of an individuality index from a (source) task to a different (target) task, allowing a task solver predict behaviours in the target task. Several neural network techniques, such as autoencoders [38, 48], hypernetworks [20], and learning-to-learn [53, 43], facilitate this transfer. Our experiments, conducted on both value-guided sequential and perceptual decision-making tasks, demonstrated the potential of the proposed EIDT framework in individuality transfer across tasks.
EIDT framework transfers individuality across tasks
The EIDT framework was originally proposed by Dezfouli et al. [10] for a bandit task. Their model was designed to predict behaviour within a session performed by a given participant, leveraging behavioural data from other sessions for model training. We extended this idea in three key ways. First, we validated that the framework is effective for previously unseen individuals who were not included in model training. Although these individuals provided behavioural data in the source task to identify their individuality indices, their data were not used for model training. Second, we demonstrated that individuality indices can be transferred even when the source and target tasks differ, indicating that decision-making tendencies are transferable across tasks. Third, while the original work focused on value-guided tasks, we validated the framework’s applicability to perceptual decision-making tasks, specifically the MNIST task. These findings establish that the EIDT framework effectively captures individual differences across both tasks and individuals.
Interpreting the individuality index remains challenging
The interpretation of the individuality index remains an open question. Since interpretation often requires task-specific considerations [13], it falls outside the primary scope of this study, whose aim is to develop a general framework for individuality transfer. Previous research [28, 18] has explored associating neural network parameters with cognitive or functional meanings. Approaches such as disentangling techniques [2] and cognitive model integration [19, 49, 44, 14] could aid in better understanding the cognitive and functional significance of the individuality index.
Regarding the individuality index, while disentanglement and separation losses [10] during the training of model training could enhance interpretability, we used only the reproduction loss, as defined in (5). Our results showed that the variables of the individuality index were correlated, suggesting that they do not independently influence behaviour. Although this dependency complicates the interpretation of the individuality index, we decided not to introduce explicit disentanglement constraints because interpretable parameters in cognitive models (e.g., [9]) are not necessarily independent (e.g., an individual with a high learning rate may also have a high inverse temperature [27], resulting these two parameters is represented with one variable).
Why can the encoder extract individuality for unseen individuals?
Our experiments, which divided participants into training and test participant pools, demonstrated that the framework successfully extracts individuality for completely new individuals. This generalization likely relies on the fact that individuality indices form clusters and individuals similar to new participants exist in the training participant pool [57]. The success of individuality transfer in our study suggests that individuals can be clustered based on behavioural patterns. Behavioural clustering has been widely discussed in relation to psychiatric conditions, medication effects, and gender-based differences (e.g., [31, 50, 40]). Our results could contribute to a deeper discussion of behavioural characteristics by clustering not only these groups but also healthy controls.
Which processes contribute to individuality?
In the MNIST task, we assumed that individuality emerged primarily from the decision-making process (implemented by an RNN [45, 6]), rather than from the visual processing system (implemented by a CNN [55]). The CNN was pretrained, and the decoder did not tune its weights. Our results do not rule out the possibility that the visual system also exhibits individuality [24, 47]; however, they imply that individual differences in perceptual decision-making can be explained primarily by variations in decision-making system [36, 51, 57, 21]. This assumption provides valuable insights for research on human perception.
Limitations
One limitation is that the source and target tasks were relatively similar. Even the cognitive model, despite its lower prediction accuracy, captured individual tendencies to some extent. Thus, our findings do not fully evaluate the generalizability of individuality transfer across diverse task domains. To broaden applicability, future studies should validate transfer across more distinct tasks that share underlying principles of individuality.
The effectiveness of individuality transfer may be influenced by dataset volume. As discussed earlier, prediction performance may depend on whether similar individuals exist in the training participant pool. In our study, 100 participants were sufficient for effective transfer. However, tasks involving greater behavioural diversity may require a substantially larger dataset.
As discussed earlier, the interpretability of the individuality index requires further investigation. Furthermore, the optimal dimensionality of the individuality index remains unclear. This likely depends on the complexity of tasks involved—specifically, the number of factors needed to represent the diversity of behaviour observed in those tasks. While these factors have been explored in cognitive modeling research (e.g., [23, 13]), a clear understanding at the individual level is still lacking. Integrating cognitive modeling with data-driven neural network approaches [10, 19] could help identify key factors underlying individual differences in decision-making.
Future directions
To further generalize our framework, a large-scale dataset is necessary, as discussed in the limitations. This dataset should include a large number of participants to ensure prediction performance for diverse individuals [32]. All participants should perform the same set of tasks, which should include a variety of tasks [56]. Building upon our framework, where the encoder currently accepts action sequences from only a single task, a more generalizable encoder should be able to process behavioural data from multiple tasks to generate a more robust individuality index. To enhance encoder, a multi-head neural network architecture [4] could be utilized. A generalized individuality index would enable transfer to a wider variety of tasks and allow accurate and detailed parameterization of individuals using data from only a single task.
Robust and generalizable parameterization of individuality enables computational modeling at the individual level. This approach, in turn, makes it possible to replicate individuals’ cognitive and functional characteristics in silico [41]. We anticipate that it offers a promising pathway toward a new frontier: artificial intelligence endowed with individuality.
4 Methods
4.1 General framework for individuality transfer across tasks
We formulate the problem of individuality transfer, which involves extracting an individuality index from a source task and predicting behaviour in a target task with preserving individuality. We consider two tasks, A and B, which are different but related. For example, task A might be a 2-step MDP task, while task B is a 3-step MDP task.
The individuality transfer across tasks is defined as follows. An individual K performs a problem within task A, with their behaviour recorded as 𝒜K. Our objective is to predict ℬK, which represents K’s behaviour when performing task B. To achieve this, we extract an individuality index z from 𝒜K, capturing the individual’s behavioural characteristics. This index z is then used to construct a task solver, enabling it to mimic K’s behaviour in task B. Since task A provides data for estimating the individuality index and task B is the target of behaviour prediction, we refer to them as the source task and target task, respectively.
Our proposed framework for the individuality transfer consists of three modules:
Task solver predicts behaviour in the target task B.
Encder extracts the individuality index from the source task A.
Decoder generates the weights of the task solver based on the individuality index.
These modules are illustrated in Figure 1. We refer to this framework as EIDT, an acronym for encoder, individuality index, decoder and task solver.
4.1.1 Data representation
For training, we assume that behaviour data from a participant pool (𝒫K ∉ 𝒫), where each participant has performed both tasks A and B. These dataset are represented as 𝒜 = {𝒜n}n∈P and ℬ = {ℬn}n∈P.
For each individual n, the set 𝒜n consists of one or more sets, each containing a problem instance ϕ (stimuli, task settings, or environment in task A) and a sequence of action(s) α (recorded behavioural responses). For example, in an MDP task, ϕ represents the Markov process (state-action-reward transition) and α consists of choices over multiple trials. In a simple object recognition task, ϕ is a visual stimulus and α is the participant’s response to the stimulus.
Similarly, ℬn consists of a problem instance ψ and an action sequence β.
4.1.2 Task solver
The task solver predicts the action sequence for task B as
where ψ is a specific problem in task B and ΘTS represents the solver’s weights. The task solver architecture is tailored to task B. For example, in an MDP task, the task solver outputs a sequence of actions in response to ψ. In a simple object recognition task, it produces an action based on a visual stimulus ψ.
4.1.3 Encoder
The encoder processes an action sequence(s) α and generates an individuality index z ∈ ℝM as
where ϕ is a problem in task A, ΘENC represents the encoder’s weights, and M is the dimensionality of the individual index. The encoder architecture is task-specific and designed for task A.
4.1.4 Decoder
The decoder receives the individuality index z and generates the task solver’s weights as
where ΘDEC represents the decoder’s weights. Since the decoder determines the task solver’s weights, it functions as a hypernetwork [20, 22].
4.1.5 Training objective
Although tasks A and B differ, an individual’s decision-making system remains consistent across tasks. We model this using the individuality index z, linking it to the task solver via the encoder and decoder. For training, we use behavioural dataset {𝒜n, ℬn }n∈𝒫 from a individual pool 𝒫.
Let α be an action sequence representing individual n’s behaviour on the source task, i.e., (α, ϕ) ∈ 𝒜n, n ∈ 𝒫. The individuality index is derived by z = ENC(α, ϕ; ΘENC). The weights of the task solver are then given by ΘTS = DEC(z; ΘDEC). Subsequently, the task solver, with the given weights, predicts an action sequence for task B as
where β is an action sequence in ℬn recorded along with the problem ψ, and O(·,·) is a suitable loss function (e.g., likelihood-based loss for probabilistic outputs). Using the datasets containing the behaviour of the individual pool 𝒫, the weights of the encoder and decoders, ΘENC and ΘDEC, are optimized by minimizing the total loss:
This section provides a general formulation of individuality transfer across two tasks. For specific details on task architectures and loss functions, see Sections 4.2 and 4.3.
4.2 Experiment on MDP task
We validated our individuality transfer framework using two different decisionmaking tasks: the MDP task and the MNIST task. This section focuses on the MDP tasks, a dynamic multi-step decision-making task.
4.2.1 Task
At the beginning of each episode, an initial state-cue is presented to the participant. For human participants, the state-cue is represented by animal images (Figure 7). For the cognitive model (Q-learning agent) and neural network-based model, the state-cue is represented numerically (e.g., (2, 1) for the first task state in the second choice). The participant makes a binary decision (denoted as action C1 or C2) for each step. In the human experiment, these actions correspond to pressing the left or right cursor key. With a certain probability (either 0.8/0.2 or 0.6/0.4), known as the state-action transition probability, the participant transitions to one of two subsequent task states. This process repeats two times for the 2-step MDP task and three times in the 3-step MDP task. After the final step, the participant receives an outcome: either a reward (r = 1) or no reward (r = 0). For human participants, rewards were displayed as symbols, as shown in Figure 7. Each sequence from initial state-cue presentation to reward delivery constitutes an episode.

The 3-step MDP task.
A. Tree diagram illustrating state-action transitions. B. Flow of a single episode in the behavioural experiment for human participants.
The state-action transition probability T (s, a, s′) from a task state s to a preceding state s′ given an action a varies gradually across episodes. With probability ptrans, one of the transition probabilities switches to a new set chosen from {(0.8, 0.2), (0.2, 0.8), (0.6, 0.4), (0.4, 0.6)}. Consequently, participants must adjust their decision-making strategy in response to these shifts in transition probabilities to maintain reward maximization.
4.2.2 Behavioural data collection
We recruited 123 participants via Prolific. All participants provided their informed consent online. This study was approved by the Committee for Human Research at the Graduate School of Engineering, The University of Osaka, and compiled with the Declaration of Helsinki. Participants received a base compensation of £4 for completing the entire experiment, A performance-based bonus (£0 to £2, average: £1) was awarded based on rewards earned in the MDP task.
Each participant completed 3 sequences for each step condition (2-step and 3-step MDP tasks), with each sequence comprising 50 episodes. The order of the 2-step and 3-step MDP tasks was randomized across sequences. State-cue assignment (animal images) were randomly determined for each sequence. Participants took a mandatory break (≥ 1 minute) between sequences.
To ensure data quality, we excluded participants based on response time, action bias, and task performance, as follows. Participants with an average response time was below 0.75 s or whose standard deviation of response time was over 10 s were excluded. Additionally, participants exhibiting a strong action bias (where 85% or more of actions were either left or right presses) were rejected. Furthermore, participants with an average reward below below (− 1.5 × interquartile range + first quantile) were excluded. This screening resulted in a final sample of 98 participants.
4.2.3 Cognitive model: Q-learning
To model decision-making in the MDP task, we employed a Q-learning agent [46]. At each step t, the agent was presented with the current task state st and selected an action at. The agent maintained Q-values, denoted as Q(s, a), for all state-action pairs, where s was a state of the set of all possible task states 𝒮 and a was an action of the set of available actions in that state 𝒞s. The probability of selecting action a was determined by a softmax policy:
where qit > 0 was a parameter called the inverse temperature or reward sensitivity, controlling the balance between exploration and exploitation.
After selecting action at, the agent received an outcome rt ∈ 0, 1 and transitioned to a new state st+1. The Q-value for the selected action was updated by
where qlr ∈ (0, 1) was the learning rate, determining how much newly acquired information replaced existing knowledge, and qdr ∈ (0, 1) was the discount rate, governing the extent to which future rewards influenced current decision. For actions not selected, the agent applied a forgetting factor qff, updating Q-values as
for
4.2.4 EIDT model
This section describes the specific models used for individuality transfer in the MDP task.
Data representation
Since MDP tasks involve sequential decision-making, each action sequence consists of multiple actions within a single session. In our experiment, each participant completed L trials per session, with L = 100 for the 2-step task and L = 150 for the 3-step MDP task. The action sequence is represented as [(s1, a1, r1), …, (sL, aL, rL)], where, st denotes the task state at trial t, at ∈ C represents the action selected from the set
Task solver
Before describing the encoder and decoder, we define the architecture of the task solver, which generates actions for the M-step MDP task. The task solver is implemented using a gated recurrent unit (GRU) [7] with Q cells, where Q = 4 for the 2-step task and Q = 8 for the 3-step task. At time-step t, the GRU takes as input the previous hidden state ht−1 ∈ ℝQ, the previous task state st−1, the previous action at−1, the previous reward rt−1, and the current task state st. It then updates the hidden state as
where Φ represents the GRU’s weights. The updated hidden state is then used to predict the probability of selecting each action through a fully-connected feed forward layer:
where vt represents the logit scores for each action (unnormalised probabilities), and W∈ ℝK×Q is the weight matrix. The probabilities of each action are computed using a softmax layer:
where π(at = Ck) represents the probability of selecting action Ck at time t, and [vt]i denotes the ith element of vt.
For input encoding, we used a 1-of-K scheme. The step of the MDP task is encoded as [1, 0, 0] for step 1, [0, 1, 0] for step 2, and [0, 0, 1] for step 3. Each task state sm is represented as [1, 0] or [0, 1] to distinguish the two state cues at each step. The participant’s action is encoded as C1: [1, 0] or C2: [0, 1], while the reward is represented as 0: [1, 0] or 1: [0, 1]. These encodings are concatenated to form input sequences.
The task solver TS(ψ; ΘTS) generates a sequence of predicted action probabilities
To evaluate prediction accuracy, the loss function
Encoder and decoder
The encoder ENC(α, ϕ; ΘENC) extracts an individuality index z from a sequence of actions α corresponding to a given environment ϕ. The first module of the encoder is a GRU, similar to the task solver, with R = 32 cells. The final hidden state hL ∈ ℝR serves as the basis for computing the individuality index z ∈ ℝM using a fully-connected feed-forward network with four layers d(·) as z = d(hL).
The decoder takes the individuality index z as input and generates the weights for the task solver by ΘTS = DEC(z; ΘDEC). The decoder is implemented as a single-layer linear network.
4.3 Experiment on MNIST task
This section describes the specific models used for individuality transfer in hand-written digit recognition (MNIST) task.
4.3.1 Task
The dataset used in this experiment was originally collected and published by Rafiei et al. [34]. In this task, participants were presented with a stimulus image depicting a handwritten digit and were required to respond by pressing the corresponding number key, as illusrtaed Figure 1. For further details regarding the task design and data collection, refer to [34].
4.3.2 EIDT model
Data representation
An action sequence, denoted as α or β, consists of a single action a and its corresponding response time b. The associated problem, represented as ϕ or ψ, corresponds to a stimulus image. The action a is selected from a set {C1, … CK}. Since the task involves recognizing digits ranging from 0 to 9, the number of possible actions is K = 10. The stimulus image, ϕ or ψ, is an image of size H× W. In this experiment, we adopted the same resolution as [34], setting H = W = 227.
Task solver
The task solver for the handwritten digit recognition task is based on the model proposed by [34]. Their model consists of a CNN and an evidence accumulation module. However, since their model represents average human behaviour and does not account for individuality differences, we replace the accumulation module with a GRU [6] to capture individuality. The CNN module processes the input image and produces an evidence vector e = CNN(ψ), where e ∈ ℝK and CNN(·) is based on the AlexNet architecture [26]. The weights of the CNN are sampled from a Bayesian neural network (BNN), introducing stochasticity in the output. This stochasticity enables the models to generate human-like, probabilistic decisions.
The stimulus image is fed into the CNN S times, generating S evidence distributions et ∈ ℝK at each time step t = 0, …, S −1. In this study, we set S = 16 to match the maximum response time, as described later. Since the CNN weights are stochastically sampled from the BNN, the CNN’s output varies even when the same image is input multiple times. To model individuality in decision-making, we introduce a GRU with Q cells (Q = 4 in our setup). The GRU receives as input the previous hidden state ht−1 ∈ ℝQ and the current evidence et, updating its hidden state as
where Φ represents the GRU’s network weights. The updated hidden state is passed through a dense layer (as defined in (10)) and a softmax layer (as defined in (11)) to generate the probability distribution over possible digit classifications [Pt(C1), …, Pt(CK)] at each time step t.
To evaluate the prediction error, we compare the action sequences generated by human participants {β, ψ} with those predicted by the task solver
In this task solver, the CNN (driven by BNN) models a visual processing system, while the RNN represents the decision-making system. We assume that the visual system (implemented by CNN and BNN) is shared across all individuals, whereas the decision-making system (implemented by RNN) captures individual differences. Based on this assumption, the CNN and BNN are pretrained using the MNIST dataset [26], and their weight distributions are fixed across individuals. The pretraining procedure followed the original methodology [34].
Encoder and decoder
Since each action sequence contains only a single action, it does not form a true “sequence.” This makes it challenging to extract individuality from a single data point. To address this, the encoder takes a set of single action sequences as input rather than a single sequence. Specifically, the encoder
Here, ϕu represent the stimulus presented in the uth trial, and αu represents the corresponding response. The number of action sequences U corresponds to the number of samples available for each individual in the dataset. Since the outputs for these action sequences are just averaged, U can be adjusted flexibly. The encoder architecture consists of a single CNN module, a single GRU, and a fully-connected feed-forward network. The CNN module is identical to the one used in the task solver. Given an input ϕu, let et,u represent the evidence output from the CNN at time step t. The GRU, which consists R cells (R = 16 in our setup), updates its hidden state based on the previous state, the current CNN evidence, and an encoding of the response action by
where Ψ represents the network weights. The function
4.4 Analysis
4.4.1 Within- and between-individual distances for evaluating the individuality index
To evaluate the disentanglement of the individuality index, we computed two types of distances: within-individual distance and between-individual-distance. Let z(α, ϕ) denote the individuality index derived from an action sequence (α, ϕ). The within-individual distance quantifies the average distance between the individuality index for each action sequence and the average individuality index for a given individual. It is defined as
where
where
4.4.2 Evaluation for uniqueness in task solvers
To assess whether a task solver generates unique behaviour for a specific participant, we compared two prediction loss scores. The first score, referred to as Original, represents the standard prediction loss for a specific participant when using the task solver designed for that participant. It is formulated as
where ΘENC and ΘDEC are omitted for simplicity, as in (4). The second score, referred to as Others, measures the prediction loss for a specific participant when using task solvers designed for other participants. It is formulated as
where 𝒫 represents the set of participant indices in the test participant pool. If the task solvers successfully captures the unique decision-making tendency of each participant, the Original score LK should be lower than the Others score ϒK for all participants K ∈ 𝒫.
5 Data availability
The behavioural data for the MDP task has been made publicly available at https://github.com/hgshrs/indiv_trans
Acknowledgements
This work was supported in part by the Japan Society for the Promotion of Science (JSPS) KAKENHI, grant number 22H05163 and 24K15047, and Japan Science and Technology Agency (JST) Advanced International Collaborative Research Program (AdCORP), grant number JPMJKB2307.
Additional information
6 Code availability
All code and trained models have been made publicly available at https://github.com/hgshrs/indiv_trans
Author contributions
H.H. designed and performed the research, collected and analyzed the data, and drafted and edited the paper.
Additional files
References
- [1]Measuring and understanding individual differences in cognitionPhilosophical Transactions of the Royal Society B: Biological Sciences 373:20170280Google Scholar
- [2]Understanding disentangling in beta-VAEarXiv Google Scholar
- [3]A contribution of cognitive decision models to clinical assessment: Decomposing performance on the Bechara gambling taskPsychological Assessment 14:253–262Google Scholar
- [4]Multi-head CNN–RNN for multi-time series anomaly detection: An industrial case studyNeurocomputing 363:246–260Google Scholar
- [5]Individual differences in cognitive abilitiesAnnual Review of Psychology 30:603–640Google Scholar
- [6]RTify: Aligning deep neural networks with human behavioral decisionsAdvances in Neural Information Processing Systems Google Scholar
- [7]Learning phrase representations using RNN encoder–decoder for statistical machine translationIn: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) Stroudsburg, PA, USA: Association for Computational Linguistics pp. 1724–1734Google Scholar
- [8]How much of reinforcement learning is working memory, not reinforcement learning? A behavioral, computational, and neurogenetic analysisEuropean Journal of Neuroscience 35:1024–1035Google Scholar
- [9]Model-based influences on humans’ choices and striatal prediction errorsNeuron 69:1204–1215Google Scholar
- [10]Disentangled behavioral representationsAdvances in Neural Information Processing Systems Google Scholar
- [11]Models that learn how humans learn: The case of decision-making and its disordersPLOS Computational Biology 15:e1006903Google Scholar
- [12]Memory states influence value-based decisionsJournal of Experimental Psychology: General 145:1420–1426Google Scholar
- [13]The interpretation of computational model parameters depends on the contexteLife 11:75474Google Scholar
- [14]Predictive and interpretable: combining artificial neural networks and classic cognitive models to understand human learning and decision makingIn: Proceedings of the 45th Annual Conference of the Cognitive Science Society pp. 928–935Google Scholar
- [15]Harmonizing the object recognition strategies of deep neural networks with humansAdvances in Neural Information Processing Systems Google Scholar
- [16]Using deep learning to predict human decisions and using cognitive models to explain deep learning modelsScientific Reports 12:4736Google Scholar
- [17]Prefrontal and striatal dopaminergic genes predict individual differences in exploration and exploitationNature Neuroscience 12:1062–1068Google Scholar
- [18]Harnessing the flexibility of neural networks to predict dynamic theoretical parameters underlying human choice behaviorPLOS Computational Biology 20:e1011678Google Scholar
- [19]Using recurrent neural network to estimate irreducible stochasticity in human choice behavioreLife 13Google Scholar
- [20]HypernetworksarXiv Google Scholar
- [21]Evidence that recurrent circuits are critical to the ventral stream’s execution of core object recognition behaviorNature Neuroscience 22:974–983Google Scholar
- [22]Probabilistic metarepresentations of neural networksarXiv Google Scholar
- [23]The relation between reinforcement learning parameters and the influence of reinforcement history on choice behaviorJournal of Mathematical Psychology 66:59–69Google Scholar
- [24]Recurrent processing in V1/V2 contributes to categorization of natural scenesThe Journal of Neuroscience 31:2488–2492Google Scholar
- [25]Deep neural networks: A new framework for modeling biological vision and brain information processingAnnual Review of Vision Science 1:417–446Google Scholar
- [26]ImageNet classification with deep convolutional neural networksAdvances in Neural Information Processing Systems Google Scholar
- [27]Intact reinforcement learning in healthy ageingbioRXiv Google Scholar
- [28]Cognitive model discovery via disentangled RNNsAdvances in Neural Information Processing Systems :61377–61394Google Scholar
- [29]Modeling individual differences using Dirichlet processesJournal of Mathematical Psychology 50:101–122Google Scholar
- [30]Model-based fMRI and its application to reward learning and decision makingAnnals of the New York Academy of Sciences 1104:35–53Google Scholar
- [31]The drift diffusion model as the choice rule in reinforcement learningPsychonomic Bulletin and Review 24:1234–1251Google Scholar
- [32]Using large-scale experiments and machine learning to discover theories of human decision-makingScience 372:1209–1214Google Scholar
- [33]BayesFlow: Learning complex stochastic models with invertible neural networksIEEE Transactions on Neural Networks and Learning Systems 33:1452–1466Google Scholar
- [34]The neural network RTNet exhibits the signatures of human perceptual decision-makingNature Human Behaviour 8:1752–1770Google Scholar
- [35]Large-scale, high-resolution comparison of the core visual object recognition behavior of humans, monkeys, and state-of-the-art deep artificial neural networksThe Journal of Neuroscience 38:7255–7269Google Scholar
- [36]The diffusion decision model: Theory and data for two-choice decision tasksNeural Computation 20:873–922Google Scholar
- [37]Artificial neural net-works for model identification and parameter estimation in computational cognitive modelsPLOS Computational Biology 20:e1012119Google Scholar
- [38]Learning internal representations by error propagationIn: Parallel Distributed Processing: Explorations in the Microstructure of Cognition: Foundations pp. 318–362Google Scholar
- [39]Reverse-engineering recurrent neural network solutions to a hierarchical inference task for miceAdvances in Neural Information Processing Systems Google Scholar
- [40]Iowa gambling task in schizophrenia: A review and new data in patients with schizophrenia and co-occurring cannabis use disordersSchizophrenia Research 92:74–84Google Scholar
- [41]Is human digital twin possible?Computer Methods and Programs in Biomedicine Update 1:100014Google Scholar
- [42]On the computational power of neural netsJournal of Computer and System Sciences 50:132–150Google Scholar
- [43]Reward-based training of recurrent neural networks for cognitive and value-based taskseLife 6Google Scholar
- [44]Using recurrent neural networks to understand human reward learningIn: Proceedings of the Annual Meeting of the Cognitive Science Society pp. 1388–1394Google Scholar
- [45]Recurrent neural networks can explain flexible trading of speed and accuracy in biological visionPLOS Computational Biology 16:e1008215Google Scholar
- [46]Reinforcement Learning: An IntroductionIn: Adaptive Computation and Machine Learning Cambridge, MA: MIT Press Google Scholar
- [47]Recurrent computations for visual pattern completionProceedings of the National Academy of Sciences 115:8835–8840Google Scholar
- [48]Wasserstein autoencodersarXiv Google Scholar
- [49]Exploration–exploitation mechanisms in recurrent neural networks and human learners in restless bandit problemsComputational Brain & Behavior Google Scholar
- [50]A critical review of sex differences in decision-making tasks: Focus on the Iowa gambling taskBehavioural Brain Research 238:95–108Google Scholar
- [51]Evidence for an accumulator model of psychophysical discriminationErgonomics 13:37–58Google Scholar
- [52]On the linear relation between the mean and the standard deviation of a response time distributionPsychological Review 114:830–841Google Scholar
- [53]Learning predictive statistics: strategies and brain mechanismsThe Journal of Neuroscience 37:144–17Google Scholar
- [54]Ten simple rules for the computational modeling of behavioral dataeLife 8:1–33Google Scholar
- [55]Using goal-driven deep learning models to understand sensory cortexNature Neuroscience 19:356–365Google Scholar
- [56]Task representations in neural networks trained to perform many cognitive tasksNature Neuroscience 22:297–306Google Scholar
- [57]Using cognitive models to map relations between neuropsychological disorders and human decision-making deficitsPsychological Science 16:973–978Google Scholar
Article and author information
Author information
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
Cite all versions
You can cite all versions using the DOI https://doi.org/10.7554/eLife.107163. This DOI represents all versions, and will always resolve to the latest one.
Copyright
© 2025, Hiroshi Higashi
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 0
- downloads
- 0
- citations
- 0
Views, downloads and citations are aggregated across all versions of this paper published by eLife.