Abstract
The brain forms certain deliberative decisions following normative principles related to how sensory observations are weighed and accumulated over time. Previously we showed that these principles can account for how people adapt their decisions to the temporal dynamics of the observations (Glaze et al., 2015). Here we show that this adaptability extends to accounting for correlations in the observations, which can have a dramatic impact on the weight of evidence provided by those observations. We tested online human participants on a novel visual-discrimination task with pairwise-correlated observations. With minimal training, the participants adapted to uncued, trial-by-trial changes in the correlations and produced decisions based on an approximately normative weighting and accumulation of evidence. The results highlight the robustness of our brain’s ability to process sensory observations with respect to not just their physical features but also the weight of evidence they provide for a given decision.
Introduction
In their efforts to break the Enigma code during World War II, Alan Turning and his colleagues at Bletchley Park recognized the importance of the concept of a “weight of evidence” for making decisions: noisy or ambiguous evidence is most useful when the influence or weight it has on the ultimate decision depends on its uncertainty. For the case of two alternatives, they used a weight of evidence in the form of the logarithm of the likelihood ratio (i.e., the ratio of the likelihoods of each of the two alternative hypotheses, given the observations), or logLR. The logLR later became central to the sequential probability ratio test (SPRT), which was proven to provide certain optimal balances between the speed and accuracy of such decisions (Barnard, 1946; Wald, 1947; Wald and Wolfowitz, 1948). Recognizing the general nature of this formulation, Turing and colleagues noted that the logLR would be “an important aid to human reasoning and … eventually improve the judgment of doctors, lawyers, and other citizens” (Good, 1979).
The logLR, or scaled versions of it, has since become ubiquitous in models of decision-making, including sequential-sampling models related to the SPRT like the drift-diffusion model (DDM). These models capture many behavioral and neural features of human and animal decision-making for a broad range of tasks (Gold and Shadlen, 2007; Smith and Ratcliff, 2004). However, it is still unclear whether decision-makers compute the normative weight of evidence—the logLR—or instead rely on approximations or other heuristics (Brown et al., 2009; Hanks et al., 2011; Ratcliff et al., 2016; Ratcliff and McKoon, 2008). Furthermore, previous findings have come from studies that have been restricted mainly to tasks that simplify the computation of the logLR by providing only observations that are statistically independent of each other. Under these conditions, the logLR can be computed separately for each observation, often by scaling the strength of that observation by its signal and noise characteristics learned from past observations from the same source (Fig. 1a). These weights of evidence are added together (i.e., the logLRs are accumulated over time) to form the aggregated decision variable that governs the final choice.
More generally, however, non-independence in the statistics of the observations can have substantial effects on how those observations should be weighed to form effective decisions (Fig. 1b,c). If not accounted for appropriately, these effects can cause a decision-maker to over- or under-estimate the weight of available evidence and make suboptimal decisions. Such suboptimalities have real-world consequences. For example, misestimation of correlation patterns in mortgage defaults is thought to have played a role in triggering the global financial crisis of 2008 (Salmon, 2009). Neglecting correlations can contribute to false beliefs and ideological extremeness in social and political settings (Denter et al., 2021; Glaeser and Sunstein, 2009; Levy et al., 2022; Ortoleva and Snowberg, 2015). Likewise, correlations in the physical environment should, in principle, be leveraged to support perception (Geisler, 2008; Parise, 2016). Yet whether and how people account for correlations when making perceptual decisions is not well understood.
The goal of this preregistered study (https://osf.io/qj92c) was to test how humans form simple perceptual decisions based on observations with different degrees of correlation. We previously showed that both theoretically optimal (i.e., an ideal observer that maximizes decision accuracy) and human observers flexibly adjust how evidence is accumulated over time to account for the temporal dynamics of the sequentially presented observations (Glaze et al., 2015; Veliz-Cuba et al., 2016). Here we assess how both ideal and human observers weigh and accumulate evidence that is based on pairs of correlated observations (Fig. 2). Under these conditions, the normative decision process involves computing a weight of evidence by scaling each paired observation by a function of the underlying correlation, then accumulating that weight of evidence across pairs until reaching a predefined bound that, like in the DDM, balances decision speed and accuracy. As we detail below, we found that people tend to follow these normative principles, accounting appropriately for the correlations (albeit based on slight misestimates of correlation magnitude) and demonstrating the robustness and flexibility with which our brains can appropriately weigh and accumulate evidence when making simple decisions.
Results
We tested 100 online participants performing a novel task that required them to form simple decisions about which of two latent sources generated the observed visual stimuli, in the presence of different correlation structures in the stimuli (Fig. 2). The task design was based on principles illustrated in Fig. 1: the normative weight of evidence for the identity of a source of paired observations from correlated, Gaussian random variables depends systematically on the magnitude and sign of the correlation (Fig. 1b). In this case, negative pairwise correlations provide, on average, stronger evidence with increasing correlation magnitude, because less overlap of the generative source distributions allows them to be more cleanly separated by the decision boundary (Fig. 1c, left inset). Positive pairwise correlations provide, on average, weaker evidence with increasing correlation magnitude, because more overlap of the generative source distributions causes them to be less cleanly separated by the decision boundary (Fig. 1c, right inset).
Participants reported the generative source (left or right) of noisy observations, depicted onscreen as the position of stars along a horizontal line (Fig. 2a). Observations were presented in pairs. Each element of the pair had the same mean value across samples from the generative source, but the noise correlation within pairs was manipulated on a per-trial basis. We assigned each participant to a correlation-magnitude group (|ρ| = 0.2, 0.4, 0.6, 0.8; 25 participants per group) in which the pairwise correlation on a given trial was drawn from three conditions: −ρ, 0, or +ρ. We equated task difficulty across participants by calibrating the means of the generative distributions (see Methods). We interleaved randomly the three correlation conditions with the two sources (left, right) and two levels of task difficulty (low, high), for 12 total conditions for each correlation-magnitude group.
Crucially, we adjusted the means of the generative distributions to ensure that the expected logLR (which we term the evidence strength) was constant across correlation conditions (Fig. 2b,c). For example, because negative correlations increase logLR (Fig. 1c), we used smaller differences in means for the negative-correlation conditions than for the zero-correlation conditions. As a result, we expect participants who make decisions by weighing the evidence according to the true logLR to produce identical distributions of choices and response times (RTs) across correlation conditions. In contrast, we expect participants who ignore the correlations to under-weigh the evidence provided by negative-correlation pairs (and thus take longer to accumulate evidence, leading to longer RTs and higher accuracy) and over-weigh the evidence provided by positive-correlation pairs (leading to shorter RTs and lower accuracy). We further expect strategies between these two extremes to have more mixed effects on choices and RTs, as we detail below.
Human response times are influenced by correlated observations
The example participant in Fig. 3a exhibited behavioral patterns that were illustrative of the overall trends we observed. Specifically, their choice accuracy was affected by evidence strength (higher accuracy for stronger evidence) but not correlation (this participant was tested using correlation values of −0.6, 0.0, and 0.6). In contrast, their RTs were affected by both evidence strength and correlation, including faster responses on correct trials using stronger evidence and more positive correlations.
Likewise, across our sample of participants choices depended strongly on evidence strength but not correlation (Fig. 3b). Logistic models fit to individual participant’s evidence-strength-dependent psychometric data demonstrated no benefit to fitting separate models per correlation condition versus a single model fit jointly to all three correlation conditions (mean ΔAIC = −4.14, protected exceedance probability [PEP] = 1.0 in favor of the joint model). This result held true at each correlation magnitude individually (all mean ΔAIC < −2.0, all PEP > 0.8).
In contrast, RTs were affected by both evidence strength and correlation, with a tendency of participants to respond faster for stronger evidence and more-positive correlations (Fig. 3b). A linear mixed-effects model fit to median RTs from correct trials confirmed these observations, indicating effects of evidence strength (F(1,98.00) = 174.24, p < 0.001), the sign of the correlation (negative, zero, positive) within participants (F(2,131.56) = 219.96, p < 0.001), and the interaction between the sign of the correlation and its magnitude between participants (F(2, 131.56) = 81.04, p < 0.001). That is, the effects of correlations on correct RTs were more pronounced in participants tested using stronger correlations. Similar effects were also present on error trials (evidence strength: (F(1,960.54) = 19.21, p < 0.001), sign of correlation: (F(2, 234.74) = 58.41, p < 0.001), correlation sign x magnitude (F(2, 233.48) = 13.50, p < 0.001).
In short, the patterns of choice data that we observed were consistent with decisions that took into account the correlations in the observations, which by design were necessary to equate the weight of evidence across correlation conditions. In contrast, the patterns of RT data that we observed rule out the possibility that the participants used normative evidence weighting (i.e., based on the true logLR) to make their decisions, because if they did, their RTs would not depend on the correlation condition. These findings leave open a broad range of possible weighing strategies between the two extremes of an ideal observer and a naïve observer who ignores the correlations (i.e., assumes independence). The analyses detailed below aimed to more precisely identify where in that range our participants’ strategies fell.
RTs are consistent with a decision bound on approximate logLR
We analyzed the RT data in more detail, based on principles from the DDM and SPRT, wherein simple decisions result from a process in which evidence is accumulated over time until reaching one of two fixed bounds, corresponding to the two choices. This process governs both the choice (which bound is reached first) and RT (when the bound is reached). We considered two forms of evidence with different scale factors (i.e., the scale value multiplied by each star position to compute its weight of evidence): 1) “naïve” were proportional to the generative mean, µg, of a pair of samples, which is a standard assumption in many implementations of the DDM (Palmer et al., 2005) but in this case ignores the correlations and thus does not produce a weight of evidence equivalent to the true logLR; and 2) “true” were proportional to , which takes into account the correlations and produces a weight of evidence equivalent to the true logLR.
Because we designed the task to present stimuli with equal expected logLR (evidence strength) across correlation conditions, decisions based on an accumulation of the true logLR to a fixed bound would have similar mean RTs across correlation conditions. In contrast, decisions based on an accumulation of the naïve logLR would be expected to have different effects for positive versus negative correlations. Ignoring positive correlations is equivalent to ignoring redundancies in the observations, which would lead to over-weighting the evidence and thus reaching the bound more quickly, corresponding to shorter RTs and lower accuracy. Ignoring negative correlations is equivalent to ignoring synergies in the observations, which would lead to under-weighting the evidence and thus reaching the bound less quickly, corresponding to longer RTs and higher accuracy (Fig. 1c).
As noted above, participants had RTs that were, on average, either relatively constant or slightly decreasing as a function of increasing correlations, particularly for larger correlations (Fig. 4a,b). These trends were not consistent with a decision process that used a fixed bound that ignored correlations. They also were not completely consistent with a decision process that used a fixed bound on the true logLR (because of the trend of decreasing RTs with increasing correlations). Instead, these results could be matched qualitatively to simulations that made decisions based on an approximation of logLR computed using underestimates of the correlation-dependent scale factor (Fig. 4c). We examined this idea more quantitatively using model fitting, detailed in the next section.
Correlation-dependent adjustments affect the bound in a drift-diffusion model
To better understand how the participants formed correlation-dependent decisions, we developed a variant of the DDM that can account for pairwise-correlated observations. The DDM jointly accounts for choices and RTs according to a process that accumulates noisy evidence over time until reaching a decision bound (Fig. 5a). The model includes two primary components that govern the decision process and can be based on just two free parameters. The drift rate governs the average rate of information accumulation (Palmer et al., 2005; Ratcliff and McKoon, 2008). This term typically depends on the product of the strength or quality of the sensory observations (generally varied via the mean, or signal, of the evidence distribution, μg) and the decision-maker’s sensitivity to those observations (the fit parameter k; i.e., drift rate(ρ) = kμ. The bound height governs the decision criterion, or rule, which corresponds to the amount of evidence required to make a decision and controls the trade-off between decision speed and accuracy (Heitz, 2014). This term is often just a single fit parameter representing symmetric bounds; i.e., bound height = B.
For our task, changing ρ affects the observation distribution (inset in Fig. 5a) in terms of both signal (the different values of μρ we used to offset the effect of the correlation affected the “drift” part of “drift-diffusion”) and noise (the different values of σρ that reflect effects of correlations on the standard deviation of the sum-of-pairs distribution affect the “diffusion” part of “drift-diffusion”). To account for these effects in the DDM, which typically expresses both the drift rate and bound height in terms of the signal-to-noise ratio of the observation distribution (Palmer et al., 2005; Shadlen et al., 2006), we scaled the drift rate by the correlation-dependent component of the evidence noise:
where k0 is the drift-rate parameter for the zero-correlation condition, which also includes the correlation-independent component of the observation noise, σg (i.e., ). Note that the correlation-dependent scale factor used here is the same factor we used to scale the generative means of the stars to ensure equal logLRs across correlation conditions (). As a result, these factors cancel in the drift-rate equation above, which depends only on subjective (fit) sensitivity (k0) and the objective strength of evidence expressed in terms of the zero-correlation condition (μ0).
Implemented this way, scaling the strength of observations to compute a weight of evidence is equivalent to scaling the bound height to account for the correlation-dependent effects on both signal and noise, which again cancel (akin to the transformation in Fig. 2c, from the left to the right panel; see Methods). However, our participants did not necessarily know the objective correlation (ρ) but instead relied on subjective internal estimates (). To account for possible misestimates, we assumed that their decision bounds were scaled by subjective estimates of correlation-dependent changes in the noise:
where B0 is the bound height for the zero-correlation condition, the numerator is the subjective component of the correlation-dependent scale factor, and the denominator reflects effect of the correlation on the diffusion process, as in the drift rate equation above. This formulation leads to the following predictions (Fig. 5b):
If , then Bρ = B0: when correlations are estimated accurately, the drift rate and bound height are equal across correlations, giving equal average choices and RTs (e.g., Fig. 5b, right-most column).
If and ρ > 0, then Bρ < B0: when positive correlations are underestimated, the bound is lower than optimal because the evidence has been over-weighed, and average performance is faster but less accurate.
If and ρ < 0, then Bρ > B0: when negative correlations are underestimated, the bound is higher than optimal because the evidence has been under-weighed, and average performance is slower but more accurate.
In short, failing to (fully) account for correlations causes the decision-maker to set their bound according to a misestimate of the weight of evidence provided by each observation. The deviation of the bound height from the normative adjustment in turn alters the speed-accuracy tradeoff, with a greater qualitative effect on RTs compared to accuracy.
Human performance is consistent with correlation-dependent bound adjustments
Qualitatively, participants’ patterns of choice and RTs across correlation conditions were consistent with bound-height adjustments that depended on slight underestimates of the objective correlation (compare Figs. 3b and 5b). To examine this idea more quantitatively and rule out alternative possibilities, such as adjustments to the drift rate, we fit four DDMs to each participant’s data: 1) a base model, with no adjustment based on the correlation; 2) a drift model, with separate drift parameters for the negative, zero, and positive correlation conditions (k−, k0, k+, respectively); 3) a bound model, which implements the normative bound scaling, where and are free parameters that estimate the subjective correlation in the negative and positive correlation conditions, respectively; and 4) a bound+drift model, which combines the adjustments from the drift and bound models. Prior to fitting these models, we confirmed that the DDM can generally account for choice and RTs in our novel task via fits to the zero-correlation condition. These fits confirmed that the DDM could account for the data, and that fits were improved with the addition of a linear collapsing bound, which we used in all subsequent fits (Figure 6—figure supplement 1).
These model fits indicated that correlation-dependent bound adjustments were more important than correlation-dependent drift adjustments for capturing differences in behavioral performance between correlation conditions (Fig. 6a, compare bound model to base and drift models). However, drift-rate adjustments were also useful for the high, but not low, correlation magnitudes (model comparison per group shown in Fig. 6b; best-fitting model per group shown in Fig. 6c; between-group random-effects model comparison (|ρ| <0.5 versus |ρ| > 0.5), p < 0.001; Rigoux et al., 2014). These drift-rate adjustments at higher correlation magnitudes accounted for the fine-scale ordering of choice behavior across correlation conditions, which was opposite to that predicted by the bound model (Fig. 6c, 0.8 correlation inset). A likely explanation for this effect relates to our task design, which involved equating evidence strength (expected logLR) across correlation conditions via changes in the generative mean of the star positions that were necessarily more dramatic for higher versus lower correlation magnitudes (see Fig. 1c). If perceived star position is not a linear function of objective star position, different drift parameters per correlation would provide a better fit (Palmer et al., 2005). Put another way, the very small generative means in the −0.6 and −0.8 correlation conditions seem to provide weaker observed stimulus strength than predicted by a linear function between screen position and stimulus strength, leading to smaller drift parameters in these conditions (Figure 6—figure supplement 2).
Human performance is consistent with a weight of evidence based on the approximated correlation
As predicted, bound adjustments best accounted for behavioral differences across correlation conditions. Because these adjustments took the form of the normative bound scaling with a subjective correlation, we leveraged the fit correlations to ask how well participants estimated the correlation, and whether deviations from equal performance across correlation conditions were caused by underestimation. To ensure that correlation effects on drift rate were accounted for in these analyses, we used the best-fitting model (either bound or bound+drift) per correlation-magnitude group.
We found a strong relationship between the objective and subjective correlations (Fig. 7a; B = 0.71, t(99) = 40.93, p < 0.001, Fisher z-transformed scale), confirming that participants were sensitive to the correlations and used them to adjust their decision process in a near-optimal manner. However, the slope of this relationship was less than one. That is, participants underestimated the objective correlation, on average (test of ) (Fisher z-transformed scale): B = −0.18, t(99) = −12.18, p < 0.001), consistent with our hypothesis that their deviations from normative behavior (i.e., unequal performance across correlation conditions) resulted from systematic underestimates of the generative correlation. These estimates also tended to be more variable for positive versus negative correlations (mean value of the standard deviation of Fisher z-transformed estimates across positive-correlation conditions = 0.22, 0.10 for negative correlation conditions), likely reflecting the weaker consequences of misestimating positive versus negative correlations on performance (Fig. 1c).
These biased, subjective correlation estimates resulted in slight but systematic deviations from optimal of the corresponding inferred decision bounds used by the participants. Specifically, inferred bounds deviated from optimal, on average (Fig. 7b; test of log10(bound scale factor): B = −0.037, t(198) = −14.83, p < 0.001). Because of the non-linear relationship between the correlation and the normative scale factor (Fig. 1c), the inferred bound scaling tended to be closer to optimal for positive correlations compared to negative correlations (B = 0.021, t(198) = 4.26, p < 0.001).
These correlation-dependent differences in the decision process did not seem to reflect ongoing adjustments that might involve, for example, feedback-driven learning specific to this task. In particular, the participants tended to exhibit some learning over the course of the task, involving substantial decreases in RT (the mean ± sem difference in RT between the first and second half of the task, measured across participants, was 0.71 ± 0.06 sec, respectively, Mann-Whitney test for H0: median difference = 0, p < 0.001) at the expense of only slight decreases in accuracy (0.02 ± 0.00% correct, p = 0.004). These trends reflected a tendency to use slightly higher drift rates (Fig. 8a) and lower decision bounds (Fig. 8b) in the latter half of the task, a pattern of results that is consistent with previous reports of practice effects for simple decisions (Balci et al., 2011; Dutilh et al., 2009). However, these adjustments were not accompanied by similar, systematic adjustments in the participants’ subjective correlation estimates, which were similar in the first versus second half of the task (Fig. 8c). This conclusion was supported by a complementary analysis showing that linear changes in RT as a function of trial number within a session tended to be the same for positive- and negative-correlation trials, as expected for stable relationships between correlation and RT (Wilcoxon rank-sum test for H0: median difference in slope = 0, p < 0.05 for just one of 8 evidence strength x correlation magnitude conditions, after accounting for multiple comparisons via Bonferroni correction). Thus, participants’ decisions appeared to be based on relatively stable estimates of the stimulus correlations that could be determined and used effectively on a trial-by-trial basis.
Discussion
This preregistered study addressed a fundamental question in perceptual decision-making: how do people convert sensory observations into a weight of evidence that can be used to form a decision about those observations? This question is important because evidence weighting affects how information is combined and accumulated over multiple sources and over time, ultimately governing the speed and accuracy of the decision process (Bogacz et al., 2006; Wald and Wolfowitz, 1948). To answer this question, we focused on correlations between observations, which are common in the real world, often ignored in laboratory studies, and can have a dramatic impact on the amount of evidence provided by a given set of observations. For simple perceptual decisions with correlated observations, the normative weight of evidence that accounts for these correlations can be expressed as a logLR. We showed that human participants make decisions that are approximately consistent with using this normative quantity, mitigating unintended shifts of the speed-accuracy tradeoff that would result from ignoring correlations. Below we discuss the implications of these findings for our understanding of the computations and mechanisms the brain uses to form simple decisions.
Previous support for the idea that human decision-makers can weigh evidence following normative principles has come from two primary lines of research. The first is studies of perceptual cue combination. Perceptual reports based on cues from multiple sensory modalities, or multiple cues from the same modality, often reflect weights of evidence that scale with the relative reliability of each cue, consistent with Bayesian theory (Ernst, 2005; Noppeney, 2021). The second is studies of evidence accumulation over time. The relationship between speed and accuracy for many decisions can be captured by models like the DDM that assume that the underlying decision process involves accumulating quantities that are often assumed to be (scaled) versions of the logLR (Bogacz et al., 2006; Edwards, 1965; Gold and Shadlen, 2001; Laming, 1968; Stone, 1960).
Central to the interpretation of these studies, and ours, is understanding the scale factors that govern evidence weights. In their simplest forms, these scale factors are scalar values that are multiplied by stimulus strength to obtain the weight of evidence associated with each observed stimulus. These weights are then combined (e.g., by adding them together if they are in the form of logLR) to form a single decision variable, which is then compared to one or more criterion values (the bounds) to arrive at a final choice, as in the DDM (see Fig. 5a). Thus, as long as there is a linear relationship between stimulus strength and logLR, then using an appropriate, multiplicative scale factor to compute the weight of evidence (either scaling the observations or the bound, depending on the particular algorithmic implementation) can support normative decision-making.
These kinds of decision processes have been studied under a variety of conditions that have provided insights into how the brain scales stimulus strength to arrive at a weight of evidence. In the simplest evidence-accumulation paradigms, stimulus strength is held constant across decisions within a block. In this case, the appropriate scale factor can be applied in the same way to each decision within a block, and normative changes in scaling across blocks are equivalent to shifting the decision bound to account for changes in stimulus strength. Results from studies using these paradigms have been mixed: some found that participants vary their bound based on changes in stimulus strength across blocks (Malhotra et al., 2017; Starns and Ratcliff, 2012), whereas others found that participants adopt a fix bound across stimulus strengths (Balci et al., 2011). Interpretation of these studies is complicated by the fact that the participant is typicall yassumed to have the goal of maximizing reward rate, which is a complicated function of multiple task parameters, including stimulus strength and timing (Bogacz et al., 2006; Zacksenhouse et al., 2010). Under such conditions, failure to take stimulus strength into account, or failure to do so optimally, could be a result of the particular strategy adopted by the decision-maker rather than a failure to accurately estimate a stimulus-strength-dependent scale factor. For example, several studies found that people deviate from optimal to a greater degree in low-stimulus-strength conditions because they value accuracy and not solely reward rate (Balci et al., 2011; Bohil and Maddox, 2003; Starns and Ratcliff, 2012). Additionally, deviations from optimal bounds have been found to depend on the uncertainty with which task timing is estimated, rather than uncertainty in estimates of stimulus strength (Zacksenhouse et al., 2010), suggesting that people can estimate stimulus strength even if they do not use it as prescribed by reward-rate maximization. By fixing expected logLR (evidence strength) across conditions, we avoided many of these potential confounds and isolated the effects of correlations on behavior.
More commonly, stimulus strength is varied from trial-to-trial in evidence-accumulation tasks. Under these conditions, the standard SPRT (and the DDM as its continuous-time equivalent; Bogacz et al., 2006) are no longer optimal (Deneve, 2012; Drugowitsch et al., 2012; Moran, 2015), because those models typically assume that the same scale factor is used on each trial, but different scale factors are needed to compute the normative weight of evidence (logLR) for different stimulus strengths. These considerations have led some to argue that it is highly unlikely that humans perform optimal computations, particularly under conditions of heterogenous stimulus strengths, because the precise stimulus statistics needed to compute the logLR are assumed to be unavailable or poorly estimated (Ratcliff et al., 2016; Ratcliff and McKoon, 2008). Relatedly, if decision-makers set their bounds according to the true logLR for each stimulus (equivalent to the goal of maintaining, on average, the same level of accuracy across stimuli), the psychometric function should be flat as a function of stimulus strength, whereas RTs should decrease with increasing stimulus strength. That decisions are both more accurate and faster with increasing stimulus strength argues strongly against the idea that people set bounds based on a fixed expected accuracy or that the accumulated evidence is scaled exactly proportional to the logLR (Hanks et al., 2011).
However, several modeling and empirical studies have shown that it is possible to adjust how decisions are formed about stimuli whose statistics vary from trial to trial, in a manner that is consistent with trying to use optimal forms of the weight of evidence. These adjustments include scaling the decision variable and/or decision bounds within a trial according to online estimates of stimulus strength or some proxy thereof, particularly when the distribution of evidence-strength levels are known (Deneve, 2012; Drugowitsch et al., 2012; Hanks et al., 2011; Huang and Rao, 2013; Malhotra et al., 2018; Moran, 2015). One possible proxy for stimulus strength is the time elapsed within a trial (Drugowitsch et al., 2012; Hanks et al., 2011; Kiani and Shadlen, 2009; Malhotra et al., 2018): the more time has passed in a trial without reaching a decision bound, the more likely that the stimulus is weak. Under certain conditions, human decision-making behavior is consistent with such adjustments (Drugowitsch et al., 2012; Malhotra et al., 2017; Palestro et al., 2018).
Our results imply that outside relatively simple cases involving statistically independent observations, elapsed time cannot serve as a sole proxy for stimulus strength. In particular, pairs of correlated observations can complicate the relationship between stimulus quality and elapsed time: in our task, negative correlations lead to slower decisions than positive correlations if the observations are treated as uncorrelated, when in fact stimulus strength is stronger for negative correlations than positive correlations. Therefore, in more general settings elapsed time should be combined with other relevant statistics, such as the correlation. In support of this idea, our data are consistent with decisions that used collapsing bounds, which can be a proxy for stimulus strength under an elapsed-time heuristic, but those collapsing bounds alone could not account for the clear behavioral adjustments to trial-to-trial differences in stimulus correlations.
Our results are in stark contrast with the literature on correlations in behavioral economics, which suggests that people fail to use correlations appropriately to inform their decision-making. When combining information from multiple sources (e.g., for financial forecasts: Budescu and Yu, 2007; Enke and Zimmermann, 2017; Hossain and Okui, 2021; Maines, 1996, 1990; or constructing portfolios of correlated assets: Eyster and Weizsacker, 2016; Laudenbach et al., 2022), most participants exhibit “correlation neglect” (i.e., partially or fully failing to account for correlations), which often leads to reduced decision accuracy. Positive correlations have also been proposed to lead to overconfidence, which has been attributed to failing to account for redundancy (Eyster and Rabin, 2010; Glaeser and Sunstein, 2009; Ortoleva and Snowberg, 2015) or to the false assumption that consistency among information sources suggests higher reliability (Kahneman and Tversky, 1973).
These discrepant results are likely a result of the vast differences in task designs between those studies and ours. Those tasks tended to present numerical stimuli representing either small samples of correlated sources or explicitly defined correlation coefficients, often in complicated scenarios. Under such conditions, participants may fail to recognize the correlation and its importance, or they may not be statistically sophisticated enough to adjust for it even if they do (Enke and Zimmermann, 2017; Maines, 1996). In contrast, highly simplified task structures increase the ability to account for correlations (Enke and Zimmermann, 2017). Nevertheless, even in simplified cases, decisions in descriptive scenarios likely rely on very different cognitive mechanisms than decisions that, like ours, are based directly on relatively low-level sensory stimuli. For example, decisions under risk can vary substantially when based on description versus direct experience (Hertwig and Erev, 2009), and giving passive exposure to samples from distributions underlying two correlated assets has been shown to alleviate correlation neglect in subsequent allocation decisions (Laudenbach et al., 2022).
These differences likely extend to how and where in the brain correlations are represented and used (or not) to inform different kinds of decisions. For certain perceptual decisions, early sensory areas may play critical roles. For example, when combining multiple visual cues to estimate slant, some observers’ estimates are consistent with assuming a correlation between cues, which is sensible because the cues derive from the same retinal image and likely overlapping populations of neurons (Oruç et al., 2003; Rosas et al., 2007). The combination of within-modality cues is thought to be encapsulated within the visual system, such that observers have no conscious access to the individual cues (Girshick and Banks, 2009; Hillis et al., 2002). These results suggest that the visual system may have specialized mechanisms for computing correlations among visual stimuli (which may or may not involve the well-studied, but different, phenomena of correlations in the patterns of firing rates of individual neurons; Cohen and Kohn, 2011) that are different than those used to support higher-order cognition.
The impact of correlations on the weight of evidence ultimately depends on the type of correlation and its relationship to other statistical features of the task environment and to intrinsic correlations in the brain (Averbeck and Lee, 2006; Bhardwaj et al., 2015; Hossain and Okui, 2021; Hu et al., 2014; Moreno-Bote et al., 2014). We showed that this impact can be substantial, and that human decision-makers’ sensitivity to correlations does not seem to require extensive, task-specific learning and can be adjusted flexibly from one decision to the next. Further work that pairs careful manipulation of task statistics with neural measurements could provide insight into how the brain tracks stimulus correlations and computes the weight of the evidence to support effective decision-making behaviors under different conditions.
Data and code availability
The datasets generated and analyzed for this article are available at https://osf.io/qygkc/. The analysis code for this article is available at https://github.com/TheGoldLab/Analysis_Tardiff_Kang_Correlated.
Acknowledgements
NT was supported by a T32 training grant from the National Institutes of Health [MH014654]. JG was supported by a CRCNS grant from the National Science Foundation [220727]. JK was funded by the Penn Undergraduate Research Mentorship program (PURM). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. We thank Long Ding for helpful comments on the manuscript.
Methods
Participants
One hundred human participants took part in this online study (42 male, 43 female, 3 other, 12 N/A; median age: 24 yrs, range 18–53, 1 N/A), each of whom provided informed consent via button press. Human protocols were approved and determined to be Exempt by the University of Pennsylvania Internal Review Board (IRB protocol 844474). Participants were recruited using the Prolific platform (https://www.prolific.com/). They were paid a base amount of $9.00 for a projected completion time of 1 hour. They also could receive a bonus of up to $8, depending on task performance (see below).
Behavioral task
The task was developed in PsychoPy (v. 2021.1.4; Peirce et al., 2019), converted to JavaScript (PyschoJS), and run on the online experiment hosting service Pavlovia (https://pavlovia.org/), via functionality integrated into PsychoPy. On each trial, the participant saw a sequence of observations. Each observation consisted of two stars displayed simultaneously. The stars’ horizontal positions were generated from a bivariate Gaussian distribution with equal means and variances for each star position and a correlation between star positions that changed from trial-trial-to-trial (the generative distribution), while their vertical position was fixed in the center of the display. The stars were generated by either a “left” source or a “right” source, varied randomly from trial-to-trial. The two sources were equidistant from the vertical midline of the screen, corresponding to equal means of the generative distribution with opposite signs. To prevent stars from being drawn past the edge of the display, their positions were truncated to a maximum value of 0.7, in units of relative window height. For a standard 16:9 monitor at full screen, this procedure implies that positions could not take on values past 78.8% of the distance from the center of the screen to the edge. Within a trial, new observations were generated from the underlying source distribution every 0.2 sec. Participants were instructed to indicate whether the stars were being generated by the left or the right source once they believed they had accumulated enough noisy information to make an accurate decision.
Each participant was assigned randomly to one of the four correlation-magnitude groups (|ρ| = 0.2, 0.4, 0.6, or 0.8; 25 participants per group) and completed 768 trials, which were divided into 4 blocks of 192 trials, with brief breaks in between. Within each block, there were 12 different stimulus conditions: 2 sources x 2 evidence strengths x 3 correlations. The source (left, right), evidence strength (high, low), and correlation (−ρ, 0.0, +ρ) were varied pseudo-randomly from trial-to-trial, per participant. Within each block, the trials were divided into 16 sets, with one trial of each condition per set. Each condition was presented in random order within a set, such that all 12 conditions were presented once before the next repetition of a given condition, resulting in 64 total repetitions of each condition across the experiment. Participants received 1 point for each correct choice and −2 points for each incorrect choice (floor of 0 points). The total number of points received by the end of the task was divided by the total possible points (768), and that proportion of $8 was awarded as the bonus.
Prior to completing the main task, each participant completed first a set of training trials, then a staircase procedure to standardize task difficulty across participants. We used a 3-down, 1-up staircase procedure to identify each participant-specific evidence-strength threshold (i.e., by varying the mean of the star-generating distribution while holding its standard deviation at a constant value of 0.1, in units of relative window height) that resulted in a target accuracy of 79.4% in the zero-correlation condition (García-Pérez, 1998). Staircase trials were presented at a fixed-duration of 1.4 sec, which in pilot data was roughly the mean RT in the free-response paradigm used in the main task, to equate the amount of information provided to each participant and avoid potential individual differences in the speed-accuracy trade-off. The high and low evidence strength conditions used in the main task were then defined as 0.4 and 2.5 times each participant’s evidence-strength threshold, respectively.
Because the staircase procedure should standardize accuracy across participants, a performance that is much lower than the target accuracy can be interpreted as a failure of the staircase procedure, a failure of the participant to maintain engagement in the task, or both. Therefore, we kept recruiting participants until we had 25 in each correlation-magnitude group with task performance at 70% or higher. No more than three candidate participants in each group were excluded due to this criterion.
Ideal-observer analysis
Bivariate-Gaussian observations (x1, x2) with equal means μg, standard deviations σg, and correlation ρ are distributed as:
where S is the generative source. For the problem of choosing between two such generative sources, S0 and S1, the normative weight of evidence can be computed using the log-likelihood ratio, . For sources with means μ0 and μ1 and equal σg and ρ, the logLR reduces to:
Our task had equal and opposite generative means, μ0 = −μ1 = μg. Under these conditions, the logLR further simplifies to:
This formulation makes clear that the logLR is a weight of evidence composed of the sum of the observations (for our task corresponding to the horizontal locations of the two stars) times a scale factor that depends on the generative properties of the sources. Because this logLR, which is expressed in terms of bivariate observations (x1, x2), depends only on the sum of the observations, it is equivalent to a logLR expressed in terms of univariate observations composed of the sum of each pair (i.e., ; see Fig. 2c). The logLR for a sequence of these (identically distributed, paired) observations is the sum of the logLRs for the individual (paired) observations.
We defined the evidence strength for a given condition as the expected value of the logLR for a single (paired) observation:
Therefore, to equate the evidence strength between two conditions with equal σg, but one with correlation = 0 and one with correlation ρ ≠ 0, we adjusted the generative mean of condition ρ to offset the correlation-dependent scale factor :
Drift-diffusion modeling (DDM)
In the DDM, noisy evidence is accumulated into a decision variable until reaching one of the two bounds, representing commitment to one of two choices (e.g., left or right). In general, the average rate of accumulation is governed by the drift rate:
where μg and σg are the mean and standard deviation of the generative distribution of the observations (which, as detailed above, for our task can be expressed as the distribution of sums of pairwise observation, x1 + x2). The drift parameter k captures subjective scaling of the stimulus strength (i.e., the signal-to-noise ratio, ), which accounts for individual differences in perceptual sensitivity and other factors.
There is an arbitrary degree of freedom in these and related models, which form equivalence classes when the decision variable and decision bound are both scaled in the same way (Green and Swets, 1966; Palmer et al., 2005). Fixing this extra degree of freedom in the DDM is typically accomplished by setting σg = 1, which causes the drift rate and bound height to be scaled implicitly by the standard deviation of the observation distribution (Palmer et al., 2005). This formulation is straightforward when stimulus strength is varied only via changes in signal, μg, and not noise (the “diffusion” in “drift-diffusion”), σg. However, our task included correlation-dependent changes in both signal and noise. In particular, the correlation scales the standard deviation of the sum distribution by . Therefore, we explicitly set the noise across correlation conditions, which is accomplished by scaling the drift rate and bound height by the relative change in the generative standard deviation induced by the correlation. Thus, for our task the drift rate for correlation ρ is:
where μρ is the correlation-specific generative mean, accounting for changes in signal; k0 is the subjective drift parameter, which is implicitly scaled by σg under the unit variance assumption ; and is the relative change in the generative standard deviation induced by the correlation. By specifying the noise with this latter term, the unit variance assumption of the DDM is maintained and k0 can accurately reflect subjective scaling of stimulus strength across correlation conditions. Because we manipulated μρ to offset the effect of the correlation on the noise, the drift rate across correlation conditions reduces to:
reflecting the equality of stimulus strength across correlation conditions (see Fig 2c).
The bound height is set by parameter B, which is equal to the amount of evidence that must be accumulated to reach a decision, from a starting point equidistant between the bounds. To maintain equal performance across correlation conditions, the bound height must be scaled to account for the normative weight of evidence. Because the drift rate is related monotonically to the logLR, it is guaranteed that there exists a bound height that satisfies this requirement (Gold and Shadlen, 2001; Green and Swets, 1966). Accordingly, the correlation-specific bound height (Bρ) for correlation ρ was adjusted relative to the bound height for the zero-correlation condition (B0) as:
where is the scale factor that ensures that the bound represents the same amount of accumulated evidence across correlation conditions, in units proportional to the true logLR. This scale factor can be derived analytically (Appendix A in the Supplementary Material).
However, decision-makers like our participants do not necessarily know the objective correlation (ρ) but instead must rely on a subjective internal estimate (). Furthermore, to set the noise under the unit variance assumption of the DDM, like the drift rate, the bound must also be scaled by the relative change in the generative standard deviation across correlation conditions. Therefore, the final correlation-dependent bound height adjustment for correlation ρ was:
Concretely, sets the observer’s subjective decision rule, and sets the noise. The ratio in this formulation makes clear that the normative correlation-dependent scale factor equalizes the weight of evidence needed to make a decision across correlation conditions (i.e., Bρ = B0). Note that the normative scale factor could also be implemented in the DDM as a scaling of both the signal and noise, as in the scaling of the observations in the formula for the logLR. We choose to scale the bound instead both for parsimony and to maintain the typical interpretation of the DDM parameters as assigning perceptual factors and decisional factors to the drift rate and bound height, respectively.
All models also included a non-decision time, ndt, that captures the contributions to RT that are not determined by decision formation (e.g., sensory or motor processing). Therefore, RT for a single simulation of the DDM is given by tS + ndt, where tS is the time at which the bound is reached. Finally, all models included a lapse rate, λ, which mixes the RT distribution determined by the drift-diffusion process with a uniform distribution in proportion to λ (i.e., λ = 0.01 computes the predicted RT distribution as a weighted average of 99% the DDM distribution and 1% a uniform distribution).
To empirically validate the ability of the DDM to account for our data (note that the DDM is the continuous-time equivalent of the discrete-time SPRT, which like the generative process in our task is a random-walk process, and one can be used to approximate the other; Edwards, 1965; Smith, 1990; Bogacz et al., 2006), we fit a basic four-parameter DDM (k0, B0, ndt, λ) to each participant’s data from the zero-correlation condition. These fits could qualitatively account for the data but were improved by the addition of a collapsing bound (Figure 6—figure supplement 1). Therefore, all models in the main analyses included a linear collapsing bound, where parameter tB determined the rate of linear collapse, such that total bound height at time t is 2(B − tBt). Choice commitment occurs when one of the bounds is reached, which happens when |x(t)| ≥ (B − tBt), where x(t) is the value of the decision variable at time t.
To isolate the mechanisms underlying behavioral adjustments to the correlations, we fit four different DDM variants to each participant’s data, jointly to all three correlation conditions. Unless otherwise specified, parameters were shared across conditions. The base model was the same as the five-parameter model described above (k0, B0, tB, ndt, λ). The bound model accounted for correlation-based adjustments using the normative form of the bound scale factor derived above with separate fit subjective correlation parameters for the −ρ and +ρ conditions ( and , respectively) for a total of seven free parameters. The drift model instead accounted for correlation-based adjustments by fitting separate drift rates to each correlation condition (k−, k0, k+); it also had seven parameters. Finally, the bound+drift model included both separate drift parameters and correlation-based bound adjustments, for a total of nine free parameters.
The DDMs were fit to participant’s full empirical RT distributions, using PyDDM (Shinn et al., 2020). Maximum-likelihood optimization was performed using differential evolution (Storn and Price, 1997), a global-optimization algorithm suitable for estimating the parameters of high-dimensional DDMs (Shinn et al., 2020). We also used PyDDM to generate predictions for the expected performance of an observer that uses the normative form of the bound-height adjustment defined above with ρ) chosen to explore different levels of correlation underestimation, where |ρ| = 0.6, and other model parameters were chosen to approximate the average parameters from participants in the 0.6 correlation-magnitude group.
Data analysis
We conducted statistical analyses in Matlab (Mathworks) and R (R Core Team, 2020). We excluded from analysis trials with RTs <0.3 sec or >15 sec, which are indicative of off-task behavior. This procedure removed 0.8% of the data across participants.
To analyze choice behavior, we fit logistic models to each participant’s choices using maximum-likelihood estimation. The basic logistic function was:
where P(R) is the probability that the subject chose the right source, βe determines the slope of the psychometric function as a function of evidence strength (expected logLR), β0 is a fixed offset, and λ is a lapse rate that sets the lower and upper asymptotes of the logistic curve. We fit two models per participant to assess whether choices were dependent on the correlations: 1) a joint model, in which the three free parameters were shared across the three correlation conditions; and 2) a separate model, in which a logistic function was fit separately to each correlation condition (nine free parameters).
To assess whether RTs were affected by the correlations, we fit linear mixed-effects models to median RTs per condition, separately for correct and error trials. The predictors included evidence strength (low, high), correlation condition (−ρ, 0.0, +ρ), and correlation magnitude (0.2, 0.4, 0.6, 0.8), as well as the interaction between evidence strength and correlation magnitude and the interaction between correlation condition and correlation magnitude. Evidence strength and correlation condition were effect coded, and correlation magnitude was z-scored and entered as a continuous covariate. The models were fit using lme4 (Bates et al., 2015b). When possible, we fit the maximal model (i.e., random intercepts for subjects and random slopes for all within-subjects variables). In cases where the maximal model failed to converge or yielded singular fits, we iteratively reduced the random-effects structure until convergence (Bates et al., 2015a). Significance was assessed via ANOVA using Kenward-Roger F-tests with Satterthwaite degrees of freedom, using the car package (Fox and Weisberg, 2019).
To assess the relationship between the objective correlations and the subjective fit correlations, we fit linear mixed-effects models to the Fisher-z-transformed correlations. To quantify the average deviation of the subjective correlation from the objective correlation and the average deviation of the bound scale factors computed with the subjective versus objective correlations, we reversed the signs of the deviations for the negative-correlation conditions so underestimates and overestimates for negative and positive correlations would have the same sign.
Model Comparison
We assessed goodness-of-fit for the logistic and drift-diffusion models using Akaike information criteria (AIC). We also used AIC values in a Bayesian random-effects analysis, which attempts to identify the model among competing alternatives that is most frequent in the population. This analysis produced a protected exceedance probability (PEP) for each model, which is the probability that the model is the most frequent in the population, above and beyond chance (Rigoux et al., 2014b). We computed PEPs using the VBA toolbox (Daunizeau et al., 2014).
References
- Effects of Noise Correlations on Information Encoding and DecodingJ Neurophysiol 95:3633–3644https://doi.org/10.1152/jn.00919.2005
- Acquisition of decision making criteria: reward rate ultimately beats accuracyAtten Percept Psychophys 73:640–657https://doi.org/10.3758/s13414-010-0049-7
- Sequential Tests in Industrial StatisticsSupplement to the Journal of the Royal Statistical Society 8https://doi.org/10.2307/2983610
- Parsimonious Mixed ModelsarXiv https://doi.org/10.48550/arXiv.1506.04967
- Fitting linear mixed-effects models using lme4J Stat Softw 67:1–48https://doi.org/10.18637/jss.v067.i01
- Visual Decisions in the Presence of Measurement and Stimulus CorrelationsNeural Comput 27:2318–2353https://doi.org/10.1162/NECO_a_00778
- The physics of optimal decision making: A formal analysis of models of performance in two-alternative forced-choice tasksPsychol Rev 113:700–765https://doi.org/10.1037/0033-295X.113.4.700
- On the generality of optimal versus objective classifier feedback effects on decision criterion learning in perceptual categorizationMem Cognit 31:181–198https://doi.org/10.3758/BF03194378
- Observing evidence accumulation during multi-alternative decisionsJ Math Psychol 53:453–462https://doi.org/10.1016/j.jmp.2009.09.002
- Aggregation of opinions based on correlated cues and advisorsJ Behav Decis Mak 20:153–177https://doi.org/10.1002/bdm.547
- Measuring and interpreting neuronal correlationsNat Neurosci 14:811–819https://doi.org/10.1038/nn.2842
- VBA: A probabilistic treatment of nonlinear models for neurobiological and behavioural dataPLoS Comput Biol 10https://doi.org/10.1371/journal.pcbi.1003441
- Making Decisions with Unknown Sensory ReliabilityFront Neurosci 6https://doi.org/10.3389/fnins.2012.00075
- Social ConnectivityMedia Bias, and Correlation Neglect. The Economic Journal 131:2033–2057https://doi.org/10.1093/ej/ueaa128
- The Cost of Accumulating Evidence in Perceptual Decision MakingThe Journal of Neuroscience 32:3612–3628https://doi.org/10.1523/JNEUROSCI.4010-11.2012
- A diffusion model decomposition of the practice effectPsychon Bull Rev 16:1026–1036https://doi.org/10.3758/16.6.1026
- Optimal strategies for seeking information: Models for statistics, choice reaction times, and human information processingJ Math Psychol 2:312–329https://doi.org/10.1016/0022-2496(65)90007-6
- Correlation Neglect in Belief FormationRev Econ Stud 86:313–332https://doi.org/10.1093/restud/rdx081
- A Bayesian View on Multimodal Cue IntegrationHuman Body Perception From The Inside Out New York: Oxford University Press New York, NY :105–132https://doi.org/10.1093/oso/9780195178371.003.0006
- Naïve Herding in Rich-Information SettingsAm Econ J Microecon 2:221–243https://doi.org/10.1257/mic.2.4.221
- Correlation Neglect in Portfolio Choice: Lab EvidenceSSRN Electronic Journal https://doi.org/10.2139/ssrn.2914526
- An R Companion to Applied RegressionThousand Oaks, CA: Sage
- Forced-choice staircases with fixed step sizes: asymptotic and small-sample propertiesVision Res 38:1861–1881https://doi.org/10.1016/S0042-6989(97)00340-4
- Visual Perception and the Statistical Properties of Natural ScenesAnnu Rev Psychol 59:167–192https://doi.org/10.1146/annurev.psych.58.110405.085632
- Probabilistic combination of slant information: Weighted averaging and robustness as optimal perceptsJ Vis 9https://doi.org/10.1167/9.9.8
- Extremism and Social LearningJournal of Legal Analysis 1:263–324https://doi.org/10.4159/jla.v1i1.10
- Normative evidence accumulation in unpredictable environmentsElife 4https://doi.org/10.7554/eLife.08825
- The Neural Basis of Decision MakingAnnu Rev Neurosci 30:535–574https://doi.org/10.1146/annurev.neuro.29.051605.113038
- Neural computations that underlie decisions about sensory stimuliTrends Cogn Sci 5:10–16https://doi.org/10.1016/S1364-6613(00)01567-9
- Studies in the History of Probability and Statistics. XXXVII A. M. Turing’s Statistical Work in World War IIBiometrika 66https://doi.org/10.2307/2335677
- Signal detection theory and psychophysicsNew York: John Wiley
- Elapsed decision time affects the weighting of prior probability in a perceptual decision taskJournal of Neuroscience 31:6339–6352https://doi.org/10.1523/JNEUROSCI.5613-10.2011
- The speed-accuracy tradeoff: history, physiology, methodology, and behaviorFront Neurosci 8https://doi.org/10.3389/fnins.2014.00150
- The description–experience gap in risky choiceTrends Cogn Sci 13:517–523https://doi.org/10.1016/j.tics.2009.09.004
- Combining Sensory Information: Mandatory Fusion Within, but Not Between, SensesScience (1979) 298:1627–1630https://doi.org/10.1126/science.1075396
- Belief Formation Under Signal CorrelationSSRN Electronic Journal https://doi.org/10.2139/ssrn.3218152
- The Sign Rule and Beyond: Boundary Effects, Flexibility, and Noise Correlations in Neural Population CodesPLoS Comput Biol 10https://doi.org/10.1371/JOURNAL.PCBI.1003469
- Reward Optimization in the Primate Brain: A Probabilistic Model of Decision Making under UncertaintyPLoS One 8https://doi.org/10.1371/journal.pone.0053344
- On the psychology of predictionPsychol Rev 80:237–251https://doi.org/10.1037/h0034747
- Representation of Confidence Associated with a Decision by Neurons in the Parietal CortexScience (1979) 324:759–764https://doi.org/10.1126/science.1169405
- Information theory of choice-reaction timesNew York: Academic Press
- How to Alleviate Correlation Neglect in Investment DecisionsManage Sci 69:3400–3414https://doi.org/10.1287/mnsc.2022.4535
- Persuasion with Correlation Neglect: A Full Manipulation ResultAm Econ Rev Insights 4:123–138https://doi.org/10.1257/aeri.20210007
- An experimental examination of subjective forecast combinationInt J Forecast 12:223–233https://doi.org/10.1016/0169-2070(95)00623-0
- The Effect of Forecast Redundancy on Judgments of a Consensus Forecast’s Expected AccuracyJournal of Accounting Research 28https://doi.org/10.2307/2491245
- Time-varying decision boundaries: insights from optimality analysisPsychon Bull Rev 25:971–996https://doi.org/10.3758/s13423-017-1340-6
- Overcoming indecision by changing the decision boundaryJ Exp Psychol Gen 146:776–805https://doi.org/10.1037/xge0000286
- Optimal decision making in heterogeneous and biased environmentsPsychon Bull Rev 22:38–53https://doi.org/10.3758/s13423-014-0669-3
- Information-limiting correlationsNat Neurosci 17:1410–1417https://doi.org/10.1038/nn.3807
- Perceptual Inference, Learning, and Attention in a Multisensory WorldAnnu Rev Neurosci 44:449–473https://doi.org/10.1146/annurev-neuro-100120-085519
- Overconfidence in Political BehaviorAmerican Economic Review 105:504–535https://doi.org/10.1257/aer.20130921
- Weighted linear cue combination with possibly correlated errorVision Res 43:2451–2468https://doi.org/10.1016/S0042-6989(03)00435-8
- Some task demands induce collapsing bounds: Evidence from a behavioral analysisPsychon Bull Rev 25:1225–1248https://doi.org/10.3758/s13423-018-1479-9
- The effect of stimulus strength on the speed and accuracy of a perceptual decisionJ Vis 5:1–1https://doi.org/10.1167/5.5.1
- Crossmodal Correspondences: Standing Issues and Experimental GuidelinesMultisens Res 29:7–28https://doi.org/10.1163/22134808-00002502
- PsychoPy2: Experiments in behavior made easyBehav Res Methods 51:195–203https://doi.org/10.3758/S13428-018-01193-Y/FIGURES/3
- R: A language and environment for statistical computingVienna, Austria: R Foundation for Statistical Computing
- The diffusion decision model: theory and data for two-choice decision tasksNeural Comput 20:873–922https://doi.org/10.1162/neco.2008.12-06-420
- Diffusion Decision Model: Current Issues and HistoryTrends Cogn Sci 20:260–281https://doi.org/10.1016/j.tics.2016.01.007
- Bayesian model selection for group studies - RevisitedNeuroimage 84:971–985https://doi.org/10.1016/j.neuroimage.2013.08.065
- Bayesian model selection for group studies - RevisitedNeuroimage 84:971–985https://doi.org/10.1016/j.neuroimage.2013.08.065
- Texture and object motion in slant discrimination: Failure of reliability-based weighting of cues may be evidence for strong fusionJ Vis 7https://doi.org/10.1167/7.6.3
- Recipe for Disaster: The Formula That Killed Wall StreetWired
- The Speed and Accuracy of a Simple Perceptual Decision: A Mathematical PrimerBayesian Brain: Probabilistic Approaches to Neural Coding The MIT Press :208–237https://doi.org/10.7551/mitpress/9780262042383.003.0010
- A flexible framework for simulating and fitting generalized drift-diffusion modelsElife 9:1–27https://doi.org/10.7554/eLife.56938
- A note on the distribution of response times for a random walk with Gaussian incrementsJ Math Psychol 34:445–459https://doi.org/10.1016/0022-2496(90)90023-3
- Psychology and neurobiology of simple decisionsTrends Neurosci 27:161–168https://doi.org/10.1016/j.tins.2004.01.006
- Age-related differences in diffusion model boundary optimality with both trial-limited and time-limited tasksPsychon Bull Rev 19:139–145https://doi.org/10.3758/s13423-011-0189-3
- Models for choice-reaction timePsychometrika 25:251–260https://doi.org/10.1007/BF02289729
- Differential Evolution - A Simple and Efficient Heuristic for Global Optimization over Continuous SpacesJournal of Global Optimization 11:341–359https://doi.org/10.1023/A:1008202821328
- Stochastic Models of Evidence Accumulation in Changing EnvironmentsSIAM Review 58:264–289https://doi.org/10.1137/15M1028443
- Sequential analysisOxford, England: John Wiley
- Optimum Character of the Sequential Probability Ratio TestThe Annals of Mathematical Statistics 19:326–339https://doi.org/10.1214/aoms/1177730197
- Robust versus optimal strategies for two-alternative forced choice tasksJ Math Psychol 54:230–246https://doi.org/10.1016/j.jmp.2009.12.004
Article and author information
Author information
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
Copyright
© 2024, Tardiff et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 161
- download
- 1
- citations
- 0
Views, downloads and citations are aggregated across all versions of this paper published by eLife.