Neural evidence accumulation persists after choice to inform metacognitive judgments
Abstract
The ability to revise one’s certainty or confidence in a preceding choice is a critical feature of adaptive decision-making but the neural mechanisms underpinning this metacognitive process have yet to be characterized. In the present study, we demonstrate that the same build-to-threshold decision variable signal that triggers an initial choice continues to evolve after commitment, and determines the timing and accuracy of self-initiated error detection reports by selectively representing accumulated evidence that the preceding choice was incorrect. We also show that a peri-choice signal generated in medial frontal cortex provides a source of input to this post-decision accumulation process, indicating that metacognitive judgments are not solely based on the accumulation of feedforward sensory evidence. These findings impart novel insights into the generative mechanisms of metacognition.
https://doi.org/10.7554/eLife.11946.001eLife digest
Reflecting on our previous choices and accurately representing our confidence in their accuracy allows us to detect, correct and learn from our errors. Yet, it remains poorly understood how such “metacognition”, or thoughts about thoughts, occurs in the human brain. In particular, a long-standing debate in this area of research concerns whether metacognitive processes in the brain occur at the same time as those that determine the actual choice, or whether they develop after the choice has been made and rely on different information.
Now, Murphy et al. have recorded brain activity in human volunteers who were carrying out a simple task in order to explore metacognition. In short, the volunteers looked at colored words and decided if each word matched its color (e.g., is the word ‘RED’ also written in a red font?). At the same time, the volunteers chose whether or not to press a button depending on the specific color/word combination shown, and most importantly reported whenever they noticed that they made an error in the task.
This approach allowed Murphy et al. to chart the development of choices and detection of errors as they occurred in the volunteers’ brains. This revealed that the metacognitive judgement about each choice relied on information that was gathered after the point the initial choice was made. Further analysis then suggested that this process relies, at least in part, on a signal generated in a region at the front of the brain.
Together, these findings suggest that metacognitive decisions rely on processes that are similar to those behind other decisions, but with a few important differences. Namely, the metacognitive process plays out at a different point in time, and likely incorporates distinct sources of information. Further work should aim to clarify the nature of these sources of information and describe their specific contributions to the process.
https://doi.org/10.7554/eLife.11946.002Introduction
The ability to detect errors is an essential feature of adaptive behavior, providing the basis for adjusting or countermanding ongoing actions and optimizing future decision-making (David et al., 2012; Fernandez-Duque et al., 2000; Fleming et al., 2012). Establishing the neurocomputational principles underpinning this key metacognitive function is therefore a major imperative. Categorical choices are thought to be made by integrating evidence over time into a decision variable that triggers action upon reaching a criterion (Gold and Shadlen, 2007; Kelly and O'Connell, 2014; Shadlen and Kiani, 2013; Smith and Ratcliff, 2004). Most theoretical models of metacognition propose that the same decision variable makes a key contribution to internal representations of choice accuracy (de Martino et al., 2013; Heath, 1984; Kiani et al., 2014; Kiani and Shadlen, 2009; Link, 2003; Moran et al., 2015; Pleskac and Busemeyer, 2010; Ratcliff and Starns, 2013; Yu et al., 2015). However, there is considerable uncertainty regarding the precise nature of this contribution.
Initial efforts to model metacognitive performance centered on the proposition that our confidence in a choice is based on a read-out of the level the decision variable has reached at the time of choice commitment (Heath, 1984; Kiani et al., 2014; Kiani and Shadlen, 2009; Link, 2003). However, the specification that the decision variable reaches a fixed threshold prior to commitment means that these models cannot account for the fact that human participants can retrospectively categorize certain choices as erroneous even in the absence of external feedback (Rabbitt and Vyas, 1981; Rabbitt, 1966; Yeung et al., 2004). To take account of this kind of observation, alternative theoretical models have been proposed in which metacognitive judgments can exploit additional evidence that is accumulated after first-order commitment (Moran et al., 2015; Pleskac and Busemeyer, 2010; Yu et al., 2015). However, a definitive neurophysiological demonstration of post-decisional evidence accumulation has yet to be definitively provided in humans or other animals. Moreover, it is unclear whether such a process would take the form of a simple continuation of first-order evidence gathering (Moran et al., 2015; Pleskac and Busemeyer, 2010; Resulaj et al., 2009; Yu et al., 2015) or might also be crucially dependent on higher-order representations of error likelihood (Yeung et al., 2004).
While the last two decades have seen intensive research on the neural signatures of decision formation in the non-human primate (Gold and Shadlen, 2007; Shadlen and Kiani, 2013), these questions have proven difficult to address because the pre-motor neurons that have been the focus of much of this work (e.g. in area LIP) fall silent upon initiation of the decision-reporting action, thus precluding measurement of post-decision activity and verification of its potential influence on metacognition. Although post-commitment neural signatures have been described in humans (Falkenstein et al., 1990; Gehring et al., 1993) and other animals (Ito et al., 2003; Narayanan et al., 2013; Pardo-Vazquez et al., 2008) which show clear sensitivity to choice accuracy and, in some cases, to the quality of metacognitive judgments (Boldt and Yeung, 2015; Nieuwenhuis et al., 2001; O'Connell et al., 2007; Steinhauser and Yeung, 2010), these signals have yet to be conclusively identified with post-decisional evidence integration. In the present study, we exploited the uniquely supramodal nature of a recently characterized build-to-threshold decision variable signal in the human brain (Kelly and O'Connell, 2013; O'Connell et al., 2012; Twomey et al., 2015) to investigate the influence of post-decisional evidence accumulation on the timing and accuracy of explicit error detection. Through a combination of electrophysiological data analysis and computational modeling, we demonstrate that neural evidence accumulation does persist after commitment to the first-order decision and that its rate determines the probability and timing of error detection. Additionally, we show that the rate of post-decision accumulation is not solely the product of feedforward sensory inputs but is influenced by the output of medial frontal structures implicated in performance monitoring and executive control.
Results
Task behavior
We analyzed 64-channel electroencephalographic (EEG) data [originally collected as part of Murphy et al. (2012); see Materials and methods], from 28 human subjects performing a Go/No-Go response inhibition task (Hester et al., 2005; Figure 1a). Subjects viewed a serial sequence of color words, each presented for 0.4 s, with the congruency between font color and semantic content varied across trials. The primary task was to execute a right-handed button press as quickly as possible when the semantic content of the word and its font color were incongruent (Go trial), and to withhold this response when either the word presented on the current trial was the same as that presented on the previous trial (‘repeat’ No-Go) or when the meaning of the word and its font color matched (‘color’ No-Go). Performance on paradigms of this nature can be readily understood in terms of a race-to-threshold between two competing accumulation processes representing the evidence in favor of a ‘Go’ or ‘Don’t Go’ decision (Gomez et al., 2007; Logan and Cowan, 1984).

Go/No-Go task and associated behavior.
(a) Subjects’ primary task was to make a speeded manual response (‘A’) to all incongruent color/word stimuli and to withhold from responding to congruent stimuli or when the same word was presented on consecutive trials. Following any commission errors, they were instructed to signal error detection as quickly as possible by pressing a secondary response button (‘B’). (b) Distribution of trial types averaged across subjects. C = correct withhold, E = error, D = detected error, U = undetected error. Error bars = s.e.m. (c) Histograms representing stimulus onset and detection response time (RTd) distributions on detected error trials, aligned to error commission and pooled across all subjects.
If subjects failed to withhold the pre-potent response to either type of No-Go stimulus, they were instructed to signal detection of this error as quickly as possible by pressing a secondary response button. The timing of this error detection report was measured relative to the initial erroneous action (detection response time; RTd). There were no significant differences in primary RT, RTd or electrophysiological signal morphology between repeat and color No-Go trials (Figure 1—figure supplement 1); therefore, we collapsed across both No-Go trial-types in all analyses.
Subjects successfully withheld from responding on an average of 56.6% (±13.0) of No-Go trials, and 68.0% (±16.7) of erroneous presses were followed by an error detection report (Figure 1b). Primary RT and RTd were not correlated across detected error trials (mean within-subject β = −0.05, s.e.m. = 0.04; t27 = −1.4, p = 0.2). The median primary RTs for detected errors (457 ± 21 ms) were faster than correct Go RTs (511 ± 22 ms; t27 = 5.5, p < 1 x 10−4), whereas undetected errors were significantly slower than Go RTs (543 ± 28 ms; t27 = 2.9, p = 0.007; detected vs. undetected errors: t27 = 5.8, p < 1 x 10−4).
A centro-parietal signature of first- and second-order evidence accumulation
A series of recent studies of perceptual decision making have established that a decision variable signal can be isolated in the human event-related potential (ERP) over centro-parietal scalp sites (Kelly and O'Connell, 2013; O'Connell et al., 2012; Twomey et al., 2015). This signal exhibits the same decision-predictive dynamics that have been reported in single-unit recordings from a variety of brain areas during perceptual decision formation (Gold and Shadlen, 2007; Shadlen and Kiani, 2013), including an evidence-dependent rate of rise and a threshold-crossing relationship with reaction time. Another important feature of this signal is that it represents the evolving decision in a domain-general fashion that is independent of motor requirements and indeed traces the decision even when no overt decision-reporting action is required (O'Connell et al., 2012). Here, we observed that the same centro-parietal positivity (CPP) was elicited by Go and No-Go stimuli (Figure 2—figure supplement 1). In what follows, we demonstrate that this signal encodes consecutive build-to-threshold processes that determine both first-order choices and second-order error detection decisions on our task. Readers familiar with the human ERP literature will note that these two processing stages incorporate stimulus- and response-evoked activity usually attributed to the ‘P300’ and ‘Pe’ components, respectively (see Discussion). We persist here with the label ‘CPP’ because it is less prescriptive about signal latency, eliciting conditions or measurement technique.
To probe the dynamics of the CPP and its relationship to the timing of the first-order decision process, we split each subject’s Go-trial RT distribution into equal-sized fast, medium and slow bins and plotted the average waveforms aligned to action execution for each bin. Consistent with previous observations (Kelly and O'Connell, 2013; O'Connell et al., 2012; Twomey et al., 2015), the CPP exhibited a gradual build-up with a rate that was inversely proportional to RT, and reached a stereotyped amplitude at the time of first-order choice commitment (Figure 2a). Thus, first-order performance on the Go/No-Go task was reliant on the same fundamental neural dynamics as have been reported for conventional perceptual decision-making paradigms (Kelly and O'Connell, 2013).

A centro-parietal decision signal for first- and second-order decision-making.
(a) Go-trial CPP waveforms aligned to primary task response and sorted by RT into three equal-sized bins, and associated scalp topography. (b) CPPs again aligned to primary response, but separately for Go trials and detected and undetected errors; gray running marker indicates significant detection effect (p < 0.05, paired t-test for detected vs. undetected). Topography illustrates scalp distribution of error detection effect. (c) Time-course of detection-predictive activity estimated as the area under the ROC curve. Permutation mean and significance threshold (1.96 s.d.) are marked as solid and dashed gray lines, respectively. (d) Detected-error CPP, aligned to primary task response and subsequent error detection report; waveforms were sorted and binned by RTd. (e) Single-trial surface plot showing temporal relationship between the CPP and RTd (curved black line); waveforms were pooled across subjects, sorted by RTd and smoothed over bins of 50 trials with Gaussian-weighted moving average. Vertical dashed lines in a and d represent median RTs. Gray markers at bottom of these plots indicate time points when linear regression of RT on signal amplitude reached significance (p < 0.05); black markers indicate center of 150 ms time windows in which regression of RT on signal slope reached significance (p < 0.05; one-tailed predicting steeper slope for faster RTs). Shaded gray areas show latencies of all associated scalp topographies. All traces were baselined to pre-stimulus period. Shaded error bars = s.e.m.
Despite the fact that the median RTd was executed 560 ms after stimulus offset, precluding the continued integration of feedforward sensory evidence in the period preceding error detection, the CPP exhibited persistent post-commitment build-up on a subset of trials, continuing its positive trajectory prior to an error detection report but gradually diminishing in amplitude following Go decisions and undetected errors (Figure 2b; see Figure 3—figure supplement 1 for explicit comparison of signals on Go and undetected error trials). Receiver operating characteristic (ROC) curve analysis (see Materials and methods) conducted in discrete temporal windows along the entire signal time course revealed that second-order performance could be reliably classified as early as 120 ms prior to error commission (Figure 2c).
We next sought to characterize the relationship between the CPP and the timing of error detection by splitting each subject’s RTd distribution into three equal-sized bins and plotting the bin-averaged waveforms aligned to both error commission and the error detection report. Mirroring our observations for the first-order responses, the build-up rate of the second-order CPP was steeper on trials characterized by faster error detection, and it again reached a fixed amplitude immediately prior to the error detection report (Figure 2d,e).
Taken together, these findings indicate that subjects engaged in evidence accumulation even after the primary task response had been executed and leveraged the new information gained by this process to make judgments about the accuracy of their preceding choices. However, CPP dynamics during the post-commitment interval were qualitatively distinct from those observed during the first-order decision process. While the positive build-up of the first-order CPP was invariant to trial-type, tracing the emerging decision irrespective of whether the evidence favored a Go or No-Go choice (Figure 2—figure supplement 1; see also Kelly and O’Connell, 2013), the second-order CPP only increased after detected errors and not, on average, after correct Go responses or undetected errors. Similarly, in a recent study that examined the post-commitment ‘Pe’ signal preceding graded confidence judgments (without interrogating this signal for accumulation-to-bound characteristics), a monotonic relationship between signal amplitude and confidence was observed whereby amplitude was greatest when subjects indicated very low confidence in the primary choice (Boldt and Yeung, 2015). Thus, rather than reflecting a continuation of the first-order decision process, these data suggest that error detection decisions were based on the selective accumulation of internal evidence that the previous choice was incorrect. In the following set of analyses, we investigated whether a higher-order neural signal that has known sensitivity to choice accuracy might influence the error detection process.
Fronto-central theta oscillations and error detection behavior
A substantial literature incorporating several species and modes of neurophysiological measurement has established that the posterior medial frontal cortex (pMFC) is highly responsive to variations in performance accuracy (Carter et al., 1998; Ito et al., 2003; Narayanan et al., 2013; Ridderinkhof et al., 2004) and its activity predicts neural and behavioral adaptation following error commission (Cavanagh et al., 2009; Danielmeier et al., 2011; Ebitz and Platt, 2015; Narayanan et al., 2013; Sheth et al., 2012). These characteristics are thought to reflect the pMFC’s role in coordinating the activity of task-relevant regions in the presence of increasing conflict between incompatible actions (Botvinick et al., 2001; Carter and van Veen, 2007; Cavanagh et al., 2009; Kerns et al., 2004), while the conflict signal generated in this brain region has also been proposed to provide a reliable basis for subsequent error detection (Yeung et al., 2004). Accordingly, we investigated the influence of pMFC on error detection decisions by interrogating a prominent oscillatory signature that provides a proxy for pMFC activity, fronto-central theta power (FCθ; 2–7 Hz; Cavanagh and Frank, 2014; Cavanagh et al., 2011; Cohen, 2014; Cohen and Donner, 2013; Narayanan, et al., 2013).
Consistent with previous reports (Cavanagh et al., 2009; Cohen and Donner, 2013; Narayanan et al., 2013), we observed a clear increase in FCθ power following stimulus onset that peaked after primary response execution (Figure 3a). The time-course of FCθ did not distinguish between correct Go responses and undetected errors (Figure 3—figure supplement 1) but underwent a sharp positive deflection at approximately the time of errors that were subsequently detected (Figure 3b). ROC analysis applied to single-trial FCθ waveforms revealed strong detection-predictive activity that, as with the CPP, achieved significant detection classification up to 120 ms before initial error commission (Figure 3c). This sensitivity to error detection was not apparent in the error-related negativity, a commonly investigated error-evoked ERP over fronto-central scalp (Falkenstein et al., 1990; Gehring et al., 1993; Figure 3—figure supplement 2), and was not due to condition-related differences in primary RT (Figure 3—figure supplement 3a).

Fronto-central θ-band (2–7 Hz) oscillatory power predicts the accuracy and timing of error detection reports.
(a) Time-frequency plot of fronto-central power, aligned to the primary task response and averaged across detected error, undetected error and RT-matched go trials; black lines enclose regions of significant power change relative to a pre-stimulus baseline (p < 0.01, paired t-test). Scalp topography shows θ power averaged over all trial types. (b) Response-aligned FCθ waveforms separately for each trial type; gray trace is the condition-averaged ERP (arbitrarily scaled). Topography illustrates scalp distribution of error detection effect. (c) Time-course of FCθ detection-predictive activity quantified by the area under the ROC curve. (d) Go-trial FCθ waveforms and associated scalp topography, aligned to primary task response; single-trial waveforms were sorted by primary RT and divided three equal-sized bins. (e) Detected-error FCθ waveforms and topographies, aligned to both the primary task response (left) and the subsequent error detection report (right); waveforms were sorted and binned by RTd. (f) Single-trial surface plot showing the temporal relationship between the FCθ power and RTd (curved black line). Conventions for b–f are the same as in Figure 2. All traces were again baselined to the pre-stimulus period. Plots in a were calculated via wavelet convolution; all other plots show filter-Hilbert transformed data (see Materials and methods).
We repeated the trial-binning analyses that were previously applied to the CPP, this time to explore the relationship between the FCθ signal and the timing of first- and second-order decision-making. The amplitude of the FCθ signal consistently distinguished between RT bins throughout both decision intervals (Figure 3d,e,f). In keeping with previous observations (Cavanagh et al., 2011; Cohen and Donner, 2013), pre-response FCθ power was consistently lower on trials with faster first-order RTs (Figure 3d). A reliable relationship also existed between pre-detection FCθ on detected errors and RTd, but here fast error detections were preceded by greater FCθ power (Figure 3e,f). The latter observation is consistent with the notion that peri-response pMFC activity serves as a form of input to the second-order decision process and thereby influences the probability and timing of error detection (Yeung et al., 2004). Next, we further explored this possibility by examining the trial-by-trial prediction of RTd by different features of the FCθ and CPP signals.
Signal interactions predict the timing of error detection
We quantitatively compared the independent contributions of the FCθ and CPP signals to variation in RTd via single-trial, within-subjects robust regressions, leveraging single-trial amplitudes, build-up rates and peak latencies of the signals as predictors in successive models (see Materials and methods; see Figure 4 for measurement approach). As expected, single-trial FCθ power was strongly negatively related to RTd (t27 = –7.2, p < 1 x 10−6). Topographically, this effect was maximal over fronto-central scalp (Figure 4a). In contrast, the pre-detection amplitude of the second-order CPP did not reliably correlate with RTd (t27 = −1.9, p = 0.08; Figure 4—figure supplement 1a), consistent with the aforementioned observation of a threshold-crossing effect prior to error detection. Statistical comparison of the associated regression weights indicated that FCθ power was a better predictor of RTd than CPP amplitude (t27 = 3.3, p = 0.003). The opposite pattern was apparent for the build-up rates and peak latencies of both signals: CPP build-up rate (t27 = -−5.0, p < 1 x 10−4) and latency (t27 = 4.9, p < 1 x 10–4, derived by permutation testing; see Materials and methods) robustly predicted RTd with both effects maximal over centro-parietal scalp (Figure 4b,c), whereas these features of the FCθ signal did not reliably account for trial-by-trial variance in RTd (build-up rate: t27 = –1.9, p = 0.07; latency: t27 = –0.7, p = 0.5; Figure 4—figure supplement 1b,c). Thus, although a relationship between FCθ build-up rate and the timing of error detection was clearly observed in the trial-averaged waveforms (Figure 3e), single-trial regressions revealed that this effect was only marginally significant when the contribution of CPP build-up rate was accounted for. Formal comparisons of the regression weights confirmed that CPP build-up rate and peak latency were superior predictors of RTd compared to the counterpart metrics derived from FCθ (build-up rate: t27 = 3.0, p=0.006; latency: t27 = 4.3, p = 0.0002).

Variance in the timing of error detection reports is explained by distinct single-trial metrics of FCθ and CPP morphology.
(a) Schematic depicting the single-trial measurement windows for FCθ power (purple) and CPP amplitude (green; left). Bar graphs (middle) show estimated effect sizes from within-subjects robust multiple regressions of RTd on both amplitude metrics (error bars = s.e.m.), and the associated topography (inset right) indicates the scalp distribution of the theta power effect. Scatterplot (right) illustrates the linear relationship between FCθ power and RTd; points and error bars are mean ± s.e.m. of data that were z-scored within subjects, pooled across subjects and grouped into 20 five-percentile bins. (b) Measurement approach, regression-estimated effect sizes, topographic distribution and scatterplot representing the single-trial relationships between RTd and signal build-up rates. (c) Similar plots representing the relationships between RTd and peak signal latencies. Points in the effect size plot (middle) indicate observed effect sizes, solid and dashed lines highlight the mean and 95% confidence intervals, respectively, of permuted distributions used for significance testing (see Materials and methods). Topographies are thresholded at p < 0.005 (a, b) and p < 0.0001 (c) for effect visualization. The solid lines in all scatterplots are simple linear robust regression fits to the unbinned data. **p < 0.01; ***p < 0.001.
Two additional features of our data point to an active role for FCθ in the error detection process. First, FCθ power began to reliably predict the speed of error detection approximately 350 ms before the CPP did (compare Figures 2d and 3c). Thus, internal fluctuations in FCθ power that were independent of the first-order evidence accumulation process reflected in the CPP were predictive of subsequent error detection behavior. Second, we tested whether the single-trial relationship between FCθ and RTd (Figure 4a) was formally mediated by a direct effect of FCθ on the rate of second-order evidence accumulation by constructing a three-variable path model with FCθ power as the predictor, RTd as the outcome and CPP build-up rate as the mediator variable (see Materials and methods). FCθ was a reliable predictor of CPP build-up rate in this model (p = 0.0007) and the mediation effect was significant (p = 0.0009), indicating that the rate of second-order evidence accumulation partially mediated the observed relationship between medial frontal signaling and the speed of error detection (Figure 5).

Mediation analysis.
Path diagram depicts the relationships between nodes in a mediation model that tested whether CPP build-up rate mediated the negative relationship between FCθ power and RTd. Lines are labeled with path coefficients, with s.e.m. shown in parentheses. FCθ power (predictor, left) predicted CPP build-up rate (mediator, middle), which in turn predicted RTd (outcome, right) controlling for FCθ power. The upper-middle coefficient indicates the formal mediation effect. The significant direct path between FCθ power and RTd, calculated controlling for the mediator, indicates partial mediation: CPP build-up rate did not explain all of the shared variance between FCθ and RTd. ***p < 0.001, bootstrapped.
Drift diffusion modelling of the second-order decision process
Informed by the above observations, we modeled error detection as a one-choice diffusion process (Ratcliff and Van Dongen, 2011) by which noisy evidence that an error has been committed is accumulated over time (at mean drift rate v with between-trial standard deviation η). Error detection is achieved in the model once the evidence tally passes a threshold (a) whereas the temporal integration process terminates if this threshold is not reached by a time deadline, thereby resulting in an undetected error (Figure 6a; Materials and methods). The decision to model second-order performance in isolation was based on the observation that higher-order signals (FCθ) play a prominent role in determining second-order decision-making, thus indicating that the input to the primary decision process is not necessarily the same as the input to the second-order process (see Discussion). For simplicity, we here focused on decomposing error detection behavior in isolation. This approach served to further identify our electrophysiological signals with distinct features of post-commitment evidence accumulation (evidential input, temporal integration) without requiring speculative assumptions about the relationship between first- and second-order decision processes.

Diffusion modelling of error detection behavior.
(a) Schematic representation of the one-choice drift diffusion model. Noisy error evidence accumulates over time at mean drift rate v until a threshold a is reached (light red traces) or a deadline on detection expires (gray traces). Drift rate is normally distributed across trials with standard deviation η, and non-decision-related processing time is captured by tnd. Within-trial noise s is fixed at 0.1 (not shown). (b) Group-level model fit. Points from left to right represent the 0.1, 0.3, 0.5, 0.7 and 0.9 RTd quantiles, estimated from the data (black) and generated by the model fit (gray). (c) Observed versus fitted error detection accuracy. Diagonal line is the identity line. (d) Group-average decision variables reflecting the accumulation of error evidence, simulated using the best-fitting model parameters for each subject and overlaid on the grand-average second-order CPP signals. Detected error traces are sorted by RTd into three equal-sized bins. CPPs are baselined to the 50 ms preceding error. (e) Between-subjects relationship between FCθ power and model-estimated drift ratio (v/η). Bar graph (inset) shows standardized regression coefficients from a multiple regression that included CPP amplitude as an additional predictor. (f, g) Scatterplots and bar graphs highlighting between-subjects relationships between drift ratio and signal build-up rates (f) and peak latencies (g). Although one extreme data point is apparent in f, this did not exert disproportionate influence over the fitted regression line (studentized deleted residual < 1) and the correlation remained marginally significant when it was removed (r = 0.37, p = 0.06). Error bars = s.e.m. *p < 0.05; **p < 0.01.
The model fit the error detection data well, capturing the shapes of the group-level and single-subject RTd distributions (Figure 6b; Table 1; Figure 6—figure supplement 1) as well as the considerable heterogeneity in individuals’ capacities for accurate second-order evaluation (Figure 6c). We then used the best-fitting model parameters for each subject to generate simulated second-order decision variable trajectories for detected and undetected errors. Despite minimal data-driven constraint on the simulation process (see Materials and methods), the temporal evolution of the simulated time-series closely traced that of the second-order CPP (Figure 6d). By allowing a proportion of trials to assume negative drift rate, the model additionally provides a parsimonious account of the apparent downward trajectory of the CPP in the averaged waveforms after undetected errors.
Parameter estimates and goodness-of-fit of error detection diffusion model.
a | tnd | v | η | χ2 | |
---|---|---|---|---|---|
Mean | 0.21 | 0.20 | 0.45 | 0.45 | 2.46 |
s.d. | 0.12 | 0.08 | 0.30 | 0.30 | 2.33 |
-
χ2 degrees of freedom = 1, critical value = 5.024.
-
24 of 28 χ2 values were below the critical value.
The fitted model parameters were also correlated with key FCθ and CPP signal characteristics across subjects. We employed the per-subject ‘drift ratio’ (v/η) as a model-based estimate of each individual’s error evidence strength (Ratcliff and Van Dongen, 2011; Materials and methods) and found that this quantity was positively correlated with FCθ power (r = 0.51, p = 0.007; Figure 6e). A partial correlation analysis verified that this effect was still present (r = 0.52, p = 0.007) when θ power over bilateral posterior electrodes (P7/P8) was included as a covariate, indicating that it is not driven by spurious global differences in oscillatory power (Cohen and Donner, 2013). Moreover, a multiple regression to quantify the independent contributions of FCθ and second-order CPP amplitude to variance in drift ratio revealed a significant effect only for the former (βFCθ = 0.47, p = 0.013; βCPP = 0.21, p = 0.24). Conversely, drift ratio was correlated with the build-up rate (r = 0.47, p = 0.013; Figure 6f) and peak latency (r = −0.52, p = 0.005; Figure 6g) of the second-order CPP, whereas no such relationships were observed for FCθ build-up rate (βFCθ = −0.12, p = 0.53; βCPP = 0.46, p = 0.017) or latency (βFCθ = −0.10, p = 0.64; βCPP = −0.46, p = 0.042). We also note that neither FCθ power nor the build-up rate nor peak latency of the second-order CPP were correlated across subjects with primary task behavior (withhold accuracy and primary RT on detected error trials; all p > 0.1), thus highlighting the particular sensitivity of these metrics to second-order decision-making. Additionally, these electrophysiological measures were not correlated with other parameters of the computational model (all p > 0.1).
Second-order decision signals are independent of error reporting requirements
Several lines of evidence indicate that the close relationship between the second-order CPP and error detection signalling cannot be attributed to motor preparation or execution. First, we have previously shown that the CPP that precedes a first-order perceptual decision is fully dissociable from motor preparation signals and is observed even when no overt decision-reporting action is required (O'Connell et al., 2012). Second, we observed no change in signal topography between the pre- and post-choice phases (Figure 2d). Third, the temporally extended and variable build-up of the second-order CPP excludes the possibility that our results can be attributed to the presence of overlapping motor execution potentials. Finally, we asked a new cohort of 12 subjects to perform the same Go/No-Go task but gave no instructions to report errors on half of the task blocks. FCθ and the CPP were clearly observed following these ‘no-report’ errors and, consistent with the fact that no-report blocks contain a mixture of detected and undetected errors, linear contrasts confirmed that FCθ and second-order CPP amplitudes were intermediate between their amplitudes following detected and undetected errors in the condition with self-initiated error reporting (FCθ, t11 = 4.2, p = 0.002; CPP, t11 = 6.9, p < 1 x 10–4; Figure 7).

Second-order decision signals persist in the absence of error reporting demands.
(a) Time-courses and topographic distributions of FCθ power from a new cohort that performed half of task blocks without any explicit instruction to self-monitor performance (‘no-report’ blocks). Left topography is the average topographic distribution of θ-power after pooling all detected and undetected error trials from blocks with self-initiated error detection reporting; right topography is θ distribution averaged across all errors in no-report blocks. (b) Time-courses and topographies of the second-order CPP component from the same cohort. Conventions are the same as in (a). Shaded gray areas in left show latencies of associated scalp topographies. Shaded error bars = s.e.m.
Discussion
Our electrophysiological and model-based analyses demonstrate that neural evidence accumulation continues after decision commitment to facilitate reflections on choice accuracy. This empirical observation has important implications for the ongoing debate on the mechanistic basis of metacognition.
A prominent hypothesis is that confidence judgments are based on internal information available at the time of choice commitment (Heath, 1984; Kiani et al., 2014; Kiani and Shadlen, 2009; Link, 2003), and models that incorporate this assumption provide a good account of behavior when choice and confidence are interrogated simultaneously. But the manner in which metacognitive evaluations are probed is likely to have a profound impact on how they are constructed in the brain. Whether participating in a laboratory experiment or everyday activity, human decision-makers are most often required to express or act upon their metacognitive evaluations some time after an initial choice. Our results reveal that if the decision-maker is allowed to report delayed, self-initiated judgments of their own performance, these judgments will incorporate new internal information available after the point of commitment, even in the absence of continued sensory input. This process of post-decision evidence accumulation has already been invoked by several computational models to successfully account for delayed confidence judgments and changes-of-mind (Moran et al., 2015; Pleskac and Busemeyer, 2010; Resulaj et al., 2009; Yu et al., 2015), while a parallel literature on error detection has long debated the mechanistic basis of post-commitment information processing (Rabbitt and Vyas, 1981; Rabbitt, 1966; Yeung et al., 2004; Yeung and Summerfield, 2012). Our findings, however, represent the first definitive neurophysiological demonstration of post-decision evidence accumulation. In so doing, we show that the critical neural dynamics that give rise to both first- and second-order decisions are captured by a single brain signal, opening up new avenues for basic and clinical investigations.
Comparison of pre- and post-decisional CPP dynamics highlighted an important qualitative distinction between the first- and second-order decision processes invoked by our paradigm. The first-order task required subjects to discriminate between the two possible stimulus categories (i.e. Go versus No-Go), a process that can be understood in terms of competitive accumulation of evidence in favor of each choice alternative (Gold and Shadlen, 2007; Gomez et al., 2007). We have previously shown that the CPP builds in response to evidence favoring each of the choice alternatives in such contexts (Kelly and O’Connell, 2013; O’Connell et al., 2012) and accordingly, we observed here that the same signal exhibited a gradual build-up prior to both Go and No-go choices. The second-order task, by contrast, was more comparable to simple signal detection with detected errors translating to hits and undetected errors translating to misses. Evidence accumulation mechanisms have also been invoked to account for signal detection decisions (e.g. Deco et al., 2007; Donner et al., 2009) but in this case the neural decision variable seems to increase towards a single detection boundary and misses occur when this boundary is not reached (Carnevale et al., 2012; Deco et al., 2007). Again, previous studies have shown that the CPP in such contexts exhibits a clear ramp-up to threshold for hit decisions that is diminished or absent on miss trials (O’Connell et al., 2012), and presently this effect was recapitulated for error detection decisions. Collectively, these findings indicate that human decision-makers are capable of quickly engaging multiple decision processes in succession that, although reliant on the same neural mechanism of evidence integration, are constructed in qualitatively distinct ways.
Our results also offer important new insights into the nature of the input to the metacognitive decision process. Several features of our results mark pMFC activation, as indexed by the power of FCθ oscillations, as a likely candidate for furnishing a source of this input. First, peri-response FCθ power was strongly predictive of subsequent error detection. Second, FCθ power predicted error detection speed from a particularly early latency relative to the CPP. Third, this predictive capacity was formally mediated by the effect of FCθ on the rate of second-order evidence accumulation, as indexed by CPP build-up rate. Fourth, FCθ correlated across-subjects with a model-based estimate of the strength of the evidence that fed into the error detection process. This specification of an important role for pMFC signaling in explicit error detection suggests that metacognitive decisions are not solely based on the accumulation of sensory evidence but are also influenced by internally generated higher-order signals.
While these observations accord with theoretical proposals that the pMFC provides a critical input to the error detection process (Botvinick et al., 2001; Yeung et al., 2004), they are not decisive about the specific nature of this input. One possibility is that pMFC furnishes a distinct source of abstracted ‘error evidence’ that directly informs second-order judgments. Extensive data suggest that θ-band activity in pMFC signals the degree to which competing action plans are simultaneously activated, commonly known as ‘conflict’ (Cavanagh and Frank, 2014; Cavanagh et al., 2011; Cohen, 2014; 2013). Computational models have shown that a post-commitment conflict signal alone would provide a reliable basis for error detection because its magnitude scales with error likelihood and can reliably classify trial-to-trial performance accuracy (Yeung et al., 2004; see also Charles et al., 2014). Such a causal role for pMFC-encoded conflict in second-order decision-making would also provide a mechanistic explanation for the recent finding that manipulating effector-specific premotor activity both before and immediately after perceptual choice has clear effects on subsequent confidence reports (Fleming et al., 2015). Alternatively, second-order decisions on our task may receive their sole or primary evidence from visual short-term memory (Smith and Ratcliff, 2009) and pMFC may serve to modulate this process by strategically tuning processing in a global network of task-relevant regions when conflict is detected (Cavanagh and Frank, 2014; Danielmeier et al., 2011; Dehaene et al., 1998; Shenhav et al., 2013). Consistent with such a non-specific influence on the post-decisional process, we found that the mediation of the FCθ/RTd relationship by CPP build-up rate was partial, indicating that pMFC may have also influenced the speed of error detection independently of its effect on the rate of evidence accumulation – perhaps via effects on other parameters of the decision process like the response threshold (Cavanagh et al., 2011) or motor execution time. The extent to which pMFC signaling directly modulates the second-order evidence accumulation process can be established in future work by examining the impact of pMFC perturbation on CPP dynamics (Hayward et al., 2004; Reinhart and Woodman, 2014; Sela et al., 2012).
In contrast to other sequential sampling accounts of confidence judgments and changes-of-mind which assume that first-order decisions and metacognitive judgments are both based on the same evidence source (Moran et al., 2015; Pleskac and Busemeyer, 2010; Resulaj et al., 2009; Yu et al., 2015), our simple diffusion model of error detection behaviour is agnostic to the nature of the evidence that drives the second-order decision process. The model thus accommodates suggestions that pMFC, or indeed other neural signals, provide modulatory inputs or additional sources of evidence that are multiplexed in a compound decision variable that determines second-order judgments (Ullsperger et al., 2014; 2010). Of course, a corresponding limitation of our modelling framework is that it does not specify the precise nature of the relationship between the first- and second-order decision processes, and thereby fails to provide explicit accounts of some aspects of our observed data. For example, our selective modelling of error detection behaviour does not provide an explanation for the similarity of the post-response CPP signals on undetected errors and correct Go trials. This similarity could perhaps be due to a common paucity of error evidence on both trial-types. On the other hand, our analysis did reveal a post-response difference in frontal ERP morphology between these trial-types (Figure 3—figure supplement 1) which suggests that they are at least partially dissociable in terms of neural dynamics. Such unresolved questions emphasize that a central goal of future research must be to build a unified model of decision-making that not only accounts for the often complex relationships between first- and second-order behavior (Moran et al., 2015; Pleskac and Busemeyer, 2010), but is also constrained by neurophysiological characterizations of the post-decision accumulation process. Moreover, a remaining challenge will be to devise innovative ways to manipulate the evidence that feeds into the second-order process in an effort to test such a model and further corroborate our identification of the CPP with post-decisional accumulation.
A substantial literature has already investigated error-related electrophysiological signals in human subjects and highlighted in particular a post-response centro-parietal ERP, labeled the Error Positivity (Pe), that reliably discriminates between detected and undetected errors (Murphy et al., 2012; Nieuwenhuis et al., 2001; O'Connell et al., 2007; Overbeek et al., 2005; Ridderinkhof et al., 2009; Wessel et al., 2011). However, a consensus regarding the precise functional significance of this signal has never been achieved. Although a series of recent studies have demonstrated that Pe amplitude correlates with confidence in perceptual decisions (Boldt and Yeung, 2015) and the criterion that subjects impose on error detection reports (Steinhauser and Yeung, 2010) and have thereby associated this component with the general quality of the metacognitive decision process [see also Steinhauser and Yeung (2012)], these studies could not identify the Pe with a specific neural mechanism. The present study, by contrast, demonstrates that the post-decisional CPP encodes a second-order decision variable that bears precisely the same dynamical properties as have been reported for first-order decision signals in single-unit and population-level neurophysiology, including an RT-predictive build-up rate and a boundary-crossing relationship to response execution – properties that have never been previously reported for the Pe. Our use of a self-initiated error detection report was a critical design feature in this regard because it facilitated the interrogation of signal dynamics leading up to the moment of error detection; this has not been possible in previous studies of the Pe, which enforced either delayed metacognitive reporting or none at all. Where we have previously demonstrated that the pre-decision build-up of the CPP encompasses activity that has traditionally been associated with the classic P300 or ‘P3b’ (Kelly and O'Connell, 2013; O'Connell et al., 2012; Twomey et al., 2015), an important implication of the present findings is that the post-decision build-up of the CPP corresponds to the activity commonly attributed to the Pe. Thus, our data suggest that the P3b and Pe reflect distinct stages of the same neurophysiological process and point to a unifying mechanistic framework for understanding both signals.
In conclusion, we have reported the first definitive neurophysiological demonstration that evidence accumulation continues after the point of decision commitment and predicts the timing and accuracy of subsequent error detection. This finding furnishes critical neurophysiological support for theoretical accounts of metacognitive decision-making that have relied on the concept of post-decisional accumulation. Moreover, we have shown that this process is informed by a higher-order neural signal generated in medial frontal cortex, which suggests that metacognitive judgments are not solely based on the feedforward accumulation of sensory evidence but also on representations of conflict or error likelihood. Collectively, these results shed significant new light on the generative mechanisms of metacognition and furnish new evidence that error detection, confidence judgments and their neural substrates can be understood in terms of the same mechanistically principled framework of evidence accumulation.
Materials and methods
Subjects
All subjects were right-handed, had normal or corrected-to-normal vision, no history of psychiatric illness or head injury, reported no color-blindness, and refrained from ingesting caffeine on the day of testing. They provided written informed consent, and all procedures were approved by the Trinity College Dublin ethics committee and conducted in accordance with the Declaration of Helsinki. Subjects received a gratuity of €20 for their participation.
Task procedures
Request a detailed protocolTesting was performed in a dark, sound-attenuated room. Stimuli were presented using the ‘Presentation’ software suite (NeuroBehavioural Systems, San Francisco, CA) and subjects responded with the thumb of their right hand using a Microsoft ‘Sidewinder’ controller. During task performance, subjects used a table-mounted head-rest which fixed their distance from the display monitor (51-cm CRT operating at 85 Hz) at 80 cm for the entire task. They were instructed to maintain gaze at a centrally-presented white fixation cross on a gray background. Color/word stimuli appeared 0.25° above fixation.
Task version 1: self-initiated error detection reporting
Request a detailed protocolThe first version of the task was administered to thirty-two individuals, four of whom were excluded from all analyses: one due to technical issues with the EEG recording, two with excessively poor task accuracy (<30% withheld No-Go trials), and a further subject with no observable CPP component. Thus, we analyzed a final sample of 28 subjects (13 male) for the primary study, with a mean age of 23.5 years (s.d. = 5.8). This pre-planned sample size is consistent with other electrophysiological studies of decision-making from our lab that interrogated similar neural signals and invoked similar analytical methods (Kelly and O'Connell, 2013; O'Connell et al., 2012; Twomey et al., 2015).
Subjects were required to respond with a single ‘A’ button press on all Go trials, and to withhold this response on both ‘repeat’ and ‘color’ No-Go trials (Figure 1). They were instructed to give equal emphasis to the speed of their Go responses and the accuracy of their No-Go withholding. In the event of any failure to withhold the pre-potent response to either type of No-Go stimulus, subjects were required to signal detection of this error as quickly as possible by pressing a second ‘B’ button. They were instructed to execute this error detection report even if they became aware of an error after the onset of the following stimulus. Although such late error detections were rare (mean = 2.5 trials, s.d. = 2.4) and excluded from analysis, this instruction mitigated against the adoption of a time-dependent decision criterion that may have obscured any threshold-like relationship between second-order decision signals and the latency of error detection.
Each subject first completed a brief automated training protocol (Murphy et al., 2012), and was then administered at least 8 blocks of the task. Where time constraints allowed, we administered more blocks to increase the number of error trials available for analysis. On average, subjects completed 9.5 ± 0.8 blocks (range 8–10). Each block consisted of 224 word presentations, 200 of which were Go stimuli and 24 of which were No-Go stimuli (12 repeat, 12 color). All stimuli were presented for 0.4 s, followed by an inter-stimulus interval of 1.6 s. The duration of each block was therefore approximately 7.5 minutes. Stimuli were presented in a pseudo-random order with a minimum of three Go trials between any two No-Go trials.
Task version 2: half task blocks without error detection reporting
Request a detailed protocolTo establish the generality of the post-decisional signals observed under the previous task version to situations in which decision-makers are not explicitly instructed to monitor their own performance, we tested an independent cohort of sixteen individuals on a version of the Go/No-go task in which error reporting instructions were manipulated. Four of these subjects were excluded from analysis due to insufficient numbers of undetected errors (<6) for reliable EEG analysis, leaving a final sample of twelve (5 male) with a mean age of 23.7 years (s.d. = 6.7). Although this sample size is lower than that employed in the first study, we here limited our analyses to dependent measures with inherently high signal-to-noise (trial-averaged waveforms) and this sample size is consistent with a previous study of error detection using the same task employed presently (Hester et al., 2005). Task design and procedures were identical to those used for the previous version of the task, with the exception that every subject was administered five task blocks with regular error detection reporting and five ‘no-report’ blocks in which subjects were not instructed to monitor for errors. The five blocks in each condition were administered contiguously, and the order in which each condition was presented was counter-balanced across subjects. Thus, half of the full sample performed the no-report blocks without knowing that they would later be required to signal their errors. The presence of both FCθ and CPP signals in the no-report condition (Figure 7) did not depend on condition order.
EEG acquisition and preprocessing
Request a detailed protocolContinuous EEG was acquired using an ActiveTwo system (BioSemi, The Netherlands) from 64 scalp electrodes, configured to the standard 10/20 setup and digitized at 512 Hz. Eye movements were recorded using two vertical electro-oculogram (EOG) electrodes positioned above and below the left eye and two horizontal EOG electrodes positioned at the outer canthus of each eye. EEG data were processed in Matlab (Mathworks) via custom scripting and subroutines from the EEGLAB toolbox (Delorme and Makeig, 2004).
Eye-blinks and other noise transients were isolated and removed from the EEG data via independent component analysis (ICA). Specifically, continuous data from each block were re-referenced to channel Fz; data were high-pass filtered to 1 Hz, low-pass filtered up to 95 Hz and notch filtered to remove 50 Hz line noise using a two-way least-squares FIR filter; noisy channels were identified by visual inspection of signal variance and removed; data were segmented into temporally contiguous epochs of 1 s duration; epochs containing values that violated amplitude (± 250 µV) and joint probability (± 4.5 s.d.; Delorme and Makeig, 2004) criteria were rejected; and, the remaining data were subjected to temporal ICA using the infomax algorithm. The ICA weights yielded by this procedure were then back-projected to the original continuous, unfiltered EEG data for the associated block. Next, independent components representing stereotyped artifactual activity such as eye blinks, saccades and individual electrode artifacts were identified by visual inspection and discarded, and the ICA-pruned data were low-pass filtered to 35 Hz. No high-pass filter was applied. Previously-identified noisy channels were then interpolated via spherical spline interpolation, and the data were re-referenced to the common average. Data epochs were extracted from 2.5 s before to 3.5 s after stimulus onset on each trial (thus minimizing edge artifacts during spectral analysis) and baseline-corrected to the 0.3 s interval preceding stimulus onset. Subsequent epoch rejection employed a dynamic window with a fixed start time of −0.3 s relative to stimulus onset and an end time that depended on the primary RT of each trial: the window ended at RT + 1.2 s for go trials and undetected errors, and RT + 0.2 s + the slowest RTd of the current participant for detected errors; end time was truncated to +2 s relative to stimulus onset if the window encroached upon onset of the subsequent stimulus on any trial. Epochs were rejected from all further analysis if any scalp channel exceeded ±100 µV at any point within this trial-specific window. Detected error trials on which RTd followed next-trial onset were also excluded. Lastly, all EEG data were converted to current source density (Kayser and Tenke, 2006) to increase spatial selectivity and minimize volume conduction (Kelly and O'Connell, 2013; Twomey et al., 2015).
A previous paper by our group investigated the relationship between the Pe component (referred to here as the second-order CPP) and error detection using the same data and reported a strong correlation between Pe peak latency and RTd (Murphy et al., 2012). However, we did not consider the influence of Pe build-up rate on performance, we did not establish functional equivalence between the Pe and CPP and we did not consider the contribution of FCθ to error detection (see below). In fact, in that paper we reported a reliable negative association between RTd and Pe amplitude immediately prior to error detection that is inconsistent with the presently reported results and difficult to reconcile with the proposal that the Pe reflects the accumulation of evidence toward a fixed decision bound. Two critical methodological distinctions between Murphy et al. and the present study account for this discrepancy. First, whereas Murphy et al. implemented a high-pass temporal filter, the present study did not. High-pass filtering is problematic in the current context because it is likely to attenuate CPP amplitude to a greater degree on trials with long RTs since decision-related neural activity is drawn out over a longer time frame on such trials. Second, the present study used a spatial filter to reduce the overlap with other temporally coincident signals because in recent work we demonstrated that boundary-crossing effects at response can be obscured by spatial overlap of the CPP with anticipatory signals emanating from frontal sites (Kelly and O'Connell, 2013; Twomey et al., 2015). Addressing these issues enabled us to make a variety of important new observations that extend our previous results in several significant ways: we provide the first demonstration of the critical build-to-threshold relationship between the second-order CPP and RTd that is characteristic of an evolving decision variable; we show that this component interacts with a frontal conflict signal to determine the probability and timing of error detection; and, we leverage computational modelling to further identify this component with the second-order evidence accumulation process.
Fronto-central theta (FCθ; 2–7 Hz) power was measured in two ways. First, EEG data were decomposed into their time-frequency representation via complex Morlet wavelet convolution (between 2 and 12 cycles per wavelet, linearly increasing across 90 linear-spaced frequencies from 1 to 30 Hz) and the resulting power estimates were normalized by the decibel (dB) transform (dB = 10*log10[power/baseline]). The baseline consisted of across-trial average power during the 0.3 s preceding stimulus onset, calculated and applied separately within each trial-type (Go, undetected error, detected error). This approach yielded a condition-averaged time-frequency plot that allowed us to select channels and frequency boundaries for FCθ analysis in a manner that was orthogonal to any potential trial-type effects. Black lines in this plot (Figure 3a) enclose regions in which contiguous time-frequency pixels were significantly different from the pre-stimulus baseline at p < 0.01, for at least 400 ms and at least 5 consecutive frequency bins. Second, power estimates for all subsequent FCθ analyses were derived by band-pass filtering the EEG data from 2 to 7 Hz (using the fir1 Matlab function to construct a narrow two-way least squares FIR filter kernel), Hilbert-transforming the filtered data to derive the analytic signal, and then converting to power. Compared to wavelet convolution this filter-Hilbert method affords greater control over the frequency characteristics of the filter, though in practice the results from both methods were qualitatively very similar. As for the previous wavelet-based approach, analyses were conducted on power estimates that were dB-normalized using a condition-specific trial-averaged pre-stimulus baseline, thus leaving within-condition trial-by-trial fluctuations in baseline power intact. Complementarily, the reported between-subjects FCθ correlation (Figure 6e) was conducted on unbaselined power estimates, thus leaving between-individual differences in baseline FCθ power intact. Both within- and between-subjects variation in baseline power emerged to be sources of variance contributing to the respective effects (Figure 3—figure supplement 3c-e; Figure 6—figure supplement 2). We also verified that the main effect of error detection on FCθ remained unchanged when no baseline was applied (Figure 3—figure supplement 3b).
Analysis of electrophysiological decision signals
Request a detailed protocolFirst- and second-order trial-averaged CPP signals were measured as the average voltage per m2 from three centro-parietal electrodes centered on the region of maximum component amplitude in the grand-average response-locked topography (Pz, P1, P2), and were low-pass filtered up to 10 Hz for analysis and display. FCθ was measured as the average power across six fronto-central electrodes, also centered on the topographic maximum (FCz, FC1, FC2, Cz, C1, C2). The relationships between RT and signal build-up rates and amplitudes were examined for the CPP on go trials (Figure 2a), the second-order continuation of this signal on detected errors (Figure 2d), and FCθ on both trial-types (Figure 3d,e). Analysis of Go-trial dynamics was restricted to trials with primary RTs > 350 ms (leading to the exclusion of 17.4 ± 12.1% of trials per subject) because the amplitude of the CPP signal was affected by visual-evoked potentials that coincided with the evolution of the decision-related activity on quicker trials (Figure 2—figure supplement 2). For each participant, single-trial waveforms were sorted into 3 equal-sized bins according to RT (primary RT for Go trials, RTd for detected errors) and averaged. To establish the timing of the relationship between signal build-up rate and RT, we measured the temporal slope of each signal in each subject’s bin-averaged waveforms using a sliding window of 150 ms, covering the entirety of both response- and detection-aligned waveforms. Build-up rate was computed as the slope of a straight line fitted to the signal within each temporal window, and a linear contrast was applied to this metric across RT bins. The centers of windows that were characterized by a significant group-level contrast in the expected direction (linear decrease in build-up rate with increasing RT; p < 0.05, one-tailed) are marked by a black running line in the associated plots. To establish the timing of the relationship between signal amplitude and RT, we conducted a linear contrast of amplitude as a function of RT bin for each temporal sample. Samples characterized by contrasts that deviated from zero at the group level (p < 0.05, two-tailed) are marked by a gray line in the associated plots. This sample-wise approach was also employed to characterize the effect of error detection on the amplitude of both centro-parietal (Figure 2b) and FCθ (Figure 3b) signals.
For FCθ and CPP ROC curve analyses (Figure 2c; Figure 3c), single-trial waveforms for detected and undetected errors were pooled across all subjects and the area under the ROC curve was calculated for the average of each signal within discrete peri-error time windows (window width of 20 ms, moving in 20 ms increments, from −0.4 to +0.6 s relative to error). Significant deviations in classification accuracy from chance levels were determined via permutation testing (1000 iterations with random trial reassignment conserving individual detected versus undetected error proportions).
All further single-trial analyses of the second-order CPP leveraged waveforms that were low-pass filtered to 6 Hz to increase signal-to-noise. Single-trial amplitude was defined as the mean power from –0.1 to +0.4 s relative to error commission for FCθ and the mean signal in the 0.2 s preceding error detection for the CPP (Figure 4a). Single-trial build-up rate was measured as the slope of a straight line fitted to each waveform using the interval 0 to +0.2 s for FCθ and +0.1 to +0.3 s for the CPP, both relative to error commission (Figure 4b). Single-trial peak latency was measured as the time of maximum signal amplitude relative to error commission within a dynamic measurement window with a start time of −0.1 s for FCθ and +0.1 s for the CPP, and an end time of the RTd for that trial +0.15 s (Figure 4c). In cases where trials were initially assigned the minimum possible latency given the above constraints, the window start time was adjusted to be the earliest latency at which the waveform next became positive-going.
The independent contributions of FCθ and the CPP toward RTd were examined via single-trial within-subjects robust regression (O'Leary, 1990). For signal amplitude, the equation RTd = β0 + β1*FCθAmp + β2*CPPAmp yielded fitted regression coefficients representing the linear relationships between RTd and both single-trial theta power (β1) and CPP amplitude (β2). For build-up rate and peak latency, two further models were constructed by replacing the predictor variables where appropriate. All coefficients in a given model were estimated simultaneously via type III sum of squares. RTd was log-transformed and peak signal latency was square root-transformed to normalize their respective distributions before coefficient estimation. Variance inflation factors for all predictors across all models were <1.8, indicating weak multi-collinearity. For the amplitude and build-up rate metrics, effect sizes of the fitted β coefficients (effect size t = β/s.e.m.) were tested for group-level statistical significance via one-sample t-test (H0: effect size = 0) and contrasted via paired t-test (H0: effect size1 − effect size2 = 0). For the latency metric, our use of a dynamic measurement window that was determined by single-trial RTd ensured an arbitrary positive correlation between single-trial signal latency and RT. To test for effect significance, we thus compared the observed effect sizes derived from the above regression model to the expected values of the same effect size metrics computed in the case in which signal latencies were randomly chosen from anywhere within each trial’s measurement window. This process was repeated 1,000 times per subject to derive subject-specific permuted distributions of the FCθ and CPP latency effects, against which we tested for statistical significance. The above procedures for amplitude, build-up rate and latency effects were also repeated on an electrode-by-electrode basis to construct topographic representations of where these effects were strongest on the scalp (Figure 4). For the latency analysis, the trial-by-trial timing of the peak positivity or negativity was extracted for each electrode, depending on whether the trial-averaged amplitude at that electrode in the 0.3 s preceding the error detection report was above or below zero, respectively.
We employed mediation analysis (M3 toolbox for Matlab; http://wagerlab.colorado.edu/tools) to establish whether second-order CPP build-up rate mediated the relationship between FCθ power and RTd (Figure 5). For this analysis, we measured CPP build-up rate from −0.3 to −0.1 s relative to error detection report in order to minimize the temporal overlap between the FCθ and CPP measures, though the results were very similar when the original error-aligned measurement window for build-up rate was employed. Single-trial values for each measure were z-scored within-subjects and pooled across-subjects. To mitigate the low signal-to-noise ratio inherent in correlating two noisy single-trial electrophysiological metrics, average values for each of FCθ, CPP build-up and RTd were computed in bins of 12 trials that were grouped after sorting trials in order of increasing RTd. Mediation effects at larger bin sizes were of comparable magnitude (Figure 5—figure supplement 1). For CPP build-up rate to be considered a significant mediator, it was required to reach statistical significance in three tests: it must be related to the predictor (FCθ), it must be related to the outcome (RTd) while controlling for the predictor, and the mediation effect (evaluating whether some covariance between predictor and outcome can be explained by the mediator) must be significant. Significance of the associated path coefficients was assessed via bias-corrected bootstrap tests with 10,000 samples (Wager et al., 2008).
Diffusion modelling
Request a detailed protocolSecond-order behavioral data (error detection accuracy and RTd) were decomposed into latent decision-making parameters via a novel application of the one-choice drift diffusion model (Ratcliff and Van Dongen, 2011; Figure 6a). Noisy evidence for an error was assumed to accumulate over time at drift rate ν until a decision bound a was met, at which point error detection was achieved. The moment-to-moment noise in the evidence is determined by the s parameter, which refers to the standard deviation of a zero-mean Gaussian distribution from which random increments to the deterministic component of the accumulation process (represented by ν) are drawn. As is common in fits of the drift diffusion model to data, s was fixed at 0.1 in order to scale all other parameters in the model across individuals. Drift rate was assumed to be normally distributed across trials with a standard deviation η, and all non-decision-related processing was assigned to a non-decision time parameter tnd. The inclusion of the η parameter accorded with previous one-choice model fits to first-order behavior which suggested that variability in drift rate is necessary to account for the various observable shapes of hazard function in one-choice data (Ratcliff and Van Dongen, 2011). We made the additional assumption that the temporal integration process terminated if a was not reached by a time deadline, thereby resulting in an undetected error. A deadline on post-decision accumulation also features in a prominent model of fast changes of mind in decision-making (Resulaj et al., 2009) and was a free parameter in that study. Here, we estimated the subject-specific detection deadline empirically in order to retain a degree of freedom when assessing model fit: any extreme outliers (> mean + 3.5 s.d.) were trimmed from each subject’s RTd distribution and the deadline was defined as the slowest remaining RT per subject. This procedure yielded an average deadline of 1.17 s relative to initial error commission (±0.19; range 0.81 to 1.47 s).
Two additional assumptions were made when plotting simulated decision variables derived from the best-fitting model parameters (Figure 6d), both of which were informed by characteristics of the second-order CPP. First, 90 ms of each subject’s fitted tnd parameter was allotted to post-threshold response preparation, and any residual tnd determined the length of the delay between error commission and the start of evidence accumulation. Second, the decision variable was subject to a linear decay to baseline over the 300 ms after the decision bound was reached.
There is no analytical solution for RT distributions with negative drift rate, which can result from including η in the one-choice model (Ratcliff and Van Dongen, 2011). The model was therefore implemented as a simulation using a random walk approximation to the diffusion process with 15,000 iterations per distribution at 1 ms step size. In order to fit the model to the observed data for each subject, five RT quantiles (0.1, 0.3, 0.5, 0.7, 0.9) were computed from that subject’s RTd distribution and the proportions of all error trials (detected + undetected) lying between those quantiles were multiplied by the total number of error trials to yield observed values (O). Thus, the defective cumulative probability distribution of error detection reports was used to derive per-quantile trial frequencies, which allowed the model to simultaneously fit both RTd and error detection accuracy. We then calculated the model-estimated proportions of trials that lay between these RT quantiles, and these were multiplied by the number of actual observations to yield the model-derived expected values (E). A χ2 statistic Σ(O – E)2/E was computed and the parameters of the model were adjusted by a particle swarm optimization routine (Birge, 2003) to minimize this value iteratively (30 particles, set at pseudorandom starting points in parameter space). Initial attempts at parameter estimation indicated that the particle swarm approach tended to be more robust to local minima than the commonly-used Simplex minimization routine.
As described elsewhere (Ratcliff and Van Dongen, 2011), the one-choice diffusion model suffers from a parameter identifiability problem whereby different estimates of the v, η and a parameters can produce similar goodness of fit but vary in magnitude by as much as 2:1. However, the ratio of v/η remains invariant across different model fits. We therefore employed this ‘drift ratio’ quantity, which is analogous to d’ in classic signal detection theory (Ratcliff and Van Dongen, 2011), as our model-based estimate of the quality of the evidence feeding into each subject’s error detection decision process. In the reported between-subjects correlations (Figure 6e,f,g), drift ratio was correlated against summary electrophysiological measures for each subject. Amplitude and peak latency measures were calculated by averaging across single-trial estimates of these metrics on detected error trials, derived via the same measurement windows that were used for the previous within-subjects regression analyses. In an effort to increase the signal-to-noise of the build-up rate metric, per-subject build-up rate for each signal was defined as the slope of a linear fit to the average error-locked waveform on detected error trials, within a window that started at a subject-specific signal onset time obtained by visual inspection and ended at the subject-specific peak signal latency. One outlier data point with an absolute studentized deleted residual value > 4 in all bivariate correlations was excluded from the reported analyses, though all relationships remained at least marginally significant when this subject was included (all p < 0.06).
References
-
PSOt: a particle swarm optimization toolbox for matlabProceedings IEEE Swarm Intelligence Symposium.
-
Shared neural markers of decision confidence and error detectionJournal of Neuroscience 35:3478–3484.https://doi.org/10.1523/JNEUROSCI.0797-14.2015
-
Conflict monitoring and cognitive controlPsychological Review 108:624–652.https://doi.org/10.1037/0033-295X.108.3.624
-
Internal signal correlates neural populations and biases perceptual decision reportsProceedings of the National Academy of Sciences of the United States of America 109:18938–18943.https://doi.org/10.1073/pnas.1216799109
-
Anterior cingulate cortex and conflict detection: an update of theory and dataCognitive, Affective, & Behavioral Neuroscience 7:367–379.https://doi.org/10.3758/CABN.7.4.367
-
Frontal theta as a mechanism for cognitive controlTrends in Cognitive Sciences 18:414–421.https://doi.org/10.1016/j.tics.2014.04.012
-
Subthalamic nucleus stimulation reverses mediofrontal influence over decision thresholdNature Neuroscience 14:1462–1467.https://doi.org/10.1038/nn.2925
-
Decoding the dynamics of action, intention, and error detection for conscious and subliminal stimuliJournal of Neuroscience 34:1158–1170.https://doi.org/10.1523/JNEUROSCI.2465-13.2014
-
Midfrontal conflict-related theta-band power reflects neural oscillations that predict behaviorJournal of Neurophysiology 110:2752–2763.https://doi.org/10.1152/jn.00479.2013
-
A neural microcircuit for cognitive conflict detection and signalingTrends in Neurosciences 37:480–490.https://doi.org/10.1016/j.tins.2014.06.004
-
Failures of metacognition and lack of insight in neuropsychiatric disordersPhilosophical Transactions of the Royal Society B: Biological Sciences 367:1379–1390.https://doi.org/10.1098/rstb.2012.0002
-
Perceptual detection as a dynamical bistability phenomenon: a neurocomputational correlate of sensationProceedings of the National Academy of Sciences of the United States of America 104:20073–20077.https://doi.org/10.1073/pnas.0709794104
-
A neuronal model of a global workspace in effortful cognitive tasksProceedings of the National Academy of Sciences of the United States of America 95:14529–14534.https://doi.org/10.1073/pnas.95.24.14529
-
EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysisJournal of Neuroscience Methods 134:9–21.https://doi.org/10.1016/j.jneumeth.2003.10.009
-
Psychophysiological Brain ResearchEffects of errors in choice reaction tasks on the Erp under focused and divided attention, Psychophysiological Brain Research, Tilberg, The Netherlands, Tilberg University Press.
-
Executive attention and metacognitive regulationConsciousness and Cognition 9:288–307.https://doi.org/10.1006/ccog.2000.0447
-
Metacognition: computation, biology and functionPhilosophical Transactions of the Royal Society B: Biological Sciences 367:1280–1286.https://doi.org/10.1098/rstb.2012.0021
-
Action-specific disruption of perceptual confidencePsychological Science 26:89–98.https://doi.org/10.1177/0956797614557697
-
A neural system for error detection and compensationPsychological Science 4:385–390.https://doi.org/10.1111/j.1467-9280.1993.tb00586.x
-
The neural basis of decision makingAnnual Review of Neuroscience 30:535–574.https://doi.org/10.1146/annurev.neuro.29.051605.113038
-
A model of the go/no-go taskJournal of Experimental Psychology 136:389–413.https://doi.org/10.1037/0096-3445.136.3.389
-
The role of the anterior cingulate cortex in the counting stroop taskExperimental Brain Research 154:355–358.https://doi.org/10.1007/s00221-003-1665-4
-
Internal and external influences on the rate of sensory evidence accumulation in the human brainJournal of Neuroscience 33:19434–19441.https://doi.org/10.1523/JNEUROSCI.3355-13.2013
-
Proceeding of the Nineteenth Annual Meeting of The International Society for PsychophysicsConfidence and random walk theory, Proceeding of the Nineteenth Annual Meeting of The International Society for Psychophysics, Stockholm, Sweden, International Society for Psychophysics.
-
On the ability to inhibit thought and action: a theory of an act of controlPsychological Review 91:295–327.https://doi.org/10.1037/0033-295X.91.3.295
-
An electrophysiological signal that precisely tracks the emergence of error awarenessFrontiers in Human Neuroscience 6:.https://doi.org/10.3389/fnhum.2012.00065
-
Common medial frontal mechanisms of adaptive control in humans and rodentsNature Neuroscience 16:1888–1895.https://doi.org/10.1038/nn.3549
-
The role of cingulate cortex in the detection of errors with and without awareness: a high-density electrical mapping studyThe European Journal of Neuroscience 25:2571–2579.https://doi.org/10.1111/j.1460-9568.2007.05477.x
-
A supramodal accumulation-to-bound signal that determines perceptual decisions in humansNature Neuroscience 15:1729–1735.https://doi.org/10.1038/nn.3248
-
Dissociable components of error processingJournal of Psychophysiology 19:319–329.https://doi.org/10.1027/0269-8803.19.4.319
-
Robust regression computation using iteratively reweighted least squaresSIAM Journal on Matrix Analysis and Applications 11:466–480.https://doi.org/10.1137/0611032
-
Neural correlates of decisions and their outcomes in the ventral premotor cortexJournal of Neuroscience 28:12396–12408.https://doi.org/10.1523/JNEUROSCI.3396-08.2008
-
Two-stage dynamic signal detection: a theory of choice, decision time, and confidencePsychological Review 117:864–901.https://doi.org/10.1037/a0019737
-
Processing a display even after you make a response to it. How perceptual errors can be correctedThe Quarterly Journal of Experimental Psychology Section A 33:223–239.https://doi.org/10.1080/14640748108400790
-
Diffusion model for one-choice reaction-time tasks and the cognitive effects of sleep deprivationProceedings of the National Academy of Sciences of the United States of America 108:11285–11290.https://doi.org/10.1073/pnas.1100483108
-
Psychology and neurobiology of simple decisionsTrends in Neurosciences 27:161–168.https://doi.org/10.1016/j.tins.2004.01.006
-
An integrated theory of attention and decision making in visual signal detectionPsychological Review 116:283–317.https://doi.org/10.1037/a0015156
-
Decision processes in human performance monitoringJournal of Neuroscience 30:15643–15653.https://doi.org/10.1523/JNEUROSCI.1899-10.2010
-
Error awareness as evidence accumulation: effects of speed-accuracy trade-off on error signalingFrontiers in Human Neuroscience 6:.
-
The classic P300 encodes a build-to-threshold decision variableThe European Journal of Neuroscience 42:.https://doi.org/10.1111/ejn.12936
-
Neural mechanisms and temporal dynamics of performance monitoringTrends in Cognitive Sciences 18:259–267.https://doi.org/10.1016/j.tics.2014.02.009
-
Conscious perception of errors and its relation to the anterior insulaBrain Structure & Function 214:629–643.https://doi.org/10.1007/s00429-010-0261-1
-
Error awareness revisited: accumulation of multimodal evidence from central and autonomic nervous systemsJournal of Cognitive Neuroscience 23:3021–3036.https://doi.org/10.1162/jocn.2011.21635
-
Metacognition in human decision-making: confidence and error monitoringPhilosophical Transactions of the Royal Society B 367:1310–1321.https://doi.org/10.1098/rstb.2011.0416
-
Dynamics of postdecisional processing of confidenceJournal of Experimental Psychology 144:489–510.https://doi.org/10.1037/xge0000062
Decision letter
-
Michael J FrankReviewing Editor; Brown University, United States
eLife posts the editorial decision letter and author response on a selection of the published articles (subject to the approval of the authors). An edited version of the letter sent to the authors after peer review is shown, indicating the substantive concerns or comments; minor concerns are not usually shown. Reviewers have the opportunity to discuss the decision before the letter is sent (see review process). Similarly, the author response typically shows only responses to the major concerns raised by the reviewers.
Thank you for submitting your work entitled "Neural evidence accumulation persists after choice to inform metacognitive judgments" for consideration by eLife. Your article has been reviewed by two peer reviewers, one of whom has agreed to reveal their identity: Jim Cavanagh. The evaluation has been overseen by a Reviewing Editor (Michael Frank) and Timothy Behrens as the Senior Editor.
The reviewers have discussed the reviews with one another and the Reviewing editor has drafted this decision to help you prepare a revised submission.
Summary:
This paper reports a novel investigation of the dynamics of post-decisional integration of evidence in error detection. There has been considerable recent interest in the idea that sensory evidence continues to accumulate after a decision, and might underpin metacognitive phenomena such as confidence judgments and changes of mind. So far this notion has only been tested with behavioral modeling, but this paper reveals a possible biological mechanism. The authors present evidence suggesting that the post-error voltage positivity in the ERP (the Pe) reflects the same sort of evidence accumulation process as the central parietal positivity (CPP). This work thus re-defines well-known neuroelectric events based on more generalized and formalized algorithmic definitions. While the authors have previously described the CPP as a first-order evidence (direct sensory evidence) accumulation process, here they argue that the post-error process described here (Pe) reflects second-order (endogenous evidence) evidence accumulation. The authors bolster this claim by applying a one-choice diffusion model to the RT data, facilitating a direct comparison between the mechanistic implementation of CPP-like algorithms (P3 and Pe) even though they have differing cognitive implications (first vs. second order decisions).
Essential revisions:
The methodological sophistication of this report leads to some compelling strengths yet also a few lingering concerns. Virtues of this approach include intelligent spatial and temporal filtering, appropriate capitalization on previously detailed understanding of the role of non-phase locked frontal theta power in post-error decision making (including a formalized mediation model of theta, CPP and RT change), and successful one-choice diffusion modeling and application of model parameters to assist in the generalization of these findings. The covariation of the CPP with the second-order detection response time is particularly clear, providing strong underpinnings for post-decision evidence accumulation frameworks. The concern that the second-order CPP response is a motor artefact is allayed by observation of a similar potential in a "no response" control experiment. But there are some lingering concerns that both reviewers highlighted, elaborated in detail below, consolidating comments from individual reviewers.
1) A limitation of the modelling framework is the omission of any relationship to objective task requirements. The error detection process is modelled as a noisy one-response accumulator with its own drift rate, but this process does not know anything about the first-order decision. I understand the benefit of keeping things simple here and not making unnecessary assumptions about the link between first and second order accumulation, but I worry that it gives a misleading picture of the underlying second order accumulation process, and what we should expect of the data. Specifically, the model predicts that objective errors that don't reach a response threshold are no different, neurally, from correct "go" trials. Instead it seems more plausible that objective features of error trials drive fluctuations in "error evidence" and that on a subset of trials this evidence reaches criterion for report. It's therefore also surprising that the CPP and FCθ signals for both "Go" and "Undetected errors" (which are matched for motor output) were also near-identical (Figure 2B and Figure 3B). Doesn't this finding go against the notion that these signals are accumulating noisy evidence to a bound? Or are these signals instead indexing the criterion for error reporting? Perhaps there are differences between undetected errors and go trials in the data that would shed light on this issue that are not apparent from Figure 2 and Figure 3?
2) More generally, caution is warranted on the claims that error detection reflects evidence accumulation per se after the primary task response has been executed. There is no manipulation of "evidence" strength (i.e. where faster accumulation would proceed for larger evidence), and no speed-accuracy trade-off manipulation to test for some kind of bounded accumulation, either of which would provide stronger support that the error detection process reflects a similar sequential sampling of evidence. Thus the authors should discuss and/or address this limitation.
3) Some analytic choices appear to be unrelated to other similar analyses, raising worrying concerns of tailoring the final report to the best outcomes of multiple choices rather than relying on the output of a single analytic framework. This decreases confidence in the robustness and thus the replicability of these full sets of findings and raises concerns of "p-hacking" based on the best of many seemingly arbitrary analytic choices (which I do not think are arbitrary at all). In particular, in paragraph four of the subsection “EEG acquisition and preprocessing” the authors describe the use of differing baseline procedures for frontal theta, which is potentially problematic. As the authors allude to ("leaving within-condition trial-by-trial fluctuations in baseline power intact"), condition-specific pre-stimulus baselines could have important influence on the appropriate level of trial-to-trial vs trial-specific inference, yet it could also simply add spurious variance into the dB ratio of post-event power. The authors do not distinguish between these possibilities or otherwise adequately justify their approach. Moreover, the between-subjects contrasts were not baseline corrected and the authors describe how both are "important source(s) of variance contributing to the reported effects". The authors need to do a more compelling job of justifying their analytic choices based on a principled, a priori framework, for example a common cross-condition baseline for all contrasts (see Cohen 2014: Analyzing Neural Time Series Data, MIT Press).
4) In the subsection “Signal interactions predict the timing of error detection” and in paragraph four of subsection “Analysis of electrophysiological decision signals”: The authors need to specify their regression models in greater detail. They report that "single-trial regressions revealed that this effect was only marginally significant after accounting for the contribution of CPP build up rate". However, it isn't made explicitly clear if they are included in a sequential or type III sum of squares manner to account for variance in main effects prior to the other main effects.
5) Paragraph one of subsection “Analysis of electrophysiological decision signals”: One-tailed test are rarely appropriate for hypotheses that could possibly go in either direction. Here, it is entirely possible that there would be a linear decline in an EEG signal associated with longer RTs.
6) The inference of a causal link between FCθ and CPP seems on shaky ground. Comparing Figure 2 and Figure 3, FCθ power differentiates between RTd earlier than CPP, but could this be due to the influence of neighbouring timepoints on the wavelet measure? This could be addressed with narrow-band wavelets or a filter-HIlbert approach. In single-trial analyses the FCθ power window (-0.1 to +0.4s) extends beyond the window at which the CPP build-up rate is calculated (+0.1 to +0.3s; Methods, paragraph three of the subsection “Analysis of electrophysiological decision signals”). It's therefore odd to enter FCθ as a cause of CPP in a mediation analysis given that the relevant part of the former potentially takes place after the latter. Putting aside the issue of timing, could the assumed causality be further bolstered by flipping the nodes in the mediation analysis, i.e. could the authors show that FCθ does not mediate the influence of CPP build-up on RTd?
7) The first-order CPP increases for both go and no-go trials, consistent with the notion of a domain-general accumulator. But the second-order CPP increases only on error trials (Figure 2B). If the second-order decision is analogous to the first (i.e. a race between neural subpopulations controlling whether or not to press the button based on fluctuating evidence), then presumably there should also be accumulation towards a no-go bound in the second-order, confident-in-being-correct case. Doesn't this mismatch between first-order and second-order CPP features weaken the conclusion that the same DV mechanism underpins first- and second-order decisions? It would be interesting to hear the author's perspective on this. For instance the second-order CPP could reflect evidence accumulation from endogenous activities towards an overt error detection response, but the authors also argue the first-order CPP is also involved in evidence accumulation leading to a non-response (Figure 2—figure supplement 1).
[Editors' note: further revisions were requested prior to acceptance, as described below.]
Thank you for resubmitting your work entitled "Neural evidence accumulation persists after choice to inform metacognitive judgments" for further consideration at eLife. Your revised article has been favorably evaluated by Timothy Behrens (Senior editor), Reviewing editor Michael Frank, and two reviewers (Steve Fleming and James Cavanagh). All were largely satisfied with your revision, but one Reviewer notes an additional analysis/discussion on the differences between go trials and undetected errors and a supplemental figure on the EEG time courses that would be helpful. Please address these issues and we will then be able to swiftly come to a final official decision.
Reviewer #1:
In my opinion, the findings presented here have been thoroughly vetted. Highly novel work raises many important questions. The authors have done a very reasonable job of fully addressing the 'solvable' issues raised here while also intelligently responding to larger-scale conceptual questions that will require multiple experiments to fully understand.
Reviewer #2:
The authors have mostly addressed my previous concerns in the first round of reviews. The one part that I don't think has been fully addressed is the question of whether there are actually differences between undetected errors and go trials in the data that would not be captured by the model (in general I'm happy with the response to point 1 that the connection between first- and second-order accumulation can be left for future work, but it seems important to fully display the data features that might inform such work). For instance the ERP analysis in Figure 3—figure supplement 1 only directly contrasts undetected and detected errors, but there appears to be a difference between go and undetected errors later in the time course. Similarly only go-trial build-up rates are documented in Figure 2A and Figure 3D. I suggest including the comparison between go and undetected CPP and FCθ timecourses as a supplemental figure to cover this issue.
https://doi.org/10.7554/eLife.11946.022Author response
1) A limitation of the modelling framework is the omission of any relationship to objective task requirements. The error detection process is modelled as a noisy one-response accumulator with its own drift rate, but this process does not know anything about the first-order decision. I understand the benefit of keeping things simple here and not making unnecessary assumptions about the link between first and second order accumulation, but I worry that it gives a misleading picture of the underlying second order accumulation process, and what we should expect of the data. Specifically, the model predicts that objective errors that don't reach a response threshold are no different, neurally, from correct "Go" trials. Instead it seems more plausible that objective features of error trials drive fluctuations in "error evidence" and that on a subset of trials this evidence reaches criterion for report. It's therefore also surprising that the CPP and FCθ signals for both "Go" and "Undetected errors" (which are matched for motor output) were also near-identical (Figure 2B and Figure 3B). Doesn't this finding go against the notion that these signals are accumulating noisy evidence to a bound? Or are these signals instead indexing the criterion for error reporting? Perhaps there are differences between undetected errors and Go trials in the data that would shed light on this issue that are not apparent from Figure 2 and Figure 3?
We agree that, without modelling first- and second-order performance within a single unified framework, we are unable to conclusively address certain aspects of the presented data. We also accept that the manuscript should do a better job of communicating what our modelling efforts do and do not tell us about the mechanisms underpinning error detection.
We do maintain that our current modelling efforts support the statement that “features of error trials drive fluctuations in ‘error evidence’ and […] on a subset of trials this evidence reaches criterion for report”. The one-choice diffusion model explicitly incorporates such variability in evidence strength (in the form of the s and η parameters, which specify within- and between-trial variability, respectively), and this variability is the primary determinant of whether or not an error will ultimately be detected. Indeed, a key strength of our modelling approach is to show that even though we make very few assumptions beyond this basic feature of sequential sampling models, our one-choice model is able to reproduce both observed error detection behavior (Figure 6B,C) and the associated neural dynamics (Figure 6D) – including, of particular note, the negative-going trial-averaged trajectory of the CPP on undetected errors. Thus, the model provides a very parsimonious account of the behavioral and neural error detection data.
On the other hand, strictly speaking our model makes no predictions about the neural equivalence (or otherwise) of undetected errors and Go trials, since behavior and neural dynamics on Go trials were not modelled. The reviewers suggest that the similarity of the neural signals on these trial-types (Figure 2B, Figure 3B) is inconsistent with the notion that the post-response CPP signal reflects post-decisional accumulation. However, we note that similar trial-averaged neural dynamics on Go and undetected error trials are not necessarily inconsistent with a post-decision accumulation account. Go trials are also likely to be associated with varying levels of error evidence (e.g. on tasks that evoke response conflict, FCθ power has been shown to vary strongly across correct trials and be predictive of task behavior, even within conflict conditions; Cohen & Cavanagh, 2011). Although in most cases the error evidence is likely to be weak, the presence of a minority of trials with elevated error evidence may be sufficient to yield equivalent trial-averaged signals to those observed on undetected error trials, which represent the subset of error trials on which the internal representation of evidence was weakest. This possibility accommodates our current modelling efforts which, as described above, specify that post-decisional accumulation takes place on both undetected and detected error trials. We do generally agree with the reviewers that our findings are not decisive about the nature of the relationship between the first- and second-order decision processes, and that this remains a key area of future inquiry. We have now attempted to further highlight this point in the Discussion section of our manuscript (paragraph six).
The reviewers also question whether the FCθ and/or CPP signals might be better understood as a reflection of the criterion for error reporting. Several aspects of our data speak against this idea. We observed that FCθ power was negatively associated with the timing of error detection (Figure 3E, Figure 4A) – opposite to the direction that would be predicted if this signal reflects trial-by-trial variation in the second-order criterion (a higher criterion should equate to slower RTs, and hence the relationship would be positive). Additionally, it is unclear to us why a signal like the CPP, which exhibits a clear RT-predictive build-up rate (Figure 4B) and whose evolution finely traces variation in the putative evidence accumulation process (Figure 6D), might reflect a static criterion as opposed to this more dynamic aspect of the decision process.
2) More generally, caution is warranted on the claims that error detection reflects evidence accumulation per se after the primary task response has been executed. There is no manipulation of "evidence" strength (i.e. where faster accumulation would proceed for larger evidence), and no speed-accuracy trade-off manipulation to test for some kind of bounded accumulation, either of which would provide stronger support that the error detection process reflects a similar sequential sampling of evidence. Thus the authors should discuss and/or address this limitation.
We of course agree that an experimental manipulation of error evidence strength would provide additional support for our conclusion that the CPP reflects the process of post-decisional accumulation. However, it is unclear how such a manipulation might be successfully implemented. One option could be to vary the duration following the first-order response that the imperative stimulus is presented for, but we question the extent to which post-response processing in such situations reflects a truly second-order metacognitive process. By contrast, a central aspect of our results is the demonstration that the decision signals that we describe persist for an extended period of time after first-order response execution even in the explicit absence of any new sensory input. This approach allowed us to verify that truly metacognitive error detection decisions are preceded by a CPP that exhibits the critical properties of a build-to-threshold decision variable as specified by sequential sampling models (an RT-predictive build-up rate and threshold amplitude at response), which itself constitutes the strongest neurophysiological support to date for the existence of post-decision accumulation. Moreover, in doing so we were able to show that this process is highly sensitive to variation in the strength of internal conflict signals (i.e. FCθ).
Although we did not manipulate the criterion that subjects imposed on their second-order decisions (e.g. via manipulation of speed-accuracy instructions or of the payoff matrix for second-order decisions), we note that this approach has been taken in a previous investigation of the Error Positivity (Pe) component (Steinhauser & Yeung, 2010) and the resulting effects were entirely consistent with our proposal that this signal reflects a second-order accumulation process (specifically, a higher criterion was associated with larger trial-averaged Pe signals). As we outline in the Discussion section (paragraph seven), however, that study could not identify this signal with the mechanism of evidence accumulation, primarily because it enforced a delay on second-order responding. This point highlights what we believe to be another important strength of our manuscript: its usefulness for clarifying the mechanistic significance of previous reports on the Pe.
We thank the reviewer for these comments and we have integrated the points above into the revised manuscript (Discussion).
3) Some analytic choices appear to be unrelated to other similar analyses, raising worrying concerns of tailoring the final report to the best outcomes of multiple choices rather than relying on the output of a single analytic framework. This decreases confidence in the robustness and thus the replicability of these full sets of findings and raises concerns of "p-hacking" based on the best of many seemingly arbitrary analytic choices (which I do not think are arbitrary at all). In particular, in paragraph four of the subsection “EEG acquisition and preprocessing” the authors describe the use of differing baseline procedures for frontal theta, which is potentially problematic. As the authors allude to ("leaving within-condition trial-by-trial fluctuations in baseline power intact"), condition-specific pre-stimulus baselines could have important influence on the appropriate level of trial-to-trial vs trial-specific inference, yet it could also simply add spurious variance into the dB ratio of post-event power. The authors do not distinguish between these possibilities or otherwise adequately justify their approach. Moreover, the between-subjects contrasts were not baseline corrected and the authors describe how both are "important source(s) of variance contributing to the reported effects". The authors need to do a more compelling job of justifying their analytic choices based on a principled, a priori framework, for example a common cross-condition baseline for all contrasts (see Cohen 2014: Analyzing Neural Time Series Data, MIT Press).
We thank the reviewers for communicating these valid concerns and now report several additional analyses that we hope will allay them. In two new figure supplements (Figure 3—figure supplement 2; Figure 6—figure supplement 3), we show the following:
1) The effect of error detection on trial-averaged FCθ power was almost identical when no baseline was applied to the theta waveforms (compare Figure 3B to Figure 3—figure supplement 2B), indicating that this effect did not depend on what baseline was employed.
2) Although the within-subjects relationship between FCθ and RTd was clearly still present when a single-trial baseline was employed (Figure 3—figure supplement 2C), the effect was weaker than that observed when a trial-averaged baseline (or equivalently, no baseline) was used (Figure 3—figure supplement 2D,E). This motivated our statement in the original submission (paragraph four of the subsection “EEG acquisition and preprocessing”) that within-subjects variation in baseline power was an important source of variance contributing to the reported effects. We do not show it here but single-trial regressions of RTd on baseline theta power alone yielded a significant effect of baseline power at the group level (p=0.014).
3) The between-subjects correlation between FCθ power and drift ratio was no longer statistically significant (p=0.16) when the FCθ signals were baselined for each subject (Figure 6—figure supplement 3). This motivated our statement in the original submission that between-subjects variation in baseline power appeared to be an important source of variance contributing to the observed between-subjects FCθ power effect.
We note that, in light of the above, our baselining protocol is principled insofar as it preserves the relevant sources of variation in baseline power for each analysis. We hope that this convinces the reviewers that we were not engaged in p-hacking.
To mitigate any remaining concerns regarding the between-subjects FCθ effect, we point out that the additional between-subjects noise that is present when not employing a baseline should in principle work to make any existing effects harder to detect – it seems less likely, then, that the lack of a baseline for this analysis is actually causing a spurious effect, and more reasonable (and consistent with at least a subset of the within-subjects analyses; Figure 3—figure supplement 2D,E) that variation in baseline power is actually meaningful. Moreover, we also now report (paragraph three of the subsection “Drift diffusion modelling of the second-order decision process”) using partial correlation that the FCθ/drift ratio correlation is still present (r=0.52, p=0.007) when unbaselined θ power averaged over bilateral posterior electrodes (P7/P8) was included as a co-variate (Cohen & Donner, 2013). This result suggests that the reported correlation is not driven by spurious ‘global’ differences in oscillatory power across subjects but is specific to the fronto-central signal of interest here.
4) In the subsection “Signal interactions predict the timing of error detection” and in paragraph four of subsection “Analysis of electrophysiological decision signals”: The authors need to specify their regression models in greater detail. They report that "single-trial regressions revealed that this effect was only marginally significant after accounting for the contribution of CPP build up rate". However, it isn't made explicitly clear if they are included in a sequential or type III sum of squares manner to account for variance in main effects prior to the other main effects.
We apologize for the lack of clarity here. We took the type III sum of squares approach – in a given model, all regressors were entered simultaneously and the associated regression coefficients represent the unique portion of variance in the dependent variable accounted for by each regressor, independent of all other regressors. We have now made this explicit in our Methods description of the single-trial regressions (paragraph four of the subsection “Analysis of electrophysiological decision signals”). We also slightly changed the phrasing in the Results section highlighted by the reviewers: rather than “[…]after accounting for the contribution[…]” (which may imply a sequential approach), we rephrase to “[…]when the contribution of CPP build-up rate was accounted for.”
5) Paragraph one of subsection “Analysis of electrophysiological decision signals”: One-tailed test are rarely appropriate for hypotheses that could possibly go in either direction. Here, it is entirely possible that there would be a linear decline in an EEG signal associated with longer RTs.
We had strong a priori motivations for using one-tailed tests in this instance: we focus on two signals (CPP, FCθ) that are invariably observed to increase rather than decrease in magnitude prior to response, and we predicted based on our previous research (Kelly & O’Connell, 2013; Twomey et al., 2015) and the logic of accumulation-to-bound sequential sampling models that if one/both of these signals reflects the process of evidence accumulation, their build-up rate should specifically be steeper for faster RTs. We also note that these analyses are primarily illustrative, and we do also provide a more rigorous test of the build-up rate/RT relationships with our two-tailed single-trial regression analysis (Figure 4B). These points notwithstanding, we have re-plotted all CPP and FCθ waveforms binned by primary RT and RTd below (Author response image 1), this time using two-tailed tests for signal build-up rates. As can be seen, this adjustment makes no substantive difference to the presence of all of the important build-up rate effects that we highlight in the manuscript.

Re-plotting of panels from figures 2A, 2D, 3D and 3E from the main manuscript, using two-tailed rather than one-tailed tests on signal build-up rates (significance indicated by black running markers).
https://doi.org/10.7554/eLife.11946.0216) The inference of a causal link between FCθ and CPP seems on shaky ground. Comparing Figure 2 and Figure 3, FCθ power differentiates between RTd earlier than CPP, but could this be due to the influence of neighbouring timepoints on the wavelet measure? This could be addressed with narrow-band wavelets or a filter-HIlbert approach. In single-trial analyses the FCθ power window (-0.1 to +0.4s) extends beyond the window at which the CPP build-up rate is calculated (+0.1 to +0.3s; Methods, paragraph three of the subsection “Analysis of electrophysiological decision signals”). It's therefore odd to enter FCθ as a cause of CPP in a mediation analysis given that the relevant part of the former potentially takes place after the latter. Putting aside the issue of timing, could the assumed causality be further bolstered by flipping the nodes in the mediation analysis, i.e. could the authors show that FCθ does not mediate the influence of CPP build-up on RTd?
In fact, all FCθ results with the exception of Figure 3A (time-frequency plot) were indeed obtained using a filter-Hilbert approach to bolster the temporal specificity of the analyses. We apologize for not making this clearer in the manuscript and have now done so in the legend for Figure 3. Note that we also reported in the original manuscript that results from the wavelet convolution and filter-Hilbert methods were qualitatively very similar (paragraph four of the subsection “EEG acquisition and preprocessing”).
The reviewers raise a good point about the time-windows employed for the measurement of FCθ power and CPP build-up rate in our original mediation analysis. To address this point, we carried out a new mediation analysis using the same measurement window for FCθ, but now a detection-locked window for CPP build-up rate (from -300 to -100ms relative to the error detection report on each trial). Thus, the new CPP build-up rate window was temporally dissociated from the FCθ power window across trials and, in the majority of cases, occurred later in time. The results from this mediation analysis, which we now report in the manuscript (in the subsection “Signal interactions predict the timing of error detection”), were essentially the same as our original attempt: all path coefficients were statistically significant at p<0.001, as was the critical mediation effect (see new Figure 5).
7) The first-order CPP increases for both go and no-go trials, consistent with the notion of a domain-general accumulator. But the second-order CPP increases only on error trials (Figure 2B). If the second-order decision is analogous to the first (i.e. a race between neural subpopulations controlling whether or not to press the button based on fluctuating evidence), then presumably there should also be accumulation towards a no-go bound in the second-order, confident-in-being-correct case. Doesn't this mismatch between first-order and second-order CPP features weaken the conclusion that the same DV mechanism underpins first- and second-order decisions? It would be interesting to hear the author's perspective on this. For instance the second-order CPP could reflect evidence accumulation from endogenous activities towards an overt error detection response, but the authors also argue the first-order CPP is also involved in evidence accumulation leading to a non-response (Figure 2—figure supplement 1).
The reviewers again highlight some very pertinent issues that warrant further exposure in the manuscript. In fact, while we do assert that the same neural signal traces first- and second-order decision processes, we do not contend that the two processes are equivalent in every respect. For their first-order task, we asked participants to make discrimination decisions about the state of a stimulus: should a prepotent task response be either executed or inhibited, given the current stimulus information. These decisions required active processing of the stimulus information presented on each trial and, as we (subsection “Task behaviour”) and the reviewers point out, are typically understood in terms of a race between two competing accumulators – one for ‘Go’ and one for ‘Don’t Go’ (Boucher et al., 2007; Gomez et al., 2007; Logan & Cowan, 1984; Verbruggen & Logan, 2009). In this kind of context, where two distinct alternatives must be arbitrated between given the current sensory information, we have previously shown that the CPP reflects the evidence accumulation process for decisions in favor of either alternative (Kelly & O’Connell, 2013). By contrast, for their second-order task of error detection, participants needed to engage in a process that is akin to signal detection: if there is sufficient evidence available that an error has been committed, then simply execute a response to indicate this. We (O’Connell et al., 2012) and others (e.g. Hillyard et al., 1971) have shown that the CPP/P3 in such detection contexts is clearly evident for hits (detected errors in the present context) but absent for misses (undetected errors).
Moreover, there is strong evidence to suggest that discrimination and detection decisions are processed differently in the brain, in ways that corroborate our above distinction between the first- (discrimination) and second-order (detection) decision processes on our paradigm. Much neurophysiological data indicates that two- or multi-alternative discriminations invoke a process of neural evidence gathering that manifests in build-to-threshold dynamics in competitive populations of neurons that are selective for each choice alternative (e.g. Gold & Shadlen, 2007) – that is, any given decision will be preceded by an evoked increase in activation of one pool of neurons. By contrast, simple perceptual detection tasks are thought to be determined by a competition between ‘No/stimulus absent’ and ‘Yes/stimulus present’ pools of neurons (e.g. Deco et al., 2007). The ‘No’ pool receives a static input and dominates the competition by default, whereas the ‘Yes’ pool is activated by stimulus information (or in our scenario, by ‘error evidence’). On hit trials (detected errors), the 'Yes' pool eventually dominates the competition and becomes highly activated; on miss trials (undetected errors), it fails to do so and there ends up being little-to-no evoked increase in firing rate in either pool. Thus, discrimination and detection decisions are clearly distinguishable in terms of neural dynamics.
We believe that the dissociable dynamics of the first-order (domain-general) and second-order (selective to error evidence) CPPs that we observed serve to highlight a critical difference in the nature of the two decision processes. The above considerations motivated our statements in the original manuscript that the second-order decision process appeared to be qualitatively distinct from the first-order process and involved the selective accumulation of error evidence (paragraph five of the subsection “A centro-parietal signature of first- and second-order evidence accumulation”). However, we now recognize that these issues, and in particular the distinction between discrimination and detection decisions, warrants further unpacking and we have now attempted to elaborate appropriately in the Discussion section of the revised submission (paragraph three).
[Editors' note: further revisions were requested prior to acceptance, as described below.]
Reviewer #2:
The authors have mostly addressed my previous concerns in the first round of reviews. The one part that I don't think has been fully addressed is the question of whether there are actually differences between undetected errors and go trials in the data that would not be captured by the model (in general I'm happy with the response to point 1 that the connection between first- and second-order accumulation can be left for future work, but it seems important to fully display the data features that might inform such work). For instance the ERP analysis in Figure 3—figure supplement 1 only directly contrasts undetected and detected errors, but there appears to be a difference between go and undetected errors later in the time course. Similarly only go-trial build-up rates are documented in Figure 2A and Figure 3D. I suggest including the comparison between go and undetected CPP and FCθ timecourses as a supplemental figure to cover this issue.
We apologize for this omission and thank the reviewer for again bringing it to our attention. We have now included an additional figure supplement as recommended by the reviewer (labelled ‘Figure 3—figure supplement 1’ in the latest submission), which displays direct comparisons of the CPP, fronto-central ERP and FCθ signals on Go trials and undetected errors. These analyses did not yield clear and robust effects for the CPP and FCθ signals (panels A and C of the new supplement), but as the reviewer correctly points out, there is a clear difference in fronto-central ERP amplitude that manifests from approximately 400ms post-response (panel B). We do not have a clear hypothesis about the functional significance of this effect and do not discuss it in detail in the revised submission. But we do acknowledge that the effect hints that these trial types may be at least partially dissociable in terms of underlying neural dynamics. We now draw attention to this interesting point in the revised manuscript (Discussion, paragraph six) in the context of our discussion about the inability of our model to provide a precise explanation for the apparent equivalence of the CPP on Go trials and undetected errors.
https://doi.org/10.7554/eLife.11946.023Article and author information
Author details
Funding
Irish Research Council (Postgraduate Fellowship)
- Peter R Murphy
European Research Council (Starting Grant 63829)
- Redmond G O'Connell
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
This work was supported by an Irish Research Council (IRC) “Embark Initiative” grant (PRM) and a European Research Council (ERC) Starting Grant (ROC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 638289). The authors thank Simon Kelly and Sander Nieuwenhuis for helpful discussions. The authors declare no conflicts of interest.
Ethics
Human subjects: Subjects provided written informed consent, and all procedures were approved by the Trinity College Dublin ethics committee and conducted in accordance with the Declaration of Helsinki.
Reviewing Editor
- Michael J Frank, Brown University, United States
Publication history
- Received: September 29, 2015
- Accepted: December 17, 2015
- Accepted Manuscript published: December 19, 2015 (version 1)
- Version of Record published: January 28, 2016 (version 2)
Copyright
© 2015, Murphy et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 3,303
- Page views
-
- 794
- Downloads
-
- 95
- Citations
Article citation count generated by polling the highest count across the following sources: Scopus, Crossref, PubMed Central.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Neuroscience
Early-life experiences may promote stereotyped behavioral alterations that are dynamic across development time, but also behavioral responses that are variable among individuals, even when initially exposed to the same stimulus. Here, by utilizing longitudinal monitoring of Caenorhabditis elegans individuals throughout development we show that behavioral effects of early-life starvation are exposed during early and late developmental stages and buffered during intermediate stages of development. We further found that both dopamine and serotonin shape the discontinuous behavioral responses by opposite and temporally segregated functions across development time. While dopamine buffers behavioral responses during intermediate developmental stages, serotonin promotes behavioral sensitivity to stress during early and late stages. Interestingly, unsupervised analysis of individual biases across development uncovered multiple individuality dimensions that coexist within stressed and unstressed populations and further identified experience-dependent effects on variation within specific individuality dimensions. These results provide insight into the complex temporal regulation of behavioral plasticity across developmental timescales, structuring shared and unique individual responses to early-life experiences.
-
- Computational and Systems Biology
- Neuroscience
Inhibition is crucial for brain function, regulating network activity by balancing excitation and implementing gain control. Recent evidence suggests that beyond simply inhibiting excitatory activity, inhibitory neurons can also shape circuit function through disinhibition. While disinhibitory circuit motifs have been implicated in cognitive processes including learning, attentional selection, and input gating, the role of disinhibition is largely unexplored in the study of decision-making. Here, we show that disinhibition provides a simple circuit motif for fast, dynamic control of network state and function. This dynamic control allows a disinhibition-based decision model to reproduce both value normalization and winner-take-all dynamics, the two central features of neurobiological decision-making captured in separate existing models with distinct circuit motifs. In addition, the disinhibition model exhibits flexible attractor dynamics consistent with different forms of persistent activity seen in working memory. Fitting the model to empirical data shows it captures well both the neurophysiological dynamics of value coding and psychometric choice behavior. Furthermore, the biological basis of disinhibition provides a simple mechanism for flexible top-down control of the network states, enabling the circuit to capture diverse task-dependent neural dynamics. These results suggest a biologically plausible unifying mechanism for decision-making and emphasize the importance of local disinhibition in neural processing.