Abstract
The formation of a subjective sense of confidence often requires the integration of signals from multiple sources of evidence. This is of particular relevance when one needs to determine with a high degree of certainty whether a multisensory stimulus is present or absent. Understanding the mechanisms underlying this ability to map evidence strength from multiple modalities into a single confidence value is therefore central to the study of metacognition. To this end, we asked healthy adults to detect the presence or absence of near-threshold stimuli that could be visual, auditory, or both, and then rate their confidence. In two pre-registered experiments (N = 48 and N = 54), audiovisual stimuli were better detected than unimodal ones, but were not associated with better metacognitive performance. Surprisingly, participants were more confident in their absence than presence judgments. To explain these results, we fitted a Bayesian evidence accumulation model in which sensory evidence is available for presence only, rendering decisions about absence dependent on counterfactual inference. The model reproduced decision patterns by assuming that a stimulus was perceived if sensory evidence from either modality exceeded a threshold (a disjunctive integration rule). In contrast, it reproduced confidence judgments by assuming that high confidence requires that the two modalities align (conjunctive for presence, disjunctive for absence). Together, these findings reveal that distinct computational mechanisms drive perception and confidence when detecting near-threshold multisensory signals.
Introduction
Imagine being outside at night with soft music playing and dim light around you. If you faintly hear the buzz of a mosquito, are you more confident that it is present if you also catch a glimpse of it, and vice versa? Or is it sufficient for you to just see it – or to just hear it – to be confident that it is around? And if you neither see nor hear a mosquito, how can you be sure that no mosquito is around?
Using audiovisual detection tasks in the laboratory, it is now clearly established that human adults typically show higher detection rates and faster reaction times for bimodal stimuli than for unimodal stimuli (Rach et al., 2011 ; Plass & Brang, 2021). This behavior is well explained by evidence accumulation models, which assume that sensory evidence from each modality is accumulated and processed together until a decision threshold is reached, with decision times reflecting the time to reach the threshold (Gondan et al., 2004; Gondan et al., 2005; Plass & Brang, 2021; Egan et al., 2025). These models account for faster decision times in bimodal than in unimodal conditions, but provide limited insight into decision accuracy. To address this, Blurton et al. (2014) introduced a drift-diffusion model with two thresholds, enabling predictions of both decision times and accuracy in a go/no-go task in which participants had to detect a target stimulus and ignore distractors. Since a stimulus was always present in this task, no comparisons could be made between decisions about the presence versus absence of sensory evidence. In addition, these models did not explain people’s ability to monitor the modality through which the stimulus was perceived.
If sensory evidence is only available for presence, a possibility is that absence is inferred from counterfactual perceptibility (i.e., “I would have perceived a stimulus if it was present”) – a form of metacognitive belief (Mazor, 2025), with metacognition referring to the ability to monitor and control our own cognitive processes (Flavell, 1979). A recent modeling study showed that optimal visual detection relies on participants integrating factual sensory evidence for presence, and inferring absence from the absence of evidence (Mazor et al., 2025). This approach differs from previous evidence accumulation models, which typically assume that the same accumulation process similarly drives decisions about presence and absence. This Bayesian model was not only capable of explaining the features of detection responses (accuracy and response times) but also the confidence associated with these decisions, showing that confidence in absence relies on the use of “counterfactual evidence”: beliefs about the evidence that would have been observed had a stimulus been present (Schipper & Mazor, 2025). This account helps clarify previously observed differences in metacognitive monitoring (i.e., the ability to make accurate confidence judgments about the results of our cognitive processes; Flavell, 1979; Fleming et al., 2012) between detection and discrimination tasks. Metacognitive judgments have been found to be more precise after discrimination than detection tasks, and this seems to be explained by the specific nature of absence judgments, which may rely on higher-order inferential processes (Meuwese et al., 2014; Kellij et al., 2020; Mazor et al., 2020; Mazor & Fleming, 2020; Mazor et al., 2023). However, in everyday life, our percepts are essentially multisensory, so a realistic model of detection – and confidence in these decisions – should also be able to explain how people judge the presence and absence of multisensory objects. This is a non-trivial problem, because in a multisensory context, observers have to combine sensory channels in order to decide if something is present or not, and such a combination can be achieved in several ways. Furthermore, it is unclear whether the same integration rules apply to detection and confidence.
We still have limited insight into these questions because how multisensory information impacts confidence judgments is only a recent domain of interest (Deroy et al., 2016). Recent studies comparing bimodal and unimodal conditions showed that, despite improvements in task performance, metacognitive efficiency was not improved in multisensory compared to unisensory conditions (Charles et al., 2020; Arbuzova et al., 2021; Faivre et al., 2018). Crucially, however, all these studies have used discrimination tasks, leaving the impact of a multisensory context on confidence in absence unknown.
Here, our goal was to understand how confidence judgments about the presence or absence of a stimulus are formed when evidence is available across multiple sensory channels. To this end, in two preregistered experiments, participants judged whether a stimulus was present or absent, irrespective of the modality, and rated their confidence in this amodal decision before providing modality-specific judgments and confidence ratings. Based on the literature relying on discrimination tasks, we hypothesized that multisensory stimuli would be associated with improved detection performance, but not metacognitive efficiency (Charles et al., 2020; Arbuzova et al., 2021; Faivre et al., 2018). We also expected participants to be more confident and have a higher metacognitive efficiency for presence than for absence judgments (Mazor et al., 2020; Mazor & Fleming, 2020; Meuwese et al., 2014; Kellij et al., 2020). We assessed whether factual and counterfactual reasoning applied to audiovisual detection by extending the model of Mazor and collaborators (2025) to the audiovisual domain. Finally, we examined the audiovisual integration rules governing confidence in presence and absence by deriving confidence from the probability of being correct at the time of the decision. Together, our behavioral and modeling results support the view that inferring audiovisual presence relies on a combination of factual and counterfactual reasoning, with distinct integration rules across sensory modalities for detection and confidence judgments.
Results
In two pre-registered online experiments (Exp.1: https://osf.io/3nvyx, Exp.2: https://osf.io/ehndv), participants performed an audiovisual detection task in which the stimulus could be either only visual, only auditory, or audiovisual. At the beginning of the experiment, stimulus intensity was calibrated for each participant to reach a 50% detection rate in visual and auditory conditions. Participants indicated whether a stimulus was present or absent irrespective of its sensory modality, before reporting their amodal confidence in their detection choice on a scale from 0 (“sure incorrect”) to 100 (“sure correct”). Finally, they reported their modality-specific detection and confidence judgments on a bi-dimensional (audio/visual) report scale, with each axis ranging from 100% sure not perceived to 100% sure perceived, and corresponding to one modality (see Fig.1). In Experiment 1, each experimental condition was equiprobable; in Experiment 2, we increased the proportion of trials without any signal to obtain an equal proportion of target absent and target present trials. Data from 48 participants were included in the main analysis for Experiment 1, and 54 for Experiment 2.

Trial structure.
1) Amodal detection: Participants were asked to indicate if they perceived or not a stimulus, irrespective of its modality. In frame: different possible stimulus types: the auditory stimulus was a sinusoidal tone of 1kHz presented in pink noise, the visual stimulus a light gray circle presented in dynamic Gaussian noise. 2) Amodal confidence: Participants indicated their confidence in their detection answer. 3) Modality-specific judgments and confidence: By moving a cursor on a bi-dimensional scale, participants indicated simultaneously whether they perceived a stimulus in each modality, and with which level of confidence.
Response bias towards absence
In both experiments, participants demonstrated good task sensitivity (Exp.1: d’ = 1.59 (SD = 0.64); Exp.2: d’ = 1.77 (SD = 0.61)). They also had a significant bias toward responding “absent”, with a Bayesian one sample t-test showing a significant positive response criterion (Exp.1: c = 0.58, SD = 0.43, BF1 > 1000; Exp.2: c = 0.43, SD = 0.48, BF1 > 1000) (see Fig.2A).

Amodal detection.
Experiment 1 is represented by circles, while Experiment 2 is represented by diamonds; error bars represent the standard error. A) Amodal performance: amodal d’ and criterion. B) Amodal response: Percentage of stimuli judged to be present as a function of the experimental condition.
Improved detection for multisensory stimuli
In both experiments, Bayesian logistic regression analyses on amodal accuracy revealed that participants were better at detecting bimodal than unimodal stimuli (Exp.1: 



Ability to monitor the source of the percept
Our bi-dimensional report scale allowed us to assess how accurately participants identified the source of their percepts. Focusing on correct presence judgments, we performed exploratory analyses to investigate the ability of participants to accurately categorize the modality in which the stimulus was presented — a form of source monitoring (“did I see or hear this stimulus?’) Participants demonstrated reliable source monitoring: auditory stimuli were correctly categorized as auditory on 85% of trials in Experiment 1 (binomial test against 50%, p < .001) and 79% in Experiment 2 (p < .001), while visual stimuli were correctly categorized as visual on 75% (p < .001) and 82% of trials (p < .001), respectively.
For audiovisual stimuli, participants categorized them as auditory in 27% (Exp.2: 17%), visual in 39% (Exp.2: 45%), and audiovisual in 33% (Exp.2: 37%) of trials. These proportions are consistent with the idea that detecting stimuli as audiovisual is based on the joint probability of detecting them visually and auditorily. A Wilcoxon signed-rank test comparing observed audiovisual reports to the predicted joint probability revealed no significant difference (Exp.1: V = 764, p = .07, Exp.2: V =811, p = .56). This indicates that audiovisual categorizations do not violate the predictions from independent detection in the visual and auditory modalities. Finally, participants did not change their minds and attributed perceived presence to at least one modality after indicating that the stimulus was present, with only 1% of detected stimuli judged as being absent in both modalities.
Perceptual multisensory interference
Our bimodal confidence scale also enabled us to investigate multisensory interference (i.e., how does one modality influence the other). We found that the presence of one modality biased participants toward reporting that a stimulus was also present in the other, especially when it was, in fact, not the case. Indeed, the presence of a visual stimulus biased participants to respond that an auditory stimulus was present only when the auditory stimulus was actually absent. Likewise, the presence of auditory stimulus also biased participants to respond that a visual stimulus was present only when it was actually absent (see Table 1 and SI).

Effect of the presence of a stimulus in the other modality
Confidence in presence vs absence
Based on previous findings (Mazor et al., 2020; Mazor et al., 2025), we hypothesized that participants would be more confident in their presence than in their absence judgments. Contrary to our hypothesis, in both experiments, participants reported higher amodal confidence following absence than presence judgments (Exp.1: 


Confidence by judgment.
Experiment 1 is represented by circles, while Experiment 2 is represented by diamonds. A) Amodal confidence as a function of hits, false alarms (FA), correct rejections (CR), and misses; error bars represent the standard error. B) Amodal metacognitive efficiency (response-conditional Mratio) as a function of the type of judgments; error bars represent the highest density interval. In both panels, presence judgements are represented in blue, and absence judgements in pink.
To evaluate metacognitive performance irrespective of task performance (Fleming & Lau, 2014, Maniscalco & Lau, 2012), we analyzed metacognitive efficiency (Mratio, estimated via response-conditional HMeta-d; Fleming, 2017) for presence versus absence judgments, both at the amodal and at the modality-specific level. In both experiments, and consistent with previous findings (Mazor et al., 2020; Mazor & Fleming, 2020), amodal metacognitive efficiency was higher for presence judgments than absence judgments (Exp.1: ΔM = 0.53 [0.27, 0.79]; Exp.2: ΔM = 0.62 [0.43, 0.81]) (see Fig.3B). Using the bidimensional scale, we found that this effect was also true at the modality-specific level: participants showed higher auditory metacognitive efficiency in auditory judged present compared to auditory judged absent trials (Exp.1: ΔM = 0.78 [0.53, 1.03]; Exp.2: ΔM = 0.67 [0.41, 0.93]), and higher visual metacognitive efficiency in visually judged present compared to visually judged absent trials (Exp.1: ΔM = 0.60 [0.40, 0.81]; Exp.2: ΔM = 0.35 [0.14, 0.56]).
Multisensory effects on metacognitive performance
To evaluate how accurately participants’ confidence tracked their accuracy, we measured metacognitive sensitivity as the slope of the Bayesian logistic regression predicting accuracy from confidence. Amodal metacognitive sensitivity was higher in bimodal than in unimodal trials (Exp.1: 



Multisensory effects.
Experiment 1 is represented by circles, while Experiment 2 is represented by diamonds. A) Amodal metacognitive sensitivity by modality as a function of the experimental condition, a slope superior to 0 indicates higher confidence in correct than incorrect responses and below 0 higher confidence in incorrect than correct responses; error bars represent the standard error. B) Amodal metacognitive efficiency (response-conditional Mratio) for presence judgments as a function of the modality of presentation; error bars represent the highest density interval.
Additionally, we analyzed metacognitive efficiency across stimulus modalities (see Fig.4B). Focusing on presence judgments, in Experiment 1, no credible difference was found between unimodal and bimodal trials (ΔM = -0.25 [-0.59, 0.10]). There was also no clear difference in metacognitive efficiency between auditory and visual trials (ΔM = 0.27 [-0.15, 0.69]), between auditory and audiovisual trials (ΔM = -0.30 [-0.67, 0.07]), or between visual and audiovisual trials (ΔM = -0.03 [-0.29, 0.26]). In contrast, in Experiment 2, we observed higher metacognitive efficiency in unimodal compared to bimodal trials (ΔM = -0.28 [-0.54, -0.02]). This effect was driven by auditory trials with both a higher metacognitive efficiency for auditory compared to visual trials (ΔM = 0.32 [0.03, 0.59]) and for auditory compared to audiovisual trials (ΔM = -0.38 [-0.63, -0.13], while there was no evidence for a difference in metacognitive efficiency between visual and audiovisual trials (ΔM = -0.06 [-0.29, 0.16]). Finally, we found strong evidence for the existence of correlations between metacognitive efficiencies in the audiovisual and auditory domain (Exp.1: rho = 0.86 [0.61, 0.99]; Exp.2: rho = 0.78 [0.42, 0.99]), between the audiovisual and visual domain (Exp.1: rho = 0.89 [0.72, 0.99]; Exp.2: rho = 0.92 [0.51, 0.99]), and between the auditory and visual domain (Exp.1: rho = 0.85 [0.61, 0.99]; Exp.2: rho = 0.76 [0.25, 0.98]), consistent with other findings on the supramodality of metacognition (Rouault et al., 2018; Faivre et al., 2018; Ais et al., 2016).
Metacognitive multisensory interference
As preregistered, we also examined multisensory interference, this time at the metacognitive level. Our results suggest a cross-modal facilitation when both modalities provide consistent cues: in both experiments, participants showed higher auditory metacognitive sensitivity for absence when both the visual and the auditory stimuli were absent and higher visual metacognitive sensitivity for presence when both the visual and the auditory stimuli were present. We also observed additional but less robust influence, present in only one of the two experiments; given this lack of reproducibility, we do not interpret these effects further (see Table 1 and SI).
Model
We extended a recent ideal-observer model of visual detection to account for our multisensory detection task (Mazor et al., 2025). In the original model, observers determine the presence of a visual object depending on the activation of a single “presence sensor”. At each time point, this sensor samples either a 1 or a 0, and the probability of sampling a 1 is higher when the target is present. Sensor activation probabilities are captured by model parameters θpresent (i.e., probability of sampling 1 when a target is present) and θabsent (i.e., probability of sampling 1 when a target is absent). Importantly, this model assumes that agents hold beliefs about these probabilities (i.e., beliefs about the probability of sampling a 1, if the target is present or if it is absent). These beliefs are captured by model parameters 

To adapt this model for audiovisual detection, we implemented two modality-specific sensors: a visual and an auditory sensor (see Fig.5). The model also allowed holding distinct prior beliefs regarding the probability of presence in each modality. At each time point, the posterior probability of a stimulus being present was updated independently in each sensory channel. Because participants were instructed to respond whenever they detected a stimulus, regardless of the sensory modality, we combined the two channels into an amodal estimate using a disjunctive integration rule, such that p(present)=p(Apresent)+p(Vpresent)-p(Apresent and Vpresent), reflecting that a stimulus can be (objectively) present in only one modality (A, V) or in both modalities (AV). We compared it to a conjunctive integration rule according to which a stimulus was judged present only if present in both modalities (see SI for comparison).

Computational model.
A) Model architecture. The observer is assumed to have access to a visual sensor and an auditory sensor, probabilistically tuned to the presence of visual and auditory evidence. The probability of activation is controlled by the parameter θ. The model agent observes the activations and updates their beliefs about the presence of a signal in each modality separately, using Bayes’ rule. The agent then integrates the two beliefs into an amodal belief in the presence of a target. Based on this belief, they decide whether to commit to a decision or accumulate more evidence by following an optimal policy, derived using backward induction. B) Example trial: modality-specific log-likelihood ratios (LLR, in green) are updated following sensor inactivations and activations. C) Integration rules: The top plot represents the disjunctive rule and the bottom plot the conjunctive rule. Amodal LLR is plotted as a function of the number of sensor activations in each modality 50 time points (i.e., 2.5s) into the trial. Black contours indicate regions in which the best action is to decide present, wait, or decide absent.
Although the model was only fitted to amodal detection accuracy and response time data, it generated qualitative predictions about amodal confidence based on the probability of being correct at the time of the decision. Furthermore, it made predictions about modality-specific effects as it computed the probability of presence separately for each modality: a stimulus was judged present if the modality-specific probability of presence was greater than 0.5 at the time of the decision, and absent otherwise. Modality-specific confidence was also read as the probability of being correct at the time of the decision, for each modality separately.
To assess the model’s ability to reproduce qualitative patterns observed in our behavioral data, we simulated datasets using a model in which belief parameters 
Reproduction of perceptual effects
The model successfully reproduced amodal detection behavior, with a mean d’ of 1.67 (SD = 0.86) and a mean criterion of 0.59 (SD = 0.49) for Experiment 1, and a mean d’ of 1.73 (SD = 0.86) and a mean criterion of 0.46 (SD = 0.47) for Experiment 2. It also reproduced higher accuracy for audiovisual trials than for unimodal trials, and for visual trials than for auditory trials in both experiments (see Fig.6A; Fig.6D; fits for the response times are shown in Fig.S6

Reproduction of perceptual effects.
Error bars represent the standard error from the data. Rectangles represent data simulated from the model, centered on the mean value and with height equal to the standard error. Left panels show the fit for Experiment 1 and right panels for Experiment 2. A-F) Percentage of stimuli judged to be present as a function of the condition of presentation for Experiments 1 and 2, at the amodal, auditory, or visual level. G-J) Observed and simulated source monitoring: Modality detected as a function of the modality of presentation in Experiments 1 and 2.
Despite being trained only to amodal detection responses, the model also reproduced the observed modality-specific detection patterns in both experiments (see Fig.6B-C; Fig.6E-F). It also reproduced participants’ capacity for accurate source monitoring (see Fig.6G-J). Notably, only 5% of stimuli judged to be present were modeled as absent in both modalities, suggesting that the model captured participants’ response consistency by attributing perceived presence to at least one modality.
Reproduction of confidence effects
Having shown that optimal amodal detection involved integrating sensory evidence according to a disjunctive rule (i.e., a stimulus is detected based on auditory or visual evidence), we sought to test whether the same rule applied for amodal confidence. To do so, we tested whether, at the time of the decision, amodal confidence was based on the probability of being correct according to a disjunctive or conjunctive integration of information, separately for presence and absence judgments. If amodal confidence follows a disjunctive integration rule, confidence in presence should be high when only one modality indicates presence, whereas confidence in absence should be high only when both modalities indicate absence (i.e., negation of the disjunction). On the other hand, if amodal confidence follows a conjunctive integration rule, confidence in presence should be high only when both modalities indicate presence, whereas confidence in absence should be high when only one modality indicates absence (i.e., negation of the conjunction).
When confidence followed a disjunctive rule, the model failed to capture important aspects of the data, such as the higher confidence for presence than absence judgments (Fig.7, left panels). On the other hand, when confidence followed a conjunctive rule, it reproduced confidence in presence judgments but failed to capture variability in confidence ratings for absence judgments (Fig. 7, middle panels). Critically, a combination of disjunctive and conjunctive integration rules, for absence and presence judgments respectively, reproduced the confidence effects we observed in both experiments (while being generally over-confident relative to human participants across the board). Namely, it predicted higher confidence in absence than in presence judgments, higher metacognitive sensitivity for bimodal compared to unimodal trials, and higher confidence for correct than for incorrect absence judgments (see Fig.7, right panels). This combined rule can be alternatively described in the following way: high confidence, whether in presence or in absence, requires that the two sensors agree: both indicate the presence of a signal for high confidence “presence” judgments, or that both indicate signal absence for high confidence “absence” judgments.

Confidence fits according to the different integration rules.
Error bars represent the standard error from the data. Rectangles represent data simulated from the model, centered on the mean value and with height equal to the standard error. Top plots represent the fit for Experiment 1 and bottom plots for Experiment 2. Each plot represents the amodal confidence as a function of condition of presentation and as a function of amodal hits, false alarms (FA), correct rejections (CR), and misses. The left panels represent the confidence based on the disjunctive rule. The middle panels represent the confidence based on the conjunctive rule. The right panels represent the confidence when absence is based on the disjunctive rule, while presence is based on the conjunctive rule.
Despite being fitted to amodal decisions only, the model captured the observed modality-specific confidence effects by reproducing confidence as a function of participants’ response and accuracy both in the visual and in the auditory modality (see Fig.8). Finally, we tested if the model captured interindividual variability in confidence asymmetries between the auditory and visual modalities, reflecting a propension to give more weight to one sensor when estimating confidence following an audiovisual stimulus. To do so, we defined for each participant a “confidence asymmetry index” capturing the difference between auditory and visual confidence in audiovisual trials, normalised by absent trials:


Reproduction of modality-specific confidence effects.
Error bars represent the standard error from the data. Rectangles represent data simulated from the model, centered on the mean value and with height equal to the standard error. Left panels represent the fit for Experiment 1 and right panels for Experiment 2. A) Auditory confidence as a function of auditory hits, false alarms (FA), correct rejections (CR), and misses for Experiment 1. B) Visual confidence as a function of visual hits, false alarms (FA), correct rejections (CR), and misses for Experiment 1. C) Correlation between observed and simulated data for the confidence asymmetry index for Experiment 1. D) Auditory confidence as a function of auditory hits, false alarms (FA), correct rejections (CR), and misses for Experiment 2. E) Visual confidence as a function of visual hits, false alarms (FA), correct rejections (CR), and misses for Experiment 2. F) Correlation between observed and simulated data for the confidence asymmetry index for Experiment 2. In all panels, presence judgements are represented in blue, and absence judgements in pink.
A significant positive correlation was observed between the observed and simulated confidence asymmetry indices (Exp.1: Pearson’s rho = .84, p < .001; Exp.2: Pearson’s rho = .61, p < .001), indicating that the model successfully captured the interindividual variability of visual and auditory weights on confidence.
Discussion
Although everyday perception is inherently multisensory, we know surprisingly little about the way people judge whether something is present or absent across multiple sensory channels, and how confident they are in such judgments. To address this, in two preregistered experiments, participants performed an audiovisual task at unimodal near-threshold intensity. On each trial, they reported whether a stimulus was present irrespective of the modality of presentation before reporting their amodal confidence in their answer, and finally, their modality-specific judgments and confidence.
To investigate audiovisual integration rules, we adapted a recent Bayesian evidence accumulation framework assuming that absence is inferred from counterfactual detectability (Mazor et al., 2025). Using a disjunctive integration rule according to which a stimulus is detected when at least one modality provides sufficient evidence, we reproduced the higher detection performance for audiovisual stimuli compared to unimodal ones. Importantly, our model showed that audiovisual signals were processed differently at the perceptual and metacognitive levels. The model successfully reproduced the observed amodal confidence patterns using a disjunctive rule for absence judgments (i.e., high confidence in absence only when neither modality provided sufficient evidence), but using a conjunctive rule for presence judgments (i.e., high confidence in presence only when both modalities provided sufficient evidence). This shows that high confidence in absence or presence occurs when the auditory and the visual channels are aligned. This suggests that, while detection decisions relied on a disjunctive process, where evidence from a single modality is sufficient to judge that something is present, at the confidence level, intersensory congruency played a critical role. This interpretation is further supported by the modality-specific confidence effects we observed, which indicated cross-modal facilitation at the metacognitive level when both modalities provided consistent cues.

Illustration of the integration rules process.
Detection decisions (red for absence, blue for presence) are based on the disjunctive integration rule (disjunction and negation of disjunction). Confidence decisions (dashed line for not sure, full line for sure) are either based on a conjunctive rule (confidence in presence) or a negation of disjunction (confidence in absence).
Looking more closely at multisensory effects, we replicated previous findings showing higher detection performance, and, for correct responses only, faster response times for audiovisual compared to unimodal stimuli (Rach et al., 2011; Plass & Brang, 2021). Using a bidimensional modality-specific scale, we also observed that participants accurately monitor the source of their percepts, by identifying the modality in which the stimulus was presented. At the metacognitive level, despite an improvement in metacognitive sensitivity, there was no boost in metacognitive efficiency of presence judgments for bimodal compared to unimodal stimuli. Experiment 2 even showed a higher metacognitive efficiency for auditory trials; given that this effect did not replicate across experiments, further work is needed to test its reliability. Prior work has similarly reported no differences in metacognitive efficiency between bimodal and unimodal stimuli during discrimination tasks (Charles et al., 2020; Arbuzova et al., 2021; Faivre et al., 2018), and our findings extend these results to detection tasks. Furthermore, we found strong correlations in metacognitive efficiency across modalities, consistent with findings suggesting the supramodality of metacognition (Rouault et al., 2018; Faivre et al., 2018; Ais et al., 2016). Finally, the modality-specific scale we developed provides a promising tool for investigating perceptual and metacognitive asymmetries in the presence or absence of multisensory stimuli, particularly in populations with sensory impairments, who may differ in both factual and counterfactual reasoning.
When investigating the metacognitive monitoring of absence, we found that participants were overall more confident in their absence than in their presence judgments, both at the amodal and modality-specific levels. This is in contradiction with some previous findings (Meuwese et al., 2014; Kellij et al., 2020; Mazor et al., 2025), but consistent with other studies (Pereira et al., 2021; Stockart et al., 2025; Dijkstra et al., 2024). As there are multiple differences between these experimental paradigms, it is difficult to pinpoint what could drive higher confidence in absence. To examine whether stimulus intensity contributes to this confidence pattern, we performed an additional experiment in which unimodal stimuli were presented at a suprathreshold intensity (see SI for detailed results). In this configuration, participants were still more confident in their absence than in their presence judgments, suggesting that increased sensory evidence is not sufficient to restore high confidence in presence. Future research will be needed to further investigate the role of the multisensory context on this effect. Indeed, if absence judgments rely on counterfactual reasoning, one can hypothesize that when multiple sensory sources are available but no stimulus is perceived in any of them, the belief that you should have perceived the stimulus if it was present could become even stronger, potentially increasing confidence in absence. Nevertheless, in all our experiments, despite higher confidence in absence, participants had a higher metacognitive efficiency for presence judgments than for absence judgments, consistent with previous findings showing lower metacognitive performance for absence (Mazor et al., 2020; Mazor & Fleming, 2020).
Finally, although our model was fitted only to amodal detection decisions, it successfully reproduced modality-specific detection performance, and therefore participants’ ability to correctly monitor the source of their percept. These results are in line with a recent study showing that information from different sensory modalities is processed separately before being integrated to reach a common decision threshold (Egan et al., 2025). It also reproduced modality-specific confidence ratings for the two modalities, with confidence defined as the probability of being correct at the time of the decision. This result corroborates previous findings showing that, in some settings, confidence closely matches the probability of being correct, especially when decision and confidence are reported simultaneously (Pouget et al., 2016; Aitchison et al., 2015).
In summary, during an audiovisual detection task at unimodal near-threshold intensity, although the presence of two sensory sources of evidence instead of one improved detection performance, it did not improve metacognitive efficiency. Our ideal observer model, equipped with two modality-specific sensors and making decisions based on a disjunctive integration rule, successfully reproduced amodal and modality-specific detection and modality-specific confidence ratings. However, amodal confidence ratings were successfully reproduced only when presence and absence judgments relied on distinct integration rules. This shows that intersensory congruency plays a critical role in confidence judgments. Overall, these findings indicate that different integration rules apply to perceptual and metacognitive decisions and underscore the importance of counterfactual reasoning for absence judgments. They further suggest that – counterintuitively – a multisensory context might not only impact the perception that something is present, but also the perception that something is absent.
Method
Protocol
In two pre-registered online experiments (Exp.1: https://osf.io/3nvyx, Exp.2: https://osf.io/ehndv), participants performed an audiovisual detection task in which they indicated whether a stimulus was present or absent regardless of the modality of presentation. Data were collected from participants recruited via the Prolific platform (N = 60 in Experiment 1; N = 61 in Experiment 2). Participants listened to auditory pink noise while observing dynamic visual Gaussian noise, updated every 33 ms. In Experiment 1, a visual stimulus (a light gray circle spanning 7.5% of the screen size and presented at the center of the screen) was embedded in visual noise on half of the trials. Independently, an auditory stimulus (a sinusoidal tone of 1 kHz) was embedded in auditory noise on half of the trials. As a result, a signal was present on 75% of the trials, with 25% of the trials including both a visual and an auditory signal presented simultaneously (AV trials), and 25% of the trials including none (absent trials). In Experiment 2, we increased the proportion of trials without any signal to obtain an equal proportion of target absent and target present trials. As a result, half of the trials contained no stimulus (absent trials), and the other half contained a stimulus in either the auditory, visual or audiovisual modality in equal proportion (present trials).
When launching the experiment, participants had to calibrate the volume of the auditory noise to a comfortable level. Then, to check if they were using headphones instead of speakers, they had to judge if a sound was presented to their left or right ear in three trials. The experiment was aborted if one error was made.
Before starting the main task, stimulus intensity was calibrated for each participant to reach a 50% detection rate in visual and auditory conditions, based on unimodal psychometric curves. To compute these psychometric curves, participants had to detect stimuli of five different intensities, each presented 20 times, with an additional 20 trials where no stimulus was presented. Auditory and visual calibrations were conducted separately, with the order of presentation counterbalanced across participants.
Each calibration could be repeated if the detection threshold could not be estimated by the fitting procedure, a sign of poor behavior. Pilot study revealed an increase in stimulus detectability between the psychometric evaluation and the main experimental phase. To account for this increase and ensure a similar number of hit and miss trials during the main experiment, we defined stimulus intensity for the main experiment as the intensity corresponding to 40% of detected stimuli during the psychometric evaluation (see SI for individual psychometric curves).
Following ten training trials, participants undertook the main task. On each trial, the stimulus could appear 200, 300, or 400ms after the noise onset and was presented for 600ms. Participants could respond from the start of the trial until a 4-s time limit after the stimulus offset. They first pressed the right or left arrow on the keyboard to indicate whether a stimulus was present or absent irrespective of its sensory modality (the response key to report the presence or absence of the stimulus was counterbalanced across participants). Critically, they were instructed to report presence if a stimulus was present visually, auditorily, or both. Following a 100 ms delay, they reported their amodal confidence in their detection choice by moving a cursor with the mouse on a scale from 0 (“sure incorrect”) to 100 (“sure correct”). Participants were instructed to report an amodal confidence judgement reflecting decision accuracy irrespective of the sensory modality. Finally, they were asked to report their modality-specific detection and confidence judgments on a bi-dimensional (audio/visual) scale, with each axis ranging from 100% sure not perceived to 100% sure perceived, and corresponding to one modality. The mapping of auditory and visual modalities to the horizontal and vertical axes was counterbalanced across participants, although this counterbalancing was not preregistered for Experiment 1.
Participants performed 288 trials in Experiment 1 and 252 in Experiment 2, divided into six experimental blocks, each experimental condition was randomly presented within each block.
In Experiment 1, six participants were excluded based on our pre-registered exclusion criteria (no variability in confidence judgments and no convergence of psychometric curves). An additional five participants were excluded as their detection accuracy was 0% in either the auditory or the visual modality. As trials with confidence ratings below 50 were not analyzed, one participant was excluded as they responded with a confidence below 50 in 286 out of 288 trials. In Experiment 2, three participants were excluded based on our pre-registered exclusion criteria and an additional four participants were excluded as their detection accuracy was 0% in either the auditory or the visual modality. As a result, 48 participants were included in the main analysis for Experiment 1, and 54 for Experiment 2.
Data analysis
All statistical analyses were performed in R. To investigate amodal detection performance and metacognitive sensitivity, we conducted a mixed-effects logistic regression on accuracy as a function of experimental conditions (stimulus absent, auditory, visual, or audiovisual), amodal confidence, and their interaction. To compare reaction times, we conducted a mixed-effects linear regression of log-transformed RTs as a function of the experimental condition and type of judgment (results in SI).
To investigate modality-specific detection and confidence effects, we pre-registered four models depending on whether the stimulus was present or absent in the given modality, and whether it was judged as absent or present in that modality. For consistency with our amodal analyses, we deviated from the original pre-registration and applied the same model structure as for the amodal effects, that is a mixed-effects logistic regression on visual accuracy as a function of experimental condition, visual confidence, and their interaction, and a mixed-effects logistic regression on auditory accuracy as a function of experimental condition, auditory confidence, and their interaction. Full details of the pre-registered analyses are available in the SI.
To investigate confidence bias, we fitted two linear models: one comparing the mean of the confidence ratings as a function of the condition of presentation and response accuracy, and the other comparing the mean of confidence judgments as a function of participants’ judgments (stimulus judged absent or present) and accuracy.
To assess participants’ ability to monitor their absence judgments, we performed a mixed-effects logistic regression on the accuracy of absence judgments (miss vs correct rejections) as a function of amodal confidence.
Finally, we estimated metacognitive efficiency using response-conditional hierarchical M-ratio (Hmeta-d’, Fleming, 2017), both for amodal and modality-specific judgments.
We additionally performed unpreregistered analyses: a Bayesian one-sample t-test on participants’ response bias and a Wilcoxon signed-rank test to investigate participants’ source monitoring.
Model
The original model (Mazor et al., 2025) considered that when a stimulus is present it remains on the screen until participants made a decision. In contrast, in our experiments, the stimulus was presented for a fixed duration of 600 ms, with temporal uncertainty in its onset. To account for this limited and variable presentation period, we divided the evidence accumulation process into three distinct temporal phases: (1) before the earliest possible stimulus onset, (2) during the window of potential stimulus presentation, and (3) after the stimulus could no longer be present. Before the earliest onset, posterior probabilities updates were driven solely by prior beliefs about presence in each modality. During the stimulus presentation window, updating was based on both prior beliefs and incoming sensory input. After this window, the probability of accumulating evidence for presence decayed. We compared this model to a model without decay, in which evidence for presence was based on random noise during periods when no stimulus could be present, corresponding to the probability of sampling a 1 if the target is absent. We found that the model with a decay mechanism fitted the data better than the model without (difference in AIC: 2593.8).
Model comparison
We compared the goodness of fit to amodal decisions across different combinations of parameters representing the believed and true likelihoods (see Table 2). Model comparisons at the group level showed that the best model was the “single belief” model, according to which the believed probability of sampling a 1 (if a target is present or absent) is the same for auditory and visual signals. The model selection was based solely on the data from Experiment 1, and we tested the best model’s ability to generalize to new data with results from Experiment 2.

Fitted models (rows) with different or identical parameters (columns) across the visual and auditory sensors.
Supplementary information
Psychometric curves
Visual psychometric curves

Visual psychometric curve for each participant of Experiment 1

Visual psychometric curve for each participant of Experiment 2
Auditory psychometric curves

Auditory psychometric curve for each participant of Experiment 1

Auditory psychometric curve for each participant of Experiment 2
Data analysis
Contrast
For the model visual accuracy ~ condition * visual confidence, we used the following contrast coding for the condition of presentation: (-1, -1, 1, 1) to compare visually present versus absent trials; (-1, 1, 0, 0) to assess the effect of auditory information on visually absent trials; (0, 0, -1, 1) to assess the effect of auditory information on visually present trials.
For the model auditory accuracy ~ condition * auditory confidence, we used the following contrast coding: (-1, 1, -1, 1) to compare auditory present versus absent trials; (-1, 0, 1,0) to test the influence of visual information on auditory-absent trials; (0, -1, 0, 1) to test the influence of visual information on auditory-present trials.
Additional results
Analysis of Reaction times
In Experiment 1, reaction times did not differ between bimodal and unimodal trials (





In Experiment 2, evidence was inconclusive for a difference between bimodal and unimodal trials (





Finally, in both experiments, we observed a significant interaction between the condition of presentation and the type of judgment, the difference in response times between absence and presence judgments was larger in present compared to absent trials indicating that participants responded faster on correct-present trials (Exp.1: 

Source monitoring: Test for a visual dominance effect
A Wilcoxon test showed no difference between the proportion of visual-only and auditory-only reports relative to their predicted probabilities assuming independent visual and auditory detection (Exp.1: V = 421, p = .09; Exp.2: V = 870, p = .27). This shows that visual categorizations occurred at the level expected from the higher visual hit rate, with no evidence for additional visual dominance.
Response bias toward absence at the modality-specific level
Although participants reported whether they detected a stimulus regardless of modality, the modality-specific scale enabled us to infer detection performance for each modality. As for amodal detection, participants had a tendency to respond that both auditory and visual stimuli were absent, with Bayesian logistic regressions on modality-specific accuracy showing that participants had a higher auditory accuracy for auditory absent compared to auditory present trials (Exp.1: 



Multisensory interference
Influence of the visual modality on auditory judgments
In Experiment 1, participants were more accurate at detecting the absence of the auditory stimulus when the visual stimulus was also absent (




Modality-specific results.
Experiment 1 is represented by circles, while Experiment 2 is represented by diamonds. A) Auditory detection performance: Percentage of stimuli judged to be present at the auditory level as a function of the experimental condition; error bars represent the standard error. B) Auditory metacognitive sensitivity as a function of the experimental condition; error bars represent the standard error. C) Auditory metacognitive efficiency (response-conditional Mratio) for auditory present judgments as a function of the visual modality; error bars represent the highest density interval. D) Auditory metacognitive efficiency (response-conditional Mratio) for auditory absent judgments as a function of the visual modality; error bars represent the highest density interval. E) Visual detection performance: Percentage of stimuli judged to be present at the visual level as a function of the experimental condition; error bars represent the standard error. F) Visual metacognitive sensitivity as a function of the experimental condition; error bars represent the standard error. G) Visual metacognitive efficiency (response-conditional Mratio) for visual present judgments as a function of the auditory modality; error bars represent the highest density interval. H) Visual metacognitive efficiency (response-conditional Mratio) for visual absent judgments as a function of the auditory modality; error bars represent the highest density interval.
At the metacognitive level, results suggest a facilitative effect of audiovisual congruency. Auditory metacognitive sensitivity for absence was higher when there was also no visual stimulus (Exp.1: 



We further assessed auditory metacognitive efficiency as a function of the visual modality. In both experiments, when the auditory stimulus was judged present, the presence of a visual stimulus had no effect on metacognitive efficiency (Exp.1: ΔM = 0.24 [-0.19, 0.68]; Exp.2: ΔM = 0.10 [-0.30, 0.48]). In contrast, in Experiment 1, when the auditory stimulus was judged absent, metacognitive efficiency was significantly higher when the visual stimulus was absent (ΔM = -0.16 [-0.30, -0.01]). However, this effect did not reach significance in Experiment 2 (M = Δ-0.08 [-0.21, 0.06]) (see Supp. Fig.5C-D).
Influence of the auditory modality on visual judgments
We tested if the auditory modality also biased participants toward responding that something is present at the visual level only when the visual stimulus was actually absent. While this effect was inconclusive in Experiment 1 (

Visual metacognitive sensitivity for absence was higher when there was also no auditory stimulus in Experiment 2 (



We also examined visual metacognitive efficiency as a function of the auditory modality. In both experiments, there was no clear effect of the auditory modality when the visual stimulus was judged absent (Exp.1: ΔM = -0.04 [-0.15, 0.08]; Exp.2: ΔM = 0.06 [-0.09, 0.20]), nor when it was judged present (Exp.1: ΔM = 0.09 [-0.26, 0.44]; Exp.2: ΔM = -0.13 [-0.46, 0.20]) (see Supp. Fig. 5G-H).
Metacognitive sensitivity between presence and absence
We preregistered the estimation of metacognitive sensitivity conditional to the stimulation conditions. We found that amodal metacognitive sensitivity was higher for absent than for present trials, indicating more accurate confidence calibration in trials where no stimulus was present (Exp.1: 

We observed a similar pattern at the modality-specific level. Participants had a higher auditory metacognitive sensitivity for auditory absent than for auditory present trials (Exp.1: 





We additionally investigated response-conditional metacognitive sensitivity to investigate differences between presence and absence judgments. Response-conditional amodal metacognitive sensitivity showed no difference between presence and absence judgements in Experiment 1 (





Preregistered modality-specific analysis
Auditory modality
In experiment 1, when the audio stimulus was present, participants were better at detecting it as being auditory present when a visual stimulus was present (



When the audio stimulus was judged present, participants were able to distinguish between their hits and the false alarms (


In experiment 2, when the audio stimulus was present, participants adapted better their auditory confidence to their auditory accuracy for audiovisual compared to audio trials (

When the audio stimulus was judged present, participants were able to distinguish between their hits and false alarms at the auditory level (


Visual modality
In experiment 1, when the visual stimulus was present, participants adapted better their visual confidence to their visual accuracy for audiovisual trials compared to visual ones (


When the visual stimulus was judged present, participants were able to distinguish between their hit and the false alarms (


In experiment 2, when the visual stimulus was present, participants adapted better their visual confidence to their visual accuracy for audiovisual trials compared to visual ones (


When the visual stimulus was judged present, participants were able to distinguish between their hit and the false alarms(

Predicting Amodal Confidence from Modality-Specific Confidence Ratings
We pre-registered the investigation of the contribution of auditory and visual confidence to amodal confidence. Beyond this, we explored which model best predicted amodal confidence from modality-specific confidence. We compared different integration rules:
Max model: amodal confidence ~ max(auditory confidence, visual confidence)
Linear model: amodal confidence ~ auditory confidence * visual confidence (equal weight for each modality)
Weighted linear model: amodal confidence ~ auditory confidence_w * visual confidence_w (the weights of the auditory and visual confidence are determined by the data, thus allowing the relative contribution of each modality to vary independently)
Optimal integration model: amodal confidence ~ auditory confidence + visual confidence - auditory confidence*visual confidence
Weighted min-max model: max(auditory confidence, visual confidence) + min(auditory confidence, visual confidence)
Based on this first comparison, we found that the min-max model was the one that best explained amodal confidence. We compared different ways to improve this min-max model by adding:
An absolute confidence score computed as (absolute(minimal_confidence −0.5)+absolute(maximal_confidence −0.5)) /2: amodal confidence ~ max(auditory confidence, visual confidence) + min(auditory confidence, visual confidence) + absolute confidence
Absolute confidence and type of judgments: amodal confidence ~ max(auditory confidence, visual confidence) + min(auditory confidence, visual confidence) + absolute confidence*judgments
Optimal integration of the min-max: amodal confidence ~ max(auditory confidence, visual confidence) + min(auditory confidence, visual confidence) - max(auditory confidence, visual confidence)*min(auditory confidence, visual confidence)
Models comparison showed that the best model was the min-max with the absolute confidence and the type of judgments taken into account.

Comparison of the different models tested.
Closer examination of this model’s parameter estimates revealed a significant main effect of the maximal confidence (Exp.1: 







Model
Disjunctive versus conjunctive rule for perceptual detection decision
We compared the goodness of fit to amodal decisions of a model integrating information based on a disjunctive rule to one integrating information based on a conjunctive rule. The best model was the one integrating information based on a disjunctive rule (difference in AIC: 334.37).
Predictions of reaction times
Our model reproduced the observed reaction times only for correct presence judgments.

Reproduction of reaction times.
Reaction times (in seconds) as a function of the condition of presentation for Experiments 1 and 2. Error bars represent the standard error from the data. Rectangles represent data simulated from the model, centered on the mean value and with height equal to the standard error
Parameters recovery
To test the recoverability of our estimated parameters, we repeated the fitting procedure for the simulated data from parameters fitted to our Experiment 1. The correlations between the two sets of parameters were generally high.

Parameter recovery.
Results Experiment 3
The goal of this experiment was to investigate the effect of stimulus intensity on confidence bias towards absence. To examine this, we replicated Experiment 2 while increasing the stimulus intensity to target a higher detection rate of approximately 70% in unimodal trials. All other aspects of the experiment remain the same as in Experiment 2. We preregistered this experiment (https://osf.io/v5kqd).
Detection effects
We compared the amodal criterion to 0 to investigate the response bias of participants. We found that participants had a significant bias toward responding that nothing was present (M = 0.21, 95% CI [0.05,0.37], t(32) = 2.70, p =.011). Consequently, they were more accurate for absent compared to present trials (


Looking at multisensory effects, we found that participants detected better bimodal than unimodal trials (
Finally, we compared the reaction times (in seconds), as a function of the experimental condition and as a function of the judgments of participants. We found no significant difference as a function of the condition of presentation. However, participants were faster following a presence judgments than an absence judgments (

Amodal confidence effects
To investigate confidence bias, we compare the mean of the confidence judgments in function of the condition and of the accuracy of the response. We found no main effect of the condition of presentation. Participants were more confident in the correct responses (

We also compare the mean of confidence judgments as a function of the participant’s judgment and of the accuracy. We found that participants were more confident for absence compared to presence judgments (



We additionally investigated metacognitive sensitivity to evaluate how accurately participants’ confidence tracked their accuracy. We found that participants adapted better their confidence to the accuracy of their answer for absent than for present trials (

Modality-specific confidence effects
At the auditory level, participants adapted better their auditory confidence to their auditory accuracy for absent than for present trials (


At the visual level, participants adapted better their visual confidence to their visual accuracy for absent than for present trials (




Metacognitive efficiency
To further examine metacognitive performance, we analyzed metacognitive efficiency (Mratio) for presence versus absence trials, and across stimulus modalities.
Metacognitive efficiency was higher for presence judgments than absence judgments (ΔM = 0.42, 95% HDI [0.24, 0.60]). This was also the case at the modality-specific level where participants showed higher auditory metacognitive efficiency when the auditory stimulus was judged present compared to absent(ΔM = 0.34, 95% HDI [0.12, 0.55]), and higher visual metacognitive efficiency when the visual stimulus was judged present compared to absent (ΔM = 0.33, 95% HDI [0.09, 0.58]).
Moreover, when looking more closely at presence judgments, no credible difference was found between unimodal and bimodal trials (ΔM = -0.09, 95% HDI [-0.32, 0.14]). There was also no clear difference in metacognitive efficiency between auditory and visual trials (ΔM = -0.06, 95% HDI [-0.32, 0.19], between auditory and audiovisual trials (ΔM < .001, 95% HDI [-0.22, 0.23]), or between visual and audiovisual trials (ΔM = -0.06, 95% HDI [-0.29, 0.16]). Finally, in order to test the supramodality of metaperception, we compute the correlation in metacognitive efficiency between modalities. We found strong evidence for correlation between the audiovisual and auditory trials (Mcorr = 0.77, 95% HDI [0.46, 0.99]), between the audiovisual and visual trials (Mcorr = 0.79, 95% HDI [0.50, 0.99]), and between the auditory and visual trials (Mcorr = 0.77, 95% HDI [0.43, 0.99]).
We also assess auditory metacognitive efficiency as a function of visual information. When the auditory stimulus was judged present, the visual stimulus had no meaningful effect on metacognitive efficiency (ΔM = 0.16, 95% HDI [-0.17, 0.48]). The visual stimulus also had no impact when the auditory stimulus was judged absent (ΔM = -0.06, 95% HDI [-0.27, 0.15]).
Finally, we examined visual metacognitive efficiency as a function of the auditory stimulus. There was no clear effect of the auditory stimulus neither when the visual stimulus was judged absent (ΔM = -0.05, 95% HDI [-0.24, 0.15]) nor judged present (ΔM = -0.009, 95% HDI [-0.39, 0.35]).
Data availability
Data, experimental protocol, analysis code, and modelling code used are available at https://gitlab.com/nfaivre/bimodal_confidence_public.
Acknowledgements
Funded by the European Union.iews and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or the European Research Council Executive Agency. Neither the European Union nor the granting authority can be held responsible for them. This work is supported by an ERC grant (Volta, 101125379) awarded to NF. The authors thank the IDEX for funding PP mobility grant.
Additional information
Contributions:
PP: Conceptualization, Methodology, Software, Validation, Formal analysis, Investigation, Data curation, Writing - Original, Writing - Review & Editing, Visualisation, Funding acquisition. MM: Methodology, Software, Formal analysis, Writing - Review & Editing, Visualisation, Supervision. CD: Conceptualization, Methodology, Writing - Review & Editing. LG: Conceptualization, Methodology, Validation, Formal analysis, Writing - Review & Editing, Visualisation, Project administration, Supervision. NF: Conceptualization, Methodology, Validation, Formal analysis, Resources, Writing - Review & Editing, Visualisation, Project administration, Funding acquisition, Supervision
Funding
EC | European Research Council (ERC)
https://doi.org/10.3030/101125379
Nathan Faivre
Idex (UGA)
Perrine Porte
References
- Doubly Bayesian Analysis of Confidence in Perceptual Decision-MakingPLOS Computational Biology 11:e1004519https://doi.org/10.1371/journal.pcbi.1004519Google Scholar
- Individual consistency in the accuracy and distribution of confidence judgmentsCognition 146:377–386https://doi.org/10.1016/j.cognition.2015.10.006PubMedGoogle Scholar
- Measuring metacognition of direct and indirect parameters of voluntary movementJournal of Experimental Psychology: General 150:2208–2229https://doi.org/10.1037/xge0000892PubMedGoogle Scholar
- Multisensory processing of redundant information in go/no-go and choice responsesAttention, Perception, & Psychophysics 76:1212–1233https://doi.org/10.3758/s13414-014-0644-0PubMedGoogle Scholar
- Optimal metacognitive control of memory recallPsychological Review https://doi.org/10.1037/rev0000441PubMed
- Evidence for metacognitive bias in perception of voluntary actionCognition 194:104041https://doi.org/10.1016/j.cognition.2019.104041PubMedGoogle Scholar
- Metacognition in Multisensory PerceptionTrends in Cognitive Sciences 20:736–747https://doi.org/10.1016/j.tics.2016.08.006PubMedGoogle Scholar
- Confidence ratings do not distinguish imagination from realityJournal of Vision 24:13https://doi.org/10.1167/jov.24.5.13PubMedGoogle Scholar
- Distinct audio and visual accumulators co-activate motor preparation for multisensory detectionNature Human Behaviour :1–15https://doi.org/10.1038/s41562-025-02280-9PubMedGoogle Scholar
- Behavioral, Modeling, and Electrophysiological Evidence for Supramodality in Human MetacoqnitionJournal of Neuroscience 38:263–277https://doi.org/10.1523/JNEUROSCI.0322-17.2017PubMedGoogle Scholar
- Metacognition and cognitive monitoring: A new area of cognitive-developmental inquiryAmerican Psychologist 34:906–911https://doi.org/10.1037/0003-066X.34.10.906Google Scholar
- HMeta-d: Hierarchical Bayesian estimation of metacognitive efficiency from confidence ratingsNeuroscience of Consciousness 2017:nix007https://doi.org/10.1093/nc/nix007PubMedGoogle Scholar
- Metacognition: Computation, biology and functionPhilosophical Transactions of the Royal Society B: Biological Sciences 367:1280–1286https://doi.org/10.1098/rstb.2012.0021PubMedGoogle Scholar
- Howto measure metacognitionFrontiers in Human Neuroscience 8:443https://doi.org/10.3389/fnhum.2014.00443PubMedGoogle Scholar
- The redundant target effect is affected by modality switch costsPsychonomic Bulletin & Review 11:307–313https://doi.org/10.3758/BF03196575PubMedGoogle Scholar
- Multisensory processing in the redundant-target effect: A behavioral and event-related potential studyPerception & Psychophysics 67:713–726https://doi.org/10.3758/BF03193527PubMedGoogle Scholar
- An investigation of how relative precision of target encoding influences metacognitive performanceAttention, Perception, & Psychophysics 83:512–524https://doi.org/10.3758/s13414-020-02190-0PubMedGoogle Scholar
- A signal detection theoretic approach for estimating metacognitive sensitivity from confidence ratingsConsciousness and Cognition 21:422–430https://doi.org/10.1016/j.concog.2011.09.021PubMedGoogle Scholar
- Inference About Absence as a Window Into the Mental Self-ModelOpen Mind 9:635–651https://doi.org/10.1162/opmi_a_00206Google Scholar
- Distinguishing absence of awareness from awareness of absencePhilosophy and the Mind Sciences 1https://doi.org/10.33735/phimisci.2020.II.69Google Scholar
- Distinct neural contributions to metacognition for detecting, but not discriminating visual stimulieLife 9:e53900https://doi.org/10.7554/eLife.53900PubMedGoogle Scholar
- Paradoxical evidence weighting in confidence judgments for detection and discriminationAttention, Perception, & Psychophysics 85:2356–2385https://doi.org/10.3758/s13414-023-02710-8PubMedGoogle Scholar
- Beliefs about perception shape perceptual inference : An ideal observer model of detectionPsychological Review https://doi.org/10.1037/rev0000552PubMed
- The subjective experience of object recognition : Comparing metacognition for object detection and object categorizationAttention, Perception, & Psychophysics 76:1057–1068https://doi.org/10.3758/s13414-014-0643-1PubMedGoogle Scholar
- Evidence accumulation relates to perceptual consciousness and monitoringNature Communications 12https://doi.org/10.1038/s41467-021-23540-yPubMedGoogle Scholar
- Multisensory stimuli shift perceptual priors to facilitate rapid behaviorScientific Reports 11https://doi.org/10.1038/s41598-021-02566-8PubMedGoogle Scholar
- Confidence and certainty : Distinct probabilistic quantities for different goalsNature neuroscience 19:366–374https://doi.org/10.1038/nn.4240PubMedGoogle Scholar
- On quantifying multisensory interaction effects in reaction time and detection ratePsychological Research 75:77–94https://doi.org/10.1007/s00426-010-0289-0PubMedGoogle Scholar
- Human Metacognition Across Domains: Insights from Individual Differences and NeuroimagingPersonality Neuroscience 1:e17https://doi.org/10.1017/pen.2018.16PubMedGoogle Scholar
- Confidence in absence as confidence in counterfactual visibilityPsyArXiv https://doi.org/10.31234/osf.io/vuytz_v1Google Scholar
- Cortical evidence accumulation for visual perception occurs irrespective of reportsNature Communications 16:8458https://doi.org/10.1038/s41467-025-63255-yPubMedGoogle Scholar
Article and author information
Author information
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
Cite all versions
You can cite all versions using the DOI https://doi.org/10.7554/eLife.110765. This DOI represents all versions, and will always resolve to the latest one.
Copyright
© 2026, Porte et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 0
- downloads
- 0
- citations
- 0
Views, downloads and citations are aggregated across all versions of this paper published by eLife.