Abstract
The sensorimotor system can recalibrate itself without our conscious awareness, a type of procedural learning whose computational mechanism remains undefined. Recent findings on implicit motor adaptation, such as over-learning from minor perturbations and swift saturation for increasing perturbation size, challenge existing theories based on sensory errors. We argue that perceptual error, arising from the optimal combination of movement-related cues, is the primary driver of implicit adaptation. Central to our theory is the linear relationship between the sensory uncertainty of visual cues and perturbation, validated through perceptual psychophysics (Experiment 1). Our theory predicts diverse features of implicit adaptation across a spectrum of perturbation conditions on trial-by-trial basis (Experiment 2) and explains proprioception changes and their relation to visual perturbation (Experiment 3). By altering visual uncertainty in perturbation, we induced unique adaptation responses (Experiment 4). Overall, our perceptual error framework outperforms existing models, suggesting that Bayesian cue integration underpins the sensorimotor system’s implicit adaptation.
Introduction
To achieve and sustain effective motor performance, humans consistently recalibrate their sensorimotor systems to adapt to both internal and external environmental disturbances (Berniker & Kording, 2008; Shadmehr et al., 2010; Wolpert et al., 2011). For instance, transitioning to a high-sensitivity gaming mouse, which drives cursor movement at an accelerated rate compared to a standard computer mouse, may initially result in decreased performance in computer-related tasks. However, humans are capable of rapidly adapting to this new visuomotor mapping within a short period of time. While conscious corrections can facilitate this adaptation process, our sensorimotor system often times adapts itself implicitly without our conscious efforts (Albert et al., 2021; Krakauer et al., 2019).
While recent research has intensively examined the interplay between explicit and implicit learning systems (Albert et al., 2022; Miyamoto et al., 2020), several characteristics of implicit motor adaptation have emerged that challenge traditional theories. Conventionally, motor adaptation is conceptualized as error-based learning, in which learning accrues in proportion to the motor error experienced (Cheng & Sabes, 2006; Donchin et al., 2003; Thoroughman & Shadmehr, 2000). However, implicit adaptation exhibits an overcompensation phenomenon where the extent of adaptation surpasses the error induced by visual perturbations (Kim et al., 2018; Morehead et al., 2017). Additionally, implicit adaptation manifests a saturation effect; it increases with perturbations but plateaus across a broad range of larger perturbations (Bond & Taylor, 2015; Kim et al., 2018; Morehead et al., 2017; Neville & Cressman, 2018). These observations of overcompensation and saturation are incongruent with prevailing state-space updating models, which presuppose that incremental learning constitutes only a fraction of the motor error (McDougle et al., 2015; Smith et al., 2006). Another aspect of implicit adaptation that remains mechanistically unexplained pertains to its impact on proprioception. In traditional motor adaptation, proprioception is biased towards the visual perturbation, maintaining a stable bias throughout the adaptation process (Ruttle et al., 2016, 2021). In contrast, implicit adaptation initially biases proprioceptive localization of the hand towards the visual perturbation, but this bias gradually drifts in the opposite direction over time (Tsay et al., 2020).
Causal inference of motor errors has been suggested to explain the discounting of large perturbations (Wei & Körding, 2009). However, the causal inference account predicts a decline in adaptation with increasing perturbation, diverging from the observed ramp-like saturation effect. (Tsay, Kim, et al., 2022) recently synthesized existing evidence to propose that implicit adaptation reaches an upper bound set by cerebellar error correction mechanisms, reflected in a ramp-like influence of vision on proprioception (Tsay, Kim, et al., 2022). While this ramp function could explain the observed saturation, the postulate of an upper bound on visual influence lacks empirical validation. Some research supports the idea of saturation in proprioceptive recalibration (Modchalingam et al., 2019), yet other studies suggest a linear increase with visual perturbations (Rossi et al., 2021; Salomonczyk et al., 2011). Additionally, current models fall short of quantitatively capturing the time-dependent shifts in proprioceptive bias associated with implicit adaptation.
In this study, we put forth a unified model that aims to account for the distinct features of implicit adaptation, based on the Bayesian combination of movement-related cues. Prior models have overlooked the fact that visual uncertainty related to the perturbation increases with the size of the perturbation as the cursor moves further from the point of fixation and into the visual periphery (Klein & Levi, 1987; Levi et al., 1987). This is particularly pertinent for implicit adaptation that is widely investigated by the so-called error-clamp paradigm, in which participants are instructed to fixate on the target and disregard the perturbing cursor. Moreover, conventional theories of motor adaptation attribute motor error to the sensory modality of the perturbation, i.e., visual errors for visual perturbations (Tsay, Kim, et al., 2022; Wei & Körding, 2009). We propose an alternative: perceptual error drives implicit adaptation, as the perturbed sensory feedback influences the perception of the effector and, subsequently, motor adaptation. Through a series of experiments, we aim to demonstrate that combining eccentricity-induced visual uncertainty (Experiment 1) with a traditional motor adaptation model (state-space model) and a classical perception model (Bayesian cue combination) can explain both over-compensation and saturation effects (Experiment 2), as well as the time-dependent changes in proprioceptive bias (Tsay et al., 2020). Finally, to provide causal evidence supporting our Perceptual Error Adaptation (PEA) model, we manipulated visual uncertainty and observed that subsequent adaptation was attenuated for large perturbations but not for small ones—a finding that contradicts existing models but aligns well with the PEA model. Across the board, our model outperforms those based on ramp error-correction (Tsay, Kim, et al., 2022) and causal inference of errors (Wei & Körding, 2009), offering a more parsimonious explanation for the salient features of implicit adaptation.
Results
The perceptual error adaptation model with varying visual uncertainty
We start by acknowledging that the perceptual estimation of effector position is dynamically updated and influenced by sensory perturbations during motor adaptation. For implicit adaptation studied via the error-clamp paradigm, participants are required to bring their hand to the target while ignoring the direction-clamped cursor (Morehead et al., 2017). Accordingly, the perceptual estimation of the hand movement direction relies on three noisy sensory cues: the visual cue from the cursor, the proprioceptive cue from the hand, and the sensory prediction of the reaching action (Figure 1A). Without loss of generality, we posit that each cue is governed by an independent Gaussian distribution: the visual cue xv follows , where μ is the cursor direction and is visual variance, the proprioceptive cue xp follows , where xℎand is the hand movement direction and is proprioceptive variance, and the sensory prediction cue xu follows , where T is the target direction and is prediction variance. Participants aim for the target, expecting their hand to reach it. Using the Bayesian cue combination framework (Berniker & Kording, 2011), the perceived hand location on trial n can be derived:
This estimated hand position is derived using maximum likelihood estimation from the three noisy cues. Given that the clamped cursor deviates the target by μ, the visual cue xv biases the hand estimate x̂Hand towards the cursor’s direction. This deviation from the target direction T constitutes the perceptual error, which drives adaptation on the subsequent trial n+1 (Eq. 2). Consisting with existing models (Albert et al., 2022; Cheng & Sabes, 2006; Herzfeld et al., 2014; McDougle et al., 2015), trial-to-trial adaptation is modeled using a state-space equation:
where A is the retention rate capturing inter-movement forgetting and B is the learning rate capturing the proportion of error corrected within a trial. The interplay between forgetting and learning dictates the overall learning extent, i.e., the asymptote of xp:
Thus, the positive influence of perturbation size μ on the adaptation extent is counterbalanced by the rise in visual uncertainty σv, since sensory uncertainty of various visual stimuli increases linearly with eccentricity (Klein & Levi, 1987; Levi et al., 1987). As participants are instructed to fixate on the target, an increase in μ lead to increased eccentricity. Hence, we model this linear increase in visual uncertainty by
where a and b are free parameters. We conducted simulations of implicit adaptation with varying error clamp size (μ). The model simulation closely resembles the saturated adaptation in three independent experiments (Kim et al., 2018; Morehead et al., 2017). In fact, our PEA model predicts a concave adaptation pattern, contrasting with the ramp pattern suggested by the PReMo model (Tsay, Kim, et al., 2022). In Experiment 1, we aim to validate the assumption of a linear increase in visual uncertainty (Eq. 1); in Experiment 2, we seek to verify whether implicit adaptation adheres to a concave pattern as prescribed by the PEA model. Subsequent experiments, namely Experiments 3 and 4, will test the model’s additional novel predictions concerning changes in proprioception and the impact of experimentally manipulated visual uncertainty on adaptation, respectively.
Experiment 1: Visual Uncertainty Increases Linearly with Perturbation Size
To quantify visual uncertainty in a standard error-clamp adaptation setting, we employed psychometric methods. Occluded from seeing their actual hand, participants (n=18) made repetitive reaches to a target presented 10 cm straight head while an error-clamped cursor moving concurrently with one of three perturbation sizes (i.e., 4°, 16° and 64°), randomized trial-by-trial. In alignment with the error-clamp paradigm, participants were instructed to fixate on the target and to ignore the rotated cursor feedback. Eye-tracking confirmed compliance with these instructions (Figure S1). Perturbation directions were counter-balanced across trials, with equal probability of clockwise (CW) and counterclockwise (CCW) rotation. Post-movement, participants were required to judge the cursor’s rotation direction (CW or CCW) relative to a briefly displayed reference point (Figure 2A & Figure 6A). Employing this two-alternative forced-choice (2AFC) task and the Parameter Estimation by Sequential Testing (PEST) procedure (Lieberman & Pentland, 1982), we derived psychometric functions for visual discrimination (Figures 6 and Figure S2). Our findings reveal a significant increase in visual uncertainty (σv) with perturbation size, for both CW and CCW rotations (Friedman test, CW direction: χ2(2) = 34.11, p = 4e-8; CCW: χ2(2) = 26.47, p = 2e-6). Given the symmetry for the two directions, we collapsed data from both directions, and confirmed the linear relationship between σv and μ by a generalized linear model: σv = a + bμ, with a = 1.853 and b = 0.309, R2 = 0.255 (F = 51.6, p = 2.53e-9). The 95% confidence intervals (CI) for a and b are [0.440, 3.266] and [0.182, 0.435], respectively. The intercept was similar to the visual uncertainty estimated in a previous study (Tsay, Avraham, et al., 2021). The linear dependency indicates a striking seven-fold increase in visual uncertainty from a 4° perturbation to a 64° perturbation (22.641 ± 6.024° vs. 3.172 ± 0.453°).
Experiment 2: Visual Uncertainty Modulated Perceptual Error Accounts for Overcompensation and Saturation in Implicit Adaptation
The critical test of the PEA model lies in its ability to employ the linear function of visual uncertainty obtained from Experiment 1 to precisely explain key features of implicit adaptation. Earlier research mostly scrutinized smaller perturbation angles when reporting saturation effects (Bond & Taylor, 2015; Kim et al., 2018; Morehead et al., 2017). In contrast, Experiment 2 involved seven participant groups (n = 84) to characterize implicit adaptation across an extensive range of perturbation sizes (i.e., 2°, 4°, 8°, 16°, 32°, 64°, and 95°). After 30 baseline training cycles without perturbations, each group underwent 80 cycles of error-clamped reaching and 10 washout cycles without visual feedback (Figure 3A). We replicated key features of implicit adaptation: it incrementally reached a plateau, and then declined during washout. Small perturbations led to overcompensation beyond visual errors (for 2°, 4°, 8°, 16° clamp sizes). Across perturbation sizes, the faster the early adaptation the larger the adaptation extent (Figure S4). Critically, the adaptation extent displayed a concave pattern: increasing steeply for smaller perturbations and tapering off for larger ones (Figure 3B). A one-way ANOVA revealed a significant group difference in adaptation extent (F(6,83) = 12.108, p = 1.543e-09). Planned contrasts indicated that 8°, 16°, and 32° perturbations did not differ from each other (all p>0.417, with Tukey-Kramer correction), consistent with earlier evidence of invariant implicit adaptation (Kim et al., 2018). However, 64° and 95° perturbations led to significantly reduced adaptation extents compared to 8° (p = 3.194e-05 and 5.509e-06, respectively), supporting the concave pattern as a more accurate portrayal of implicit adaptation across varying perturbation size.
Importantly, the PEA model, when augmented with visual uncertainty data from Experiment 1, precisely predicts this size-dependent adaptation behavior (Figure 3B). Beyond adaptation extent, the model also accurately predicts the trial-by-trial adaptation across all seven participant groups, employing a single parameter set (R2 = 0.975; Figure 3A). The model had only four free parameters (A = 0.974, B = 0.208, σp= 11.119°, σu = 5.048°; Table S1). Remarkably, both the retention rate A and learning rate B are consistent with previous studies focusing on visuomotor rotation adaptation (Albert et al., 2022). We also quantified proprioceptive uncertainty (σp) in a subset of participants (n=13) using a similar 2AFC procedure as in Experiment 1. We found that σp was 9.737°±5.598° (Figure S6), which did not statistically differ from the σp value obtained from the model fitting (two-tailed t-test, p = 0.391). In summary, the perceptual parameters obtained in Experiment 1, when incorporated into the PEA model, effectively explain the implicit adaptation behaviors observed in different participant groups in Experiment 2.
In comparative analysis, the PReMo model yields a substantially lower R2 value of 0.749 (Figure S3B). It tends to underestimate adaptation for medium-size perturbations and overestimate it for large ones (Figure 3C; see also Figure S3B for trial-by-trial fitting). Another alternative is the causal inference model, previously shown to account for nonlinearity in motor learning (Mikulasch et al., 2022; Wei & Körding, 2009). Although this model has been suggested for implicit adaptation (Tsay, Avraham, et al., 2021), it fails to reproduce the observed concave adaptation pattern (Figures S3C and 3D). The model aligns well with adaptations to medium-size perturbations (8°, 16°, and 32°) but falls short for small and large ones, yielding an R2 value of 0.711 (see Figure S3C for trial-by-trial fits). Model comparison metrics strongly favor the PEA model over both the PReMo and causal inference models, as evidenced by AIC scores of 2255, 3543, and 3283 for the PEA, PReMo, and causal inference models, respectively (Table S2). In summary, it is the eccentricity-induced visual uncertainty that most accurately accounts for the implicit adaptation profile across a broad spectrum of perturbation sizes, rather than saturated visual influence or causal inference of error.
Experiment 3: Cue Combination Accounts for Changes in Proprioception During Implicit Adaptation
Motor adaptation not only recalibrates the motor system but also alters proprioception (Rossi et al., 2021) and even vision (Simani et al., 2007). In traditional motor adaptation involving both explicit and implicit components, the perceived hand location is initially biased towards the visual perturbation and subsequently stabilizes (Ruttle et al., 2016). However, in implicit adaptation, the perceived hand location initially aligns with but later drifts away from the visual feedback (Tsay et al., 2020). The PReMo model proposes that this drift comprises two phases: initial proprioceptive recalibration and subsequent visual recalibration (Tsay, Kim, et al., 2022), however, this assumption is lack of empirical validation. In contrast, we suggest that the perceived hand location is based on the same Bayesian cue combination principle. In this framework, the perceived hand location at the end of each reach is influenced by both the proprioceptive cue (xp) and the estimated hand position under the influence of clamped feedback (x̂Hand, Eq. 1).
During early adaptation, x̂Hand is biased towards the clamped feedback, while xp remains near the target as the motor system has yet to adapt (Figure 4A). This results in an initial negative proprioceptive bias. As adaptation progresses, although x̂Hand remains biased, xp gradually shifts in the positive direction due to adaptation, resulting in an increasingly positive proprioceptive bias. Remarkably, the PEA model can predict these temporal changes in proprioception with high accuracy (R2 = 0.982; Figure 4A).
If the hand estimate x̂Handindeed influences proprioceptive recalibration during adaptation, our PEA model can make specific quantitative predictions about the relationship between proprioception changes and visual perturbation size. While traditional visuomotor paradigms suggest either invariant (Modchalingam et al., 2019) or linear increases in proprioceptive recalibration with visual-proprioceptive discrepancy (Salomonczyk et al., 2011), the PEA model prescribes a concave function in relation to visual perturbation size (Figure 4B).
To empirically test this prediction, Experiment 3 (n=11) measured participants’ proprioceptive recalibration during implicit adaptation, using a procedure similar to the error-clamp perturbations in Experiment 2. After each block of six adaptation trials, participants’ right hands were passively moved by a robotic manipulandum, and they indicated the perceived direction of their right hand using a visually represented “dial” controlled by their left hand (Figure 7B). This method quantifies proprioceptive recalibration during adaptation (Cressman & Henriques, 2009). Each adaptation block was followed by three such proprioception test trials. The alternating design between adaptation and proprioception test blocks allowed us to assess proprioceptive biases across varying perturbation sizes, which consisted of ±10°, ±20°, ±40°, and ±80°, to covering a wide range (Figure 4D).
Our findings confirmed a typical proprioceptive recalibration effect, as the perceived hand direction was biased towards the visual perturbation (Figure 4E). Importantly, the bias in the initial proprioception test trial exhibited a concave function of perturbation size. A one-way repeated-measures ANOVA revealed a significant effect of perturbation size (F(3,30)=3.603,p=0.036), with the 20° and 40° conditions displaying significantly greater proprioceptive bias compared to the 80° condition (pairwise comparisons: 20° v.s. 80°, p = 0.034; 40° v.s. 80°, p = 0.003). The bias was significantly negative for 20° and 40° conditions (p = 0.005 and p = 0.007, respectively with one-tailed t-test), but not for 10° and 80° condition (p = 0.083 and p = 0.742, respectively). This concave pattern aligns well with the PEA model’s predictions (Figure 4B), further consolidating its explanatory power.
This stands in contrast to the PReMo model, which assumes a saturation for the influence of the visual cue on the hand estimate (Eq. 12-13). As a result, PReMo’s predicted proprioceptive bias follows a ramp function, deviating substantially from our empirical findings (Figure 4C). The causal inference model, which mainly focuses on the role of visual feedback in error correction, lacks the capability to directly predict changes in proprioceptive recalibration.
Interestingly, we observed that the proprioceptive bias reduced to insignificance by the third trial in each proprioception test block (one-tailed t-test, all p > 0.18; Figure 4E, yellow line). This suggests that the influence from implicit adaptation – manifested here as trial-by-trial updates of the perceived hand estimate x̂Hand – decays rapidly over time.
Experiment 4: Differential Impact of Upregulated Visual Uncertainty on Implicit Adaptation Across Perturbation Sizes
Thus far, we have presented both empirical and computational evidence underscoring the pivotal role of perceptual error and visual uncertainty in implicit adaptation. It is crucial to note, however, that this evidence is arguably correlational, arising from natural variations in visual uncertainty as a function of perturbation size. To transition from correlation to causation, Experiment 4 (n = 19) sought to directly manipulate visual uncertainty by blurring the cursor, thereby offering causal support for the role of multimodal perceptual error in implicit adaptation.
By increasing visual uncertainty via cursor blurring, we hypothesized a corresponding decrease in adaptation across all perturbation sizes. Notably, the PEA model predicts a size-dependent attenuation in adaptation: the reduction is less marked for smaller perturbations and more pronounced for larger ones (Figure 5A). This prediction diverges significantly from those of competing models. The PReMo model, operating under the assumption of a saturation effect for large visual perturbations, predicts that cursor blurring will only influence adaptation to smaller perturbations, leaving adaptation to larger perturbations unaffected (Figure 5B). The causal inference model makes an even more nuanced prediction: it anticipates that the blurring will lead to a substantial reduction in adaptation for small perturbations, a diminishing effect for medium perturbations, and a potential reversal for large perturbations (Figure 5C). This prediction results from the model’s core concept that causal attribution of the cursor to self-action—which directly dictates the magnitude of adaptation—decreases for small perturbations but increases for large ones when overall visual uncertainty is elevated.
Starting from the above predictions, Experiment 4 was designed to assess the impact of elevated visual uncertainty across small (4°), medium (16°), and large (64°) perturbation sizes. Visual uncertainty was augmented by superimposing a Gaussian blurring mask on the cursor (Burge et al., 2008). Each participant performed reaching tasks with either a standard or blurred clamped cursor for a single trial, bracketed by two null trials devoid of cursor feedback (Figure 5D). These three-trial mini-blocks permitted the quantification of one-trial learning as the directional difference of movements between the two null trials. To preclude the cumulative effect of adaptation, perturbation sizes and directions were randomized across mini-blocks.
Crucially, our findings corroborated the predictions of the PEA model: visual uncertainty significantly diminished adaptation for medium and large perturbations (16° and 64°), while leaving adaptation for small perturbations (4°) largely unaffected (Figure 5E). A two-way repeated-measures ANOVA, with two levels of uncertainty and three levels of perturbation size, revealed a significant main effect of increased visual uncertainty in reducing implicit adaptation (F(1,18) = 42.255, p = 4.112e-06). Furthermore, this effect interacted with perturbation size (F(2,36) = 5.391, p = 0.012). Post-hoc analyses demonstrated that elevated visual uncertainty significantly attenuated adaptation for large perturbations (p = 2.877e-04, d = 0.804 for 16°; p = 1.810e-05, d = 1.442 for 64°) but exerted no such effect on small perturbations (p = 0.108, d = 0.500). These empirical outcomes are not congruent with the predictions of either the PReMo or the causal inference models (Figure 5B and 5C). This lends compelling empirical support to the primacy of perceptual error in driving implicit adaptation, as posited by our PEA model.
Discussion
In this study, we elucidate the central role of perceptual error, derived from multimodal sensorimotor cue integration, in governing implicit motor adaptation. Utilizing the classical error-clamp paradigm, we uncover that the overcompensation observed in response to small perturbations arises from a sustained perceptual error related to hand localization, and the saturation effect commonly reported in implicit adaptation is not an intrinsic characteristic but is attributable to increasing sensory uncertainty with increasing visual perturbation eccentricity—a factor hitherto neglected in existing models of sensorimotor adaptation. Contrary to conventional theories that describe implicit adaptation as either saturated or invariant (Kim et al., 2018; Tsay, Kim, et al., 2022), our data reveal a concave dependency of implicit adaptation on visual perturbation size, characterized by diminishing adaptation in response to larger perturbations. Notably, our Perceptual Error Adaptation (PEA) model, calibrated using perceptual parameters from one set of participants, provides a robust account of implicit adaptation in separate groups subjected to varying perturbations. The model further successfully captures the perceptual consequences of implicit adaptation, such as the continuous shifts in proprioceptive localization during the adaptation process (Tsay et al., 2020) and its correlation with perturbation size. Lastly, we manipulated visual uncertainty independently of perturbation size and demonstrated that this selectively attenuated adaptation in the context of larger perturbations while leaving smaller perturbations unaffected. These empirical results, inconsistent with predictions from existing models, underscore the conceptual and quantitative superiority of our PEA model. In summary, our findings advocate for a revised understanding of implicit motor adaptation, suggesting that it is governed by Bayesian cue combination-based perceptual estimation of effector localization.
Bayesian cue combination has been established as a foundational principle in various perceptual phenomena, both intra- and inter-modally (Seilheimer et al., 2014). It has also been implicated in motor adaptation (Burge et al., 2008; He et al., 2016; Körding & Wolpert, 2004; Wei & Körding, 2010). However, previous studies have largely focused on experimentally manipulating sensory cue uncertainty to observe its effects on adaptation (Burge et al., 2008; Wei & Körding, 2010), similar to our Experiment 4. What has been largely overlooked is the natural covariance between visual uncertainty and perturbation size, which, when incorporated into classical state-space models, provides a compelling explanation for implicit adaptation.
The causal inference framework (Wei & Körding, 2009) fails to adequately predict sensorimotor changes in implicit adaptation. For instance, it underestimates the adaptation extent for large perturbations and incorrectly predicts that increasing visual uncertainty would augment, rather than reduce, adaptation to large perturbations. We postulate that causal inference is more relevant to motor learning dominated by explicit processes, such as traditional visuomotor rotations, rather than in implicit adaptations where cue combination is obligatory.
Similar to our PEA model, the PReMo model also incorporates the integration of multiple sensory cues. But two models differ fundamentally in their conceptualization of how these cues contribute to the error signal. The PReMo model posits two intermediate perceptual variables with Bayesian cue integration: a visual estimate of the cursor and a proprioceptive estimate of the hand (Tsay, Kim, et al., 2022). The final error signal in PReMo is presumed to be a proprioceptive error, not from further Bayesian cue combination, but from a visual-to-proprioceptive bias that is governed by a predetermined, ramp-like visual influence that saturates around a 6–7° visual-proprioceptive discrepancy (Eq. 13). These assumptions lack empirical validation. Our findings in Experiment 3 indicate that proprioceptive recalibration follows a concave function with respect to visual perturbation size, contradicting the ramp-like function assumed by PReMo. Moreover, the presupposed ramp-like visual influence generates a rigid prediction for a ramp-like adaptation extent profile, which is at odds with the concave adaptation pattern we observed in Experiment 2 and in a similar study involving trial-by-trial learning (Tsay, Avraham, et al., 2021). Furthermore, PReMo predicts that increasing visual uncertainty will selectively reduce adaptation to small perturbations while sparing large ones. This is inconsistent with our findings in Experiment 4, which demonstrated that increased visual uncertainty substantially impacted adaptation more to larger perturbations than to small ones. Lastly, PReMo’s reliance on a proprioceptive bias constrains its ability to account for the temporal shifts in perceived hand location during adaptation (Tsay et al., 2020). In contrast to PEA’s unified approach, PReMo must resort to separate mechanisms of proprioceptive and visual recalibration at different phases of adaptation to explain these shifts. In summary, the PReMo model’s assumptions introduce limitations that make it less consistent with empirical observations, particularly concerning the nonlinearities observed in both motoric and perceptual aspects of implicit adaptation.
Our research contributes to an ongoing debate concerning the driving forces behind error-based motor learning, specifically addressing the question of whether implicit adaptation is driven by target error or sensory prediction error (Albert et al., 2022; Izawa & Shadmehr, 2011; Leow et al., 2020; Mazzoni & Krakauer, 2006; McDougle et al., 2015; Miyamoto et al., 2020; Taylor & Ivry, 2011; Tseng et al., 2007). Most empirical data fueling this debate stem from traditional motor adaptation paradigms where explicit and implicit learning co-occur and interact. In these paradigms—visuomotor rotation being a prime example— target error is defined as the disparity between the target and the perturbed cursor, while sensory prediction error is the disparity between the predicted and actual cursor. Both types of error are sensory (specifically, visual) in nature, yet they differ due to the misalignment between the predicted or desired cursor direction and the target direction, which is induced by explicit learning (Taylor et al., 2014).
By employing the error-clamp paradigm, our study was able to isolate implicit learning, thereby eliminating potential confounds from explicit learning. Interestingly, in this paradigm, the target error and sensory prediction error effectively refer to the same visual discrepancy, as both the predicted and target directions are aligned. Despite this, classical state-space models, which utilize this visual error, fail to account for the nuanced features of implicit adaptation (Tsay, Kim, et al., 2022). In contrast, our PEA model reframes the perturbing cursor as a visual cue influencing the perceptual estimation of hand location, rather than as a source of visual error. The resultant bias in hand estimation from the desired target serves as the actual error signal. This leads us to posit that the error signal driving implicit sensorimotor adaptation is fundamentally perceptual, rather than sensory. From a normative standpoint, this perceptual error could be construed either as a predictive or performance error (Albert et al., 2022), but importantly, it is not tied to a specific modality (i.e., vision or proprioception). Instead, it directly pertains to the perceptual estimate that is crucial for task execution, i.e., bringing the hand to the target. The concept of perceptual error-driven learning can be extrapolated to various motor adaptation paradigms, including those involving explicit learning. For instance, in visuomotor rotation tasks, explicit learning manifests as a deviation in the aiming direction from the visual target, whereas implicit learning manifests as a further deviation the actual hand position from this aiming direction (Taylor et al., 2014). Even in the presence of explicit learning, the perturbed cursor continues to bias the perceptual estimate of the hand, thereby potentially driving implicit adaptation. In this scenario, the perceptual error is defined as the difference between the perceptual estimate of the hand and the altered aiming direction, which serves as the new “target” when explicit learning is in play. Our PEA model would predict similar saturation effects in implicit adaptation for this conventional adaptation paradigm, similar to for the error-clamp paradigm. Indeed, evidence from the conventional adaptation paradigm suggests that its implicit adaptation follows either a saturation effect (Bond & Taylor, 2015; Neville & Cressman, 2018) or a concave pattern (Tsay, Haith, et al., 2022) across a range of perturbation sizes. Furthermore, according to the PEA framework, this perceptual error is anchored on the aiming target, thereby naturally predicting that implicit and explicit adaptations should interact in a complementary manner, a notion that aligns with recent theories on their interaction (Albert et al., 2022; Miyamoto et al., 2020). Future research is warranted to further investigate the role of perceptual error in driving implicit learning across diverse motor learning paradigms.
Our study provides a new angle on explaining proprioceptive changes during motor adaptation, advocating for a Bayesian cue combination framework. Previously, the change in proprioceptive hand localization during motor adaptation has been ascribed to visual-proprioceptive discrepancy-induced recalibration (Ruttle et al., 2018; Salomonczyk et al., 2013) and/or altered sensory prediction driven by the adapted forward internal model (Mostafa et al., 2019; ‘t Hart & Henriques, 2016). To dissect these components, researchers have often compared proprioceptive localization in actively moved (Tsay et al., 2020) versus passively placed (passive localization, e.g., Experiment 3) hands during adaptation, attributing the smaller bias in passive localization to recalibration alone. The difference between the two is then considered to reflect altered sensory prediction due to motor adaptation (Mostafa et al., 2019; Rossi et al., 2021). But these conceptual divisions lack computational models for validation. For instance, researchers have proposed that proprioceptive recalibration in visuomotor adaptation is either a fixed proportion (e.g., 20%) of the visual-proprioceptive discrepancy (Henriques & Cressman, 2012; Ruttle et al., 2021) or largely invariant (Modchalingam et al., 2019). In fact, cross-sensory calibration typically follows the Bayesian principle, as shown in other task paradigms other than motor adaptation (Stetson et al., 2006; Wozny & Shams, 2011). Our Experiment 3 shows that proprioceptive recalibration exhibits a concave, instead of invariant or proportional, dependency to visual perturbation size, a finding follows the Bayesian principles of cue combination. Our results also confirm that the critical cue for passive localization is the biased perceived hand position (x̂Hand) fueled by adaptation.
The same Bayesian framework applies to active localization, though this time x̂Hand is to be combined with the proprioceptive cue from the adapted hand. In this sense, active localization indeed serves as a multifaceted reflection of both the internal model and proprioceptive recalibration (Mostafa et al., 2019; Rossi et al., 2021). Specifically, the proprioceptive cue continuously drifts by the adapted internal model, while the perceived hand position encapsulates the effects of proprioceptive recalibration. During the initial stages of perturbation, the immediate negative bias in active localization is predominantly attributable to rapid proprioceptive recalibration. This is evidenced by a sudden shift in the estimated hand position (x̂Hand; Figure 4A), occurring before the internal model has had sufficient time to adapt.
Then, why does active localization in traditional motor adaptation paradigms yield a largely stable bias (Ruttle et al., 2016, 2021)? We postulate that the rapid explicit learning leads to a quick asymptotic adaptation, while previous investigations have predominantly measured active localization after adaptation has plateaued (Henriques & Cressman, 2012; Modchalingam et al., 2019; Mostafa et al., 2019; Salomonczyk et al., 2011, 2013; Tsay, Kim, et al., 2021). Consequently, these studies may overlook the evolving effect of the adaptation. In contrast, the gradual nature of implicit adaptation provides a unique opportunity to uncover the underlying mechanisms governing changes in proprioception during the adaptation process.
Notably, our model aligns with previous findings that show a positive correlation between proprioceptive recalibration and motor adaptation based on individual differences (Ruttle et al., 2021; Salomonczyk et al., 2013; Tsay, Kim, et al., 2021). Unlike existing theories that posit proprioceptive recalibration either as a component of (Modchalingam et al., 2019; Mostafa et al., 2019; Ruttle et al., 2021) or a driver for implicit adaptation (Tsay, Kim, et al., 2022), our PEA model provides a mechanistic and empirically testable framework. It posits that the misestimation of hand position (x̂Hand) —induced by the recent perturbation—serves as the driving factor for both implicit adaptation and changes in proprioception. This misestimation is perturbation-dependent, resulting in both implicit adaptation and proprioceptive recalibration exhibiting a concave profile relative to perturbation size. Updated on a trial-by-trial basis, this misestimation exerts immediate effects, manifesting as an abrupt negative bias (Figure 4A). Additionally, its influence decays rapidly, becoming negligible within three trials (Figure 6C). These converging lines of evidence strongly suggest that perceptual misestimation of hand position is central to the process of proprioceptive recalibration during adaptation.
Our findings contribute nuanced perspectives to the modulation of implicit learning rate by factors beyond visual perturbation size. Previous studies have shown that environmental inconsistency -- defined as the inconsistency of visual errors -- reduced the rate (Herzfeld et al., 2014; Hutter & Taylor, 2018) or asymptote (Albert et al., 2021) of implicit adaptation. Baseline motor variance in unperturbed conditions has been shown to increase implicit adaptation rate, proposed as a sign of better exploratory learning (Wu et al., 2014). These studies interpret such phenomena as parametric changes in the learning rate in relation to visual errors, conceptualized as alterations to the B parameter in existing models. However, apparent change in learning rate to visual errors does not necessarily signify parametric modification, but may attribute to other factors that influence the use of visual cues (He et al., 2016), such as visual uncertainty in our case. Previous research has also pointed to alternative factors like error discounting based on causal inference of error (Wei & Körding, 2009), proprioceptive uncertainty (Ruttle et al., 2021; Tsay, Kim, et al., 2021), and state estimation uncertainty (He et al., 2016; Wei & Körding, 2010). Our work suggests a shift in perspective: the driving error signal for implicit learning should be considered as perceptual, rather than merely visual. This paradigmatic shift could serve as a cornerstone for future research aimed at understanding how learning rates adapt under varying conditions.
Our new framework opens avenues for exploring the memory characteristics of implicit learning. Traditional motor adaptation often exhibits ‘savings,’ or accelerated relearning upon re-exposure to a perturbation (Della-Maggiore & McIntosh, 2005; Huberdeau et al., 2019; Krakauer et al., 2005; Landi et al., 2011). In contrast, implicit adaptation has been found to exhibit a decreased learning rate during re-adaptation (Avraham et al., 2021), a phenomenon attributed to conditioning (Avraham et al., 2021) or associative learning mechanisms (Avraham et al., 2022). Investigating this ‘anti-saving’ effect will yield insights into the unique memory properties of implicit learning. Although our current PEA model is structured around single-epoch learning and does not directly address this question, it does raise new, testable hypotheses. For example, is the reduced adaptation rate during relearning attributable to a down-weighting of perturbed visual feedback in cue combination, or does it reflect a parametric alteration in the learning rate? Another noteworthy aspect of implicit learning is its remarkably slow decay rate. It has been observed that the number of trials required to washout the implicit adaptation exceeds the number of trials needed to establish it (Avraham et al., 2021; Tsay et al., 2020). In the context of our perceptual error framework, this raises the possibility that washout phases might be governed by state updating involving a distinct set of sensorimotor cues or an alternative updating mechanism, such as memory formation and selection (Oh & Schweighofer, 2019).
Methods Participants
We recruited 115 college students from Peking University (77 females, 38 males, 22.05 ± 2.82 years, mean ± SD). Participants were all right-handed according to the Edinburgh handedness inventory (Oldfield, 1971) and had normal or corrected-to-normal vision. Participants were naïve to the purpose of the experiment and provided written informed consent, which was approved by the Institutional Review Board of the School of Psychological and Cognitive Sciences, Peking University. Participants received monetary compensation upon completion of the experiment.
Apparatus
In Experiment 1, 2 and 4, participants were seated in front of a vertically-placed LCD screen (29.6 x 52.7 cm, Dell, Round Rock, TX, US). They performed the movement task with their right hand, holding a stylus and slide it on a horizontally placed digitizing tablet (48.8 x 30.5 cm, Intuos 4 PTK-1240, Wacom, Saitama, Japan). In Experiment 1, a keyboard was provided to the participants’ left hand to enable them to report the direction of visual stimuli in the discrimination task. A customized wooden shelter was placed above the tablet to block the peripheral vision of the right arm. In Experiment 1 and 4, participants placed their chin on a chin rest attached on the wooden shelter to stabilize their head. Their eye movement was recorded by an eye tracker (Tobii pro nano, Tobii, Danderyd Municipality, Sweden) affixed at the lower edge of the screen. The sampling rate was 160-200 Hz for the tablet and 60 Hz for the eye tracker.
Experiment 3 was conducted using the KINARM planar robotic manipulandum with a virtual-reality system (BKIN Technologies Ltd., Kingston, Canada). Participants seated in a chair and held the robot handles with their left and right hands (Figure 7). The movement task was performed with the right handle and the left handle was used to indicate the perceived direction of right hand in the proprioception test. A semi-silvered mirror was placed below the eye level to block the vision of the hands and the robotic manipulandum; it also served as a display monitor.
Experiment 1: measuring visual uncertainty in error-clamp adaptation
Eighteen among twenty participants finished the reaching with clamped error feedback and visual discrimination task in 3 consecutive days, two participants withdrew during the experiment. Participants made reaching movement by sliding the stylus from a start position at the center of the workspace to towards a target (Figure 6A). The start position, the target, and the cursor were represented by a gray dot, a blue cross and a white dot on the screen, respectively. All these elements had a diameter of 5mm. The procedure of the motor and visual discrimination task is illustrated in Figure 2A. To initiate a trial, participates moved the cursor into the start position. Following an 800ms holding period, a target appeared 10 cm away in twelve o’clock direction and participants were instructed to slide through the target rapidly while maintaining a straight hand trajectory. The trial terminated when the distance between the hand and the start position exceeded 10 cm, regardless of whether the target was hit. A warning message, “too slow”, would appear on the screen if participants failed to complete the trial within 300 ms after initiating the movement. Each practice day began with 60 standard reaching trials, during which veridical feedback about hand location was provided by the cursor. The target would change from blue to green if the cursor successfully passed through it. In subsequent visual clamp trials, the cursor moved along a predetermined direction set by the perturbation angle, while its position was updated in real-time based on the hand’s location. The cursor’s distance from the start position was equal to the distance between the hand and the start position until the end of the trial.
Following each trial, the cursor remained frozen at its final position for an additional 800 ms before disappearing. The visual discrimination task commenced 1000 ms thereafter. A yellow reference point, located 10 cm from the start position, was displayed for 150 ms near the cursor’s final position (Figure 2A & Figure 6A). Subsequently, all visual stimuli, except for the blue cross at the start position, were removed from the screen. Participants were then required to judge whether the reference point was situated in a clockwise (CW) or counterclockwise (CCW) direction relative to the cursor’s final position and to report their judgment by pressing a key on the keyboard. Participants were informed that they no longer controlled the direction of cursor movement during the task. They were instructed to fixate their gaze on either the start position or the blue cross during the motor task, while actively ignoring the white cursor. During the discrimination task, they were required to maintain their gaze on the blue cross. Eye movements were monitored in real-time using an eye tracker. Participants received a warning if their gaze was detected outside a 75-pixel-wide band-shaped region centered on the line of gaze four consecutive times during the experiment (Figure S1).
In each trial, the angular deviation between the error-clamped cursor and the reference point was determined using a PEST procedure (Lieberman & Pentland, 1982). Figure 6C- D illustrates the evolution of the deviation angle and step size for an exemplary participant experiencing a -16° perturbation. In each round, the deviation commenced at 30° (indicated by yellow points in Figure 6C-D) and was altered by one step size following each trial. The initial step size was set at 10° and was halved whenever the direction judgment changed (i.e., from “CW” to “CCW” or vice versa). For a specific perturbation angle, the initial deviation always started from the CW direction for the first round and flipped the direction at the beginning of the next round. A round terminated either when the step size fell below a predefined criterion (indicated by the red line in Figure 6D) or when the trial count exceeded 30. Six perturbation angles were randomly interleaved (Figure 6B), and the experiment concluded when four complete rounds of the PEST procedure had been completed for each perturbation angle. Consequently, the total number of trials varied among participants and across practice days. Additionally, for some perturbation angles, more than four complete rounds could be conducted in a single day.
Experiment 2: Motor adaptation with different perturbation size
Eighty-four participants were randomly allocated into seven groups, each comprising 12 individuals. Each group performed a motor adaptation task featuring clamped visual feedback at different perturbation angles: 2°, 4°, 8°, 16°, 32°, 64°, and 95°. As in Experiment 1, participants were instructed to slide rapidly and directly through the target, which was represented by a blue dot rather than a cross. In each trial, the target appeared at one of four possible locations (45°, 135°, 225° or 315° counter-clockwise from the positive x-axis). The sequence of target locations was randomized, yet constrained so that all four positions appeared in cycles of four trials. Each group commenced with a baseline session that included 15 cycles of reaching trials with veridical feedback, followed by 15 cycles without visual feedback. Subsequently, during the perturbation session, participants completed 80 cycles of training trials featuring the error-clamped cursor with one perturbation angle (i.e., clamp size), depending on their group assignment. To assess the aftereffect, a session comprising 10 cycles of movement without visual feedback was administered.
Experiment 3: Proprioception test with different perturbation sizes
Eleven participants were recruited for testing their proprioception recalibration. This experiment incorporated two types of trials: reaching trials and proprioception test trials. During the reaching trials, participants were instructed to aim for a target, which could appear at one of three possible locations (25°, 45°, or 65° counter-clockwise from the positive x-axis, as represented by light blue dots in Figure 4C, right panel). The task was similar to those in Experiments 1 and 2, with the key difference being that participants performed the task using KINARM robots (as depicted in Figure 7A). The dimensions and relative distances of the visual stimuli remained consistent with those used in Experiments 1 and 2. As in previous experiments, three kinds of visual feedback were provided during different sessions: no visual feedback, veridical feedback, and feedback featuring an error-clamped cursor.
In the proprioception test, participants were instructed to hold the robot’s right handle and wait for passive movement by the robot to one of six proprioception targets (small red dots in Figure 4C, right panel). These targets were spaced at 10° intervals, ranging from 20° to 70° counter-clockwise from the positive x-axis, and flanked the three reaching targets. The passive movement lasted for 1,000 ms and followed a straight-line path at a speed consistent with a minimum jerk velocity profile. During this movement, a ring with a 10 cm radius, centered at the start position, was displayed on the screen (depicted as a red arc in Figure 7B). The cursor was also replaced by a ring, its radius expanding as the hand moved toward the proprioception target.
After the right hand reached the proprioception target, participants were instructed to maintain their right hand’s position. Using the left handle, they were then asked to indicate the perceived location of their right hand. The position of the left handle was mapped to the rotation of a “dial,” which was constrained to the target arc.
The position of hp was displayed on the target arc as a small red rectangle (a visual “dial,” as shown in Figure 7B). Participants were instructed to indicate the location of their right hand by moving the red rectangle to the position they perceived as accurate. The final position of hp was recorded when its angular velocity remained below 1 degree/second for a duration exceeding 1000 ms. The proprioceptive bias was then calculated as the angular deviation between the actual hand position (hR) and the perceived hand position (hp).
Reaching trials and proprioception test trials were organized into blocks (Figure 4D). Each reaching block consisted of 6 trials, targeting 3 different locations with 2 repetitions each. Each reaching block was followed by a proprioception test block consisting of 3 trials. In these test trials, the robot moved the participant’s right hand toward a target position near one of the three reaching targets. These test targets were randomly chosen from six possible locations (Figure 4C, right panel). The entire experiment comprised 40 reaching blocks and 40 subsequent proprioception test blocks. The first four reaching blocks provided veridical cursor feedback, the next four offered no cursor feedback, and the remaining 32 featured one of eight possible perturbation sizes (±10°, ±20°, ±40°, and ±80°). The size of the perturbation was randomized between blocks.
Experiment 4: upregulating visual uncertainty affects implicit adaptation
Nineteen participants from Experiment 1 completed Experiment 4. The reaching task employed the same setup as in Experiment 1. However, instead of performing perceptual judgments of cursor motion direction, participants engaged in movements with one of three types of cursor feedback: veridical feedback, no feedback, and feedback with clamped perturbation. To assess the influence of visual uncertainty on implicit learning, we modified the cursor to appear blurred in half of the clamped trials. The blurring mask had a diameter of 6.8 mm, and the color intensity decreased from the cursor’s center following a two-dimensional Gaussian distribution with σx = σy = 1.4 mm. As depicted in Figure 5D, participants underwent the same procedures across three consecutive days. Each day consisted of 60 baseline trials, followed by 15 training blocks designed to assess single-trial learning. Within each training block, 12 trials featured an error-clamped cursor, each flanked by a trial without feedback. The difference between two adjacent no-feedback trials served as a measure of single-trial learning at specific perturbation sizes. Each of the 12 perturbation trials was randomly assigned one of 12 possible perturbations, comprising two cursor presentations (blurred or clear) and six clamp sizes (±4°, ±16°, ±64°).
Data analysis
Processing of kinematic data
In Experiments 1, 2, and 4, hand kinematic data were collected online at a sampling rate ranging between 160 and 200 Hz and subsequently resampled offline to 125 Hz. The movement direction of the hand was determined by the vector connecting the start position to the hand position at the point where it crossed 50% of the target distance, i.e., 5 cm from the start position.
In Experiment 3, hand positions and velocities were directly acquired from the KINARM robot at a fixed sampling rate of 1 kHz. The raw kinematic data were smoothed using a fifth-order Savitzky-Golay filter with a window length of 50 ms. Owing to the high temporal resolution and reliable velocity profiles provided by the KINARM system, the heading direction in Experiment 3 was calculated as the vector connecting the start position to the hand position at the point of peak velocity.
Psychometric curve
For the visual discrimination task, data of all three days were pooled together, the probability of responding that “the reference point was in the counter-clockwise direction of the cursor” was calculate as p for all angle differences (Figure S2). At each perturbation size, a logistic function was used to fit the probability distribution for individual participants:
where k is the slope and x0 is the origin of the logistic function. The visual uncertainty was defined as the angle differences between 25% and 75% of the logistic function:
Statistical analysis
In Experiment 1, since the visual uncertainty σvfollows a non-negative skewed distribution among participants, it violated the assumption of the ANOVA test. We thus applied Friedman’s nonparametric test to determine whether σv changes with the perturbation angle μ. Specifically, σvfor both positive and negative μ were subjected to Friedman’s test separately, with μ serving as the factor. Given the symmetry between positive and negative μ, we pool the data to quantify the linear dependency of σv on the absolute μ (Eq. 4). Because σv is expected to be always positive, we assume that it is generated from a gamma distribution rather than a normal distribution. Thus, the data was fitted by a generalized linear regression model with the absolute value of μ as independent variable and σv as dependent variable.
In Experiment 2, the adaptation extent was defined as the mean hand angles in the last 10 cycles in the perturbation phase (cycle 101-110). A one-way ANOVA with perturbation size serving as the factor to examine its influence on the adaptation extent. Pairwise post-hoc comparisons were conducted using Tukey-Kramer correction.
In Experiment 3, proprioceptive recalibration was quantified as the angular difference between the perceived and actual hand directions. A one-way repeated-measures ANOVA was conducted on the data of first trial, using perturbation size as the within-subject factor. Greenhouse-Geisser corrections were applied when the assumption of sphericity was violated (Kirk, 1968). Multiple pairwise comparisons were conducted among different perturbation sizes for the first proprioception test. To determine if the proprioceptive biases were significantly different from zero, one-tailed (left) t-tests were conducted separately for the first and third proprioception test trials at each perturbation size.
In Experiment 4, the single-trial learning data was subjected to a 2 (visual uncertainty) x 3 (perturbation size) repeated-measures ANOVA. Greenhouse-Geisser corrections were applied as above, and the simple main effect of visual uncertainty was tested for each of the three perturbation sizes.
Model fitting and simulations
Perceptual Error Adaptation (PEA) model
Model fitting for adaptation extent as a function of perturbation size
To fit the adaptation extent data from three different experiments in previous studies in (Kim et al., 2018; Morehead et al., 2017), Eq. 3 and Eq. 4 were modified for simplification. To avoid overfitting of the small dataset, we reduced the number of model parameters by assuming that x̂Hand asymptote to the target direction in the final adaptation trials that are used for computing adaptation extent, thus the retention rate A ≡ 1. Insert Eq. 4 to Eq. 3, the asymptote hand angle with different perturbation size is:
Two ratio parameters R1,ext = σp⁄a and R2,ext = b⁄a were used in data fitting. Three datasets were fitted separately.
Model fitting for trial-by-trial adaptation and proprioception changes
The trial-by-trial changes of adaptation (Figure 3A) and of proprioceptive localization (Figure 4A) was fitted with Eq. 1, Eq. 2, and Eq. 4 based on the mean performance of all participants. The PEA model only had four free parameters, Θ = [σu, σp, A, B]. The slope a and intercept b in Eq. 1 were obtained by psychometric tests from Experiment 1 (see statistical analysis). The reported hand position (xreport, blue dots in Figure 4A) was based on the proprioceptive cue xpand the estimated hand x̂Handfrom the reaching trial. With the Bayesian cue combination assumption, the reported hand position was biased by xpwith a ratio determined by the variance of xp and x̂Hand:
where and are the variance of x̂Hand and xp respectively. To verify if the slope b and intercept a obtained from Experiment 1 are consistent across experiments, they were also estimated by fitting data from Experiment 2 (Figure 3). In this case, the model fitting was performed with 6 free parameters, Θ = [σu, σp, a, b, A, B]. The fitted values of a and b are fallen into the 95% CI of estimated parameters in Experiment 1 (purple line in Figure 2C, see details in Table S1).
The dependence of proprioceptive recalibration on perturbation size (Figure 4B) were simulated by the PEA model with the parameter values estimated from Experiment 2. We assumed that the proprioceptive bias results from the influence of a biased hand estimate (x̂Hand) during adaptation and the influence is quantified as a percentage of its deviation from the true hand location:
where the actual hand location is 0, Rp is the percentage of influence, and x̂Hand is determined by Eq.1. In simulation, Rp varied from 0.05 to 0.8 to estimate the overall dependence of proprioceptive recalibration on perturbation size.
Model fitting and simulation for single-trial learning
In the single-trial learning paradigm (Figure S5), the average movement direction across trials aligns with the target direction since the visual perturbations are evenly distributed in both directions. Thus, the sensory cue xu and xp have the same mean. For modeling single-trial learning, instead of having two separate cues, we assume a combined cue of xu and xp to follow where T is the target direction, represents the variance of integrated sensory signal of xu and xp. Single-trial learning was quantified as the difference between the two null trials before and after the perturbation trial. As the perturbation size in the triplet of trials varied randomly, we assume that the effects of different perturbations are independent. Thus, single-trial learning was modeled as learning from the current perturbation without history effect. It follows the equations modified from Eq. 1 and 2:
where xvis the visual perturbation, Wintand Wvare the weights of the cues, σvis the standard deviation of the visual cue specified by Eq. 4. Parameter set Θ = [σint, a, b, B] was fitted to the average data from all participants. Model simulations (Figure 5A) were performed with the same single-trial learning equations. For the clear cursor condition, we used the same parameter values estimated from Experiment 2 (see details in Table S1). For the blurred cursor condition, the standard deviation of visual cue was changed to:
for the simulation of the increase in visual uncertainty, the ratio Rv varied from 1.1 to 3.
PReMo model
We used the PReMo model to fit the average adaptation extent obtained from Experiment 2 (Figure 3C & Figure S3B). Following the study by (Tsay, Kim, et al., 2022), the hand position at trial n+1 is:
where
In data fitting, we used two parameters to represent the ratio between sensory cues: R1 = . The data were fitted with the parameter set Θ = [R, , where is the saturation angle, ηp is a scaling factor, A is the retention rate and B is the learning rate. For simulating the proprioceptive localization of the hand (Figure 4C), the parameter values estimated from Experiment 2 were used. The bias of hand estimation in the proprioception trials is determined as: xbias = −(0 −xper)Rp, where ratio Rp varies from 0.05 to 0.8. Thus, similar to the PEA model simulation, the proprioceptive bias is a fraction of the bias in the hand estimation from the adaptation trials. Single-trial learning (Figure 5B) was simulated with:
where xperis determined by Eq. 12 and Eq. 13. For the clear condition, we used the parameter values estimated from Experiment 2 with PReMo. For the blurred cursor condition, the standard deviation of visual signal σv,blur increases with a ratio Rv, as in Eq. 12.
Causal inference model
The causal inference model by (Wei & Körding, 2009) was used to fit the data of Experiment 2 (Figure 3D & Figure S3C). The hand position at trial n+1 is updated by learning from visual error at trial n:
where A and B are the retention and learning rates, respectively; T is the target direction. Specifically for this model, the learning from error is modulated by the probability (p) of causal attribution of visual error to the action or proprioception:
where xv,n is the visual cue at trial n. S and C are the scaling factors, and σ is the standard deviation of the integrated cue combining visual and proprioceptive cues, following
Thus, the data were fitted with five parameters Θ = [σ, S, C, A, B]. For simulating single-trial learning with cursor blurring (Figure 5C), the ratio between σv and σpis fixed as ½. The single-trial leaning was determined as:
where p is determined by Eq. 18. Put Eq. 12 and Eq. 19 into , we can calculate the standard deviation of the integrated sensory signal for the blurred cursor:
Simulation was performed with R ranging from 1.1 to 3.
Data fitting
All data were fitted using MATLAB (2022b, MathWorks, Natick, MA, US) build-in function fmincon with 100 randomly sampled initial values of parameter sets. See Table S1 and Table S2 for the fitted parameter values and comparisons between different models.
Data availability
Data presented in this work are available at: https://doi.org/10.6084/m9.figshare.24503926.v1.
Supplementary Materials
References
- Competition between parallel sensorimotor learning systemsELife 11
- An implicit memory of errors limits human sensorimotor adaptationNature Human Behaviour :1–15
- Reexposure to a sensorimotor perturbation produces opposite effects on explicit and implicit learning processesPLoS Biology 19
- Contextual effects in sensorimotor adaptation adhere to associative learning rulesELife 11
- Estimating the sources of motor errors for adaptation and generalizationNature Neuroscience 11:1454–1461
- Bayesian approaches to sensory integration for motor controlWiley Interdisciplinary Reviews. Cognitive Science 2:419–428
- Flexible explicit but rigid implicit learning in a visuomotor adaptation taskJournal of Neurophysiology 113:3836–3849
- The statistical determinants of adaptation rate in human reachingJournal of Vision 8:1–19
- Modeling sensorimotor learning with linear dynamical systemsNeural Computation 18:760–793
- Sensory recalibration of hand position following visuomotor adaptationJournal of Neurophysiology 102:3505–3518
- Time course of changes in brain activity and functional connectivity associated with long-term adaptation to a rotational transformationJournal of Neurophysiology 93:2254–2262
- Quantifying generalization from trial-by-trial behavior of adaptive systems that learn with basis functions: theory and experiments in human motor controlThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 23:9032–9045
- The statistical determinants of the speed of motor learningPLoS Computational Biology 12
- Visuomotor adaptation and proprioceptive recalibrationJournal of Motor Behavior 44:435–444
- A memory of errors in sensorimotor learningScience 345:1349–1353
- Practice induces a qualitative change in the memory representation for visuomotor learningJournal of Neurophysiology 122:1050–1059
- Relative sensitivity of explicit reaiming and implicit motor adaptationJournal of Neurophysiology 120:2640–2648
- Learning from sensory and reward prediction errors during motor adaptationPLoS Computational Biology 7
- Invariant errors reveal limitations in motor correction rather than constraints on error sensitivityCommunications Biology 1
- Experimental design: Procedures for the behavioral sciences (Vol. 1–f)Brooks/Cole
- Position sense of the peripheral retinaJOSA A 4:1543–1553
- Bayesian integration in sensorimotor learningNature 427:244–247
- Adaptation to visuomotor transformations: consolidation, interference, and forgettingThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 25:473–478
- Motor learningComprehensive Physiology 9:613–663
- One week of motor adaptation induces structural changes in primary motor cortex that predict long-term memory one year laterThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 31:11808–11813
- Task errors drive memories that improve sensorimotor adaptationJournal of Neuroscience 40:3075–3088
- Positional uncertainty in peripheral and amblyopic visionVision Research 27:581–597
- Microcomputer-based estimation of psychophysical thresholds: the best PESTBehavior Research Methods & Instrumentation 14:21–25
- An implicit plan overrides an explicit strategy during visuomotor adaptationThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 26:3642–3645
- Explicit and Implicit Processes Constitute the Fast and Slow Processes of Sensorimotor LearningThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 35
- Visuomotor mismatch responses as a hallmark of explaining away in causal inferenceNeural Computation 35:27–37
- Implicit adaptation compensates for erratic explicit strategy in human motor learningNature Neuroscience 23:443–455
- The effects of awareness of the perturbation during motor adaptation on hand localizationPloS One 14
- Characteristics of implicit sensorimotor adaptation revealed by task-irrelevant clamped feedbackJournal of Cognitive Neuroscience 29:1061–1074
- Motor learning without moving: proprioceptive and predictive hand localization after passive visuoproprioceptive discrepancy trainingPloS One 14
- The influence of awareness on explicit and implicit contributions to visuomotor adaptation over timeExperimental Brain Research. Experimentelle Hirnforschung. Experimentation Cerebrale 236:2047–2059
- Minimizing Precision-Weighted Sensory Prediction Errors via Memory Formation and Switching in Motor AdaptationThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 39:9237–9250
- The assessment and analysis of handedness: the Edinburgh inventoryNeuropsychologia 9:97–113
- Mechanisms of proprioceptive realignment in human motor learningCurrent Opinion in Physiology 20:186–197
- Time course of reach adaptation and proprioceptive recalibration during visuomotor learningPloS One 11
- The fast contribution of visual-proprioceptive discrepancy to reach aftereffects and proprioceptive recalibrationPloS One 13
- Implicit motor learning within three trialsScientific Reports 11
- Proprioceptive recalibration following prolonged training and increasing distortions in visuomotor adaptationNeuropsychologia 49:3053–3062
- The role of the cross-sensory error signal in visuomotor adaptationExperimental Brain Research. Experimentelle Hirnforschung. Experimentation Cerebrale 228:313–325
- Models and processes of multisensory cue combinationCurrent Opinion in Neurobiology 25:38–46
- Error correction, sensory prediction, and adaptation in motor controlAnnual Review of Neuroscience 33:89–108
- Visual-shift adaptation is composed of separable sensory and task-dependent effectsJournal of Neurophysiology 98:2827–2841
- Interacting adaptive processes with different timescales underlie short-term motor learningPLoS Biology 4
- Motor-sensory recalibration leads to an illusory reversal of action and sensationNeuron 51:651–659
- Separating predicted and perceived sensory consequences of motor learningPloS One 11
- Flexible cognitive strategies during motor learningPLoS Computational Biology 7
- Explicit and implicit contributions to learning in a sensorimotor adaptation taskThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 34:3023–3032
- Learning of action through adaptive combination of motor primitivesNature 407
- The effect of visual uncertainty on implicit motor adaptationJournal of Neurophysiology 125:12–22
- Interactions between sensory prediction error and task error during implicit motor learningPLoS Computational Biology 18
- Individual differences in proprioception predict the extent of implicit sensorimotor adaptationJournal of Neurophysiology 125:1307–1321
- Understanding implicit sensorimotor adaptation as a process of proprioceptive re-alignmentELife 11
- Continuous reports of sensed hand position during sensorimotor adaptationJournal of Neurophysiology 124:1122–1130
- Sensory prediction errors drive cerebellum-dependent adaptation of reachingJournal of Neurophysiology 98
- Relevance of error: what drives motor adaptation?Journal of Neurophysiology 101:655–664
- Uncertainty of feedback and state estimation determines the speed of motor adaptationFrontiers in Computational Neuroscience 4
- Principles of sensorimotor learningNature Reviews. Neuroscience 12
- Recalibration of auditory space following milliseconds of cross-modal discrepancyJournal of Neuroscience 31:4607–4612
- Temporal structure of motor variability is dynamically regulated and predicts motor learning abilityNature Neuroscience 17
Article and author information
Author information
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
- Reviewed Preprint version 2:
- Version of Record published:
Copyright
© 2024, Zhang et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 792
- downloads
- 89
- citations
- 4
Views, downloads and citations are aggregated across all versions of this paper published by eLife.