Abstract
Mesolimbic dopamine activity occasionally exhibits ramping dynamics, reigniting debate on theories of dopamine signaling. This debate is ongoing partly because the experimental conditions under which dopamine ramps emerge remain poorly understood. Here, we show that during Pavlovian and instrumental conditioning, mesolimbic dopamine ramps are only observed when the inter-trial interval is short relative to the trial period. These results constrain theories of dopamine signaling and identify a critical variable determining the emergence of dopamine ramps.
Introduction
Mesolimbic dopamine activity was classically thought to operate in either a “phasic” or a “tonic” mode(1–3). Yet, recent evidence points to a “quasi-phasic” mode in which mesolimbic dopamine activity exhibits ramping dynamics(4–16). This discovery reignited debate on theories of dopamine function because it appeared inconsistent with the dominant theory that dopamine signaling conveys temporal difference reward prediction error (RPE)(17), since ramping dopamine would paradoxically be a “predictable prediction error”(18). Recent work has hypothesized that dopamine ramps reflect the value of ongoing states, serving as a motivational signal(1, 4–6). Others have argued that ramping dopamine indeed reflects RPE under some assumptions, namely correction of uncertainty via sensory feedback(11, 15, 19), representational error(20), or memory lapses(21). Still others have proposed that dopamine ramps reflect a causal influence of actions on rewards in instrumental tasks(12). This debate has been exacerbated in part because there is no clear understanding of why dopamine ramps appear only under some experimental conditions. Accordingly, uncovering a unifying principle of the conditions under which dopamine ramps appear will provide important constraints on theories of dopamine function(1, 12, 18–42).
To investigate the necessary conditions for dopamine ps, we turned to our recent work proposing that dopamine acts as a teaching signal for causal learning by representing the Adjusted Net Contingency for Causal Relations (ANCCR)(32, 42, 43). The crux of the ANCCR model is that learning to predict a future meaningful event (e.g., reward) can occur by looking back in time for potential causes of that event. Critically, the ability to learn by looking backwards depends on how long one holds on to past events in memory. If the past is quickly forgotten, there is little ability to identify causes that occurred long before a reward. On the other hand, maintaining memory for too long is computationally inefficient and would allow illusory associations across long delays. Thus, for optimal learning, the timescale for memory maintenance should flexibly depend on “environmental timescales” set by the overall rates of events. In ANCCR, this is achieved by controlling the duration of a memory trace of past events with the “eligibility trace” time constant (illustrated in Extended Data Fig 1). When there is a high average rate of events, the eligibility trace time constant is small. Accordingly, we successfully simulated dopamine ramping dynamics assuming two conditions: a dynamic progression of cues that signal temporal proximity to reward, and a small eligibility trace time constant relative to the trial period(32). However, whether these conditions are sufficient to experimentally produce mesolimbic dopamine ramps in vivo remains untested.
In this study, we designed experiments to address the influence of environmental timescales on dopamine ramps. Specifically, we sought to test the key prediction that dopamine ramps would be observed for a small, but not large, eligibility trace time constant, which is hypothesized to emerge for high overall event rate. To do so, we manipulated the intertrial interval (ITI) duration in both an auditory Pavlovian conditioning paradigm and a virtual reality navigation task. The results confirmed our key prediction from ANCCR, providing a clear constraint on theoretical explanations for the controversial phenomenon of dopamine ramps.
Results
We first measured mesolimbic dopamine release in the nucleus accumbens core using a dopamine sensor (dLight1.3b)(44) in an auditory cue-reward task. We varied both the presence or absence of a progression of cues indicating reward proximity (“dynamic” vs “fixed” tone) and the inter-trial interval (ITI) duration (short vs long ITI). Varying the ITI was critical because our theory predicts that the ITI is a variable controlling the eligibility trace time constant, such that a short ITI would produce a small time constant relative to the cue-reward interval (Supplementary Note 1, Fig 1a-e). In all four experimental conditions, head-fixed mice learned to anticipate the sucrose reward, as reflected by anticipatory licking (Fig 1f-g). In line with our earlier work, we showed that simulations of ANCCR exhibit a larger cue onset response when the ITI is long and exhibit ramps only when the ITI is short (Fig 1h). Consistent with these simulated predictions, experimentally measured mesolimbic dopamine release had a much higher cue onset response for long ITI (Fig 1i-j). Furthermore, dopamine ramps were observed only when the ITI was short and the tone was dynamic (Fig 1i, k-m, Extended Data Fig 2). Indeed, dopamine ramps—quantified by a positive slope of dopamine response vs time within trial over the last five seconds of the cue—appeared on the first day after transition from a long ITI/dynamic tone condition to a short ITI/dynamic tone condition and disappeared on the first day after transition from a short ITI/dynamic tone condition to a short ITI/fixed tone condition (Fig 1l). These results confirm the key prediction of our theory in Pavlovian conditioning.

Pavlovian conditioning dopamine ramps depend on ITI.
a. Top, fiber photometry approach schematic for nucleus accumbens core (NAcC) dLight recordings. Bottom, head-fixed mouse. b. Pavlovian conditioning experimental setup. Trials consisted of an 8 s auditory cue followed by sucrose reward delivery 1 s later. c. Cumulative Distribution Function (CDF) of ITI duration for long (solid line, mean 55 s) and short ITI (dashed line, mean 8 s) conditions. d. Experimental timeline. Mice were divided into groups receiving either a 3 kHz fixed and dynamic up↑ tone or a 12 kHz fixed and dynamic down↓ tone. e. Tone frequency over time. f. Peri-stimulus time histogram (PSTH) showing average licking behaviors for the last 3 days of each condition (n = 9 mice). g. Average anticipatory lick rate (baseline subtracted) for 1 s preceding reward delivery (two-way ANOVA: long ITI vs short ITI F(1) = 9.3, **p = 0.0045). h. ANCCR simulation results from an 8 s dynamic cue followed by reward 1 s later for long ITI (teal) and short ITI (pink) conditions. Bold lines show the average of 20 iterations. i. Left, average dLight dopamine signals. Vertical dashed lines represent the ramp window from 3 to 8 s after cue onset, thereby excluding the influence of the cue onset and offset responses. Solid black lines show linear regression fit during window. Right, closeup of dopamine signal during window. j. Average peak dLight response to cue onset for LD and SD conditions (paired t-test: t(8) = 6.3, ***p = 2.3×10−4). k. dLight dopamine signal with linear regression fit during ramp window for example SD trials. Reported m is slope. l. Session average per-trial slope during ramp window for the first day and last 3 days of each condition (one-sided [last day LD < first day SD] paired t-test: t(8) = −2.1, *p = 0.036; one-sided [last day SD > first day SF] paired t-test: t(8) = 2.4, *p = 0.023). m. Average per-trial slope for last 3 days of each condition (Tukey HSD test: q = 3.8, LD vs SD **p = 0.0027, SD vs SF *p = 0.011). All data presented as mean ± SEM. See Supplementary Table 1 for full statistical details.
While these results are consistent with the idea that dopamine ramps are shaped by the ITI, an alternative explanation could be differences in behavioral learning across experimental conditions. To test this possibility, we repeated the same Pavlovian conditioning paradigm with a counterbalanced training order in a second cohort of mice (Fig 2a). Despite the shuffled training order, this cohort behaved similarly and showed robust anticipatory licking across conditions (Extended Data Fig 3a-b). As with the previous cohort, dopamine ramps were only observed in the short ITI/dynamic tone condition, rapidly appearing on the first day of this condition and disappearing on the first day of the subsequent long ITI/dynamic tone condition (Fig 2b-d, Extended Data Fig 4, Extended Data Fig 5). Critically, the presence of dopamine ramps during the last five seconds of the cue could not be explained by variations in behavior; during this period, anticipatory licking was similar across all conditions, and there was no difference in the slope of the lick rate between the dynamic tone conditions (Extended Data Fig 3c-d). Taken together, these results rule out any effects of differential learning across conditions on dopamine ramps.

Pavlovian conditioning dopamine ramps do not depend on training order.
a. Experimental timeline in which the SD condition occurs before the LD condition. b. Left, average dLight dopamine signals for SD and LD conditions. Vertical dashed lines represent the ramp window from 3 to 8 s after cue onset. Solid black lines show linear regression fit during window. Right, closeup of dopamine signal during window (n = 9 mice). c.Session average per-trial slope during ramp window for the first day and last 3 days of each condition (last day LF to first day SD: **p = 0.0046, last day SD to first day LD: **p = 0.0067). d. Average per-trial slope for last 3 days of each condition (LF vs SD: ***p = 9.1 ×10−4, SD vs LD: *p = 0.010).
Although the difference in dopamine ramp slope seems to be well explained by the ITI condition, it might instead reflect differences in post-reward dopamine dynamics, which drop below baseline. As dopamine levels recover from this drop over several seconds, it could appear as a dopamine ramp on the subsequent trial given a sufficiently short ITI. The lack of dopamine ramps in the short ITI/fixed tone condition serves as a control for this, however (Fig 1l-m). Furthermore, there is no significant difference in the pre-cue dopamine slope between conditions nor is there a correlation between the precue dopamine slopes and the dopamine ramp slopes during the cue in the short ITI/dynamic condition (Extended Data Fig 6). As such, our results cannot be captured by a natural ramp in the dopamine signal following reward.
Given the speed with which dopamine ramps appeared and disappeared, we next tested whether the slope of dopamine ramps in the short ITI/dynamic tone condition depended on the previous ITI duration on a trial-by-trial basis. We found that there was indeed a statistically significant trial-by-trial correlation between the previous ITI duration and the current trial’s dopamine response slope in the short ITI/dynamic condition with ramps, but not in the long ITI/dynamic condition without ramps (Fig 3a-c). The dependence of a trial’s dopamine response slope with previous ITI was significantly negative, meaning that a longer ITI correlates with a weaker ramp on the next trial. This finding held when analyzing either animal-by-animal (Fig 3a-b) or the pooled trials across animals while accounting for mean animal-by-animal variability (Fig 3c). This relationship was only significant for a single previous trial, however, and did not hold for a broader estimate of average previous ITIs (Extended Data Fig 7). In addition, we quantified how the relative change in ITI duration between consecutive trials correlates with changes in dopamine ramp slope (Fig 3d). We found a significantly negative relationship between the change in dopamine slope and change in ITI (Fig 3e-f). Furthermore, the change in slope was significantly greater for relative decreases in ITI compared to relative increases in ITI, indicating that a relatively shorter ITI tends to have a stronger ramp (Fig 3g). These results suggest that the eligibility trace time constant adapts rapidly to changing ITI in Pavlovian conditioning.

Per-trial dopamine ramps correlate with previous ITI.
a. Scatter plot for an example animal showing the relationship between dopamine response slope within a trial and previous ITI for all trials in the last 3 days of SD condition. Plotted with linear regression fit (black line) used to find this animal’s β coefficient of −0.045. b. Linear regression β coefficients for previous ITI vs. trial slope calculated per animal (***p = 5.6 ×10−4). c. Scatter plot of Z-scored trial slope vs. previous ITI pooled across mice for all trials in the last 3 days of SD condition (***p = 6.6 ×10−10). The Z-scoring per animal removes the effect of variable means across animals on the slope of the pooled data. d.dLight dopamine signal for two consecutive example SD trials showing the change in ITI and change in slope. The grey shaded regions indicate ITIs, and the vertical dashed lines mark the ramp window period. Reported m is slope. e. Scatter plot for the same example animal in a showing the relationship between the change in dopamine slopes and the chance in ITI across all trials in the last 3 days of SD condition. Plotted with linear regression fit (black line). Dot colors indicate magnitude of Δ ITI: light pink for Δ ITI below −1 s; grey for Δ ITI between −1 s and 1 s; dark pink for Δ ITI above 1 s. f. Linear regression β coefficients for Δ ITI vs. Δ slope calculated per animal (***p = 3.2 ×10−4). g. Comparison of the average Δ slope for Δ ITI below −1 s vs above 1 s (*** p = 2.3 ×10−4).
Due to the robust relationship between ITI and dopamine ramp slope on a per-trial basis, we next sought to explore the potential relationships between other important dopaminergic and behavioral variables. Though the dopamine cue onset response is significantly greater in the long compared to short ITI/dynamic tone condition, there is no apparent relationship between the cue onset response and dopamine ramp slope in either condition (Extended Data Fig 8a-b). Furthermore, neither the cue onset response nor the dopamine ramp slope correlates with the per-trial behavior quantified as lick slope (Extended Data Fig 8c-f). Finally, unlike the ramping dopamine slope, this ramping lick slope did not correlate with ITI duration (Extended Data Fig 8g-h). The fact that this exploration of additional variables yielded no significant relationships highlights the unique, specific influence of ITI on dopamine ramp slope.
We next tested whether the results from Pavlovian conditioning could be reproduced in an instrumental task. In keeping with prior demonstrations of dopamine ramps in head-fixed mice, we used a virtual reality (VR) navigational task in which head-fixed mice had to run towards a destination in a virtual hallway to obtain sucrose rewards(11, 15, 19, 45) (Fig 4a-b, Extended Data Fig 9). At reward delivery, the screen turned blank during the ITI and remained so until the next trial onset. After training animals in this task using a medium ITI, we changed the ITI duration to short or long for eight days before switching to the other (Fig 4c). We found evidence that mice learned the behavioral requirement during the trial period, as they significantly increased their running speed during trial onset (Fig 4d-e) and reached a similarly high speed prior to reward in both ITI conditions (Fig 4f-g). Consistent with the results from Pavlovian conditioning, the dopamine response to the onset of the hallway presentation was larger during the long ITI compared to the short ITI condition (Fig 4h-i), and dopamine ramps were observed only in the short ITI condition (Fig 4j-m). Unlike the Pavlovian conditioning, the change in the ITI resulted in a more gradual appearance or disappearance of ramps (Fig 4l), but there was still a weak overall correlation between dopamine response slope on a trial and the previous inter-reward interval (Extended Data Fig 10). These results are consistent with a more gradual change in the eligibility trace time constant in this instrumental task. Collectively, the core finding from Pavlovian conditioning that mesolimbic dopamine ramps are present only during short ITI conditions was reproduced in the instrumental VR task.

VR navigation dopamine ramps depend on ITI.
a. Head-fixed VR approach schematic. b. VR navigation task experimental setup. Trials consisted of running down a patterned virtual hallway to receive sucrose reward. VR monitor remained black during the ITI. c. Experimental timeline. Following training, mice were assigned to either long or short ITI conditions for 8 days before switching. d. Velocity PSTH aligned to trial onset for long (teal) and short (pink) ITI conditions (n = 9 mice). e Average change in velocity at trial onset. Bottom asterisks indicate both conditions significantly differ from zero (long: ***p = 1.0 ×10−4; short: ***p = 2.3 ×10−5). Top asterisks indicate significant difference between conditions (**p = 0.0028). f. Velocity PSTH aligned to reward delivery. g. Average velocity during 1 s preceding reward (p = 0.50). h. PSTH showing average dLight dopamine signal aligned to trial onset. i. Comparison of peak dLight onset response (***p = 6.3 ×10−5). j. Left, average dLight dopamine signal across distances spanning the entire virtual corridor. Vertical dashed lines represent the ramp window from 20 to 57 cm (10 cm before end of track). Solid black lines show linear regression fit during window. Right, closeup of dopamine signal during window. k. dLight dopamine signal with linear regression fit during ramp window for example short ITI trials. Reported m is slope. l. Session average per-trial slope during ramp window for all days of each condition. m. Comparison of average per-trial slope during ramp window for last 3 days of both conditions (*p = 0.035).
Discussion
Our results provide a general framework for understanding past results on dopamine ramps. According to ANCCR, the fundamental variable controlling the presence of ramps is the eligibility trace time constant. Based on first principles, this time constant depends on the ITI in common task designs (Supplementary Note 1). Thus, the ITI is a simple proxy to manipulate the eligibility trace time constant, thereby modifying dopamine ramps. In previous navigational tasks with dopamine ramps, there was no explicitly programmed ITI(4, 10, 16). As such, the controlling of the pace of trials by these highly motivated animals likely resulted in short effective ITI compared to trial duration. An instrumental lever pressing task with dopamine ramps similarly had no explicitly programmed ITI(7), and other tasks with observed ramps had short ITIs(8, 11, 13–15). One reported result that does not fit with a simple control of ramps by ITI is that navigational tasks produce weaker ramps with repeated training(10). These results are generally inconsistent with the stable ramps that we observed in Pavlovian conditioning across eight days (Fig 1l). A speculative explanation might be that when the timescales of events vary considerably (e.g., during early experience in instrumental tasks due to variability in action timing), animals use a short eligibility trace time constant to account for the potential non-stationarity of the environment. With repeated exposure, the experienced stationarity of the environment might increase the eligibility trace time constant, thereby complicating its relationship with the ITI. Alternatively, as suggested previously(7, 10), repeated navigation may result in automated behavior that ignores the progress towards reward, thereby minimizing the calculation of associations of spatial locations with reward.
While the focus of this study was environmental timescales set by the ITI, we also assumed that a dynamic sequence of external cues signaling temporal or spatial proximity to reward would be required for dopamine ramps to occur. Indeed, our results that dopamine ramps occur in the short ITI/dynamic, but not fixed, tone condition corroborate this assumption, as well as results from other Pavlovian conditioning experiments utilizing dynamic cues in a head-fixed setup(11, 19). In contrast, experiments involving freely moving animals do not require explicitly dynamic cues because the sensory feedback from navigating though the environment presumably functions in the same way to indicate proximity to reward. Future experiments can investigate this further by characterizing the role of sensory feedback indicating reward proximity in mediating dopamine ramps. One set of observations superficially inconsistent with our assumption of the necessity of a sequence of external cues is that ramping dopamine dynamics can be observed even when only internal states signal reward proximity (e.g., timing a delayed action)(10, 13, 14). In these cases, however, animals were required to actively keep track of the passage of time, which therefore strengthens an internal progression of neural states signaling temporal proximity to reward. We speculate that once learned, these internal states could serve the role of external cues in the ANCCR framework. Previously, we argued against this kind of assumption in learning theories(46). Our earlier position was that it is problematic to assume fixed internal states that pre-exist and provide a scaffold for learning, such as in temporal difference learning. This is because these pre-existing states would need to already incorporate information that can only be acquired during the course of learning(46). Unlike this position, here we are merely speculating that after learning, an internal progression of states can serve the function of externally signaled events. Similarly, we have previously postulated that such an internal state exists during omission of a predicted reward, but only after learning of the cue-reward association(32).
Though the experiments in this study were motivated by the ANCCR framework, they were not conducted to discriminate between theories. As such, it is also possible to rationalize these results in the context of other models of dopamine ramps. In the value model, dopamine is thought to represent the discounted sum of future rewards(4–6, 18). The shape of this value function, and thus the predicted dopamine dynamics, is determined by the discount factor, γ. If there is a low γ (i.e., greater discounting), then the corresponding value function produces a steeper ramp. Consequently, it is possible to use the value model to explain our results if one assumes that a shorter ITI causes greater temporal discounting. The basis for such an assumption is unclear, though it has been suggested that the overall temporal discounting in an environment depends on reward rates(47, 48). Furthermore, we do not find evidence for dopamine ramps acting as a value signal to directly increase motivation. This is because we find similar trial-related behaviors in conditions with and without dopamine ramps. Thus, ITI dependent emergence of dopamine ramps for the same trial parameters provides strong constraints for the motivational role of dopamine ramps.
Substantial efforts have been made to account for the phenomenon of dopamine ramps as a temporal difference RPE(11, 18–20, 36). As with the value model, simulated dopamine responses in temporal difference RPE models are also modulated by the discount factor, γ. It has been proposed that temporal discounting in the dopamine system depends on the cue-reward delay(35). In our experiment, however, the cue-reward delay is not the key variable determining the presence of ramps; instead, it is the ITI. Another work has also proposed that a spectrum of discount factors can explain diverse activity profiles of single dopamine neurons(36). Specifically, monotonic upward ramps were simulated using a high γ (i.e., weaker discounting). Therefore, in this model, one would need to assume that shorter ITIs cause weaker temporal discounting to produce steeper ramps. Notably, this is in the opposite direction as the value model. Overall, it is unclear whether any fundamental principle predicts an ITI dependent change in temporal discounting in the dopamine system to allow RPE to explain our results. Similarly, whether other models of dopamine ramps(12, 18, 33) can capture an ITI dependent emergence of dopamine ramps remains to be explored.
While it is thus possible to rationalize our results using alternative theories of dopamine, ANCCR provides a principled and parsimonious explanation. Given that the foundation of ANCCR is looking back in time for causes of rewards, it is clear that differences in memory maintenance via eligibility traces will have profound implications on predicted dopamine signaling. For example, when the ITI is short and rewards are being frequently delivered, it intuitively makes sense that the eligibility trace time constant needs to be small; this is because the time window over which one would want to search for potentially causal cues is going to be shorter in this situation. We formalize this intuition by postulating that the eligibility trace time constant adapts to the overall event rates for efficient coding (Supplementary Note 1). In the case of a dynamic progression of cues, the cues closer in time to the reward will have higher causal power, and thus higher ANCCR, resulting in a dopamine ramp.
Our ANCCR simulations motivated the experiments, but we did not explicitly intend to fit the data. Accordingly, there are several details of the experiments that we did not include in the simulations. First, animals were trained initially using a long ITI (Pavlovian) or medium ITI (VR). This may explain a discrepancy between the simulations and experimental results: the cue onset response in the short ITI condition is small but positive in the experiment but negative in ANCCR. This discrepancy may be because the cue onset was already learned to be meaningful prior to the short ITI condition, thereby resulting in a stronger cue onset response in the experimental data. Further, we did not explicitly model potential trial-by-trial changes in eligibility trace time constant, sensory noise, internal threshold, local mechanisms controlling dopamine release, or sensor dynamics. Thus, we did not expect to capture all experimental observations in the motivating simulations. Regardless of such considerations, the current results provide a clear constraint for dopamine theories and demonstrate that an underappreciated experimental variable determines the emergence of mesolimbic dopamine ramps.
Methods
Animals
All experimental procedures were approved by the Institutional Animal Care and Use Committee at UCSF and followed guidelines provided by the NIH Guide for the Care and Use of Laboratory Animals. A total of 27 adult wild-type C57BL/6J mice (#000664, Jackson Laboratory) were divided between experiments: nine mice (4 females, 5 males) were used for the first cohort of Pavlovian conditioning, nine mice (4 females, 5 males) were used for the second cohort of Pavlovian conditioning, and nine mice (6 females, 3 males) were used for the VR task. Following surgery, mice were single housed in a reverse 12-hour light/dark cycle. Mice received environmental enrichment and had ad libitum access to standard chow. To increase motivation, mice underwent water deprivation. During deprivation, mice were weighed daily and given enough fluids to maintain ~85% of their baseline weight.
Surgeries
Surgical procedures were always done under aseptic conditions. Induction of anesthesia was achieved with 3% isoflurane, which was maintained at 1-2% throughout the duration of the surgery. Mice received subcutaneous injections of carprofen (5 mg/kg) for analgesia and lidocaine (1 mg/kg) for local anesthesia of the scalp prior to incision. A unilateral injection (Nanoject III, Drummond) of 500 nL of dLight1.3b(44) (AAVDJ-CAG-dLight1.3b, 2.4 × 1013 GC/mL diluted 1:10 in sterile saline) was targeted to the NAcC using the following coordinates from bregma: AP 1.3, ML ±1.4, DV −4.55. The glass injection pipette was held in place for 10 minutes prior to removal to prevent the backflow of virus. After viral injection, an optic fiber (NA 0.66, 400 µm, Doric Lenses) was implanted 100 µm above the site of injection. Subsequently, a custom head ring for head-fixation was secured to the skull using screws and dental cement. Mice recovered and were given at least three weeks before starting behavioral experiments. After completion of experiments, mice underwent transcardial perfusion and subsequent brain fixation in 4% paraformaldehyde. Fiber placement was verified using 50 µm brain sections under a Keyence microscope for subsequent visualization (Extended Data Fig 2a, Extended Data Fig 4a, Extended Data Fig 9a).
Behavior
All behavioral experiments took place during the dark cycle in dark, soundproof boxes with white noise playing to minimize any external noise. Prior to starting Pavlovian conditioning, water-deprived mice underwent 1-2 days of random rewards training to get acclimated to our head-fixed behavior setup(49). In a training session, mice received 100 sucrose rewards (~3 µL, 15% in water) at random time intervals taken from an exponential distribution averaging 12 s. Mice consumed sucrose rewards from a lick spout positioned directly in front of their mouths. This same spout was used for lick detection. After completing random rewards, mice were trained on Pavlovian conditioning. An identical trial structure was used across all conditions, consisting of an auditory tone lasting 8 s followed by a delay of 1 s before sucrose reward delivery. Two variables of interest were manipulated—the length of the ITI (long or short) and the type of auditory tone (fixed or dynamic)—resulting in four conditions: long ITI/fixed tone (LF), long ITI/dynamic tone (LD), short ITI/dynamic tone (SD), and short ITI/fixed tone (SF). In the first cohort (Fig 1), mice began with the LF condition (mean 7.4 days, range 7-8) before progressing to the LD condition (mean 6.1 days, range 5-11), the SD condition (8 days), and finally the SF condition (8 days). In the second cohort (Fig 2), the experimental order was switched such that mice began with the LF condition before moving on to the SD condition and ending with the LD condition (8 days for each condition). The ITI was defined as the period between reward delivery and the subsequent trial’s cue onset. In the long ITI conditions, the ITI was drawn from a truncated exponential distribution with a mean of 55 s, maximum of 186 s, and minimum of 6 s. The short ITIs were similarly drawn from a truncated exponential distribution, averaging 8 s with a maximum of 12 s and minimum of 6 s. While mice had 100 trials per day in the short ITI conditions, long ITI sessions were capped at 40 trials due to limitations on the amount of time animals could spend in the head-fixed setup. For the fixed tone conditions, mice were randomly divided into groups presented with either a 3 kHz or 12 kHz tone. While the 12 kHz tone played continuously throughout the entire 8 s, the 3 kHz tone was pulsed (200 ms on, 200 ms off) to make this lower frequency tone more obvious to the mice. For the dynamic tone conditions, the tone frequency either increased (dynamic up↑ starting at 3 kHz) or decreased (dynamic down↓ starting at 12 kHz) by 80 Hz every 200ms, for a total change of 3.2 kHz across 8 s. Mice with the 3 kHz fixed tone had the dynamic up↑ tone, whereas mice with the 12 kHz fixed tone had the dynamic down↓ tone. This dynamic change in frequency across the 8 s was intentionally designed to indicate to the mice the temporal proximity to reward, which is thought to be necessary for ramps to appear in a Pavlovian setting.
For the VR task, water-deprived mice were head fixed above a low-friction belt treadmill. A magnetic rotary encoder attached to the treadmill was used to measure the running velocity of the mice. In front of the head-fixed treadmill setup, a virtual environment was displayed on a high-resolution monitor (20” screen, 16:9 aspect ratio) to look like a dead-end hallway with a patterned floor, walls, and ceiling. The different texture patterns in the virtual environment were yoked to running velocity such that it appeared as though the animal was travelling down the hallway. Upon reaching the end of the hallway, the screen would turn fully black and mice would receive sucrose reward delivery from a lick spout positioned within reach in front of them. The screen remained black for the full duration of the ITI until the reappearance of the starting frame of the virtual hallway signaled the next trial onset. To train mice to engage in this VR task, they began with a 10 cm long virtual hallway. This minimal distance requirement was chosen to make it relatively easy for the mice to build associations between their movement on the treadmill, the corresponding visual pattern movement displayed on the VR monitor, and reward deliveries. Based on their performance throughout training, the distance requirement progressively increased by increments of 5-20 cm across days until reaching a maximum distance of 67 cm. Training lasted an average of 21.4 days (range 11-38 days), ending once mice could consistently run down the full 67 cm virtual hallway for three consecutive days. The ITIs during training (“med ITI”) were randomly drawn from a truncated exponential distribution with a mean of 28 s, maximum of 90 s, and minimum of 6 s. Following training, mice were randomly divided into two groups with identical trials but different ITIs (long or short). Again, both ITIs were randomly drawn from truncated exponential distributions: long ITI (mean 62 s, max 186 s, min 6 s) and short ITI (mean 8 s, max 12 s, min 6 s). After 8 days of the first ITI condition, mice switched to the other condition for an additional 8 days. There were 50 trials per day in both the long and short ITI conditions.
Fiber Photometry
Beginning three weeks after viral injection, dLight photometry recordings were performed with either an open-source (PyPhotometry) or commercial (Doric Lenses) fiber photometry system. Excitation LED light for wavelengths of 470 nm (dopamine dependent dLight signal) and 405 nm (dopamine independent isosbestic signal) were sinusoidally modulated via an LED driver and integrated into a fluorescence minicube (Doric Lenses). The same minicube was used to detect incoming fluorescent signals at a 12 kHz sampling frequency before demodulation and downsampling to 120 Hz. Excitation and emission light passed through the same low autofluorescence patchcord (400 µm, 0.57 NA, Doric Lenses). Light intensity at the tip of this patchcord was consistently ~40 µW across days. For Pavlovian conditioning, the photometry software received a TTL signal for the start and stop of the session to align the behavioral and photometry data. For alignment in the VR task, the photometry software received a TTL signal at each reward delivery.
Data Analysis
Behavior: Licking was the behavioral readout of learning used in Pavlovian conditioning. The lick rate was calculated by binning the number of licks every 100 ms. A smoothed version produced by Gaussian filtering is used to visualize lick rate in PSTHs (Fig 1f, Extended Data Fig 5a, Extended Data Fig 9d). Anticipatory lick rate for the last three days combined per condition was calculated by subtracting the average baseline lick rate during the 1 s before cue onset from the average lick rate during the trace period 1 s before reward delivery (Fig 1g, Extended Data Fig 5b). The same baseline subtraction method was used to calculate the average lick rate during the 3 to 8 s post cue onset period (Ext Data Fig 5c).
Running velocity, rather than licking, was the primary behavioral readout of learning for the VR task. Velocity was calculated as the change in distance per time. Distance measurements were sampled every 50 ms throughout both the trial and ITI periods. Average PSTHs from the last three days per condition were used to visualize velocity aligned to trial onset (Fig 4d) and reward delivery (Fig 4f). The change in velocity at trial onset was calculated by subtracting the average baseline velocity (baseline being 1 s before trial onset) from the average velocity between 1-2 s after trial onset (Fig 4e). Pre-reward velocity was the mean velocity during the 1 s period before reward delivery (Fig 4g).
The inter-trial interval (ITI) used throughout is defined as the time period between the previous trial reward delivery and the current trial onset (Fig 1c, Ext Data Fig 9b). The inter-reward interval (IRI) is defined as the time period between the previous trial reward delivery and the current trial reward delivery (Ext Data Fig 9b). For the previous IRI vs trial slope analysis (Ext Data Fig 10), IRI outliers were removed from analysis if they were more than three standard deviations away from the mean of the original IRI distribution. Finally, trial durations in the VR task were defined as the time it took for mice to run 67 virtual cm from the start to the end of the virtual hallway (Ext Data Fig 9b-c).
Dopamine: To analyze dLight fiber photometry data, first a least-square fit was used to scale the 405 nm signal to the 470 nm signal. Then, a percentage dF/F was calculated as follows: dF/F = (470 – fitted 405) / (fitted 405) * 100. This session-wide dF/F was then used for subsequent analysis. The onset peak dF/F (Fig 1j, Fig 4i) was calculated by finding the maximum dF/F value within 1 s after onset and then subtracting the average dF/F value during the 1 s interval preceding onset (last three days per condition combined). For each trial in Pavlovian conditioning, the time aligned dLight dF/F signal during the “ramp window” of 3 to 8 s after cue onset was fit with linear regression to obtain a per-trial slope. These per-trial slopes were then averaged for each day separately (Fig 1l) or for the last three days in each condition (Fig 1m) for subsequent statistical analysis. A smoothing Gaussian filter was applied to the group average (Fig 1i, Fig 2b) and example trial (Fig 1k) dLight traces for visualization purposes.
Distance, rather than time, was used to align the dLight dF/F signal in the VR task. Virtual distances were sampled every 30 ms, while dF/F values were sampled every 10 ms. To sync these signals, the average of every three dF/F values was assigned to the corresponding distance value. Any distance value that did not differ from the previous distance value was dropped from subsequent analysis (as was its mean dF/F value). This was done to avoid issues with averaging if the animal was stationary. For each trial in the VR task, the distance aligned dLight dF/F signal during the “ramp window” of 20 to 57 cm from the start of the virtual hallway was fit with linear regression to obtain a per-trial slope. These per-trial slopes were then averaged for each day separately (Fig 4l) or for the last three days in each condition (Fig 4m) for subsequent statistical analysis. To visualize the group averaged distance aligned dLight trace (Fig 4j) and example trial traces (Fig 4k), the mean dF/F was calculated for every 1 cm after rounding all distance values to the nearest integer.
Simulations
We previously proposed a learning model called Adjusted Net Contingency of Causal Relation (ANCCR)(32), which postulates that animals retrospectively search for causes (e.g., cues) when they receive a meaningful event (e.g., reward). ANCCR measures this retrospective association, which we call predecessor representation contingency (PRC), by comparing the strength of memory traces for a cue at rewards (M←cr; Equation 1) to the baseline level of memory traces for the same cue updated continuously (M←c−; Equation 2).
α and α0 are learning rates and the baseline samples are updated every dt seconds. E←ci represents eligibility trace of c) at the time of event i and E←c− represents eligibility trace of cue (c) at baseline samples updated continuously every dt seconds. The eligibility trace (E) decays exponentially over time depending on decay parameter T (Equation 4).
where ti ≤ t denotes the moments of past occurrences of event i. In Supplementary Note 1, we derived a simple rule he setting of T based on event rates. For the tasks considered here, this rule translated to a constant multiplied by IRI. We have shown in a revised version of a previous study(43) that
Statistics
All statistical tests were run on Python 3.11 using the scipy (version 1.10) package. Full details related to statistical tests are included in Supplementary Table 1. Data presented in figures with error bars represent mean ± SEM. Significance was determined using 0.05 for α. *p < 0.05, **p < 0.01, ***p < 0.001, ns p > 0.05.
Supplementary Note 1
Setting of eligibility trace time constant
It is intuitively clear that the eligibility trace time constant T needs to be set to match the timescales operating in the environment. This is because if the eligibility trace decays too quickly, there will be no memory of past events, and if it decays too slowly, it will take a long time to correctly learn event rates in the environment. Further, the asymptotic value of the baseline memory trace of event x, M←x− for an event train at a constant rate λx with average period tx is T /tx = T λx. This means that the neural representation of M←x− will need to be very high if T is very high and very low if T is very low. Since every known neural encoding scheme is non-linear at its limits with a floor and ceiling effect (e.g., firing rates can’t be below zero or be infinitely high), the limited neural resource in the linear regime should be used appropriately for efficient coding. A linear regime of operation for M←x− is especially important in ANCCR since the estimation of the successor representation by Bayes’ rule depends on the ratio of M←x− for different event types. Such a ratio will be highly biased if the neural representation of M←x− is in its non-linear range. Assuming without loss of generality that the optimal value of M←x− is Mopt for efficient linear coding, we can define a simple optimality criterion for the eligibility trace time constant T. Specifically, we postulate that the net sum of squared deviations of M←x− from Mopt for all event types should be minimized at the optimal T. The net sum of squared deviations, denoted by SS, can be written as
Where the second equality assumes asymptotic values of M←x−. The minimum of SS with respect to T will occur when
For typical cue-reward experiments with each cue predicting reward at 100% probability,
Thus, in typical experiments with 100% reward probability, the eligibility trace time constant should be proportional to the IRI or the total trial duration, which is determined by the ITI—the experimental proxy that we manipulate. Please do note, however, that the above relationship is not strictly controlled by the ITI, but by the frequency of repeating events in the environment (i.e., environmental timescale).




Statistical Details

Dependence of ANCCR on eligibility trace time constant.
a. Schematic showing exponential decay of cue eligibility traces for two-cue sequential conditioning (left) and multi-cue conditioning (right) with a long inter-trial interval (ITI). In this case, a long ITI results in a proportionally large eligibility trace time constant, T, producing slow eligibility trace decay (Supplementary Note 1). Reward delivery time indicated by vertical dashed line. b. Schematized ANCCR magnitudes (arbitrary units) for cues in the two-cue (left) and multi-cue (right) conditioning paradigms with a long ITI. Since the eligibility trace for the first cue is still high at reward time, there is a large ANCCR at this cue. The remaining cues are preceded consistently by earlier cues associated with the reward, thereby reducing their ANCCR. c. Same conditioning trial structure as in a, but with a short ITI and smaller T, producing rapid eligibility trace decay. d. Schematized ANCCR magnitudes for cues in both conditioning paradigms with a short ITI. Since the eligibility trace for the first cue is low at reward time, there is a small ANCCR at this cue. Though the remaining cues are preceded consistently by earlier cues associated with the reward, the eligibility traces of these earlier cues decay quickly, thereby resulting in a higher ANCCR for the later cues.

Pavlovian conditioning cohort 1 histology and dopamine responses.
a. Mouse coronal brain sections showing reconstructed locations of optic fiber tips (red circles) in NAcC for Pavlovian conditioning cohort 1. b. Example average dLight traces for the last three days of all conditions. Vertical dashed lines at 3 and 8 s represent the ramp window period. Black lines display the linear regression fit during this period. c. Same as in b but for the average dLight traces across all animals (n = 9 mice).

Pavlovian conditioning licking behavior data.
a. PSTH showing average licking behavior for the last 3 days of each condition for Pavlovian conditioning cohort 2 (n = 9 mice). b. Average anticipatory lick rate (baseline subtracted) for 1 s trace preceding reward delivery for cohort 2 (p = 0.77). c. Comparison of average baseline subtracted lick rate during the ramp window (3 to 8 s after cue onset) across all conditions for both cohorts (p = 0.093, n = 18 mice). d. Comparison of average lick slope during the ramp window across all conditions (fixed tone vs dynamic tone: ***p = 9.8 ×10−4).

Pavlovian conditioning cohort 2 histology and dopamine responses.
a. Mouse coronal brain sections showing reconstructed locations of optic fiber tips (red circles) in NAcC for Pavlovian conditioning cohort 2. b. Example average dLight traces for the last three days of all conditions. Vertical dashed lines at 3 and 8 s represent the ramp window period. Black lines display the linear regression fit during this period. c. Same as in b but for the average dLight traces across all animals (n = 9 mice).

Pavlovian conditioning cumulative dopamine response data.
a. Individual plots for each mouse from Pavlovian conditioning cohort 1 displaying the cumulative distribution of per-trial slopes for the last three days in all conditions. Vertical dashed lines indicate the average trial slope for LF (grey), LD (teal), SD (pink), and SF (purple) conditions. b. Same as in a but for Pavlovian conditioning cohort 2 mice.

Pavlovian conditioning dopamine ramps do not correlate with pre-cue dopamine activity.
a. Average dLight dopamine signal for Pavlovian conditioning cohort 1. Black lines represent linear regression fit during the pre-cue window and ramp window, each marked with grey shaded regions. b. Average per-trial dLight slope during the pre-cue window for the last 3 days of each condition (p = 0.37). c. Linear regression β coefficients for per-trial ramp dLight slope vs. pre-cue dLight slope in SD condition calculated per animal (p = 0.082). d. Scatter plot with linear regression fit (black line) of Z-scored ramp slope vs pre-cue slope pooled across mice for all trials in the last 3 days of SD condition (p = 0.17).

Pavlovian conditioning dopamine responses do not correlate with broader estimates of ITI.
a. Linear regression β coefficients for trial dLight slope vs. average previous ITI for the past 1 through 10 ITIs calculated per animal for all trials in the last 3 days of the SD condition (**p = 0.0056, ns p > 0.05; using Benjamini-Hochberg Procedure). b.Same as in a but for the LD condition (*p = 0.019, ns p > 0.05).

No significant correlations exist between additional dopamine and behavior measurements.
a.Left, linear regression β coefficients for dLight slope vs. dLight onset peak calculated per animal in SD condition (p = 0.54). Right, scatter plot with linear regression fit (black line) of Z-scored dLight slope vs dLight onset peak pooled across mice for all trials in the last 3 days of SD condition (p = 0.84). b. Same as in a but for LD condition (left p = 0.99, right p = 0.84). c. Left, linear regression β coefficients for dLight slope vs. lick slope calculated per animal in SD condition (p = 0.52). Right, scatter plot with linear regression fit (black line) of Z-scored dLight slope vs lick slope pooled across mice for all trials in the last 3 days of SD condition (p = 0.15). d. Same as in c but for LD condition (left p = 0.52, right p = 0.44). e. Left, linear regression β coefficients for dLight slope vs. dLight onset peak calculated per animal in SD condition (p = 0.54). Right, scatter plot with linear regression fit (black line) of Z-scored dLight slope vs dLight onset peak pooled across mice for all trials in the last 3 days of SD condition (p = 0.84). f. Same as in e but for LD condition (left p = 0.52, right p = 0.091). g. Left, linear regression β coefficients for dLight slope vs. dLight onset peak calculated per animal in SD condition (p = 0.54). Right, scatter plot with linear regression fit (black line) of Z-scored dLight slope vs dLight onset peak pooled across mice for all trials in the last 3 days of SD condition (p = 0.86). h. Same as in g but for LD condition (left p = 0.54, right p = 0.84). Benjamini-Hochberg Procedure used for p values from all t-tests and p values from all linear regression separately for all comparisons in this figure.

VR navigation task histology and responses.
a. Mouse coronal brain sections showing reconstructed locations of optic fiber tips (red circles) in NAcC for VR navigation task. b. Left, CDF of ITI duration for long (teal), medium (grey), and short (pink) ITI conditions. Middle, CDF plot of inter-reward interval (IRI) durations for each condition. Right, CDF plot of trial durations for each condition. c.Comparison of average trial duration for long and short ITI conditions (p = 0.34). d. Lick rate PSTH aligned to reward delivery indicates minimal anticipatory licking behavior. e. Scatter plot showing relationship between average per-session slope and inter-reward interval for the last 3 days in long (teal) and short (pink) ITI conditions. Black line indicates linear regression fit (*p = 0.012). f. CDF plots for each mouse separately showing the distribution of per-trial slopes for the last three days in both conditions. Vertical dashed lines indicate the average trial slope for long (teal) and short (pink) ITI conditions.

Trial-by-trial correlation of dopamine response slope vs previous inter-reward interval (IRI) in the VR task.
a. Scatter plot for an example animal showing the relationship between dopamine response slope within a trial and previous inter-reward interval (IRI) for all trials in the last 3 days of the short ITI condition. Plotted with linear regression fit (black line) used to find this animal’s β coefficient of −0.0014. Here, we are measuring the environmental timescale using IRI instead of ITI because the trial duration in this task (see Supplementary Note 1) depends on the running speed of the animals, which varies trial to trial. Thus, IRI measures the net time interval between successive trial onsets. In the Pavlovian conditioning experiment, IRI and ITI differ by a constant since the trial duration is fixed. b. Linear regression β coefficients for previous IRI vs trial slope calculated per animal (p = 0.32). c. Scatter plot of Z-scored trial slope vs. previous IRI pooled across mice for all trials in the last 3 days of the Short ITI condition (*p = 0.035).
Acknowledgements
We thank J. Berke and members of the Namboodiri laboratory for helpful discussions. This project was supported by the NIH (grants R00MH118422 and R01MH129582 to V.M.K.N.), the NSF (graduate research fellowship to J.R.F.), the UCSF Discovery Fellowship (J.R.F.), and the Scott Alan Myers Endowed Professorship (V.M.K.N.). The authors have no competing interests.
Additional information
Author contributions
J.R.F. and V.M.K.N. conceived the project. J.R.F. performed experiments and analyses. H.J. performed simulations. A.M. helped with the design and instrumentation of the VR experiment. V.M.K.N. oversaw all aspects of the study. J.R.F. and V.M.K.N. wrote the manuscript with help from all authors.
Funding
National Institute of Mental Health (R00MH118422)
National Institute of Mental Health (R01MH129582)
National Science Foundation
References
- 1.What does dopamine mean?Nature Neuroscience 21:787–793Google Scholar
- 2.Phasic versus tonic dopamine release and the modulation of dopamine system responsivity: a hypothesis for the etiology of schizophreniaNeuroscience 41:1–24Google Scholar
- 3.Tonic dopamine: opportunity costs and the control of response vigorPsychopharmacology 191:507–520Google Scholar
- 4.Prolonged dopamine signalling in striatum signals proximity and value of distant rewardsNature 500:575–579Google Scholar
- 5.Mesolimbic dopamine signals the value of workNature Neuroscience 19:117–126Google Scholar
- 6.Dissociable dopamine dynamics for learning and motivationNature 570:65–70Google Scholar
- 7.Dynamic mesolimbic dopamine signaling during action sequence learning and expectation violationScientific Reports 6:20231Google Scholar
- 8.Action initiation shapes mesolimbic dopamine encoding of future rewardsNature Neuroscience 19:34–36Google Scholar
- 9.Specialized coding of sensory, motor and cognitive variables in vta dopamine neuronsNature 570:509–513Google Scholar
- 10.Ramping activity in midbrain dopamine neurons signifies the use of a cognitive mapbioRxiv Google Scholar
- 11.A unified framework for dopamine signals across timescalesCell 183:1600–1616Google Scholar
- 12.Wave-like dopamine dynamics as a mechanism for spatiotemporal credit assignmentCell 184:2733–2749Google Scholar
- 13.Slowly evolving dopaminergic activity modulates the moment-to-moment probability of reward-related self-timed movementseLife 10:e62583https://doi.org/10.7554/eLife.62583Google Scholar
- 14.The neural basis of delayed gratificationScience Advances 7:eabg6611Google Scholar
- 15.Midbrain dopamine neurons signal phasic and ramping reward prediction error during goal-directed navigationCell Reports 41Google Scholar
- 16.Dual credit assignment processes underlie dopamine signals in a complex spatial environmentNeuron 111:3465–3478Google Scholar
- 17.Dopamine ramps upNature 500:533–535Google Scholar
- 18.Tamping ramping: Algorithmic, implementational, and computational explanations of phasic dopamine signals in the accumbensPLOS Computational Biology 11:e1004622Google Scholar
- 19.The role of state uncertainty in the dynamics of dopamineCurrent Biology 32:1077–1087Google Scholar
- 20.Dopamine ramps are a consequence of reward prediction errorsNeural Computation 26:467–471Google Scholar
- 21.Striatal dopamine ramping may indicate flexible reinforcement learning with forgetting in the cortico-basal ganglia circuitsFrontiers in Neural Circuits 8Google Scholar
- 22.The short-latency dopamine signal: a role in discovering novel actions?Nature Reviews Neuroscience 7:967–975Google Scholar
- 23.The debate over dopamine’s role in reward: the case for incentive saliencePsychopharmacology 191:391–431Google Scholar
- 24.Dopamine in motivational control: rewarding, aversive, and alertingNeuron 68:815–834Google Scholar
- 25.Reinforcement signalling in drosophila; dopamine does it all after allCurrent Opinion in Neurobiology 23:324–329Google Scholar
- 26.Rethinking dopamine as generalized prediction errorProceedings of the Royal Society B: Biological Sciences 285:20181645Google Scholar
- 27.Dopamine neurons create pavlovian conditioned stimuli with circuit-defined motivational propertiesNature Neuroscience 21:1072–1083Google Scholar
- 28.Dopamine transients do not act as model-free prediction errors during associative learningNature Communications 11:106Google Scholar
- 29.Ventral tegmental dopamine neurons control the impulse vector during motivated behaviorCurrent Biology 30:2681–2694Google Scholar
- 30.Causal evidence supporting the proposal that dopamine transients function as temporal difference prediction errorsNature Neuroscience 23:176–178Google Scholar
- 31.Dopamine release in the nucleus accumbens core signals perceived saliencyCurrent Biology 31:4748–4761Google Scholar
- 32.Mesolimbic dopamine release conveys causal associationsScience 378:eabq6740Google Scholar
- 33.Dopamine mediates the bidirectional update of interval timingBehavioral Neuroscience 136:445Google Scholar
- 34.Mesolimbic dopamine adapts the rate of learning from actionNature 614:294–302Google Scholar
- 35.Dopamine neurons encode a multidimensional probabilistic map of future rewardbioRxiv Google Scholar
- 36.Multi-timescale reinforcement learning in the brainbioRxiv Google Scholar
- 37.Dopaminergic prediction errors in the ventral tegmental area reflect a multithreaded predictive modelNature Neuroscience 26:830–839Google Scholar
- 38.Spontaneous behaviour is structured by reinforcement without explicit rewardNature 614:108–117Google Scholar
- 39.Does phasic dopamine release cause policy updates?European Journal of Neuroscience Google Scholar
- 40.Dynamic behaviour restructuring mediates dopamine-dependent credit assignmentNature 626:583–592Google Scholar
- 41.Dopamine projections to the basolateral amygdala drive the encoding of identity-specific reward memoriesNature Neuroscience :1–9Google Scholar
- 42.Mesostriatal dopamine is sensitive to changes in specific cue-reward contingenciesScience Advances 10:eadn4203Google Scholar
- 43.Few-shot learning: temporal scaling in behavioral and dopaminergic learningbioRxiv Google Scholar
- 44.Ultrafast neuronal imaging of dopamine dynamics with designed genetically encoded sensorsScience 360:eaat4422Google Scholar
- 45.Creating and controlling visual environments using bonvisioneLife 10Google Scholar
- 46.How do real animals account for the passage of time during associative learning?Behavioral Neuroscience 136:383–391Google Scholar
- 47.Rationalizing decision-making: understanding the cost and perception of timeTiming & Time Perception Reviews 1Google Scholar
- 48.Intertrial unconditioned stimuli differentially impact trace conditioningLearning & Behavior 45:49–61Google Scholar
- 49.An open-source behavior controller for associative learning and memory (b-calm)Behavior Research Methods :1–16Google Scholar
Article and author information
Author information
Version history
- Sent for peer review:
- Preprint posted:
- Reviewed Preprint version 1:
- Reviewed Preprint version 2:
Cite all versions
You can cite all versions using the DOI https://doi.org/10.7554/eLife.98666. This DOI represents all versions, and will always resolve to the latest one.
Copyright
© 2024, Floeder et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 538
- downloads
- 27
- citations
- 5
Views, downloads and citations are aggregated across all versions of this paper published by eLife.