Abstract
The processing of temporally-extended sequences of stimuli critically relies on Working Memory (WM). Yet, how WM supports the encoding and retrieval of novel sequences is unknown. Existing theories rely on associative learning driven by repetitions and are, thus, unable to explain how people can reproduce novel sequences of stimuli immediately. Here, we propose that detailed temporal information about a novel sequence can be rapidly stored in WM by short-term synaptic plasticity over multiple time scales. To substantiate this proposal, we extend our previously-proposed synaptic theory of WM to include synaptic augmentation, besides more short-lived depression and facilitation, consistently with experimental observations. The long time scales associated with augmentation naturally lead to the emergence of a temporal gradient in the synaptic efficacies, which can be used to immediately replay, at normal speed or in a time-compressed way, novel sequences. The theory is consistent with behavioral and neurophysiological observations.
1 Introduction
Purposeful behavior requires storing and retrieving relevant information over multiple time scales. Typically, this information also includes a temporal component that is key to achieve the goal. For instance, to reach the closest coffee place we just asked directions to, we have to turn left at the next corner, walk one block, and then turn right. We’ll get no espresso following the directions in the wrong order.
The Working Memory (WM) – a specialized, low-capacity component of the memory system – is believed to be responsible for rapidly encoding and maintaining novel information (e.g., the directions we just asked) over short time scales Cowan [2001], Baddeley [2003]. But, exactly, which information about a novel sequence of stimuli is stored in WM?
People effortlessly remember short but otherwise arbitrary (i.e., novel) sequences of familiar stimuli. For instance, people routinely hum short tunes they have just heard while musicians can even replay them with fidelity. A person can finger-tap, out of memory and with good accuracy, a pattern of irregularly spaced clicks spread over a few seconds which has been just experienced. The encoding of serial order, in particular, has been extensively investigated in the serial recall task Kahana [2012]. In this task, a list of randomly chosen items (e.g., words) is presented sequentially to the subject that, then, has to recall them in the presented order. This task is thought to rely on WM and, indeed, the number of correctly recalled items – typically about 4 items – is a standard measure of WM capacity. Interestingly, people almost invariably recall short lists (i.e., within WM capacity) in the presented order, even without explicit instructions to do so, as in the free recall task Dimperio et al. [2005], Ward et al. [2010], Grenfell-Essam and Ward [2012].
These observations suggest that WM rapidly and automatically stores quite detailed temporal information about a novel sequence of familiar stimuli, besides information about the stimuli themselves. Furthermore, this temporal information can be retrieved without any special training.
The models originally proposed for the computational architecture of WM have no mechanism for the encoding of temporal information Cowan [2001], Baddeley [2003]. As to neuronal models of WM (i.e., short-term memory maintenance), there have been different proposals. The most popular idea is that active maintenance relies on the co-existence of stable steady states of activity in the memory network (attractors) that are selected by stimulus presentations Amit [1995], Amit and Brunel [1997], Wang [2021]. The current state of activity, thus, reflects the recent history of stimulation. This mechanism can store the identity of the stimuli in the sequence but not information about their relative timing (e.g., the order of occurrence). Instead, this information needs to be learned in the course of multiple repetitions of the same sequence Kleinfeld [1986], Sompolinsky and Kanter [1986].
Partly to address the inability of attractor networks to rapidly store temporal information, an alternative account has been proposed Maass et al. [2002], Buonomano and Maass [2009]. In this account, active maintenance relies on the transitory, but high-dimensional, responses elicited by stimulus presentation in the memory network (liquid state machine). Such a mechanism can rapidly store the identify of the stimuli as well as detailed information about their times of occurrence, thanks to the high-dimensionality of the response. However, it is now the read-out of this information that needs to be learned, again in the course of multiple repetitions Cueva et al. [2020], Zhou et al. [2023].
Somehow surprisingly in view of their profound differences, these two accounts make one identical prediction; temporal information about a novel sequence of stimuli is not immediately available (e.g., to produce some behavior), either because it has not been stored yet (attractor networks) or because it cannot yet be read-out (liquid state machines). We have just discussed evidence contrary to this prediction.
We have proposed that information is maintained in WM by synaptic facilitation within the neuronal populations that code for the items, rather than by the enhanced, persistent activity of those populations Mongillo et al. [2008]. Facilitation is an experimentally well-characterized transient enhancement of the synaptic efficacy that is quickly induced by pre-synaptic spiking activity and can last for up to several seconds Markram et al. [1998], Zucker and Regehr [2002]. In particular, facilitation was reported at inter-pyramidal connections in the prefrontal cortex, a region heavily implicated in WM Hempel et al. [2000], Wang et al. [2006]. The theory is compatible with multiple experimental observations and motivated further experiments aimed at disentangling persistent activity and information maintenance Rose et al. [2016], Wolff et al. [2017], Panichello et al. [2024].
In the framework of the synaptic theory of WM, the maintenance of information can be achieved by different regimes of neuronal activity, depending on the background input to the network; at increasing levels of the background input, these regimes are: (i) activity-silent, where the information is transiently maintained without enhanced spiking activity; (ii) low-activity, where the information is periodically refreshed, at low rate, by brief spontaneous reactivations of corresponding neuronal populations (i.e., population spikes, PSs); (iii) persistent-activity, where the information is maintained by tonically active neuronal populations.
Facilitation is not the only form of transient synaptic enhancement induced by repetitive pre-synaptic activity. Exper-iments reveal other forms, such as augmentation and potentiation, which build up more slowly than facilitation but are significantly more long-lived Fisher et al. [1997], Thomson [2000], Fioravante and Regehr [2011]. As a result, the instantaneous value of the synaptic efficacy can reflect the history of pre-synaptic activation over tens of seconds (i.e., the time scale of augmentation) or even minutes (i.e., the time scale of potentiation) rather than just seconds (i.e., the time scale of facilitation). In the present contribution, we propose that such a transient synaptic enhancement over multiple time scales allows the encoding of both stimulus and temporal information in the instantaneous synaptic efficacies.
To substantiate this proposal, we extend the synaptic theory of WM to include synaptic augmentation, observed in the prefrontal cortex at the same synapses that exhibit significant short-term facilitation Hempel et al. [2000], Wang et al. [2006].
2 Results
To illustrate the putative role of synaptic augmentation in the encoding of temporal information, we consider the simplified setting used in Mi et al. [2017]. The network is composed of P distinct excitatory populations, that represent the memory items, and one inhibitory population, that prevents simultaneous activity at enhanced rates in the excitatory populations. The recurrent synaptic connections within each excitatory population display short-term synaptic plasticity according to the Tsodyks-Markram (TM) model Markram et al. [1998]. The population-averaged synaptic input to population a (a = 1, …, P), ha, evolves in time according to
where τ is the neuronal time constant; Ia(t), the external input to population a, is the sum of two components: a background input, to control the activity regime of the network, and a selective input, to elicit enhanced activity during the presentation of the corresponding item; Aa is the average strength of the synapses within excitatory population a; ra, the average activity of population a, is a smoothed threshold-linear function of ha, i.e.,
where α > 0 is a parameter controlling the smoothing; ua and xa are, respectively, the levels of short-term facilitation and depression of the recurrent synapses within population a; AEI is the strength of the synapses from the inhibitory population to any excitatory population; rI = ϕ (hI) is the average activity of the inhibitory population, and
where II is the constant background input to the inhibitory population and AIE is the strength of the synapses from any excitatory population to the inhibitory population.
The levels of short-term facilitation and depression, ua and xa, evolve in time according to Tsodyks et al. [1998]:
where U is the baseline release probability; τF and τD are the facilitation and depression time constants, respectively. In words: Activity in the population induces both facilitation, i.e., it increases ua, and depression, i.e., it decreases xa, while, in the absence of activity (i.e., ra = 0), facilitation and depression decay to their respective baseline levels, ua = U and xa = 1.
In Mi et al. [2017], the Aa’s in Equation 1 are time-independent parameters with the same value for all the excitatory populations. By contrast here, to model synaptic augmentation, the Aa’s are activity-dependent dynamic variables that increase with the ra’s according to
where Ā is the basal average strength of the synapses within an excitatory population (i.e., following a long period of synaptic inactivity), τA is the augmentation time constant, KA controls how fast the average strength of the synapses, Aa, increases with the activity, and AM is the maximal increase of the basal synaptic strength that can be induced by augmentation. In the following, we take AM = 3Ā.
The physiological mechanisms responsible for synaptic augmentation are poorly understood Fisher et al. [1997], Thomson [2000], Fioravante and Regehr [2011]. Equation 6 provides a minimal phenomenological description of synaptic augmentation in the spirit of the original TM model Markram et al. [1998]. However, as it will become clear in the following, our results do not depend critically on this modeling choice. For instance, one would obtain the same results by modeling augmentation as an activity-dependent increase in the baseline release probability U (data not shown). Facilitating synaptic transmission observed at inter-pyramidal synapses in the prefrontal cortex is well described by the above model with the following choice of synaptic parameters: U ∼ 0.2, τF ∼ 1s, τD ∼ 0.1s, τA ∼ 10s and KA ≪ 1 Hempel et al. [2000], Wang et al. [2006], Barri et al. [2016]. The full set of network and short-term plasticity parameters used in the simulations can be found in the caption of Fig. 1.

Network activity encodes sequence information.
Network responses to 3 sequentially presented items with (A) and without synaptic augmentation (B). The bottom panel in (A) shows the level of synaptic augmentation in the corresponding synaptic populations. The presentation of an item is simulated by a 10-fold increase of the background input selectively to the corresponding neuronal population for 250ms (gray areas). The background input to the remaining populations is kept constant at its baseline level. Network parameters: P = 16, τ = 8ms, α = 1.5Hz, AEE = 6.0, AEI = 1.1, AIE = 1.75, Ibkg = 8.0Hz; Short-term plasticity parameters: U = 0.3, KA = 0.01, τD = 0.3s, τF = 1.5s, τA = 20s.
In Fig. 1A, we show the response of the model network to a sequence of 3 items with variable inter-item intervals. The interval between the onset of the first and second item is 1 second, while the interval between the onset of the second and the third item is 2 seconds. Following the presentation of the last item, the neuronal populations that have been stimulated reactivate in a repeating cycle, indicating that the corresponding items are being actively maintained in WM. Importantly, this regime of activity does not correspond to a steady state of the network dynamics. This is evident from the amplitudes of the PS and from the levels of synaptic augmentation in the reactivating neuronal populations (Fig. 1A, bottom panel) that are still changing with time. Note that the amplitudes of the PS are different for the different populations.
For comparison, we show in Fig. 1B the response of the network to the same sequence in the absence of synaptic augmentation (i.e., Aa = Ā and KA = 0). As can be seen, in this case the network dynamics rapidly (on a time scale ∼ τF) converge to a steady, attractor state; the amplitude of the PS is the same for all the reactivating populations. Once this state reached, the network activity carries no information about the sequence beyond the identity of the stimuli composing the sequence.
The transient regime exhibited by the model network in the presence of augmentation is long-lived because the level of augmentation grows slowly with time. As can be seen in the bottom panel of Fig. 1A, significant augmentation only occurs during the reactivations. The increase of the augmentation level with each reactivation is mainly controlled by KA and KA is small, consistently with the experiments. As a result, the levels of synaptic augmentation in the reactivating neuronal populations are still changing with time long after the presentation of the last item in the sequence. At the same time, the decay of the level of augmentation between two consecutive reactivations of the same population (∼τD) is negligible, because τD ≪ τA. Therefore, the longer an item has been active in WM – that is, the larger the number of reactivations – the larger the corresponding level of augmentation. Indeed, the level of augmentation encodes, quite accurately, the time elapsed since item’s presentation (Fig. 1A, bottom panel).
In summary, in the presence of synaptic augmentation, WM activity naturally encodes the temporal structure of a novel sequence of familiar items, besides encoding information about the identity of the single items. As this information is present in the levels of synaptic augmentation and in the amplitudes of the PSs during the reactivations, it is readily accessible to a downstream read-out network.
To illustrate this important point, we consider a very simple read-out mechanism to reconstruct/replay the sequence stored in WM (Fig. 2A).

Replay of a novel sequence.
(A) Architecture of the memory and read-out network (see main text for details). (B) Top panel: Input to the read-out network from the memory network. Middle panel: Activation state of the item-selective populations in the read-out network, as determined by comparing the sum of the input from the memory network and of the background input to the threshold. Bottom panel: Time course of the background input to the read-out network. (C) Same as (B) for a different sequence. Note that the time course of the background input to the read-out network is the same in (B) and (C).
Each item-selective population in the memory network sends excitatory inputs to the corresponding item-selective population in a read-out network. For simplicity, we assume that the excitatory synapses between the memory and the read-out network feature the same dynamics as the excitatory synapses within the memory network. The activation of a population in the read-out network signals the retrieval of the corresponding item. The item-selective populations in the read-out network do not interact with each other. Rather, they receive a uniform background input (i.e., the same for all populations) that effectively sets the threshold for their activation. To read-out the contents of WM, following the presentation of the last stimulus, the background input ramps up until the first population in the read-out network activates, and then stays constant.
This mechanism results in the approximate replay of the sequence stored in WM. We illustrate this in Fig. 2B and C for two different sequences and the same read-out network. This is because the temporal evolution of the level of augmentation and of the amplitude of the PSs in the active populations are approximately time-translation invariant; that is, they are approximately the same when aligned to stimulus onset. Hence, the inputs from the memory to the read-out network will reach the same level (i.e., the same threshold) at time intervals that (approximately) match the time intervals between the presentations of the corresponding items. For the same reason, the accuracy of the replay is rather robust against (reasonable) changes in the rate of increase of the input to the read-out network.
The augmentation gradient can also be used to fast-replay the sequence stored in WM (i.e., in a time-compressed way). Fast replay has been suggested as a mechanism for consolidating the storage of information in the long-term memory, by bringing patterns of neuronal activity representing temporally distant events within a time window in which associative, long-term synaptic plasticity can most effectively operate Melamed et al. [2004], Jensen and Lisman [2005]. The time-compressed replay of the sequence is initiated by decreasing the level of background input to the WM network for (at least) a time ∼ τF (3, bottom panel). This prevents further reactivations (Fig. 3, top panel) and the synaptic variables start decaying toward their baseline levels (Fig. 3, middle panel). The background input is then raised again to a suitably larger level. The levels of augmentation, i.e., the Aa’s, have hardly changed because τF ≪ τA. However, short-term depression and facilitation will be close to their corresponding baseline levels (i.e., xa ≃ 1 and ua ≃ U for a = 1, …, P). For the once-active neuronal populations, however, the steady, low-rate state of activity becomes unstable when the background input is raised above a critical level. Hence, they will start reactivating, with the most unstable one (i.e., the one with the larger Aa) reactivating first, the next most unstable one reactivating second, and so on Mi et al. [2017]. As a result, the reactivations follow the temporal gradient encoded in the augmentation levels (Fig. 2, top panel). Note that the network replays the sequence in about 250 ms, thus achieving a compression factor of about 10.

Fast-replay of a novel sequence.
The top panel shows the response of the network to the external inputs depicted in the bottom panel. After the presentation of the sequence, the background input is first decreased for 3 seconds and then increased for 250 milliseconds. The middle panel shows the resulting time course of Aux in the corresponding synaptic populations. Immediately before the background input is increased again, Aux ≃ AU.
There is significant experimental evidence that items can be maintained in WM in different representational states and that these states can be rapidly altered by task demand LaRocque et al. [2014], Oberauer and Awh [2022]. A case in point is the study of Rose et al. [2016], who used a retro-cue design to manipulate these putative representational states. Briefly, human subjects performed a two-item delayed recognition task with two retro-cues and two recognition probes per trial. Following items’ presentation and an initial delay period, the first retro-cue informed the subject about which of the two items will be probed in the impending (i.e., after a delay period) recognition test. The cued item is considered prioritized for upcoming behavior. After the first recognition test, a second retro-cue indicated the item to be probed, following another delay period, in the second recognition test. Importantly, each retro-cue randomly prioritized either of the two items with the same probability and, hence, both items had to be kept in WM until the second retro-cue. Rose et al. [2016] found that, during the initial delay period, both items could be reliably decoded from the fMRI signal. By contrast, during the delay period between a retro-cue and the subsequent recognition test, only the prioritized item could be reliably decoded. However, the decodability of the de-prioritized item could be recovered by transcranial magnetic stimulation.
The experimental observations of Rose et al. [2016] (see also Wolff et al. [2017]) support the idea that there exist (at least) two states of ‘maintenance in WM’ with distinct neurophysiological signatures. Though the relationship between the fMRI signal and neural activity is not an obvious one, these results have been interpreted as indicating that these two states differ in their level of neural activity. In particular, the failure to decode would indicate the absence of enhanced spiking activity (but see Barbosa et al. [2021]).
Our model network can reproduce these experimental observations, as illustrated in Fig. 4. Similarly to the experiment of Rose et al. [2016], two items are presented and, after the initial delay period, one of the two is prioritized. We simulate the effect of the retro-cue on neural activity by assuming that the de-prioritized item receives a lower external input as compared to the prioritized item (Fig. 4, bottom panel). Note that, as in the experiment of Rose et al. [2016] there are only two items, prioritizing one item is the same as de-prioritizing the other item. Following the first retro-cue, the neuronal population encoding the prioritized item keeps reactivating while the neuronal population encoding the de-prioritized item stops reactivating (Fig. 4, top panel). We account for the results of Rose et al. [2016] by assuming that only reactivating items can be decoded.

Reactivation of an activity-silent memory.
The top panel shows the response of the network to the external inputs depicted in the bottom panel. 8 seconds after the presentation of the two items, the background input to the neuronal population selective to second item (red) is decreased for 8 seconds, and then restored to its original level, while decreasing the background input to the neuronal population selective to the first item (blue). The middle panel shows the resulting time course of the level of augmentation in the corresponding synaptic populations.
In the absence of neural activity, the ‘working memory’ of the de-prioritized item is kept by the augmentation level (Fig. 4, middle panel). To illustrate this, we prioritize with the second retro-cue the previously de-prioritized item and, indeed, the corresponding neuronal population resumes its reactivating dynamics (Fig. 4, top panel). In the presence of synaptic augmentation, the time span of the activity-silent regime becomes of the order of τA, which is of the order of 10 seconds, and hence fully consistent with the experimental observations of Rose et al. [2016]. We note that in the model originally proposed in Mongillo et al. [2008], instead, the time span of the activity-silent regime was of the order of τF (∼ 1 s), quantitatively inconsistent with the results of Rose et al. [2016].
3 Discussion
We propose that transient, non-associative synaptic plasticity over multiple time scale can support the temporally-structured encoding of a sequence of stimuli. We have illustrated this idea in a minimal model network that extends the synaptic theory of WM to include synaptic augmentation, besides synaptic depression and facilitation. In the low-activity regime, where items are maintained by short-lived reactivations of the corresponding neuronal populations, the presence of synaptic augmentation naturally leads to a temporal gradient in the synaptic efficacies that encodes both the items and their relative times of occurrence. This gradient can then be used to replay the sequence either at normal speed or in a time-compressed way. The mechanism that generates the temporal gradient is robust, because it relies on the order-of-magnitude differences between the build-up and the decay time of the augmentation and those of the depression and facilitation.
Our model allows the storage and the retrieval of short sequences of items by relying on synaptic plasticity mechanisms that are well-characterized experimentally, that is, the transient enchancement on multiple time scales of the synaptic efficacy driven solely by pre-synaptic activity Fisher et al. [1997], Thomson [2000], Fioravante and Regehr [2011]. Alternative models, as already pointed out, rely instead on some form of associative learning Botvinick and Watanabe [2007], Gillett et al. [2020], Ryom et al. [2021], Zhou et al. [2023]. At the physiological level, associative learning is thought to entail long-term synaptic plasticity. This is because the induction of long-term synaptic plasticity is dependent on the joint pattern of pre- and post-synaptic activity, as required for associativity. However, there is presently no evidence that long-term synaptic plasticity can be induced and/or expressed on the time scales relevant to explain the experimental observations, that is, a few seconds Lansner et al. [2023].
A key prediction of our theory is that items are maintained in the low-activity regime. Indeed, if the items are maintained either in the activity-silent regime or in the persistent-activity regime, the proposed mechanism fails. In the first case, because in the absence of reactivations the gradient does not build up; in the second case, because the augmentation levels quickly saturate due to the enhanced firing rates. This prediction is consistent with multiple experimental observations Siegel et al. [2009], Fuentemilla et al. [2010], Lundqvist et al. [2016], Panichello et al. [2024]. In multi-item working memory tasks, the neuronal activity during the maintenance period is characterized by short episodes of spiking synchrony, detected as brief gamma bursts in the local field potential Siegel et al. [2009], Lundqvist et al. [2016] or in the MEG/EEG signal Fuentemilla et al. [2010]. These episodes, which we identify with the population spikes in our model, are associated with the reactivation of the neural representation of the items, as evidenced by the fact that item’s identity can be reliably decoded only during the gamma bursts. Importantly, during a given gamma burst, only information about one of the maintained items can be reliably decoded Fuentemilla et al. [2010], Lundqvist et al. [2016], suggesting that the items are reactivated one at a time, as required by our theory.
Behavioral data in serial-recall tasks strongly support the notion that the encoding of serial order relies on a primacy gradient that prioritizes recall Grossberg [1978], Farrell and Lewandowsky [2004], Hurlstone and Hitch [2015, 2018]. Our theory makes an explicit proposal as to its neurophysiological substrate: The primacy gradient is encoded by the augmentation levels and its generation depends on a specific interplay of the synaptic and neuronal dynamics (as described above). As such, the generation of the gradient is an inescapable consequence of the active maintenance of an item in WM. This would naturally explain why the recall order tends to be the same as the presentation order also in free-recall tasks, provided that the sequence does not exceed WM capacity.
In the behavioral context, our theory also makes novel predictions. For instance, the temporal gradient builds up gradually with the reactivations of the corresponding neuronal populations between consecutive presentations. This requires a presentation rate that is slow enough for these reactivations to occur in sufficient number. Hence, as the presentation rate is increased, the theory predicts that encoding of the serial order should degrade. Consistently with this prediction, increasing the presentation rate of the items results in a larger number of transposition errors, that is, some items are recalled at the wrong serial position (see, e.g., Farrell and Lewandowsky [2004]). Experiments with very rapid serial visual presentation (RSVP) of the items show that the subjects are unable to report the correct order of presentation, even when the number of items is below capacity Reeves and Sperling [1986]. At the other extreme, if the presentation rate is too slow, or the list is too long, then the primacy gradient will also degrade because of the saturation of the synaptic augmentation, or the inability of actively maintaining all the items. Consistently with this prediction, in serial-recall tasks, the spontaneous tendency in recalling the items with the presented order rapidly degrades with increasing list lengths Grenfell-Essam and Ward [2012].
We have used, to illustrate our theory, a minimal model network that neglects many physiological details. Hence, a comparison between the model behavior and the electrophysiological observations cannot be completely direct. Nevertheless, the model provides a unified account of diverse features of the experimental data, at least qualitatively.
The model naturally generates ramping activity as a consequence of active maintenance, that is, the average level of activity in a reactivating neuronal population increases with time (see, e.g., Fig. 1). Ramping activity has been proposed as a potential neuronal mechanism to encode time and it is commonly observed in electrophysiological studies of WM. A case in point is the recent study of Cueva et al. [2020], who found ramping activity during the maintenance period of different delayed-response tasks, regardless of whether or not timing was important to perform the task. Interestingly, the non-steady dynamics of neuronal activity during the delay period was almost entirely explained by the ramping activity Cueva et al. [2020]. We note that these observations are fully consistent with our model.
In neurophysiological studies of WM for sequences, the conjunctive coding of item identity and serial-order information at the single-neuron level has also commonly been observed Barone and Joseph [1989], Funahashi et al. [1997], Xie et al. [2022]. Conjunctive coding refers to the modulation of neuron’s activity by both item and order information, so that, for instance, the average firing rate of the neuron during the delay period following different sequences with the same item changes depending on the position of the item in the sequence Xie et al. [2022]. Also in our model the firing rate of a neuron is naturally sensitive to the temporal order due to the augmentation gradient (Fig. 2B). It remains to be seen whether our model, in a more physiologically-detailed setting, is able to quantitatively account for some features of conjunctive coding as observed in experiments Xie et al. [2022]. In this respect, an important caveat is that animals in these studies have been extensively trained on the task with a limited number of sequences. Extensive training and sequences’ repetition could lead to the emergence of stimulus-adapted neuronal representations via associative plasticity mechanisms Botvinick and Watanabe [2007], Gillett et al. [2020], Ryom et al. [2021].
The long time scales brought about by synaptic augmentation significantly extend the time span of memories maintained in the activity-silent state. As discussed in the Results section, the time scales of synaptic augmentation are fully compatible with the experimental observations, such as those of Rose et al. [2016], suggesting that a memory that has been silent for 10 seconds can still be successfully retrieved upon cueing (see Fig. 4). Accordingly, one would expect a very large storage capacity in the activity-silent mode. For instance, by assuming that an item is to be refreshed every ∼ 10 seconds to prevent its loss (based on Rose et al. [2016]), and that the refresh takes 100 milliseconds (based on Fuentemilla et al. [2010], Lundqvist et al. [2016]), one would estimate a storage capacity of 100 items for the activity-silent WM. This raises the possibility that WM capacity, which is an experimental estimate of (uncued) recall, could result from the inability to retrieve the information, rather than from the inability to encode and/or maintain it. In this scenario, WM capacity is ultimately determined by the degree of selectivity that the background control – that we identify with the “central executive” or the “focus of attention” of cognitive theories – can attain.
Acknowledgements
G.M. work is supported by grants ANR-19-CE16-0024-01 and ANR-20-CE16-0011-02 from the French National Research Agency and by a grant from the Simons Foundation (891851, G.M.). M.T. is supported by the Israeli Science Foundation grant 1657/19 and Foundation Adelis.
References
- The magical number 4 in short-term memory: A reconsideration of mental storage capacityBehavioral and Brain Sciences 24:87–114Google Scholar
- Working memory: looking back and looking forwardNature reviews neuroscience 4:829–839Google Scholar
- Foundations of human memoryOxford University Press Google Scholar
- A comparative analysis of serial and free recallMemory & Cognition 33:833–839Google Scholar
- Examining the relationship between free recall and immediate serial recall: the effects of list length and output orderJournal of Experimental Psychology: Learning, Memory, and Cognition 36:1207–1241Google Scholar
- Examining the relationship between free recall and immediate serial recall: The role of list length, strategy use, and test expectancyJournal of Memory and Language 67:106–148Google Scholar
- The hebbian paradigm reintegrated: Local reverberations as internal representationsBehavioral and Brain Sciences 18:617–626Google Scholar
- Model of global spontaneous activity and local structured activity during delay periods in the cerebral cortexCerebral cortex (New York, NY: 1991) 7:237–252Google Scholar
- 50 years of mnemonic persistent activity: quo vadis?Trends in Neurosciences 44:888–902Google Scholar
- Sequential state generation by model neural networksProceedings of the National Academy of Sciences 83:9469–9473Google Scholar
- Temporal association in asymmetric neural networksPhysical review letters 57:2861Google Scholar
- Real-time computing without stable states: A new framework for neural computation based on perturbationsNeural computation 14:2531–2560Google Scholar
- State-dependent computations: spatiotemporal processing in cortical networksNature Reviews Neuroscience 10:113–125Google Scholar
- Low-dimensional dynamics for working memory and time encodingProceedings of the National Academy of Sciences 117:23021–23032Google Scholar
- Multiplexing working memory and time in the trajectories of neural networksNature Human Behaviour 7:1170–1184Google Scholar
- Synaptic theory of working memoryScience 319:1543–1546Google Scholar
- Differential signaling via the same axon of neocortical pyramidal neuronsProceedings of the National Academy of Sciences 95:5323–5328Google Scholar
- Short-term synaptic plasticityAnnual Review of Physiology 64:355–405Google Scholar
- Multiple forms of short-term plasticity at excitatory synapses in rat medial prefrontal cortexJournal of neurophysiology 83:3031–3041Google Scholar
- Heterogeneity in the pyramidal network of the medial prefrontal cortexNature neuroscience 9:534Google Scholar
- Reactivation of latent working memories with transcranial magnetic stimulationScience 354:1136–1139Google Scholar
- Dynamic hidden states underlying working-memory-guided behaviorNature neuroscience 20:864–871Google Scholar
- Intermittent rate coding and cue-specific ensembles support working memoryNature :1–8Google Scholar
- Multiple overlapping processes underlying short-term synaptic enhancementTrends in neurosciences 20:170–177Google Scholar
- Facilitation, augmentation and potentiation at central synapsesTrends in neurosciences 23:305–312Google Scholar
- Short-term forms of presynaptic plasticityCurrent opinion in neurobiology 21:269–274Google Scholar
- Synaptic correlates of working memory capacityNeuron 93:323–330Google Scholar
- Neural networks with dynamic synapsesNeural Computation 10:821–835Google Scholar
- Quantifying repetitive transmission at chemical synapses: a generative-model approachENeuro 3Google Scholar
- Coding and learning of behavioral sequencesTrends in neurosciences 27:11–14Google Scholar
- Hippocampal sequence-encoding driven by a cortical multi-item working memory bufferTrends in neurosciences 28:67–72Google Scholar
- Multiple neural states of representation in short-term memory? it’sa matter of attentionFrontiers in human neuroscience 8:5Google Scholar
- Is There an Activity-silent Working Memory?Journal of Cognitive Neuroscience 34:2360–2374Google Scholar
- Pinging the brain with visual impulses reveals electrically active, not activity-silent, working memoriesPLoS biology 19:e3001436Google Scholar
- From numerosity to ordinal rank: a gain-field model of serial order representation in cortical working memoryJournal of Neuroscience 27:8636–8642Google Scholar
- Characteristics of sequential activity in networks with temporally asymmetric hebbian learningProceedings of the National Academy of Sciences 117:29948–29958Google Scholar
- Latching dynamics as a basis for short-term recallPLoS computational biology 17:e1008809Google Scholar
- Fast hebbian plasticity and working memoryCurrent Opinion in Neurobiology 83:102809Google Scholar
- Phase-dependent neuronal coding of objects in short-term memoryProceedings of the National Academy of Sciences 106:21341–21346Google Scholar
- Theta-coupled periodic replay in working memoryCurrent Biology 20:606–612Google Scholar
- Gamma and beta bursts underlie working memoryNeuron 90:152–164Google Scholar
- Behavioral contrast in short term memory: serial binary memory models or parallel continuous memory models?Journal of Mathematical Psychology 17:199–219Google Scholar
- Modelling transposition latencies: Constraints for theories of serial order memoryJournal of Memory and Language 51:115–135Google Scholar
- How is the serial order of a spatial sequence represented? insights from transposition latenciesJournal of Experimental Psychology: Learning, Memory, and Cognition 41:295–324Google Scholar
- How is the serial order of a visual sequence represented? Insights from transposition latenciesJournal of Experimental Psychology: Learning, Memory, and Cognition 44:167–192Google Scholar
- Attention gating in short-term visual memoryPsychological review 93:180–206Google Scholar
- Prefrontal cortex and spatial sequencing in macaque monkeyExperimental brain research 78:447–464Google Scholar
- Delay-period activity in the primate prefrontal cortex encoding multiple spatial positions and their order of presentationBehavioural brain research 84:203–223Google Scholar
- Geometry of sequence working memory in macaque prefrontal cortexScience 375:632–639Google Scholar
Article and author information
Author information
Version history
- Sent for peer review:
- Preprint posted:
- Reviewed Preprint version 1:
Cite all versions
You can cite all versions using the DOI https://doi.org/10.7554/eLife.107005. This DOI represents all versions, and will always resolve to the latest one.
Copyright
© 2025, Gianluigi Mongillo & Misha Tsodyks
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 0
- downloads
- 0
- citations
- 0
Views, downloads and citations are aggregated across all versions of this paper published by eLife.