Cognitive Neuroscience: Memorable first impressions
Look out the window and see what stands out. Perhaps you notice some red and pink azaleas in full bloom. Now close your eyes and picture that scene in your mind. Initially, the colors and silhouettes linger vividly, but the details wither rapidly, leaving only a faded version of the image. As time passes, the accuracy with which an image can be recalled drops abruptly.
Memory is a critical, wonderful, multifaceted mental capacity that relies on many structures and mechanisms throughout the brain (Baddeley, 2003; Squire and Wixted, 2011; Schacter et al., 2012). This is not surprising, given the diversity of timescales and data types – such as images, words, facts and motor skills – that we can remember. Studies have shown that our visual memories are strongest immediately after an image disappears, remaining reliable for about half a second. This has traditionally been attributed to ‘iconic’ memory, which is thought to rely on a direct readout of stimulus-driven activity in visual circuits in the brain. In this case, the memory remains vivid because, after the stimulus (i.e., the image) has been removed, the visual activity takes some time to decay (Sperling, 1960; Pratte, 2018; Teeuwen et al., 2021).
In contrast, recalling an image a second or so after it has disappeared engages a different type of memory – visual working memory – that relies on information stored in different circuits in the frontal lobe (Pasternak and Greenlee, 2005; D’Esposito and Postle, 2015). Although not as vivid, the stored image remains stable for much longer. This is because, despite being more robust, the storage capacity for visual working memory is more limited: fewer items and less detail can be recalled from a remembered image. Together, these findings led to the idea that there are two distinct short-term memory mechanisms. Now, in eLife, Ivan Tomić and Paul Bays report strong evidence indicating that iconic memory and visual working memory are part of the same recall mechanism (Tomić and Bays, 2023).
Tomić and Bays – who are based at the University of Zagreb and the University of Cambridge – first constructed a detailed computational model to describe how sensory information is passed to a visual working memory circuit for storage and later recall (Figure 1). In this model, visual neurons respond to the presentation of an image containing a few items. This stimulus causes sensory activity to rise smoothly while the input lasts, and to decay once the stimulus ceases, consistent with previous experiments (Teeuwen et al., 2021). This sensory response then drives a population of visual working memory neurons that can sustain their activity in the absence of a stimulus, although this activity will eventually be corrupted due to noise (Wimmer et al., 2014; DePasquale et al., 2018). An important feature of the model is that each remembered item is allocated an equal fraction of the maximum possible working memory activity.
The model constructed by Tomić and Bays can make specific testable predictions. For example, it predicts that if an item is cued for later recall while the sensory signal is still present, the working memory activity associated with the non-targets will decay rapidly, freeing up resources and thus increasing the working memory activity associated with the cued item. This leads to more accurate recall of the item. In contrast, if an item is cued for later recall once the sensory signal has approached zero, this ‘boost’ does not happen, and the item is not recalled as accurately. In addition, the working memory activity should increase with longer exposure to the stimulus and should decrease as the number of remembered items increases.
These predictions were confirmed through experiments with humans. Participants were shown visual stimuli while several factors were varied, including the number of items to be remembered, the duration of the stimulus, the time at which the item to be recalled was identified, and the time of the actual recall. The results of these experiments are consistent with the notion that, during recall, visual information is always read out from the same population of neurons.
The findings of Tomić and Bays are satisfying for their simplicity; what seemed to require two separate mechanisms is explained by a single framework aligned with many prior studies. However, models always require simplifications and shortcuts. For instance, much evidence indicates that both frontal lobe circuits and sensory areas contribute to the self-sustained maintenance of activity that underlies the short-term memory of sensory events (Pasternak and Greenlee, 2005). Therefore, visual working memory is likely the result of continuous recurrent dynamics across areas (DePasquale et al., 2018; Stroud et al., 2024). Furthermore, there is still debate about the degree to which visual working memory implies equal sharing of resources, as opposed to some items receiving larger or smaller shares (Ma et al., 2014; Pratte, 2018). Nevertheless, the proposed model is certainly an important advance that future studies can build upon.
References
-
Working memory: looking back and looking forwardNature Reviews Neuroscience 4:829–839.https://doi.org/10.1038/nrn1201
-
The cognitive neuroscience of working memoryAnnual Review of Psychology 66:115–142.https://doi.org/10.1146/annurev-psych-010814-015031
-
Working memory in primate sensory systemsNature Reviews Neuroscience 6:97–107.https://doi.org/10.1038/nrn1603
-
Iconic memories die a sudden deathPsychological Science 29:877–887.https://doi.org/10.1177/0956797617747118
-
The information available in brief visual presentationsPsychological Monographs 74:1–29.https://doi.org/10.1037/h0093759
-
The cognitive neuroscience of human memory since H.MAnnual Review of Neuroscience 34:259–288.https://doi.org/10.1146/annurev-neuro-061010-113720
-
The computational foundations of dynamic coding in working memoryTrends in Cognitive Sciences 4:S1364-6613(24)00053-6.https://doi.org/10.1016/j.tics.2024.02.011
-
A neuronal basis of iconic memory in macaque primary visual cortexCurrent Biology 31:5401–5414.https://doi.org/10.1016/j.cub.2021.09.052
Article and author information
Author details
Publication history
Copyright
© 2024, Salinas and Sheikh
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 574
- views
-
- 45
- downloads
-
- 0
- citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Neuroscience
During rest and sleep, memory traces replay in the brain. The dialogue between brain regions during replay is thought to stabilize labile memory traces for long-term storage. However, because replay is an internally-driven, spontaneous phenomenon, it does not have a ground truth - an external reference that can validate whether a memory has truly been replayed. Instead, replay detection is based on the similarity between the sequential neural activity comprising the replay event and the corresponding template of neural activity generated during active locomotion. If the statistical likelihood of observing such a match by chance is sufficiently low, the candidate replay event is inferred to be replaying that specific memory. However, without the ability to evaluate whether replay detection methods are successfully detecting true events and correctly rejecting non-events, the evaluation and comparison of different replay methods is challenging. To circumvent this problem, we present a new framework for evaluating replay, tested using hippocampal neural recordings from rats exploring two novel linear tracks. Using this two-track paradigm, our framework selects replay events based on their temporal fidelity (sequence-based detection), and evaluates the detection performance using each event's track discriminability, where sequenceless decoding across both tracks is used to quantify whether the track replaying is also the most likely track being reactivated.
-
- Neuroscience
The remarkable ability of the motor system to adapt to novel environments has traditionally been investigated using kinematically non-redundant tasks, such as planar reaching movements. This limitation prevents the study of how the motor system achieves adaptation by altering the movement patterns of our redundant body. To address this issue, we developed a redundant motor task in which participants reached for targets with the tip of a virtual stick held with both hands. Despite the redundancy of the task, participants consistently employed a stereotypical strategy of flexibly changing the tilt angle of the stick depending on the direction of tip movement. Thus, this baseline relationship between tip-movement direction and stick-tilt angle constrained both the physical and visual movement patterns of the redundant system. Our task allowed us to systematically investigate how the motor system implicitly changed both the tip-movement direction and the stick-tilt angle in response to imposed visual perturbations. Both types of perturbations, whether directly affecting the task (tip-movement direction) or not (stick-tilt angle around the tip), drove adaptation, and the patterns of implicit adaptation were guided by the baseline relationship. Consequently, tip-movement adaptation was associated with changes in stick-tilt angle, and intriguingly, even seemingly ignorable stick-tilt perturbations significantly influenced tip-movement adaptation, leading to tip-movement direction errors. These findings provide a new understanding that the baseline relationship plays a crucial role not only in how the motor system controls movement of the redundant system, but also in how it implicitly adapts to modify movement patterns.