Emotional vocalizations alter behaviors and neurochemical release into the amygdala
Abstract
The basolateral amygdala (BLA), a brain center of emotional expression, contributes to acoustic communication by first interpreting the meaning of social sounds in the context of the listener’s internal state, then organizing the appropriate behavioral responses. We propose that modulatory neurochemicals such as acetylcholine (ACh) and dopamine (DA) provide internal-state signals to the BLA while an animal listens to social vocalizations. We tested this in a vocal playback experiment utilizing highly affective vocal sequences associated with either mating or restraint, then sampled and analyzed fluids within the BLA for a broad range of neurochemicals and observed behavioral responses of adult male and female mice. In male mice, playback of restraint vocalizations increased ACh release and usually decreased DA release, while playback of mating sequences evoked the opposite neurochemical release patterns. In non-estrus female mice, patterns of ACh and DA release with mating playback were similar to males. Estrus females, however, showed increased ACh, associated with vigilance, as well as increased DA, associated with reward-seeking. Experimental groups that showed increased ACh release also showed the largest increases in an aversive behavior. These neurochemical release patterns and several behavioral responses depended on a single prior experience with the mating and restraint behaviors. Our results support a model in which ACh and DA provide contextual information to sound analyzing BLA neurons that modulate their output to downstream brain regions controlling behavioral responses to social vocalizations.
eLife assessment
This important study advances our understanding of how distinct types of communication signals differentially affect mouse behaviors and amygdala cholinergic/dopaminergic neuromodulation. The evidence supporting the authors' claims is solid. Researchers interested in the complex interaction between prior experience, sex, behavior, hormonal status, and neuromodulation should benefit from this study.
https://doi.org/10.7554/eLife.88838.4.sa0Important: Findings that have theoretical or practical implications beyond a single subfield
- Landmark
- Fundamental
- Important
- Valuable
- Useful
Solid: Methods, data and analyses broadly support the claims with only minor weaknesses
- Exceptional
- Compelling
- Convincing
- Solid
- Incomplete
- Inadequate
During the peer-review process the editor and reviewers write an eLife Assessment that summarises the significance of the findings reported in the article (on a scale ranging from landmark to useful) and the strength of the evidence (on a scale ranging from exceptional to inadequate). Learn more about eLife Assessments
Introduction
In social interactions utilizing vocal communication signals, the acoustic features of the signals carry emotional information (Altenmüller et al., 2013; Darwin, 1872; Seyfarth and Cheney, 2003). Listeners receive and analyze the acoustic information, compare it with previous experiences, identify the salience and valence of such information, and respond with appropriate behaviors. These integrated functions depend on brain circuits that include the amygdala, a region located within the temporal lobe that is recognized to play a role in orchestrating emotional responses to salient sensory stimuli (LeDoux, 2000; McGaugh, 2002; Sah et al., 2003). The amygdalar target of auditory inputs from the thalamus and cortex is the basolateral amygdala (BLA) (LeDoux et al., 1984; Romanski and LeDoux, 1993; Shi and Cassell, 1997; Tsukano et al., 2019). By integrating this auditory input with other sensory inputs and inputs from other limbic areas, BLA neurons shape appropriate behavioral responses (Beyeler et al., 2016; Gründemann et al., 2019; Namburi et al., 2015; Namburi et al., 2016) via projections to downstream targets such as the nucleus accumbens (Ambroggi et al., 2008; Stuber et al., 2011) and central nucleus of the amygdala (Ciocchi et al., 2010).
The BLA processes vocal and other acoustic information in a context-dependent manner (Gadziola et al., 2016; Grimsley et al., 2013; Matsumoto et al., 2016; Wenstrup et al., 2020; Wiethoff et al., 2009). Contextual information may arise from inputs associated with other sensory modalities (e.g., somatic sensation or olfaction) (Grimsley et al., 2013; Lanuza et al., 2004; McDonald, 1998), but in other cases the contextual information is associated with an animal’s internal state. Sources of internal state cues to BLA include brain circuits involving modulatory neurochemicals (i.e., neuromodulators), known to affect the processing of sensory signals, thus shaping attention, emotion, and goal-directed behaviors (Bargmann, 2012; Likhtik and Johansen, 2019; Schofield and Hurley, 2018). While previous work has demonstrated a role for some neuromodulators—dopamine (DA) and acetylcholine (ACh)—in the production of social vocalizations (Inagaki et al., 2020, Rojas-Carvajal et al., 2022; Silkstone and Brudzynski, 2020), it remains unclear how these and other neuromodulators contribute to vocal processing in social interactions.
This study investigates contextual information provided by neuromodulatory inputs to the BLA in response to vocal communication signals. Our hypothesis is that these salient vocalizations elicit distinct patterns of neuromodulator release into the BLA, by which they shape the processing of subsequent meaningful sensory information. We further hypothesize that these neuromodulatory patterns may depend on longer-term processes that are critical to vocal communication: to experience with these behaviors and the accompanying vocalizations, to sex, and to estrous stage in females. To test these hypotheses, we conducted playback experiments in a mouse model to understand the behavioral and neuromodulator responses to salient vocalizations associated with very different behavioral states.
Results
To study how vocalizations affect behaviors and release of neurochemicals within BLA, we first developed highly salient vocal stimuli associated with appetitive (mating) and aversive (restraint) behaviors of CBA/CaJ mice. From more intense, mating-related interactions between adult male and female mice that included female head-sniffing and attempted or actual mounting (Gaub et al., 2016; Ghasemahmad, 2020, see Materials and methods), we selected several sequences of vocalizations to form a 20-min mating vocal stimulus. These sequences included ultrasonic vocalizations (USVs) with harmonics, steps, and complex structure, mostly emitted by males, and low-frequency harmonic calls (LFHs) emitted by females (Figure 1A, C; Finton et al., 2017; Gaub et al., 2016; Ghasemahmad, 2020; Hanson and Hurley, 2012). During short periods of restraint, mice produce distinctive mid-frequency vocalizations (MFVs) that are associated with anxiety-related behaviors and increased release of the stress hormone corticosterone (Dornellas et al., 2021; Grimsley et al., 2016; Niemczura et al., 2020). From vocal sequences produced by restrained mice, we created a 20-min vocal stimulus, primarily containing MFVs and fewer USV and LFH syllables (Figure 1B, C).
We next asked whether these salient vocal stimuli, associated with very different behavioral states, could elicit distinct behaviors and patterns of neuromodulator release into the BLA. We focused on the neuromodulators ACh and DA, since previous work suggests that these neuromodulatory systems interact in the emission of positive and negative vocalizations (Rojas-Carvajal et al., 2022; Silkstone and Brudzynski, 2020). Our experiments combined playback of the vocal stimuli, behavioral tracking and observations (see Table 1 for descriptions), and microdialysis of BLA extracellular fluid in freely moving mice (Figure 1D).
Prior to the study, male and female mouse subjects had no experience with sexual or restraint behaviors. On the first 2 days of the experiment, mice in an experienced group (EXP, n = 31) were each exposed to 90-min sessions with mating and restraint behaviors in a counterbalanced design (Figure 1D). Mice were then implanted with a guide cannula for microdialysis. On the playback/sample collection day (Day 6), a microdialysis probe was inserted into the guide cannula. After a 4-hr period of mouse habituation and probe equilibration, we recorded behavioral reactions and sampled extracellular fluid from the BLA before (Pre-Stim) and during a 20-min playback period, divided into two 10-min stimulation/collection/observation periods that are designated Stim 1 and Stim 2 (Figure 1D, E; see also Materials and methods). Each mouse received playback of either the mating or restraint stimuli, but not both: same-day presentation of both stimuli would require excessively long playback sessions, the condition of the same probe would likely change on subsequent days, and quality of a second implanted probe on a subsequent day was uncertain. Fluids were analyzed using a liquid chromatography/mass spectrometry (LC/MS) technique that allowed simultaneous measurement of several neurochemicals and their metabolites in the same dialysate samples, including ACh, DA, and the serotonin metabolite 5-hydroxyindoleacetic acid (5-HIAA) (see Materials and methods). All neurochemical results during Stim 1 and Stim 2 periods are expressed as a percentage relative to the Pre-Stim control period. However, raw values of ACh, DA, and 5-HIAA are reported in source data files (Figure 3—source data 1, Figure 3—source data 2, Figure 3—source data 3). Data are reported only from mice with more than 75% of the microdialysis probe implanted within the BLA (Figure 2).
We first describe tests to examine whether playback of mating and restraint vocalizations results in different behavioral and neurochemical responses in male mice. We observed that two behaviors, still-and-alert and flinching, showed differential effects of playback type, increasing during restraint playback relative to mating playback during the Stim 1 period (Figure 3A, B). There were also distinct patterns of ACh and DA release into the BLA depending on the type of vocalization playback (Figure 3C, D). Thus, in response to restraint vocalizations, we observed an increase in ACh concentration during both Stim 1 (M ± SD: +24 ± 31% re Pre-Stim baseline) and Stim 2 (+25 ± 16%) playback windows. Mating vocalizations, however, resulted in a decrease in ACh release during both playback periods (Stim 1: −19 ± 21.0%; Stim 2: −16.4 ± 24.0%) (Figure 3C). DA release displayed opposite patterns to ACh, increasing during playback of mating vocalizations during Stim 1 (+12.0 ± 27.4%) and Stim 2 (+25 ± 39% re baseline), but decreasing during playback of restraint vocal sequences (Stim 1: −11.0 ± 11.2%; Stim 2: −22 ± 25.6%) (Figure 3D). For ACh, levels differed significantly between the mating and restraint groups during both Stim 1 and Stim 2, while DA levels differed significantly only during Stim 2 (Figure 3C, D). In contrast, the serotonin metabolite 5-HIAA showed no distinct pattern over time following playback of either vocal stimulus, nor significant differences between groups (Figure 3E). These findings suggest that both behavioral responses and ACh and DA release are modulated in listening male mice by the affective content of social vocalizations.
-
Figure 3—source data 1
- https://cdn.elifesciences.org/articles/88838/elife-88838-fig3-data1-v1.xlsx
-
Figure 3—source data 2
- https://cdn.elifesciences.org/articles/88838/elife-88838-fig3-data2-v1.xlsx
-
Figure 3—source data 3
- https://cdn.elifesciences.org/articles/88838/elife-88838-fig3-data3-v1.xlsx
-
Figure 3—source data 4
- https://cdn.elifesciences.org/articles/88838/elife-88838-fig3-data4-v1.xlsx
As male and female mice emit different vocalizations during courtship and mating (Finton et al., 2017; Grimsley et al., 2013; Neunuebel et al., 2015; Sales, 1972), we tested whether playback of vocal interactions associated with mating (Figure 1A) results in different behavioral and neurochemical responses in listening male and female mice. Since our testing included both estrus and non-estrus females, we further examined the estrous effect on neurochemical release and behavioral reactions.
Playback of mating vocalizations resulted in some general and some sex-based differences in behavioral responses. For instance, all groups displayed increased attending behavior (Figure 4A). In females, regardless of estrous stage, rearing decreased and still-and-alert behavior increased significantly relative to male mice (Figure 4B, C). This change in motor activity was further supported by video tracking results showing a significant decrease in distance traveled by female mice in response to mating vocal playback (significant sex effect (F(1,15) = 9.3, p = 0.008, ŋ2 = 0.4)). However, one behavioral change was estrous dependent: females in estrus displayed a strikingly higher number of flinching behaviors compared to males and non-estrus females during both Stim 1 and Stim 2 periods (Figure 4D). Our analysis of neuromodulator responses to mating vocalization playback revealed an estrous-dependent modulation of ACh levels during playback. ACh concentration in estrus females increased during both Stim 1 and Stim 2 periods, whereas ACh decreased in both males and non-estrus females (Figure 4E). Moreover, during playback, post hoc comparison with Bonferroni correction showed that the ACh in estrus females was significantly higher (Stim 1, +29.3 ± 27.0%; Stim 2, +25.5 ± 22%) than both males (Stim 1, −18 ± 21.0%; Stim 2, −16.0 ± 24.0%) and non-estrus females (Stim 1, −6.0 ± 14%; Stim 2, −35.0 ± 29.6%). DA release showed increases during mating playback within all three experimental groups (males: Stim 1, +12.0 ± 27.0%, Stim 2, +26.0 ± 39.1%; estrus females: Stim 1, +24.1 ± 37.0%, Stim 2, +23.0 ± 27.0%; non-estrus females: Stim 1, +27.0 ± 17.0%, Stim 2, +49.0 ± 48.0%), but no significant differences among groups (Figure 4F). Similar to male groups in restraint and mating playback, the 5-HIAA release patterns in females showed no clear modulation pattern during mating vocal playback or differences among groups (Figure 4G).
Like male mice exposed to restraint vocalizations, estrus females showed robust and significant increases in flinching behavior in response to mating vocalizations. Both groups displayed increased ACh release. This supports the possible involvement of ACh in shaping such behavior in both males listening to restraint calls and in estrus females listening to mating vocalizations. Note that neuromodulator release, including ACh, has been previously linked to motor behaviors (Wall and Woolley, 2020), but the changes in ACh that we observed showed no relationship with behaviors involving motor activity such as rearing (restraint males: Stim 1: n = 6, r = 0.1, p = 0.8; Est females: n = 6, r = 0.7, p = 0.08), locomotion (Stim 1: restraint males, n = 6, r = −0.2, p = 0.7; Est females: n = 6, r = 0.3, p = 0.6), or distance traveled (Stim 1: restraint males, n = 6, r = 0.16, p = 0.8; Est females: n = 6, r = −0.6, p = 0.2). This suggests that the observed changes in ACh reflect the valence of these vocalizations.
All EXP mice used in the above experiments had undergone a single session each to experience mating and restraint conditions prior to the playback session on Day 6 (Figure 1D). Does such experience shape the release patterns of these neuromodulators in response to vocal playback? We tested male and female mice under identical vocal playback conditions as previous groups, except that they did not receive the restraint and mating experiences (INEXP groups). Since only one INEXP female was in a non-estrus stage during the playback session, our analysis of the effect of experience included only estrus females and males.
Several behavioral responses to vocalization playback differed between EXP and INEXP mice in a sex- or context-dependent manner. For example, EXP estrus females showed significantly increased sound-evoked flinching behaviors compared to INEXP estrus females (Figure 5A) in response to mating vocal sequences. These experience effects were not observed in males in response to mating or restraint vocal playback. Furthermore, the differences in flinching behavior among EXP groups (male-mating vs male-restraint or estrus female-mating) were not evident among INEXP groups. For rearing behavior, EXP males responding to mating vocalizations showed an increase compared to INEXP males (Figure 5B). This pattern was not observed during restraint playback for males or mating playback in estrus females. These findings indicate that behavioral responses to salient vocalizations result from interactions between sex of the listener or context of vocal stimuli with the previous behavioral experience associated with these vocalizations.
A major finding is that prior experience with mating and restraint behaviors shaped patterns of ACh release in response to vocal playback; there was a significant interaction of context and experience for ACh (F(1,39) = 12.7, p < 0.001; Figures 5C and 6A). Thus, ACh release was significantly different between INEXP and EXP males in restraint playback groups. A similar pattern was observed in INEXP estrus females in response to mating vocal playback, that is, the concentration of ACh re baseline failed to show such pronounced increases as observed in EXP females and showed a significant experience effect (sex*experience interaction: F(1,39) = 7.8, p = 0.008; Figures 5C and 6A). Furthermore, INEXP males in mating playback group failed to show the reduced ACh concentration observed in EXP males (INEXP Stim 1: +9.6 ± 15.0%; INEXP Stim 2: +6.1 ± 30.0; EXP Stim 1: −19.0 ± 21.0%; EXP Stim 2: −16.0 ± 24.0%), even though these differences were not significant (Figures 5C and 6A).
DA release patterns, which were weaker in EXP groups compared to ACh patterns (Figure 3C vs D ; Figure 4E vs F) showed no significant experience-dependent changes in any group (Figure 5D). There was no evidence that vocal playback altered DA release in any of the three groups of INEXP mice (Figure 6B), unlike what was observed in EXP groups (Figures 3D and 4F). 5-HIAA concentrations, which were unaffected by sex, estrous stage, or playback type (Figures 3E and 4G), were also unaffected by experience (Figure 6C).
Collectively, these data suggest that the playback vocalization type and estrous effects observed in ACh release patterns and behavioral reactions depend on previous experience with the corresponding behaviors.
Discussion
Functional imaging studies in humans and mechanistic studies in other species provide substantial evidence that the amygdala participates in circuits that process vocalizations (Frühholz et al., 2016; Liebenthal et al., 2016; Sander et al., 2003; Wenstrup et al., 2020; Voytenko et al., 2023), assess the valence of appetitive and aversive cues (Kyriazi et al., 2018; O’Neill et al., 2018; Pignatelli and Beyeler, 2019; Smith and Torregrossa, 2021), and shape appropriate behavioral responses to these cues (Gründemann et al., 2019; Lim et al., 2009; Schönfeld et al., 2020; Zhang and Li, 2018). Nonetheless, how contextual information is delivered to the amygdala and contributes to vocal processing is not well understood. Since the amygdala receives strong projections from neuromodulatory brain centers (Aitta-Aho et al., 2018; Asan, 1997; Asan, 1998; Bigos et al., 2008; Bocchio et al., 2016; Carlsen et al., 1985), and since the role of these neurochemicals in providing internal state and contextual information is well documented (Bocchio et al., 2016; Jiang et al., 2016; Likhtik and Johansen, 2019), we hypothesized that release patterns of neuromodulators into the BLA provide contextual information during processing of affective vocalizations. Our results show that these emotionally charged vocalizations result in distinct release patterns of ACh and DA into the BLA of male and female mice. Furthermore, female hormonal state appears to influence ACh but not DA release into the BLA when processing mating vocalizations. Such context- or state-dependent changes were not observed in patterns of other neurochemicals (e.g., 5-HIAA), nor in the absence of experience with behaviors associated with the vocalizations. In particular, we showed that a single 90-min experience with intense behaviors is sufficient to establish strong, consistent patterns of ACh release into the amygdala. These data indicate that during analysis of affective vocalizations in the BLA, ACh, and DA provide experience-dependent, state- and context-related information that can potentially modulate sensory processing within the BLA and thus shape an individual’s response to these vocalizations.
The linkage between neuromodulator release and behavioral responses to vocalizations varies. The strongest case is for a link between increased ACh release and flinching behavior. Males in the restraint playback group and estrus females in the mating playback group both displayed significantly increased ACh release and flinching behavior, occurring in both playback periods. This suggests a mechanistic relationship. Other significant behavioral responses to vocal stimuli did not closely match the timing of neurochemical changes, suggesting a weak mechanistic link. Overall, it is not surprising that most behaviors and neurochemicals do not match well, given the poor temporal resolution of the neurochemical sampling. We believe that our results show the need for a finer analysis of DA and ACh release based on techniques for high temporal resolution of neurochemical measurements (e.g., genetically encoded ACh indicator; Jing et al., 2018), coupled with short interval behavioral observations. Only then can the role of the BLA and neuromodulator inputs in behavioral vocal responses be established more strongly. Furthermore, we note that no conclusion can be made regarding relationships between noradrenalin and serotonin concentrations and vocal playback, since these were not detected by the neurochemical analysis.
The BLA receives strong cholinergic projections from the basal forebrain (Aitta-Aho et al., 2018; Carlsen et al., 1985) that contribute to ACh-dependent processing of aversive cues and fear learning in the amygdala (Baysinger et al., 2012; Gorka et al., 2015; Jiang et al., 2016; Tingley et al., 2014). Our findings support these studies by demonstrating increased ACh release in BLA in EXP animals in response to playback of aversive vocalizations. Although the exact mechanism by which ACh affects vocal information processing in BLA is not clear yet, the result of ACh release onto BLA neurons seems to enhance arousal during emotional processing (Likhtik and Johansen, 2019). Our results, in conjunction with previous work, suggest mechanisms by which vocalizations affect ACh release and in turn drive behavioral responses (Figure 7A). Cholinergic modulation in the BLA is mediated via muscarinic and nicotinic ACh receptors on BLA pyramidal neurons and inhibitory interneurons (Aitta-Aho et al., 2018; Mesulam et al., 1983; Pidoplichko et al., 2013; Unal et al., 2015). During the processing of sensory information in the BLA, partially non-overlapping populations of neurons respond to cues related to positive or negative experiences (Namburi et al., 2015; Paton et al., 2006; Smith and Torregrossa, 2021). These neurons then project to different target areas involved in appetitive or aversive behaviors—the nucleus accumbens or central nucleus of the amygdala, respectively (Namburi et al., 2015).
In response to an aversive cue or experience (Figure 7A), released ACh affects BLA neurons according to their activity. If projection neurons are at rest, ACh may exert an inhibitory effect by activating nicotinic ACh receptors on local GABAergic interneurons, which in turn synapse onto the quiescent pyramidal neurons. This results in GABA-A-mediated inhibitory postsynaptic potentials in the pyramidal neurons. Alternately, direct activation of M1 ACh receptors on pyramidal neurons, activating inward rectifying K+ currents, may result in additional ACh-mediated inhibition (Figure 7A; Aitta-Aho et al., 2018; Pidoplichko et al., 2013; Unal et al., 2015). When BLA pyramidal neurons are already active due to strong excitatory input associated with aversive cues, M1 receptor activation can result in long afterdepolarizations that produce persistent firing lasting as long as ACh is present (Jiang et al., 2016; Unal et al., 2015). Such a process may explain persistent firing observed in single neuron responses to aversive social vocalizations in bats (Gadziola et al., 2012; Peterson and Wenstrup, 2012). Through this process of inhibiting quiescent neurons and enhancing activation and persistent firing in active neurons, ACh sharpens the population signal-to-noise ratio (SNR) during the processing of salient, aversive signals in the BLA. These neurons, processing negative cues, likely project to the central nucleus of the amygdala (CeA) to regulate defensive behaviors such as escape and avoidance (Figure 7A; Beyeler et al., 2016; Davis et al., 2010; Namburi et al., 2015). In agreement, our behavioral findings show increased behaviors such as flinching along with increased release of ACh during processing of aversive vocalizations. Such prolonged afterdepolarizations provide the appropriate condition for associative synaptic plasticity (Likhtik and Johansen, 2019) and underlie an increase in the ratio of alpha-amino-3-hydroxy-5-methyl-4-isoxazole propionic acid (AMPA) receptor current to N-methyl-D-aspartate (NMDA) receptor current among CeA-projecting neurons during processing of aversive cues.
Dopaminergic innervation from the ventral tegmental area (Asan, 1998) acts on BLA neurons via D1 and D2 receptors, both G-protein-coupled receptors. DA is important in reward processing, fear extinction, decision making, and motor control (Ambroggi et al., 2008; Di Ciano and Everitt, 2004; Lutas et al., 2019). We observed increased DA release in the BLA in EXP groups in response to mating vocalizations both for males and for females across estrous stages. Electrophysiological studies show that DA enhances sensory processing in BLA neurons by increasing the population response SNR in a process like ACh (Kröner et al., 2005; Vander Weele et al., 2018). Thus, during processing of mating vocalizations or those related to other rewarding experiences, DA presence in the vicinity of BLA pyramidal neurons and interneurons is enhanced (Figure 7B). For neurons with elevated spiking activity during processing of appetitive vocalizations or other sensory stimuli, DA acts on D2 receptors of pyramidal cells to further enhance neuronal firing and result in persistent firing of these projection neurons. Conversely, in BLA projection neurons that do not respond to such positive cues, such as CeA-projecting neurons, DA exerts a suppressive effect directly via D1 receptors and indirectly by activating inhibitory interneuron feedforward inhibition (Kröner et al., 2005). The net result of this process in response to appetitive vocalizations is an enhancement of activity in the reward-responding neurons and suppression of activity in aversive-responding neurons. This process likely depends on the increase in synaptic plasticity via an enhanced ratio of AMPA receptor current to NMDA receptor current during processing such cues in NAc-projecting neurons in the BLA (Namburi et al., 2015; Otani et al., 2003; van Vugt et al., 2020). Our findings suggest that this may occur in the BLA in response to appetitive vocalizations.
As the results with males listening to restraint vocalizations demonstrate, increased ACh release in BLA is associated with processing aversive cues. How, then, should the increased ACh release in estrus females during mating vocal playback be interpreted? Previous work shows that neuromodulation of amygdalar and other forebrain activity is altered by sex hormone/receptor changes in males and females (Egozi et al., 1986; Kalinowski et al., 2023; Kirry et al., 2019; Matsuda et al., 2002; Mizuno et al., 2022; van Huizen et al., 1994). For instance, the cholinergic neurons that project to the BLA, originating in the basal forebrain, exhibit high expression of estrogen receptors that is influenced by a female’s hormonal state (Shughrue et al., 2000). During estrus, the enhanced circulating estrogen affects release of ACh and may influence neuronal networks and behavioral phenotypes in a distinct manner (Gibbs, 1996; McEwen, 1998). Thus, increased ACh release in estrus females may underlie increased attentional and risk-assessment behaviors in response to vocalization playback, as we observed with flinching behavior in this experimental group. Combined with DA increase, it may trigger both NAc and CeA circuit activation, resulting in both reward-seeking and cautionary behaviors in estrus females.
Our results demonstrate the strong impact of even limited experience in shaping behavioral and neuromodulatory responses associated with salient social vocalizations. In the adult mice, a single 90-min session of mating and of restraint, occurring 4–5 days prior to the playback experiment, resulted in consistent behavioral responses and consistent and enhanced ACh release into BLA for both vocalization types. As previous work shows (Huang et al., 2012; Nadim and Bucher, 2014; Pawlak et al., 2010), neuromodulatory inputs play crucial roles in regulating experience-dependent changes in the brain. However, it remains unclear whether the experience shapes neuromodulator release, or whether neuromodulators deliver the experience-related effect into the BLA. It is also unknown whether these experiences were specific to the type of vocalization playback associated with the experience, or whether either experience had a more generalized effect on responses to all vocalization playback, as suggested by the broad physiological effects of mating or restraint (Leuner et al., 2010; Arnold et al., 2019).
The interaction between ACh and DA is thought to shape motor responses to external stimuli (Lester et al., 2010). Our results support the view that a balance of DA and ACh may regulate the proper behavioral response to appetitive and aversive auditory cues. For instance, increased reward-seeking behavior (rearing and locomotion) in EXP males during mating vocalizations playback may result from the differential release of the two neuromodulators—decreased ACh and increased DA. Furthermore, the lack of this differential release may be the underlying cause for the lack of such responses in INEXP male mice. This supports the role of experience in tuning interactions of these two neuromodulators throughout the BLA, for shaping appropriate behaviors. Overall, the behavioral changes orchestrated by the BLA in response to emotionally salient stimuli are most likely the result of the interaction between previous emotional experiences, hormonal state, content of sensory stimuli, and sex of the listening animals.
Materials and methods
Animals
Experimental procedures were approved by the Institutional Animal Care and Use Committee at Northeast Ohio Medical University (protocol 18-09-207). A total of 83 adult CBA/CaJ mice (Jackson Labs, p90-p180), male and female, were used for this study. Animals were maintained on a reversed dark/light cycle and experiments were performed during the dark cycle. The mice were housed in same-sex groups until the week of the experiments, during which they were singly housed. Food and water were provided ad libitum except during the experiment.
The estrous stage of female mice was evaluated based on vaginal smear samples obtained by sterile vaginal lavage. Samples were collected using glass pipettes filled with double distilled water, placed on a slide, stained using crystal violet, and coverslipped for microscopic examination. Estrous stage was determined by the predominant cell type: squamous epithelial cells (estrus), nucleated cornified cells (proestrus), or leukocytes (diestrus) (McLean et al., 2012). To confirm that the stage of estrous did not change during the experiment day, samples obtained prior to and after data collection on the experimental day were compared.
Experimental overview
Request a detailed protocolThe basic experimental structure that occurred over 6 days is illustrated in Figure 1D. Briefly, each mouse was placed in an arena on Days 1 and 2 to provide one-time experiences of mating and sustained restraint. After the experience on Day 2, the subject was anesthetized for implantation of a guide tube for the microdialysis probe. On Day 6, the vocalization playback session occurred. The mouse was briefly anesthetized for insertion of the microdialysis probe into the amygdala, followed by several hours recovery. Before, during, and after vocal playback, we sampled extracellular fluid from the amygdala through the microdialysis probe and recorded video to analyze the subject’s behavior. An LC/MS method was used to measure concentrations of several neurochemicals.
The microdialysis sampling interval of 10 min was used to establish the temporal framework for vocal playback and analysis of neurochemicals and behaviors (Figure 1E). The analysis compared neurochemicals and behaviors in a period immediately preceding the vocal stimuli (Pre-Stim) and in two 10-min periods featuring playback of the same sequence of vocal stimuli (Stim 1 and Stim 2). Each playback session featured only one category of vocalizations (associated with mating or restraint) and each animal participated in a single playback session. Additional details of this experimental structure are described in the sections below.
Acoustic methods
Vocalization recording and analysis
Request a detailed protocolTo record vocalizations for use in playback experiments, mice were placed in an open-topped plexiglass chamber (width, 28 cm; length, 28 cm; height, 20 cm), housed within a darkened, single-walled acoustic chamber (Industrial Acoustics, New York, NY) lined with anechoic foam (Grimsley et al., 2016; Niemczura et al., 2020). Acoustic signals were recorded using ultrasonic condenser microphones (CM16/CMPA, Avisoft Bioacoustics, Berlin, Germany) connected to a multichannel amplifier and A/D converter (UltraSoundGate 416H, Avisoft Bioacoustics). The gain of each microphone was independently adjusted once per recording session to optimize the SNR while avoiding signal clipping. Acoustic signals were digitized at 500 kHz and 16-bit depth, monitored in real time with RECORDER software (Version 5.1, Avisoft Bioacoustics), and Fast Fourier Transformed (FFT) at a resolution of 512 Hz. A night vision camera (VideoSecu Infrared CCTV), centered 50 cm above the floor of the test box, recorded the behaviors synchronized with the vocal recordings (VideoBench software, DataWave Technologies, version 7).
To record mating vocalizations, 10 animals (5 male–female pairs) were used in sessions that lasted for 30 min. A male mouse was introduced first into the test box, followed by a female mouse 5 min later. Vocalizations were recorded using two ultrasonic microphones placed 30 cm above the floor of the recording box and 13 cm apart. See below for analysis of behaviors during vocal recordings.
To record vocalizations during restraint, six mice (four male, two female) were briefly anesthetized with isoflurane and then placed in a restraint jacket as described previously (Grimsley et al., 2016). Vocalizations were recorded for 30 min while the animal was suspended in the recording box. Since these vocalizations are usually emitted at lower intensity compared to mating vocalizations, the recording microphone was positioned 2–3 cm from the snout to obtain the best SNR.
Vocal recordings were analyzed offline using Avisoft-SASLab Pro (version 5.2.12, Avisoft Bioacoustics) with a hamming window, 1024 Hz FFT size, and an overlap percentage of 98.43. For every syllable, the channel with the higher amplitude signal was extracted using a custom-written Python code (https://github.com/GavazziDA/Wenstrup_Lab_Ghasemahmad_2023, copy archived at Gavazzi, 2024) and analyzed. Since automatic syllable tagging did not allow distinguishing some syllable types such as noisy calls and MFVs from background noise, we manually tagged the start and end of each syllable, then examined spectrograms to measure several acoustic features and classify syllable types based on Grimsley et al., 2011; Grimsley et al., 2016.
Vocalization playback
Request a detailed protocolVocalization playback lasting 20 min was constructed from a set of seven repeating stimulus blocks lasting 2:50 min each (Figure 1E). Each block was composed of a set of vocal sequence exemplars that alternated with an equal duration of background sound (Silence) associated with the preceding exemplar. The exemplars were recorded during mating interactions and restraint, and selected based on high SNR, correspondence with behavioral category by video analysis, and representation of the spectrotemporal features of vocalizations emitted during mating and restraint (Ghasemahmad, 2020; Grimsley et al., 2016). Mating stimulus blocks contained five exemplars of vocal sequences emitted during mating interactions. These exemplars ranged in duration from 15.0 to 43.6 s. Restraint stimulus blocks included seven vocal sequences, emitted by restrained male or female mice, with durations ranging from 5.7 to 42.3 s. Across exemplars, each stimulus block associated with both mating and restraint included different sets of vocal categories (Figure 1A–C).
Playback sequences, that is, exemplars, were conditioned in Adobe Audition CC (2018), adjusted to a 65-dB SNR level, then normalized to 1 V peak-to-peak for the highest amplitude syllable in the sequence. This maintained relative syllable emission amplitude in the sequence. For each sequence, an equal duration of background noise (i.e., no vocal or other detected sounds) from the same recording was added at the end of that sequence (Figure 1E). A 5-ms ramp was added at the beginning and the end of the entire sequence to avoid acoustic artifacts. A MATLAB app (EqualizIR, Sharad Shanbhag; https://github.com/TytoLogy/EqualizIR, Shanbhag, 2018) compensated and calibrated each vocal sequence for the frequency response of the speaker system. Vocal sequences were converted to analog signals at 500 kHz and 16-bit resolution using DataWave (DataWave SciWorks, Loveland, CO), anti-alias filtered (TDT FT6-2, fc = 125 kHz), amplified (HCA-800II, Parasound, San Francisco, CA), and sent to the speaker (LCY, K100, Ying Tai Audio Company, Hong Kong). Each sequence was presented at peak level equivalent to 85 dB SPL.
Behavioral methods
Behaviors during both vocalization recording and playback sessions were recorded using a night vision camera (480TVL 3.6 mm, VideoSecu), centered 50 cm above the floor of the test box, and SciWorks (DataWave, VideoBench version 7) for video acquisition and analysis.
Analysis of mating behaviors during vocal recordings
Request a detailed protocolInteractions between male and female mice were video-recorded, then analyzed second-by-second. Among more general courtship interactions, we identified a set of behaviors that are mating-related, as described previously (Gaub et al., 2016; Heckman et al., 2016). All vocal sequences selected as exemplars for playback of ‘mating’ vocalizations were recorded in association with these male-mating behaviors: head-sniffing, attempted mounting, or mounting. Vocalizations during these behaviors included chevron, stepped, and complex USVs emitted with longer durations and higher repetition rates, and more LFH calls (Gaub et al., 2016; Ghasemahmad, 2020; Hanson and Hurley, 2012).
Experience and playback sessions
Request a detailed protocolPrior to playback experiments, each animal underwent 90-min sessions on two consecutive days (Days 1 and 2) that provided both mating and restraint experiences to the EXP group (n = 31 animals) or no experiences of these to the INEXP group (n = 22 animals). EXP sessions were presented in a counterbalanced pattern across subjects (Figure 1D). For the mating experience, mounting or attempted mounting was required for the animal to be included in the remainder of the experiment. However, we did not record detailed behaviors or track estrous stage during the mating experience session. After the Day 2 session, mice underwent surgery for implantation of a microdialysis guide canula (see below), then recovered for 4 days.
On Day 6, the day of the playback experiment, male mice were randomly assigned to either restraint or mating vocal playback groups. Females were only tested with mating playback, since a preliminary behavioral study suggested no difference in male and female responses to restraint. Video recording was performed simultaneously with microdialysis, beginning 10 min before vocal playback (pre-stimulus) and continuing for 20 min during playback (Figure 1D).
Analysis of behaviors during vocal playback
Request a detailed protocolBehavioral analysis was based on previous descriptions (Bakshi and Kelley, 1993; Bakshi and Kelley, 1994; Blanchard et al., 2003; Füzesi et al., 2016; Grimsley et al., 2015; Lezak et al., 2017; Saibaba et al., 1996) (Mouse Ethogram: https://mousebehavior.org). The list of behaviors assessed was determined through pilot analysis and previous studies characterizing rodent defense and fear behaviors. They are defined as the following: abrupt attending, flinching, locomotion, rearing, grooming, still-and-alert, and stretch-attend (description in Table 1; Blanchard et al., 2003). A behavior was only counted when its occurrence lasted two or more seconds, except for flinching, which could take place more quickly.
For analysis of behaviors during vocal playback, all videos were examined blind to the sex, estrous state, or experience of the animal and the context of vocalizations. Video recordings were analyzed manually (n = 48 mice) in 10-s intervals for the seven behaviors identified above. The number of occurrences of every behavior per 10-s block was marked. The numbers were added for every block of 2:50 min of vocal playback (see Figure 1E), then averaged for the Pre-Stim, Stim 1, and Stim 2 periods separately.
Videos were further analyzed automatically using the video tracker within VideoBench (DataWave Technologies, version 7) for total distance traveled and time spent in the periphery and center of the test arena. For video-tracking analysis of time spent, the floor was divided into 16 equal squares with 4 central squares identified as ‘center’ and the 4 corner squares identified as periphery. The two squares in the middle of each wall were excluded from the analysis to avoid attraction to the holes affect the animal’s preference for the periphery vs center.
Procedures related to microdialysis
Surgery
Request a detailed protocolSince both left and right amygdala are responsive to vocal stimuli in human and experimental animal studies (Wenstrup et al., 2020), we implanted microdialysis probes into the left amygdala to maintain consistency with other studies in our laboratory. Mice were anesthetized with Isoflurane (2–4%, Abbott Laboratories, North Chicago, IL) and hair overlying the skull was removed using depilatory lotion. A midline incision was made, then the skin was moved laterally to expose the skull. A craniotomy (≃1 mm2) was made above the BLA on the left side (stereotaxic coordinates from bregma: −1.65 mm rostrocaudal, +3.43 mm mediolateral). A guide cannula (CMA-7, CMA Microdialysis, Sweden) was implanted to a depth of 2.6 mm below the cortical surface and above the left BLA (Figure 1D, Day 2), then secured using dental cement. After surgery, the animal received a subcutaneous injection of carprofen (4 mg/kg, s.c.) and topical anesthetic (lidocaine) and antibiotic cream (Neosporin). It was returned to its cage and placed on a heating pad until fully recovered from anesthesia. Recovery from the surgical procedures occurred in the animal’s home cage and was assessed by resumption of normal eating, drinking, exploratory, and grooming behaviors and postures. Playback experiments occurred in all recovered animals 4 days after surgery.
Microdialysis
Request a detailed protocolOn the day before microdialysis, the probe was conditioned in 70% methanol and artificial cerebrospinal fluid (aCSF) (CMA Microdialysis, Sweden). On playback day (Day 6), the animal was briefly anesthetized and the probe with 1-mm membrane length and 0.24-mm outer diameter (MWCO 6 kDa) was inserted into the guide cannula (Figure 1D).
Using a spiral tubing connector (0.1 mm ID × 50 cm length) (CT-20, AMUZA Microdialysis, Japan), the inlet and outlet tubing of the probe was connected to the inlet/outlet Teflon tubing of the microdialysis lines. A swivel device for fluids (TCS-2-23, AMUZA Microdialysis, Japan), secured to a balance arm, held the tubing, and facilitated the animal’s free movement during the experiment.
After probe insertion, a 4-hr period allowed animals to habituate and the neurochemicals to equilibrate between aCSF fluid in the probe and brain extracellular fluid. Sample collection then began, in 10-min intervals, beginning with four background samples, then two samples during playback of restraint or mating vocal sequences (total 20 min), then one or more samples after playback ended. To account for the dead volume of the outlet tubing, a flow rate of 1.069 µl/min was established at the syringe pump to obtain a 1-µl sample per minute. Samples collected during this time were measured for volume to assure a consistent flow rate. To prevent degradation of collected neurochemicals, the outlet tubing was passed through ice to a site outside the sound-proof booth where samples were collected on ice, then stored in a −80°C freezer. The choice of 10-min sampling intervals was based on the minimum collection time needed to provide 5 µl samples both for immediate testing and for a backup sample in the event of catastrophic failure of the processes associated with the neurochemical analysis.
Neurochemical analysis
Request a detailed protocolSamples were analyzed using an LC/MS technique (Figure 1D) at the Vanderbilt University Neurochemistry Core. This analysis was blind to the sex, estrous state, or experience of the animal and the context of vocalizations. This method allows simultaneous measurement of several neurochemicals in the same dialysate sample: ACh, DA, and its metabolites (3,4-dihydroxyphenylacetic acid (DOPAC) and homovanillic acid (HVA)), serotonin (5-HT) and its metabolite (5-HIAA), norepinephrine (NE), gamma aminobutyric acid (GABA), and glutamate. Due to low recovery of NE and 5-HT from the mouse brain, we were unable to track these two neuromodulators in this experiment.
Before each LC/MS analysis, 5 μl of the sample was derivatized using sodium carbonate, benzoyl chloride in acetonitrile, and internal standard (Wong et al., 2016). LC was performed on a 2.0 × 50 mm, 1.7 μM particle Acquity BEH C18 column (Waters Corporation, Milford, MA, USA) with 1.5% aqueous formic acid as mobile phase A and acetonitrile as mobile phase B. Using a Waters Acquity Classic UPLC, samples were separated by a gradient of 98–5% of mobile phase A over 11 min at a flow rate of 0.6 ml/min with delivery to a SCIEX 6500+Qtrap mass spectrometer (AB Sciex LLC, Framingham, MA, USA). The mass spectrometer was operated using electrospray ionization in the positive ion mode. The capillary voltage and temperature were 4 kV and 350°C, respectively (Wong et al., 2016; Yohn et al., 2020). Chromatograms were analyzed using MultiQuant 3.0.2 Software (AB SCIEX, Concord, Ontario, Canada).
Verification of probe location
Request a detailed protocolWe verified placement of microdialysis probes to minimize variability that could arise due to placement in regions surrounding BLA that receive different sources of neurochemical inputs (e.g., cholinergic inputs to putamen and central amygdala). To verify probe placement within the BLA, each probe was perfused with 2% dextran-fluorescein (MW 4 kDa) (Sigma-Aldrich Inc, Atlanta, GA) at the end of the experiment. The location of the probe was then visualized in adjacent cleared and Nissl-stained sections (e.g., Figure 2, insets). Sections were photographed using a SPOT RT3 camera and SPOT Advanced Plus imaging software (version 4.7) mounted on a Zeiss Axio Imager M2 fluorescence microscope. Adobe Photoshop CS3 was used to adjust brightness and contrast globally. Animals were included in statistical analyses only if ≥75% of the probe membrane was located within the BLA (Figure 2).
Data analysis
Request a detailed protocolFor neurochemical analyses, the total numbers of animals used were NEXP = 31 and NINEXP = 22. Only mice with usable neurochemical data were further evaluated for behaviors. This included 46 mice: (NEXP = 25 and NINEXP = 21). Note that some animals used in neurochemical analyses were not included in the behavioral analyses because their behavioral data were unavailable. Since only one INEXP female was in a non-estrus stage during the playback session, our analysis of the effect of experience included only estrus females and males. Furthermore, one INEXP female without detectable DA levels was removed from the INEXP group neurochemical analysis.
Only the following 10-min sample collection windows were used for statistical analysis: Pre-Stim, Stim 1, and Stim 2 (Figure 1E). The purpose of the two collection windows during vocal playback was to allow detection of different epochs of neurochemical release, as may occur in ACh release into the amygdala (Aitta-Aho et al., 2018). All neurochemical data were normalized to the background level, obtained from a single pre-stimulus sample immediately preceding playback. This provided clarity in representations and did not result in different outcomes of statistical tests compared to use of three pre-stimulus samples. The fluctuation in all background samples was ≤20%. The percentage change from background level was calculated based on the formula: % change from background = (100 × stimulus sample concentration in pg)/background sample concentration in pg. Values are represented as mean ± one standard error unless stated otherwise. Raw neurochemical values are reported in Figure 3—source data 1–3 for ACh,DA, and 5-HIAA, respectively.
All statistical analyses were performed using SPSS (IBM, V. 26 and 27). To examine vocalizations in different stages of mating, we used a linear mixed model to analyze the changes of interval and duration (dependent variables) of vocalizations based on mating interaction intensity (fixed effect). For behavioral and neurochemical analyses in playback experiments, we initially compared the output using a linear mixed model and a generalized linear model (GLM) with a repeated measure for neurochemical data. Since both statistical methods resulted in similar findings, we chose to use the GLM for statistical comparisons of the neurochemicals and behaviors. Where Mauchly’s test indicated that the assumption of sphericity had been violated, degrees of freedom were corrected using Greenhouse–Geisser estimates of sphericity.
To further clarify the differences observed between groups for every comparison in the GLM repeated measure, multivariate contrast was performed. All multiple comparisons were corrected using Bonferroni post hoc testing. 95% confidence intervals were used to compare values between timepoints (Julious, 2004).
For both microdialysis and behavioral data, we tested the hypotheses that the release of neurochemicals into the BLA and the number of behaviors is differently modulated by the vocal playback type (mating and restraint) in male mice, by estrous stage of females or sex of the animals during mating vocal playback, and by previous experience in any of these groups. In the GLM model for Figure 3, context of vocalizations was defined as the fixed factor and time (Stim 1 and Stim 2) as the within-subject factor. The model in Figure 4 tested the effect of sex and estrous state as a nested variable within sex as fixed factors, which were considered in the model design (design = intercept + sex + estrous (sex)). The GLM model in Figure 5 tested sex, context, and experience as fixed factors with the interaction terms of sex*experience and context*experience. Here, we avoided using a full factorial model due to the absence of females in the restraint group, as that would have made the model unbalanced (design = intercept + sex + context + experience + sex*experience + context*experience). Furthermore, estrous state was not tested in examining the experience effect due to lack of INEXP non-estrus females. Dependent (response) variables included the normalized concentration of ACh, DA, 5-HIAA in Stim 1 and Stim 2 or the numbers of behaviors in Pre-Stim, Stim 1, and Stim 2 as previously described in the behavioral analysis section.
Throughout the figures, the distributions of data points are visualized by box plots, computed using the ‘inclusive median’ formula in Excel. The box indicates first (Q1) and third (Q3) quartiles with the median line between the two. The whiskers normally stretch between minimum and maximum values, but this computation sometimes shows data points outside of the whiskers. Excel identifies these values as outliers in the boxplot function. However, our GLM statistical analysis includes these values because there was no compelling reason to exclude them. The mean value is marked with an ‘x’.
Data availability
Numerical data used to generate the figures are contained in source data files provided for Figures 1 and 3. Software tools used in preparing sounds or analyzing sounds/behaviors are identified in the Materials and methods and available on provided links to GitHub.
References
-
BookEvolution of Emotional CommunicationOxford University Press.https://doi.org/10.1093/acprof:oso/9780199583560.001.0001
-
Sexual experience confers resilience to restraint stress in female ratsHormones and Behavior 107:61–66.https://doi.org/10.1016/j.yhbeh.2018.12.003
-
The catecholaminergic innervation of the rat amygdalaAdvances in Anatomy, Embryology, and Cell Biology 142:1–118.https://doi.org/10.1007/978-3-642-72085-7
-
Feeding induced by opioid stimulation of the ventral striatum: role of opiate receptor subtypesThe Journal of Pharmacology and Experimental Therapeutics 265:1253–1260.
-
Acute 5-HT reuptake blockade potentiates human amygdala reactivityNeuropsychopharmacology 33:3221–3225.https://doi.org/10.1038/npp.2008.52
-
The Mouse Defense Test Battery: pharmacological and behavioral assays for anxiety and panicEuropean Journal of Pharmacology 463:97–116.https://doi.org/10.1016/s0014-2999(03)01276-7
-
Serotonin, amygdala and fear: Assembling the puzzleFrontiers in Neural Circuits 10:24.https://doi.org/10.3389/fncir.2016.00024
-
Cholinergic projections from the basal forebrain to the basolateral amygdaloid complex: A combined retrograde fluorescent and immunohistochemical studyThe Journal of Comparative Neurology 234:155–167.https://doi.org/10.1002/cne.902340203
-
BookThe Expression of the Emotions in Man and AnimalsAmerican Psychological Association.https://doi.org/10.1037/10001-000
-
Phasic vs sustained fear in rats and humans: role of the extended amygdala in fear vs anxietyNeuropsychopharmacology 35:105–135.https://doi.org/10.1038/npp.2009.109
-
The sound of emotions-Towards a unifying neural network perspective of affective sound processingNeuroscience and Biobehavioral Reviews 68:96–110.https://doi.org/10.1016/j.neubiorev.2016.05.002
-
Hypothalamic CRH neurons orchestrate complex behaviours after stressNature Communications 7:11937.https://doi.org/10.1038/ncomms11937
-
A novel coding mechanism for social vocalizations in the lateral amygdalaJournal of Neurophysiology 107:1047–1057.https://doi.org/10.1152/jn.00422.2011
-
Two distinct representations of social vocalizations in the basolateral amygdalaJournal of Neurophysiology 115:868–886.https://doi.org/10.1152/jn.00953.2015
-
Ultrasonic vocalizations of adult male Foxp2-mutant mice: behavioral contexts of arousal and emotionGenes, Brain, and Behavior 15:243–259.https://doi.org/10.1111/gbb.12274
-
SoftwareWenstrup_Lab_Ghasemahmad_2023, version swh:1:rev:38079544a46716ca6009ac67609be089dfe13bd0Software Heritage.
-
Basal forebrain moderates the magnitude of task-dependent amygdala functional connectivitySocial Cognitive and Affective Neuroscience 10:501–507.https://doi.org/10.1093/scan/nsu080
-
Coding the meaning of sounds: contextual modulation of auditory responses in the basolateral amygdalaThe Journal of Neuroscience 33:17538–17548.https://doi.org/10.1523/JNEUROSCI.2205-13.2013
-
An improved approach to separating startle data from noiseJournal of Neuroscience Methods 253:206–217.https://doi.org/10.1016/j.jneumeth.2015.07.001
-
Contextual modulation of vocal behavior in mouse: Newly identified 12 khz “mid-frequency” vocalization emitted during restraintFrontiers in Behavioral Neuroscience 10:38.https://doi.org/10.3389/fnbeh.2016.00038
-
Determinants of the mouse ultrasonic vocal structure and repertoireNeuroscience and Biobehavioral Reviews 65:313–325.https://doi.org/10.1016/j.neubiorev.2016.03.029
-
Molecular characterization of frog vocal neurons using constellation pharmacologyJournal of Neurophysiology 123:2297–2310.https://doi.org/10.1152/jn.00105.2020
-
A genetically encoded fluorescent acetylcholine indicator for in vitro and in vivo studiesNature Biotechnology 36:726–737.https://doi.org/10.1038/nbt.4184
-
Using confidence intervals around individual means to assess statistical significance between two meansPharmaceutical Statistics 3:217–222.https://doi.org/10.1002/pst.126
-
Estrous cycle stage gates sex differences in prefrontal muscarinic control of fear memory formationNeurobiology of Learning and Memory 161:26–36.https://doi.org/10.1016/j.nlm.2019.03.001
-
Dopamine modulates excitability of basolateral amygdala neurons in vitroJournal of Neurophysiology 93:1598–1610.https://doi.org/10.1152/jn.00843.2004
-
Emotion circuits in the brainAnnual Review of Neuroscience 23:155–184.https://doi.org/10.1146/annurev.neuro.23.1.155
-
Acetylcholine-dopamine interactions in the pathophysiology and treatment of CNS disordersCNS Neuroscience & Therapeutics 16:137–162.https://doi.org/10.1111/j.1755-5949.2010.00142.x
-
Behavioral methods to study anxiety in rodentsDialogues in Clinical Neuroscience 19:181–191.https://doi.org/10.31887/DCNS.2017.19.2/wcarlezon
-
Neuromodulation in circuits of aversive emotional learningNature Neuroscience 22:1586–1597.https://doi.org/10.1038/s41593-019-0503-3
-
State-specific gating of salient cues by midbrain dopaminergic input to basal amygdalaNature Neuroscience 22:1820–1833.https://doi.org/10.1038/s41593-019-0506-0
-
Cortical pathways to the mammalian amygdalaProgress in Neurobiology 55:257–332.https://doi.org/10.1016/s0301-0082(98)00003-3
-
Multiple ovarian hormone effects on brain structure and functionThe Journal of Gender-Specific Medicine 1:33–41.
-
Memory consolidation and the amygdala: a systems perspectiveTrends in Neurosciences 25:456–461.https://doi.org/10.1016/s0166-2236(02)02211-7
-
Performing vaginal lavage, crystal violet staining, and vaginal cytological evaluation for mouse estrous cycle staging identificationJournal of Visualized Experiments 67:e4389.https://doi.org/10.3791/4389
-
The role of the cannabinoid system in fear memory and extinction in male and female micePsychoneuroendocrinology 138:105688.https://doi.org/10.1016/j.psyneuen.2022.105688
-
Neuromodulation of neurons and synapsesCurrent Opinion in Neurobiology 29:48–56.https://doi.org/10.1016/j.conb.2014.05.003
-
Architectural representation of valence in the limbic systemNeuropsychopharmacology 41:1697–1715.https://doi.org/10.1038/npp.2015.358
-
Physiological and behavioral responses to vocalization playback in miceFrontiers in Behavioral Neuroscience 14:155.https://doi.org/10.3389/fnbeh.2020.00155
-
Basolateral amygdala circuitry in positive and negative valenceCurrent Opinion in Neurobiology 49:175–183.https://doi.org/10.1016/j.conb.2018.02.012
-
Timing is not everything: Neuromodulation opens the STDP gateFrontiers in Synaptic Neuroscience 2:146.https://doi.org/10.3389/fnsyn.2010.00146
-
Valence coding in amygdala circuitsCurrent Opinion in Behavioral Sciences 26:97–106.https://doi.org/10.1016/j.cobeha.2018.10.010
-
The amygdaloid complex: anatomy and physiologyPhysiological Reviews 83:803–834.https://doi.org/10.1152/physrev.00002.2003
-
The human amygdala: an evolved system for relevance detectionReviews in the Neurosciences 14:303–316.https://doi.org/10.1515/revneuro.2003.14.4.303
-
BookCircuits for modulation of auditory functionIn: Oliver D, Cant N, Fay R, Popper A, editors. The Mammalian Auditory Pathways. Springer Handbook of Auditory Research. Springer. pp. 235–267.https://doi.org/10.1007/978-3-319-71798-2_9
-
Lesions of the rat basolateral amygdala reduce the behavioral response to ultrasonic vocalizationsBehavioural Brain Research 378:112274.https://doi.org/10.1016/j.bbr.2019.112274
-
Meaning and emotion in animal vocalizationsAnnals of the New York Academy of Sciences 1000:32–55.https://doi.org/10.1196/annals.1280.004
-
Cortical, thalamic, and amygdaloid projections of rat temporal cortexThe Journal of Comparative Neurology 382:153–175.https://doi.org/10.1002/(SICI)1096-9861(19970602)382:23.0.CO;2-2
-
Dissimilar interaction between dopaminergic and cholinergic systems in the initiation of emission of 50-kHz and 22-kHz vocalizationsPharmacology, Biochemistry, and Behavior 188:172815.https://doi.org/10.1016/j.pbb.2019.172815
-
Valence encoding in the amygdala influences motivated behaviorBehavioural Brain Research 411:113370.https://doi.org/10.1016/j.bbr.2021.113370
-
Task-phase-specific dynamics of basal forebrain neuronal ensemblesFrontiers in Systems Neuroscience 8:174.https://doi.org/10.3389/fnsys.2014.00174
-
Impact of basal forebrain cholinergic inputs on basolateral amygdala neuronsThe Journal of Neuroscience 35:853–863.https://doi.org/10.1523/JNEUROSCI.2706-14.2015
-
Muscarinic receptor characteristics and regulation in rat cerebral cortex: changes during development, aging and the oestrous cycleThe European Journal of Neuroscience 6:237–243.https://doi.org/10.1111/j.1460-9568.1994.tb00266.x
-
Intracellular recordings reveal integrative function of the basolateral amygdala in acoustic communicationJournal of Neurophysiology 129:1334–1343.https://doi.org/10.1152/jn.00103.2023
-
BookThe amygdala – A Hub of the social auditory brainIn: Wenstrup JJ, editors. The Senses: A Comprehensive Reference. Elsevier. pp. 812–837.https://doi.org/10.1016/b978-0-12-809324-5.24194-1
-
Population coding of valence in the basolateral amygdalaNature Communications 9:5195.https://doi.org/10.1038/s41467-018-07679-9
Article and author information
Author details
Funding
National Institute on Deafness and Other Communication Disorders (R01DC00937)
- Jeffrey J Wenstrup
Kent State University (Graduate Student Senate Research Award)
- Zahra Ghasemahmad
Northeast Ohio Medical University (BMS Dissertation Completion Award)
- Zahra Ghasemahmad
The funders had no role in study design, data collection, and interpretation, or the decision to submit the work for publication.
Acknowledgements
We are indebted to Sharad Shanbhag and Daniel Gavazzi for software used to condition vocal sequences and analyze audio and behavioral data, respectively, to Kristin Yeager of the Kent State University Statistical Consulting Center for statistical consulting, and to Sheila Fleming for advice on behavioral assessments. We thank Anthony Zampino, Austin Poth, Debin Lei, and Krish Nair for technical assistance. We further thank Vanderbilt University Neurochemistry Core, supported by Vanderbilt Brain Institute and the Vanderbilt Kennedy Center, that performed the neurotransmitter sample analyses. We are grateful to Drs. Alexander Galazyuk, Brett Schofield, Merri Rosen, Mahtab Tehrani, and Sharad Shanbhag for their comments on previous versions of this manuscript. National Institutes of Health grant R01DC00937 (JJW, Alexander Galazyuk). Kent State University Graduate Student Senate (ZG). Northeast Ohio Medical University Biomedical Sciences Program Committee (ZG).
Ethics
This study was performed in accordance with recommendations in the Guide for the Care and Use of Laboratory Animals of the National Institutes of Health. All procedures utilizing animals conformed to those approved by the Institutional Animal Care and Use Committee of the Northeast Ohio Medical University (#15-033 and #18-09-207). All surgery was performed using isoflurane anesthetic. Our procedures and practice sought to minimize animal suffering.
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
- Reviewed Preprint version 2:
- Reviewed Preprint version 3:
- Version of Record published:
Cite all versions
You can cite all versions using the DOI https://doi.org/10.7554/eLife.88838. This DOI represents all versions, and will always resolve to the latest one.
Copyright
© 2023, Ghasemahmad et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 878
- views
-
- 52
- downloads
-
- 2
- citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Computational and Systems Biology
- Neuroscience
Diffusional kurtosis imaging (DKI) is a methodology for measuring the extent of non-Gaussian diffusion in biological tissue, which has shown great promise in clinical diagnosis, treatment planning, and monitoring of many neurological diseases and disorders. However, robust, fast, and accurate estimation of kurtosis from clinically feasible data acquisitions remains a challenge. In this study, we first outline a new accurate approach of estimating mean kurtosis via the sub-diffusion mathematical framework. Crucially, this extension of the conventional DKI overcomes the limitation on the maximum b-value of the latter. Kurtosis and diffusivity can now be simply computed as functions of the sub-diffusion model parameters. Second, we propose a new fast and robust fitting procedure to estimate the sub-diffusion model parameters using two diffusion times without increasing acquisition time as for the conventional DKI. Third, our sub-diffusion-based kurtosis mapping method is evaluated using both simulations and the Connectome 1.0 human brain data. Exquisite tissue contrast is achieved even when the diffusion encoded data is collected in only minutes. In summary, our findings suggest robust, fast, and accurate estimation of mean kurtosis can be realised within a clinically feasible diffusion-weighted magnetic resonance imaging data acquisition time.
-
- Neuroscience
The retina transforms patterns of light into visual feature representations supporting behaviour. These representations are distributed across various types of retinal ganglion cells (RGCs), whose spatial and temporal tuning properties have been studied extensively in many model organisms, including the mouse. However, it has been difficult to link the potentially nonlinear retinal transformations of natural visual inputs to specific ethological purposes. Here, we discover a nonlinear selectivity to chromatic contrast in an RGC type that allows the detection of changes in visual context. We trained a convolutional neural network (CNN) model on large-scale functional recordings of RGC responses to natural mouse movies, and then used this model to search in silico for stimuli that maximally excite distinct types of RGCs. This procedure predicted centre colour opponency in transient suppressed-by-contrast (tSbC) RGCs, a cell type whose function is being debated. We confirmed experimentally that these cells indeed responded very selectively to Green-OFF, UV-ON contrasts. This type of chromatic contrast was characteristic of transitions from ground to sky in the visual scene, as might be elicited by head or eye movements across the horizon. Because tSbC cells performed best among all RGC types at reliably detecting these transitions, we suggest a role for this RGC type in providing contextual information (i.e. sky or ground) necessary for the selection of appropriate behavioural responses to other stimuli, such as looming objects. Our work showcases how a combination of experiments with natural stimuli and computational modelling allows discovering novel types of stimulus selectivity and identifying their potential ethological relevance.