A decentralised neural model explaining optimal integration of navigational strategies in insects
Abstract
Insect navigation arises from the coordinated action of concurrent guidance systems but the neural mechanisms through which each functions, and are then coordinated, remains unknown. We propose that insects require distinct strategies to retrace familiar routes (route-following) and directly return from novel to familiar terrain (homing) using different aspects of frequency encoded views that are processed in different neural pathways. We also demonstrate how the Central Complex and Mushroom Bodies regions of the insect brain may work in tandem to coordinate the directional output of different guidance cues through a contextually switched ring-attractor inspired by neural recordings. The resultant unified model of insect navigation reproduces behavioural data from a series of cue conflict experiments in realistic animal environments and offers testable hypotheses of where and how insects process visual cues, utilise the different information that they provide and coordinate their outputs to achieve the adaptive behaviours observed in the wild.
Introduction
Central-place foraging insects navigate using a ‘toolkit’ of independent guidance systems (Wehner, 2009) of which the most fundamental are path integration (PI), whereby foragers track the distance and direction to their nest by integrating the series of directions and distances travelled (for reviews see Heinze et al., 2018; Collett, 2019), and visual memory (VM), whereby foragers derive a homing signal by comparing the difference between current and stored views (for reviews see Zeil, 2012; Collett et al., 2013). Neurophysiological and computational modelling studies advocate the central complex neuropil (CX) as the PI centre (Heinze and Homberg, 2007; Seelig and Jayaraman, 2015; Stone et al., 2017), whereas the mushroom body neuropils (MB) appear well suited to assessing visual valence as needed for VM (Heisenberg, 2003; Ardin et al., 2016; Müller et al., 2018). Yet, two key gaps in our understanding remain. Firstly, although current VM models based on the MB architecture can replicate route following (RF) behaviours whereby insects visually recognise the direction previously travelled at the same position (Ardin et al., 2016; Müller et al., 2018), they cannot account for visual homing (VH) behaviours whereby insects return directly to their familiar surroundings from novel locations following a displacement (e.g. after being blown off course by a gust of wind) (Wystrach et al., 2012). Secondly, despite increasing neuroanatomical evidence suggesting that premotor regions of the CX coordinate navigation behaviour (Pfeiffer and Homberg, 2014; Heinze and Pfeiffer, 2018; Honkanen et al., 2019), a theoretical hypothesis explaining how this is achieved by the neural circuitry has yet to be developed. In this work, we present a unified neural navigation model that extends the core guidance modules from two (PI and VM) to three (PI, RF, and VH) and by integrating their outputs optimally using a biologically realistic ring attractor network in the CX produces realistic homing behaviours.
The foremost challenge in realising this goal is to ensure that the core guidance subsystems provide sufficient directional information across conditions. Contemporary VM models based on the MBs can replicate realistic RF behaviours in complex visual environments (ant environments: Kodzhabashev and Mangan, 2015; Ardin et al., 2016, bee environments: Müller et al., 2018) but do not generalise to visual homing scenarios whereby the animal must return directly to familiar terrain from novel locations (ants: Narendra, 2007, bees: Cartwright and Collett, 1982, wasps: Stürzl et al., 2016). Storing multiple nest-facing views before foraging, inspired by observed learning walks in ants (Müller and Wehner, 2010; Fleischmann et al., 2016) and flights in bees and wasps (Zeil et al., 1996; Zeil and Fleischmann, 2019), provides a potential solution (Graham et al., 2010; Wystrach et al., 2013), but simulation studies have found this approach to be brittle due to high probabilities of aligning with the wrong memory causing catastrophic errors (Dewar et al., 2014). Moreover, ants released perpendicularly to their familiar route do not generally align with their familiar visual direction as predicted by the above algorithms (Wystrach et al., 2012), but instead move directly back towards the route (Fukushi and Wehner, 2004; Kohler and Wehner, 2005; Narendra, 2007; Mangan and Webb, 2012; Wystrach et al., 2012), which would require a multi-stage mental alignment of views for current models. New computational hypothesis are thus required that can guide insects directly back to their route (often moving perpendicularly to the habitual path), but also allow for the route direction to be recovered (now aligned with the habitual path) upon arrival at familiar surroundings (see Figure 1A ‘Zero Vector’).
With the necessary elemental guidance systems defined, a unifying model must then convert the various directional recommendations into a single motor command appropriate to the context (Cruse and Wehner, 2011; Hoinville et al., 2012; Collett et al., 2013; Webb, 2019). Behavioural studies show that when in unfamiliar visual surroundings (‘Off-Route’) insects combine the outputs of their PI and VH systems (Collett, 1996; Bregy et al., 2008; Collett, 2012) relative to their respective certainties consistent with optimal integration theory (Legge et al., 2014; Wystrach et al., 2015; Figure 1A ‘Full Vector’). Upon encountering their familiar route, insects readily recognise their surroundings, recover their previous bearing and retrace their familiar path home (Harrison et al., 1989; Kohler and Wehner, 2005; Wystrach et al., 2011; Mangan and Webb, 2012). Thus, the navigation coordination model must posses two capabilities: (a) output a directional signal consistent with the optimal integration of PI and VH when Off-Route (b) switch from Off-Route (PI and VH) to On-Route (RF) strategies when familiar terrain is encountered. Mathematical models have been developed that reproduce aspects of cue integration in specific scenarios (Cruse and Wehner, 2011; Hoinville and Wehner, 2018), but to date no neurobiologically constrained network revealing how insects might realise these capabilities has been developed.
To address these questions a functional modelling approach is followed that extends the current base model described by Webb, 2019 to (a) account for the ability of ants to home from novel locations back to the familiar route before retracing their familiar path the rest of the journey home, and (b) propose a neurally based model of the central complex neuropil that integrates competing cues optimally and generates a simple steering command that can drive behaviour directly. Performance is bench-marked by direct comparison to behavioural data reported by Wystrach et al., 2012 (showing different navigation behaviours on and off the route), Legge et al., 2014; Wystrach et al., 2015 (demonstrating optimal integration of PI and VM), and through qualitative comparison to extended homing paths where insects switch between strategies according to the context (Narendra, 2007). Biological realism is enforced by constraining models to the known anatomy of specific brain areas, but where no data exists an exploratory approach is taken to investigate the mechanisms that insects may exploit. Figure 1A depicts the adaptive behaviours observed in animals that we wish to replicate accompanied by a functional overview of our unified model of insect navigation (Figure 1B) mapped to specific neural sites (Figure 1C).
Results
Mushroom bodies as drivers of rotational invariant visual homing
For ants to return directly to their familiar route after a sideways displacement (Figure 1A 'Zero Vector') without continuous mental or physical realignment they require access to rotational invariant visual cues. Stone et al., 2018 recently demonstrated that binary images of panoramic skylines converted into their frequency components can provide such a rotationally-invariant encoding of scenes in a compact form (see Image processing for an introduction to frequency transformations of images). Moreover, they demonstrated that the difference between the rotationally invariant features (the amplitudes of the frequency coefficients) between two locations increases monotonically with distance producing an error surface reminiscent of the image difference surfaces reported by Zeil et al., 2003 which can guide an agent back to familiar terrain. Here we investigate whether the MB neuropils shown capable of assessing the visual valence of learned rotationally-varying panoramic skylines for RF (Ardin et al., 2016; Müller et al., 2018), might instead assess the visual valence of rotationally-invariant properties of views sampled along a familiar route supporting visual homing.
To this end, the intensity sensitive input neurons of Ardin et al., 2016’s MB model are replaced with input neurons encoding rotational invariant amplitudes (Figure 2A left, blue panel). The network is trained along an curved route in a simulated world that mimics the training regime of ants in Wystrach et al., 2012 (see Materials and methods and Reproduce visual navigation behaviour for details on simulated world, image processing, model architecture and training and test regime). After training, the firing rate of the MB output neuron (MBON) when placed at locations across the environment at random orientations reveals a gradient that increases monotonically with distance from the familiar route area, providing a homing signal sufficient for VH independent of the animal’s orientation (Figure 2C).
Motor output is then generated by connecting the MBON to a steering network recently located in the fan-shaped body (FB/CBU) of the CX that functions by minimising the difference between the animal’s current and desired headings (Stone et al., 2017). Stone et al., 2017’s key insight was that the anatomically observed shifts of activity in the columnar neurons that encode the desired heading in essence simulate 45° turns left and right, and thus by comparing the summed differences between the activity profiles of these predicted headings to the current heading then the appropriate turning command can be computed (see Figure 2B). We adopt this circuit as the basis for computing steering commands for all strategies as suggested by Honkanen et al., 2019.
In the proposed VH model the current heading input to the steering circuit uses the same celestial global compass used in Stone et al., 2017’s PI model. Insects track their orientation through head-direction cells Seelig and Jayaraman, 2015 whose concurrent firing pattern forms a single bump of activity that shifts around the ring as the animal turns (measured through local visual [Green et al., 2017; Turner-Evans et al., 2017], global visual (Heinze and Homberg, 2007) and proprioceptive (Seelig and Jayaraman, 2015) cues). Neuroanatomical data (Kim et al., 2017; Turner-Evans et al., 2019; Pisokas et al., 2019) supports theoretical predictions (Cope et al., 2017; Kakaria and de Bivort, 2017) that the head-direction system of insects follows a ring attractor (RA) connectivity pattern characterised by local excitatory interconnections between direction selective neurons and global inhibition. In this work, the global compass RA network is not modelled directly but rather we simulate its sinusoidal activity profile in a ring of I-TB1 (locusts and of flies) neurons found in the protocerebral bridge (PCB/PB) (Figure 2A green ring) (see Current headings).
A desired heading is then generated by copying the current activity pattern of the global compass neurons to a new neural ring which we speculate could reside in either a distinct subset of I-TB1 neurons (Beetz et al., 2015) or in the FB. Crucially, the copied activity profile also undergoes a leftward shift proportional to any increase in visual novelty (a similar shifting mechanisms has been proposed for the head-direction system [Green et al., 2017; Turner-Evans et al., 2017]) which we propose is measured by neurons in the superior medial protocerebrum (SMP) (Aso et al., 2014; Plath et al., 2017) (see Figure 2A centre and activity of red rings). The result is a mechanism that recommends changing direction when the agent moves away from familiar terrain (visual novelty increases) but recommends little change to the current heading when the visual novelty is decreasing (see Figure 2C for a schematic of the VH mechanism). We note that there is a distinction between a ring network which describes a group of neurons whose pattern of activity forms a circular representation regardless of actual physical arrangement and RA networks which follow a specific connectivity pattern (all modelled RAs labelled in figures). Taken together the model iteratively refines it’s orientation to descend the visual novelty gradient and thus recover familiar terrain (see Figure 2A for full model).
Figure 2D demonstrates that the proposed network accurately replicates both the directed initial paths as in Wystrach et al., 2012 (see the inserted black arrow), and extended homing paths as in Narendra, 2007 observed in ants displaced to novel locations perpendicular to their familiar routes. We note that upon encountering the route the model is unable to distinguish the direction in which to travel and thus meanders back and forth along the familiarity valley, unlike real ants, demonstrating the need for additional route recognition and recovery capabilities.
Optimally integrating visual homing and path integration
We have demonstrated how ants could use visual cues to return to the route in the absence of PI but in most natural scenarios (e.g. displacement by a gust of wind) ants will retain a home vector readout offering an alternative, and often conflicting, guidance cue to that provided by VH. In such scenarios, desert ants strike a comprise by integrating their PI and VH outputs in a manner consistent with optimal integration theory by weighting VH relative to the familiarity of the current view (Legge et al., 2014) and PI relative to the home vector length (a proxy for directional certainty) (Wystrach et al., 2015).
Various ring-like structures of the CX represent directional cues as bumps of activity with the peak defining the specific target direction, and the spread providing a mechanism to encode cue certainty as required for optimal integration (for an example see increased spread of HD cell activity when only proprioceptive cues are present [Seelig and Jayaraman, 2015]). Besides their excellent properties to encode the animal’s heading ring attractors also provide a biologically realistic means to optimally weight cues represented in this format (Touretzky, 2005; Mangan and Yue, 2018) without the need for dedicated memory circuits to store means and uncertainties of each cue.
Thus we introduce a pair of integrating ring-attractor networks to the CX model (Figure 3A grey neural rings: RA_L and RA_R) that take as input the desired headings from the above proposed VH model (red neural rings: VH_L and VH_R) and Stone et al., 2017’s PI model (orange neural rings: PI_L and PI_R) and output combined Off Route desired heading signals that are sent to the steering circuits (blue neural rings: CPU_L and CPU_R). Stone et al., 2017 mapped the home vector computation to a population of neurons (CPU4) owing to their dual inputs from direction selective compass neurons (I_TB1) and motion-sensitive speed neurons (TN2) as well as their recurrent connectivity patterns facilitating accumulation of activity as the animal moves in a given direction. Wystrach et al., 2015 showed that the certainty of PI automatically scales with the home-vector length owing to the accumulating effect of the memory neurons which correlates with directional uncertainty, and thus the output PI network is directly input to the ring attractor circuits. In our implementation the VH input has a fixed height and width profile and influences the integration through tuning neurons (TUN) (see the plotted activation function in Figure 3B and Optimal cue integration) that we suggest reside in the SMP and modulate the PI input to the integration network. Altering the weighting in this manner rather than by scaling the VH input independently allows VH to dominate the integrated output at sites with high visual familiarity even in the presence of a large home vector without having large stored activity. We note, however, that both approaches remain feasible and further neuroanatomical data is required to clarify which, if either, mechanism is employed by insects.
-
Figure 3—source data 1
- https://cdn.elifesciences.org/articles/54026/elife-54026-fig3-data1-v2.zip
-
Figure 3—source data 2
- https://cdn.elifesciences.org/articles/54026/elife-54026-fig3-data2-v2.zip
-
Figure 3—source data 3
- https://cdn.elifesciences.org/articles/54026/elife-54026-fig3-data3-v2.zip
Figure 3C shows the initial headings produced by the model which replicates the trends reported in cue-conflict experiments by Legge et al., 2014 and Wystrach et al., 2015 when the uncertainty of PI and VH cues were altered independently. Example extended paths of independent PI and VH models and the ring-attractor-based combined PI and VH model are plotted in Figure 3D with the combined model showing the most ant-like behaviour (Kohler and Wehner, 2005; Mangan and Webb, 2012) by initially following predominantly the home-vector direction before switching to visual homing when the home-vector length drops leading the simulated ant back to familiar terrain. Note that the PI-only and PI+VH models are drawn back towards their fictive nest sites indicated by their home vectors which if left to run would likely result in emergent search-like patterns as in Stone et al., 2017. Moreover, upon encountering the route the VH-based models (VH-only and PI+VH) are unable to distinguish the direction in which to travel and hence again (see meander around the valley of familiarity Figure 2D and Figure 3D) further demonstrating a need for a route recovery mechanism.
Route following in the insect brain
The model described above can guide insects back to their familiar route area, but lacks the means to recover the route direction upon arrival as observed in homing insects. This is not surprisingly as VH relies upon translationally-varying but rotational-invariant information whereas RF requires rotationally varying cues. Thus we introduce a new elemental guidance system that makes use of the rotationally-varying phase coefficients of the frequency information derived from the panoramic skyline which tracks the orientation of specific features of the visual surroundings (see Materials and methods). Here, we ask whether by associating the rotationally invariant amplitudes (shown useful for place recognition) with the rotationally-varying phases experienced at those locations, insects might recover the familiar route direction.
Neuroanatomical data with which to constrain a model remains sparse and therefore a standard artificial neural network (ANN) architecture is used to investigate the utility of phase-based route recovery with biological plausibility discussed in more detail below. A three-layer ANN was trained to associate the same 81 rotational-invariant amplitudes as used in the VH model with the rotational varying phase value of single frequency coefficient experienced when travelling along the habitual route which we encode in an eight neuron-ring (see Figure 4A and Route Following for detailed model description). Thus, when the route is revisited the network should output the orientation that the phase converged upon when at the same location previously, which we note is not necessarily aligned with the actual heading of the animal (e.g. it may track the orientation to vertical bar [Seelig and Jayaraman, 2015]). Realignment is possible using the same steering mechanism as described above but which seeks to reduce the offset between the current phase readout (e.g. a local compass locked onto visual features of the animals surroundings), and the recalled phase readout from the ANN.
-
Figure 4—source data 1
- https://cdn.elifesciences.org/articles/54026/elife-54026-fig4-data1-v2.mat
-
Figure 4—source data 2
- https://cdn.elifesciences.org/articles/54026/elife-54026-fig4-data2-v2.mat
-
Figure 4—source data 3
- https://cdn.elifesciences.org/articles/54026/elife-54026-fig4-data3-v2.mat
We speculate that the most likely neural pathways for the new desired and current headings are from Optic Lobe via Anterior Optic Tubercle (AOTU) and Bulb (BU) to EB (CBL) of the CX (Homberg et al., 2003; Omoto et al., 2017) (see Figure 4A) with the desired heading terminating in the EB, whereas the current heading continues to the PB forming a local compass that sits beside the global compass used by PI and VH systems. This hypothesis is further supported by the recently identified parallel pathways from OL via AOTU to the CX in Drosophila (Timaeus et al., 2020). That’s to say that, firstly, there are two parallel pathways forming two compass systems- the global (here based on celestial cues) and the local (based on terrestrial cues) compasses modelled by the activation of I-TB1 and II-TB1 neurons, respectively. Four classes of CL1 neurons (or E-PG and P-EG neurons) Heinze and Homberg, 2009; Xu et al., 2020 and three classes of independent TB1 neurons Beetz et al., 2015 have been identified that provide potential sites for the parallel recurrent loops encoding independent local and global compasses. Secondly, the desired heading, which is the recalled phase of a specific view, is generated through the neural plasticity from AOTU to BU and BU to EB, which is line with recent evidence of associative learning between the R-neurons transmitting visual information from BU to EB and the compass neurons (CL1a or E-PG neurons) that receive input from EB (Kim et al., 2019; Fisher et al., 2019). This kind of learning endows the animal with the ability to flexibly adapt their local compass and also desired navigational orientation according to the changing visual surroundings. Hanesch et al., 1989 reported a direct pathway from EB to FB neurons which we model to allow comparison of the local compass activity (II-TB1) with the desired heading. However, we note that this connectivity has not been replicated in recent studies Heinze and Homberg, 2008 and thus further investigation of potential pathways is required.
The RF model accurately recovers the initial route heading in a similar manner to real ants returned to the start of their familiar route (Wystrach et al., 2012; Figure 4B, insert), and then follows the remaining route in its entirety back to the nest again reflecting ant data (Kohler and Wehner, 2005; Mangan and Webb, 2012; Figure 4B). The quiver plots displayed in the background of Figure 4B show the preferred homing direction output by the ANN when rotated on the spot across locations in the environment. The noise in the results are due to errors in the tracking performance (see examples Figure 4B right) yet as these errors are in largely confined to the magnitude, the steering circuit still drives the ant along the route. We note that this effect is primarily a function of the specific frequency transformation algorithm used which we borrow from computer graphics to investigate the utility of frequency encoding of visual information. The biological realism of such transforms and their potential implementation in the insect vision system are addressed in the Discussion. The displaced routes also highlight the danger of employing solely RF which often shadows rather than converges with the route when displaced sideways, further demonstrating the necessity for integration with the Off-Route strategies that promote route convergence.
Route recovery through context-dependent modulation of guidance systems
Homing insects readily recognise familiar route surroundings, recover their bearing, and retrace their habitual path home, irrespective of the status of other guidance system such as PI. Replicating such context-dependent behavioural switching under realistic conditions is the final task for the proposed model. The visual novelty measured by the MBON provides an ideal signal for context switching with low output when close to the route when RF should dominate versus high output further away from the route when PI and VH should be engaged (see Figure 2D). Also the fact that Off-route strategies (PI and VH) compute their turning angles with reference to the global compass whereas the On-route RF strategy is driven with reference to a local compass provides a means to modulate their inputs to the steering circuit independently. This is realised through a non-linear weighting of the On and Off-route strategies which we propose acts through the same SMP pathway as the VH model (see the SN1 and SN2 neurons in Figure 5A) (see Context-dependent switch for neuron details and Figure 6 for a force-directed graph representation of the final unified model).
The activity of the proposed switching circuit and the paths that it generates in simulated zero vector and full vector displacement trials are shown in Figure 5B and C respectively. In the full vector trial (Figure 5B (upper), Figure 5C (solid line)) as visual novelty is initially high (see high TUN activity until step 78) SN2 is activated which enables Off-Route strategies (PI and VH) while SN1 (always the inverse of SN2) is deactivated which disables On-Route strategies. Note that it is the integration of PI and VH that generates the direct path back to the route area in the FV trial: PI recommends moving at a 45° bearing but VH prevents ascension of the visual novelty gradient that this would cause with the compromise being a bearing closer to 90° that is toward the route. As the route is approached the visual novelty decreases (again see TUN activity), until at step 78 SN2 falls below threshold and deactivates the Off-Route strategies while conversely SN1 activates and engages On-Route strategies. After some initial flip-flopping while the agents converges on the route (steps 78–85) RF becomes dominant and drives the agent back to the nest via the familiar path. In the zero vector trial (Figure 5B (lower), (Figure 5B (dashed line)) Off-route strategies (here only VH) largely dominate (some false positive route recognition (e.g step 60)) until the route is recovered (step 93), at which point the same flip-flopping during route convergence occurs (steps 93–96) followed by RF alone which returns the agent to the nest via the familiar path. It should be noted that the data presented utilised different activation functions of the TUN neuron that weights PI and VH (see Table 1 for parameter settings across trials and Discussion for insights into model limitations and potential extensions), yet the results presented nevertheless provide a proof-of-principle demonstration that the proposed unified navigation model can fulfil all of the criteria defined for replication of key adaptive behaviour observed in insects (Figure 1A).
Discussion
This work addresses two gaps in the current understanding of insect navigation: what are the core visual guidance systems required by the insect navigational toolkit? And how are they coordinated by the insect brain?
We propose that the insect navigation toolkit (Wehner, 2009; Webb, 2019) should be extended to include independent visual homing (VH) and route following (RF) systems (see Figure 1B for updated Insect Navigation Toolkit). We show how VH and RF can be realised using frequency-encoding of panoramic skylines to separate information into rotationally invariant amplitudes for VH and rotationally varying phases for RF. The current model utilises frequency encoding schema from the computer graphics but behavioural studies support the use of spatial frequency by bees (Horridge, 1997; Lehrer, 1999), with neurons in the lobula of dragonflies (O'Carroll, 1993) and locusts James and Osorio, 1996 found to have receptive fields akin to basis functions, providing a mechanism by which to extract the frequency information necessary for the local compass system. Our model allows for this information extraction process to happen at multiple stages ahead of its usage in the central learning sites such as the MBs opening the possibility for its application in either the optic lobes or subsequent pathways through regions such as the AOTU. Further, neurophysiological data is required to pinpoint both the mechanisms and sites of this data processing in insects. Similarly, following Stone et al., 2017 the global compass signal directly mimics the firing pattern of compass neurons in the CX without reference to sensory input but Gkanias et al., 2019 recently presented a plausible neural model of the celestial compass processing pipeline that could be easily integrated into the current model to fill this gap. Follow-on neuroanatomically constrained modelling of the optic lobes presents the most obvious extension of this work allowing the neural pathway from sensory input to motor output signal to be mapped in detail. Conversely, modelling the conversion of direction signals into behaviour via motor generating mechanisms such as central pattern generators (see Steinbeck et al., 2020) will then allow closure of the sensory-motor loop.
Visual homing is modelled on neural circuits found along the OL-MB-SMP pathway (Ehmer and Gronenberg, 2002; Gronenberg and López-Riquelme, 2004) before terminating in the CX steering circuit (Stone et al., 2017) and shown capable of producing realistic homing paths. In this schema, the MBs do not measure rotationally varying sensory valence as recently used to replicate RF (Ardin et al., 2016; Müller et al., 2018), but rather the spatially varying (but rotationally invariant) sensory valence more suited to gradient descent strategies such as visual homing (Zeil et al., 2003; Stone et al., 2018) and other taxis behaviours (Wystrach et al., 2016). This is inline with the hypothesis forwarded by Collett and Collett, 2018 that suggest that the MBs output ‘whether’ the current sensory stimulus is positive or negative and the CX then adapts the animal heading, the ‘whither’, accordingly.
Route following is shown possible by learned associations between the amplitudes (i.e. the place) and the phase (the orientation) experienced along a route, allowing realignment when later at a proximal location. This kind of neural plasticity-based correlation between the visual surroundings and the orientations fits with data recently observed in fruit flies (Kim et al., 2019; Fisher et al., 2019). These studies provide the neural explanation for the animal’s ability to make flexible use of visual information to navigate while the proposed model gives a detailed implementation of such ability in the context of insect’s route following schema. Neurophysiological evidence suggests that the layered visual pathway from OL via AOTU and BU to the EB of the CX (Barth and Heisenberg, 1997; Homberg et al., 2003; Omoto et al., 2017) with its suggested neural plasticity properties (Barth and Heisenberg, 1997; Yilmaz et al., 2019) provides a possible neural pathway but further analysis is needed to identify the circuit structures that might underpin the generation of RF desired heading. In addition to the desired heading, the current heading of RF is derived from the local compass system anchored to animal’s immediate visual surroundings. This independent compass system may be realised parallel to the global compass system in an similar but independent circuit (Heinze and Homberg, 2009; Beetz et al., 2015; Xu et al., 2020). Our model therefore hypothesises that insects possess different compass systems based on varied sensory information and further that insects possess the capability (via CX-based RAs) to coordinate their influence optimally according to the current context. Since the global compass, the local compass and the desired heading of RF share the same visual pathway (OL->AOTU->BU->CX), distinct input and output patterns along this pathway may be found by future neuroanatomical studies. In addition, in the proposed model, the activation of current heading and desired heading of RF overlap in the EB, and therefore separation of activation profiles representing each output (e.g. following methods in Seelig and Jayaraman, 2015) presents another meaningful topic for future neurophysiological research.
Closed-loop behavioural studies during which the spatial frequency information of views is altered (similar to Paulk et al., 2015) coincident with imaging of key brain areas (Seelig and Jayaraman, 2013) offers a means to investigate which neural structures make use of what visual information. Complimentary behavioural experiments could verify the distinct VH and RF systems by selectively blocking the proposed neural pathways with impacts on behaviour predicted by Figure 2C and Figure 4B, respectively. Ofstad et al., 2011 report that visual homing abilities are lost for fruit flies with a blocked EB of the CX but not MB, which is predicted by our model if animals have learned target-facing views to which they can later align using their RF guidance system. Analysis of animal’s orientation during learning is thus vital to unpacking precisely how the above results arise.
With the elemental guidance strategies defined, we propose that their outputs are coordinated through the combined action of the MBs and CX. Specifically, we demonstrate that a pair of ring attractor networks that have similar connectivity patterns of the CX-based head-direction system (Kim et al., 2017; Turner-Evans et al., 2019; Pisokas et al., 2019), are sufficient for optimally weighting multiple directional cues from the same frame of reference (e.g. VH and PI). The use of a pair of integrating RAs is inspired by the column structure of the FB which has 16 neural columns divided into two groups of 8 neural columns that each represent the entire 360°space. The optimal integration of PI and VH using a ring attractor closely matches the networks theorised to govern optimal directional integration in mammals (Jeffery et al., 2016) and supports hypothesis of their conserved use across animals (Mangan and Yue, 2018). Optimality is secured either through adapting the shape of the activity profile of the input as is the case for PI which naturally scales with distance, or by using a standardised input activity profile with cross-inhibition of competing cues as is the case for VH in the model. The later schema avoids the need for ever increasing neural activity to maintain relevance.
To replicate the suite of navigational behaviours described in Figure 1, our network includes three independent ring attractor networks: the global compass head direction system Pisokas et al., 2019; the local compass head direction system (Seelig and Jayaraman, 2015; Kim et al., 2017; Turner-Evans et al., 2019); and an Off-route integration system (modelled here). We would speculate that it is likely that central place foraging insects also possess a similar integration network for ‘On-Route’ cues (not modelled here) bringing the total number of RAs to four. The utility of RAs for head-direction tracking arises from their properties in converging activity to a signal bump that can easily be shifted by sensory input and is maintained in the absence of stimulation. In addition, RAs also possess the beneficial property that they spontaneously weight competing sensory information stored as bumps of activity in an optimal manner. Thus, there are excellent computational reasons for insects to invest in such neural structures. Yet, it should be clear that the model proposed here represents a proof-of-concept demonstrating that the underlying network architectures already mapped to the CX (directional cues encoded as bumps of activity Seelig and Jayaraman, 2015; Heinze and Homberg, 2007; various lateral shifting mechanisms (Stone et al., 2017; Green et al., 2017; Turner-Evans et al., 2017); RAs [Kim et al., 2017; Turner-Evans et al., 2019; Pisokas et al., 2019]) are sufficient to generate adaptive navigation but further studies are required to critique and refine the biological realism of this hypothesis.
While this assemblage recreates optimal integration of strategies that share a compass system, it does not easily extend to integration of directional cues from other frames of reference (e.g. VH and PI reference the global compass versus RF that references a local compass). Indeed as the CX steering network seeks to minimise the difference between a current and a desired heading, calibrating input signals from different frames of reference would require a similar calibration of their respective compass systems. Rather, the proposed model incorporates a context-dependent non-linear switching mechanism driven by the output of the MB that alternates between strategies: global compass based PI and VH are triggered when the surroundings are unfamiliar, but when in familiar surroundings engage local compass-based RF. In summary, the adaptive behaviour demonstrated is the result of distinct guidance systems that converge in the CX, with their relative weighting defined by the output of the MB. This distributed architecture is reminiscent of mechanisms found in the visual learning of honeybees (Plath et al., 2017), and supports the hypothesis that the CX is the navigation coordinator of insects (Heinze, 2017; Honkanen et al., 2019) but shows how the MB acts as a mediator allowing the CX to generate optimal behaviour according to the context.
The resultant unified model of insect navigation Figure 1B and C represents a proof-of-principle framework as to how insects might co-ordinate core navigational behaviours (PI, VH and RF) under standard field manipulations Figure 1A. Neuroanatomical data has been drawn from across insect classes (see Table 2) to ensure neural realism where possible with performance compared to ant navigation behaviour in a single simulated desert ant habitat. The framework can be easily extended to new navigation behaviours observed in other insects from idiothetic PI (Kim and Dickinson, 2017) to straight line following El Jundi et al., 2016 to migrations (Reppert et al., 2016) as well as more nuanced strategies that flexibly use directional cues from different sensory modalities (Wystrach et al., 2013; Schwarz et al., 2017; Dacke et al., 2019). A priority of future works should be the investigation of the differences and commonalities in sensory systems, neural structures and ecology of different insect navigators and how they impact behaviour allowing for extension and refinement of the framework for different animals. Complementary stress-testing of models across different environments in both simulation and robotic studies are also required to ensure that model performance generalises across species and habitats and to provide guidance to researchers seeking the sensory, processing and learning circuits underpinning these abilities.
Materials and methods
All source code related to this publication is available for download at https://github.com/XuelongSun/InsectNavigationToolkitModelling (Sun et al., 2020 ; copy archived at https://github.com/elifesciences-publications/InsectNavigationToolkitModelling). All simulations and network models are implemented by Python 3.5 and make use of external libraries-numpy, matplotlib, scipy, PIL and cv2.
Simulated 3D world
Request a detailed protocolThe environment used in this study is that provided by Stone et al., 2018 which is itself adapted from Baddeley et al., 2012 (see Figure 7C). It is a virtual ant-like world consisting of randomly generated bushes, trees and tussocks based on triangular patches (for more details see Baddeley et al., 2012). Therefore, the data of this simulated world is stored in a matrix with the size of , defining the three dimensional coordinates (x,y,z) of the three vertices of (number of patches) triangle patches. Agent movement was constrained to a training and test area allowing free movement without the requirement of an additional obstacle avoidance mechanism.
Image reconstruction
Request a detailed protocolThe agent’s visual input at location with the heading direction is simulated from a point 1 cm above from the ground plane with field of view wide by high (centred on the horizon). This panoramic image () is then wrapped onto a sky-centred disk as required by the Zernike Moments transformation algorithm used with the size of ready for image processing (see Figure 7D upper).
Image processing
Frequency encoding conceptual overview
Request a detailed protocolImage compression algorithms such as JPEG encoding Hudson et al., 2018 have long utilised the fact that a complex signal can be decomposed into a series of trigonometric functions that oscillate at different frequencies. The original signal can then be reconstructed by summing all (for prefect reconstruction) or some (for approximate reconstruction) of the base trigonometric functions. Thus, compression algorithms seek a balance between using the fewest trigonometric functions to encode the scene (for example, by omitting high frequencies that humans struggle to perceive), and the accuracy of the reconstructed signal (often given as an option when converting to JPEG format). Figure 7A provides a cartoon of the frequency decomposition process for a panoramic view.
When such transforms are applied to fully panoramic images, or skylines, benefits beyond compression arise. Specifically, discrete transformation algorithms used to extract the frequency information generate a series of information triplets to describe the original function: frequency coefficients describe the frequency of the trigonometric function with associated amplitudes and phase values defining the vertical height versus the mean and the lateral position of the waveform respectively (Figure 7A). For panoramic views, regardless of the rotational angle of the image capturing device (eye or camera) the entire signal will always be visible and hence the amplitudes of the frequency coefficients do not alter with rotation (Figure 7B). This information has been used for successful place recognition in a series of robot studies (Pajdla and Hlaváč, 1999; Menegatti et al., 2004; Stone et al., 2016). Most recently Stone et al., 2018 demonstrated that the difference between the amplitudes of the frequency coefficients recorded at two locations increases monotonically with distance producing an error surface suitable for visual homing. This feature of the frequency encoding underlies the visual homing results described in Mushroom bodies as drivers of rotational invariant visual homing.
In addition, as the phase of each coefficient describes how to align the signal this will naturally track any rotation in the panoramic view (Figure 7B) providing a means to realign with previous headings. The phase components of panoramic images have been utilised previously to derive the home direction in a visual homing task (Stürzl and Mallot, 2006). This feature of the frequency encoding underlies the route following results described in Route following in the insect brain.
The image processing field has created an array of algorithms for deriving the frequency content of continuous signals (Jiang et al., 1996; Gonzalez et al., 2004). To allow exploration of the usefulness of frequency information, and how it could be used by the known neural structures, we adopt the same Zernike Moment algorithm used by Stone et al., 2018, but the reader should be clear that there are many alternate and more biologically plausible processes by which insects could derive similar information. It is beyond the scope of this proof of concept study to define precisely how this process might happen in insects but future research possibilities are outlined in the Discussion.
Zernike Moments encoding
Request a detailed protocolZernike Moments (ZM) are defined as the projection of a function onto orthogonal basis polynomials called Zernike polynomials (Teague, 1980; Khotanzad and Hong, 1990). This set of functions are defined on the unit circle with polar coordinates shown as:
where is the order and is the repetition meeting the condition: , and is even to ensure the rotational invariant property is met. is the radial polynomial defined as:
For a continuous image function , the ZM coefficient can be calculated by:
For a digital image, summations can replace the integrals to give the ZM:
ZM are extracted from the simulated insect views in wrapped format (Figure 7D) whose centre is taken to be the origin of the polar coordinates such that all valid pixels lie within the unit circle. For a given image (P1 in Figure 7D) and the rotated version of this image , (P2 in Figure 7D), the amplitude and phase of ZM coefficients of these two images will satisfy:
From which we can see that the amplitude of the ZM coefficient remains the same while the phase of ZM carries the information regarding the rotation (see Figure 7A and D). This property is the cornerstone of the visual navigation model where the amplitudes encode the features of the view while the phase defines the orientation.
Amplitudes for ZM orders ranging from to were selected as they appeared to cover the majority of information within the image. From Equation 1, we know that , so we limited to reduce the computational cost, which sets the total number of ZM coefficients () to which was input to the visual navigation networks. For training the ANN network for RF, in Equation 5, if we set , such that which means that all ZM coefficients will provide the same information when the image is rotated. Further, the difference between the phase of ZM coefficients of the current view with those of the memorised view, will inherently provide the angle with which to turn to realign oneself, that is:
where the order of this ZM is selected to be manually by comparing the performance with different orders in this specific virtual environment, is the current heading of the agent while is the memorised heading direction (desired heading direction).
Neural networks
We use the simple firing rate to model the neurons in the proposed networks, where the output firing rate is a sigmoid function of the input if there is no special note. In the following descriptions and formulas, a subscript is used to represent the layers or name of the neuron while the superscript is used to represent the value at a specific time or with a specific index.
Current headings
Request a detailed protocolIn the proposed model, there are two independent compass systems based on the global and the local cues respectively so named global and local compass correspondingly. These two compass systems have similar neural pathways from OL via AOTU and BU to the CX but ended distinct groupings of TB1 neurons: I-TB1 and II-TB1 in the PB.
Global compass
Request a detailed protocolThe global compass neural network applied in this study is the same as that of Stone et al., 2017, which has three layers of neurons: TL neurons, CL1 neurons and I-TB1 neurons. The 16 TL neurons respond to simulated polarised light input and are directly modelled as:
where is the angular preference of the 16 TL-neurons. The 16 CL1-neurons are inhibited by TL-neuron activity which invert the polarisation response:
The 8 I-TB1 neurons act as a ring attractor creating a sinusoidal encoding of the current heading. Each I-TB1 neuron receives excitation from the CL1 neuron sharing the same directional preference and inhibition from other I-TB1 neurons via mutual connections:
where is a balance factor to modify the strength of the inhibition and the CL1 excitation. Finally, the population coding represents the heading of global compass of the agent at time .
Local compass
The local compass is derived from the terrestrial cues through a similar visual pathway as the global compass and also ends in a ring attractor network. As for the global compass, the local compass heading is directly modelled by the population encoding of II-TB1 neurons:
where is the angular preference of the II-TB1 neurons and is the phase of ZM. Therefore, the firing rate of encodes the heading of the local compass.
Visual homing
Request a detailed protocolThe neural network of visual homing is an associative network constrained by the anatomical structure of the mushroom body (MB) of the insects. In contrast to Ardin et al., 2016 where a spiking neural network is implemented to model the MB, we apply a simple version of MB where the average firing rates of neurons are used.
The visual projection neurons (vPNs) directly receive the amplitudes of the ZM coefficients as their firing rates:
where is the number of the vPN neurons which is the same as the total number of ZM amplitudes applied and in this study . The denotes the amplitudes of ZM coefficients.
The vPNs project into Kenyon cells (KC) through randomly generated binary connections , which result in the scenario wherein one KC receives 10 randomly selected vPNs’ activation:
where denotes the total input current of KC from the vPN and the KCs are modelled as binary neurons with the same threshold :
The MBON neuron sums all the activation of Kenyon cells via plastic connections :
An anti-Hebbian learning rule is applied for the plasticity of in a simple way:
where is the learning rate. The learning process will happen only when the reward signal is turned on. The activation of EN represents the familiarity of the current view and the change of the is defined as:
is used to track the gradient of the familiarity to guide the agent to the more familiar locations by shifting the I-TB1 neurons’ activation .
The relationship between the and the is shown as following:
Path integration
Request a detailed protocolThe PI model implemented is that published by Stone et al., 2017. The core functionality arises from the CPU4 neurons that integrate the activation of TN2 neurons that encode the speed of the agent and the inverted activation of direction-sensitive I-TB1 neurons. The result is that the population of CPU4 neurons iteratively track the distance and orientation to the nest (a home vector) in a format akin to a series of directionally locked odometers.
The firing rate of the CPU4 neurons are updated by:
where the rate of the memory accumulation ; the memory loss ; the initial memory charge of CPU4 neurons .
The input of the TN2 neurons encoding the speed is calculated by:
where is the velocity (see Equation 39) of the agent and is the preference angle of the TN2 neurons. In this study . The activation function applied to TN2 neurons is the rectified linear function given by:
As CPU4 neurons integrate the speed and direction of the agent, the desired heading of PI can be represented by the population encoding of these neurons, thus:
Route following
Request a detailed protocolThe route following model is based on a simple artificial neural network (ANN) with just one hidden layer. The input layer directly takes the amplitudes of the ZM coefficients as the activation in the same way as that of visual projection neurons in MB network. This is a fully connected neural network with the sigmoid activation function, so the forward propagation is ruled by:
where and denote the input and output of the neuron in layer, thus the input is the same as the MB network and the output of the ANN is consequently the population coding of the RF desired heading, that is:
For a fast and efficient implementation, the learning method applied here is back propagation with gradient descend. Training data is derived from the amplitudes and the population encoded phases of the ZM coefficients of the images reconstructed along a habitual route. As shown in Equation 11 the II-TB1 neurons encode the heading of local compass, therefore, the training pair for the RF network can be defined as . After training, this network will correlate the desired ZM phase with the specific ZM amplitudes, and when RF is running, the output of this neural network will represent the desired heading with respect to the current heading of the local compass represented by the population encoding of II-TB1 neurons.
Coordination of elemental guidance strategies
Request a detailed protocolThe coordination of the three main navigation strategies PI, VH and RF are realised in distinct stages. Firstly, Off-route strategies (PI and VH) are optimally integrated by weighing according to the certainly of each before a context-dependent switch activates either On-route (RF) or Off-route strategies depending on the current visual novelty.
Optimal cue integration
Request a detailed protocolA ring attractor neural network is used to integrate the cues from the VH and PI guidance systems. As reported in Hoinville and Wehner, 2018 summation of directional cues represented in vector format leads to optimal angular cue integration which is the same case as real insects. Mangan and Yue, 2018 gave a biology plausible way to do this kind of computation based on a simple ring attractor neural network. There are two populations of neurons in this network, the first is the integration neurons (IN) which is the output population of the network. Constrained by the number of columns in each hemisphere of the insects CX, we set the number of the IN to be 8, and its firing rate is updated by:
where is the recurrent connections from neuron to neuron, is the activation function that provides the non-linear property of the neuron:
where denotes the offset of the function.
In Equation 26, and generally denote the cues that should be integrated. In this study, and represent the desired heading of path integration () and visual homing (). The desired heading of PI is also tuned by the tuning neuron (TUN) in SMP which is stimulated by the MBON of MB (see Figure 3A) and its activation function is defined by a rectified linear function, that is:
where is the scaling factor.
Thus, the and for this ring attractor network can be calculated by:
The second population of the ring attractor is called the uniform inhibition (UI) neurons modelled by:
After arriving at a stable state, the firing rate of the integration neurons in this ring attractor network provides the population encoding of the optimal integrated output :
Context-dependent switch
The model generates two current/desired headings pairs: the current heading of global compass decoded by with the desired heading optimally integrated by the integration neurons of the ring attractor network and the current heading of local compass decoded by II-TB1 neurons with the desired heading decoded by the output of the RF network . These two pairs of signal both are connected to the steering circuit (see Figure 5A and Steering circuit) but are turned on/off by two switching neurons (SN1 and SN2) in the SMP (Figure 5A). SN2 neuron receives the activation from MBON neuron and is modelled as:
While SN1 will always fire unless SN2 fires:
Therefore, the context-depend switch is achieved according to the current visual novelty represented by the activation of MBON.
Steering circuit
Request a detailed protocolThe steering neurons, that is CPU1 neurons () receive excitatory input from the desired heading () and inhibitory input from the current heading () to generate the turning signal:
The turning angle is determined by the difference of the activation summations between left () and right () set of CPU1 neurons:
which corresponds to the difference of the length of the subtracted left and right vectors in Figure 2A. In addition, as it is illustrated in Figure 2A, another key part of steering circuit is the left/right shifted desired heading, in this paper, this is achieved by the offset connectivity pattern ( and ) from the desired heading to the steering neurons (Heinze and Homberg, 2008; Stone et al., 2017):
Where the and are:
which defines the connection pattern realising the left/right shifting of the desired headings used throughout our model (Figure 2A, Figure 3A, Figure 4A, Figure 5A and Figure 6A.
The current heading input to the steering circuit is also switched between global and local compass input via the SN1 and SN2 neuron:
Detailed neural connectivity of unified model
Request a detailed protocolFigure 6A shows a complete picture of the proposed model. Specifically, it highlights the final coordination system showing that CX computing the optimal navigation output with the modulation from the MB and SMP. In addition, offset connectivity pattern from the desired heading to the steering circuit that underpin the left/right shifting is clearly shown. Figure 6B and C shows the network generating the desired heading of RF and VH respectively.
In addition, Table 2 provides details of all modelled neural circuits with their function and naming conventions with links to biological evidence for these neural circuits where it exists and the animal that they were observed in.
Simulations
Equation 35 gives the turning angle of the agent, thus the instantaneous "velocity" () at every step can be computed by:
where is the step length with the unit of centimetres. Note that we haven’t defined the time accuracy for every step of the simulations, thus the unit of the velocity in this implementation is rather than . Then the position of agent in the Cartesian coordinates for the is updated by:
The main parameter settings for all the simulations in this paper can be found in Table 1.
Reproduce visual navigation behaviour
Request a detailed protocolInspired by the benchmark study of real ants in Wystrach et al., 2012, we test our model of VH and RF by reproducing the homing behaviours in that study. This is achieved by constructing a habitual route with a similar shape (arc or banana shape) in our simulated 3D world. The position and heading along that route is manually generated by:
where the is the radius of the arc and in this case is the number of the sampling points where view images are reconstructed along the route. The reconstructed views then be wrapped and decomposed by ZM into amplitudes and phases are used to train the ANN network of RF and MB network of VH.
Visual homing
Request a detailed protocolAfter training, 12 agents with different initial headings that were evenly distributed in were released at the sideways release point () for the simulation of VH (Figure 2D). The headings of the agents at radius 2.5 m from the release point (manually selected to ensure that the all the agents have completed any large initial loop) are taken as the initial headings.
Route following
After training, 2 agents with 0° and 180° are released at the different release points () for the simulation of RF (see Figure 4B) to generate the homing path. And then, we release 12 agents on the route () with different initial headings that is evenly distributed in to compare the results with the real ant data in Wystrach et al., 2012. The heading of each agent at the position that is 0.6m from the release point is taken as the initial heading.
Reproduce the optimal cue integration behaviour
Request a detailed protocolWe evaluated the cue integration model by reproducing the results of Wystrach et al., 2012 and Legge et al., 2014. The ants’ outbound routes in Wystrach et al., 2015 is bounded by the corridor, so here we simulate the velocity of the agent by:
where the function generates a random value from the uniform distribution of , thus the speed of x-axis will be in and will cancel each other during the forging. The speed of y-axis is constant so it will accumulated and be recorded by the PI model. And is the basic speed of the agent and is the total time for outbound phase determining the length of the outbound route. As for the simulated homing route, we duplicate the outbound route when but with a inverted heading direction. And then the visual navigation network was trained with images sampled along a simulated route (grey curve in Figure 3B).
Tuning PI uncertainty
Request a detailed protocolThe agent in this simulation was allowed to forage to different distances of 0.1m, 1m, 3m or 7m from the nest to accrue different PI states and directional certainties before being translated to a never-before-experienced test site 1.5m from the nest. (RP1 in Figure 3B). For each trial, we release 20 agents with different initial headings that is evenly distributed in . The headings of every agent at the position that is 0.6m from the start point is taken as the initial headings, and the mean direction and the 95% confidential intervals are calculated. As in the biological experiment, the angle between the directions recommended by the PI and visual navigation systems differed by approximately 130°.
As the length of the home vector increase (0.1m -> 7m) the activation of PI memory becomes higher (Figure 3B), and increasingly determines the output of the ring attractor integration. Since the length of the home vector is also encoded in the activation of the PI memory neurons, the ring attractor can extract this information as the strength of the cue. As the visual familiarity is nearly the same in the vicinity of the release point, the strength of visual homing circuit remains constant and has more of an influence as the PI length drops.
Tuning visual uncertainty
Request a detailed protocolThe agent in this simulation was allowed to forage up to 1m from the nest to accrue its PI state and directional certainty before being translated to three different release points (RP1, RP2 and RP3 in Figure 3B). As the distance from nest increases (RP1->RP2->RP3) so does the visual uncertainty. For each trial, we release 12 agents with different initial headings that is evenly distributed in . The headings of each agent at the position that is 0.3m from the start point is taken as the initial headings, and the mean direction and the 95% confidential intervals are calculated.
Whole model
Request a detailed protocolThe simulated habitual route remains the same as in the simulation of visual navigation (Reproduce visual navigation behaviour) as is the learning procedure. The zero- and full- vector agents are both released at with the heading 0° and 90°, respectively. The full-vector agent’s PI memory is generated by letting the agent forage along the route from nest to feeder.
Data availability
All the source code of the implementation and part of the data are uploaded to Github and are available via https://github.com/XuelongSun/InsectNavigationToolkitModelling (copy archived at https://github.com/elifesciences-publications/InsectNavigationToolkitModelling).
References
-
Using an insect mushroom body circuit to encode route memory in complex natural environmentsPLOS Computational Biology 12:e1004683.https://doi.org/10.1371/journal.pcbi.1004683
-
A model of ant route navigation driven by scene familiarityPLOS Computational Biology 8:e1002336.https://doi.org/10.1371/journal.pcbi.1002336
-
Topographic organization and possible function of the posterior optic tubercles in the brain of the desert locust Schistocerca gregariaJournal of Comparative Neurology 523:1589–1607.https://doi.org/10.1002/cne.23736
-
Nest-mark orientation versus vector navigation in desert antsJournal of Experimental Biology 211:1868–1873.https://doi.org/10.1242/jeb.018036
-
Insect navigation en route to the goal: multiple strategies for the use of landmarksThe Journal of Experimental Biology 199:227–235.
-
How navigational guidance systems are combined in a desert antCurrent Biology 22:927–932.https://doi.org/10.1016/j.cub.2012.03.049
-
Spatial memory in insect navigationCurrent Biology 23:R789–R800.https://doi.org/10.1016/j.cub.2013.07.020
-
Path integration: how details of the honeybee waggle dance and the foraging strategies of desert ants might help in understanding its mechanismsThe Journal of Experimental Biology 222:jeb205187.https://doi.org/10.1242/jeb.205187
-
No need for a cognitive map: decentralized memory for insect navigationPLOS Computational Biology 7:e1002009.https://doi.org/10.1371/journal.pcbi.1002009
-
Segregation of visual input to the mushroom bodies in the honeybee (Apis mellifera)The Journal of Comparative Neurology 451:362–373.https://doi.org/10.1002/cne.10355
-
Mushroom body volumes and visual interneurons in ants: comparison between sexes and castesThe Journal of Comparative Neurology 469:198–213.https://doi.org/10.1002/cne.11014
-
A Snapshot-Based mechanism for celestial orientationCurrent Biology 26:1456–1462.https://doi.org/10.1016/j.cub.2016.03.030
-
Ontogeny of learning walks and the acquisition of landmark information in desert ants, Cataglyphis fortisThe Journal of Experimental Biology 219:3137–3145.https://doi.org/10.1242/jeb.140459
-
Navigation in wood ants Formica japonica: context dependent use of landmarksJournal of Experimental Biology 207:3431–3439.https://doi.org/10.1242/jeb.01159
-
From skylight input to behavioural output: a computational model of the insect polarised light compassPLOS Computational Biology 15:e1007123.https://doi.org/10.1371/journal.pcbi.1007123
-
Animal cognition: multi-modal interactions in ant learningCurrent Biology 20:R639–R640.https://doi.org/10.1016/j.cub.2010.06.018
-
Multisensory convergence in the mushroom bodies of ants and beesActa Biologica Hungarica 55:31–37.https://doi.org/10.1556/ABiol.55.2004.1-4.5
-
Neuronal architecture of the central complex in Drosophila melanogasterCell and Tissue Research 257:343–366.https://doi.org/10.1007/BF00261838
-
Unraveling the neural basis of insect navigationCurrent Opinion in Insect Science 24:58–67.https://doi.org/10.1016/j.cois.2017.09.001
-
Principles of insect path integrationCurrent Biology 28:R1043–R1058.https://doi.org/10.1016/j.cub.2018.04.058
-
Neuroarchitecture of the central complex of the desert locust: intrinsic and columnar neuronsThe Journal of Comparative Neurology 511:454–478.https://doi.org/10.1002/cne.21842
-
Editorial: the insect central Complex-From sensory coding to directing movementFrontiers in Behavioral Neuroscience 12:156.https://doi.org/10.3389/fnbeh.2018.00156
-
Mushroom body memoir: from maps to modelsNature Reviews Neuroscience 4:266–275.https://doi.org/10.1038/nrn1074
-
ConferenceLearning and retrieval of memory elements in a navigation taskConference on Biomimetic and Biohybrid Systems. pp. 120–131.
-
Organization and neural connections of the anterior optic tubercle in the brain of the Locust, Schistocerca gregariaThe Journal of Comparative Neurology 462:415–430.https://doi.org/10.1002/cne.10771
-
The insect central complex and the neural basis of navigational strategiesThe Journal of Experimental Biology 222:jeb188854.https://doi.org/10.1242/jeb.188854
-
Pattern discrimination by the honeybee: disruption as a cueJournal of Comparative Physiology A: Sensory, Neural, and Behavioral Physiology 181:267–277.https://doi.org/10.1007/s003590050113
-
JPEG-1 standard 25 years: past, present, and future reasons for a successJournal of Electronic Imaging 27:040901.https://doi.org/10.1117/1.JEI.27.4.040901
-
Characterisation of columnar neurons and visual signal processing in the medulla of the locust optic lobe by system identification techniquesJournal of Comparative Physiology A 178:183–199.https://doi.org/10.1007/BF00188161
-
Optimal cue combination and landmark-stability learning in the head direction systemThe Journal of Physiology 594:6527–6534.https://doi.org/10.1113/JP272945
-
Optical image reconstruction using frequency-domain data: simulations and experimentsJournal of the Optical Society of America A 13:253–266.https://doi.org/10.1364/JOSAA.13.000253
-
Ring attractor dynamics emerge from a spiking model of the entire protocerebral bridgeFrontiers in Behavioral Neuroscience 11:8.https://doi.org/10.3389/fnbeh.2017.00008
-
Invariant image recognition by Zernike momentsIEEE Transactions on Pattern Analysis and Machine Intelligence 12:489–497.https://doi.org/10.1109/34.55109
-
Idiothetic path integration in the fruit fly Drosophila melanogasterCurrent Biology 27:2227–2238.https://doi.org/10.1016/j.cub.2017.06.026
-
ConferenceRoute following without scanningConference on Biomimetic and Biohybrid Systems. pp. 199–210.
-
Idiosyncratic route-based memories in desert ants, Melophorus bagoti: how do they interact with path-integration vectors?Neurobiology of Learning and Memory 83:1–12.https://doi.org/10.1016/j.nlm.2004.05.011
-
Combining sky and earth: desert ants (Melophorus bagoti) show weighted integration of celestial and terrestrial cuesJournal of Experimental Biology 217:4159–4166.https://doi.org/10.1242/jeb.107862
-
Shape perception in the honeybee: symmetry as a global frameworkInternational Journal of Plant Sciences 160:S51–S65.https://doi.org/10.1086/314216
-
ConferenceAn analysis of a ring attractor model for cue integrationConference on Biomimetic and Biohybrid Systems. pp. 459–470.
-
Image-based memory for robot navigation using properties of omnidirectional imagesRobotics and Autonomous Systems 47:251–267.https://doi.org/10.1016/j.robot.2004.03.014
-
A neural network model for familiarity and context learning during honeybee foraging flightsBiological Cybernetics 112:113–126.https://doi.org/10.1007/s00422-017-0732-z
-
Homing strategies of the australian desert ant Melophorus bagoti. II. interaction of the path integrator with visual cue informationJournal of Experimental Biology 210:1804–1812.https://doi.org/10.1242/jeb.02769
-
ConferenceZero phase representation of panoramic images for image based localizationInternational Conference on Computer Analysis of Images and Patterns. pp. 550–557.
-
Closed-Loop behavioral control increases coherence in the fly brainJournal of Neuroscience 35:10304–10315.https://doi.org/10.1523/JNEUROSCI.0691-15.2015
-
Organization and functional roles of the central complex in the insect brainAnnual Review of Entomology 59:165–184.https://doi.org/10.1146/annurev-ento-011613-162031
-
Neurobiology of monarch butterfly migrationAnnual Review of Entomology 61:25–42.https://doi.org/10.1146/annurev-ento-010814-020855
-
Anatomy of the mushroom bodies in the honey bee brain: the neuronal connections of the alpha-lobeThe Journal of Comparative Neurology 334:444–465.https://doi.org/10.1002/cne.903340309
-
How ants use vision when homing backwardCurrent Biology 27:401–407.https://doi.org/10.1016/j.cub.2016.12.019
-
Connecting brain to behaviour: a role for general purpose steering circuits in insect orientation?The Journal of Experimental Biology 223:jeb212332.https://doi.org/10.1242/jeb.212332
-
ConferenceSkyline-based localisation for aggressively manoeuvring robots using uv sensors and spherical harmonicsIEEE International Conference on Robotics and Automation. pp. 5615–5622.https://doi.org/10.1109/ICRA.2016.7487780
-
An anatomically constrained model for path integration in the bee brainCurrent Biology 27:3069–3085.https://doi.org/10.1016/j.cub.2017.08.052
-
Rotation invariant visual processing for spatial memory in insectsInterface Focus 8:20180010.https://doi.org/10.1098/rsfs.2018.0010
-
How wasps acquire and use views for homingCurrent Biology 26:470–482.https://doi.org/10.1016/j.cub.2015.12.052
-
Efficient visual homing based on Fourier transformed panoramic imagesRobotics and Autonomous Systems 54:300–313.https://doi.org/10.1016/j.robot.2005.12.001
-
Image analysis via the general theory of moments*Journal of the Optical Society of America 70:920–930.https://doi.org/10.1364/JOSA.70.000920
-
BookAttractor network models of head direction cellsIn: Wiener SI, Taube JS, editors. Head Direction Cells and the Neural Mechanisms of Spatial Orientation. MIT Press. pp. 411–432.
-
The internal maps of insectsThe Journal of Experimental Biology 222:jeb188094.https://doi.org/10.1242/jeb.188094
-
The architecture of the desert ant’s navigational toolkitMyrmecological News 12:85–96.
-
Ants might use different view-matching strategies on and off the routeJournal of Experimental Biology 215:44–55.https://doi.org/10.1242/jeb.059584
-
Snapshots in ants? new interpretations of paradigmatic experimentsJournal of Experimental Biology 216:1766–1770.https://doi.org/10.1242/jeb.082941
-
Optimal cue integration in antsProceedings of the Royal Society B: Biological Sciences 282:20151484.https://doi.org/10.1098/rspb.2015.1484
-
Distributed plasticity in ant visual pathways following colour learningProceedings of the Royal Society B: Biological Sciences 286:20182813.https://doi.org/10.1098/rspb.2018.2813
-
Structure and function of learning flights in ground-nesting bees and waspsThe Journal of Experimental Biology 199:245–252.
-
Catchment Areas of panoramic snapshots in outdoor scenesJournal of the Optical Society of America A 20:450–469.https://doi.org/10.1364/JOSAA.20.000450
-
Visual homing: an insect perspectiveCurrent Opinion in Neurobiology 22:285–293.https://doi.org/10.1016/j.conb.2011.12.008
-
The learning walks of ants (hymenoptera: formicidae)Myrmecological News 29:93–110.https://doi.org/10.25849/myrmecol.news_029:093
Article and author information
Author details
Funding
Horizon 2020 Framework Programme (ULTRACEPT 778062)
- Xuelong Sun
- Shigang Yue
Horizon 2020 Framework Programme (STEP2DYNA 691154)
- Xuelong Sun
- Shigang Yue
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
This research has received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Sklodowska-Curie grant agreement No 778062, ULTRACEPT and No 691154, STEP2DYNA.
Thanks to Barbara Webb and Insects Robotics Group at the Univ of Edinburgh, Hadi Maboudi, Alex Cope and Andrew Philippedes for comments on early drafts, and to Antoine Wystrach for provision of data from previous works. Thanks for proof readers Anne and Mike Mangan (Snr). Finally, thanks to our editor and reviewers who helped improve the model and manuscript through their excellent feedback.
Copyright
© 2020, Sun et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 2,504
- views
-
- 395
- downloads
-
- 67
- citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Computational and Systems Biology
- Neuroscience
Hypothalamic kisspeptin (Kiss1) neurons are vital for pubertal development and reproduction. Arcuate nucleus Kiss1 (Kiss1ARH) neurons are responsible for the pulsatile release of gonadotropin-releasing hormone (GnRH). In females, the behavior of Kiss1ARH neurons, expressing Kiss1, neurokinin B (NKB), and dynorphin (Dyn), varies throughout the ovarian cycle. Studies indicate that 17β-estradiol (E2) reduces peptide expression but increases Slc17a6 (Vglut2) mRNA and glutamate neurotransmission in these neurons, suggesting a shift from peptidergic to glutamatergic signaling. To investigate this shift, we combined transcriptomics, electrophysiology, and mathematical modeling. Our results demonstrate that E2 treatment upregulates the mRNA expression of voltage-activated calcium channels, elevating the whole-cell calcium current that contributes to high-frequency burst firing. Additionally, E2 treatment decreased the mRNA levels of canonical transient receptor potential (TPRC) 5 and G protein-coupled K+ (GIRK) channels. When Trpc5 channels in Kiss1ARH neurons were deleted using CRISPR/SaCas9, the slow excitatory postsynaptic potential was eliminated. Our data enabled us to formulate a biophysically realistic mathematical model of Kiss1ARH neurons, suggesting that E2 modifies ionic conductances in these neurons, enabling the transition from high-frequency synchronous firing through NKB-driven activation of TRPC5 channels to a short bursting mode facilitating glutamate release. In a low E2 milieu, synchronous firing of Kiss1ARH neurons drives pulsatile release of GnRH, while the transition to burst firing with high, preovulatory levels of E2 would facilitate the GnRH surge through its glutamatergic synaptic connection to preoptic Kiss1 neurons.
-
- Computational and Systems Biology
Degree distributions in protein-protein interaction (PPI) networks are believed to follow a power law (PL). However, technical and study bias affect the experimental procedures for detecting PPIs. For instance, cancer-associated proteins have received disproportional attention. Moreover, bait proteins in large-scale experiments tend to have many false-positive interaction partners. Studying the degree distributions of thousands of PPI networks of controlled provenance, we address the question if PL distributions in observed PPI networks could be explained by these biases alone. Our findings are supported by mathematical models and extensive simulations and indicate that study bias and technical bias suffice to produce the observed PL distribution. It is, hence, problematic to derive hypotheses about the topology of the true biological interactome from the PL distributions in observed PPI networks. Our study casts doubt on the use of the PL property of biological networks as a modeling assumption or quality criterion in network biology.