A Bayesian approach to single-particle electron cryo-tomography in RELION-4.0
Abstract
We present a new approach for macromolecular structure determination from multiple particles in electron cryo-tomography (cryo-ET) data sets. Whereas existing subtomogram averaging approaches are based on 3D data models, we propose to optimise a regularised likelihood target that approximates a function of the 2D experimental images. In addition, analogous to Bayesian polishing and contrast transfer function (CTF) refinement in single-particle analysis, we describe the approaches that exploit the increased signal-to-noise ratio in the averaged structure to optimise tilt-series alignments, beam-induced motions of the particles throughout the tilt-series acquisition, defoci of the individual particles, as well as higher-order optical aberrations of the microscope. Implementation of our approaches in the open-source software package RELION aims to facilitate their general use, particularly for those researchers who are already familiar with its single-particle analysis tools. We illustrate for three applications that our approaches allow structure determination from cryo-ET data to resolutions sufficient for de novo atomic modelling.
Editor's evaluation
Single-particle tomography (SPT) is a useful method to determine the structure of proteins imaged in situ. This important work presents an easy-to-use tool for SPT that approximates the use of 2D tomographic projections using a ‘pseudo-subtomogram’ data structure, chosen to facilitate implementation within the existing RELION codebase. The examples shown provide solid support for the claims about the efficacy of the approach.
https://doi.org/10.7554/eLife.83724.sa0Introduction
In recent years, electron cryo-microscopy (cryo-EM) has allowed the 3D imaging of an increasing number of biological macromolecules at resolutions sufficient for de novo atomic modelling. This development, originally driven by advances in detector technology, was further facilitated by novel, robust image processing algorithms. In single-particle analysis, images of multiple copies of isolated macromolecular complexes, or particles, that are suspended in random orientations in a thin layer of vitreous water are combined in a 3D reconstruction. Nowadays, many aspects of single-particle analysis workflows can be performed with only minimal human supervision, for example, the detection, extraction, and initial classification of particles in the images (Zivanov et al., 2018; Bepler et al., 2019; Wagner et al., 2019), 3D reconstruction (Zivanov et al., 2018; Punjani et al., 2017), as well as refinement of the optical parameters (Zivanov et al., 2018; Zivanov et al., 2020; Punjani et al., 2017; Tegunov et al., 2021) and per-particle tracking of electron beam-induced motion (Zheng et al., 2017; Zivanov et al., 2019). Many of the algorithms that underlie these modern methods are built on solid statistical foundations that require few tunable parameters. This decreases the need for operator expertise and provides objectivity, as well as robustness, in obtaining optimal structures.
The single-particle approach is, however, limited to investigating isolated protein complexes that are purified to relative homogeneity. To examine these complexes in their crowded physiological environment, electron cryo-tomography (cryo-ET) may be used instead. In the tomographic approach, the sample is tilted multiple times during image acquisition, yielding a so-called tilt series of images from which a 3D tomogram can be computed. In the same manner as single-particle analysis, repeated occurrences of particles in those tomograms can then be aligned and averaged to obtain higher-resolution reconstructions. This process is referred to as subtomogram averaging. Unlike the field of single-particle analysis, labs use many different tools for subtomogram averaging (e.g. Kremer et al., 1996; Nickell et al., 2005; Heumann et al., 2011; Castaño-Díez et al., 2012; Hrabe et al., 2012; Chen et al., 2013; Galaz-Montoya et al., 2016; Sanchez et al., 2019; Chen et al., 2019; Jiménez de la Morena et al., 2022) and many of the tools used require considerable levels of expertise from the operator, often in order to tune parameters that arise from heuristics in the underlying algorithms. This not only provides a barrier for new scientists entering the field, but can also lead to the calculation of suboptimal structures.
Compared to single-particle analysis, subtomogram averaging faces several unique challenges. In addition to estimating the position and orientation of each particle, the algorithm also has to consider the geometry of the tilt series. Typically, this is solved through a set of preprocessing steps that include estimation of contrast transfer function (CTF) parameters and alignment of the tilt series, followed by the reconstruction of, often inconveniently large, tomograms for the entire field of view. Smaller subtomograms, centred around selected particles, are then extracted from the tomograms and used in a separate process of subtomogram alignment and averaging. The separation between tomogram reconstruction and subtomogram averaging can lead to an accumulation of errors, because errors in the CTF estimation or tilt-series alignments are hard to correct. In addition, because the sample cannot be rotated 180° within the microscope, the subtomograms contain empty regions in Fourier space, the so-called missing wedge, which are difficult to deal with in subtomogram averaging (e.g. see Schmid and Booth, 2008; Förster et al., 2008; Bartesaghi et al., 2008; Frank et al., 2012).
A fundamental problem with subtomogram averaging as described above is that it transforms the original 2D image data into 3D subtomograms, which are then used as a substitute for experimental data in the alignment algorithm. RELION-2 introduced the concept of a 3D CTF to describe the transfer of information from the 2D images to the subtomograms, which dealt to some extent with the missing wedge and the loss of information through interpolations in the reconstruction algorithm (Bharat and Scheres, 2016). A drawback of the 3D CTF approach is that it does not deal correctly with the lower resolution regions of Fourier space, where information from different tilt images overlaps. A statistically more attractive approach would be to formulate the optimisation target function directly as a function of the actual 2D images that are measured in the microscope. This has been proposed in an approach called constrained single-particle cryo-ET (Bartesaghi et al., 2012), where individually boxed particles from the tilt-series images are processed as in single-particle analysis, but their relative orientations are kept fixed. A similar approach was also implemented in the program emClarity (Himes and Zhang, 2018). To deal with unknowns in the relative orientations of the particles from the tilt-series images, as well as their CTFs, the program M recently introduced new optimisation approaches that compare reference projections against the 2D particle images (Tegunov et al., 2021). M relies on RELION for alignment and classification of 3D subtomograms that are recalculated from the optimised parameters in M. Nevertheless, this iterative approach allows subtomogram averaging to resolutions that approach those observed for single-particle analysis, even for particles in complex cellular environments (Tegunov et al., 2021).
Here, we describe a new approach to subtomogram averaging in RELION-4.0 that optimises a regularised likelihood function that approximates the direct use of the 2D images of the tilt series. In order to do so at acceptable computational and implementation costs, we have altered the main refinement program in RELION-4.0 to work with so-called pseudo-subtomograms: explicitly constructed sets of 3D data arrays that contain sums of CTF pre-multiplied 2D tilt-series images, together with auxiliary arrays that contain the corresponding sum of squared CTFs and how often each 3D voxel has been observed. Pseudo-subtomograms no longer aim to represent the actual scattering potential of the underlying particles, in the way that conventional subtomograms would. Instead, they represent a convenient way to implement an approximation to the 2D approach within the existing RELION code. Evaluation of the pseudo-subtomograms by RELION-4.0 approximates the likelihood of observing a hypothetical particle in the images of the entire tilt series, given the current model. Using that likelihood as a metric, operations equivalent to those in single-particle analysis can now be performed on tomographic data, for example, 3D initial model generation, 3D classification, or high-resolution refinement. In addition, we describe new methods for optimising parameters of the tilt series that exploit the increased signal-to-noise ratio in the average structure. Besides optimisation of the tilt-series alignment itself, we also describe methods analogous to CTF refinement (Zivanov et al., 2018; Zivanov et al., 2020) for refining descriptors of the optical properties (defocus, astigmatism, and higher-order aberrations) and a method akin to Bayesian polishing (Zivanov et al., 2019) to model beam-induced particle motion throughout the tilt series. Once all these parameters have been optimised, new pseudo-subtomograms can be constructed and the alignment can be repeated. The resulting iterative image processing workflow is similar to existing approaches for single-particle analysis in RELION.
Methods
Particle alignment and averaging
RELION performs maximum a posteriori estimation to find the set of model parameters that maximise the probability of observing the experimental images . Using Bayes’ theorem, we define a regularised likelihood optimisation target function as
where expresses prior information about the model, that is, that the reconstructed map has limited power in Fourier space, and is the likelihood of observing the data given the model. A marginalised likelihood function is used, where one integrates over the unknown alignments of each individual particle. For simplicity, these integrals are omitted from the notations used in this article.
The data model assumes independent Gaussian noise on the Fourier components of the cryo-EM images of individual particles . We therefore write the negative log-likelihood of observing a particle in a hypothetical alignment as a sum over a grid of 2D Fourier pixels :
where is the Fourier transform of the experimental particle image, its contrast-transfer function, denotes the 2D slice out of the 3D Fourier transform of the known map into the view of the particle, and is the noise variance of the frequency band of given by
for a 2D vector and a matrix that respectively encapsulate the particle’s position and orientation, and the evaluation of is achieved through linear interpolation.
In tomography, our aim is to approximate that same likelihood on tilt-series data. The equivalent is a sum over the pixels of the relevant regions of all images from the tilt series:
We model the shifts and rotations as compositions of per-particle and per-image components:
where we keep the per-particle rotation component, , identical for all images in the tilt series, and only vary , the rotational alignment of the tilt-series images. In turn, the tilt-series alignment is shared among all particles in a given tilt image. The per-particle part of the translation is modelled as a 3D vector, , that can vary over different tilt images . This contrasts with single-particle analysis, where beam-induced motion of the particle can be corrected for as a preprocessing step (Li et al., 2013; Scheres, 2014; Zheng et al., 2017; Zivanov et al., 2019), so that each particle is associated with a single 2D translation in a motion-corrected image.
For our pseudo-subtomogram approach, we now approximate the sum over 2D pixels and tilt images in Equation 4 by a sum over 3D voxels in the pseudo-subtomogram:
Here, the data term , the weight term , and the multiplicity term are 3D arrays in the Fourier domain. Together, they constitute a pseudo-subtomogram. They are constructed as follows:
where represents linear interpolation with forward mapping, that is, each 2D Fourier pixel is projected into 3D Fourier space, updating the eight closest voxels.
Ignoring the difference of pre-multiplying the images with their CTFs, Equation 7 aims to be equivalent of Equation 4. The variance is equivalent to , the power of the noise in the individual Fourier components in the 2D images.
We then optimise Equation 1 by expectation-maximisation (Dempster et al., 1977), using Equation 7 to construct the likelihood function and using a prior , based on the expected frequency-dependent power of the signal . This leads to the following iterative algorithm:
where (n) denotes the iteration; the divisions by and in Equation 11 are evaluated element-wise; and and are calculated by averaging over and , respectively, in hollow spheres of radius and thickness 1, described by . The ratio of the terms containing and in Equation 13 corrects the estimate for the power of the signal from the CTF-corrected map by the average CTF2 to account for the fact that the likelihood in Equation 7 was calculated for CTF pre-multiplied images.
Pre-oriented priors
Many proteins are organised in imperfect 2D arrays inside the tomograms, for example, inside membranes or as part of capsid-like structures. Often, the individual protein molecules inside these arrays exhibit limited rotational freedom with respect to the surface normal of the array, although they may be able to rotate freely around that normal. This knowledge is often exploited in subtomogram averaging approaches through local orientational searches, for example, see Förster et al., 2005. This not only accelerates the refinement, as fewer orientations need to be evaluated, it also makes it possible to solve more challenging structures because fewer solutions are allowed. In RELION, local orientational searches are implemented as Gaussian priors on the Cartesian translations and on the three Euler angles that describe rotations (Scheres, 2012). One advantage of using pseudo-subtomogram alignment is that the coordinate system of the pseudo-subtomograms themselves can be chosen arbitrarily. By default, pseudo-subtomograms are created in the same orientation as the tomogram, but the user can choose to orient them in a more meaningful way. For example, by constructing the pseudo-subtomograms with their -axis parallel to the 2D array, using a rotational prior of approximately 90° on the tilt angle will limit the amount of rocking of the particles inside the array, while avoiding singularities in the definition of the Euler angles that occur when the tilt angle is close to 0°.
Tilt-series refinement
Averaging over multiple particles leads to an increased signal-to-noise ratio in the estimated density map . We also implemented procedures that exploit for subsequent re-estimation of parameters that describe the tilt series. These procedures do not require pseudo-subtomograms, but are performed by comparing projections of the density maps directly with the (Fourier) pixels of 2D boxes that are extracted from the tilt-series images, with a sufficient size to hold the CTF-delocalised signal. The various tilt-series parameters are then estimated by minimising the negative log-likelihood as defined in Equation 4, that is, the sum over noise-weighted square differences between the prediction and the observation.
The tilt-series properties that can be refined fall into two broad categories: optical and geometrical. The optical refinement concerns the different parameters of the CTF, while the geometrical refinement aims to optimise the alignment of the tilt series, as well as the beam-induced motion of the individual particles. Both sets of algorithms are closely related to the corresponding single-particle algorithms in RELION: optical-aberration refinement (Zivanov et al., 2018; Zivanov et al., 2020) and Bayesian polishing (Zivanov et al., 2019), respectively. In spite of the similarity between the algorithms, the models that are optimised differ significantly from single-particle analysis. Details of the implementation of the optical and geometrical refinement algorithms are given in Appendix 1. We also note that Bayesian polishing in SPA describes particle motions between individual movie frames. Although our approach for tomography can also consider movie frames, the current implementation uses the same regularisation of particle motions between movie frames within each tilt image as between the movie frames from other tilt images. Because preliminary tests showed limited benefits in considering the movie frames in this manner, only the functionality to model particle motions between the tilt-series images was exposed on the GUI.
CTF refinement for tomographic data in RELION-4.0 includes optimisation of scale factors that model frequency-dependent radiation damage, defocus, astigmatism, and higher-order symmetrical and antisymmetrical aberrations. Although individual particles within a field of view are at distinct defoci in the tilt-series images, their relative defoci are known from the geometry of the tilt series and the known 3D positions of the particles in the tomogram. Therefore, one can efficiently perform defocus estimation in a single pass, considering all particles in a tilt-series image simultaneously. In order to do so, we modified procedures that were developed for higher-order aberration estimation in single-particle analysis (Zivanov et al., 2020), where the information from all particles in each tilt-series image is condensed into two images that are used to estimate a common phase shift (see Appendix).
Similar procedures can also be used to model higher-order symmetrical and antisymmetrical aberrations in the tomographic data. Analogously to our single-particle approach, they are modelled using Zernike polynomials and estimated in the same way. Because the higher-order aberrations are often only a limiting factor at relatively high spatial frequencies, a large number of particles are needed to estimate them reliably. Optimally, higher-order aberrations would thus be estimated globally, over the entire data set, and only for cases that yield high-resolution averages. If aberrations change during data collection, data sets may be split into optics groups, for which aberrations are estimated separately. Typically, the third-order antisymmetrical aberrations are the most important ones, that is, trefoil and axial coma, which can both be caused by a tilted beam. The resolution gains that these optimisations will yield depend on the microscope (mis)alignment. Provided alignment has been performed reasonably well, higher-order aberration correction will probably be most useful for reconstructions that extend beyond 3 Å resolution.
The geometric alignment includes both (rigid) rotational and translational re-alignment of the tilt-series images, as well as the modelling of beam-induced motion of individual particles throughout the tilt series. For the latter, we neglect rotations of the particles, and only model beam-induced translations. By doing so, we can precompute the likelihood of each particle being in each position around its original one, and then look for an alignment that simultaneously maximises the sum of those likelihoods over all tilt-series images and all particles, as well as a prior that ensures spatially coherent motion. This allows us to evaluate the likelihood of a hypothetical particle position by looking up a single interpolated value in an image. In this formulation, the problem becomes equivalent to the Bayesian polishing approach that we originally developed for single-particle analysis, except for the inclusion of a third spatial dimension for the motion.
Results
We tested our approach on three test cases. Appendix 2—table 1 provides experimental details for each of the data sets; Appendix 2—table 2 provides details on the image processing.
HIV-1 immature capsid
We tested the workflow above on the cryo-ET data set that was used to determine the structure of the immature capsid lattice and spacer peptide 1 (CA-SP1) regions of the Gag polyprotein from human immunodeficiency virus 1 (HIV-1) (Schur et al., 2016) (EMPIAR-10164). We used the same subset of five tomograms that were also used to assess the NovaCTF (Turoňová et al., 2017), emClarity (Himes and Zhang, 2018), and Warp (Tegunov and Cramer, 2019) programs. Introducing 3D CTF correction, and using the alignment parameters from the original analysis by Schur et al., NovaCTF reported a resolution of 3.9 Å (Turoňová et al., 2017). The Warp program introduced local and global motion correction in the tilt-series images, as well as optimisation of CTF parameters. The combination of Warp and subtomogram alignment and averaging in RELION-3 led to a resolution of 3.8 Å. A recent application of emClarity led to a reconstruction to 3.3 Å resolution (Ni et al., 2022).
We used tilt-series projections after movie frame alignment from the original analysis (Schur et al., 2016), without any other preprocessing step, along with the tilt-series alignment data, performed with IMOD package (Kremer et al., 1996), and CTF parameters estimation using CTFFIND4 (Rohou and Grigorieff, 2015). We used 12,910 particles from the five tomograms subset, reconstructed an initial reference map using the original published particle alignment, and filtered it to 5 Å. A first alignment in 3D auto-refine, followed by averaging of the initial pseudo-subtomograms, led to a resolution of 3.6 Å. This average was then used for a full cycle of pseudo-subtomogram improvement and realignment. We first applied CTF refinement to optimise the defoci of all particles. This improved the resolution only marginally. Subsequent optimisation of the tilt-series geometry, including modelling local particle motion, improved the resolution to 3.5 Å. Finally, realignment of newly generated pseudo-subtomograms led to a resolution of 3.4 Å. A second cycle of these three steps provided 3.3 Å, while a third cycle converged to 3.2 Å (Figure 1a). Geometrical refinement was performed estimating local particle motion. The consideration of deformations did not show additional improvement. In the first cycle, where improvements in both CTFs and geometry are most obvious, the order of applying those optimisations did not alter the final result for this data set. These data and results are also distributed as part of the subtomogram tutorial in RELION-4.0, as described on https://relion.readthedocs.io/en/release-4.0/. Figure 1—figure supplement 1 shows the improvement in map quality during the iterative refinement process; Figure 1—figure supplement 2 shows a comparison with the 3.3 Å map from emClarity.
Analysis of the complete data set generated a structure at 3.0 Å resolution (Figure 1—figure supplement 1), which is the same resolution obtained using the M and RELION-3 workflow (Tegunov et al., 2021; Figure 1—figure supplement 3), and is likely limited by flexibility and asymmetry in the CA hexamer.
Caulobacter crescentus S-layer
We also applied our approach to thin cellular appendages of C. crescentus bacteria known as stalks, which have previously been imaged using cryo-ET (Bharat et al., 2017). The cell body and cell stalks of C. crescentus cells are covered by a nearly hexagonal, paracrystalline array known as the surface layer (S-layer) (Smit et al., 1992). The structure of the S-layer was solved using a combination of X-ray crystallography, cryo-EM single-particle analysis, and subtomogram averaging, revealing how the S-layer is attached to bacterial cells by an abundant glycolipid called lipopolysaccharide (LPS) (Bharat et al., 2017; von Kügelgen et al., 2020). Previously, cryo-ET of the S-layer, using 110 tilt series collected with a dose-symmetric scheme, yielded 51,866 hexamers of the S-layer. This study used a subtomogram averaging approach that is based on a constrained cross-correlation approach implemented in the AV3 MATLAB suite (Förster and Hegerl, 2007), and which was specifically optimised for the analysis of macromolecules arranged in a lattice (Wan et al., 2017). A 7.4 Å reconstruction of the S-layer was obtained, in which alpha-helices were resolved (Bharat et al., 2017). This reconstruction was improved by application of NovaCTF (Turoňová et al., 2017), leading to a 4.8 Å reconstruction, in which large amino acid residue side chains were resolved (von Kügelgen et al., 2020). Moreover, density for an LPS molecule was observed near the putative LPS-binding residues of the S-layer, in agreement with a cryo-EM single-particle structure of an in vitro reconstituted coplex (von Kügelgen et al., 2020). We used the tilt series after movie frame alignment from the initial analysis (Bharat et al., 2017), along with the tilt-series alignments performed within IMOD (Kremer et al., 1996), CTF parameters from CTFFIND4 (Rohou and Grigorieff, 2015), and the Euler angle assignments and subtomogram coordinates from the original analysis. These parameters were imported into RELION-4.0, followed by multiple cycles of pseudo-subtomogram generation and refinement, analogous to the immature HIV-1 data set described above, leading to a 5.6 Å reconstruction of the S-layer hexamer (Figure 2a). Next, we defined a mask around the central pore of the S-layer, corresponding to the inner domain bound to LPS, to perform focused refinements. Another cycle of pseudo-subtomogram reconstruction, CTF refinement, and refinement within the new mask improved the resolution to 4.4 Å. Accounting for per-particle motions with additional cycles of pseudo-subtomogram improvements and refinements increased the resolution of the central pore to 4.0 Å, and the inner domain of the S-layer to 3.7 Å. Further 3D classification without alignments identified a subset of 42,990 subtomograms that gave a 3.5 Å resolution reconstruction of the inner S-layer.
The 3.5 Å map is in excellent agreement with the single-particle structure of the in vitro reconstituted complex, including the LPS binding site (von Kügelgen et al., 2020; see Figure 2—figure supplement 1). Furthermore, divalent metal ions, known to be tightly bound to the inner S-layer (Matthew, 2021), are resolved (Figure 2b). Surprisingly, at lower isosurface contour levels, we also observed a second LPS binding site (Figure 2c and d). The size and shape of this density agree with the structure of the LPS O-antigen, illustrating how improved subtomogram averaging in RELION-4.0 can help uncover new biology.
Coat protein complex II
Finally, we applied our approach to the Saccharomyces cerevisiae coat protein complex II (COPII), which mediates the transport of newly synthesised proteins from the endoplasmic reticulum (ER) to the Golgi apparatus as part of the secretory pathway. COPII is formed by five proteins that assemble sequentially on the ER membrane to induce remodelling of the bilayer into coated carriers in a process known as COPII budding, while simultaneously selectively recruiting cargo into these budding membrane carriers. COPII budding can be reconstituted in vitro from purified proteins and artificial membranes, to form small, spherical vesicles, or long, straight tubes. Cryo-ET has previously been used to visualise the architecture of COPII on reconstituted tubules (Hutchings et al., 2018; Hutchings et al., 2021). The coat assembles into two concentric layers; the inner layer forms a pseudo helical lattice, which has previously been solved to 4.6 Å resolution using Dynamo-based subtomogram averaging protocols (Castaño-Díez et al., 2012).
We used the tilt series after movie alignment from the initial analysis (Hutchings et al., 2021), along with the tilt-series alignments performed in Dyname (Castaño-Díez et al., 2012) and CTF parameters from CTFFIND4 (Rohou and Grigorieff, 2015). COPII-coated tubes were manually traced in the resulting tomograms, and particles were extracted by randomly oversampling their surface, with approximate initial orientations assigned based on the cylindrical geometry. Dynamo was used for initial alignment of 8× binned subtomograms to define the centre of the particles and the directionality of individual tubes. We then imported the particle coordinates for processing in RELION-4.0 using 3D refinement at 4× and 2× binning. Since we expect inner coat subunits to arrange in a lattice, we cleaned the data set by excluding any subtomograms that did not conform to the expected geometrical relationship with their neighbouring particles. A first 3D refinement of the unbinned data set gave a map at 4.4 Å resolution, which was further improved to 4.2 Å and 4.0 Å by tilt-series frame alignment and CTF refinement, respectively. Two further rounds of 3D refinement, followed by tilt-series frame alignment and CTF refinement, yielded a final map with a resolution of 3.8 Å (Figure 3).
At this resolution, most side chains are visible in the map, enabling us to build and refine an atomic model. The improved model will allow the design of point mutants to precisely disrupt interfaces between coat subunits and test their effects in COPII budding.
Discussion
We formulate the problem of averaging over multiple identical particles in tomographic tilt series in an empirical Bayesian framework that is based on a statistical model that approximates one for two-dimensional experimental data. The Bayesian framework has proven effective in reducing the number of tunable parameters and in obtaining high-quality reconstructions from single-particle data (Fernandez-Leiro and Scheres, 2016). The two-dimensional data model describes the experimental images better than alternative approaches that use 3D reconstructed subtomograms as an intermediate. One example of a problem with the intermediate 3D data model is the need for missing wedge correction, which arises from the observation that the experimental images were acquired, incompletely, in three dimensions. Artefacts related to suboptimal missing wedge correction may affect both alignment and classification of particles. By using an approximation to the 2D data model, missing wedge correction is no longer required. Instead, the problem approaches that of single-particle analysis, where projections from different orientations and of different structural states are sorted out simultaneously. Provided the 3D Fourier transform of the distinct classes is fully sampled through the orientation distribution of the raw particles, likelihood optimisation techniques have been highly successful in tackling this problem in single-particle analysis (Scheres et al., 2007; Fernandez-Leiro and Scheres, 2016).
In practice, the implementation in RELION-4.0 does not use stacks of 2D projection images as input for the refinement program that performs alignment and classification. Instead, the concept of 3D pseudo-subtomograms is introduced, where the tilt-series images are Fourier transformed, pre-multiplied with their CTF, and inserted as a slice into a 3D Fourier volume according to the best current estimates for the tilt-series geometry. The use of 3D pseudo-subtomograms allowed reusing existing code for subtomogram averaging in RELION, while input stacks of 2D images would have required significant software development efforts. Nevertheless, in the future we might still choose to implement a true 2D version of the code, which would be more efficient, both in terms of processing time and disk storage requirements. In cases where the number of tilt images is small in comparison to the box size, fewer Fourier pixels need to be examined in a stack of 2D images than in a pseudo-subtomogram, with a corresponding decrease in processing time. Moreover, the likelihood calculation from the 3D pseudo-subtomogram approach requires separate storage of the accumulated squares of the CTFs, and the corresponding multiplicity terms. In contrast, in the 2D approach, only the 2D images need to be stored, as CTF parameters can be calculated on-the-fly and there is no need for a multiplicity term, giving a corresponding decrease in storage requirements. However, if one were to collect tilt series with very fine angular increments or in a continuous manner (Chreifi et al., 2019), then the current implementation may still be preferable.
Besides the alignment and classification of individual particles, the methods described here also deal with re-estimation of parameters that describe the optical and geometrical features of the tilt series. As soon as an initial average structure has been obtained, its increased signal-to-noise ratio can be exploited to determine these parameters more accurately than what is possible from the raw tilt-series images alone. The implementations in RELION-4.0 again follow those previously implemented for single-particle analysis, where CTF refinement is used for re-estimation of the tilt-series images CTFs, and algorithms akin to Bayesian polishing are used to re-estimate the tilt-series alignment, as well as the movement of individual particles throughout the tilt-series acquisition process. As better tilt-series parameters will allow better pseudo-subtomograms, particle alignment and classification are iterated with the optimisation of the tilt-series parameters.
Similar tilt-series and CTF optimisation approaches have been implemented in the program M (Tegunov et al., 2021). Compared to M, RELION-4.0 uses computationally more efficient algorithms; M uses GPUs to accelerate the calculations. In both tomography and SPA, RELION-4.0 only models beam-induced translations of the particles, whereas M also models beam-induced rotations. Since SPA routinely reaches 2 Å resolutions without modelling beam-induced rotations, we assumed that the effect of rotations of individual particles throughout the tilt series is not large enough to warrant their correction at typical tomography resolutions. In cases where the data do allow for better than 2 Å resolutions, M could still be used to correct for beam-induced rotations in a postprocessing step, following alignment and classification of the individual particles in RELION. It is likely that adaptation of M, in order to function with the pseudo-subtomograms proposed here, would lead to increased synergy between the two programs. In the meantime, external tools to convert from M parameters to RELION-4.0 are already available (https://github.com/joton/reliontomotools; Zivanov, 2022 copy archived at swh:1:rev:bfa43828876ceb77bed0c7eb72f794c79c9de5e6).
Besides the reduction in tunable parameters that is characteristic of the Bayesian approach, its uptake by researchers that are new to the field is further facilitated through the implementation of a graphical user interface. This interface is already widely used for single-particle analysis and has been extended for the processing of tomographic data in RELION-4.0. Apart from the calculations that will be familiar to users of single-particle analysis, for example, 3D classification, 3D initial model generation, and 3D auto-refinement, the new interface also provides convenient access to the tomography-specific versions for CTF refinement and Bayesian polishing, as well as preprocessing operations to calculate the pseudo-subtomograms. However, tilt-series alignment, tomogram reconstruction, and particle picking are not yet part of the RELION workflow. Efforts to also implement solutions for those steps in a single tomography processing pipeline are ongoing and will be part of future RELION releases. Meanwhile, current import procedures rely on specific preprocessing operations in IMOD (Kremer et al., 1996), and particle coordinate conversion tools to use in RELION-4.0 are available for a range of third-party software packages (Pyle et al., 2022). To further facilitate the uptake of this new software by the community, we have provided an online tutorial that uses the publicly available HIV-1 immature capsid data set to describe and illustrate all steps necessary to obtain the results described in Figure 1.
In summary, we introduce new methods for subtomogram averaging to resolutions that are sufficient for de novo atomic modelling and increase the accessibility of this emerging technique. We envision that our methods will allow more researchers to calculate better structures from tomographic data, which will aid the next revolution in structural biology, where macromolecular complexes are imaged, not in isolation, but in their biologically relevant environment.
Appendix 1
CTF refinement
CTF refinement in RELION-4.0 optimises the following parameters: scale, defocus, astigmatism, and higher-order (even and odd) optical aberrations. Since, save for the difference in defocus, the same CTF needs to be valid for an entire micrograph of particles, similar optimisations can be applied as in our single-particle algorithms. All the relevant information is first consolidated into a minimal form using linear transformations, and the final, typically non-linear, optimisation is then performed on that minimal form.
We formulate the CTF for tilt-series frame of particle as follows:
where describes the overall scaling factor, the empirical radiation-damage model as defined by Grant and Grigorieff, 2015, the symmetrical phase delay component, and the antisymmetrical one. Note that only varies between particles. This is because it contains the quadratic defocus term that depends on the position of the particle. The phase delays are parametrised the same way as in single-particle analysis in RELION-3 – as a combination of explicitly named low-order terms and higher-order Zernike polynomials:
As before, the astigmatic-defocus matrix is decomposed into a defocus term and two linear astigmatism terms, a1 and a2, while describes the spherical aberration of the microscope, a constant phase offset (owing to amplitude contrast and a phase plate, if one is used), is the wavelength of the electron, and are the higher-order even Zernike terms. One key difference to single-particle analysis is that the defocus term is no longer a free parameter for each particle, but it instead depends on the already known 3D position of the particle. Therefore, in tomography, the defocus term is only estimated once per tilt image, and all the particles contribute to that estimate.
The scaling factor is estimated by computing the following two sums for each micrograph and dividing them (the † symbol indicates complex conjugation):
Note that the used in Equation 19 is missing its scale factor:
Alternatively, we also allow the user to fit the parameters of Lambert’s extinction model to the data instead, assuming perfectly flat samples of constant thickness. In that case, the CTF scale in image of tomogram is expressed as a function of the beam luminance , sample normal , and optical sample thickness :
If this option is used, then the CTF scales of all the tilt series in the data set are estimated together. The beam luminance is modelled globally, while the sample thickness and normal are allowed to differ among tomograms, but not between the images of a tilt series. The vector points in viewing direction of tilt image . Note that this model does not allow for separating the geometrical sample thickness from its extinction factor, so we can only estimate the product of the two. Also, the ice normal is required to be perpendicular to the estimated tilt axis of the tilt series since its component pointing in the direction of the axis is indistinguishable from an increase in ice thickness or opacity. This global optimisation is performed using the sums and computed in Equations 18 and 19, where the subscript indicates tilt-series . This is done by finding a global value of and values of and for all tomograms that produce which minimise the following quantity and thus maximise the overall likelihood in Equation 4:
To perform defocus estimation efficiently, we apply the optimisations we originally developed for the estimation of higher-order aberrations in single-particle analysis (Zivanov et al., 2020). It allows us to determine a collective offset to for a large set of particles that all have different values of . Specifically, it allows the change to the log-likelihood arising from changing the value of at any Fourier pixel to be expressed as a pair of 2D images, independently of the number of particles. Therefore, each pixel of each particle only needs to be considered once. After that, the log-likelihood can be evaluated by iterating over the pixels of a single image.
In single-particle analysis, this approach is used to estimate the higher-order aberrations that are shared among all the particles in a data set. In tomography, we also use this approach to condense the information from all the particles in a tilt image (all of which exhibit slightly different defoci), into two such images, and to then determine the optimal change to efficiently using a nonlinear algorithm.
The two condensed images and that we compute are the same as the ones in single-particle analysis, except for the inclusion of the noise power . The definitions are repeated here for the sake of completeness. Note that each pixel of contains a real symmetrical matrix and each pixel of a vector:
where describes the point on the unit circle corresponding to the initial phase angle , which is given by the initial CTF parameters:
The predicted 2D images contain the effect of the initial CTFs, except for the symmetrical aberration:
The vector-valued condensed image describes the most likely phase shift for each pixel , expressed as a point on a circle, while the matrix-valued one, , describes the anisotropic weight of that information. With these two condensed images computed for a given tilt image , the change to the likelihood defined in Equation 4 resulting from the change to the phase delay at any pixel can be expressed as a quadratic form. Therefore, we look for a change to the astigmatic-defocus matrices which produces phase delays that minimise that quadratic form:
where the per-pixel error is given by the deviation from the optimal phase shift :
As an alternative to fitting independently for each tilt-series image, our program also allows the user to apply an regulariser to the of the different images in the same series. In that case, the sum in Equation 27 runs over all the pixels of all the tilt-series frames . This helps to stabilise the CTFs of the higher tilts, but it risks impairing the estimates of the CTFs of the more important lower tilts. Formally, this is done by minimising the following cost:
Since the early tilt-series images carry more information than the later ones, their values in are typically significantly greater. Therefore, using this formulation, they automatically assume a greater weight in the estimation. The optimal weight for the regulariser itself, , cannot be measured from the data. Its optimal value depends on how reproducible the defocus values are for each of the tilt-series images, which in turn depends on the microscope setup, such as the stability of the stage.
Geometric refinement
Analogously to Bayesian polishing, the log-likelihood of a particle being observed at a position is given by twice its cross-correlation with the predicted image:
where IFT stands for inverse Fourier transform.
To keep the problem differentiable, the cross-correlation CC is always accessed using cubic interpolation. After the inverse Fourier transformation, each such cross-correlation table is cropped to a smaller size to make storing many such tables feasible, and the memory throughput efficient. The size of the tables can be controlled by the user, and should be set to the maximal positional error expected in the data set.
The geometrical model that is optimised this way projects points in the tomogram to 2D positions in each tilt image:
The initial linear projection is obtained by multiplying with a matrix , and then optionally shifted by the non-linear image distortion . The cost that is being minimised consists of the sum over all (negative) cross-correlation values of all particles in all images plus all regularisers for all regularised parameters:
Although our framework supports arbitrary projection matrices , our optimisation algorithm only looks for orthogonal rotations to the initial projection matrix. This is achieved by parametrising that rotation using Tait–Bryan angles, not Euler angles. The disadvantage of Euler angles is that they are gimbal locked in the initial configuration where all three angles are zero, that is, the first and third angles refer to the same axis. The rigid alignment of the tilt image is never regularised because we do not assume to have any prior information on it.
The distortion field can take different forms. We have implemented models that express the distortion using the Fourier basis, a cubic spline basis, and an affine linear one. The intended purpose of these deformations is to model distortions of the image that arise at the image forming lenses at the bottom of the optical system. An imperfect calibration of these lenses is likely to go unnoticed as long as the microscope is only used for single-particle analysis because the same particle is never observed at starkly different positions during the collection of a single-particle movie. In tomography, a given particle may appear at any image position in any tilt image, so arbitrary deformations to the 2D image become relevant. We expect these deformations to be stable over time, so the intended purpose of the deformation field is to model only one such deformation per tilt series. Optionally, we also allow the user to instead model a different deformation for each tilt image, but we have not encountered any data sets where this has produced an improvement. The deformation fields are optionally regularised by penalising the squared coefficients of the respective model. This limits the extent of deformation, and it forces the system to explain changes in position through particle motion, rather than image deformations.
The quantity that we do expect to change during the collection of the tilt series is instead the 3D position of the particle, . Analogously to Bayesian polishing, we model this change as motion over time. The position in image is given by the sum over its per-tilt-series-frame velocities up to that point. Note that the velocity vector refers to the motion between images and
It is important to note that the tilt images are implicitly assumed to be ordered chronologically. In practice, this is usually not given, so the images are reordered by the program based on the cumulative radiation dose of each image.
As in Bayesian polishing, the motion vectors themselves are expressed in a collective coordinate system for all the particles. This allows the spatial smoothness of a hypothetical motion field to be evaluated and used as a prior. For a more detailed derivation, we refer to the paper on Bayesian polishing for single-particle analysis (Zivanov et al., 2019). The formal details will be given in the following for the sake of completeness and to highlight differences to the original formulation.
The collective coordinate system for particle motion is obtained through a low-rank approximation of a Gaussian process. This is done by constructing and then diagonalising the covariance matrix for a set of initial particle positions (where is the number of particles). The entries of contain the value of the following square-exponential covariance function for each pair of particles and :
Optionally, the user can instead also use the original formulation without the square inside the exponential:
The former option forces particles in immediate proximity to move more similarly, but it allows for a greater discrepancy at greater distances. Both the single-particle and the tomography implementations allow the user to choose either function, but the default has changed from the latter to the former in tomography. This choice was motivated by both empirical observations and the fact that the square-exponential kernel produces fewer meaningful deformation components, which speeds up the optimisation for tomograms with hundreds of particles.
We perform a singular-value decomposition of the covariance matrix ,
which allows us to construct the coordinate system as follows:
where is the singular value and the corresponding singular vector. Basis vectors with small are discarded here to speed up the optimisation. Let represent the number of remaining basis vectors. In this coordinate system, the set of all particle velocities in a tilt image can be expressed as , where is a coefficient matrix that encodes the velocity in three spatial dimensions. Note that the same basis is shared between all three dimensions and all tilt-series frames. In this coordinate system, the negative log-likelihood of a configuration of particle velocities is given by the Frobenius norm of the coefficient vector, , that is, the sum of the squares of its entries. Therefore, the motion regulariser takes a simple form:
The acceleration regulariser that would penalise changes in velocity from one tilt-series frame to the next in single-particle analysis has been omitted from tomography. This is because, unlike a single-particle movie, the tilt images are not collected in one continuous exposure. Since they are being exposed individually, there is no reason to assume that the particle velocities will be continuous between them. Two further differences to the original Bayesian polishing are hidden in the notation: the covariance is now based on the 3D distances between the particles, and the coefficient matrix contains three columns instead of two.
As in the original Bayesian polishing approach, the complete alignment of the tilt series is performed by finding parameters that minimise from Equation 36 using L-BFGS (Liu and Nocedal, 1989). The set of parameters that are being optimised always includes the three Tait–Bryan angles for each tilt image and the set of initial particle positions. The latter are essential because all the information we have about their 3D positions is derived from the tilt images themselves, so changing the alignment requires the particles to be able to shift to more likely positions. Estimating the image deformations and particle motion is optional. If they are being estimated, then a set of deformation coefficients is fitted either to each tilt image or to each tilt series, while a set of motion coefficients is fitted to each image transition.
In addition to this local, L-BFGS-based refinement, we also offer two methods to align only the 2D shifts of all tilt images globally. This means that instead of trying to obtain the optimal alignment through small changes to the initial one, we instead look for the best possible image shift overall, keeping all other parameters constant. This is helpful when individual tilt images are so badly aligned that a local optimisation cannot converge to the globally optimal position. Note that the initially assumed angles are rarely as incorrect as the image shifts since the angles can be controlled more effectively through the experimental setup.
There are two variants to this method. If the sample contains few particles per tomogram, then the best results are obtained by predicting an entire micrograph and computing its cross-correlation with the original one. The maximum value in that large cross-correlation image then indicates the optimal image shift. This approach can in theory deal with arbitrarily large misalignments. If the sample is dense, however, then this whole-micrograph approach can fail. In that case, better results are obtained by instead adding up the small, per-particle cross-correlation images defined in Equation 31, and finding the maximum in that sum. This latter approach can only correct for misalignments smaller than half the box size of the particle, and it often produces inferior results on samples with few particles per tomogram.
Appendix 2
Data availability
We have only used previously published cryo-EM data sets for testing our software. Reconstructed maps and atomic models generated in this study have been submitted to the EMDB and PDB, with entry codes as indicated in Table 1.
-
EmpiarID EMPIAR-10164. Cryo-electron tomography of immature HIV-1 dMACANC VLPs.
References
-
Real-Space refinement in phenix for cryo-EM and crystallographyActa Crystallographica. Section D, Structural Biology 74:531–544.https://doi.org/10.1107/S2059798318006551
-
Classification and 3D averaging with missing wedge correction in biological electron tomographyJournal of Structural Biology 162:436–450.https://doi.org/10.1016/j.jsb.2008.02.008
-
Fast and accurate reference-free alignment of subtomogramsJournal of Structural Biology 182:235–245.https://doi.org/10.1016/j.jsb.2013.03.002
-
Rapid tilt-series acquisition for electron cryotomographyJournal of Structural Biology 205:163–169.https://doi.org/10.1016/j.jsb.2018.12.008
-
ISOLDE: a physically realistic environment for model building into low-resolution electron-density mapsActa Crystallographica. Section D, Structural Biology 74:519–530.https://doi.org/10.1107/S2059798318002425
-
Maximum likelihood from incomplete data via the em algorithmJournal of the Royal Statistical Society 39:1–22.https://doi.org/10.1111/j.2517-6161.1977.tb01600.x
-
Structure determination in situ by averaging of tomogramsMethods in Cell Biology 79:741–767.https://doi.org/10.1016/S0091-679X(06)79029-X
-
Classification of cryo-electron sub-tomograms using constrained correlationJournal of Structural Biology 161:276–286.https://doi.org/10.1016/j.jsb.2007.07.006
-
Computational separation of conformational heterogeneity using cryo-electron tomography and 3D sub-volume averagingJournal of Structural Biology 178:165–176.https://doi.org/10.1016/j.jsb.2012.01.004
-
Alignment algorithms and per-particle CTF correction for single particle cryo-electron tomographyJournal of Structural Biology 194:383–394.https://doi.org/10.1016/j.jsb.2016.03.018
-
Clustering and variance maps for cryo-electron tomography using wedge-masked differencesJournal of Structural Biology 175:288–299.https://doi.org/10.1016/j.jsb.2011.05.011
-
PyTom: a python-based toolbox for localization of macromolecules in cryo-electron tomograms and subtomogram analysisJournal of Structural Biology 178:177–188.https://doi.org/10.1016/j.jsb.2011.12.003
-
ScipionTomo: towards cryo-electron tomography software integration, reproducibility, and validationJournal of Structural Biology 214:107872.https://doi.org/10.1016/j.jsb.2022.107872
-
Computer visualization of three-dimensional image data using imodJournal of Structural Biology 116:71–76.https://doi.org/10.1006/jsbi.1996.0013
-
SerialEM: a program for automated tilt series acquisition on tecnai microscopes using prediction of specimen positionMicroscopy and Microanalysis 9:1182–1183.https://doi.org/10.1017/S1431927603445911
-
Tom software toolbox: acquisition and analysis for electron tomographyJournal of Structural Biology 149:227–234.https://doi.org/10.1016/j.jsb.2004.10.006
-
CTFFIND4: fast and accurate defocus estimation from electron micrographsJournal of Structural Biology 192:216–221.https://doi.org/10.1016/j.jsb.2015.08.008
-
ConferenceFast Alignment of Limited Angle Tomograms by projected Cross Correlation2019 27th European Signal Processing Conference (EUSIPCO.https://doi.org/10.23919/EUSIPCO.2019.8903041
-
RELION: implementation of a bayesian approach to cryo-EM structure determinationJournal of Structural Biology 180:519–530.https://doi.org/10.1016/j.jsb.2012.09.006
-
Methods for aligning and for averaging 3D volumes with missing dataJournal of Structural Biology 161:243–248.https://doi.org/10.1016/j.jsb.2007.09.018
-
Real-Time cryo-electron microscopy data preprocessing with warpNature Methods 16:1146–1152.https://doi.org/10.1038/s41592-019-0580-y
-
Efficient 3d-ctf correction for cryo-electron tomography using novactf improves subtomogram averaging resolution to 3.4åJournal of Structural Biology 199:187–195.https://doi.org/10.1016/j.jsb.2017.07.007
-
Ctf determination and correction for low dose tomographic tilt seriesJournal of Structural Biology 168:378–387.https://doi.org/10.1016/j.jsb.2009.08.016
-
SoftwareReliontomotools, version swh:1:rev:bfa43828876ceb77bed0c7eb72f794c79c9de5e6Software Heritage.
Article and author information
Author details
Funding
UK Research and Innovation (MC_UP_A025_1013)
- Sjors HW Scheres
UK Research and Innovation (MC_UP_1201/16)
- John AG Briggs
European Research Council (ERC-CoG-2014 grant 648432)
- John AG Briggs
European Research Council (ERC-StG-2019 grant 852915)
- Giulia Zanetti
Swiss National Science Foundation (205321_179041/1)
- Daniel Castaño-Díez
UK Research and Innovation (BBSRC grant BB/T002670/1)
- Giulia Zanetti
European Research Council (ERC-AdG-2015 grant 692726)
- Jasenko Zivanov
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
We are grateful to the MRC-LMB EM facility for help with data acquisition and to Jake Grimmett, Toby Darling, and Ivan Clayson for help with high-performance computing. This work was funded by the UK Research and Innovation (UKRI) Medical Research Council (MC_UP_A025_1013 to SHWS; and MC_UP_1201/16 to JAGB), the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program (ERC-CoG-2014, grant 648432, MEMBRANEFUSION to JAGB and ERC StG-2019, grant 852915 CRYTOCOP to GZ); the Swiss National Science Foundation (grant 205321_179041/1 to DC-D), the Max Planck Society (to JAGB) and the UKRI Biotechnology and Biological Sciences Research Council (grant BB/T002670/1 to GZ). TAMB is a recipient of a Sir Henry Dale Fellowship, jointly funded by the Wellcome Trust and the Royal Society (202231/Z/16/Z). JZ was partially funded by the European Union’s Horizon 2020 research and innovation program (ERC-ADG-2015, grant 692726, GlobalBioIm to Michael Unser). TAMB thanks the Vallee Research Foundation, the Leverhulme Trust, and the Lister Institute of Preventative Medicine for support.
Copyright
© 2022, Zivanov et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 3,980
- views
-
- 618
- downloads
-
- 150
- citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Structural Biology and Molecular Biophysics
Polycystin-1 (PC1) is the protein product of the PKD1 gene whose mutation causes autosomal dominant Polycystic Kidney Disease (ADPKD). PC1 is an atypical G protein-coupled receptor (GPCR) with an autocatalytic GAIN domain that cleaves PC1 into extracellular N-terminal and membrane-embedded C-terminal (CTF) fragments. Recently, activation of PC1 CTF signaling was shown to be regulated by a stalk tethered agonist (TA), resembling the mechanism observed for adhesion GPCRs. Here, synthetic peptides of the first 9- (p9), 17- (p17), and 21-residues (p21) of the PC1 stalk TA were shown to re-activate signaling by a stalkless CTF mutant in human cell culture assays. Novel Peptide Gaussian accelerated molecular dynamics (Pep-GaMD) simulations elucidated binding conformations of p9, p17, and p21 and revealed multiple specific binding regions to the stalkless CTF. Peptide agonists binding to the TOP domain of PC1 induced close TOP-putative pore loop interactions, a characteristic feature of stalk TA-mediated PC1 CTF activation. Additional sequence coevolution analyses showed the peptide binding regions were consistent with covarying residue pairs identified between the TOP domain and the stalk TA. These insights into the structural dynamic mechanism of PC1 activation by TA peptide agonists provide an in-depth understanding that will facilitate the development of therapeutics targeting PC1 for ADPKD treatment.
-
- Structural Biology and Molecular Biophysics
BTB (bric-a-brack, Tramtrack, and broad complex) is a diverse group of protein-protein interaction domains found within metazoan proteins. Transcription factors contain a dimerizing BTB subtype with a characteristic N-terminal extension. The Tramtrack group (TTK) is a distinct type of BTB domain, which can multimerize. Single-particle cryo-EM microscopy revealed that the TTK-type BTB domains assemble into a hexameric structure consisting of three canonical BTB dimers connected through a previously uncharacterized interface. We demonstrated that the TTK-type BTB domains are found only in Arthropods and have undergone lineage-specific expansion in modern insects. The Drosophila genome encodes 24 transcription factors with TTK-type BTB domains, whereas only four have non-TTK-type BTB domains. Yeast two-hybrid analysis revealed that the TTK-type BTB domains have an unusually broad potential for heteromeric associations presumably through a dimer-dimer interaction interface. Thus, the TTK-type BTB domains are a structurally and functionally distinct group of protein domains specific to Arthropodan transcription factors.