Hostname: page-component-586b7cd67f-t8hqh Total loading time: 0 Render date: 2024-11-24T03:55:18.613Z Has data issue: false hasContentIssue false

Multiple-image super-resolution of cryo-electron micrographs based on deep internal learning

Published online by Cambridge University Press:  09 February 2023

Qinwen Huang
Affiliation:
Department of Computer Science, Duke University, Durham, North Carolina, USA
Ye Zhou
Affiliation:
Department of Computer Science, Duke University, Durham, North Carolina, USA
Hsuan-Fu Liu
Affiliation:
Department of Biochemistry, Duke University School of Medicine, Durham, North Carolina, USA
Alberto Bartesaghi*
Affiliation:
Department of Computer Science, Duke University, Durham, North Carolina, USA Department of Biochemistry, Duke University School of Medicine, Durham, North Carolina, USA Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina, USA
*
*Corresponding author. E-mail: [email protected]

Abstract

Single-particle cryo-electron microscopy (cryo-EM) is a powerful imaging modality capable of visualizing proteins and macromolecular complexes at near-atomic resolution. The low electron-doses used to prevent radiation damage to the biological samples, however, result in images where the power of the noise is 100 times greater than the power of the signal. To overcome these low signal-to-noise ratios (SNRs), hundreds of thousands of particle projections are averaged to determine the three-dimensional structure of the molecule of interest. The sampling requirements of high-resolution imaging impose limitations on the pixel sizes that can be used for acquisition, limiting the size of the field of view and requiring data collection sessions of several days to accumulate sufficient numbers of particles. Meanwhile, recent image super-resolution (SR) techniques based on neural networks have shown state-of-the-art performance on natural images. Building on these advances, here, we present a multiple-image SR algorithm based on deep internal learning designed specifically to work under low-SNR conditions. Our approach leverages the internal image statistics of cryo-EM movies and does not require training on ground-truth data. When applied to single-particle datasets of apoferritin and T20S proteasome, we show that the resolution of the 3D structure obtained from SR micrographs can surpass the limits imposed by the imaging system. Our results indicate that the combination of low magnification imaging with in silico image SR has the potential to accelerate cryo-EM data collection by virtue of including more particles in each exposure and doing so without sacrificing resolution.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press

Impact Statement

This research paper describes an image super-resolution method that improves the quality of single-particle cryo-EM images and results in higher resolution reconstructions of protein structures. By leveraging internal image statistics of cryo-EM movies, we propose to use a deep-learning framework that is self-supervised and does not require training on ground-truth images. This work is addressed to people working at the interface between biological imaging and computer vision. The proposed approach is validated on real single-particle cryo-EM datasets and we show that the resolution of 3D structures can surpass the limits imposed by the imaging system. This advance can potentially accelerate cryo-EM data collection and pave the way for improving the throughput of structure determination using single-particle cryo-EM.

1. Introduction

Single-particle cryo-electron microscopy (cryo-EM) is a powerful imaging modality used to determine the three-dimensional structure of proteins and macromolecular complexes at near-atomic resolution(Reference Bendory, Bartesaghi and Singer1Reference Singer and Sigworth4). By combining hundreds of thousands of noisy projection images of identical copies of the molecule of interest taken from different orientations, 3D reconstructions can be obtained where molecular level details can be visualized. While the signal contributed by each individual projection is very weak, averaging the contribution from many particles allows to overcome the extremely low signal-to-noise ratios (SNRs). Acquiring such large datasets, however, is time consuming and can take several days to complete, becoming a bottleneck in the structure determination pipeline. One strategy to improve the throughput of data collection is to increase the size of the field of view by acquiring images at lower magnification. For example, doubling the pixel size will increase the imaging area fourfold, resulting in four times as many particles per exposure. While this strategy will in principle limit the attainable resolution due to the coarser spatial sampling, it does not mean a permanent loss of high-frequency information because images are collected in movie-mode and motion exists between the acquired frames. Super-resolution (SR) is a widely studied problem in the field of natural image photography. In recent years, the SR field has focused on single-image super-resolution (SISR) where a high-resolution (HR) image is obtained from a single low-resolution (LR) input. However, SISR is mostly limited to adding high-frequency information from learned image priors. On the other hand, multi-image super-resolution (MISR) aims to reconstruct the original HR signal using multiple LR images. When sub-pixel motion exists, each LR image provides different LR samples of the underlying higher-resolution scene. MISR approaches exploit this additional information to achieve the recovery of the HR signal. This same principle presents the opportunity of using MISR to overcome resolution constraints in single-particle cryo-EM.

Many SR algorithms based on machine learning have been proposed that achieve state-of-the-art (SotA) performance on natural images. Here, we set out to explore whether these strategies can be extended to work on low-SNR images such as the noisy projections obtained in single-particle cryo-EM. We propose to use a MISR algorithm that utilizes deep internal learning as initially presented in the zero-shot super-resolution (ZSSR) framework(Reference Shocher, Cohen and Irani5) for SISR. Indeed, learning based on internal statistics has shown promising results for cryo-EM image denoising(Reference Krull, Buchholz and Jug6,Reference Palovcak, Asarnow, Campbell, Yu and Cheng7) . Internal data repetition at different frequency levels occurs naturally in single-particle cryo-EM: first, movie frames that are acquired earlier in the exposure contain more high-frequency signal than movie frames acquired later in the exposure; and second, each exposure area contains hundreds of naturally occurring projections of the same macromolecule. Our algorithm, cryo-zero-shot super-resolution (cryo-ZSSR), exploits cross scale internal data repetition in noisy movies obtained from frozen hydrated protein samples imaged under an electron microscope. We train a movie-specific neural network that takes in multiple frames from each LR movie and reconstructs a single 2× SR image per exposure area. These SR images are then fed into the standard cryo-EM data processing pipeline and used to generate a 3D reconstruction of the protein of interest (Figure 1). Our SR algorithm is self-supervised and does not require training on ground-truth HR data, which is nevertheless not available in cryo-EM. We evaluate the performance of our approach on two real datasets of apoferritin and T20S proteasome and show that the SR images can produce higher resolution reconstructions compared to the LR data. Used in combination with low-magnification imaging, our approach can be used to accelerate data collection while still producing high-quality 3D reconstructions.

Figure 1. Super-resolution single-particle structure determination pipeline and example micrographs from two cryo-EM datasets. (a) Cryo-EM movies are collected using a large pixel size and subsequently upsampled by a factor of 2 using our self-supervised cryo-zero-shot super-resolution (cryo-ZSSR) approach. Super-resolved micrographs are then fed into the standard single-particle reconstruction workflow producing three-dimensional structures at resolutions surpassing the Nyquist rate. The 2× upsampling factor effectively results in a 4x speedup in the rate of data acquisition allowing the collection of four times more particles in the same amount of time. (b) Left: Example of a single raw frame from a movie of apoferritin from EMPIAR-10146 collected at 2  $ {e}^{-}/ $ Å2. Right: Average of 50 frames corresponding to a total dose of 100  $ {e}^{-}/ $ Å2. (c) Left: Example of a single raw frame from a movie of T20S proteasome from EMPIAR-10025 collected at 1.4  $ {e}^{-}/ $ Å2. Right: Average of 38 frames corresponding to a total dose of 53.2  $ {e}^{-}/ $ Å2.

2. Related Work

There has been a tremendous amount of work on image SR in the past couple of decades, both for SISR and for MISR. Traditional SISR algorithms can be divided into three categories: interpolation-based methods (e.g., bilinear, Lanczos kernels, etc.), reconstruction-based methods(Reference Zeyde, Elad and Protter8,Reference Schulter, Leistner and Bischof9) , and example-based methods(Reference Freeman, Jones and Pasztor10,Reference Timofte, De and Gool11) . Interpolation-based algorithms are straightforward and fast but suffer from limited accuracy. Reconstruction-based methods make use of prior knowledge about images by restricting the possible solution space to generate high-quality images. However, these methods are usually time-consuming and their performance degrades rapidly as the upsampling factors increase. Example-based methods usually leverage machine learning to analyze relationships between the LR and its corresponding HR counterparts from training examples. More recently, deep-learning-based SISR algorithms(Reference Ulyanov, Vedaldi and Lempitsky12,Reference Lim, Son, Kim, Nah and Lee13) built upon example-based learning received wide attention and demonstrated great superiority compared to more traditional approaches.(Reference Ulyanov, Vedaldi and Lempitsky12,Reference Lim, Son, Kim, Nah and Lee13) Generative adversarial network (GAN)(Reference Goodfellow, Pouget-Abadie, Mirza, Xu, Warde-Farley, Ozair, Courville and Bengio14) based deep-learning approaches, such as Super Resolution Generative Adversarial Network (SRGAN)(Reference Ledig, Theis, Huszár, Caballero, Aitken, Tejani, Totz, Wang and Shi15), and Photo Sampling via Latent Space Exploration (PULSE)(Reference Menon, Damian, Hu, Ravi and Rudin16) are able to perform extremely well on certain natural and facial images. However, most of these networks are trained in a supervised manner and require knowledge of ground-truth images. Results reconstructed using GANs, even though visually appealing, tend to generate information that does not exist in the actual HR pictures. In addition, the formation of training datasets, specifically the LR images, are usually generated using predetermined ideal processes (e.g., bicubic downsampling, Gaussian blurring, etc.). In reality, LR images rarely follow this model, resulting in poor performance of previously mentioned SotA methods. To overcome this limitation, ZSSR was proposed by Shocher et al.(Reference Shocher, Cohen and Irani5). Instead of relying on prior training, this method exploits the internal recurrence of information inside a single image and trains an image-specific CNN on examples extracted solely from the input image itself. Thus, ZSSR is able to achieve and outperform SotA methods on LR images generated under nonideal downsampling models.

Turning to the problem of MISR, which involves the extraction of information from many LR observations of the same scene to reconstruct HR images, the earliest method developed by Tsai and Huang(Reference Tsai and Huang17) used a frequency domain technique to improve the spatial resolution of images by combining multiple LR images with sub-pixel accuracy displacements. Later on, other spatial domain MISR methods were proposed that include nonuniform interpolation such as adaptive kernel regression(Reference Takeda, Farsiu and Milanfar18), Bayesian modeling algorithms(Reference Tipping and Bishop19), and projection onto convex sets (POCSs)(Reference Fan, Wu, Li and Ma20). Most of these SR methods assume a priori knowledge about the motion model, blur kernel, and noise level. However, there are many cases where the actual image degradation process is unknown.

For this reason, many blind SR image reconstruction methods were developed. These methods usually involve two steps: (a) motion estimation for LR images, followed by (b) simultaneous estimation of both the HR image and the blurring function. Since separating image registration and HR estimation tends to produce sub-optimal results, some researchers have developed methods that jointly estimate motion parameters and the HR reconstruction(Reference Robinson, Toth, Lo and Farsiu21). Recently, similar to SISR problems, deep-learning-based methods have been proposed to simultaneously solve video SR and MISR problems. Most of the existing work is focused on video SR, such as frame recurrent SR(Reference Sajjadi, Vemulapalli and Brown22) which utilizes previous inferred HR frames to super-resolve subsequent frames in an end-to-end trainable framework that incorporates both frame registration and HR estimation. More recently, several deep-learning-based algorithms are proposed to solve MISR problems in satellite imaging and burst photography. HighRes-net(Reference Deudon, Kalaitzis, Goytom, Arefin, Lin, Sankaran, Michalski, Kahou, Cornebise and Bengio23) learns to co-register, fuse, and upsample multiple frames into one super-resolved image in an end-to-end manner. Residual attention model (RAMS) utilizes 3D convolutions to exploit spatial and temporal relationships across images for HR reconstruction(Reference Salvetti, Mazzia, Khaliq and Chiaberge24) of satellite images. Deep burst SR(Reference Bhat, Danelljan, Gool and Timofte25) combines both pixel-wise optical flow alignment and attention based fusion module to achieve HR reconstruction from image bursts.

Despite all the previous work of SR on natural and satellite images, little work has been done on MISR methods in the context of cryo-EM. Preliminary work done by Chen et al.(Reference Chen26) demonstrated that MISR reconstruction surpassing the Nyquist frequency is possible by using a noiseless synthetic dataset and without considering the modulation effects of the contrast transfer function (CTF). Real micrographs acquired with an electron microscope, however, inevitably suffer from low-SNR due to the small doses used during imaging and are modulated by the CTF. Meanwhile, deep-learning techniques have been applied to cryo-EM imaging in a variety of other contexts, including particle picking(Reference Wagner, Merino and Stabrin27Reference Sanchez-Garcia, Segura, Maluenda, Carazo and Sorzano30), automated micrograph and class selection(Reference Li, Cash, Tesmer and Cianfrocco31), CNNs for segmentation of cryo-electron tomograms(Reference Chen, Dai and Sun32), map denoising and local resolution estimation(Reference Tegunov, Xue, Dienemann, Cramer and Mahamid33Reference Avramov, Vyenielo, Gomez-Blanco, Adinarayanan, Vargas and Si35), and more recently the study of conformational heterogeneity during 3D reconstruction(Reference Zhong, Bepler, Berger and Davis36). Inspired by the recent success of deep-learning-based MISR approaches, we aim to tackle the problem of SR using cryo-EM images by jointly registering LR images and reconstructing SR images, all within an end-to-end trainable network based on the ZSSR framework(Reference Shocher, Cohen and Irani5). An early version of our approach was reported in Huang et al.(Reference Huang, Zhou, Du, Chen, Wang, Rudin and Bartesaghi37)

3. Method

3.1. Deep internal statistics of cryo-EM micrographs

The key assumptions of our approach are that cryo-EM micrographs have data internal repetition and that due to radiation damage, the amount of signal available at different frequencies varies for each frame, as high-frequency information is degraded in frames with more radiation damage. Figure 1 shows an example of a single-particle cryo-EM micrograph (average from all raw movie frames). Since each micrograph contains hundreds of projections of the same protein-of-interest, data repetition occurs naturally. In addition, as data is collected in movie mode, early frames and late frames are subject to different radiation damage regimes. Earlier frames in the exposure contain more high-frequency information than later frames which are affected by radiation damage. As shown in Figure 2, at higher frequencies, Thon rings are more visible in micrographs obtained by averaging the first half of frames in the movie, compared to micrographs obtained by averaging the second half of the frames. This phenomenon was also empirically verified by Bartesaghi et al.(Reference Bartesaghi, Matthies, Banerjee, Merk and Subramaniam38) and Grant and Grigorieff(Reference Grant and Grigorieff39) which compared 3D reconstructions of proteins obtained using frames from different exposure ranges, showing that frames from lower exposures achieve 3D reconstructions with higher resolutions compared to those obtained using later frames in the exposure.

Figure 2. Internal predictive power of movie-specific information. (a) Power spectrum calculated from the average of the first half of frames (less radiation damage) and from the second half of frames (more radiation damage). As indicated by the white arrows, Thon rings are more visible in the first image which has less radiation damage compared to the second image that presents more radiation damage. As reported earlier, this shows that earlier frames in the exposure carry more high-frequency information than the later frames. (b) Cross-correlation between the fitted contrast transfer function (CTF) and the measured power spectrum. Similar to panel (a), the power spectrum computed from the early part of the exposure has higher cross-correlation compared with the theoretical CTF. The better cross-correlation fit confirms that the high-frequency signal is stronger in the first half of the exposure.

3.2. Problem formulation

MISR aims at recovering an HR image $ {I}^{HR} $ from a set of $ M $ LR images $ {I}_i^{LR},i\in \left[1,\dots, M\right] $ of the same scene acquired during a certain time window. Typically, a LR image $ {I}_i^{LR} $ is related to the HR image $ {I}^{HR} $ through motion shift, blurring, downsampling, and noise corruption:

(1) $$ {I}_i^{LR}= DC\left({\phi}_{m_i}\left({I}^{HR}\right)\right)+{\sigma}_i, $$

where $ D $ is the downsampling process, $ C $ represents blurring, $ {\phi}_{m_i} $ is the relative motion for each frame, and $ {\sigma}_i $ is the noise corruption. In the context of cryo-EM, $ {I}_i^{LR} $ are the collected LR movie frames and $ {I}^{HR} $ is the ground-truth HR image, without CTF modulation and free of noise. Motion shift mainly comes from beam-induced motion that occurs during data acquisition, blurring is modeled by the CTF, which describes how contrast (information) is transferred to the image in terms of the spatial frequency. Severe noise corruption is a result of low electron dosage during imaging. It is worth noting that the CTF, unlike other commonly used blurring kernels such as Gaussian, has multiple zero crossings, which means information at certain frequencies is completely lost, making direct inversion impossible. Denoising techniques applied at the single image-level are also prone to the removal of high-frequency information along with the actual noise. Therefore, in the standard cryo-EM data processing pipeline, CTF inversion and denoising is not applied until the final step of 3D reconstruction. In order to cope with the special characteristics of cryo-EM images described above, unlike most standard SR algorithms, instead of recovering an HR image $ {I}^{HR} $ that is free of blurring and noise from LR images, we aim to generate $ {\tilde{I}}^{SR} $ subject to the modulation by the CTF and noise using our proposed methodology $ G $ :

(2) $$ {\displaystyle \begin{array}{l}{\tilde{I}}^{SR}=G\left({I}_{1,\dots, K}^{LR}\right),\\ {}{\tilde{I}}^{SR}=C\ast {I}^{HR}+\sigma .\end{array}} $$

Since raw frames $ {I}_j^{LR},\hskip0.5em j\in \left[1,\dots, M\right] $ , where $ M $ is the total number of raw frames in cryo-EM movies, have extremely low SNR and errors of SR reconstruction from LR images grow in proportion to the noise variance(Reference Robinson and Milanfar40), and frames collected under greater electron dosage have less high-frequency information due to radiation damage, we use moving averages of raw frames aligned to different reference frames (instead of using raw frames as LR inputs):

(3) $$ {\hat{I}}_i^{LR}=F\left({I}_{m,\dots n}^{LR},j\right),i=1,\dots, K $$

where $ F $ aligns and averages raw frames $ {I}_{m,\dots, n}^{LR} $ from the mth frame to the nth frame in the movie with respect to the reference frame $ j $ . Reference frame $ j $ is selected at random from all possible frames $ M $ . These generated LR frame averages $ {\hat{I}}_i^{LR} $ have higher SNR, and since each $ {\hat{I}}_i^{LR} $ is aligned to a different reference frame, relative motions exist between these frame averages. Therefore, they are more suitable as LR inputs for SR reconstruction. In addition, as these inputs are obtained by averaging frames with high electron exposure, they contain limited high-frequency information.

3.3. Proposed framework

Our proposed framework (Figure 3a,b) leverages both the strong internal predictive power and the generalization capabilities of deep neural networks. Given an input movie stack $ {I}_{0,\dots, M}^{LR} $ with $ M $ frames, we first divide the stack into two parts: (a) early frames in the exposure $ {I}_{0,\dots, M/2}^{LR} $ (first half of the frames), and (b) late frames in the exposure $ {I}_{M/2,\dots, M}^{LR} $ . We transform $ {I}_{0,\dots, M/2}^{LR} $ into $ {\hat{I}}_i^{LR} $ using Equation 3 and $ {I}_{0,\dots, M/2}^{LR} $ into $ {I}_{avg}^{LR} $ by aligning and averaging all frames in the low exposure stack. We further downsample $ {\hat{I}}_i^{LR} $ by a factor of 2 and treat the further downsampled frames as LR inputs to the network. We treat $ {I}_{avg}^{LR} $ as a pseudo HR micrograph. The network learns to reconstruct $ {I}_{avg}^{LR} $ using further downsampled $ {\hat{I}}_i^{LR} $ . Using this pseudo HR–LR pair, we are able to train our movie-specific SR Net without the need for any ground-truth HR images. To summarize, a movie-specific SR Net is trained in the following ways:

  1. 1. Extract example patches of fixed size from the input LR frames $ {\hat{I}}_i^{LR} $ and the input pseudo HR image $ {I}_{avg}^{LR} $ .

  2. 2. Further downsample the extracted examples from $ {\hat{I}}_i^{LR} $ by a factor of $ s $ (we use 2). These downsampled examples now become temporary LRs.

  3. 3. Temporary HR–LR pair is formed using the downsampled extracted patches from $ {\hat{I}}_i^{LR} $ and its corresponding extracted patch from the pseudo HR image $ {I}_{avg}^{LR} $ .

  4. 4. Feed temporary LR images obtained in step 2 into SR Net, a SR output is generated and compared with the temporary HR.

Figure 3. Overall cryo-ZSSR framework. (a) During the training stage, pseudo LR–HR pairs are formed using further downsampled frames that have more radiation damage (second half of frames in a movie) ( $ {\hat{I}}_i^{LR} $ ) and averages of frames with less radiation damage ( $ {I}_{avg}^{LR} $ , first half of frames in a movie). Extracted patches of frames from further downsampled $ {\hat{I}}_i^{LR} $ are fed into SR Net which produces a $ 2\times $ super-resolved image $ {\tilde{I}}^{SR} $ . SR Net learns to recover $ {I}_{avg}^{LR} $ from the coarser input $ {\hat{I}}_i^{LR} $ . (b) During the inference stage, the resulting self-supervised SR Net is then applied to the full $ {\hat{I}}_i^{LR} $ to produce its SR output. (c) Architecture of SR Net: input frames are first upsampled to the desired output size. The interpolated frames are used as inputs to SR Net. These frames are encoded, fused, and decoded to generate the final SR output.

Once the network is trained, instead of using $ {\hat{I}}_i^{LR} $ obtained using frames from the second half of the exposure, a new set of $ {\tilde{I}}_i^{LR} $ is formed by using averages of all raw frames aligned to different reference frames:

(4) $$ {\tilde{I}}_i^{LR}=F\left({I}_{0,\dots M}^{LR},j\right),\hskip0.2em i=1,\dots, K, $$

where F represents the application of a motion correction/alignment algorithm such as MotionCor2 (Reference Zheng, Palovcak, Armache, Verba, Cheng and Agard41). In this new setup, LR frames are used as input to the trained network and the desired SR output $ {\tilde{I}}^{SR} $ is constructed. By using $ {I}^{HR} $ from the first half of the frames (that contain more high-frequency information) and learning to recover these high-frequency information from LR inputs, the movie-specific SR Net is able to leverage the power of cross-frequency internal recurrence of image-specific information. To further enrich the training dataset, data augmentation is applied to the set of LR images to extract more pairs of HR–LR to train on, including mirror reflections in the vertical and horizontal directions.

The overall architecture of SR Net is based on HighResNet(Reference Deudon, Kalaitzis, Goytom, Arefin, Lin, Sankaran, Michalski, Kahou, Cornebise and Bengio23), which includes three main steps: encoding, fusion, and decoding (Figure 3c). The network learns to implicitly co-register multiple LR frames $ {I}_i^{LR} $ and fuse them into a single SR view. Unlike HighResNet, which upscales input LR views during the decoding step using a deconvolution layer, we first upscale LR inputs $ {I}_i^{LR} $ to the desired SR output size $ {I}_i^{Inter} $ before feeding it into the encoder using bilinear interpolation. The network thus learns the residual between the interpolated LR and the HR images. A detailed description of the architecture of HighResNet is given in Deudon et al.(Reference Deudon, Kalaitzis, Goytom, Arefin, Lin, Sankaran, Michalski, Kahou, Cornebise and Bengio23).

Encode. The encoding stage contains two steps: first, compute a reference micrograph and second, embed each frame jointly with the reference. The reference micrograph is computed as the median of all input LR frames $ {I}_{i,\dots, K}^{LR} $ . The reference micrograph is concatenated to each frame and the concatenated reference-frame representations serve as inputs to the embedding layer. The shared reference micrograph serves as anchor for implicit alignment and encourages the network to learn differences across multiple frames.

Fuse. Encoded outputs $ {s}_{i,..,K}^0 $ from the embedding layer are then fused recursively. At each time step $ t $ , after fusion, the number of encoded outputs is reduced by half. Given a pair of hidden states $ {s}_i^t $ and $ {s}_j^t $ , the fusion step merges these two representations by first concatenating $ {s}_i^t $ and $ {s}_j^t $ and then projecting to a new representation.

Decode. After $ T={\log}_2K $ fusion steps, the final LR encoded representation $ {s}_i^T $ is fed into the decoder and the decoder outputs the final super-resolved micrograph, $ {I}^{SR} $ .

3.3.1. Loss function

In addition to computing the mean absolute error (MAE) between the generated SR and the actual pseudo-HR images, we also use a Fourier domain frequency loss to help preserve CTF modulation.

Fourier domain frequency loss. For an image and its Fourier representation $ F $ , denote $ F\left(u,v\right) $ as the Fourier coefficient at spectrum coordinate $ \left(u,v\right) $ . Let $ R\left(u,v\right) $ and $ I\left(u,v\right) $ be its real and imaginary parts, we can rewrite $ F\left(u,v\right) $ as

(5) $$ F\left(u,v\right)=R\left(u,v\right)+ iI\left(u,v\right). $$

Let $ {F}_{HR}\left(u,v\right) $ be the Fourier coefficient of the ground-truth HR image and $ {F}_{SR}\left(u,v\right) $ be the Fourier coefficient of the reconstructed SR image. Denote $ {\overrightarrow{r}}^{HR} $ and $ {\overrightarrow{r}}^{SR} $ as respective vectors mapped from $ {F}_{HR}\left(u,v\right) $ and $ {F}_{SR}\left(u,v\right) $ . By the definition of amplitude and phase of each Fourier coefficient, $ \mid \overrightarrow{r}\mid $ corresponds to the amplitude and $ \theta $ corresponds to the phase. Therefore, the frequency distance can be represented as the distance between $ {\overrightarrow{r}}^{HR} $ and $ {\overrightarrow{r}}^{SR} $ , which can be calculated using the $ {L}_2 $ Euclidean distance:

(6) $$ d\left({\overrightarrow{r}}_{\left(u,v\right)}^{HR},{\overrightarrow{r}}_{\left(u,v\right)}^{SR}\right)={\left|{F}_{HR}\Big(u,v\left)-{F}_{SR}\right(u,v\Big)\right|}^2. $$

As both generated SR and pseudo-HR are corrupted by noise, this means that at higher frequency, its corresponding coefficients contain both information from underlying signal and noise. Therefore, minimizing distance at the high-frequency portion of the spectrum may lead to undesired learning of noise. To mitigate this effect, we reweight the loss at each frequency level using a Gaussian kernel. For higher frequencies, its resulting loss is down-weighted in the overall loss calculation. The final loss is calculated as the sum of the MAE loss and the weighted frequency loss.

3.3.2. Implementation details

We use the ADAM optimizer, starting with a learning rate of .001 and we adaptively decrease the learning rate based on the training procedure proposed in Shocher et al.(Reference Shocher, Cohen and Irani5). Training stops when the learning rate reaches $ {10}^{-5} $ , at around 200 iterations. The network is trained to learn upscaling by a factor of 2. At each iteration, a fixed crop size of $ 256\times 256 $ is used, while the 2× downsampled versions have size of $ 128\times 128 $ . This way, training time is independent of the input size. During the inference stage, the generated SR is further combined with the back-projection technique(Reference Irani and Peleg42,Reference Glasner, Bagon and Irani43) . The final image is corrected by back-projection. Each set of LR images takes around 2 min to train for an upsampling factor of 2, the final SR image takes about 30 s to generate on an NVIDIA Tesla V100 GPU with 32GB of memory.

3.4. Overall data processing pipeline with cryo-ZSSR

As shown in Figure 1a, cryo-ZSSR serves as a pre-processing step to the overall cryo-EM 3D reconstruction process. To summarize, the new processing pipeline is as follows:

  • For each movie stack $ {I}_{0,\dots, M}^{LR} $ containing M frames, divide each stack into two parts: (Reference Bendory, Bartesaghi and Singera) early frames in the exposure $ {I}_{0,\dots, M/2}^{LR} $ (first half of the frames), and (Reference Bartesaghi, Merk and Banerjeeb) late frames in the exposure $ {I}_{M/2,M}^{LR} $ (second half of the frames).

  • For $ {I}_{M/2,M}^{LR} $ , align a subset of these frames with respect to different reference frames (we used four frames) using a motion correction algorithm such as MotionCorr to obtain $ {\hat{I}}_{i=1,\dots, 4}^{LR} $ , Equation 3. For $ {I}_{0,\dots, M/2}^{LR} $ , align these frames with respect to the center frame to obtain $ {I}_{avg}^{LR} $ .

  • Using $ {\hat{I}}_{i=1,\dots, 4}^{LR} $ and $ {I}_{avg}^{LR} $ , generate pseudo HR–LR pairs and train the network by following steps in Section 3.3.

  • Align all frames with respect to the reference frames used in the second step above and obtain a new set of $ {\hat{I}}_{i=1,\dots, 4}^{LR} $ . Feed the new $ {\hat{I}}_{i=1,\dots, 4}^{LR} $ into trained network and obtain the final super resolved output $ {\tilde{I}}^{SR} $ .

  • Perform all the following data processing steps (CTF estimation, particle picking, orientation estimation, and 3D reconstruction) on $ {\tilde{I}}^{SR} $ generated micrographs from each movie stack.

4. Experiments and Validation on Single-Particle Cryo-EM Data

To validate our approach, we used cryo-EM movies of apoferritin and the T20S proteasome from the Electron Microscopy Public Image Archive (EMPIAR) under accession codes 10146 and 10025(Reference Bartesaghi, Aguerrebere and Falconieri3,Reference Iudin, Korir, Salavert-Torres, Kleywegt and Patwardhan44) .

EMPIAR-10146. Apoferritin is a commonly used test sample that has a molecular weight of 440 kDa and octahedral symmetry (O). This dataset consists of 20 movies with 50 frames each and $ 1,240\times 1,200 $ pixels in size. The physical pixel size is 1.5 Å and the images were acquired using a beam energy of 300 kV and an exposure rate of 2  $ {e}^{-}/ $ Å2 (equivalent to a total dose of 100  $ {e}^{-} $ 2), Figure 1b. The original movies are subjected to the standard single-particle pipeline resulting in a 3.5 Å resolution reconstruction from 1,200 particles that was used as ground truth. We then binned the original movies by a factor of 2 using the IMOD program(Reference Kremer, Mastronarde and McIntosh45) (resulting in a pixel size of 3 Å). Downsampled frames using uncropped frames had a size of $ 620\times 600 $ pixels and were aligned using MotionCorr2(Reference Zheng, Palovcak, Armache, Verba, Cheng and Agard41) and averaged. The frames were aligned to four different reference frames and the corresponding frame averages were generated.

EMPIAR-10025. The T20S proteasome has a molecular weight of 750 kDa and D7 symmetry. We used a subset of this dataset which consists of 47 movies with 38 frames each and $ 7,420\times 7,676 $ pixels in size. As the movies are acquired using microscope’s camera SR mode, the pixel size is 0.66 Å. The images were acquired using a beam energy of 300 kV and an exposure rate of 1.4  $ {e}^{-}/ $ Å2 (equivalent to a total dose of 58  $ {e}^{-} $ 2), Figure 1c. Instead of using the super-resolved movies as reference, we used movies binned by a factor of 2 (1.32 Å pixel size, $ 3,710\times 3,838 $ image size) as ground truth. LR movie frames are obtained by downsampling the original movies by a factor of 4 (2.64 Å pixel size, $ 1,855\times 1,919 $ image size). We followed the same alignment and downsampling procedure as in EMPIAR-10406. For simplicity, we cropped sub-micrograph patches of size $ 1,024\times 1,024 $ from the resulting LR frames and these sub-micrograph patches are used as the input LR images.

Movie-specific SR Nets were trained for each movie in both datasets. Once fully trained, LR images were upscaled by a factor of 2 through our framework. For EMPIAR-10406, the resulting SR image has size $ 1,240\times 1,200 $ , which is the same as the original unbinned micrographs. For EMPIAR-10025, the resulting SR image has size $ 2,048\times 2,048 $ . The original HR reference movies were never used or seen during any part of the training or testing steps. The resulting SR micrographs now replaced the downsampled LR images and were used as inputs to the single-particle cryo-EM structure determination pipeline. The CTF of each SR micrograph was estimated using CTFFIND4(Reference Rohou and Grigorieff46) and particles were extracted and subjected to iterative 3D refinement using the cisTEM package(Reference Grant, Rohou and Grigorieff47). This process was repeated for the LR images, and the SR micrographs upsampled using bilinear interpolation and cryo-ZSSR. In both cases, we used the exact same particle stacks and estimated orientation parameters for 3D reconstruction.

4.1. Cryo-ZSSR improves the quality of individual micrographs

To test the performance of our algorithm on individual micrographs, we estimated the CTF of each of the micrographs in both datasets using three sets of images (LR, upsampled using bilinear interpolation, and upsampled using cryo-ZSSR) and the ground-truth original image. Specifically, we quantified the overall improvement in image quality by measuring the estimated fit resolution for images in both datasets. Estimated fit resolution gives an indication of how far the signal extends (lower numbers are better). As shown in Figure 4, left, in both datasets, SR images reconstructed using our approach have better fit resolution than both the LR and bilinear interpolated images, indicating that cryo-ZSSR can effectively recover HR information present in the LR movie stacks. In addition, we show 1D CTF radial profiles of a representative cryo-ZSSR upsampled image and its corresponding CTF cross correlation fit compared to LR, bilinear interpolated and original images (Figure 4, right). As shown, the cross-correlation fit results indicate that the strength of the signal present in the cryo-ZSSR result is higher when compared to the LR and bilinear interpolated images, even though it is not as good as the original input. This implies that although cryo-ZSSR can recover some high-frequency information present in the LR movie stacks, it cannot achieve perfect recovery. In addition, some high-frequency information is permanently lost during the downsampling process, making full recovery impossible.

Figure 4. Cryo-ZSSR improves image quality metrics for individual micrographs. To evaluate the performance of cryo-ZSSR at the micrograph level, we estimated the CTF of movies in the EMPIAR-10146 and EMPIAR-10025 datasets before and after upsampling. (a) CTF statistics of EMPIAR-10146. Right: Histogram of estimated fit resolution showing the net improvement in image quality obtained by cryo-ZSSR (lower fit resolutions represent better results). Middle: Example 1D CTF radial profiles of cryo-ZSSR upsampled image. Left: Corresponding CTF Fit cross correlation score. As shown, the output from cryo-ZSSR has better cross correlation score compared to both bilinear interpolation and the low-resolution image. (b) CTF statistics of EMPIAR-10025. Similar to EMPIAR-10146, cryo-ZSSR is able to achieve better fit resolution, cross correlation score compared to LR input and the bilinear interpolated image. Right: Histogram of estimated fit resolution. Middle: Example 1D CTF radial profiles of cryo-ZSSR upsampled images. Left: Corresponding CTF fit cross correlation score for LR, bilinear interpolation, cryo-ZSSR and original inputs.

4.2. Cryo-ZSSR improves the resolution of 3D reconstructions

We also evaluated the downstream effects of our SR interpolation algorithm by measuring the quality of the final 3D reconstructions. The three sets of movies (LR, upsampled using bilinear interpolation, and cryo-ZSSR) were used as input to the standard single-particle refinement pipeline implemented in cisTEM. For EMPIAR-10406, $ \mathrm{1,200} $ particles were selected and aligned against an external reference of apoferritin using iterative projection matching. We repeated this process using the three sets of images and the ground-truth data (Figure 5). Consistent with the CTF estimation results, the features and resolution of the cryo-ZSSR map are better than the ones obtained using bilinear interpolation and the LR data, with estimated resolutions according to the 0.143-FSC criteria of 3.9 Å, 4.8 Å, and 6.0 Å, respectively (Figure 5a). The resolution obtained using the ground-truth images is 3.5 Å. Lower numbers indicate better reconstruction quality. The resolution obtained by cryo-ZSSR clearly surpasses the 6 Å Nyquist limit imposed by the original physical pixel size of 3 Å, and the reconstruction shows clear density for side chains, in agreement with the atomic model and corresponding structural features in the ground-truth map. For EMPIAR-10025, $ \mathrm{7,810} $ particles were selected as inputs to ab-initio reconstruction and homogeneous refinement. Similarly, the features and resolution of the cryo-ZSSR map are better than the ones obtained using bilinear interpolation and LR images, with estimated resolutions according to the 0.143-FSC criteria of 3.9 Å, 4.0 Å, and 5.5 Å (Figure 5b). The resolution obtained using the ground-truth images was 3.1 Å. While our proposed method is able to obtain better reconstruction compared to bilinear interpolation, the improvement is not as significant as compared to EMPIAR-10146. It should be noted that for LR images with a pixel size of 2.64 Å, the Nyquist limit is 5.28 Å. However, the current reconstruction method is only able to obtain a resolution of 5.53 Å, indicating that the resolution is limited by other factors (less symmetry) in addition to the pixel size. This reveals a potential limitation of our proposed method: when there is less symmetry, which leads to less internal data repetition, even though it is still able to outperform traditional interpolation-based upsampling methods, the degree of improvement is less significant.

Figure 5. Cryo-ZSSR upsampled images improve the resolution of 3D structures. To evaluate the performance of cryo-ZSSR at the 3D level, we performed 3D reconstruction for both apoferritin (EMPIAR-10146) and T20S proteasome (EMPIAR-10025) datasets. In each case, reconstructions were obtained using the same set of particles. (a) Overall structure of apoferritin and zoomed-in view of an alpha helix with fitted atomic model, for maps obtained from the LR images (top left), upsampled images using bilinear interpolation (top right), upsampled images using cryo-ZSSR (bottom left), and ground-truth images (bottom right). Fourier shell correlation (FSC) curves for maps obtained using LR images (gray), upsampled using bilinear interpolation (green), and upsampled using cryo-ZSSR (magenta) against ground-truth reconstruction (bottom). Estimated resolutions are 6.0 Å, 4.8 Å, and 3.9 Å, respectively, based on the 0.143-cutoff (dotted line). Lower numbers represent better reconstruction quality. (b) Overall structure of T20S proteasome and zoomed-in view with fitted atomic model. Similar to EMPIAR-10146, cryo-ZSSR is able to achieve better 3D resolution. FSC curves for maps obtained using LR images (gray), upsampled using bilinear interpolation (green), and upsampled using cryo-ZSSR (magenta) against ground-truth reconstruction (bottom). Estimated resolutions are 3.9 Å, 4.0 Å, and 5.5 Å, respectively, based on the 0.143-cutoff (dotted line). Due to low sampling rate, the FSC for the LR reconstruction has a rapid decay at 5.5 Å.

5. Discussion and Conclusion

We present a neural network framework to upsample low-SNR single-particle cryo-EM movies using a MISR algorithm based on self-supervised deep internal learning. By leveraging information repetition across multiple frequencies in collected cryo-EM movies, we are able to train the network without the need of ground-truth or prior training using HR images. Applications of this technique to a LR dataset of apoferritin sampled at 3 Å/pixel resulted in a three-dimensional reconstruction at 3.9 Å resolution where side chains could be visualized at a similar level of detail seen in the ground-truth map. On a LR dataset of T20S proteasome sampled at 2.64 Å/pixel, after SR upsampling, a three-dimensional reconstruction at 3.9 Å resolution is obtained. These experiments suggest that cryo-ZSSR is an effective strategy to recover HR information contained in low-SNR, LR cryo-EM movies. The proposed framework is most suitable for datasets with relatively high symmetry that achieve the Nyquist frequency during 3D reconstruction, where the main limitation factor is the pixel size. For datasets where the resolution is limited by factors other than the pixel size, the resolution improvements brought about by our method will be less significant. Admittedly, our proposed algorithm has some limitations: first, it does not have an explicit motion compensation component. While implicit registration already results in SR images that contains more high-frequency information, accurate sub-pixel motion estimation is key in further improving the quality of the SR image. Second, the algorithm does not presently account for the resolution-lowering effects caused by radiation damage affecting cryo-EM samples, as current fusion and decoding steps do not incorporate dose weighting. Previous research shows that by accounting for radiation damage through dose weighting, the resolution of 3D reconstructions can be improved. Therefore, incorporating frequency-domain Fourier coefficient reweighting has the potential of obtaining further improvements in resolution. As cryo-EM datasets typically contain thousands of micrographs, training of our image-specific network can take a long time. To this end, we are investigating the possibility of training a dataset-specific network. In addition, we will continue investigating how protein symmetry can affect the performance of the algorithm, as symmetry plays an important role in internal statistics recurrence. Overall, the proposed SR strategy may be used in conjunction with lower magnification imaging to accelerate data collection without sacrificing image quality.

Acknowledgments

This study utilized the computational resources offered by Duke Research Computing (http://rc.duke.edu). We thank T. Futhey, C. Kneifel, K. Kilroy, M. Newton, V. Orlikowski, T. Milledge, and D. Lane from the Duke Office of Information Technology and Research Computing for providing assistance with the computing environment.

Competing Interests

The authors declare no competing interests exist.

Authorship Contributions

Conceptualization: Q.H., A.B.; Data processing and visualization: Y.Z., H.L.; Methodology and implementation: Q.H.; Writing original draft: Q.H. All authors approved the final submitted draft.

Funding Statement

This work was supported by a Visual Proteomics Imaging grant from the Chan Zuckerberg Initiative (CZI) to A.B.

Data Availability Statement

Replication data and code can be found in https://gitlab.cs.duke.edu/bartesaghilab/cryo-zssr.

References

Bendory, T, Bartesaghi, A & Singer, A (2020) Single-particle cryo-electron microscopy: Mathematical theory, computational challenges, and opportunities. IEEE Signal Process Mag 37, 5876.CrossRefGoogle ScholarPubMed
Bartesaghi, A, Merk, A, Banerjee, S, et al. (2015) 2.2 Å resolution cryo-EM structure of β-galactosidase in complex with a cell-permeant inhibitor. Science 348, 11471151.CrossRefGoogle ScholarPubMed
Bartesaghi, A, Aguerrebere, C, Falconieri, V, et al. (2018) Atomic resolution cryo-EM structure of β-galactosidase. Structure 26, 848856.e3.CrossRefGoogle ScholarPubMed
Singer, A & Sigworth, F (2020) Computational methods for single-particle electron cryomicroscopy. Annu Rev Biomed Data Sci 3, 163190.CrossRefGoogle ScholarPubMed
Shocher, A, Cohen, N & Irani, M (2017) “Zero-Shot” super-resolution using deep internal learning. Computer Vision and Pattern Recognition.CrossRefGoogle Scholar
Krull, A, Buchholz, T & Jug, F. (2018) Noise2Void – Learning denoising from single noisy images. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 21242132.Google Scholar
Palovcak, E, Asarnow, D, Campbell, M, Yu, Z & Cheng, Y (2020) Enhancing the signal-to-noise ratio and generating contrast for cryo-EM images with convolutional neural networks. IUCrJ 7, 11421150.CrossRefGoogle ScholarPubMed
Zeyde, R, Elad, M & Protter, M (2010) On single image scale-up using sparse-representations. In Proceedings of the 7th International Conference on Curves and Surfaces, Avignon, France: Springer, June 24-30 2010 pp. 711730.Google Scholar
Schulter, S, Leistner, C & Bischof, H (2015) Fast and accurate image upscaling with super-resolution forests. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, June 7 2015 to June 12, 2015 pp. 37913799.CrossRefGoogle Scholar
Freeman, W, Jones, T & Pasztor, E (2002) Example-based super-resolution. IEEE Comput Graphics Appl 22, 5665.CrossRefGoogle Scholar
Timofte, R, De, V & Gool, L (2013) Anchored neighborhood regression for fast example-based super-resolution. In 2013 IEEE International Conference on Computer Vision, Sydney, Australia: IEEE, 1-8 Dec. 2013, pp. 19201927.CrossRefGoogle Scholar
Ulyanov, D, Vedaldi, A & Lempitsky, V (2020) Deep image prior. Int J Computr Vision 128, 18671888.CrossRefGoogle Scholar
Lim, B, Son, S, Kim, H, Nah, S & Lee, KM (2017) Enhanced Deep Residual Networks for Single Image Super-Resolution. 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 11321140.CrossRefGoogle Scholar
Goodfellow, I, Pouget-Abadie, J, Mirza, M, Xu, B, Warde-Farley, D, Ozair, S, Courville, A Bengio, Y. (2020) Generative adversarial networks. Communications of The ACM.CrossRefGoogle Scholar
Ledig, C, Theis, L, Huszár, F, Caballero, J, Aitken, AP, Tejani, A, Totz, J, Wang, Z, & Shi, W. (2016). Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 105114.Google Scholar
Menon, S, Damian, A, Hu, S, Ravi, N & Rudin, C (2020) PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of Generative Models. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 24342442.CrossRefGoogle Scholar
Tsai, RY & Huang, TS (1980) Moving image restoration and registration. IEEE International Conference on Acoustics, Speech, and Signal Processing.CrossRefGoogle Scholar
Takeda, H, Farsiu, S & Milanfar, P (2007) Kernel regression for image processing and reconstruction. IEEE Trans Image Process 16, 349366.CrossRefGoogle ScholarPubMed
Tipping, M & Bishop, C (2003) Bayesian image super-resolution. In Advances in Neural Information Processing Systems 15, pp. 13031310. http://papers.nips.cc/paper/2315-bayesian-image-super-resolution.pdf Accessed 2020 November.Google Scholar
Fan, C, Wu, C, Li, G & Ma, J (2017) Projections onto convex sets super-resolution reconstruction based on point spread function estimation of low-resolution remote sensing images. Sensors (Basel, Switzerland) 17, 362.CrossRefGoogle ScholarPubMed
Robinson, M, Toth, C, Lo, J & Farsiu, S (2010) Efficient Fourier-wavelet super-resolution. IEEE Trans Image Process 19, 26692681.CrossRefGoogle ScholarPubMed
Sajjadi, MS, Vemulapalli, R & Brown, MA (2018) Frame-Recurrent Video Super-Resolution. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 66266634. Accessed 2021 JanuaryCrossRefGoogle Scholar
Deudon, M, Kalaitzis, A, Goytom, I, Arefin, MR, Lin, Z, Sankaran, K, Michalski, V, Kahou, SE, Cornebise, J, & Bengio, Y (2020). HighRes-net: Recursive Fusion for Multi-Frame Super-Resolution of Satellite Imagery. ArXiv, abs/2002.06460.Google Scholar
Salvetti, F, Mazzia, V, Khaliq, A & Chiaberge, M (2020) Multi-image super resolution of remotely sensed images using residual attention deep neural networks. Remote Sens 12, 2207.CrossRefGoogle Scholar
Bhat, G, Danelljan, M, Gool, LV & Timofte, R (2021) Deep Burst Super-Resolution. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 92059214. Accessed 2022 MarchCrossRefGoogle Scholar
Chen, J (2018) Single-particle 3D reconstruction beyond the Nyquist frequency. In 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid: IEEE, 3-6 Dec. 2018, pp. 24422445.CrossRefGoogle Scholar
Wagner, T, Merino, F, Stabrin, M, et al. (2019) SPHIRE-crYOLO is a fast and accurate fully automated particle picker for cryo-EM. Commun Biol 2, 113.CrossRefGoogle ScholarPubMed
Wang, F, Gong, H, Liu, G, et al. (2016) DeepPicker: a deep learning approach for fully automated particle picking in cryo-EM. J Struct Biol 195, 325336.CrossRefGoogle ScholarPubMed
Bepler, T, Morin, A, Rapp, M, et al. (2019) Positive-unlabeled convolutional neural networks for particle picking in cryo-electron micrographs. Nat Methods 16, 11531160.CrossRefGoogle ScholarPubMed
Sanchez-Garcia, R, Segura, J, Maluenda, D, Carazo, J & Sorzano, C (2018) Deep consensus, a deep learning-based approach for particle pruning in cryo-electron microscopy. IUCrJ 5, 854865.CrossRefGoogle ScholarPubMed
Li, Y, Cash, J, Tesmer, J & Cianfrocco, M (2020) High-throughput cryo-EM enabled by user-free preprocessing routines. Structure 28, 858869.e3.CrossRefGoogle ScholarPubMed
Chen, M, Dai, W, Sun, S, et al. (2017) Convolutional neural networks for automated annotation of cellular cryo-electron tomograms. Nat Methods 14(10), 983985.CrossRefGoogle ScholarPubMed
Tegunov, D, Xue, L, Dienemann, C, Cramer, P & Mahamid, J (2021) Multi-particle cryo-EM refinement with M visualizes ribosome-antibiotic complex at 3.7 Å inside cells. Nat Methods 18, 186193.CrossRefGoogle Scholar
Ramírez-Aportela, E, Mota, J, Conesa, P, Carazo, J & Sorzano, C (2019) DeepRes: a new deep-learning- and aspect-based local resolution method for electron-microscopy maps. IUCrJ 6, 10541063.CrossRefGoogle ScholarPubMed
Avramov, T, Vyenielo, D, Gomez-Blanco, J, Adinarayanan, S, Vargas, J & Si, D (2019) Deep learning for validating and estimating resolution of cryo-electron microscopy density maps. Molecules 24, 1181.CrossRefGoogle ScholarPubMed
Zhong, E, Bepler, T, Berger, B & Davis, J (2021) CryoDRGN: Reconstruction of heterogeneous structures from cryo-electron micrographs using neural networks. Nat Methods 18, 176185.CrossRefGoogle Scholar
Huang, Q, Zhou, Y, Du, X, Chen, R, Wang, J, Rudin, C, & Bartesaghi, A (2020). Cryo-ZSSR: multiple-image super-resolution based on deep internal learning. ArXiv, abs/2011.11020.Google Scholar
Bartesaghi, A, Matthies, D, Banerjee, S, Merk, A & Subramaniam, S (2014) Structure of β -galactosidase at 3.2-Ã resolution obtained by cryo-electron microscopy. Proc Natl Acad Sci USA 111, 1170911714.CrossRefGoogle ScholarPubMed
Grant, T & Grigorieff, N (2015) Measuring the optimal exposure for single particle cryo-EM using a 2.6 Ã reconstruction of rotavirus VP6. eLife 4, e06980.CrossRefGoogle ScholarPubMed
Robinson, D & Milanfar, P (2006) Statistical performance analysis of super-resolution. IEEE Trans Image Process 15, 14131428.CrossRefGoogle ScholarPubMed
Zheng, S, Palovcak, E, Armache, J, Verba, K, Cheng, Y & Agard, D (2017) MotionCor2: anisotropic correction of beam-induced motion for improved cryo-electron microscopy. Nat Methods 14, 331332.CrossRefGoogle ScholarPubMed
Irani, M & Peleg, S (1991) Improving resolution by image registration. CVGIP 53, 231239.Google Scholar
Glasner, D, Bagon, S & Irani, M (2009) Super-resolution from a single image. In Proceedings of the IEEE International Conference on Computer Vision, Kyoto: IEEE, Sept. 27 2009 to Oct. 4 2009, pp. 349–356.Google Scholar
Iudin, A, Korir, P, Salavert-Torres, J, Kleywegt, G & Patwardhan, A (2016) EMPIAR: a public archive for raw electron microscopy image data. Nat Methods 13, 387388.CrossRefGoogle ScholarPubMed
Kremer, J, Mastronarde, D & McIntosh, J (1996) Computer visualization of three-dimensional image data using IMOD. J Struct Biol 116, 7176.CrossRefGoogle ScholarPubMed
Rohou, A & Grigorieff, N (2015) CTFFIND4: fast and accurate defocus estimation from electron micrographs. J Struct Biol 192, 216221.CrossRefGoogle ScholarPubMed
Grant, T, Rohou, A & Grigorieff, N (2018) cisTEM, user-friendly software for single-particle image processing. eLife 7, e35383.CrossRefGoogle ScholarPubMed
Figure 0

Figure 1. Super-resolution single-particle structure determination pipeline and example micrographs from two cryo-EM datasets. (a) Cryo-EM movies are collected using a large pixel size and subsequently upsampled by a factor of 2 using our self-supervised cryo-zero-shot super-resolution (cryo-ZSSR) approach. Super-resolved micrographs are then fed into the standard single-particle reconstruction workflow producing three-dimensional structures at resolutions surpassing the Nyquist rate. The 2× upsampling factor effectively results in a 4x speedup in the rate of data acquisition allowing the collection of four times more particles in the same amount of time. (b) Left: Example of a single raw frame from a movie of apoferritin from EMPIAR-10146 collected at 2 $ {e}^{-}/ $Å2. Right: Average of 50 frames corresponding to a total dose of 100 $ {e}^{-}/ $Å2. (c) Left: Example of a single raw frame from a movie of T20S proteasome from EMPIAR-10025 collected at 1.4 $ {e}^{-}/ $Å2. Right: Average of 38 frames corresponding to a total dose of 53.2 $ {e}^{-}/ $Å2.

Figure 1

Figure 2. Internal predictive power of movie-specific information. (a) Power spectrum calculated from the average of the first half of frames (less radiation damage) and from the second half of frames (more radiation damage). As indicated by the white arrows, Thon rings are more visible in the first image which has less radiation damage compared to the second image that presents more radiation damage. As reported earlier, this shows that earlier frames in the exposure carry more high-frequency information than the later frames. (b) Cross-correlation between the fitted contrast transfer function (CTF) and the measured power spectrum. Similar to panel (a), the power spectrum computed from the early part of the exposure has higher cross-correlation compared with the theoretical CTF. The better cross-correlation fit confirms that the high-frequency signal is stronger in the first half of the exposure.

Figure 2

Figure 3. Overall cryo-ZSSR framework. (a) During the training stage, pseudo LR–HR pairs are formed using further downsampled frames that have more radiation damage (second half of frames in a movie) ($ {\hat{I}}_i^{LR} $) and averages of frames with less radiation damage ($ {I}_{avg}^{LR} $, first half of frames in a movie). Extracted patches of frames from further downsampled $ {\hat{I}}_i^{LR} $ are fed into SR Net which produces a $ 2\times $ super-resolved image $ {\tilde{I}}^{SR} $. SR Net learns to recover $ {I}_{avg}^{LR} $ from the coarser input $ {\hat{I}}_i^{LR} $. (b) During the inference stage, the resulting self-supervised SR Net is then applied to the full $ {\hat{I}}_i^{LR} $ to produce its SR output. (c) Architecture of SR Net: input frames are first upsampled to the desired output size. The interpolated frames are used as inputs to SR Net. These frames are encoded, fused, and decoded to generate the final SR output.

Figure 3

Figure 4. Cryo-ZSSR improves image quality metrics for individual micrographs. To evaluate the performance of cryo-ZSSR at the micrograph level, we estimated the CTF of movies in the EMPIAR-10146 and EMPIAR-10025 datasets before and after upsampling. (a) CTF statistics of EMPIAR-10146. Right: Histogram of estimated fit resolution showing the net improvement in image quality obtained by cryo-ZSSR (lower fit resolutions represent better results). Middle: Example 1D CTF radial profiles of cryo-ZSSR upsampled image. Left: Corresponding CTF Fit cross correlation score. As shown, the output from cryo-ZSSR has better cross correlation score compared to both bilinear interpolation and the low-resolution image. (b) CTF statistics of EMPIAR-10025. Similar to EMPIAR-10146, cryo-ZSSR is able to achieve better fit resolution, cross correlation score compared to LR input and the bilinear interpolated image. Right: Histogram of estimated fit resolution. Middle: Example 1D CTF radial profiles of cryo-ZSSR upsampled images. Left: Corresponding CTF fit cross correlation score for LR, bilinear interpolation, cryo-ZSSR and original inputs.

Figure 4

Figure 5. Cryo-ZSSR upsampled images improve the resolution of 3D structures. To evaluate the performance of cryo-ZSSR at the 3D level, we performed 3D reconstruction for both apoferritin (EMPIAR-10146) and T20S proteasome (EMPIAR-10025) datasets. In each case, reconstructions were obtained using the same set of particles. (a) Overall structure of apoferritin and zoomed-in view of an alpha helix with fitted atomic model, for maps obtained from the LR images (top left), upsampled images using bilinear interpolation (top right), upsampled images using cryo-ZSSR (bottom left), and ground-truth images (bottom right). Fourier shell correlation (FSC) curves for maps obtained using LR images (gray), upsampled using bilinear interpolation (green), and upsampled using cryo-ZSSR (magenta) against ground-truth reconstruction (bottom). Estimated resolutions are 6.0 Å, 4.8 Å, and 3.9 Å, respectively, based on the 0.143-cutoff (dotted line). Lower numbers represent better reconstruction quality. (b) Overall structure of T20S proteasome and zoomed-in view with fitted atomic model. Similar to EMPIAR-10146, cryo-ZSSR is able to achieve better 3D resolution. FSC curves for maps obtained using LR images (gray), upsampled using bilinear interpolation (green), and upsampled using cryo-ZSSR (magenta) against ground-truth reconstruction (bottom). Estimated resolutions are 3.9 Å, 4.0 Å, and 5.5 Å, respectively, based on the 0.143-cutoff (dotted line). Due to low sampling rate, the FSC for the LR reconstruction has a rapid decay at 5.5 Å.