115
ROBUST IMAGE PROCESSING FOR CRYO-ELECTRON TOMOGRAPHY USING SPARSE PRIORS A DISSERTATION SUBMITTED TO THE DEPARTMENT OF ELECTRICAL ENGINEERING AND THE COMMITTEE ON GRADUATE STUDIES OF STANFORD UNIVERSITY IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF DOCTOR OF PHILOSOPHY Ka Hye Song March 2016

ROBUST IMAGE PROCESSING FOR CRYO-ELECTRON TOMOGRAPHY …

  • Upload
    others

  • View
    6

  • Download
    0

Embed Size (px)

Citation preview

ROBUST IMAGE PROCESSING FOR CRYO-ELECTRON TOMOGRAPHY

USING SPARSE PRIORS

A DISSERTATION

SUBMITTED TO THE DEPARTMENT OF ELECTRICAL ENGINEERING

AND THE COMMITTEE ON GRADUATE STUDIES

OF STANFORD UNIVERSITY

IN PARTIAL FULFILLMENT OF THE REQUIREMENTS

FOR THE DEGREE OF

DOCTOR OF PHILOSOPHY

Ka Hye Song

March 2016

http://creativecommons.org/licenses/by/3.0/us/

This dissertation is online at: http://purl.stanford.edu/rt579ds2887

© 2016 by Ka Hye Song. All Rights Reserved.

Re-distributed by Stanford University under license with the author.

This work is licensed under a Creative Commons Attribution-3.0 United States License.

ii

I certify that I have read this dissertation and that, in my opinion, it is fully adequatein scope and quality as a dissertation for the degree of Doctor of Philosophy.

Mark Horowitz, Primary Adviser

I certify that I have read this dissertation and that, in my opinion, it is fully adequatein scope and quality as a dissertation for the degree of Doctor of Philosophy.

Emmanuel Candes

I certify that I have read this dissertation and that, in my opinion, it is fully adequatein scope and quality as a dissertation for the degree of Doctor of Philosophy.

Farshid Moussavi

Approved for the Stanford University Committee on Graduate Studies.

Patricia J. Gumport, Vice Provost for Graduate Education

This signature page was generated electronically upon submission of this dissertation in electronic format. An original signed hard copy of the signature page is on file inUniversity Archives.

iii

Abstract

Cryo-electron tomography(CET ) is the only imaging modality that can image 3D density

maps of cells and viruses at their native state. It covers the resolution range of 4 - 8nm, and

can reach resolutions as low as 2nm and under by using subtomogram averaging. As such, it

bridges the gap between high resolution techniques such as the X-ray crystallography, and

low resolution ones such as the light microscopy. Thanks to this unique property, CET has

been extensively used to reveal the molecular organization of cellular structures in bacterial

cells and viruses.

Although CET can provide a higher resolution reconstruction of macromolecules than

other imaging modalities, there are several challenges to overcome to achieve high-quality

3D reconstructions of these structures. First, raw CET projections and tomograms have

very low SNR typically less than 1. Second, due to the limitations of the arrangement of

the sample holder and the transmission electron microscope, it is not possible to obtain

informative tomographic projections from all angles, which distorts the 3D reconstructions.

Due to these difficulties, conventional image processing techniques often fail to achieve

their goals. To make the image processing pipeline for CET more robust, utilizing the prior

information known about tomographic projections and reconstructions of interest is crucial.

In this thesis, two such examples are presented. One is an image in-painting algorithm using

l1 norm minimization to remove interferences in CET . This particular example exploits

the fact that CET projections are sparse in the discrete cosine domain (DCT ). The other

example is subtomogram averaging via nuclear norm minimization where we exploited the

observation that aligned structures span a very low dimensional space. Both examples

iv

deliver promising results even when original density maps are heavily distorted and covered

with significant noise.

v

Acknowledgments

One can easily forget that one has reached beyond one�s ability and expanded one�s horizon

thanks to all the support from the very people right next one. Here, I want to remember

the people who enabled me to finish my thesis successfully.

First of all, it was one of my life time fortune to have Prof. Mark Horowitz as my Ph.D.

thesis advisor. He has shown me a great example of an encouraging and supportive advisor

as well as an educator to all of us in his research group. I really appreciate that he has

offered me a research opportunity where I can grow together with group members with very

different background and interest. In addition, I need to mention that without his patient

guidance, I would not be able to achieve the same level of excellence in my Ph.D. degree as

I did.

Another important person who shaped my thesis is Prof. Emmanuel Candes. When I

started out my Ph.D. candidacy, he offered a course in the compressed sensing theory which

inspired me to apply the ideas to improve the image processing techniques in cryo Electron

Tomography. After taking the class, Emmanuel graciously continued to advise me on my

thesis and introduced me smart colleagues who also contributed to this thesis as well.

I would also like to give my deepest appreciation to Dr. Farshid Moussavi for sharing

his insights and experiences on developing image processing techniques for cryo Electron

Tomography. In addition, he has been a very supportive mentor as well as a dependable

friend who readily helped me to finish my thesis.

Another important supporters for my Ph.D. degree are Prof. Dwight Nishimura and

Prof. John Pauly. I am grateful for their support not only because they served in my

vi

defense committee but also they gave great classes in medical imaging. The first time I

learned about the mathematical principles of tomography was in Prof. Nishimura�s class,

and in Prof. Pauly�s classes, I went deeper into understanding how medical images are

formed and reconstructed.

Even though I am a EE major, thanks to Dr. Luis Comolli, I was able to publish in cryo

Electron Tomography. Although I did not have much background knowledge in structural

biology, Luis helped me to learn and appreciate how important advanced image processing

techniques are to study how cells work. Not only did he provide most of the data sets I have

used in this thesis, he also gave me advice on how to write to the audiences in the field. He

was very open to learn and use new image processing techniques that can bring out more

information from tomographic images, and motivated me to pursue directions that nobody

else has taken.

I am also very grateful to Dr. Ewout van den Berg who taught me how to use convex op-

timization techniques in the real world. Together with him, I was able to actually formulate

image processing problems into convex optimization problems that can be efficiently solved.

He has also been a great friend and a colleague with whom I spent great time working with.

I would like to also thank Dr. Fernando Amat for introducing me to cryo Electron

Tomography. Like Farshid, he has been working with Mark on cryo Electron Tomography

when I first joined Mark�s research group, and ever since then he has been a supportive

mentor and a great friend who I can discuss about a lot of things from research to life.

Before I became more interested in image processing and biomedical imaging, I was

interested in more theoretical statistical signal processing. While working on my Master�s

degree, Prof. Robert Gray graciously advised me on vector quantization and offered me

a research opportunity with him. I am also grateful that he supported me to pursue my

Ph.D.

Throughout my Ph.D. candidacy, one of my pleasures was having lunch meetings with

my VLSI research group members where we talked about a lot of different things happening

around the world. Often, Mark would start conversations by asking “What is new and

interesting?” and since all of the members are very unique, intelligent and fun to talk to,

vii

I had great time and learned so much from everyone. I really appreciated their fresh ideas

from different perspectives, and candid friendship.

Another group of people that I enjoyed meeting every week was my pseudo sisters, Yeo-

myoung, Eunah and Bora. We shared a lot of lows and highs in our lives and they truly

feel like my second family. And of course, I am fortunate to have all my friends I met on

Stanford campus. Some of whom also had spent my high school and college years together

as well, and they all became my extended family.

Last but not least, I would like to thank my family here in the U.S. and back in Korea.

First of all, I have to mention that my husband Erhan has been the most committed

supporter in my Ph.D. studies, and he has been there to hold me whenever I have doubts

on what I was pursuing. One of the memorable advice he gave was that ‘Research is like

farming. What you need to do is to get up everyday and do things diligently.’ This was

quite interesting analogy to me since I was thinking that research is something you need to

show off your skills and intelligence at. And all of his other advices also helped me going

through the whole journey with patience and tenacity. I am also grateful to our daughter

Elif for showing me a great example of being strong and resilient. She always reminds me

that there is going to be tomorrow and tomorrow is another day.

Among all my family members, I believe that I probably did not even think about doing

graduate studies without my parent�s encouragements. My mother has been especially vocal

about pursuing Ph.D. and how it can enrich my life and my father always showed deep trust

in my decisions. I am truly grateful to what they have provided for me to finish my Ph.D.

studies as well as to lead my own life as an independent being.

viii

Contents

Abstract iv

Acknowledgments vi

1 Introduction 2

2 Background 6

2.1 Structure of Electron Microscope . . . . . . . . . . . . . . . . . . . . . . . . 6

2.2 Electron Microscope Image Formation . . . . . . . . . . . . . . . . . . . . . 8

2.3 Challenges in Cryo Electron Tomography . . . . . . . . . . . . . . . . . . . 10

2.4 Image Processing Pipeline for 3D CET . . . . . . . . . . . . . . . . . . . . 12

2.5 Strengthening Image Processing Pipeline using Sparse Prior . . . . . . . . . 18

3 Digital In-painting via l1 Norm Minimization 22

3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

3.2 Theoretical background and previous work . . . . . . . . . . . . . . . . . . . 24

3.3 Proposed algorithm: Digital inpainting via compressed sensing . . . . . . . 29

3.4 Materials . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

3.5 Results and discussions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33

3.6 Conclusions and future work . . . . . . . . . . . . . . . . . . . . . . . . . . 48

4 Subtomogram Averaging via Nuclear Norm Minimization 49

4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49

ix

4.2 Previous Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

4.3 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53

4.4 Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59

4.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83

5 Conclusions 85

Bibliography 88

x

List of Tables

3.1 Quantitative comparison of inpainting fidelity using artificial fiducial markers

among different inpainting methods. . . . . . . . . . . . . . . . . . . . . . . 45

4.1 Clustering accuracy for GroEL & GroEL/ES and Helicases data sets . . . . 80

xi

List of Figures

1.1 Examples of cellular structures of bacterial cells and viruses using CET . . 3

1.2 Missing wedge problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4

2.1 Structure of electron microscope . . . . . . . . . . . . . . . . . . . . . . . . 7

2.2 Contrast transfer function examples . . . . . . . . . . . . . . . . . . . . . . 11

2.3 Missing wedge description . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

2.4 CET image processing pipeline . . . . . . . . . . . . . . . . . . . . . . . . . 14

2.5 Principles of CET . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

3.1 Survey of energy loss in compressed 2D-DCT reconstructions of tomographic

projections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

3.2 B. sphaericus S-layers and C. crescentus projections and theirDCT -compressed

2D-reconstructions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

3.3 Inpainting result of CET projections of isolated wild type B. sphaericus S-layer 35

3.4 CET reconstructions (tomogram slices) of isolated wild type B. sphaericus

S-layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

3.5 CET reconstructions (tomogram slices) of inpainted recombinant B. sphaer-

icus S-layer tomographic projections . . . . . . . . . . . . . . . . . . . . . . 38

3.6 Inpainting result of CET projections of C. crescentus . . . . . . . . . . . . 40

3.7 CET reconstructions (tomogram slices) of inpainted C. crescentus projections. 41

3.8 CET reconstructions (tomogram slices) of C. crescentus projections II . . . 42

xii

3.9 Inpainting artificial fiducial markers in CET reconstructions of isolated wild

type B. sphaericus S-layer and C. crescentus . . . . . . . . . . . . . . . . . 44

3.10 Comparison of the averaged B. sphaericus S-layer unit models using the

original and the inpainted tomograms. . . . . . . . . . . . . . . . . . . . . . 47

4.1 Nuclear Norm Minimization Example . . . . . . . . . . . . . . . . . . . . . 57

4.2 Cross sections and radon measurements of GroEL and GroEL/ES . . . . . 61

4.3 Cross sections of the original Helicases . . . . . . . . . . . . . . . . . . . . . 62

4.4 Noisy Radon measurements of Helicases . . . . . . . . . . . . . . . . . . . . 62

4.5 XYZ cross sections of the averaged GroEL structures . . . . . . . . . . . . . 67

4.6 XYZ cross sections of the averaged GroEL/ES . . . . . . . . . . . . . . . . 68

4.7 Alignment accuracy for GroEL and GroEL/ES data set . . . . . . . . . . . 69

4.8 Fourier shell correlation curves of GroEL and GroEL/ES . . . . . . . . . . 70

4.9 XYZ cross sections of the averaged Hel1 . . . . . . . . . . . . . . . . . . . . 73

4.10 XYZ cross sections of the averaged Hel2 . . . . . . . . . . . . . . . . . . . . 74

4.11 XYZ cross sections of the averaged Hel3 . . . . . . . . . . . . . . . . . . . . 75

4.12 Alignment accuracy for Helicases data set . . . . . . . . . . . . . . . . . . . 76

4.13 Pairwise angles between the error rotation matrices for Helicases at SNR = 1 77

4.14 Fourier shell correlation curves of Helicases . . . . . . . . . . . . . . . . . . 78

4.15 Confusion matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81

4.16 Distributions of the top ten normalized singular values of GroEL, GroEL/ES

and Helicases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

xiii

1

List of Symbols

CET Cryo-Electron Tomography

CCD Charge Couple Device

ET Electron Tomography

SNR Signal-to-Noise Ratio

2D Two-dimensional

3D Three-dimensional

FSC Fourier Shell Correlation

CTF Contrast Transfer Function

TEM Transmission Electron Microscope

DCT Discrete Cosine Transform

CS Compressed Sensing

SV D Singular Value Decomposition

Chapter 1

Introduction

Cryo-electron tomography(CET ) has become an important imaging technology in many

subfields of biology thanks to its ability to image intact cells, viruses and large molecular

complexes in their near-native frozen-hydrated state ([MNM04, LFB05, JB07, LJ09, Fer12,

GJ12]). Because samples are frozen within milliseconds in a controlled environment (humid-

ity and temperature), the aqueous suspension is vitrified without crystalline ice formation,

and samples’ subcellular structures and macromolecules are preserved with minimal dam-

age. Therefore, CET is an ideal tool to study their near-intact structure and relationship

to their native environment in the 3D volumetric reconstructions (known as tomograms)

at a resolution of ∼ 4 nm (∼ 2 nm with averaging techniques). This technique bridges

the resolution gap between the lower resolution imaging technologies, such as fluorescent

light microscopy, and the higher resolution ones such as X-ray crystallography and NMR

spectroscopy[GJ12, MS09, LFB05, MNM04].

Thanks to this unique property, CET has been extensively used to reveal the molecular

organization of cellular structures in bacterial cells and viruses. Four different examples of

such building blocks of cells are displayed in Figure. 1.1: (a) partially ordered hexagonal

arrangement of chemoreceptor arrays in Caulobacter crescentus cells in [KWS08] (b) two

predominant receptor conformations of the Escherichia coli serine receptor Tsr in [KWZ+08],

(c) the flagellar motor in cells of the Treponema primitia in [MLJ06], (d) distribution and

2

CHAPTER 1. INTRODUCTION 3

structure of ribosomes in Spiroplasma melliferum in [OFK+06]. In addition, CET has

been an indispensable vehicle to uncover AIDS virus envelope spikes [ZLBJ+06], HIV-1

capsids [SHR+15], the Ebola virus[BNR+12].

Figure 1.1: Examples of cellular structures of bacterial cells and viruses using CET :(a) Visualization of the three-dimensional structure (3D) and partially ordered hexago-nal arrangement of chemoreceptor arrays in Caulobacter crescentus cells. (b) The twopredominant receptor conformations of the Escherichia coli serine receptor Tsr derivedby 3D averaging, corresponding to the ’kinase-off’ and ’kinase-on’ signaling states. (c)The flagellar motor in cells of the Treponema primitia (scale bar 20 nm). (d) Distri-bution and structure of ribosomes (indicated by the arrow) in Spiroplasma melliferumwith green and yellow indicating higher or lower levels of accuracy of detection (scalebar 100 nm). Source: Jacqueline L.S. Milne & Sriram Subramaniam Nature ReviewsMicrobiology 7, 666-675 (September 2009) [MS09]

Although CET can provide a high resolution reconstruction of macromolecules, there are

two fundamental challenges that make it difficult to achieve high-quality 3D reconstructions

of these structures. First, raw CET projections and tomograms have a very low SNR—

typically less than 1. The image quality of CET reconstructions has certainly been much

improved with advances in imaging technologies ([MNM04, LFB05, JB07, LJ09]). However,

due to the low electron dose allowed in biological and organic materials to avoid significant

CHAPTER 1. INTRODUCTION 4

damage, the low SNR problem is not going to be easily overcome.

The other challenge is the missing frequency information (missing wedge or cone), which

is the consequence of the sampling geometries and the limited angular range within which

standard TEM stages can rotate (typically ± 72 deg). This is because the electron path

length within a sample increases as the tilt angle increases, which degrades the projection

image quality significantly beyond a certain tilt angle. The projection-slice theorem [Bra56]

states that the Fourier transformation of the Radon projection of a structure at a given

angle corresponds to a slice through the Fourier transformation of the structure. From this

it follows that projections at a limited range of angles lead to a wedge of missing data in

the frequency space. Direct reconstruction based on such partial data introduces severe dis-

tortion in the resulting density map, as illustrated in Figure 1.2. More specifically, features

are elongated along the direction of the missing wedge (or, in a direction orthogonal to the

stage rotation axis) and sometimes distorted features of undesired objects can create streak

artifacts that cast shadow on the biological specimen of interest. The missing frequency

domain can be minimized by different sample geometries and dual tilt data acquisition.

However, these limitations are also likely to remain for the near future on a vast range of

specimens and standard CET data acquisition techniques.

↓ F ↑ F−1

−→

Figure 1.2: Counter-clockwise from top left: original structure; its representation infrequency domain; the missing wedge in frequency due to limitations in tilt angles; andthe effect this has on the structure reconstructed by means of an inverse Fourier trans-formation.

CHAPTER 1. INTRODUCTION 5

Because of these fundamental image quality constraints, analyzing tomograms is still a

challenging task. Researchers have been trying to manage these problems and obtain high

resolution structures by applying various image processing techniques as well as intelligent

image analysis techniques. A set of these techniques are routinely applied and formed a

image processing pipeline for CET .

Although, all components in this pipeline exist, they can be improved to better cope

with low SNR, missing wedge as well as various artifacts that are present in CET . In this

thesis, we propose two techniques that can widen the scope of CET imaging by utilizing

sparse priors. More specifically, we propose robust image processing techniques that find

solutions in a restricted space where we assume certain statistical properties are valid for

CET images. In particular, we assume that CET images and reconstructions have a sparse

representation in a specific domain. This assumption is tested on two different techniques.

The first is to seamlessly remove high contrast objects to minimize artifacts in the 3D den-

sity map (digital inpainting) and the second is to find a set of refined reconstructions of

multi-class macromolecules (subtomogram averaging).

To demonstrate and evaluate the applicability of spare prior inspired images processing

techniques on enhancing CET image analysis: in Chapter 2, we first discuss how CET

images are formulated as well as the challenges extracting information from CET projec-

tions and reconstructions. Next, the image processing pipeline for CET and robust image

processing techniques using sparse priors, compressive sensing, and low rank matrix approx-

imation, are introduced. Chapter 3 then presents an algorithm that removes high contrast

artifacts in CET projections and reconstructions via compressive sensing. Chapter 4 de-

scribes how tomograms of sub-cellular structures are simultaneously aligned and clustered

via nuclear norm minimization. Finally, Chapter 5 discusses the future perspective of robust

image processing techniques for CET .

Chapter 2

Background

To gain insights about the signal characteristics of 2D cryo Electron Images as well as 3D

CET Reconstructions, this chapter provides a brief introduction to how images are formed

and acquired by Transmission Electron Microscope (TEM). For more background on how

Electron Microscope operates as well as how the images are formed by TEM , readers are

referred to Ch. 6 of [RK08] and Ch. 3 and 4 of [Fra06].

2.1 Structure of Electron Microscope

TEM is a projector that captures the 3D structure of the specimen of interest. As shown

in Fig 2.1, TEM is almost like an upside down light microscope that uses an electron

gun instead of light source and condenser lenses instead of optical lenses. In addition, the

entire path that electrons pass through is in a vacuum tube so the electron waves only

capture the interaction with the specimen. Highly coherent electrons are emitted from the

electron gun by thermal, Schottky, or field emission[RK08]. The directions of electrons

are controlled by multi stage condenser lenses to capture the specimen at various aperture

and magnification. The electrons that passed through the specimen are recorded by the

detectors and the resulting images can be directly recorded on a film or to a Charge Couple

Device (CCD) camera.

6

CHAPTER 2. BACKGROUND 7

Figure 2.1: Structure of electron microscope. Image Source: http://www.microbiologyinfo.com/differences-between-light-microscope-and-electron-microscope/

CHAPTER 2. BACKGROUND 8

2.2 Electron Microscope Image Formation

When electrons emitted from the electron gun, they interact with the ordered array of atoms

of the specimen and generate scattered electrons that are either elastically scattered or in-

elastically scattered. Among these, the elastically electrons carry the structural information

of the specimen. This interaction between the electrons and the specimen can be described

as a linear system, and we can see how the projected image of the specimen is formed on

the image plane by calculating the transferred wave out of the microscope according to the

Fourier optics. More specifically, the transfer function S(x, y) from the specimen can be

approximated in a following form:

S(x, y) = |S(x, y)| exp{iϕ(x, y)} (2.1)

where

ϕ(x, y) � − π

λeU

�V (x, y, z)dz. (2.2)

|S(x, y)| is the absorption and ϕ(x, y) is the phase shift caused by the specimen. V (x, y, z)

is the electron potential distribution of the specimen and λ is the electron wavelength, e is

the absolute electron charge, U is the accelerating voltage of the incoming electrons.

Thus, the wave diffracted by the atoms passing the specimen can be written as:

ψo(x, y) = S(x, y)ψs(x, y) (2.3)

where ψs(x, y) is the incoming planar wave function. According to the weak scattering

phase object approximation, we approximate the specimen transfer function as following.

S(x, y) ∼ (1− s(x, y))(1+ iϕ(x, y)− · · · ) ∼ 1− s(x, y)+ iϕ(x, y), s << 1 and ϕ << 1,

(2.4)

where s(x, y) is the inelastic scattering factor. By assuming that the incoming electron wave

is coherent and uniform magnitude, i.e. ψs(x, y) = 1, the Fourier transform of the exit wave

CHAPTER 2. BACKGROUND 9

from the specimen becomes

ψo(kx, ky) =1

M(δ(kx, ky)− s(kx, ky) + iϕ(kx, ky)). (2.5)

This exit wave travels through the microscope whose transfer function can be approximated

as following in the frequency domain.

P (kx, ky) = a exp(−iγ(kx, ky)) (2.6)

where γ(kx, ky) =π2(Csλ3r4 − 2Δzλr2), r =

�k2x + k2y, Cs is a spherical aberration of the lens and

Δz is defocus. This lens characteristic function is known as a contrast transfer function

(CTF).

Then the wave out of the microscope can be defined in the frequency domain as

ψi(kx, ky) = P (kx, ky) · ψo(kx, ky) (2.7)

=1

MP (kx, ky)(δ(kx, ky)− s(kx, ky) + iϕ(kx, ky)) (2.8)

where M is the magnification factor of the microscope. Then the resulting image on the

sensor can be written as

j(Mx,My) = M2|ψi(Mx,My)|2

= 1− 2

� �as(kx, ky) cos(γ(kx, ky)) exp 2πi(kxx+ kyy)dkxdky+

2

� �aϕ(kx, ky) sin(γ(kx, ky)) exp 2πi(kxx+ kyy)dkxdky. (2.9)

According to Eq. 2.9, in the electron microscope images, both amplitude contrast and

phase contrast terms are present. If we assume that the specimen of interest has very little

amplitude contrast because it is very thin, the resulting image can be further approximated

CHAPTER 2. BACKGROUND 10

as

j(Mx,My) �� �

2aϕ(kx, ky) sin(γ(kx, ky)) exp 2πi(kxx+ kyy)dkxdky (2.10)

= ϕ(kx, ky) ∗ 2a�sin(γ(kx, ky)), (2.11)

where ∗ is a convolution operator and · is a Fourier Transform operator. �sin(γ(kx, ky))is a Fourier transform of sin(γ(kx, ky)). Since ϕ(kx, ky) � − π

λeU

�V (x, y, z)dz, the final

image can be interpreted as a microscope transferred version of the projected 3D potential

function of the specimen along Z axis. For phase contrast signals, the CTF often only refers

to 2a sin(γ(kx, ky)). As seen in Figure 2.2, this function changes the phase of frequency

components periodically depending on the defocus valueΔz. When the defocus is 1, the mid

frequency components are mostly amplified and as the defocus increases, the lower frequency

components are amplified as well as high frequency components. However, due to the nature

of the sinusoid function, the amplification sign oscillates and this phase incoherence created

by CTF typically limits the resolution of the 3D reconstruction to the first zero crossing of

CTF where all the frequency components have the same sign. Although, we do not have

the phase incoherence problem when the defocus is 1, we also do not have enough gain

on the low frequency components that contains most of the large structural information.

As a result, the CET projection images with defocus 1 tend to be poor. They highlight

the amplified noise in the mid frequency range. In reality, to achieve good amplification

through out the frequency range that contains the structural information of the specimen,

often the defocus has to be well adjusted or multiple tomograms taken at different defocus

values have to be combined.

2.3 Challenges in Cryo Electron Tomography

One of the most critical challenges of extracting structural information of the specimen from

CET images is a very low signal to noise ratio. The signal contrast from the phase contrast

imaging produced by elastically scattered electrons is very small. To this small signal, a

CHAPTER 2. BACKGROUND 11

Figure 2.2: Plots for the microscope contrast transfer function sin(γ(kx, ky)) withdifferent defocus values. (a) Δz =

√Csλ = 1. (b) Δz =

√3Csλ =

√3. (c) Δz =√

5Csλ =√5. The x-axis corresponds to the spatial frequency up to p =

√2(Csλ

2)14 .

Figure from [Fra06] Ch. 3.

large level of noise is introduced by various sources, such as inelastic scattered electrons and

the detector noise added in the read out process. Normally the signal to noise ratio could

be boosted by increasing the illumination intensity. However, in CET this is not possible

since the amount of electrons that can be used without damaging samples is limited. In

addition, since we need to take multiple projections to create a tomographic reconstruction,

the electron dose is very limited per image. Acquisition techniques to improve SNR by

energy filtering and adjusting the defocus to steer the large gain frequency range of CTF

function to overlap with the frequency range that contains the most information about the

specimen have fundamental physical limits. As a result often conventional image processing

techniques have difficulty extracting the necessary information from CET images.

Another critical challenge to analyze 3D CET reconstructions is the missing wedge

problem. This refers to the fact that the projections of the specimen cannot be taken

throughout 180 degrees to complete all the frequency information. This is due to the

fixed electron beam geometry where the thin samples have to rotate around the sample

holders. When the thin sample is perpendicular to the beam facing the thin thickness,

the images have a higher SNR since the signal contrast from phase contrast imaging is the

strongest and the noise contributed by inelastically scattered electrons is minimal. However,

as the tilt angle increases, the actual thickness that electrons have to pass through increase

CHAPTER 2. BACKGROUND 12

quickly producing a lot more inelastically scattered electrons, thus resulting in low signal

contrast. Therefore, the projection images taken at angles over ±70 degrees usually does

not contribute much information to 3D reconstruction due to the limited amount of electron

dose. This will result in missing frequency information which appears as significant artifacts

in the 3D reconstructions thus complicating the analysis at the later stage.

We will use a 2D problem to demonstrate the effect of the missing wedge. Given an

original object, assume that our electron beam is parallel to left diagonal angle of the image,

therefore creating a missing wedge around this diagonal line as seen in Figure 2.3. In this

case, the reconstruction is missing the frequency components along the left diagonal line,

thus creating blurry boundaries. This distortion of the appearance of the specimen is a

critical challenge in CET . Since, often there are many copies of the same specimen that

exist in the sample set and each copy of the same specimen will appear differently if they

happen to orient in different directions. For example, in this artificial example of a head

phantom in Figure 2.3, when the beam is aligned along the right diagonal line, it will create

the blur in the right diagonal direction, resulting in a different image than the previous

one. This can complicate the conventional image analysis techniques such as alignment and

clustering.

2.4 Image Processing Pipeline for 3D CET

To cope with the challenges aforementioned and extract sound structural information of

the specimen of interest, an extensive image processing pipeline has been developed over

time. This procedure starting from acquiring electron microscope images to producing a

final analysis is known as an ‘Image processing pipeline for 3D CET ’, and it contains

higher level functional stages such as image acquisition, 3D reconstruction and analysis as

described in Figure 2.4.

CHAPTER 2. BACKGROUND 13

OriginalProjectionAngles

Reconstructions

Figure 2.3: Effect of missing wedge: Left: The original head phantom image. Cen-ter: Two different sets of Radon projection angles. Right: Reconstructions of the headphantom using the Radon projections produced by two sets of Radon projections withdifferent missing wedge.

CHAPTER 2. BACKGROUND 14

Figure 2.4: Four stages in ‘Image processing pipeline for 3D CET ’: 1. Image processingtechniques to correct for imperfections introduced in tomographic image acquisition. 2.3D Reconstruction and refinement techniques 3. Techniques to identify and segment thespecimen of interest and further refining the 3D structure of the specimen. 4. The finalgoal of the image processing pipeline

CHAPTER 2. BACKGROUND 15

2.4.1 Image Acquisition

This step involves taking a series of projections of a specimen for 3D tomographic recon-

struction as well as any post-processing required to enhance projection images. As shown

in the previous section, by taking a single image using an electron microscope, we can take

a projection of the specimen along Z axis as in Figure 2.5 B. To create 3D tomogram,

the specimen holder rotates along Y axis to capture projections of the specimen at multi-

ple angles. Typically, the sample holder rotates about ±70 degrees at 1 or 2 degree steps

apart as shown in Figure 2.5 A, and these projections are post-processed if necessary. Post-

processing in this step is very important to obtain a high resolution 3D reconstruction. One

of the most critical steps is aligning projection images. Often, the sample holder does not

keep the sample at the exact center when it rotates, the center of rotation axis is moving

in raw projection images. Since this motion error created by the sample holder rotation

can degrade the resolution of the 3D reconstruction, these images have to be aligned be-

fore reconstruction. Broadly speaking, there are two categories of alignment methods, one

utilizing a high contrast gold markers[BHE01, AMC+08] and the others utilizing the local

features of samples[SME+09, CDSAAF10]. Since the contrast of the biological samples in

CET is not very strong, often the markers are utilized to align projection images accu-

rately. After alignment, the projection images can be CTF corrected. As mentioned in the

previous section, CTF creates incoherent frequency components, therefore the resolution

of the final reconstruction is heavily limited. However, recent progress in CTF correction

techniques [FLC06, XMS+09, VSS+11] can help to further improve the resolution of the

final reconstruction. To avoid CTF correction and related issues, Zernike phase contrast

imaging [DKMN10] that has uniform transfer characteristics in the frequency domain can

be utilized instead of the conventional phase contrast imaging for CET .

2.4.2 3D Reconstruction

Once the projections are aligned and corrected, 3D reconstruction can be calculated using

well known methods such as filtered-back projection[Fra06], ART[GBH70] or SIRT[AK84].

CHAPTER 2. BACKGROUND 16

Figure 2.5: Principles of CET : A. Sample holder rotation. B. Projection through thespecimen to create each image. C. Back-projection from the images to reconstruct the3D density function. Image source: http://www.ana.unibe.ch/forschung/experimentellemorphologie/index ger.html

CHAPTER 2. BACKGROUND 17

The principals of these reconstructions are pictorially described in Figure 2.5 B and C. In B,

projection images contain the 2D slice of the 3D frequency components of a specimen are

taken, then by populating the 3D frequency characteristic function of a specimen using these

2D slices as in C. Then, by transforming this 3D frequency domain function to the image

space, the 3D density function of a specimen can be calculated[DK68, KS01]. All these

methods are implemented in a widely used CET software packages such as IMOD[KMM96]

and Bsoft[HCWS08]. For more comprehensive list of resources, refer to [Fer12].

To highlight certain features, 3D denoising algorithms can be applied after recon-

struction, which can be categorized into anisotropic diffusion based[FH01, FL05], wavelets

based[MHL+05], and linear filter based methods. Among these, anisotropic diffusion meth-

ods are most widely used and known to bring out the signal better than other methods for

CET [NAB+08].

2.4.3 Analysis

Once the 3D volume of a specimen is calculated and denoised, the structure of a specimen

can be analyzed using segmentation and subtomogram averaging. The goal of segmentation

is to isolate a set of pixels that contain specific structural features from the background.

There are two main properties of the structural components that can be utilized to separate

them. One is the property of a region such as color or texture, and the other is a prop-

erty of edges that separates two different structural components. The former is known as

region clustering or classification approach and the other is known as edge detection. For

CET , segmentation methods that utilizes the gradients of edges have been implemented

and have shown successes for certain types of specimen. For example, watershed method

for 3D structure described in [Vol02] has successfully segmented macromolecules and sub-

cellular structures. Another notable methods are to characterize the membrane features

as ridges [MSGF13] or structural tensors [MSGA+14] and detected pixels that has strong

membrane characteristics. Another type of approach is to combine template matching and

tracing to follow the pixels that contains particular characteristics conveyed in templates of

membrane[MHA+10] or filaments[RGH+12]. As shown here, there is not a single dominant

CHAPTER 2. BACKGROUND 18

approach that can segment any type of specimen yet. This is due to the very low signal

contrast and the low SNR as well as other artifacts such as missing wedge.

Subtomogram averaging method is a method to refine the 3D density map of a specimen

by combining multiple observations, and is a critical step to obtain a high resolution model.

This method requires individual observations to be classified and aligned within each class

at the same time. This challenging task has been tackled in various ways for CET and

existing methods are discussed in detail in Chapter 4.

2.5 Strengthening Image Processing Pipeline using Sparse

Prior

So far, the image processing pipeline described above has been successfully utilized to

produce high resolution 3D density maps of various cellular structures and to analyze whole

cell images. However, the final analysis step still suffers from the fundamental challenge of

processing CET images, very low SNR with artifacts which requires human intervention

to obtain good results. To reduce the needed intervention, the following chapters present

a way to remove artifacts using a sparse prior (Chapter 3) and to enhance subtomogram

averaging (Chapter 4).

Sparse prior can be interpreted as a prior knowledge that the signal of interest has a

sparse representation, which means that most information carried by the signal lies within

a low dimensional space. For example, a continuous sinusoidal signal does not have a sparse

representation in the time domain, but it has a sparse representation in the frequency

domain. Another example is that a set of the images of the same person taken under various

illuminations span a low dimensional space[BJ03]. This section introduces the compressive

sensing theory and the matrix rank minimization problem as a practical way to frame and

solve image processing problems using sparse priors.

CHAPTER 2. BACKGROUND 19

2.5.1 Compressed sensing

Compressed sensing (CS) theory claims that a signal can be reconstructed using a smaller

number of measurements than the number given by Nyquist sampling theory if the signal

meets the conditions below ([CDS98, CRT06a, CRT06c]).

1. A signal of interest is sparse in a known domain. (a sparse domain)

2. The measurements are sampled incoherently to the domain where a signal of interest

is sparsely represented.

For example, a sinusoid signal, y = sin(t), is sparse in the frequency domain because

it has only two nonzero frequency components. We can quantify sparsity, s (0 < s ≤ 1),

as a ratio between the number of nonzero elements of a transformed signal in the sparse

domain and the dimension of a signal in its original domain. Often, signals are not exactly

sparse but rather compressible. A sparse signal has few nonzero elements in the sparse

domain and all other elements are exactly zero. A compressible signal has few large el-

ements that contain the most energy and information of the signal in the sparse domain

and all other elements in this domain are not exactly zero but negligible compared to the

large elements. Hence, by using only these few large elements and discarding the smallest

ones, compressible signals can also be sparsely represented without much error. Due to

this characteristic of compressible signals, these signals can be regarded as sparse signals

in many applications. The classic examples of exploiting signal compressibility are JPEG

image compression schemes, which decompose images by DCT or discrete wavelet trans-

form (DWT) and compress them by setting the negligible coefficients in the DCT or DWT

domain to zero. In many applications of compressed sensing theory, compressible signals

can be reconstructed using the same methods used for sparse signals within bounded dis-

crepancy ([CRT06c]).

Incoherent sensing is a strategy to decrease the number of measurements needed to re-

construct a signal. It ensures that the information contained in the original signal is spread

CHAPTER 2. BACKGROUND 20

out over a set of incoherent measurement functions that are not correlated to each other

and have small correlation with unit vectors in the domain where a signal of interest is

sparse. For example, one of the most commonly used incoherent functions are a set of ran-

domly selected rows of discrete Fourier transform (DFT) matrix. Every DFT coefficient is

a linear combination of all signal elements in its original domain, thus, carries information

of all signal elements. According to CS theory, if a signal is sparse in the DFT domain, by

randomly measuring small number of samples of the signal in the original domain, we can

fully reconstruct the original signal with high probability because these random measure-

ments carry fraction of all signal elements [CRT06a]. If a signal is compressible in the DFT

domain, it still can recover the dominant components of the signal in the sparse domain

but not the negligible ones.

If a signal of interest is sparse in a known domain and its measurements are incoherent,

a sparse signal x0 ∈ Rn can be reconstructed by solving the convex optimization problem

below:

minx

||Ux||1 s.t. ||Ax− b||2 < � (2.12)

Here U ∈ Cn×n is a sparsifying transform and A ∈ Cm×n is a sensing matrix (m < n) where

m is the number of measurements. b ∈ Cm is the measurement of x0 using A (b = Ax0 if

there is no measurement noise) and � is the upper bound on the amount of noise present

in the measurement b.1 If the measurement is not corrupted by noise, � can be set to 0

and the solution of this optimization problem is the exact reconstruction of x0 if we have a

sufficient number of measurements [CRT06a].

2.5.2 Low rank matrix approximation

In the current image processing and computer vision problems, often we are interested in

learning an underlying structure that is common among the images while extracting the

discrepancies between the images. For example, when recognizing faces, it is very often

1If U is an identity matrix, A has to satisfy the incoherent sensing properties. Otherwise AU−1 has tobe an incoherent sensing matrix.

CHAPTER 2. BACKGROUND 21

that a face of the same person does not appear the same in different images due to dif-

ferent color balancing and illuminations in addition to other cosmetic changes. Therefore,

a robust face recognizer learns the underlying structure from a set of training images and

try to recognize the same person from test images[BJ03, WYG+09]. One way to effectively

learn the underlying common structure from a set of different observations is to find the

low rank approximation of a matrix whose columns are vectorized observations. This low

rank approach is more general than face recognition and appears frequently in many differ-

ent applications such as removing occlusion and restoring low-rank texture[ZGLM12], and

detecting anomalies in video surveillances[CLMW11].

The problem of finding a matrix with minimum rank can be formulated as:

minX

rank(X) s.t. ||A(X)−B||2 < �, (2.13)

where X is the row rank structure of the signals that measures to be B when applied to a

function A up to a noise limit �. If the measurement constraint is a convex set, it is well

documented that this problem can efficiently be solved using a nuclear norm heuristic for a

rank metric and convex programming [FHB01]. A nuclear norm of a matrix X ∈ Rm×n is

defined as

||X||∗ =min{m,n}�

i=1

σi(X), (2.14)

where σi(X) are the singular values of X.

Chapter 3

Digital In-painting via l1 Norm

Minimization

3.1 Introduction

The low SNR and the limited angle tomography make analyzing tomograms a challeng-

ing task. Researchers have been trying to manage these problems in various ways. One

way to overcome the low SNR is utilizing colloidal gold beads as fiducial markers to align

tomographic projections precisely. Researchers have also been actively looking for geneti-

cally modifiable labels, equivalent to green fluorescent protein in the light microscopy, that

can help detect specific proteins or macromolecules in noisy tomograms [DFG+09, MD07,

WML11]. The benefits of using markers and labels1 are clear, but there are also disad-

vantages to using them, especially when markers are larger than objects of interest and

significantly denser than the background. A common problem that results from the pres-

ence of colloidal gold beads throughout the sample is the occlusion of features or regions of

interest in projections. Another, perhaps more serious problem, is that 3D reconstruction

algorithms are unable to perfectly handle the large and abrupt contrast difference between

1In this chapter, markers and labels are used interchangeably because they create artifacts of the samenature. However, they have different roles in image analysis. Typically markers are used as landmarks foraccurate alignment and labels are targeted to certain macromolecules of interest for better localization.

22

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 23

a sample and fiducial markers. As a result, a halo or a shadow created by markers is elon-

gated by missing wedge effects, and projects onto neighboring (and not occluded) sample

features. In addition, very serious ripple effects are also very common, in particular when

clusters of markers are imaged only through part of the angular range. Finally it is some-

times impossible not to have markers around objects if they are used as labels.

There are numerous cases where markers themselves or their artifacts interfere with

detecting objects of interest or analyzing them. For example, in CET reconstructions of

intact bacteria, the goal of the image analysis can be to segment filaments through the

length of the cell, or to understand the coherence range of repetitive structures. These

tasks are not possible in the presence of dense artifacts that distort the shape of filaments

or the repeating structures. Also, in CET sub-volumetric averaging of macromolecules and

viruses, the ripples and halos caused by nearby fiducial markers, together with the distor-

tions created by the missing wedge, hinder the precise alignment. When macromolecules

are nanogold-labeled, these electron-dense labels themselves can bias aligning the molecules

as well. Finally, in CET of aqueous suspensions of inorganic nanoparticles, quantitative

analysis of reconstructed volumes can also be biased by fiducial markers. These phenom-

ena are especially pronounced when specimens are thin and markers lie very close to the

target objects, or when markers lump together. Without these artifacts, the specimen of

interest can be studied with greater clarity, and valuable data sets that may include unique

structures or conformations that are short-lived and difficult to capture can be saved from

the shadowing artifacts. Therefore, to utilize the high SNR markers and labels for ana-

lyzing CET reconstructions without side effects, it is necessary to come up with ways to

reduce artifacts created by dense markers, or to erase markers during the image analysis

(skeletonization, segmentation, sub-volumetric alignment, quantitative image analysis of

nanoparticles suspensions, etc).

In this chapter, we propose a new algorithm that removes high contrast objects by

digital inpainting (defined in Sec. 3.2), which utilizes the fact that CET projections can

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 24

be decomposed into a sparse representation in the DCT domain. We fill in the missing

regions occluded by high contrast objects in projections and demonstrate that the resulting

inpainted projections and the reconstructed volume show minimal artifacts around the

regions near high contrast objects. To help readers solidly understand our new algorithm,

we start by reviewing the existing digital inpainting methods and the CS theory which

underlies some of them. Then we show that CET projections are sparse in the DCT

domain and how we can exploit this information to inpaint occluded regions in projections

in the framework of CS. For evaluation, we first examine the inpainted projections and

tomograms of the surface-layer (S-layer) of Bacillus sphaericus, which has a natural short-

range order property that results in projections with a sparse power spectral density. After

confirming the success of our proposed method with S-layers, we move on to inpaint whole

cell data sets of Caulobacter crescentus, which are not naturally sparse but compressible

(defined in Sec. 2.5.1). In each experiment, we compare the perceptual quality of inpainted

projections and reconstructions produced by our algorithm to those produced by the existing

inpainting methods. To assist the perceptual comparison of different inpainting techniques,

we also compute image similarity metrics for the inpainted tomograms. To verify whether

inpainting can also help quantitative analysis of tomograms by removing fiducial marker

artifacts, we average the S-layer units located near the inpainted regions and compare

these averaged volumes to a reference S-layer structure. These qualitative and quantitative

evaluations show that our algorithm can reduce artifacts from high contrast objects and

reveal their neighboring regions more clearly than the conventional algorithms without

creating secondary inpainting artifacts.

3.2 Theoretical background and previous work

High contrast metal artifacts are not a new problem in tomography and there are solu-

tions developed for different imaging modalities that suffer from the same problem. In

the X-ray tomography community, many researchers have been looking for solutions for

this problem, known as metal artifact reduction(MAR) ([WSOV96, ZRW+00, DND+00,

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 25

XRY+05, WK04]) because metal prostheses in patients’ body create artifacts that hinder

accurate medical diagnosis. The most common strategy is to fill in the regions occluded by

metal objects with locally interpolated values or maximum likelihood estimates in projec-

tions. Recently, approaches that specifically formulate MAR in a constrained optimization

framework have been introduced as well ([GZY+06, ZWX10]). They often minimize the

total variation of the reconstruction because typical X-ray images are well approximated

as piecewise constant functions. Metal objects can also be removed in a reconstruction

domain. This approach is less popular because it requires tracing all the artifacts that are

created in the process of reconstruction.

While it appears that these algorithms developed in the X-ray tomography community

can be easily applied to remove high contrast artifacts created by electron-dense markers in

CET , differences in image characteristics make this task challenging. X-ray projections and

tomograms have a higher SNR than CET ones, and X-ray tomograms do not suffer from a

missing wedge problem. In addition to these differences in image quality, X-ray tomograms

have different image statistics due to different contrast mechanisms. As previously men-

tioned, they often contain piecewise constant features while CET tomograms have more

textural ones. Therefore, we need to tailor these methods to remove high contrast objects

for CET .

Researchers in the CET community have also been removing fiducial markers by fill-

ing gaps with locally interpolated pixel values or with random numbers drawn from local

statistics. Although these methods rely on local statistics, boundaries are usually visible

and corrected regions look artificial both in projections and reconstructed volume. This is

because these methods rely on a very simple image model that does not account for the

overall signal property of a projection.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 26

3.2.1 Digital Inpainting

Digital inpainting is a technique to fill in missing regions in images without visible traces

([BBC+01]). It is basically an interpolation scheme in which image models play an im-

portant role ([CKS02]). Images can be decomposed into two components, structure and

texture, and different image models are applied for these components because they exhibit

different statistical and perceptual characteristics. The first component, structure (also

known as cartoon), is defined as piecewise constant regions of images that also contain

sharp edges. The most common method to inpaint the structural part is a variational

approach, which tries to propagate information from existing parts of an image to missing

regions as smoothly as possible. This approach assumes that an image is a function that lies

in the Bounded Variational (BV) space; one way to implement this approach is to pose it

as a Total Variation (TV) minimization problem ([CS01]). The second component, texture,

is defined as rapidly varying or oscillating regions of images in terms of intensity. Texture

inpainting is often carried out by texture synthesis techniques that analyze the local statis-

tics of observed regions and predict missing ones. Structures and textures can overlay, and

there are approaches which try to inpaint both parts simultaneously ([BVSO03, ESQD05]).

Elad et al. in ([ESQD05]) showed how to use the compressed sensing technique to

digitally inpaint natural images. In this section, we show that CET projections satisfy the

conditions on signals that can be inpainted using the compressed sensing framework.

3.2.2 Sparsity of CET projections

In order to apply the compressed sensing framework to inpaint CET projections, we first

need to find a domain where projections have sparse representations. In general, it is very

difficult to analytically prove that projections of a specimen are sparse or compressible in

any known domain without any prior knowledge of an atomic structure of a specimen. How-

ever, the sparsity or compressibility of arbitrary signals can be demonstrated empirically

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 27

[LDP07, PL09]. For example, there have been extensive studies on the statistics of natu-

ral images and it has been generally accepted that natural images are compressible in the

DCT or DWT domain [TM02]. Following this practice, DCT has been used for sparsifying

textural parts and DWT for structural parts of natural images in [ESQD05]. This section

shows that CET projections are sparse in the DCT domain by evaluating the fidelity of

reconstructions of CET projections using only 1% and 5 % of their largest DCT coefficients

in terms of magnitude. We visually inspect the fidelity of compressed reconstructions and

also quantify the loss in signal energy from compression. We choose DCT as a sparsifying

transform because it sparsifies textural parts of images well, and CET projections contain

large texture parts due to the dense nature of biological specimens. DWT can be another

option. However, DWT is mostly used to sparsify piecewise smooth images ([SED05]), and

the resulting inpainted regions also tend to be flat and texture-less. Therefore, we focused

on DCT in this chapter to preserve the continuity in the image statistics in the inpainting

regions and the neighboring regions.

We surveyed the sparsity of 2D CET projections of B. sphaericus S-layers and of C.

crescentus of various shapes in the 2D-DCT domain using 704 projections from 11 tomo-

grams and 1527 projections from 13 tomograms respectively. Their projection angles vary

between −65◦ and 65◦. We evaluated the closeness between the compressed reconstruction

and the original image using normalized mean squared error (NMSE), which is defined as

||x0−xαDCT ||2F||x0||2F

. x0 is an original image, xαDCT is its compressed reconstruction using DCT

transform at compression rate at α and || · ||F is a Frobenious norm. This metric measures

the portion of signal energy lost from compression. When projections are compressed at

95%, it means that they are reconstructed by using only top 5% of 2D-DCT coefficients

in magnitude. The median loss of energy by compressing projections of S-layers at the

compression rate of 95% and 99% are 0.0011 and 0.0030 respectively. (See Fig. 3.1 plot (a)

and (b).) For the whole cells, the median NMSE is 0.0030 when compressed at 95% and

0.0051 when compressed at 99%. (See Fig. 3.1. plot (c) and (d).) This is very little loss

in signal energy given the high compression rate of 95 % and 99 %. Notice that we can

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 28

also indirectly see that the projections of whole cells are not as sparse as the projections of

S-layers by comparing the loss of energy from compression.

Upper 5% reconstruction(a)

Upper 1% reconstruction(b)

RMSE

(c)

RMSE

(d)

RMSE RMSE

Slayer

NumberofProjections

WholeCell

NumberofProjections

Figure 3.1: Survey of energy loss in compressed 2D-DCT reconstructions of tomo-graphic projections.: In (a) and (b) B. sphaericus S-layers ; In (c) and (d) C. crescentuswhole cells : In (a) and (c), NMSE of reconstructions using only top 5% magnitude2D-DCT coefficients. In (b) and (d) using top 1% coefficients.

When we visually compare the compressed reconstructions with the original projections,

it is very difficult to differentiate the reconstructions from the originals with naked eye. (See.

Fig. 3.2.) In addition to little signal energy loss and high visual fidelity, the statistics of the

residuals, differences between the original image and the reconstructed image, closely follow

the Gaussian statistics, shown in the last column of Fig. 3.2, which implies that the infor-

mation embedded in the discarded coefficients is mostly corrupted by noise.2 From these

2The noise of CET projections have Poisson statistics which can be safely approximated as a additiveGaussian noise when the mean value is sufficiently large.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 29

observations, we can conclude that projections of B. sphaericus S-layers and C. crescentus

are compressible in the DCT domain.

3.2.3 Incoherent sensing of CET projections

In addition to the sparsity of projections, we also need to prove that the domain where

a signal is measured is incoherent to the domain where the signal is sparse to justify ap-

plying compressed sensing theory for inpainting CET projections. According to [CRT06a],

randomly selected orthonormal basis functions form an incoherent sensing matrix. In our

application, the overall sensing matrix is AU−1, where A, is an identity matrix with missing

rows at the locations where pixels are randomly missing (because gold beads are randomly

located.) and U is a 2D-DCT matrix, which is orthonormal. Therefore our sensing matrix,

AU−1, satisfies the incoherent sensing condition. In addition, we are inpainting small re-

gions of CET projections, our CS-based digital inpainting algorithm satisfies the minimum

measurement constraint for faithful reconstruction.

3.3 Proposed algorithm: Digital inpainting via compressed

sensing

The inpainting procedure starts with identifying objects to be removed. In this chapter,

we choose colloidal fiducial markers (diameter ∼ 10nm) as target high contrast objects to

be removed.3 We find fiducial markers in all projections to avoid creating inconsistencies

between projections. To do this, we select fiducial markers in the projection taken at 0

degree tilt angle and track them through out the tilt series using imodfindbeads and track

functions in IMOD ([KMM96]). When IMOD fails to detect or track the markers to be

removed, we hand-label markers. Based on locations and a radius of fiducial markers, we

create a binary mask for each projection that discards pixels occluded by fiducial markers.

3This algorithm can inpaint any missing or corrupted regions in CET projections, such as X-ray damagedpixels, as long as the problem satisfies the compressed sensing constraints mentioned in Sec. 2.5.1.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 30

Figure 3.2: B. sphaericus S-layers (1) and C. crescentus (2-3) projections and theirDCT -compressed 2D-reconstructions: (a) Original projections at tilt angle 0 degree. (b-c) Compressed 2D-reconstructions of projection of (a) using only 5% and 1% of the largestDCT coefficients. (d) Normal probability plot of residuals of 5%, 1% reconstructions. Inthis plot, the residuals are plotted against the theoretical normal distribution to assessthe normality of the residuals. If the line is straight, the residuals follow the normalstatistics.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 31

Once the fiducial markers are all detected and masks are created, we can inpaint the

projections by solving the optimization problem in Eq. 2.12. Here x ∈ Rn is a vectorized

square image where n = N2 and it is the inpainted image when this problem is successfully

solved. N is the number of rows or columns of a square image. We define U ∈ Rn×n as a re-

arranged 2D-DCT matrix that carries out 2D-DCT on a vectorized image x. A ∈ {0, 1}n×n

is a diagonal matrix whose diagonal elements are binary. If the ith element of x is to be

inpainted, A(i, i) = 0 and A(i, i) = 1 otherwise. y ∈ Rm is a vector of observed pixel

values that are not occluded in the original image. � is an upper-bound on the amount of

discrepancy in unoccluded pixels while inpainting occluded ones. To preserve unoccluded

regions as intact as possible, we choose a small value of � from a range of [0.00001, 0.0005]

% of the l2 norm of the original image. By setting � as the noise floor in measurements,

we can also denoise projections ([CDS98]). However, we chose a very small � which is well

below the noise floor not to denoise projections because tomograms are often averaged to

reveal finer details that are often buried in noise.

Large scale convex optimization problems with a non-differentiable objective can take

long time to solve and the optimization problem formulated in Eq. 2.12 certainly belongs

to this category. Therefore, we use a l1 norm minimization solver, NESTA (a shorthand

for Nesterovs algorithm), which is specially developed for solving large scale compressed

sensing problems ([BBC11]). This solver minimizes a smoothened version of l1 norm by

a first-order method. This technique provides a good trade-off between computational ef-

ficiency and numerical accuracy. NESTA is also easy to configure because its parameters

can be intuitively determined. In addition to the sparsifying transform and sensing matrix

in Eq. 2.12, it requires a user to specify only two additional parameters. One is the final

smoothing factor, which determines the accuracy of the solution and the other is the error

bound, � in Eq. 2.12. In our case, the error bound is set to be small to avoid denoising and

the final smoothing parameter is set to be 10−5 to obtain accurate solutions. According

to our experience, if the final smoothing factor is small enough, smaller than 10−3, the

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 32

inpainted regions do not differ much perceptually while a smaller smoothing factor does

provide a solution with smaller l1 norm in the DCT domain. The computational time

increases as the smoothing factor decreases but the increase was in the order of tens of sec-

onds to solve a 2048 × 2048 matrix. NESTA has other parameters that have default values

and according to our experience, these default values provide accurate solutions within a

reasonable amount of time. Therefore, we have not fine-tuned these parameters.

Given these parameters, our inpainting algorithm takes about 5 minutes to inpaint an

image with 2048 × 2048 pixels using NESTA on a PC with Intel(R) Core(TM)2 Duo CPU

E7500 2.93GHz and 8 GB RAM. The whole algorithm including the solver, NESTA, is

implemented in Matlab(TM), and we used an optimized 2D-DCT , not the default one

provided in Matlab, to minimize the run time. Projecting on to and recovering from the

sparse signal domain is the most expensive part of the digital inpainting computation,

totaling about 55% of the runtime. This rather long processing time (compared to seconds

using IMOD) is the price for more sophisticated inpainting. Inpainting a set of images can

be naturally parallelized for each image because each inpainting task is independent of each

other. Therefore, the runtime for inpainting a whole stack of tomographic projections can

be kept constant around the time to inpaint a single image.

3.4 Materials

Two kinds of datasets are used to test CS-based digital inpainting for CET projections:

an in-vitro, ordered self-assembled macromolecular system, and intact bacteria. For the

former, one tomogram of wild type S-layer protein from B. sphaericus (wtSbpA) and two

tomograms of in-vitro recombinant (truncated sequence) S-layer protein (rSbpA) were used

for digital inpainting. 8 additional tomograms of in-vitro recombinant rSbpA (total of 11

tomograms) were used to survey the sparsity of S-layer projections. The intact bacteria con-

sist of three tomograms of C. crescentus intact cells were used for inpainting and additional

10 tomograms of C. crescentus (total of 13 tomograms) were used to survey the sparsity

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 33

of CET projections of intact bacteria. Cell cultures, cryo-grids preparation, cryo-EM data

acquisition and processing were done as previously described in [ACN+10, BCG+10]. B.

sphaericus sample preparation and data acquisition were also done using the identical meth-

ods described in [ACN+10, BCG+10].

All the fiducial alignments were done with RAPTOR ([AMC+08]) and the reconstruc-

tions were performed with the weighted-back-projection provided by IMOD ([KMM96]).

3.5 Results and discussions

In this section, we present inpainted projections and tomograms of isolated S-layers of B.

sphaericus and whole cells of C. crescentus using the proposed method and compare these

images with the results obtained using the existing methods in the CET community. To

the best of our knowledge, there have not been many inpainting algorithms introduced in

the CET community and the most commonly used ones are polynomial interpolation in-

painting implemented in IMOD and random noise inpainting. To evaluate the quality of

inpainting performed by our method, we also inpainted all data sets using IMOD (ccderaser

function) with polynomial orders 0,1,2,4 and Poisson random noise inpainting, which fills

missing regions with randomly drawn pixel values from an estimated Poisson distribution

based on the local statistics.5

First, we visually evaluate the inpainting performance of all inpainting algorithms men-

tioned above by comparing the original projections and tomograms with their inpainted

pairs. Comparing reconstructed volume is very important because inpainting should not

create any secondary artifacts while removing high contrast objects. Secondly, we examine

whether inpainting high contrast objects can actually reveal the structure shadowed by ar-

tifacts by introducing artificial markers and inpainting them. Lastly, we perform a small

4For each data set, the best performing polynomial order is selected for figures in this section.5This method is to replace marker regions with artificially created backgrounds that has the same statis-

tical property of the background noise in CET projections.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 34

scale S-layer subtomogram averaging experiment to see whether digital inpainting actually

enhances signals of interest in the uncovered regions, and to verify that these part of tomo-

grams can produce meaningful quantitative analysis. The result shows that the projections

inpainted by the proposed method are visually more natural and the corresponding regions

in tomograms show less severe artifacts. There are a few cases where we can see halos of

fiducial markers in reconstructions; however, in those cases, our method still creates less se-

vere artifacts than the existing methods. In addition, artificial fiducial marker experiments

and subtomogram averaging experiments confirm that digital inpainting enhances signals

of interest shadowed by high contrast artifacts by removing them, and both visual and

statistical fidelity of CS-inpainted volume is superior to the ones produced by the existing

methods.

3.5.1 Surface layers

S-layers of B. sphaericus are naturally formed 2D paracrystals that have a sparse structure

in the frequency domain, and their CET projections are also sparse in the DCT domain.

Therefore, our method should be able to inpaint the occluded regions with minimal artifacts.

We confirm that this hypothesis is true by visually inspecting the inpainted projections in

Fig. 3.3. The proposed method inpaints the areas occluded by fiducial markers without

artificial boundaries and the overall regions seem very natural while other methods simply

replace fiducial markers with artificial disks. This difference stems from the unrealistic

image models that the other methods are based upon. The IMOD model assumes that CET

projections locally form polynomials and random noise inpainting assumes that pixel values

within a close range in CET projections are drawn from an I.I.D. Poisson distribution. On

the other hand, our CS-based inpainting method does not create obvious discontinuities in

projections because it assumes that CET projections consist of a few DCT basis functions

which are continuous within the support of images.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 35

Figure 3.3: CET projections of isolated wild type B. sphaericus S-layer: (a) Beforeinpainting: Fiducial markers are occluding S-layer surfaces. (b) CS inpainting: Markersare removed and no artificial boundaries are visible. (c) IMOD inpainting: Polynomialorder 0 is used, and the markers are replaced by a constant value disk. (d) Random noiseinpainting: The markers are replaced by I.I.D. Poisson random variables, which do notpreserve the continuity of the local pixel values.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 36

Because our algorithm can inpaint CET projections without visible boundaries and ar-

tificial traces, we also expect that 3D tomograms reconstructed from projections inpainted

by our method will have less severe artifacts than those produced from projections inpainted

by the conventional methods. However, the apparent differences among the inpainted re-

gions of 3D tomograms produced by these different algorithms are not as striking as the

differences among the corresponding areas of the inpainted projections. We can see that

all methods remove markers without many visible traces of inpainting in the reconstructed

S-layers in Fig. 3.4. Although all inpainted reconstructions appear very similar to each

other, those inpainted by IMOD and random noise show more visible boundaries than the

tomogram inpainted by the proposed method. This phenomenon is more clearly visible in

Fig. 3.5 Slayer 1 plots, where we only have two fiducial markers that occlude the surface

of the S-layer. In Fig. 3.5, plot Slayer 1 (b), our method seamlessly unveils the underlying

S-layer unit without any artifacts while the other methods create obvious traces of inpaint-

ing. However, when there are many gold beads that are lumped together and occluding a

large area, all methods are rather unsuccessful at recovering the underlying structure. In

Fig. 3.5, in the second row, the surface of the S-layer is not exactly recovered by any of the

inpainting methods, instead they appear to shrink the area affected by the high contrast

artifacts. This different inpainting between the small and the large inpainted regions seen in

Fig. 3.5 can be attributed to the paracrystalline nature of S-layers. Unit cells of paracrys-

tals preserve their regularity only locally not globally. Therefore, if the occluded region is

small, our method can recover the regular structure in the occluded region without much

difficulty. If the occluded region is large, the unit cells around the occluded region do not

carry enough information to recover the missing structure because the regularity of these

unit cells are not preserved for a long range. Although any of the methods cannot suppress

the obvious inpainting artifacts when the inpainted area is large, the characteristic S-layer

lattice is more clearly visible and the boundaries of the recovered region look more seamless

in the tomogram inpainted by the proposed method than in the tomograms inpainted by

the other methods.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 37

Figure 3.4: CET reconstructions (tomogram slices) of isolated wild type B. sphaericusS-layer: (a) Before inpainting: The high-contrast artifacts created by colloidal fiducialmarkers are occluding surrounding regions. (b) CS inpainting: Most of the markers andtheir artifacts are removed. The structure of the S-layer around and inside the inpaint-ing regions seems to be better recovered than in the other inpainted reconstructions in(c) and (d). No obvious discontinuities are visible around the inpainting regions. (c)IMOD inpainting: Polynomial order 0 inpainting. Some inpainted areas look artificiallymonotone and smoothened. (d) Random noise inpainting: No artificially smoothenedareas are visible but the structure of the S-layer around and inside the inpainted regionsdoes not seem to be well recovered.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 38

Figure 3.5: CET reconstructions (tomogram slices) of recombinant B. sphaericus S-layer: In the first row, (a) Before inpainting: The high-contrast artifacts created by twocolloidal fiducial markers are occluding surrounding regions. (b) CS inpainting: Themarkers and their artifacts are removed. The structure of the S-layer around and insidethe inpainting regions appears to be better recovered than in the other inpainted re-constructions in (c) and (d). No obvious discontinuities and artifacts are visible aroundthe inpainting regions. (c) IMOD inpainting: Polynomial order 0 inpainting. The in-painted areas look artificially smoothened. (d) Random noise inpainting: No artificiallysmoothened areas are visible but the structure of the S-layer around and inside the in-painted regions does not appear to be well recovered. In the second row, (a) Markers arelumped together and occluding a large region. (b) All the markers and their artifacts areremoved. The structure of the S-layer around and inside the inpainting regions appearsto be slightly recovered. No obvious discontinuities are visible around the inpaintingregions. (c) The inpainted areas look artificially monotone and smoothened. (d) Noartificially smoothened areas visible but the structure of the S-layer around and insidethe inpainted regions does not appear to be well recovered.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 39

3.5.2 Whole cells

The ultrastructure of C. crescentus, our model for bacteria in general, is not sparse in any

known domain. However, we have empirically demonstrated that CET projections of C.

crescentus are compressible in the DCT domain; therefore, we expect that our inpainting

method can remove high contrast objects without creating severe artifacts, as it does with

S-layers. This hypothesis is validated by the seamlessly inpainted projections in Fig. 3.6.

The projections inpainted by our CS-based method appear to be more natural to naked

eye, with less visible artifacts, than the projections created by other methods.

As seen in the previous section on S-layers, well-inpainted projections result in tomo-

grams with less severe artifacts. In Fig. 3.7 and Fig. 3.8. we see markers whose high

contrast artifacts cast shadow on the cell membrane. The challenge here is to inpaint the

occluded regions without disrupting the integrity of the neighboring cell membrane. Since

the membrane creates discontinuities that are not very sparse in the DCT domain, the oc-

cluded membranes are not fully recovered after inpainting in the projections as in Fig. 3.6.

However, we still do not find any artificial boundaries around the inpainted regions in the

CS-inpainted reconstructions in both Fig. 3.7 and Fig. 3.8 because our method inpaints

missing pixel values by estimates that blend well with the surroundings without disconti-

nuities, while other methods simply replace markers with artificial disks. If the markers

occlude the membrane in a small number of projections, the back-projected volume does

recover the membrane from the unoccluded projections. In this case, the contribution from

the inpainted membrane averages out as long as inpainting does not create gross errors.

Therefore, we can see membranes clearly without any artifacts in tomograms as in Fig. 3.7

and Fig. 3.8. If markers are occluding specific regions containing sharp edges in most of

the projections, these edges may not be clearly visible in the reconstructions because our

method does not necessarily recover the edges in the projections.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 40

Figure 3.6: CET projections of C. crescentus: (a) Before inpainting: Fiducial markersocclude the cell membrane. (b) CS inpainting: Markers are removed and no artificialboundaries are visible. At the projection angle of -30 degree, we see markers that lieon the different plane of the tomogram. These markers were not labeled to be removed.(c) IMOD inpainting: Polynomial order 0 was used, and the markers are replaced by aconstant value disk. (d) Random noise inpainting: The markers are replaced by I.I.D.Poisson random variables which do not preserve the continuity of the local pixel values.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 41

Figure 3.7: CET reconstructions (tomogram slices) of C. crescentus I: (a) Beforeinpainting: Markers are casting shadow on the cell membrane. (b) CS inpainting: Allthe markers and their artifacts are removed and the nearby membrane areas are clearlyvisible without noticeable artifacts. (c) IMOD inpainting: Polynomial order 1 inpainting.The inpainted areas look artificially monotone and smoothened but the membrane is alsovisible. (d) Random noise inpainting: The halos from inpainting still occlude the nearbymembrane. (1-3) Different C. crescentus data set ID.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 42

Figure 3.8: CET reconstructions (tomogram slices) of C. crescentus II: (a) Beforeinpainting: Markers are casting shadow on the cell membrane. (b) CS inpainting: Allthe markers and their artifacts are removed and the nearby membrane region is clearlyvisible without noticeable artifacts. (c) IMOD inpainting: Polynomial order 1 inpainting.The inpainted areas look artificially monotone and smoothened but the membrane is alsovisible. (d) Random noise inpainting: The halos from inpainting still occlude the nearbymembrane.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 43

3.5.3 Artificial fiducial marker experiment

The purpose of inpainting high contrast objects is to reduce the associated artifacts and

reveal the underlying structure of a specimen. To verify whether removing fiducial markers

does uncover the nearby structure, we create artificial fiducial markers on actual CET

projections and remove them by inpainting. We then compare the revealed structure to

the corresponding one in the original tomogram visually as well as using image quality

metrics, NMSE and mutual information (MI). Artificial markers are created in projections

by applying a 2D Gaussian mask at locations calculated according to the single-axis data

acquisition geometry. In Fig. 3.9, we see the 3D reconstructions of the artificial markers

as well as the original tomograms and the inpainted reconstructions. While all inpainting

methods can remove the high contrast artifacts in tomograms, the proposed method creates

less obvious artificial traces from inpainting than others. Also the areas corrupted by

artificial markers appear to be smaller in the CS-inpainted tomograms than in others,

thus revealing the areas that are right next to the fiducial markers. Being able to contain

inpainting artifacts within a small area is especially important when markers are located

very closely to the objects of interest.

The numerical fidelity of the inpainted volumes are also evaluated by two image quality

metrics, NMSE and MI. NMSE measures the discrepancy in each pixel value relative to l2

norm of the original volume, and MI measures the discrepancy in the overall pixel value

distribution between two volumes. According to the information theory, MI quantifies the

amount of information one volume contains about the other ([PMV03]). The larger the MI

between two volumes, the more information of one volume can be explained by using the

information in the other. All metrics are computed using only voxels near the inpainted

regions. The size of a cropped cube-shape volume is 3 times of the diameter of the inpainted

fiducial marker. In Table 3.1, while all inpainting produces a similar amount of pixel-wise

discrepancies according to the median and mean NMSE, the CS-inpainted volumes have

higher MI than other inpainted volumes. This result can be explained by the nature of these

two metrics. While NMSE is blind to the overall composition of volumes, MI compares the

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 44

Figure 3.9: Artificial fiducial markers in CET reconstructions of isolated wild typeB. sphaericus S-layer (bead 1 and 2) and C. crescentus (bead 3 and 4). From left toright, (a) Original tomogram. (b) Artificial Fiducial markers. (c) CS-inpainted. (d)IMOD-inpainted(Polynomial order 0) (e) Random noise inpainted. The CS-inpaintedreconstructions have unnoticeable boundaries while IMOD and random noise inpaintedreconstructions have artificial regions with noticeable boundaries. The proposed methodalso recovers some structure of the S-layer occluded by the fiducial markers.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 45

NMSE Mutual InformationBead ID CS IMOD Random CS IMOD Random

3DBrdU #1 0.0001 0.0001 0.0002 0.7066 0.5735 0.34403DBrdU #2 0.0001 0.0001 0.0002 0.4035 0.4306 0.27983Dwt1 #1 0.0004 0.0004 0.0004 4.2246 4.1835 4.11113Dwt1 #2 0.0003 0.0004 0.0004 4.1326 3.9970 3.9889Slayer9 #1 0.0003 0.0004 0.0003 1.7839 1.4507 1.7569Slayer9 #2 0.0003 0.0004 0.0003 1.8236 1.4791 1.8002

mean 0.0003 0.0003 0.0003 2.1792 2.0191 2.0468median 0.0003 0.0004 0.0003 1.8038 1.4649 1.7786

Table 3.1: Quantitative comparison of inpainting fidelity using artificial fiducial markersamong different inpainting methods. According to NMSE, all inpainting methods resultin a similar amount of pixel-wise discrepancies on average. However, according to MI,CS-inpainted volumes contain more information that can explain the original volume inmost of the cases.

statistics of pixel values in volumes, and CS-based inpainting does fill in the missing pixel

values using continuous waves that are inferred to preserve the global consistency of each

projection, which also resulted in the consistent tomographic reconstruction.

3.5.4 S-layer subtomogram averaging experiment

One of the advantages of removing fiducial markers is that post processing algorithms such

as subtomogram averaging are less affected by false patterns generated by high contrast

artifacts. To verify this conjecture, we average 77 carefully selected subtomograms of S-

layer units that are right next to fiducial markers from a single tomogram. We choose a

small number of subtomograms for two reasons: 1. We only choose those units near fiducial

markers, and 2. This experiment is intended not to solve a structure of S-layer units but

to see whether inpainted volumes actually contain enhanced signal from the specimen of

interest than the volumes with high contrast artifacts. We repeat the identical averaging

process using the same selection of subtomograms from the original tomogram and the

inpainted tomograms processed by the CS, IMOD and random inpainting methods. The

original tomogram was taken under ∼ −9µm defocus and it has a pixel size of 0.6854 nm.

Each subtomogram has the size of 77 × 77 × 77 pixels (∼ 53nm each side), and they are

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 46

centered on the green spheres shown in the original tomogram (Fig.3.10 plot (a)) and in

the CS-inpainted tomogram (Fig.3.10 plot B). The centers of subtomograms are located

20 voxels (∼ 14nm) away from the fiducial markers. We used the maximum likelihood

method ([SMVC09]) implemented in the Xmipp package for the iterative alignment and

averaging, and the initial reference was chosen by the algorithm. Since we use a single S-

layer reconstruction whose surface is relatively flat without too much shape distortion, we

search only a maximum of 45 degrees around each axis by 10 degree steps for 10 iterations.

In Fig.3.10 plot (b), the subtomograms from the original volume with fiducial markers

failed to align the S-layer structure because of the interfering artifacts created by nearby

fiducial markers. However, when the subtomograms at the same locations from the inpainted

volumes are iteratively aligned and averaged, they can reveal the grid arrangement of S-

layer units. In Fig.3.10 plot (c), We can see that all three averaged subtomograms using

the inpainted volume processed by the CS, IMOD and Random inpainting methods quickly

converged to a meaningful first averaged model or template that could be further refined,

and it is very difficult to tell the differences between these averaged models with naked

eye. Therefore, to achieve more accurate assessment, we compare these averaged models

to an independent averaged S-layer structure derived from different B. sphaericus S-layer

data sets, using Fourier Shell Coefficient criterion. As shown in Fig.3.10 plot (d), we can

see that the averaged S-layer volume from the CS-inpainted tomogram actually has higher

FSC than the ones from the other inpainted volumes especially near the spatial frequency

that corresponds to the periodicity between S-layer units. The periodicity of S-layer unit

is reported to be ∼ 13.5nm, and the diameter of each unit is ∼ 9.375nm. Also, the CS-

inpainted volume can produce an averaged S-layer structure comparable to the one using

the subtomograms from the fiducial marker free regions of the tomogram in terms of FSC.6

We computed FSC using bresolve from Bsoft ([HCWS08]). This result confirms that digital

6We also averaged 77 subtomograms from the fiducial markerless regions in the same tomogram usedfor this experiment described in the paragraph above. Except the location of the subtomograms, all otherexperiment parameters are controlled to be the same as the ones used for averaging subtomograms nearfiducial markers and their inpainted regions.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 47

Figure 3.10: Comparison of the averaged B. sphaericus S-layer unit models using theoriginal and the inpainted tomograms. : (a) S-layer unit selections( marked by greencircles) in the original volume before inpainting (left) and in the CS-inpainted volume(right): The inset in the left hand side plot contains the fiducial markers that create theartifacts (located 20 voxels above the S-layer). (b) Iso-surfaces of the averaged S-layermodel from the original tomogram (left) and the CS-inpainted one (right): It is obviousthat the alignment procedure cannot align the S-layer lattice in the original tomogramwhile it can produce a fair initial S-layer lattice model from the CS-inpainted tomogram.(c) Averaged S-layer volumes in gray scale volume slices: 1. The original raw volumefailed to converge to a S-layer lattice structure. 2-4. All inpainted volumes convergeto a S-layer lattice structure with subtle differences. However, we can still observe thatthe averaged volume using the CS-inpainted tomogram show clearer barrel shape of aS-layer unit. (d) Quantitative comparison using FSC: Near the periodicity of S-layerunits, spatial frequency = 1/13.5 ∼ 0.0741(1/nm), the averaged S-layer volume from theCS-inpainted volume has the highest FSC which is very close to the averaged S-layervolume using fiducial markerless part of the original volume. The vertical line refers tothe calculated location of the first zero crossing created by the contrast transfer function(CTF) with ∼ −9µm defocus for our microscope.

CHAPTER 3. DIGITAL IN-PAINTING VIA L1 NORM MINIMIZATION 48

inpainting can enhance signals from the structures occluded by removing high contrast

artifacts; in addition, by utilizing the structural information such as the sparsity of CET

projections in the DCT domain, our CS-based inpainting method can actually remove

artifacts more intelligently, which helps to align and refine the underlying structure.

3.6 Conclusions and future work

We have presented a digital inpainting algorithm based on compressed sensing theory, which

effectively removes artifacts created by high contrast objects such as fiducial markers. Our

algorithm estimates missing pixel values that sparsify the inpainted CET projections in

the DCT domain. This technique shows better performance in removing artifacts than the

conventional algorithms because it exploits the sparse or compressible nature of CET pro-

jections in the DCT domain. The proposed method successfully fills missing regions with

minimal distortions or artifacts, such as halos, shadows or ripples, in both projections and

reconstructions. It always provides projections and reconstructions that are perceptually

better than or as desirable as those provided by the conventional methods.

Without these adversary artifacts created by high contrast objects such as fiducial mark-

ers or individual macromolecular tags, specimen areas in the immediate vicinity of these

objects can be studied with greater clarity and the specimen can be used for further analy-

sis. For example, we can crop sub-volumes from these regions and use them for alignment

and averaging. This is very well illustrated in the case of the S-layer shown in Fig. 3.10.

Moreover, if our goal is to compute sub-volume averaging of tagged macromolecules, we can

inpaint their high contrast tags that can bias the alignment of the target macromolecules.

Chapter 4

Subtomogram Averaging via

Nuclear Norm Minimization

4.1 Introduction

To address the low SNR and the missing wedge distortion of cryo-ET reconstructions, one

acquires many sets of projections from multiple copies of the structure of interest and com-

bines them into a single reconstruction: by averaging multiple observations we can reduce

noise and thereby effectively raise the SNR. Moreover, because the specimens in the sample

holder are flash frozen in essentially random orientations, their projection data will have

missing wedges in different locations when aligned to a common orientation. Consequently,

combining sufficiently many observations allows us to fill in the frequency domain and to

eliminate the missing wedge. Although the general idea is simple, this process is by no

means trivial and has itself two main difficulties. The first difficulty is that the orientations

and the centers of each of the macromolecular complexes in the sample are unknown. The

second difficulty, when the sample contains a mixture of heterogeneous structures, is that

the class or type of each macromolecular complex is unknown. The subtomogram averaging

problem thus consists of two sub-problems: alignment and clustering/classification.

49

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION50

In this chapter we propose a new approach for the reference-free alignment and 3D to-

mographic reconstruction from a heterogeneous data set that may be partial in the sense

of the missing wedge. The approach is based on rank minimization and its nuclear-norm

relaxation, and we therefore refer to it as NN-CET (nuclear-norm based cryo-ET). The

inspiration of this algorithm comes from the robust image alignment work by Peng et

al. [PGW+12], in which a set of translated and rotated images with distortions, such as

different lighting conditions and partial occlusion, is aligned by means of iterative lineariza-

tion and nuclear-norm minimization. Our proposed algorithm is novel in several respects.

In particular, it

1. jointly solves the alignment and clustering problem;

2. introduces matrix rank as a collaborative metric to evaluate the alignment and clus-

tering accuracy;

3. estimates the alignment parameters by solving a series of convex optimization prob-

lems; and

4. is reference-free and does not even generate a reference model internally.

While this approach is promising, its current implementation has high computational

requirements and some convergence issues. As a result we were not able to align volumes

larger than 32 × 32 × 32 voxels and it requires coarse pre-alignment before starting our

algorithm.

To understand the context of this approach, we review existing works in Section 4.2 and

then provide a detailed discussion and derivation of the proposed method in Section 4.3. To

examine the effectiveness of the new method we describe experiments on two reconstruction

tasks using synthetic data sets in Section 4.4 and discuss the current algorithm’s limitations.

4.2 Previous Work

In the case of reconstructing a single structure, subtomogram averaging reduces to an

alignment problem, which has been well studied. Optimal alignment parameters are often

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION51

recovered by means of an exhaustive search over a predefined discrete range of parameters

[KFB+13, SMVC09, FMZ+05]. These methods work well for coarse alignment, but can

suffer from high computational complexity for accurate alignment where the search is over

a larger and finer parameter grid. Several approaches based on harmonic analysis have been

proposed to reduce this complexity [CPH+13, XBA12, BSL+08].

Single particle methods can also provide very high resolution density maps of isolated

and purified macromolecules. However, these methods require a reliable reference against

which all electron micrographs are aligned in order to find the correct alignment angles as

well as classes if the sample micrographs contain multiple classes. Since macromolecules are

purified, their structure in their native state cannot be easily resolved using this method.

Often, however, the structures of interest are not homogeneous, especially when they

are in their native state [MS09]. For the high-resolution reconstruction of each structure

it is therefore critical that both classification and alignment be accurate. Unfortunately,

these two problems are intricately linked: without good classification it is hard to find

meaningful alignment parameters; likewise, accurate classification is extremely challenging

without good alignment parameters. Pairwise comparison between subtomograms may seem

to be a solution, but this is compounded by an extremely low SNR and the missing wedge

problem. Various methods have been proposed to overcome these challenges by jointly

or alternatingly solving these subproblems. Indeed, such methods have been successfully

applied to uncover high-resolution macromolecular structures in their native state [BNR+12,

KCW+10, BS09, NSP+06, ZLBJ+06, FMZ+05]. Nevertheless, the subtomogram averaging

problem is far from solved and remains an active area of research.

By far the most widely used method for the subtomogram averaging problem is to solve

the clustering and the alignment problems iteratively by alternating between them (see

for example [KFB+13, BSL+08, SB08, Win07, FMZ+05] among many others). During the

alignment stage, these algorithms try to find the alignment parameters that are optimal

with respect to the current clustering. In most of these algorithms, this is done by means of

class reference models: a set of alignment parameters is determined by minimizing the dis-

similarity between a class reference and a subtomogram. This is complicated by the missing

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION52

frequency information in each subtomogram, and complex similarity metrics that take this

into account have been proposed. Perhaps the most commonly used metric between two

reconstructions with missing wedge is the constrained correlation between the angular com-

ponents that are observed in both reconstructions[CPH+13, XBA12, BSL+08, FMZ+05].

Another metric is proposed by Kuybeda et al. [KFB+13], who use the nuclear norm as a

metric to evaluate the overall alignment.

The reference classes used in alignment are typically determined by one of two ap-

proaches: they can either be calculated internally by averaging randomly selected subtomo-

grams in the data set, or they can be three-dimensional density maps of the target structures

that have been estimated externally. Once subtomograms are aligned to a reference, the

next clustering iteration can start, where all subtomograms are reclustered using well-known

methods such as hierarchical clustering or other methods based on multivariate statistical

analysis [vHF81].

One method that does not alternate between the alignment and clustering is the maximum-

likelihood (ML) approach by Scheres et al. [SMVC09]. This approach explicitly assumes

that tomograms contain heterogeneous structures and incorporates alignment and cluster-

ing into a single likelihood optimization problem. Nevertheless, this approach still relies

on a metric that quantifies the pairwise difference between a reference and an observation.

In addition, like many of the algorithms mentioned above, it also exhaustively searches for

rotation and translation parameters within a user-defined range.

Image alignment and clustering problems are also actively researched in various other

fields [PMV03, ZF03, Bro92]. In the computer vision community, for instance, there is an

increased interest in analyzing uncontrolled images and videos shared by users online. Ana-

lyzing such images shares some of the difficulties encountered in the subtomogram averaging.

First, user-generated images are heterogeneous and not aligned. Second, these images can

exhibit various degrees of missing information due to occlusion or varying lighting con-

ditions. To align and classify these natural images, various robust alignment algorithms

have been introduced [PGW+12, WN12, ClTCB11, lTB03, VGS08]. Many of these algo-

rithms exploit certain inherent low-dimensional structures in the data set, such as sparsity

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION53

or low-rankness. By taking advantage of this, it is possible to robustly align or classify

noisy and corrupted images. These approaches have been successful in recognizing faces in

natural images [PGW+12], analyzing 4D computed tomographic images [GCSZ11], and in

the generation of magnetic resonance images [WN12].

4.3 Methodology

This section shows how a low-rank matrix optimization problem can be used to align,

reconstruct and classify 3D structures. The motivation for the proposed approach and its

derivation is best explained starting from the data generation process.

4.3.1 Formation of the Signal

In cryo-ET we would like to reconstruct the density function of a small number of different

structures. Ideally, we would be able to generate a density function sl : R3 → R for each of

the l = 1, . . . , k different structures. Assuming that the density of the macromolecular struc-

tures in the sample are sufficiently separated from each other and that the tilt-series data

has been appropriately segmented, we can consider the process that generates a tilt-series

for an individual macromolecular structure as following. Suppose that a macromolecular

structure i ∈ [1, N ] belongs to a class ci ∈ [1, k] where N is the number of all observations.

Then we can describe each observation’s density in a normalized coordinate system by ap-

plying a suitable rigid transformation to the coordinate system, consisting of a rotation and

a translation. Denoting this transformation by τθi , with rotation and translation parameters

θi, we obtain the density function

fi = sci ◦ τ−1θi

. (4.1)

This assumes that all macromolecular structures of the same type have the exact same

density. In practice this will not hold exactly, but we can view density function s as an

average density function for the class and model deviations and slight perturbations from

this average as noise.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION54

With the given macromolecular structure transformed to its physical orientation and

location we can now generate a tilt series. In practice this is done by collecting projection

data with the sample holder tilted at different angles. Here we equivalently fix the macro-

molecular structure and rotate the projection plane around the y axis instead. For a given

rotation angle α the projection function P : R3 → R2 for density function f is given by

Pα(fi)(u, v) =

� T

−Tfi

u cos(α)− γ sin(α)

v

u sin(α) + γ cos(α)

dγ. (4.2)

There are several simplifying assumptions we make here. First, we have chosen some-

what arbitrary integration limits. Mathematically we can choose (−∞,∞), but in practice

we only integrate from the irradiation source to the detector. Second, we assume that the

contrast-transfer function (CTF) is the identity. When a good model for the CTF is known,

however, this would be easy to integrate in (4.2). The second assumption is that there is no

degradation of the macromolecular structure in subsequent projections; i.e., the function fi

remains constant during the acquisition process. The sampling data d(i,j) for macromolec-

ular structure fi is obtained by evaluating Pα at a tilt angle of αj where j ∈ [1,M ], and

applying a suitable noise process. M is the number of projections.

Given only discrete data, the best we can hope for is to reconstruct discretized volumes

xl of the density functions sl. We can transform xl back to a function using for example (tri-

linear) interpolation: sl := I(xl). The resulting function can then be used for projections

by substituting sl in (4.1) by sl, followed by application of (4.2).

4.3.2 Rank Minimization

By reordering the elements we can express each 3D volume as a column vector xi and form

a matrix

X∗ = [x1, x2, . . . , xN ], (4.3)

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION55

where X∗ ∈ Rn×N , with n the number of voxels in each discretized volume xi. If we knew

the transformation parameters θ = {θ1, . . . , θN}, then we would be able to approximately

estimate the measured data, d(i,j), by evaluating the projections of the interpolated data

I(xi) under transformation τ−1θi

at locations (u, v) ∈ G:

d(i,j) ≈�Pαj (I(xi) ◦ τ−1

θi)(u, v)

�(u,v)∈G

,

where d(i,j) ∈ Rm×1, with m is the number of pixels in each projection image. For fixed θi

and αj , and assuming tri-linear interpolation, we can write the projection of the transformed

xi as the matrix vector product A(i,j)xi where

A(i,j) :=�Pαj (I(·) ◦ τ−1

θi)(u, v)

�(u,v)∈G

,

where A(i,j) ∈ Rm×n. From this, it may seem that we can reconstruct xi from data points

d(i,j) by solving (if feasible):

find x such that �A(i,j)x− d(i,j)�2 ≤ σ, ∀j, (4.4)

where j ∈ [1,M ] indexes the tilt angles, and σ is an estimate of the noise/misfit level. This

formulation requires that projections of the reconstructed volume x, match the observed

projection data up to the given noise level. Even though the transformation parameters

are given, this procedure would not give us the desired result. The reason for this is that

the reconstruction of each macromolecular structure i is treated entirely independently and

consequently suffers from the missing wedge problem. Moreover, even if two instances i1

and i2 are from the same class, i.e., ci1 = ci2 , there is absolutely no guarantee that xi1 = xi2 .

This is where low-rank matrix optimization comes in. If we carefully look at (4.3) we

can see that the rank of X∗ is at most equal to the number of distinct columns. In typical

cryo-ET settings the number of distinct classes k is very small compared to the number of

instances N , and much smaller than the length of each vector xi. Taking this into account

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION56

we can modify (4.4) to the problem

minimizeX

rank(X) s.t. �A(i,j)xi − d(i,j)�2 ≤ σ, ∀i, j, (4.5)

where xi is the i-th column of X. Using the rank takes care of two things. First, it

provides the heuristic on how heterogeneous the structures are in the entire sample set,

which can help us decide on the number of classes (more on this below). Second, the rank

structure links columns together, which means that, in effect, all the data for a given class

are combined, thereby solving the missing-wedge problem (provided of course that the data

covers the entire frequency space).

There are two problems with (4.5). The first problem is that it assumes that the

transformation parameters θi are known. This, of course, is a fundamental problem in

subtomogram averaging, where even a rough approximation may be difficult to obtain. The

second problem is that, even if we did know all parameters, rank minimization would still

be an NP-hard problem [MJCD08, RFP10]. We discuss each of these in more detail below.

4.3.3 Nuclear-Norm Approximation

One of the well-known approaches of avoiding the complexity associated with minimizing

the rank is to replace rank(X) by the nuclear-norm �X�∗, which is defined as the sum of the

singular values of X [RFP10]. This convex approximation can be seen as a generalization of

the highly-successful use of �1 instead of the cardinality, often referred to as �0, in compressed

sensing [CRT06b, CT06, Don06] (applied, in this case, to the vector of singular values).

Applied to (4.5) this gives

minimizeX

�X�∗ s.t. �A(i,j)xi − d(i,j)�2 ≤ σ, ∀i, j. (4.6)

As an example of this, we generated four different two-dimensional classes xi of size 33×33

(see Figure 4.1(a)). From this we generated 300 instances that were rotated uniformly at

random, and translated by up to ±10%. A tilt series, consisting of 11 equally-spaced

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION57

projections at angles ranging from −10◦ to +10◦ and sampled at 50 points, was then

generated for each instance. We then solved (4.6) by writing it as a series of penal-

ized formulations and solving them using, a fast iterative shrinkage-thresholding algorithm

(FISTA) [BT09a, BT09b]. The images corresponding to several of the columns xi of the

solution X∗ are illustrated in Figure 4.1(b). Despite the fact that no class information

was given, nuclear-norm minimization is still capable of combining the information from

the various instances of each class and obtaining accurate reconstructions. In other words,

nuclear-norm minimization is able to find a parsimonious model that explains the data.

Without the low-rank structure each reconstruction would simply have constituted a back-

projection from very incomplete data.

(a) (b)

Figure 4.1: Density of (a) four ground-truth structures, and (b) image representationof several columns of the solution X∗ to the nuclear-norm minimization problem whengiven a limited number of projections for each instance, along with their rotation andtranslation parameters, but without any information about the structure classes. Fourdistinct classes matching the ground-truth structures are clearly recovered.

4.3.4 Finding the Transformation Parameters

It now remains to incorporate the transformation parameters into the problem formulation.

For simplicity of notation and to make the dependency on θi explicit, we denote by Aθi ∈RmM×n the linear operator obtained by stacking projection matrix A(i,j) and likewise denote

by di ∈ RmM×1 the stacking of projection d(i,j). We further write θ = {θ1, . . . , θN}. Addingthe transformation parameters to (4.6) and combining the misfit of all data points for a

given xi, we obtain

minimizeX,θ

�X�∗ s.t. �Aθixi − di�2 ≤ σi ∀i, (4.7)

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION58

with misfit parameters σi modified accordingly. Despite its friendly appearance this formu-

lation is still difficult to solve due to the nonlinearity of Aθi . Inspired by the work of Peng

et al. [PGW+12] we apply iterative linearization of A. The idea is that for sufficiently small

changes Δθi and Δxi we can approximate

A(θi+Δθi)(xi +Δxi) ≈ Aθi(xi +Δxi) + Jθi(xi)Δθi,

where Jθi(xi) ∈ RmM×6 denotes the Jacobian of Aθixi with respect to θi; that is, the changes

in the projection with respect to the transformation parameters of the transformed volume

xi. In our application, the observed tomograms are aligned by finding six transformation

parameters corresponding to the rotation and translation along X, Y, and Z axes. Note

that, due to the use of linear interpolation, Aθixi is not differentiable everywhere, although

the set of parameters θi at which this occurs is measure zero. For our experiments we use

finite-difference approximation for the Jacobian, thereby avoiding this problem. Adding

superscript iteration numbers k to the parameters we summarize the iterative algorithm as

follows. At every step we solve:

Xk+1,Δθk := argminX,Δθ

�X�∗ (4.8)

s.t. �Aθki(xi) + Jθki

(xki )Δθi − di�2 ≤ σi, ∀i

and then update θk+1 = θk + βΔθk, with stepsize 0 < β ≤ 1.

4.3.5 Solving the Subproblem

We solve each subproblem (4.8) by first reformulating it as a conic problem and then solving

it using templates for first-order conic solvers (TFOCS) [BCG11]. For the reformulation we

use the Lorentz cone, which is defined as K := {(x, σ) ∈ Rn × R | �x�2 ≤ σ}. With this it

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION59

is easy to see that we can rewrite (4.8) as

minimizeX,Δθ

�X�∗

subject to

Aθki

Jθki(xki )

0 0

xi

Δθi

+

−di

σi

∈ K, ∀i,

which is the final formulation of the subproblem.

4.3.6 Configuring the Solver

The proposed solver requires estimates for the noise level σi as well as the initial rotation

and translation parameters. The noise level needs to be chosen with care; choosing a σ

value that is too large results in the trivial solution X = 0, whereas choosing σi too small

leads to overly restrictive conditions and no feasible solutions. The noise level can either

be based on the experimental setup, or in a more heuristic way by choosing σi equal to

the norm of the projections and then gradually reducing it. Due to the non-convex nature

of the overall problem, the initial alignment parameters cannot deviate too much from the

desired alignment parameters. The suggested procedure therefore requires estimating initial

alignment parameters.

4.4 Experiments

We now evaluate the alignment and clustering performance of our proposed algorithm, NN-

CET, by comparing it against the performance of ML-TOMO [SMVC09], which is one of

several publicly available and widely used subtomogram algorithms in the community. We

used this particular method since it does joint alignment and clustering in a single framework

like our approach. In particular, we used the ML-TOMO implementation included in the

Xmipp Version 2.4 software package [SMVM+04].

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION60

4.4.1 Experiment Setup

Data generation

For test problems we used two sets of 3D density maps of macromolecules that have been

studied by single particle electron microscopy. The first set contains GroEL (EMD-1042)

and GroEL/ES (EMD-1180) [RFR+01]. The second set contains three different conforma-

tions of Helicase molecules (EMD-1135,1146,1148) [NRRM+06]. The density maps for these

structures were downloaded from the Electron Microscopy Data Bank (EMDB) [emd]. For

the experiments, we rescaled the model structures while retaining the relative scales. The

original GroEL molecule in EMDB has a 3D volume of size 128 × 128 × 128 voxels and a

voxel width of 1.74 A, and GroEL/ES has a size of 192 × 192 × 192 voxels and a width

of 1.4 A. The first two Helicase molecules are of size 80 × 80 × 80 voxels, the third one

is of size 64 × 64 × 64 voxels. All Helicase molecules have a voxel width of 4.1 A. Given

the computational complexity of this method, all models were re-sampled using tri-linear

interpolation to a 32× 32× 32 volume of voxels of size 8.2 A in each dimension preserving

their relative size. Example density maps for GroEL and GroEL/ES, as well as Helicases

are illustrated in the top row of Figure 4.2.

To generate each simulation data, we sampled N = 100 instances with classes inde-

pendently chosen at random with uniform probability of 0.5 for each of the classes GroEL

or GroEL/ES. We then rotated the density function of each instance by (φxi , φyi , φ

zi ) ∈

[−180, 180]× [−180, 180]× [−180, 180] degrees at uniformly random and then translated the

rotated volume by (τxi , τyi , τ

zi ) ∈ [−5, 5]× [−5, 5]× [−5, 5] pixels at uniformly random. After

these rigid-body transformations, we applied the Radon transform to each of the instances

at 25 projection angles α = {−60,−55, . . . , 55, 60} degrees. Two-dimensional projections of

size 46×46 pixels1 were generated using the projxyz function in IMOD[KMM96]. The pixel

size in the projection was chosen to match that of the individual voxels and the grid size was

chosen slightly larger than the data volume dimension to ensure that all relevant projection

1The projection size of 46 pixels corresponds to the maximum projection width of a 32× 32× 32 volumewhen rotated along a single axis.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION61

information was captured. The clean projections were then corrupted using i.i.d. Gaussian

measurement noise to obtain SNR levels of 1, 0.1, 0.03, and 0.01. Our definition of SNR is

var(x)var(n) , where x is the noiseless original projection image which is normalized to have zero

mean and unit variance and n is additive Gaussian noise with zero mean. Examples of the

observed projections of the GroEL and GroEL/ES density maps are shown in Figure 4.2

GroEL GroEL/ES

X Y Z X Y Z

SNR=1 0.1 0.03 0.01

GroEL

GroEL/E

S

Figure 4.2: Top row: Cross sections of the original GroEL and GroEL/ES volumesalong the X, Y , and Z axes. Second and third rows: Radon measurements of GroELand GroEL/ES, respectively, for SNR 1, 0.1, 0.03, and 0.01.

For Helicases we have 3D density maps for each of the three conformation types, cross

sections of which can be seen in Figure 4.3. From these density maps we again generated

a set of N = 100 instances with conformation types chosen uniformly at random. Then

we successively applied the rotation, translation, and projection steps as described for the

GroEL and GroEL/ES dataset, and added measurement noise. Examples of the resulting

observations are shown in Figure 4.4.

Solver Setup and Parameter Initialization

The NN-CET algorithm requires three sets of initial parameters: matrix X, translation

and rotation parameters, and noise level estimates. For the initial X we simply generate

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION62

Hel1

Hel2

Hel3

X Y Z

Figure 4.3: Cross sections of the original Helicases along the X, Y , and Z axes.

SNR=1 0.1 0.03 0.01

Hel1

Hel2

Hel3

Figure 4.4: Noisy Radon measurements of Helicases at SNR = 1, 0.1 0.03, 0.01.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION63

a random matrix with entries sampled i.i.d from the standard normal distribution N (0, 1).

The initial translation parameters are calculated by finding the center of gravity2 of each to-

mographic reconstruction, and the initial rotation parameters are randomly selected within

a range of [−15, 15] degrees from the correct value.

We have also experimented with running the application without the initial centering,

for situations where mass centering does not provide reliable estimates. In this case the

method does find a reasonable estimate of the centers, but the initial translation errors can

increase errors in the angular estimates. For this situation we use the proposed method

in a two-stage approach. First, the algorithm is run to estimate the center of structures

instead of mass centering. In the second stage we reinitialize and rerun the algorithm

using the calculated translation parameters from the first stage. Both approaches, with

mass centering and with two step processing, produce equivalent alignment and clustering

accuracy.

Finally, we need to provide noise level estimates σi. In our experiments, we used the

actual noise levels in each observation. It is important to have a reasonable value for σi to

obtain high alignment accuracy. When σi is too small there may not be a feasible solution

to (4.8). Likewise, we should certainly choose σi to be smaller than �di�2 otherwise the

corresponding solution xi will be zero.

Similarly for testing ML-TOMO, the initial rotation parameters were set within a range

±15 degrees around the ground truth rotation parameters for both the GroEL & GroEL/ES

and Helicases experiments. ML-TOMO was configured to search over a range of ±30 de-

grees around the initial alignment to allow more freedom in the alignment procedure. The

granularity for the rotation angles was set to be five degrees. The number of references is

set to two for GroEL and GroEL/ES, and three for Helicases. ML-TOMO simulations ran

for 20 iterations. The number of iterations for ML-TOMO is shorter because we observed

that the estimated alignment parameters as well as classes did not change much after about

2The center of mass of each tomogram is calculated by weighted sum of voxel coordinates, where theweight for each voxel is defined as the voxel’s gray level value divided by the sum of gray values of all voxels.For mass centering, a user does not need to provide 3D reconstructions but the algorithm will automaticallyreconstruct 3D volumes given tilt series information and calculate mass centers.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION64

15 iterations.

Evaluation Metrics

We use two metrics to evaluate the quality of the resulting alignments: 1. Difference in the

recovered and the ground truth alignment parameters. 2. Difference in the recovered density

functions and the ground truth ones. The first metric quantifies the discrepancy between the

recovered alignment parameters and the true alignment parameters. We measure the error

in rotation parameters by comparing the rotation matrices. Given two rotation matrices S

and R in SO(3), we can define their inner-product as the trace of the product: �R,S� =Tr(RTS). Based on this we can derive two quantities. First, the inner product induces a

squared norm �S�2 = �S, S�, where �·� can be verified to be the Frobenius norm. Second, we

can define angles αr(S,R) (in radians) between two rotation matrices by using the standard

definition

cos(αr(R,S)) :=�R,S�

�R�F �S�F.

The Frobenius norm of a rotation matrix in SO(d) is√d, and we therefore define the angles

in degree between two rotations as

α(R,S) :=180

π· acos

�Tr(RTS)

3

�.

Our proposed method recovers rotation and translation parameters without any fixed

reference, which makes measuring the error to the ground truth challenging. As an example,

we can rotate all density maps (or at least those within the same class) arbitrarily without

affecting the projection data, provided that the associated rotation matrices are modified

accordingly. In other words, all that matters is the relative orientation and position between

the volumes. We therefore report on the variance of the translation and rotation parameters

within each class and remove any bias present. Given a set of translation parameter errors

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION65

{�i}Ni=1 in one class, we compute squared standard deviation:

[STD�]2 =

1

N

n�

i=1

(�i − �)2,

with mean translation error � given by

� := argmin�

1

N

N�

i=1

��− �i�22 =1

N

N�

i=1

�i.

For rotation matrices {Ri}Ni=1 with corresponding ground truth {Si}Ni=1, we can define the

pairwise deviation Qi as the rotation matrix that satisfies QiSi = Ri, giving Qi = RiSTi .

The squared standard deviation of the rotation matrices is then given by

[STDQ]2 =

1

N

N�

i=1

�Q−Qi�2F ,

where the mean rotation matrix Q is defined as follows [Moa02]:

Q := argminQ∈SO(3)

1

N

N�

i=1

�Q−Qi�2F . (4.9)

When relaxed to the orthogonal group O(3) instead of SO(3) the solution of this is given

by Q = UV T , where UΣV T is the singular-value decomposition of 1N

�Ni=1Qi. In all our

experiments Q had a determinant of +1, thus indicating that Q lies in SO(3) and therefore

solves (4.9).

The standard deviation of translations or rotation matrices are sensitive to outliers. We

therefore also look at the pairwise difference, in term of angles or Frobenius-norm distances,

between deviations Qi and Qj , which can be plotted as matrices.

The second evaluation metric looks at the correlation between the recovered density

function and the true density function in terms of Fourier shell correlation (FSC) [HvH86].

FSC can provide information on the correlation between the recovered and the reference

function at each spatial frequency up to the Nyquist rate. Typically, a good reconstruction

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION66

has high FSC in both the high and low spatial frequency ranges.

4.4.2 Alignment Accuracy

The proposed method returns two types of refined reconstructions. The first uses the

estimated alignment parameters to align individual subtomograms from which averaged

structures of macromolecular complexes are formed. The second consists of the columns in

the recovered matrix X∗, each of which represents a volume in vectorized form. The first

type of reconstruction is used to evaluate alignment accuracy, and the second type is used

to cluster individual observations.

GroEL and GroEL/ES

Figures 4.5 and 4.6 show the recovery of GroEL and GroEL/ES structures using ML-

TOMO and our proposed method. It can be seen that both methods recovered the six-

fold symmetric structures when SNR = 1 and 0.1. As the SNR decreases, however, both

algorithms have difficulty recovering the underlying structure and the alignment parameters.

The standard deviation of the recovered rotation and translations parameters are plotted in

Figure 4.7 as a function of NN-CET iterations. It can be seen that the standard deviation of

the rotation parameter errors decreases steadily when the SNR is relatively high. For lower

SNR values of 0.03 and 0.01 the algorithm is unable to find the good rotational alignment

for either of the two classes. The variance in the translation errors decreases steadily for

all given noise levels, although for SNR of 0.01, the variance decreases more slowly. Note

in particular that, although the mass centering was not able to center structures very well

when SNR = 0.03, the proposed algorithm can still find the correct translation parameters

gradually as iterations progress. As mentioned in the prior section, this allows the proposed

method to center the observations without initial mass-based centering. From the FSC

plotted in Figure 4.8 for SNR = 1 and 0.1, we see that both ML-TOMO and NN-CET

recover the structures at a similar level of resolution for GroEL. However, as a result of

misclassification, ML-TOMO fails to recover GroEL/ES. We will discuss this in more detail

in Section 4.4.3.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION67

X Y Z

Original

SNR

=1

NN-C

ET

ML

SNR

=0.1

NN-C

ET

ML

SNR

=0.03

NN-C

ET

ML

SNR

=0.01

NN-C

ET

ML

Figure 4.5: XYZ cross sections of the averaged GroEL structures for different SNRlevels. Each column contains X, Y, Z cross sections at the center of each side of thevolume. The rows, from top to bottom, correspond to: 1. Ground truth GroEL. Therest of even-numbered rows (2, 4, 6, 8) show the GroEL reconstructions of NN-CET atSNR = 1, 0.1, 0.03, 0.01 respectively, and the rest of odd-numbered rows (3, 5, 7, 9)show the GroEL reconstructions of ML-TOMO at SNR = 1, 0.1, 0.03, 0.01 respectively.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION68

X Y Z

Original

SNR

=1

NN-C

ET

ML

SNR

=0.1

NN-C

ET

ML

SNR

=0.03

NN-C

ET

ML

SNR

=0.01

NN-C

ET

ML

Figure 4.6: XYZ cross sections of the averaged GroEL/ES structures for differentSNR levels. Each column contains X, Y, Z cross sections at the center of each side of thevolume. The rows, from top to bottom, correspond to: 1. Ground truth GroEL. The restof even-numbered rows (2, 4, 6, 8) show the GroEL/ES reconstructions of NN-CET atSNR = 1, 0.1, 0.03, 0.01 respectively, and the rest of odd-numbered rows (3, 5, 7, 9) showthe GroEL/ES reconstructions of ML-TOMO at SNR = 1, 0.1, 0.03, 0.01 respectively.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION69

Rotation Translation

0 20 40 60 80 1000

0.5

1

1.5

2

2.5

3

3.5

4

4.5

Rot

atio

n ST

D

Class 1Class 2

0 20 40 60 80 1000

1

2

3

4

5

6

Tran

slat

ion

STD

Class 1Class 2

SNR = 1

0 20 40 60 80 1002

2.5

3

3.5

4

4.5

5

5.5

6

Rot

atio

n ST

D

Class 1Class 2

0 20 40 60 80 1000

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

Tran

slat

ion

STD

Class 1Class 2

SNR = 0.1

0 20 40 60 80 1002.5

3

3.5

4

4.5

5

5.5

6

6.5

7

Rot

atio

n ST

D

Class 1Class 2

0 20 40 60 80 1001

1.5

2

2.5

3

3.5

4

4.5

5

Tran

slat

ion

STD

Class 1Class 2

SNR = 0.03

0 20 40 60 80 1003.5

4

4.5

5

5.5

6

6.5

7

7.5

Rot

atio

n ST

D

Iteration

Class 1Class 2

0 20 40 60 80 1002.5

3

3.5

4

4.5

5

Tran

slat

ion

STD

Class 1Class 2

SNR = 0.01

Figure 4.7: Alignment accuracy for GroEL and GroEL/ES data set quantified by therotation angle error STD (left) in degrees and the translation error STD (right) in pixelsas a function of NN-CET iteration, as defined in Section 4.4.1. Each row corresponds toa given SNR level.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION70

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

SNR = 1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1NN−CETML

SNR = 0.1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

freq

SNR = 0.03

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

freq

SNR = 0.01

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

SNR = 1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1NN−CETML

SNR = 0.1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

freq

SNR = 0.03

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

freq

SNR = 0.01

Figure 4.8: Fourier shell correlation curves of GroEL (rows 1,2) and GroEL/ES (row3,4) at SNR = 1, 0.1 0.03, and 0.01. The curves are calculated between the ground truthstructure and the 1. NN-CET aligned reconstruction (our method, cyan); 2. ML-TOMOreconstruction (green).

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION71

Helicases

Similar to the GroEL and GroEL/ES case, the reconstruction accuracy for Helicases suffers

from low SNR, as can be seen from Figures 4.9 to 4.11.

The standard deviation of the translation and rotation estimation errors for Helicases

are plotted in Figure 4.12. At SNR = 1, the in-class variation in rotation errors decreases

strongly until around 50 iterations, after which it increases again. When looking at the

pairwise difference between rotation matrix estimation errors, shown in Figure 4.13, we

see that the initial rotation alignment is poor. At iteration 50, the angles between all

pairs of rotation parameter misfits are very small, indicating good alignment. The pairwise

distances do increase again, but overall the level at iteration 100 is still smaller than at

the beginning. This means that although we do not have the most optimal alignment at

iteration = 100, the observations are still relatively better aligned than at the beginning of

the simulations (at iteration = 0) still maintaining the classes all reasonably aligned within

themselves. This particular observation indicates a few issues: 1. the algorithm can find the

well estimated aligned volumes as well as the alignment parameters, however, the current

formulation is not strongly constrained enough to stop at the optimal solution but keep

exploring different solutions when the optimal solution does not lie in a well constrained

region. 2. the current stopping criterion is not able to detect the best aligned moment for

each class. To cope with these issues, one can record the entire solution path and choose the

best solution for an external metric such as FSC. In this section, we provide the simulation

results obtained at iteration 50 for SNR = 1 to compare the best result, and for SNR = 0.1,

0.3 and 0.01, the simulation results are obtained at iteration 100. At SNR = 0.1 and SNR

= 0.03, the STD of rotation parameter errors fluctuate but overall do seem to decrease. At

SNR = 0.01, the proposed method was not able to find the good angular alignment for any

class.

The FSC plots in Figure 4.14 show that NN-CET and ML-TOMO return reconstructions

that have a comparable spatial resolution to each other for SNR as low as 0.03. The

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION72

resolution of the averaged Hel2 of ML-TOMO, however, is much lower than that of NN-

CET due to misclassification as shown in the confusion matrix in Figure 4.15.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION73

X Y Z

Original

SNR

=1

NN-C

ET

ML

SNR

=0.1

NN-C

ET

ML

SNR

=0.03

NN-C

ET

ML

SNR

=0.01

NN-C

ET

ML

Figure 4.9: XYZ cross sections of the averaged Hel1 structures for different SNR levels.Each column contains X, Y, Z cross sections at the center of each side of the volume.The first row contains the ground truth. The rest of even-numbered rows (2, 4, 6, 8)show the Hel1 reconstructions of NN-CET at SNR = 1, 0.1, 0.03, 0.01 respectively, andthe rest of odd-numbered rows (3, 5, 7, 9) show the Hel1 reconstructions of ML-TOMOat SNR = 1, 0.1, 0.03, 0.01 respectively.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION74

X Y Z

Original

SNR

=1

NN-C

ET

ML

SNR

=0.1

NN-C

ET

ML

SNR

=0.03

NN-C

ET

ML

SNR

=0.01

NN-C

ET

ML

Figure 4.10: XYZ cross sections of the averaged Hel2 structures for different SNRlevels. Each column contains X, Y, Z cross sections at the center of each side of thevolume. The first row contains the ground truth. The rest of even-numbered rows (2, 4,6, 8) show the Hel2 reconstructions of NN-CET at SNR = 1, 0.1, 0.03, 0.01 respectively,and the rest of odd-numbered rows (3, 5, 7, 9) show the Hel2 reconstructions of ML-TOMO at SNR = 1, 0.1, 0.03, 0.01 respectively.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION75

X Y Z

Original

NN-C

ET

ML

NN-C

ET

ML

NN-C

ET

ML

NN-C

ET

ML

Figure 4.11: XYZ cross sections of the averaged Hel3 structures for different SNRlevels. Each column contains X, Y, Z cross sections at the center of each side of thevolume. The first row contains the ground truth. The rest of even-numbered rows (2, 4,6, 8) show the Hel3 reconstructions of NN-CET at SNR = 1, 0.1, 0.03, 0.01 respectively,and the rest of odd-numbered rows (3, 5, 7, 9) show the Hel3 reconstructions of ML-TOMO at SNR = 1, 0.1, 0.03, 0.01 respectively.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION76

Rotation Translation

0 20 40 60 80 1000

0.5

1

1.5

2

2.5

3

3.5

4

Rot

atio

n ST

D

Class 1Class 2Class 3

0 20 40 60 80 1000

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

Tran

slat

ion

STD

Class 1Class 2Class 3

SNR = 1

0 20 40 60 80 1000

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

Rot

atio

n ST

D

Class 1Class 2Class 3

0 20 40 60 80 1000

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

Tran

slat

ion

STD

Class 1Class 2Class 3

SNR = 0.1

0 20 40 60 80 1002

3

4

5

6

7

8

9

Rot

atio

n ST

D

Class 1Class 2Class 3

0 20 40 60 80 1000

1

2

3

4

5

6

Tran

slat

ion

STD

Class 1Class 2Class 3

SNR = 0.03

0 20 40 60 80 1003

4

5

6

7

8

9

10

11

12

13

Rota

tion

STD

Iteration

Class 1Class 2Class 3

0 20 40 60 80 1002

2.5

3

3.5

4

4.5

5

5.5

Tran

slat

ion

STD

Class 1Class 2Class 3

SNR = 0.01

Figure 4.12: Alignment accuracy for Helicases data set quantified by the rotation angleerror STD (left) in degrees and the translation error STD (right) in pixels as a functionof NN-CET iteration, as defined in Section 4.4.1. Each row corresponds to a given SNRlevel.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION77

Iteration: 1 50 100

Class

1Class

2Class

3

0

5

10

15

20

25

30

35

Figure 4.13: Pairwise angles between the error rotation matrices between observationswithin classes in degrees for Helicases data set at SNR = 1.

4.4.3 Clustering Accuracy

In the previous section, we have seen that the alignment accuracy of the averaged struc-

tures of NN-CET and ML-TOMO is equivalent when ML-TOMO can accurately cluster

subtomograms. However, when ML-TOMO does not cluster instances accurately, those

clusters that do not contain many examples suffer significantly in terms of resolution. Our

algorithm does not explicitly cluster the samples, but provides a row-rank approximation

of the aligned reconstructions. By clustering the columns of the solution matrix X∗ us-

ing k-means, we can assign classes to each sample. By contrast, ML-TOMO explicitly

assigns each sample to a class that maximizes the overall likelihood of observing the data

set. We evaluate the clustering accuracy, which we define to be the percentage of correctly

labeled samples. As shown in Table 4.1, both NN-CET and ML-TOMO cluster GroEL and

GroEL/ES samples with a very high accuracy at SNR = 1 and 0.1, which gradually reduces

as the SNR decreases. For Helicases we see that ML-TOMO is unable to obtain accurate

clustering for any of the SNR levels, whereas our method clusters Helicases without any

error for SNR levels as low as 0.03. This can be attributed to the dimension reduction that

follows from the nuclear-norm minimization. The proposed approach aligns observations

considering the fact that the heterogeneity in the samples has to be much smaller than the

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION78

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

SNR = 1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1NN−CETML

SNR = 0.1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1FS

C

freq

SNR = 0.03

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

freq

SNR = 0.01

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

SNR = 1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1NN−CETML

SNR = 0.1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

freq

SNR = 0.03

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

freq

SNR = 0.01

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

SNR = 1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1NN−CETML

SNR = 0.1

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

FSC

freq

SNR = 0.03

0 0.1 0.2 0.3 0.4 0.50

0.2

0.4

0.6

0.8

1

freq

SNR = 0.01

Figure 4.14: Fourier shell correlation curves of Hel1 (rows 1,2), Hel2 (rows 3,4), Hel3(rows 5,6) at SNR = 1, 0.1 0.03, and 0.01. The curves are calculated between theground truth structure and the 1. NN-CET aligned reconstruction (our method, cyan);2. ML-TOMO reconstruction (green).

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION79

dimension of the space containing the observations. The efficacy of dimension reduction in

clustering tomograms has also been demonstrated in other works, where all subtomograms

are assumed to be aligned [HHM11, YSRR10].

The confusion matrices in Figure 4.15 provide a more detailed analysis on the perfor-

mance of the two different methods. As shown in the first two rows, the confusion matrices

of the instances clustered using NN-CET show a strong diagonal pattern (or permutation

thereof) until the noise level becomes too high (0.01 for Helicases and 0.03 for GroEL

and GroEL/ES). By contrast, the confusion matrices for the instances clustered using ML-

TOMO only show strong diagonal patterns for GroEL and GroEL/ES cases when SNR =

1 and 0.1. For Helicases, ML-TOMO failed to cluster accurately at any of the given SNR

levels even though the correct number of references is provided. Instead, it tends to create

one or two large clusters not three as provided.

By examining the distribution of the singular values in the nuclear-norm minimized

reconstruction X∗, we can also get an insight on the degree of heterogeneity of the sample

set. Singular values can be interpreted as a scaling factor of each orthogonal component

contained in the image space spanned by the reconstructions; the larger the singular value,

the stronger the component. In this light, if we see a set of singular values that are uniform,

this means that the image space contains a variety of structural components, therefore is

more heterogeneous. On the other hand, if we see a set of singular values with only a few

large values and all others small, we conclude that the set will have few distinct structural

components. There are three factors that affect the singular values of a subtomogram

set: 1. the number of different conformations in the set; 2. how well the reconstructions

are aligned; and 3. the amount of noise in the reconstruction. When the tomograms are

not well aligned or contain high level of noise, the singular values will have a rather flat

distribution. On the other hand, if the subtomogram averaging is successful in aligning

the structures and removing the noise, the only factors affecting the distribution of singular

values should be the number of conformation classes in the set and their similarity. As such,

the distribution of singular values provides a good heuristic to determine how well aligned

the samples are and how many distinct classes are present in the dataset. Figure 4.16 shows

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION80

GroEL/ES HelicasesSNR Unaligned X ML Unaligned X ML

1 59% 100% 100% 43% 100% 64%0.1 60% 100% 91% 43% 100% 69%0.03 54% 72% 70% 46% 100% 58%0.01 55% 57% 53% 50% 43% 38%

Table 4.1: Clustering accuracy for GroEL & GroEL/ES and Helicases data sets: Thehighest accuracy for each experiment is highlighted in bold. 1. The unaligned columncontains the accuracy of clustering subtomograms before aligning them. 2. The X columncontains the accuracy of clustering the solved X matrix using the proposed method. 3.The ML column contains the accuracy of labels returned by the Maximum Likelihoodmethod.

the sorted singular values of the nuclear-norm minimized reconstruction X∗ normalized

by the largest singular value on logarithmic scale, and likewise for matrices formed by

concatenating the raw subtomograms before subtomogram averaging (unaligned). For NN-

CET we see that the resulting distribution is extremely sparse with a sharp decrease after

1–3 significant singular values. The number of dominant singular values of GroEL and

GroEL/ES reconstructions at SNR = 1 and 0.1, and for Helicases at SNR = 1, 0.1, and

0.03 matches exactly the true number of conformational classes contained in the data set.

For the unaligned subtomograms prior to any processing, we see that the ten largest

singular values are at most 1–5 times smaller than the largest one. This indicates that the

set is not well aligned and/or severely affected by noise, and it is clearly not possible to

estimate the number of classes from this set of singular values.

In summary, we conclude that the dimension reduction carried out by the nuclear-

norm minimization in NN-CET enables more accurate clustering, and also provides a good

heuristic for determining the number of conformation classes contained in the data. The

procedures used in ML-TOMO do not provide any information equivalent to this.

4.4.4 Computational Complexity and Runtime

The proposed method was implemented in Matlab with a dependency on TFOCS [BCG11]

to solve the sub-problems. The computation was parallelized among 12 cores on a single

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION81

SNR=1 0.1 0.03 0.01

NN-C

ET

GroEL/ES

Helicases

ML-T

OMO

GroEL/ES

Helicases

Figure 4.15: Confusion matrices. Each row contains the breakdown ratio of instancesof each true class into its estimated classes. ith row corresponds to the true class i andjth column corresponds to the estimated class j. Perfect classification is indicated by amatrix with “1.00” on the diagonal and “0.00” everywhere else.

machine, resulting in a per-iteration time of about 275 seconds and an overall runtime of

7.5 hours for 100 iterations. Around 25% of the compute time is spent on evaluating the

nuclear norm and applying singular-value soft thresholding. The remaining 75% is spent

on evaluating the measurement constraint, three-quarters of which was used for evaluating

projections and back-projections using the 3D Radon operator and its adjoint; amounting

to roughly half of the total computation. We ran Xmipp ML-TOMO (version 2.4) on a

single core taking about 36 hours for 20 iterations. The proposed method took around 53

hours for 100 iterations when ran on a single core.

As reflected in the lengthy run time, this initial implementation does not scale up well

to handle high resolution volumes. In our test, the convergence of parameter estimates

for high resolution data sets were also slower especially estimating accurate translation

parameters. This slow convergence and lengthy run time can be overcome by speeding

up the computation intensive 3D Radon operators as well as more efficient nuclear norm

calculation.

The 3D Radon implementation used in the NN-CET solver has a complexity of O(N ·

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION82

GroEL/ES Helicases

2 4 6 8 10−1.5

−1

−0.5

0UnalignedX

SNR = 12 4 6 8 10

−1.5

−1

−0.5

0UnalignedX

SNR = 1

2 4 6 8 10−1.5

−1

−0.5

0

SNR = 0.12 4 6 8 10

−1.5

−1

−0.5

0

SNR = 0.1

2 4 6 8 10−1.5

−1

−0.5

0

SNR = 0.032 4 6 8 10

−1.5

−1

−0.5

0

SNR = 0.03

2 4 6 8 10−1.5

−1

−0.5

0

SNR = 0.012 4 6 8 10

−1.5

−1

−0.5

0

SNR = 0.01

Figure 4.16: Distributions of the top ten normalized singular values on base-ten loga-rithmic scale of: (Unaligned) the matrix containing subtomograms before aligning them;and (X) the solution matrix X∗ returned by NN-CET. Results for GroEL and GroEL/ESare on the left, those for Helicases on the right. From top to bottom: SNR= 1, 0.1, 0.03,and 0.01.

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION83

M ·m · s), where, N is the number of volumes, M is the number of projections per volume,

m is the number of samples/observations per projection (number of pixels per projection),

and s is the number of voxels along one side of a 3D cube, thus giving a total of n = s3

voxels. For the Radon operator and its adjoint, processing a single projection line takes O(s)

and can be parallelized over individual volumes as well as projection angles in the forward

case. We therefore expect that the computational cost could be significantly reduced by a

multi-threaded implementation of the Radon operator.

Further reductions in the compute time could be achieved by speeding up the nuclear

norm evaluation, which takes up around 25% of the total compute time. The complexity of

SVD, which is the core operation to evaluate nuclear norms is O(max(a2b, a2b)) for matrices

of size a× b. In our experiments we have a = n and b = N with n > N , giving a complexity

of O(n2 ·N) for evaluating the SVD. Techniques to reduce the computational complexity of

the SVD or methods that avoid it altogether (see for example [WYZ12]) could help further

reduce the compute time of NN-CET.

4.5 Conclusions

In this chapter, we have proposed NN-CET, an approach to the subtomogram averaging

problem based on matrix norm minimization. The proposed approach does not require any

prior information such as the number of classes or references of the target structures. In-

stead, it formulates the alignment and classification problem as a low-rank matrix recovery

problem that explicitly recovers the alignment parameters by calculating the linearized effect

of translation and rotation of the underlying structure on the tomographic measurements.

This approach implicitly addresses the missing wedge issue by formulating tomographic

sensing operators at each projection angle, which is known a priori. The nuclear-norm min-

imized solution matrix spans the principal components of the heterogeneous conformations

embedded in the observed samples, and we can effectively cluster reconstructions in this

low dimensional space.

Our experiments indicate that NN-CET successfully clusters structures during alignment

CHAPTER 4. SUBTOMOGRAMAVERAGING VIA NUCLEAR NORMMINIMIZATION84

and manages to do so more accurately than ML-TOMO. The alignment accuracy of the

averaged structures of NN-CET and ML-TOMO are equivalent for those cases where ML-

TOMO can accurately cluster subtomograms. While our results are promising, there are

a number of issues that need to be addressed before it can be used as a reference free

subtomogram averaging algorithm with high clustering accuracy. The first is computational

cost. While we think our algorithm could easily leverage the parallel computation of a

cluster, we have not yet created a cluster version. The more significant issue is the one of

convergence. Further work is needed to improve the convergence of this algorithm, so larger

data sets can be run, as well as create a better stopping condition, to avoid the issues we

saw when dealing with Helicases.

Chapter 5

Conclusions

CET is an indispensable tool to discover the high resolution structure of the building

blocks of cells and viruses thanks to its capability of imaging in high resolution in 3D. This

success has not been possible without a significant research effort to cope with challenging

image characteristics of CET projections and reconstructions, which are very low SNR and

missing frequency components. The low SNR and contrast of CET images are here to

stay since it is limited by electron dose and voltage that we can apply to organic materials

without damaging it. The missing wedge problem is difficult to overcome as well since the

quality of CET projections are limited by the increase in the thickness of the sample when

it is rotated up to high tilt angles. The longer the path that electrons have to path, the

weaker the phase-contrast signal from the electrons becomes due to the increase in inelastic

scattering. Fortunately, these challenges have been managed by automated acquisition

techniques and image processing pipeline, that enables researchers take large amount of

samples and assemble them into high resolution models of macromolecules at higher SNR.

In this thesis, we have proposed robust image processing enhancements to the CET

image processing pipeline using sparse priors. The idea behind robust image processing

techniques proposed here is that often a single image or a set of images exhibit sparsity pat-

tern in an alternate signal domain. The first example is the digital inpainting algorithm for

CET . Here, we have empirically demonstrated that CET projections are sparse in theDCT

85

CHAPTER 5. CONCLUSIONS 86

domain, and utilized this information to remove high contrast artifacts in the projections

images as well as 3D reconstructions in the CS framework. The algorithm can seamlessly

replace pixels occupied by high contrast objects such as gold markers in projection images.

Then, by reconstructing 3D volume using inpainted projections, we can produce artifact-free

tomograms. By removing high contrast artifacts casting shadow on specimens of interest,

we can study them with greater clarity as well as analyze them analytically with more

confidence. Artificial marker test and subtomogram averaging test have confirmed that the

proposed method preserves the structural integrity of the unpainted regions and their neigh-

boring regions free from artifacts better than the conventional inpainting methods used in

the CET community.

The presented method inpaints each individual projection separately, because this al-

lowed us to create a number of large, but tractable optimization problems. Unfortunately

optimizing each projection separately does not exactly solve the desired problem because

each optimization can result in an inpainted projection that is not necessarily consistent

with the 3D structure of a specimen, creating secondary inpainting artifacts. One straight-

forward application of CS techniques to provide a tomogram that is free from artifacts and

consistent with 3D structural information of a specimen can be jointly minimizing the 2D

sparsity of all projections in the DCT domain as well as the 3D sparsity in the Fourier

domain if applicable. However, this approach is not yet computationally tractable.

Another example of sparsity utilizing robust image processing technique demonstrated

in this thesis is a batch alignment and clustering algorithm for subtomograms. In this

case, we utilized the fact that a set of well aligned images span a low dimensional space.

Even a set of images containing multiple classes, if they are all well aligned within each

class, the whole image space spanned by these images should be as low dimension as the

number of classes. This means that an image matrix whose columns are vectorized version

of a subtomogram should have a low rank if all subtomograms are aligned within each

class. Utilizing this information, we framed the subtomogram averaging problem, basically

a simultaneous batch alignment and clustering problem, into a matrix norm minimization

problem, that can align and cluster subtomograms without prior references nor the number

CHAPTER 5. CONCLUSIONS 87

of class information. This feature, that we do not have to impose a prior class information,

is very unique, since all subtomogram algorithms currently being used in the CET commu-

nity require one or both of this information. However, using prior references can bias the

resulting averaged structure. In our limited scale simulations, we have demonstrated that

this algorithm can efficiently produce clustered and aligned averaged models from noisy

and missing-wedge corrupted tomograms even using only a small number of observations.

The resulting averaged structure produced by the proposed method has higher FSC than

ones produced by the benchmark method. However, this algorithm has been only evaluated

using synthetic data sets, and it needs to be fully evaluated on large scale high resolution

real data set to prove its efficacy in the field.

Although, both examples in this thesis demonstrated positive opportunities for sparsity

inspired image processing algorithms for robust CET image analysis, there are a few im-

provements that can enhance these algorithms and be widely used on a large set of CET

images. One of the most significant improvement can come from expediting convex pro-

gramming used in these algorithms. Currently, l1 norm minimization used in the digital

inpainting algorithm can be solved quite quickly, therefore, the computational complexity is

less of a concern here. However, subtomogram averaging problem can be quite large. For ex-

ample, these days it has become quite common to align a set of 1000 or more high resolution

subtomograms of size 128× 128× 128 and to solve this scale of matrix norm minimization

problem, the current solver requires large amount of memory in addition to many compute

cores to solve large scale SVD and to evaluate each subtomogram’s measurement require-

ments. Fortunately, there are interesting developments in distributed memory computing

for large scale dense linear algebra [PMvdG+13, BCC+97] as well as GPU processing[LN09],

so this problem may become feasible in the near future.

Bibliography

[ACN+10] Fernando Amat, Luis R. Comolli, John F. Nomellini, Farshid Moussavi, Ken-

neth H. Downing, John Smit, and Mark Horowitz. Analysis of the intact

surface layer of Caulobacter crescentus by cryo-electron tomography. Journal

of Bacteriology, 192(22):5855–5865, 2010.

[AK84] A.H. Andersen and A.C. Kak. Simultaneous algebraic reconstruction tech-

nique (sart): A superior implementation of the {ART} algorithm. Ultrasonic

Imaging, 6(1):81 – 94, 1984.

[AMC+08] Fernando Amat, Farshid Moussavi, Luis R. Comolli, Gal Elidan, Kenneth H.

Downing, and Mark Horowitz. Markov random field based automatic image

alignment for electron tomography. Journal of Structural Biology, 161(3):260

– 275, 2008. The 4th International Conference on Electron TomographyThe

4th International Conference on Electron Tomography.

[BBC+01] C. Ballester, M. Bertalmio, V. Caselles, G. Sapiro, and J. Verdera. Filling-in

by joint interpolation of vector fields and gray levels. Image Processing, IEEE

Transactions on, 10(8):1200–1211, August 2001.

[BBC11] Stephen Becker, Jerome Bobin, and Emmanuel J. Candes. Nesta: A fast and

accurate first-order method for sparse recovery. SIAM Journal on Imaging

Sciences, 4(1):1–39, 2011.

88

BIBLIOGRAPHY 89

[BCC+97] L Susan Blackford, Jaeyoung Choi, Andy Cleary, Eduardo D’Azevedo, James

Demmel, Inderjit Dhillon, Jack Dongarra, Sven Hammarling, Greg Henry,

Antoine Petitet, et al. ScaLAPACK users’ guide, volume 4. siam, 1997.

[BCG+10] Grant R. Bowman, Luis R. Comolli, Guido M. Gaietta, Michael Fero, Sun-

Hae Hong, and et al. Caulobacter popz forms a polar subdomain dictating

sequential changes in pole composition and function. Molecular Microbiology,

76(1):173–189, 2010.

[BCG11] Stephen Becker, Emmanuel J. Candes, and Michael Grant. Templates for con-

vex cone problems with applications to sparse signal recovery. Mathematical

Programming Computation, 3(3), 2011.

[BHE01] Sami Brandt, Jukka Heikkonen, and Peter Engelhardt. Multiphase method

for automatic alignment of transmission electron microscope images using

markers. Journal of Structural Biology, 133(1):10 – 22, 2001.

[BJ03] R. Basri and D.W. Jacobs. Lambertian reflectance and linear subspaces. Pat-

tern Analysis and Machine Intelligence, IEEE Transactions on, 25(2):218–

233, Feb 2003.

[BNR+12] Tanmay A. M. Bharat, Takeshi Noda, James D. Riches, Verena Kraehling,

Larissa Kolesnikova, Stephan Becker, Yoshihiro Kawaoka, and John A. G.

Briggs. Structural dissection of Ebola virus and its assembly determinants

using cryo-electron tomography. Proceedings of the National Academy of Sci-

ences of the United States of America, 109(11):4275–4280, 2012.

[Bra56] R. N. Bracewell. Strip integration in radio astronomy. Australian Journal of

Physics, 9:198–217, 1956.

[Bro92] Lisa Gottesfeld Brown. A survey of image registration techniques. ACM

Computing Surveys, 24(4):325–376, 1992.

BIBLIOGRAPHY 90

[BS09] Alberto Bartesaghi and Sriram Subramaniam. Membrane protein structure

determination using cryo-electron tomography and 3D image averaging. Cur-

rent Opinion in Structural Biology, 19(4):402–407, 2009.

[BSL+08] A. Bartesaghi, P. Sprechmann, J. Liu, G. Randall, G. Sapiro, and S. Subra-

maniam. Classification and 3D averaging with missing wedge correction in

biological electron tomography. Journal of Structural Biology, 162(3):436–

450, 2008.

[BT09a] Amir Beck and Marc Teboulle. A fast iterative shrinkage-thresholding al-

gorithm for linear inverse problems. SIAM Journal on Imaging Sciences,

2(1):183–202, 2009.

[BT09b] Amir Beck and Marc Teboulle. A fast iterative shrinkage-thresholding algo-

rithm with application to wavelet-based image deblurring. In International

Conference on Acoustics, Speech and Signal Processing, pages 693–696, 2009.

[BVSO03] M. Bertalmio, L. Vese, G. Sapiro, and S. Osher. Simultaneous structure and

texture image inpainting. Image Processing, IEEE Transactions on, 12(8):882

– 889, 2003.

[CDS98] Scott Shaobing Chen, David L. Donoho, and Michael A. Saunders. Atomic

decomposition by basis pursuit. SIAM J. Sci. Comput., 20:33–61, December

1998.

[CDSAAF10] Daniel Castao-Dez, Margot Scheffer, Ashraf Al-Amoudi, and Achilleas S.

Frangakis. Alignator: A {GPU} powered software package for robust fiducial-

less alignment of cryo tilt-series. Journal of Structural Biology, 170(1):117 –

126, 2010.

[CKS02] Tony F. Chan, Sung Ha Kang, and Jianhong Shen. Euler’s elastica

and curvature-based inpainting. SIAM Journal on Applied Mathematics,

63(2):564–592, 2002.

BIBLIOGRAPHY 91

[CLMW11] Emmanuel J Candes, Xiaodong Li, Yi Ma, and John Wright. Robust principal

component analysis? Journal of the ACM (JACM), 58(3):11, 2011.

[ClTCB11] Ricardo S. Cabral, Fernando D. De la Torre, Joao P. Costeira, and Alexan-

dre Bernardino. Matrix completion for multi-label image classification. In

J. Shawe-Taylor, R.S. Zemel, P. Bartlett, F.C.N. Pereira, and K.Q. Wein-

berger, editors, Advances in Neural Information Processing Systems 24, pages

190–198. 2011.

[CPH+13] Yuxiang Chen, Stefan Pfeffer, Thomas Hrabe, Jan Michael Schuller, and

Friedrich Forster. Fast and accurate reference-free alignment of subtomo-

grams. Journal of Structural Biology, 182(3):235–245, 2013.

[CRT06a] E.J. Candes, J. Romberg, and T. Tao. Robust uncertainty principles: exact

signal reconstruction from highly incomplete frequency information. Infor-

mation Theory, IEEE Transactions on, 52(2):489 – 509, 2006.

[CRT06b] Emmanuel J. Candes, Justin Romberg, and Terence Tao. Robust uncertainty

principles: Exact signal reconstruction from highly incomplete frequency in-

formation. IEEE Transactions on Information Theory, 52(2):489–509, 2006.

[CRT06c] Emmanuel J. Candes, Justin K. Romberg, and Terence Tao. Stable signal

recovery from incomplete and inaccurate measurements. Communications on

Pure and Applied Mathematics, 59(8):1207–1223, August 2006.

[CS01] Tony F. Chan and Jianhong Shen. Mathematical models for local nontexture

inpaintings. SIAM Journal on Applied Mathematics, 62(3):1019–1043, 2001.

[CT06] Emmanuel J. Candes and Terence Tao. Near-optimal signal recovery from

random projections – universal encoding strategies. IEEE Transactions on

Information Theory, 52(2), 2006.

BIBLIOGRAPHY 92

[DFG+09] Elia Diestra, Juan Fontana, Paul Guichard, Sergio Marco, and Cristina Risco.

Visualization of proteins in intact cells with a clonable tag for electron mi-

croscopy. Journal of Structural Biology, 165(3):157 – 168, 2009.

[DK68] DJ DeRosier and A Klug. Reconstruction of three dimensional structures

from electron micrographs. Nature, 217(5124):130–134, 1968.

[DKMN10] Radostin Danev, Shuji Kanamaru, Michael Marko, and Kuniaki Nagayama.

Zernike phase contrast cryo-electron tomography. Journal of Structural Biol-

ogy, 171(2):174 – 181, 2010.

[DND+00] B. De Man, J. Nuyts, P. Dupont, G. Marchal, and P. Suetens. Reduction

of metal streak artifacts in x-ray computed tomography using a transmission

maximum a posteriori algorithm. Nuclear Science, IEEE Transactions on,

47(3):977–981, June 2000.

[Don06] David L. Donoho. Compressed sensing. IEEE Transactions on Information

Theory, 52(4):1289–1306, 2006.

[emd] Emdatabank. http://emdatabank.org/.

[ESQD05] M. Elad, J.-L. Starck, P. Querre, and D.L. Donoho. Simultaneous cartoon

and texture image inpainting using morphological component analysis (mca).

Applied and Computational Harmonic Analysis, 19(3):340 – 358, 2005.

[Fer12] Jose-Jesus Fernandez. Computational methods for electron tomography. Mi-

cron, 43(10):1010–1030, 2012.

[FH01] Achilleas S. Frangakis and Reiner Hegerl. Noise reduction in electron to-

mographic reconstructions using nonlinear anisotropic diffusion. Journal of

Structural Biology, 135(3):239 – 250, 2001.

[FHB01] Maryam Fazel, Haitham Hindi, and Stephen P Boyd. A rank minimiza-

tion heuristic with application to minimum order system approximation. In

BIBLIOGRAPHY 93

American Control Conference, 2001. Proceedings of the 2001, volume 6, pages

4734–4739. IEEE, 2001.

[FL05] J.-J. Fernndez and S. Li. Anisotropic nonlinear filtering of cellular structures

in cryoelectron tomography. Computing in Science and Engineering, 7(5):54–

61, 2005. cited By 24.

[FLC06] J.J. Fernndez, S. Li, and R.A. Crowther. {CTF} determination and correction

in electron cryotomography. Ultramicroscopy, 106(7):587 – 596, 2006.

[FMZ+05] Friedrich Forster, Ohad Medalia, Nathan Zauberman, Wolfgang Baumeister,

and Deborah Fass. Retrovirus envelope protein complex structure in situ

studied by cryo-electron tomography. Proceedings of the National Academy

of Sciences of the United States of America, 102(13):4729–4734, 2005.

[Fra06] J. Frank. Electron Tomography: Methods for Three-dimensional Visualization

of Structures in the Cell. Second Edition. Springer, 2006.

[GBH70] Richard Gordon, Robert Bender, and Gabor T. Herman. Algebraic recon-

struction techniques (art) for three-dimensional electron microscopy and x-

ray photography. Journal of Theoretical Biology, 29(3):471 – 481, 1970.

[GCSZ11] Hao Gao, Jian-Feng Cai, Zuowei Shen, and Hongkai Zhao. Robust principal

component analysis-based four-dimensional computed tomography. Physics

in Medicine and Biology, 56(11):3181–3198, 2011.

[GJ12] Lu Gan and Grant J. Jensen. Electron tomography of cells. Quarterly Reviews

of Biophysics, 45:27–56, 2012.

[GZY+06] Jianwei Gu, Li Zhang, Guoqiang Yu, Yuxiang Xing, and Zhiqiang Chen. X-

ray ct metal artifacts reduction through curvature based sinogram inpainting.

Journal of X-ray Science and Technology, 14:73–82, 2006.

BIBLIOGRAPHY 94

[HCWS08] J. Bernard Heymann, Giovanni Cardone, Dennis C. Winkler, and Alasdair C.

Steven. Computational resources for cryo-electron tomography in bsoft. Jour-

nal of Structural Biology, 161(3):232–242, 2008.

[HHM11] John M. Heumann, Andreas Hoenger, and David N. Mastronarde. Clustering

and variance maps for cryo-electron tomography using wedge-masked differ-

ences. Journal of Structural Biology, 175(3):288–299, 2011.

[HvH86] G. Harauz and M. van Heel. Exact filters for general geometry three dimen-

sional reconstruction. Optik, 73(4):146–156, 1986.

[JB07] Grant J Jensen and Ariane Briegel. How electron cryotomography is opening

a new window onto prokaryotic ultrastructure. Current Opinion in Structural

Biology, 17(2):260–267, 2007.

[KCW+10] Mikhail Kudryashev, Marek Cyrklaff, Reinhard Wallich, Wolfgang Baumeis-

ter, and Friedrich Frischknecht. Distinct in situ structures of the Borrelia

flagellar motor. Journal of Structural Biology, 169(1):54–61, 2010.

[KFB+13] Oleg Kuybeda, Gabriel A. Frank, Alberto Bartesaghi, Mario Borgnia, Sri-

ram Subramaniam, and Guillermo Sapiro. A collaborative framework for

3D alignment and classification of heterogeneous subvolumes in cryo-electron

tomography. Journal of Structural Biology, 181(2):116–127, 2013.

[KMM96] James R. Kremer, David N. Mastronarde, and J. Richard McIntosh. Com-

puter visualization of three-dimensional image data using IMOD. Journal of

Structural Biology, 116(1):71–76, 1996.

[KS01] Avinash C.. Kak and Malcolm Slaney. Principles of computerized tomographic

imaging. Society for Industrial and Applied Mathematics, 2001.

[KWS08] Cezar M. Khursigara, Xiongwu Wu, and Sriram Subramaniam. Chemore-

ceptors in caulobacter crescentus: Trimers of receptor dimers in a partially

BIBLIOGRAPHY 95

ordered hexagonally packed array. Journal of Bacteriology, 190(20):6805–

6810, 2008.

[KWZ+08] Cezar M. Khursigara, Xiongwu Wu, Peijun Zhang, Jonathan Lefman, and

Sriram Subramaniam. Role of hamp domains in chemotaxis signaling by

bacterial chemoreceptors. Proceedings of the National Academy of Sciences,

105(43):16555–16560, 2008.

[LDP07] Michael Lustig, David Donoho, and John M. Pauly. Sparse mri: The appli-

cation of compressed sensing for rapid mr imaging. Magnetic Resonance in

Medicine, 58(6):1182–1195, December 2007.

[LFB05] Vladan Lucic, Friedrich Forster, and Wolfgang Baumeister. Structural studies

by electron tomography: from cells to molecules. Annual Review of Biochem-

istry, 74(1):833–865, 2005.

[LJ09] Zhuo Li and Grant J Jensen. Electron cryotomography: a new view into

microbial ultrastructure. Current Opinion in Microbiology, 12(3):333–340,

2009.

[LN09] Sheetal Lahabar and PJ Narayanan. Singular value decomposition on gpu

using cuda. In Parallel & Distributed Processing, 2009. IPDPS 2009. IEEE

International Symposium on, pages 1–10. IEEE, 2009.

[lTB03] Fernando De la Torre and Michael J. Black. Robust parameterized component

analysis: theory and applications to 2D facial appearance models. Computer

Vision and Image Understanding, 91(1–2):53–71, 2003.

[MD07] Christopher P. Mercogliano and David J. DeRosier. Concatenated metalloth-

ionein as a clonable gold label for electron microscopy. Journal of Structural

Biology, 160(1):70 – 82, 2007.

[MHA+10] Farshid Moussavi, Geremy Heitz, Fernando Amat, Luis R. Comolli, Daphne

Koller, and Mark Horowitz. 3d segmentation of cell boundaries from whole

BIBLIOGRAPHY 96

cell cryogenic electron tomography volumes. Journal of Structural Biology,

170(1):134 – 145, 2010.

[MHL+05] W. Moss, S. Haase, J.M. Lyle, D.A. Agard, and J.W. Sedat. A novel 3d

wavelet-based filter for visualizing features in noisy biological data. Journal

of Microscopy, 219(2):43–49, 2005. cited By 13.

[MJCD08] Raghu Meka, Prateek Jain, Constantine Caramanis, and Inderjit S. Dhillon.

Rank minimization via online learning. In ICML, pages 656–663, 2008.

[MLJ06] Gavin E Murphy, Jared R Leadbetter, and Grant J Jensen. In situ structure

of the complete treponema primitia flagellar motor. Nature, 442(7106):1062–

1064, 2006.

[MNM04] Richard McIntosh, Daniela Nicastro, and David Mastronarde. New views of

cells in 3D: an introduction to electron tomography. Trends in cell biology,

15(1):43–51, 2004.

[Moa02] Maher Moakher. Means and averaging in the group of rotations. SIAM

journal on matrix analysis and applications, 24(1):1–16, 2002.

[MS09] Jacqueline L. S. Milne and Sriram Subramaniam. Cryo-electron tomogra-

phy of bacteria: progress, challenges and future prospects. Nature Reviews

Microbiology, 7(9):666–675, 2009.

[MSGA+14] Antonio Martinez-Sanchez, Inmaculada Garcia, Shoh Asano, Vladan Lucic,

and Jose-Jesus Fernandez. Robust membrane detection based on tensor vot-

ing for electron tomography. Journal of Structural Biology, 186(1):49 – 61,

2014.

[MSGF13] Antonio Martinez-Sanchez, Inmaculada Garcia, and Jose-Jesus Fernandez. A

ridge-based framework for segmentation of 3d electron microscopy datasets.

Journal of Structural Biology, 181(1):61 – 70, 2013.

BIBLIOGRAPHY 97

[NAB+08] Rajesh Narasimha, Iman Aganj, Adam E. Bennett, Mario J. Borgnia, Daniel

Zabransky, Guillermo Sapiro, Steven W. McLaughlin, Jacqueline L.S. Milne,

and Sriram Subramaniam. Evaluation of denoising algorithms for biological

electron tomography. Journal of Structural Biology, 164(1):7 – 17, 2008.

[NRRM+06] Rafael Nunez-Ramrez, Yolanda Robledo, Pablo Mesa, Silvia Ayora, Juan Car-

los Alonso, Jos Mara Carazo, and Luis Enrique Donate. Quaternary poly-

morphism of replicative Helicase G40P: structural mapping and domain re-

arrangement. Journal of Molecular Biology, 357(4):1063–1076, 2006.

[NSP+06] Daniela Nicastro, Cindi Schwartz, Jason Pierson, Richard Gaudette, Mary E.

Porter, and J. Richard McIntosh. The molecular architecture of axonemes

revealed by cryoelectron tomography. Science, 313(5789):944–948, 2006.

[OFK+06] Julio O. Ortiz, Friedrich Frster, Julia Krner, Alexandros A. Linaroudis, and

Wolfgang Baumeister. Mapping 70s ribosomes in intact cells by cryoelec-

tron tomography and pattern recognition. Journal of Structural Biology,

156(2):334 – 341, 2006.

[PGW+12] Yigang Peng, Arvind Ganesh, John Wright, Wenli Xu, and Yi Ma. RASL: Ro-

bust alignment via sparse and low-rank decomposition for linearly correlated

images. IEEE Transactions on Pattern Analysis and Machine Intelligence,

34(11):2233–2246, 2012.

[PL09] J. Provost and F. Lesage. The application of compressed sensing for photo-

acoustic tomography. Medical Imaging, IEEE Transactions on, 28(4):585

–594, 2009.

[PMV03] J.P.W. Pluim, J.B.A. Maintz, and M.A. Viergever. Mutual-information-based

registration of medical images: a survey. IEEE Transactions on Medical

Imaging, 22(8):986–1004, 2003.

BIBLIOGRAPHY 98

[PMvdG+13] Jack Poulson, Bryan Marker, Robert A. van de Geijn, Jeff R. Hammond, and

Nichols A. Romero. Elemental: A new framework for distributed memory

dense matrix computations. ACM Trans. Math. Softw., 39(2):13:1–13:24,

February 2013.

[RFP10] Benjamin Recht, Maryam Fazel, and Pablo A. Parrilo. Guaranteed minimum-

rank solutions of linear matrix equations via nuclear norm minimization.

SIAM Review, 52(3):471–501, 2010.

[RFR+01] Neil A. Ranson, George W. Farr, Alan M. Roseman, Brent Gowen, Wayne A.

Fenton, Arthur L. Horwich, and Helen R. Saibil. ATP-bound states of GroEL

captured by cryo-electron microscopy. Cell, 107(7):869–879, 2001.

[RGH+12] Alexander Rigort, David Gnther, Reiner Hegerl, Daniel Baum, Britta Weber,

Steffen Prohaska, Ohad Medalia, Wolfgang Baumeister, and Hans-Christian

Hege. Automated segmentation of electron tomograms for a quantitative de-

scription of actin filament networks. Journal of Structural Biology, 177(1):135

– 144, 2012. Ueli Aebi Festschrift.

[RK08] L. Reimer and H. Kohl. Transmission Electron Microscopy, Fifth Edition.

Springer Series in Optical Sciences. Springer, 2008.

[SB08] Michael F. Schmid and Christopher R. Booth. Methods for aligning and

for averaging 3D volumes with missing data. Journal of Structural Biology,

161(3):243–248, 2008.

[SED05] J.-L. Starck, M. Elad, and D.L. Donoho. Image decomposition via the combi-

nation of sparse representations and a variational approach. Image Processing,

IEEE Transactions on, 14(10):1570 –1582, Oct. 2005.

[SHR+15] Florian K. M. Schur, Wim J. H. Hagen, Michaela Rumlova, Tomas Ruml,

Barbara Muller, Hans-Georg Krausslich, and John A. G. Briggs. Structure

BIBLIOGRAPHY 99

of the immature HIV-1 capsid in intact virus particles at 8.8 A resolution.

Nature, 517:505–508, 2015.

[SME+09] Carlos OS Sorzano, Cedric Messaoudi, Matthias Eibauer, Jose Roman Bilbao-

Castro, Reiner Hegerl, S Nickell, Sergio Marco, and Jose Marıa Carazo.

Marker-free image registration of electron tomography tilt-series. BMC bioin-

formatics, 10(1):124, 2009.

[SMVC09] Sjors H.W. Scheres, Roberto Melero, Mikel Valle, and Jose-Maria Carazo.

Averaging of electron subtomograms and random conical tilt reconstructions

through likelihood optimization. Structure, 17(12):1563–1572, 2009.

[SMVM+04] C.O.S. Sorzano, R. Marabini, J. Velazquez-Muriel, J.R. Bilbao-Castro,

S.H.W. Scheres, J.M. Carazo, and A. Pascual-Montano. Xmipp: a new gen-

eration of an open-source image processing package for electron microscopy.

Journal of Structural Biology, 148(2):194–204, 2004.

[TM02] D.S. Taubman and M.W. Marcellin. JPEG 2000: Image Compression Funda-

mentals, Standards and Practice. Kluwer International Series in Engineering

and Computer Science, 2002.

[VGS08] A. Vedaldi, G. Guidi, and S. Soatto. Joint data alignment up to (lossy)

transformations. In Proceedings of the IEEE Conference on Computer Vision

and Pattern Recognition, pages 1–8, June 2008.

[vHF81] Marin van Heel and Joachim Frank. Use of multivariate statistics in analysing

the images of biological macromolecules. Ultramicroscopy, 6(2):187–194, 1981.

[Vol02] Niels Volkmann. A novel three-dimensional variant of the watershed trans-

form for segmentation of electron density maps. Journal of Structural Biology,

138(12):123 – 129, 2002.

BIBLIOGRAPHY 100

[VSS+11] Lenard M. Voortman, Sjoerd Stallinga, Remco H.M. Schoenmakers, Lucas J.

van Vliet, and Bernd Rieger. A fast algorithm for computing and cor-

recting the {CTF} for tilted, thick specimens in {TEM}. Ultramicroscopy,

111(8):1029 – 1036, 2011.

[Win07] Hanspeter Winkler. 3D reconstruction and processing of volumetric data

in cryo-electron tomography. Journal of Structural Biology, 157(1):126–137,

2007.

[WK04] Oliver Watzke and WilliA. Kalender. A pragmatic approach to metal ar-

tifact reduction in ct: merging of metal artifact reduced images. European

Radiology, 14:849–856, 2004.

[WML11] Qing Wang, Christopher P. Mercogliano, and Jan Lwe. A ferritin-based label

for cellular electron cryotomography. Structure, 19(2):147 – 154, 2011.

[WN12] Christian Wachinger and Nassir Navab. Entropy and Laplacian images: struc-

tural representations for multi-modal registration. Medical Image Analysis,

16(1):1–17, 2012.

[WSOV96] Ge Wang, D.L. Snyder, J.A. O’Sullivan, and M.W. Vannier. Iterative deblur-

ring for ct metal artifact reduction. Medical Imaging, IEEE Transactions on,

15(5):657–664, October 1996.

[WYG+09] J. Wright, A.Y. Yang, A. Ganesh, S.S. Sastry, and Yi Ma. Robust face recog-

nition via sparse representation. Pattern Analysis and Machine Intelligence,

IEEE Transactions on, 31(2):210–227, Feb 2009.

[WYZ12] Zaiwen Wen, Wotao Yin, and Yin Zhang. Solving a low-rank factorization

model for matrix completion by a nonlinear successive over-relaxation algo-

rithm. Mathematical Programming Computation, 4(4):333–361, 2012.

[XBA12] Min Xu, Martin Beck, and Frank Alber. High-throughput subtomogram

alignment and classification by Fourier space constrained fast volumetric

BIBLIOGRAPHY 101

matching. Journal of Structural Biology, 178(2):152–164, 2012. Special Issue:

Electron Tomography.

[XMS+09] Quanren Xiong, Mary K. Morphew, Cindi L. Schwartz, Andreas H. Hoenger,

and David N. Mastronarde. {CTF} determination and correction for low dose

tomographic tilt series. Journal of Structural Biology, 168(3):378 – 387, 2009.

[XRY+05] Dan Xia, John C. Roeske, Lifeng Yu, Charles A. Pelizzari, Arno J. Mundt,

and et al. A hybrid approach to reducing computed tomography metal arti-

facts in intracavitary brachytherapy. Brachytherapy, 4(1):18–23, 2005.

[YSRR10] Lingbo Yu, Robert R. Snapp, Teresa Ruiz, and Michael Radermacher. Proba-

bilistic principal component analysis with expectation maximization (PPCA-

EM) facilitates volume classification and estimates the missing data. Journal

of Structural Biology, 171(1):18–30, 2010.

[ZF03] Barbara Zitova and Jan Flusser. Image registration methods: a survey. Image

and Vision Computing, 21(11):977–1000, 2003.

[ZGLM12] Zhengdong Zhang, Arvind Ganesh, Xiao Liang, and Yi Ma. Tilt: Trans-

form invariant low-rank textures. International Journal of Computer Vision,

99(1):1–24, 2012.

[ZLBJ+06] Ping Zhu, Jun Liu, Julian Bess Jr, Elena Chertova, Jeffrey D. Lifson, Henry

Gris, Gilad A. Ofek, Kenneth A. Taylor, and Kenneth H. Roux. Distribu-

tion and three-dimensional structure of AIDS virus envelope spikes. Nature,

441(7095):847–852, 2006.

[ZRW+00] S. Zhao, D.D. Robeltson, G. Wang, B. Whiting, and K.T. Bae. X-ray ct

metal artifact reduction using wavelets: an application for imaging total hip

prostheses. Medical Imaging, IEEE Transactions on, 19(12):1238–1247, 2000.

BIBLIOGRAPHY 102

[ZWX10] Xiaomeng Zhang, Jing Wang, and Lei Xing. Metal artifact reduction in

computed tomography by constrained optimization. Medical Imaging 2010:

Physics of Medical Imaging, 7622(1):76221T, 2010.