Isotropic Reconstruction of 3D Fluorescence Microscopy Images Using Convolutional Neural Networks
Isotropic Reconstruction of 3D Fluorescence Microscopy Images Using Convolutional Neural Networks
1 Introduction
Fluorescence microscopy is a standard tool for imaging biological samples [15].
Acquired images of confocal microscopes [3] as well as light-sheet microscopes [4],
however, are inherently anisotropic, owing to a 3D optical point-spread function
(PSF) that is elongated along the axial (z) direction which typical leads to a 2
to 4-fold lower resolution along this axis. Furthermore, due to the mechanical
plane-by-plane acquisition modality of most microscopes, the axial sampling is
reduced as well, further reducing the overall resolution by a factor of 4 to 8. These
effects later render downstream data analysis, e.g. cell segmentation, difficult.
To circumvent this problem, multiple techniques are known and used: Classical
deconvolution methods [9,12] are arguably the most common of these. They
can be applied on already acquired data, however, their performance is typically
inferior to other more complex techniques. Some confocal systems, e.g. when
using two-photon excitation with high numerical aperture objectives and an
isotropic axial sampling, can acquire almost isotropic volumes [3,10] (cf. Fig. 3).
Downsides are low acquisition speed, high photo toxicity/bleaching, and large file
c Springer International Publishing AG 2017
M. Descoteaux et al. (Eds.): MICCAI 2017, Part II, LNCS 10434, pp. 126–134, 2017.
DOI: 10.1007/978-3-319-66185-8 15
Isotropic Reconstruction of 3D Fluorescence Microscopy Images 127
Fig. 1. (a) 3D images acquired on light microscopes are notoriously anisotropic due
to axial undersampling and optical point spread function (PSF) anisotropy. (b) The
IsoNet-2 architecture has a U-net [13] like topology and is trained to restore anisotrop-
ically blurred/downsampled lateral patches. After training it is applied to the axial
views.
and then applies the concept of self super-resolution [5] by learning from the
very same dataset for which we restore isotropic resolution.
2 Methods
Given the true fluorophore distribution f (x, y, z) the acquired volumetric image
g of a microscope can be approximated by the following process
g = P Sσ (h ⊗ f ) + η (1)
where h = h(x, y, z) is the point spread function (PSF) of the microscope, ⊗ is
the 3D convolution operation, Sσ is the axial downsampling/slicing operator by a
factor σ, P is the signal dependent noise operator (e.g. poisson noise) and η is the
detector noise. As the PSF is typically elongated along z and σ > 1, the lateral
slices gxy of the resulting volumetric images show a significant higher resolution
and structural contrast compared to the axial slices gxz and gyz (cf. Fig. 1a).
with a suitable chosen 3d rotated PSF h̃. To learn the inverse mapping pxy → gxy
n
we assemble lateral patches (gxy , pnxy )n∈N and train a fully convolutional neural
network [8] to minimize the pixel wise PSNR loss
L= −[20 log10 max gxy n
− 10 log10 |gxy
n
− g̃xy
n 2
| ] (3)
n
where g̃xy
n
is the output of the network when applied to pnxy . For choosing the
best h̃ we consider the two choices (i) full : h̃ = hrot where hrot is a rotated
version of the original PSF that is aligned with the lateral planes, and (ii) split:
h̃ = hsplit which is the solution to the deconvolution problem hrot = hiso ⊗
hsplit and hiso is the isotropic average of h. The later choice is motivated by
the observation that convolving lateral slices with hsplit leads to images with a
resolution comparable to the axially ones. After training we apply the network
on the unseen, anisotropically blurred, bicubic upsampled axial slices gxz of the
whole volume to get the final estimation output.
Isotropic Reconstruction of 3D Fluorescence Microscopy Images 129
IsoNet-1. Which is the proposed network architecture of [1] used for super res-
olution: C64,9,9 − C32,5,5 − C1,5,5 − C1,1,1 . Here the first layer acts as a feature
extractor whose output is mapped nonlinearly to the resulting image estimate
by the subsequent layers. After each convolutional layer a rectifying activation
function (ReLu) is applied.
IsoNet-2. Which is similar to the proposed network architecture of [13] for seg-
mentation which consists of a contractive part: C16,7,7 − M2,2 − C32,7,7 − M2,2 −
C64,7,7 − U2,2 − C32,7,7 − U2,2 − C16,7,7 − C1,1,1 and symmetric skip connections.
The contractive part of the network learns sparse representations of the input
whereas skip connections are sensitive to image details (cf. Fig. 1b). In contrast
to [13], however, the network learns the residual to the input.
For all datasets, both architectures were trained for 100 epochs with the
Adam optimizer [7] and a learning rate 5 · 10−3 . We furthermore use a dropout
of 20% throughout and apply data augmentation (flipped and rotated images)
where it is compatible with the symmetries of the PSF (i.e. whenever the latter
commutes with the augmentation symmetry).
3 Results
3.1 Synthetic Data
We use 3 different types of synthetic datasets of size 5123 that resemble typical
biological structures, as shown in Fig. 2: The uppermost row shows small axial
crops from a volume containing about 1500 simulated nuclei. The middle row
shows crops of membrane structures as they are frequently seen in tightly packed
cell epithelia. The last row shows both, simulated cell nuclei and surrounding
labeled membranes. All volumes were created in-silico by combining plausible
structure distributions, perlin-noise based textures and realistic camera noise.
Note that the first column shows the ground truth images that were used to
generate the isotropic ground truth, by convolving with the isotropic PSF, and
the blurred images that were subsampled and convolved with realistic PSFs in
order to resemble microscopic data. This third column (blurred) is then used as
the input to all our and other tested methods. The subsequent 6 columns show
the results of (i) Richardson-Lucy deconvolution [9], (ii) pure SRCNN [1], i.e.
disregarding the PSF, (iii) the IsoNet-1 using the full PSF, (iv) the IsoNet-1
using the anisotropic component of the PSF hsplit , (v) the IsoNet-2 using the
130 M. Weigert et al.
Table 1. Computed PSNR values against isotropic GT (upper rows), and against
GT (lower rows). PSF types are: gaussian (σxy /σz = 2/8); confocal with numerical
aperture NA = 1.1; light-sheet with NAdetect = 0.8 and NAillum = 0.1. Bold values
indicate best. Standard deviation in brackets (n = 10).
full PSF, and (vi) the IsoNet-2 using the split PSF. In addition to the visuals
given in the figure, Table 1 compares the PSNR of the full volumes with the two
ground truth versions, averaged over 10 different randomly created stacks per
dataset type. As can be seen, our method performs significantly (p < 0.01) best
in all cases. Note that failing to incorporate the PSF (as with pure SRCNN)
results in an inferior reconstruction.
First, we apply a global threshold using the intermodes method [14]. Then, holes
in thresholded image regions are closed. Finally, cells that clump together are
separated by applying a 3D watershed algorithm on the 3D euclidian distance
transform. This pipeline is freely available to a large audience in tools like Fiji
or KNIME. We applied this pipeline to the isometric ground truth data, the
blurred and subsampled input data, and the result produced by the IsoNet-2.
As evaluation metric we used SEG (ISBI Tracking Challenge 2013), the average
intersection over union of matching cells when compared to the ground truth
labels, that takes values in [0, 1], where 1 corresponds to a perfect voxel-wise
matching. The results for the different conditions SEGGT = 0.923 (isotropic
ground truth), SEGblurred = 0.742 (blurred input), and SEGIsoN et−2 = 0.913
(network output), demonstrate the effectiveness of our approach.
Fig. 3. Results on fluorescence microscopy images of liver tissue (data taken from [10]).
Nuclei (DAPI) and membrane (Phalloidin) staining of hepatocytes, imaged with a two-
photon confocal microscope (excitation wavelength 780 nm, NA = 1.3, oil immersion,
n = 1.49). We start from an isotropic acquisition (ground truth), simulate an anisotropic
acquisition (by taking every 8th slice), and compare the isotropic image to the IsoNet-2
recovered image. Scalebar is 50 µm.
where we trained IsoNet-2 to restore the raw axial (yz) slices. The final results
exhibit perceptual sharpness close to that of the higher quality raw lateral (xy)
slices, even when compared to multi deconvolution, demonstrating the ability
to restore isotropic resolution from a single volume in different experimental
settings.
Fig. 4. IsoNet-2 applied to (a) Drosophila and (b) C. elegans volumes. The image
quality of the recovered IsoNet-2 axial (yz) slices is significantly improved and shows
similar isotropic resolution when compared to the lateral (xy) slices. In (b) we addi-
tionally compare to the result of multiview deconvolution [11]. Scalebar (a) 50 µm, (b)
10 µm.
4 Discussion
course, not fill in missing information: If the axial sample rate would drop below
the Shannon limit (with respect to the smallest structures we are interested in
resolving), the proposed networks will not be able to recover the data. Source
code will be released at https://github.com/maweigert/isonet.
References
1. Dong, C., Loy, C.C., He, K., Tang, X.: Learning a deep convolutional network for
image super-resolution. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.)
ECCV 2014. LNCS, vol. 8692, pp. 184–199. Springer, Cham (2014). doi:10.1007/
978-3-319-10593-2 13
2. Dong, C., Loy, C.C., He, K., Tang, X.: Image super-resolution using deep convo-
lutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 38(2), 295–307 (2016)
3. Economo, M.N., Clack, N.G., Lavis, L.D., Gerfen, C.R., Svoboda, K., Myers, E.W.,
Chandrashekar, J.: A platform for brain-wide imaging and reconstruction of indi-
vidual neurons. Elife 5, e10566 (2016)
4. Huisken, J., Swoger, J., Del Bene, F., Wittbrodt, J., Stelzer, E.H.K.: Optical sec-
tioning deep inside live embryos by selective plane illumination microscopy. Science
(New York, N.Y.) 305, 1007–1009 (2004)
5. Jog, A., Carass, A., Prince, J.L.: Self super-resolution for magnetic resonance
images. In: Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W. (eds.)
MICCAI 2016. LNCS, vol. 9902, pp. 553–560. Springer, Cham (2016). doi:10.1007/
978-3-319-46726-9 64
6. Kim, J., Kwon Lee, J., Mu Lee, K.: Accurate image super-resolution using very
deep convolutional networks. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 1646–1654 (2016)
7. Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint
arXiv:1412.6980 (2014)
8. Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic
segmentation. In: Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, pp. 3431–3440 (2015)
9. Lucy, L.B.: An iterative technique for the rectification of observed distributions.
The Astron. J. 79, 745 (1974)
10. Morales-Navarrete, H., Segovia-Miranda, F., Klukowski, P., Meyer, K., Nonaka, H.,
Marsico, G., Chernykh, M., Kalaidzidis, A., Zerial, M., Kalaidzidis, Y.: A versatile
pipeline for the multi-scale digital reconstruction and quantitative analysis of 3D
tissue architecture. Elife 4, e11214 (2015)
11. Preibisch, S., Amat, F., Stamataki, E., Sarov, M., Singer, R.H., Myers, E.,
Tomancak, P.: Efficient bayesian-based multiview deconvolution. Nat. Methods
11, 645–648 (2014)
12. Richardson, W.H.: Bayesian-based iterative method of image restoration. JOSA
62(1), 55–59 (1972)
134 M. Weigert et al.
13. Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomed-
ical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F.
(eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). doi:10.
1007/978-3-319-24574-4 28
14. Sezgin, M., Sankur, B.I.: Survey over image thresholding techniques and quantita-
tive performance evaluation. J. Electron. Imaging 13(1), 146–168 (2004)
15. Tsien, R.Y.: The green fluorescent protein. Annu. Rev. Biochem. 67, 509–544
(1998)