0% found this document useful (0 votes)
8 views

Self-supervised Visualisation of Medical Image Datasets

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views

Self-supervised Visualisation of Medical Image Datasets

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 13

1–13, 2024

Self-supervised Visualisation of Medical Image Datasets

Ifeoma Veronica Nwabufo1,2 [email protected]


Jan Niklas Böhm1,2 [email protected]
Philipp Berens1,2 [email protected]
Dmitry Kobak1,2,3 [email protected]

1
Hertie Institute for AI in Brain Health, University of Tübingen, Germany
2
Tübingen AI Center, University of Tübingen, Germany
arXiv:2402.14566v2 [cs.CV] 24 Jul 2024

3
IWR, Heidelberg University, Germany

Abstract
Self-supervised learning methods based on data augmentations, such as SimCLR, BYOL,
or DINO, allow obtaining semantically meaningful representations of image datasets and
are widely used prior to supervised fine-tuning. A recent self-supervised learning method,
t-SimCNE, uses contrastive learning to directly train a 2D representation suitable for
visualisation. When applied to natural image datasets, t-SimCNE yields 2D visualisations
with semantically meaningful clusters. In this work, we used t-SimCNE to visualise medical
image datasets, including examples from dermatology, histology, and blood microscopy. We
found that increasing the set of data augmentations to include arbitrary rotations improved
the results in terms of class separability, compared to data augmentations used for natural
images. Our 2D representations show medically relevant structures and can be used to aid
data exploration and annotation, improving on common approaches for data visualisation.
Keywords: Self-supervised learning, augmentations, contrastive learning, data visualisation

1. Introduction
Medical image datasets have been quickly growing in size and complexity (Litjens et al.,
2017; Topol, 2019; Zhou et al., 2021). Whereas medical professionals can analyse, annotate,
and classify individual images, tasks involving large batches of images, ranging from data
curation and quality control to exploratory analysis, remain challenging.
Self-supervised learning (SSL) has recently emerged in computer vision as the dominant
paradigm for learning image representations suitable for downstream tasks (Balestriero et al.,
2023), and it has increasingly been adopted in medical imaging (Huang et al., 2023). In
contrastive learning methods, such as SimCLR (Chen et al., 2020), BYOL (Grill et al., 2020),
or DINO (Caron et al., 2021), data augmentation is used to generate different views of each
image, and a deep network is trained to keep these views close together in the representation
space. However, the learned representations are typically high-dimensional.
Recently, Böhm et al. (2023) suggested a self-supervised contrastive method, called
t-SimCNE, for 2D visualisation of image datasets. Using natural image datasets, the authors
demonstrated that t-SimCNE obtains semantically meaningful visualisations, representing
rich cluster structure and highlighting artefacts in the data. Their methods clearly outper-
formed existing 2D embedding methods like t-SNE (Van der Maaten and Hinton, 2008) and
UMAP (McInnes et al., 2020) for natural image data.

© 2024 I.V. Nwabufo, J.N. Böhm, P. Berens & D. Kobak.


Nwabufo Böhm Berens Kobak

a b
2

augnmentation
z1
ResNet H-Flip Crop Jitter Grayscale
+
data
proj. head z2 c

V-Flip 90° Rot. Rand Rot.

Figure 1: (a) In t-SimCNE, the network is trained to map two random augmentations of an
input image to close locations in the 2D output space. (b) Augmentations used for natural
images in t-SimCNE. (c) Additional augmentations suggested here for medical images.

Here we apply t-SimCNE to several medical microscopy datasets, and demonstrate that
it yields medically relevant visualisations, outperforming t-SNE visualisations of pretrained
networks. Furthermore, we show that the results improve when using rotational data
augmentations (Figure 1) informed by the rotational invariance of microscropy images. Our
code is available at https://github.com/berenslab/medical-t-simcne.

2. Related work
Contrastive learning methods have been widely applied to medical image datasets (for
a review, see Huang et al., 2023) but usually as pre-training for downstream tasks such
as classification or segmentation. Some recent works visualised high-dimensional SSL
representations; e.g. Cisternino et al. (2023) used UMAP of DINO to visualise histopathology
data. In contrast, our focus is on self-supervised visualisations trained end-to-end.
Contrastive learning relies on data augmentations to create several views of each image,
and the choice of data augmentations plays a crucial role in methods’ success (Tian et al.,
2020). A large number of works explored data augmentations for medical images in a
supervised setting (reviewed by Chlap et al., 2021; Goceri, 2023). In the self-supervised
context, van der Sluijs et al. (2023) studied the effect of augmentations on the representation
of X-ray images. For histopathology images, Kang et al. (2023) advocated for using rotations
and vertical flips, as well as staining-informed color transformations, while some other works
used neighbouring patches as positive pairs (Li et al., 2021; Wang et al., 2021).

3. Background: SimCLR and t-SimCNE


SimCLR (Chen et al., 2020) produces two augmentations for each image in a given mini-
batch of size b, resulting in 2b augmented images. Each pair of augmentations forms a
so-called positive pair, whereas all other possible pairs in the mini-batch form negative pairs.
The model is trained to maximise the similarity between the positive pair elements while
simultaneously minimising the similarity between the negative pair elements.
An augmented image xi is passed through a ResNet (He et al., 2016) backbone to give the
latent representation hi , which is then passed through a fully-connected projection head with
one hidden layer to yield the final output zi . SimCLR employs the InfoNCE loss function

2
Self-supervised Visualisation of Medical Image Datasets

Table 1: Used datasets.

Dataset Image dim. Sample size Classes Reference


Leukemia 28 × 28 18 365 7 Matek et al. (2019a)
Bloodmnist 28 × 28 17 092 8 Yang et al. (2023)
Dermamnist 28 × 28 10 015 2 Yang et al. (2023)
Pathmnist 28 × 28 107 180 9 Yang et al. (2023)
PCam16 96 × 96 327 680 2 Veeling et al. (2018)

(van den Oord et al., 2019), which for one positive pair (i, j) can be written as

exp sim(zi , zj )/τ
ℓij = − log P2b , (1)
k̸=i exp sim(zi , zk )/τ

where sim(x, y) = x⊤ y/ ∥x∥ · ∥y∥ is the cosine similarity and τ is a hyperparameter that


was set to 1/2 in Chen et al. (2020). Even though the loss function operates on zi (typically
128-dimensional), for downstream tasks, SimCLR uses the representations hi (Bordes et al.,
2022), typically at least 512-dimensional.
The idea of t-SimCNE (Böhm et al., 2023) is to make the network output (zi ) two-
dimensional so that it is directly suitable for data visualization. It does not make sense
to apply the cosine similarity to this representation in R2 as it would effectively normalise
the embeddings to lie on a one-dimensional circle. t-SimCNE replaces the exponential of
the scaled cosine similarity with the Cauchy similarity used in t-SNE (Van der Maaten and
Hinton, 2008): (1 + ∥x − y∥2 )−1 . The resulting loss function is
2b
1 X 1
ℓij = −log 2
+ log . (2)
1 + ∥zi − zj ∥ 1 + ∥zi − zk ∥2
k̸=i

Böhm et al. (2023) found that directly optimizing this loss is difficult, and suggested a
three-stage process. The first stage (1000 epochs) used a 128-dimensional output which was
then replaced with a 2D output and fine-tuned in the subsequent two stages (500 epochs).
For their experiments on CIFAR datasets, the authors used a ResNet18 with a modified
first layer kernel size of 3×3, and a projection head with hidden layer size of 1024 (Figure 1a).

4. Experimental setup
Datasets We used five publicly available medical image datasets with sample sizes ranging
from 10 000 to over 300 000 (Table 1). Three datasets we took from the Medmnistv2
collection (Yang et al., 2023), all consisting of 28 × 28 RGB images. Dermamnist is based
on the HAM10000 dataset (Tschandl et al., 2018), a collection of multi-source dermatoscopic
images of common pigmented skin lesions. The images are categorised into 7 classes, which
we reduced to binary labels: melanocytic nevi and other skin conditions. Bloodmnist is
based on a dataset of microscopy images of individual blood cells from healthy donors
(Acevedo et al., 2020), categorised into 8 classes corresponding to cell types. Pathmnist
is based on a dataset of non-overlapping patches from colorectal cancer histology slides

3
Nwabufo Böhm Berens Kobak

t-SNE in pixel space t-SNE of pre-trained ResNet18 t-SimCNE w/default augm. t-SimCNE with 90° rotations

EBO
LYT
MYO
MON
EOS
NGS
OTH

kNN acc. = 69.0 kNN acc. = 82.0 kNN acc. = 87.2 kNN acc. = 94.4
sil score = -0.09 sil score = -0.11 sil score = 0.13 sil score = 0.34

Figure 2: Visualisations of the Leukemia dataset. Small classes shown in black (‘OTH’ in the
legend). kNN accuracy and silhouette scores shown in each panel. (a) t-SNE of the original
images in the pixel space. (b) t-SNE of the 512-dimensional representation obtained via an
ImageNet-pretrained ResNet18 network. (c) t-SimCNE using the same augmentations as in
Böhm et al. (2023). (d) t-SimCNE using augmentations including 90° rotations and flips.
Note that the EBO class is well separated here, despite only consisting of 78 images.

(Kather et al., 2019), categorized into 9 classes corresponding to tissue types. The Leukemia
dataset (Matek et al., 2019b) contains microscopy images of white blood cells taken from
patients, some of which were diagnosed with acute myeloid leukemia. We resized 224 × 224
images to 28 × 28 and merged 9 rare classes (< 80 cells) into one, obtaining 7 classes.
The Patch Camelyon16 (PCam16) dataset (Veeling et al., 2018), adapted from the
Camelyon16 challenge (Bejnordi et al., 2017), consists of 96 × 96 patches from breast cancer
histology slides. There are two classes: metastases and non-metastases. A patch was
classified as metastases if there was any amount of tumor tissue in its central 32 × 32 region.

Augmentations Böhm et al. (2023) worked with natural images and used the same
data augmentations as Chen et al. (2020): cropping, horizontal flipping, color jittering,
and grayscaling (Figure 1b). Here we used all of these augmentations with the same
hyperparameters and probabilities (see Table S1 for ablations). We reasoned that the
semantics of microscopy or pathology images should be invariant to arbitrary rotations and
arbitrary flips (Kang et al., 2023). For that reason we considered two additional sets of
augmentations: (i) vertical flips and arbitrary 90° rotations; (ii) rotations by any arbitrary
angle (Figure 1c). In each case, all possible rotations were equally likely. When rotating an
image by an angle that is not a multiple of 90°, the corners need to be filled in (Figure 1c,
right). For this we used the average color of all border pixels across all images in a given
dataset. This color was dataset specific, but the same for all images in a dataset.

Architecture and training We used the original t-SimCNE implementation (Böhm


et al., 2023) with default parameters unless stated otherwise. For PCam16, we used the
unmodified ResNet18 (He et al., 2016) without the fully-connected layer. All networks were
trained from scratch on an NVIDIA RTX A6000 GPU with the batch size of 1024, except
for PCam16 where we had to reduce the batch size to 512 to fit it into GPU memory.

4
Self-supervised Visualisation of Medical Image Datasets

Figure 3: (a) t-SimCNE visualisation of the Leukemia dataset. Only a subset of classes is
listed in the legend. (b) t-SimCNE visualisation of the Bloodmnist dataset. (c) t-SimCNE
visualisation of the Dermamnist dataset. In all three cases, we used augmentations including
90° rotations and vertical flips.

Baselines For comparison, we applied t-SNE to images in pixel space, in pretrained


ResNet representation, and in SimCLR representation. The SimCLR models had the same
architecture as t-SimCNE models but with 128D output and were trained with SimCLR
loss (Eq. 1) for 1000 epochs. We then applied t-SNE to the 512-dimensional SimCLR
representation before the projector head. We took ImageNet-pretrained ResNet18 and
ResNet152 models from the PyTorch library (Paszke et al., 2019). To pass our images
through these networks, we resized all images to 256 × 256, center cropped to 224 × 224,
and normalized (following He et al., 2016). The resulting representations had 512 and 2048
dimensions respectively. We used openTSNE 1.0.1 (Poličar et al., 2019) with default settings
to reduce to 2D. When doing t-SNE of the PCam16 data in pixel space, we first performed
principal component analysis and only used the first 100 PCs as input to t-SNE.

Evaluation We used two metrics to evaluate the quality of 2D embeddings, with classi-
fication and clustering being two possible downstream tasks: kNN classification accuracy
(Pedregosa et al., 2011) with k = 15 and a 9:1 training/test split, and silhouette score
(Rousseeuw, 1987). For a single point x, the silhouette score s ∈ [−1, 1] is defined as
(b − w)/ max(w, b) where w is the average distance between x and points from the same
class, and b is the average distance between x and points from the closest other class. The
silhouette score of the entire embedding is the average s across all points. These two measures
are complementary: The kNN accuracy measures how well the classes are separated from
each other, while the silhouette score measures how far they are separated from each other.

5. Results
In this study, we asked (i) how the contrastive visualisation method t-SimCNE (Böhm et al.,
2023) could be applied to medical image datasets, and (ii) if the set of data augmentations
could be enriched compared to what is typically used on natural images.

5
Nwabufo Böhm Berens Kobak

Table 2: The kNN accuracy of 2D embeddings. Means ± standard deviations over three
runs; PCam16 experiments had only one run due to its large size.

Dataset
Method
Leukemia Bloodmnist Dermamnist Pathmnist PCam16
def. augm. 86.3 ± 0.7% 90.4 ± 0.3% 77.3 ± 0.6% 97.2 ± 0.2% 92.6%
t-SimCNE + 90° rot. 94.4 ± 0.1% 93.0 ± 0.3% 77.5 ± 0.3% 98.0 ± 0.0% 93.1%
+ rand. rot. 95.1 ± 0.2% 92.9 ± 0.1% 80.1 ± 0.7% 97.3 ± 0.0% 90.8%
def. augm. 95.0 ± 0.1% 94.0 ± 0.1% 81.9 ± 0.1% 98.1 ± 0.0% 96.3%
t-SNE of
+ 90° rot. 95.9 ± 0.1% 95.8 ± 0.1% 80.8 ± 0.6% 98.4 ± 0.0% 96.4%
SimCLR
+ rand. rot. 95.6 ± 0.1% 95.4 ± 0.1% 82.2 ± 0.2% 97.9 ± 0.1% 94.9%
pixel space 69.0% 73.2% 78.0% 56.9% 76.9%
t-SNE ResNet18 82.0% 78.1% 81.9% 87.2% 86.7%
ResNet152 72.9% 72.9% 81.0% 88.8% 86.4%

Table 3: Silhouette scores (Section 4) of 2D embeddings. Same format as in Table 2.

Dataset
Method
Leukemia Bloodmnist Dermamnist Pathmnist PCam16
def. augm. 0.13 ± 0.00 0.40 ± 0.00 0.13 ± 0.01 0.45 ± 0.02 0.04
t-SimCNE + 90° rot. 0.33 ± 0.01 0.44 ± 0.03 0.11 ± 0.00 0.48 ± 0.06 0.05
+ rand. rot. 0.52 ± 0.02 0.50 ± 0.01 0.13 ± 0.06 0.41 ± 0.03 0.05
def. augm. 0.21 ± 0.01 0.37 ± 0.00 0.14 ± 0.00 0.23 ± 0.01 0.16
t-SNE of
+ 90° rot. 0.23 ± 0.01 0.35 ± 0.02 0.14 ± 0.01 0.25 ± 0.01 0.13
SimCLR
+ rand. rot. 0.21 ± 0.00 0.37 ± 0.02 0.16 ± 0.00 0.26 ± 0.00 0.06
pixel space −0.09 0.07 0.08 −0.05 0.02
t-SNE ResNet18 −0.11 0.13 0.14 0.17 0.04
ResNet152 −0.15 0.03 0.14 0.19 0.05

We considered the Leukemia dataset as our first example (Figure 2). Naive application of
t-SNE to the raw images in pixel space resulted in an embedding with little class separation
and low kNN accuracy of 67.4% (Figure 2a). Passing all images through an ImageNet-
pretrained ResNet and then embedding them with t-SNE improved the kNN accuracy to
82.2% but visually the classes were still separated poorly (Figure 2b). Training t-SimCNE
with default data augmentations gave embeddings with 86.7% kNN accuracy (Table 2) and
much better visual class separation (Figure 2c and Table 3). This shows that t-SimCNE can
produce meaningful 2D visualizations of medical image datasets.
We reasoned that the set of data augmentations could be enriched to include 90° rotations
and flips because the semantics of blood microscopy images is rotationally invariant. When
training t-SimCNE with this set of data augmentations, the kNN accuracy increased to 94.0%.
Additionally including all possible rotations by an arbitrary angle as data augmentations
yielded the highest kNN accuracy (95.2%) and the highest silhouette score (0.24), indicating
that domain-specific augmentations can further improve t-SimCNE embeddings.
Across the five datasets considered in this study, we saw three different outcomes.
On microscropy datasets (Leukemia and Bloodmnist), t-SimCNE with random rotations

6
Self-supervised Visualisation of Medical Image Datasets

Adipose

Mucus

Background
Col-adenocarcinoma

Colon mucosa Smooth muscle

Cancer-ass. stroma

Lymphocytes
Debris

Figure 4: t-SimCNE visualisation of the Pathmnist dataset. Colours correspond to classes.


Images correspond to three random points close to the tip of the annotation line.

performed the best: it had by far the best silhouette score (Table 3) and visually the
most separated classes (Figure 3a,b). SimCLR followed by t-SNE has also benefited from
rotational augmentations. Compared to t-SimCNE, it had slightly higher kNN accuracies
(Figure 2), but much lower silhouette scores.
On pathology datasets (Pathmnist and PCam16), t-SimCNE with 90° rotations per-
formed the best. On Pathmnist, it had the highest silhouette score (Figure 4). On PCam16,
t-SimCNE showed clearer structures compared to SimCLR + t-SNE, but this difference was
not captured by the silhouette scores which on this dataset were all close to zero (Table 3).
This is because it only had two classes, whereas t-SimCNE separated images not only by
class but also by tissue types (Figure 5); this led to large within-class distances and hence
misleadingly low silhouette scores.
Finally, on the dermatology dataset (Dermamnist), performance of all methods was
similarly poor: SimCLR and t-SimCNE resulted in embeddings not very different from
t-SNE in pixel space (Figure 3c).
As a control experiment, we applied t-SimCNE with 90° rotations and vertical flips to
the CIFAR-10 dataset (Krizhevsky et al., 2009). It decreased the kNN accuracy from 89% to
76%. This confirms that rotation augmentations are not helpful for natural images because
they are not invariant to rotations, unlike microscopy and pathology images.
In the pathology datasets, t-SimCNE showed meaningful subclass structure. For example,
in Pathmnist, the debris class separated into three clearly distinct subsets (Figures 4), one
of which had markedly different staining colour. In the PCam16 dataset, the embedding
clearly split patches with and without metastasis, based on the density of chromatin and
variation in the size of the cells. The difference in visual appearance (different shades of
violet) between top-right and bottom-left likely reveals a technical artefact resulting from
different staining durations.

7
Nwabufo Böhm Berens Kobak

a b

no metastasis
metastasis

Figure 5: (a) t-SimCNE visualisation of the PCam16 dataset. (b) We superimposed a


10 × 10 grid over the embedding and selected one image in each square. Frame colours show
image classes. If a square had fewer than 100 images, no image was shown.

6. Discussion
In this paper, we showed that t-SimCNE (Böhm et al., 2023) can be successfully applied to
medical image datasets, yielding semantically meaningful visualisations, and benefits from
rotational data augmentations, leveraging rotational invariance of microscropy images. In
agreement with Böhm et al. (2023), t-SimCNE performed better than SimCLR + t-SNE
combination. Even though SimCLR tended to have slightly higher kNN accuracy, the
silhouette score was typically much lower: t-SimCNE achieved visually much stronger cluster
separation, which is useful for practical visualisations. Furthermore, parametric nature of
t-SimCNE allows to embed new (out-of-sample) images into an existing embedding.
We found that blood microscopy datasets benefited the most from random rotations,
while pathology datasets showed the best results with 90° rotations and flips. We believe it
is because in blood microscopy images, the semantically meaningful part is always in the
center (Figure 3a,b) and so the corners of the image may not be important. In contrast,
in histopathology images, the edges of the image may contain relevant information, which
may get rotated out of the image and replaced by solid-color triangles (Figure 1c). One of
the datasets, Dermamnist, exhibited poor results with all analysis methods. This may be
because in this dataset the images are too small to convey biomedically relevant information,
or because the sample size was insufficient (Table 1).
In conclusion, we argue that t-SimCNE is a promising tool for visualisation of medical
image datasets. It can be useful for quality control, highlighting artefacts and problems in
the data. It can also create a 2D map of cell types, tissue types, or medical conditions, which
can be useful not only for clinical purposes but also education and research, potentially
combined with an interactive image exploration tool. In the future, it may be interesting to
extend t-SimCNE to learn representations invariant to technical (e.g. staining) artefacts.

8
Self-supervised Visualisation of Medical Image Datasets

Acknowledgments
We thank Christian Schürch for discussion on histopathology data. This work was supported
by the German Science Foundation (Excellence Cluster 2064 “Machine Learning — New
Perspectives for Science”, project number 390727645), the Hertie Foundation, and the
Cyber Valley Research Fund (D.30.28739). The authors thank the International Max
Planck Research School for Intelligent Systems (IMPRS-IS) for supporting Jan Niklas Böhm.
Philipp Berens is a member of the Else Kröner Medical Scientist Kolleg “ClinbrAIn: Artificial
Intelligence for Clinical Brain Research”.

References
Andrea Acevedo, Anna Merino, Santiago Alférez, Ángel Molina, Laura Boldú, and José
Rodellar. A dataset of microscopic peripheral blood cell images for development of
automatic recognition systems. Data in Brief, 30, 2020.

Randall Balestriero, Mark Ibrahim, Vlad Sobal, Ari S. Morcos, Shashank Shekhar, Tom Gold-
stein, Florian Bordes, Adrien Bardes, Grégoire Mialon, Yuandong Tian, Avi Schwarzschild,
Andrew Gordon Wilson, Jonas Geiping, Quentin Garrido, Pierre Fernandez, Amir Bar,
Hamed Pirsiavash, Yann LeCun, and Micah Goldblum. A Cookbook of Self-Supervised
Learning. ArXiv, abs/2304.12210, 2023.

Babak Ehteshami Bejnordi, Mitko Veta, Paul Johannes Van Diest, Bram Van Ginneken,
Nico Karssemeijer, Geert Litjens, Jeroen AWM Van Der Laak, Meyke Hermsen, Quirine F
Manson, Maschenka Balkenhol, et al. Diagnostic assessment of deep learning algorithms
for detection of lymph node metastases in women with breast cancer. Jama, 318(22):
2199–2210, 2017.

Jan Niklas Böhm, Philipp Berens, and Dmitry Kobak. Unsupervised visualization of
image datasets using contrastive learning. In International Conference on Learning
Representations, 2023.

Florian Bordes, Randall Balestriero, Quentin Garrido, Adrien Bardes, and Pascal Vincent.
Guillotine Regularization: Improving Deep Networks Generalization by Removing their
Head. ArXiv, abs/2206.13378, 2022.

Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski,
and Armand Joulin. Emerging properties in self-supervised vision transformers. In
Proceedings of the IEEE/CVF International Conference on Computer Vision, pages
9650–9660, 2021.

Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework
for contrastive learning of visual representations. In International Conference on Machine
Learning, pages 1597–1607. PMLR, 2020.

Phillip Chlap, Hang Min, Nym Vandenberg, Jason Dowling, Lois Holloway, and Annette
Haworth. A review of medical image data augmentation techniques for deep learning
applications. Journal of Medical Imaging and Radiation Oncology, 65(5):545–563, 2021.

9
Nwabufo Böhm Berens Kobak

Francesco Cisternino, Sara Ometto, Soumick Chatterjee, Edoardo Giacopuzzi, Adam P


Levine, and Craig A Glastonbury. Self-supervised learning for characterising histomor-
phological diversity and spatial RNA expression prediction across 23 human tissue types.
bioRxiv, pages 2023–08, 2023.
Evgin Goceri. Medical image data augmentation: techniques, comparisons and interpreta-
tions. Artificial Intelligence Review, pages 1–45, 2023.
Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond,
Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad
Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised
learning. Advances in Neural Information Processing Systems, 33:21271–21284, 2020.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image
recognition. In Proceedings of the IEEE Conference on Computer vision and Pattern
Recognition, pages 770–778, 2016.
Shih-Cheng Huang, Anuj Pareek, Malte Jensen, Matthew P Lungren, Serena Yeung, and
Akshay S Chaudhari. Self-supervised learning for medical image classification: a systematic
review and implementation guidelines. NPJ Digital Medicine, 6(1):74, 2023.
Mingu Kang, Heon Song, Seonwook Park, Donggeun Yoo, and Sérgio Pereira. Benchmarking
Self-Supervised Learning on Diverse Pathology Datasets. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition, pages 3344–3354, 2023.
Jakob Nikolas Kather, Johannes Krisam, Pornpimol Charoentong, Tom Luedde, Esther
Herpel, Cleo-Aron Weis, Timo Gaiser, Alexander Marx, Nektarios A Valous, Dyke Ferber,
et al. Predicting survival from colorectal cancer histology slides using deep learning: A
retrospective multicenter study. PLoS Medicine, 16(1):e1002730, 2019.
Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny
images. 2009.
Bin Li, Yin Li, and Kevin W Eliceiri. Dual-stream multiple instance learning network for
whole slide image classification with self-supervised contrastive learning. In Proceedings of
the IEEE/CVF conference on computer vision and pattern recognition, pages 14318–14328,
2021.
Geert Litjens, Thijs Kooi, Babak Ehteshami Bejnordi, Arnaud Arindra Adiyoso Setio,
Francesco Ciompi, Mohsen Ghafoorian, Jeroen Awm Van Der Laak, Bram Van Ginneken,
and Clara I Sánchez. A survey on deep learning in medical image analysis. Medical Image
Analysis, 42:60–88, 2017.
C. Matek, S. Schwarz, C. Marr, and K. Spiekermann. A Single-cell Morphological Dataset
of Leukocytes from AML Patients and Non-malignant Controls. The Cancer Imaging
Archive, 2019a.
Christian Matek, Simone Schwarz, Karsten Spiekermann, and Carsten Marr. Human-level
recognition of blast cells in acute myeloid leukaemia with convolutional neural networks.
Nature Machine Intelligence, 1(11):538–544, 2019b.

10
Self-supervised Visualisation of Medical Image Datasets

Leland McInnes, John Healy, and James Melville. UMAP: Uniform Manifold Approximation
and Projection for Dimension Reduction, 2020.

Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan,
Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, An-
dreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank
Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. PyTorch: An
Imperative Style, High-Performance Deep Learning Library. In H. Wallach, H. Larochelle,
A. Beygelzimer, F. d’Alché Buc, E. Fox, and R. Garnett, editors, Advances in Neural
Information Processing Systems 32, pages 8024–8035. Curran Associates, Inc., 2019.

F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel,


P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau,
M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine Learning in Python.
Journal of Machine Learning Research, 12:2825–2830, 2011.

Pavlin G. Poličar, Martin Stražar, and Blaž Zupan. openTSNE: a modular Python library
for t-SNE dimensionality reduction and embedding. bioRxiv, 2019.

Peter J. Rousseeuw. Silhouettes: A graphical aid to the interpretation and validation of


cluster analysis. Journal of Computational and Applied Mathematics, 20:53–65, 1987.
ISSN 0377-0427.

Yonglong Tian, Chen Sun, Ben Poole, Dilip Krishnan, Cordelia Schmid, and Phillip Isola.
What makes for good views for contrastive learning? Advances in Neural Information
Processing Systems, 33:6827–6839, 2020.

Eric J Topol. High-performance medicine: the convergence of human and artificial intelligence.
Nature medicine, 25(1):44–56, 2019.

Philipp Tschandl, Cliff Rosendahl, and Harald Kittler. The HAM10000 dataset, a large col-
lection of multi-source dermatoscopic images of common pigmented skin lesions. Scientific
Data, 5(1):1–9, 2018.

Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation Learning with Contrastive
Predictive Coding, 2019.

Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-SNE. Journal of
Machine Learning Research, 9(11), 2008.

Rogier van der Sluijs, Nandita Bhaskhar, Daniel Rubin, Curtis Langlotz, and Akshay
Chaudhari. Exploring Image Augmentations for Siamese Representation Learning with
Chest X-Rays, 2023.

Bastiaan S Veeling, Jasper Linmans, Jim Winkens, Taco Cohen, and Max Welling. Rotation
equivariant CNNs for digital pathology. In Medical Image Computing and Computer
Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain,
September 16-20, 2018, Proceedings, Part II 11, pages 210–218. Springer, 2018.

11
Nwabufo Böhm Berens Kobak

Xiyue Wang, Sen Yang, Jun Zhang, Minghui Wang, Jing Zhang, Junzhou Huang, Wei Yang,
and Xiao Han. Transpath: Transformer-based self-supervised learning for histopathological
image classification. In Medical Image Computing and Computer Assisted Intervention–
MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October
1, 2021, Proceedings, Part VIII 24, pages 186–195. Springer, 2021.

Jiancheng Yang, Rui Shi, Donglai Wei, Zequan Liu, Lin Zhao, Bilian Ke, Hanspeter Pfister,
and Bingbing Ni. MedMNIST v2-A large-scale lightweight benchmark for 2D and 3D
biomedical image classification. Scientific Data, 10(1):41, 2023.

S Kevin Zhou, Hayit Greenspan, Christos Davatzikos, James S Duncan, Bram Van Ginneken,
Anant Madabhushi, Jerry L Prince, Daniel Rueckert, and Ronald M Summers. A review
of deep learning in medical imaging: Imaging traits, technology trends, case studies with
progress highlights, and future promises. Proceedings of the IEEE, 109(5):820–838, 2021.

12
Self-supervised Visualisation of Medical Image Datasets

Appendix

Table S1: Ablation study, removing individual augmentations from t-SimCNE. The full set
of augmentations included the default t-SimCNE augmentations plus arbitrary rotations
(kNN accuracy is given in percents).

Augmentations Leukemia BloodMNIST PathMNIST


kNN acc. Silhouette kNN acc. Silhouette kNN acc. Silhouette
All 95.1 ± 0.2 0.52 ± 0.02 92.9 ± 0.1 0.50 ± 0.01 97.3 ± 0.0 0.41 ± 0.03
No crops 79.7 ± 0.6 0.14 ± 0.00 76.0 ± 1.1 0.20 ± 0.01 59.8 ± 1.1 −0.02 ± 0.03
No color jitter 82.0 ± 0.2 −0.01 ± 0.01 90.0 ± 0.1 0.45 ± 0.02 94.3 ± 0.3 0.24 ± 0.02
No grayscaling 95.6 ± 0.4 0.52 ± 0.02 92.1 ± 0.3 0.44 ± 0.01 98.5 ± 0.0 0.39 ± 0.05

13

You might also like