Scarf
Scarf
A BSTRACT
arXiv:2106.15147v2 [cs.LG] 15 Mar 2022
1 I NTRODUCTION
In many machine learning tasks, unlabeled data is abundant but labeled data is costly to collect,
requiring manual human labelers. The goal of self-supervised learning is to leverage large amounts
of unlabeled data to learn useful representations for downstream tasks such as classification. Self-
supervised learning has proved critical in computer vision (Grill et al., 2020; Misra & Maaten, 2020;
He et al., 2020; Tian et al., 2019) and natural language processing (Song et al., 2020; Wang et al.,
2019; Raffel et al., 2019). Some recent examples include the following: Chen et al. (2020) showed
that training a linear classifier on the representations learned by their proposed method, SimCLR,
significantly outperforms previous state-of-art image classifiers and requires 100x fewer labels to do
so; Brown et al. (2020) showed through their GPT-3 language model that by pre-training on a large
corpus of text, only few labeled examples were required for task-specific fine-tuning for a wide range
of tasks.
A common theme of these advances is learning representations that are robust to different views or
distortions of the same input; this is often achieved by maximizing the similarity between views of
the same input and minimizing those of different inputs via a contrastive loss. However, techniques
to generate views or corruptions have thus far been, by and large, domain-specific (e.g. color
distortion (Zhang et al., 2016) and cropping (Chen et al., 2020) in vision, and token masking (Song
et al., 2020) in NLP). Despite the importance of self-supervised learning, there is surprisingly little
work done in finding methods that are applicable across domains and in particular, ones that can be
applied to tabular data.
In this paper, we propose S CARF, a simple and versatile contrastive pre-training procedure. We
generate a view for a given input by selecting a random subset of its features and replacing them by
random draws from the features’ respective empirical marginal distributions. Experimentally, we
test S CARF on the OpenML-CC18 benchmark (Vanschoren et al., 2013; Bischl et al., 2017; Feurer
et al., 2019), a collection of 72 real-world classification datasets. We show that not only does S CARF
pre-training improve classification accuracy in the fully-supervised setting but does so also in the
presence of label noise and in the semi-supervised setting where only a fraction of the available
training data is labeled. Moreover, we show that combining S CARF pre-training with other solutions
to these problems further improves them, demonstrating the versatility of S CARF and its ability to
1
Published as a conference paper at ICLR 2022
Maximize
Similarity
f g
Input Unchanged
Pre-training
Encoder f head g
(shared (shared InfoNCE
weights) weights)
Sample half
the indices to Corrupt by drawing
f g
corrupt. from the resp.
feature’s marginal.
Supervised Fine-Tuning
Input
Classification head
Classification
Loss
x f h
Figure 1: Diagram showing unsupervised S CARF pre-training (Top) and subsequent supervised fine-
tuning (Bottom). During pre-training, networks f and g are learned to produce good representations
of the input data. After pre-training, g is discarded and a classification head h is applied on top of the
learned f and both f and h are subsequently fine-tuned for classification.
learn effective task-agnostic representations. We then conduct extensive ablation studies, showing the
effects of various design choices and stability to hyperparameters. Our ablations show that S CARF’s
way of constructing views is more effective than alternatives. We show that S CARF is less sensitive
to feature scaling and is stable to various hyperparameters such as batch size, corruption rate, and
softmax temperature.
2 R ELATED W ORKS
A number of self-supervised learning techniques have been proposed in computer vision (Zhai
et al., 2019; Tung et al., 2017; Jing & Tian, 2020). One framework involves learning features based
on generated images through various methods, including using a GAN (Goodfellow et al., 2014;
Donahue et al., 2016; Radford et al., 2015; Chen et al., 2018), predicting pixels (Larsson et al.,
2016), predicting colorizations (Zhang et al., 2016; Larsson et al., 2017), ensuring local and global
consistency (Iizuka et al., 2017), and learning synthetic artifacts (Jenni & Favaro, 2018). Most related
to our approach are contrastive learning ones (Tian et al., 2019; Hassani & Khasahmadi, 2020; Oord
et al., 2018; Henaff, 2020; Li et al., 2016; He et al., 2020; Bojanowski & Joulin, 2017; Wang & Gupta,
2015; Gidaris et al., 2018). In particular, our framework is similar to SimCLR (Chen et al., 2020),
which involves generating views of a single image via image-based corruptions like random cropping,
color distortion and blurring; however, we generate views that are applicable to tabular data.
Self-supervised learning has had an especially large impact in language modeling (Qiu et al., 2020).
One popular approach is masked language modeling, wherein the model is trained to predict input
tokens that have been intentionally masked out (Devlin et al., 2018; Raffel et al., 2019; Song
et al., 2019) as well as enhancements to this approach (Liu et al., 2019; Dong et al., 2019; Bao
et al., 2020; Lample & Conneau, 2019; Joshi et al., 2020) and variations involving permuting the
tokens (Yang et al., 2019; Song et al., 2020). Denoising autoencoders have been used by training them
to reconstruct the input from a corrupted version (produced by, for example, token masking, deletion,
and infilling) (Lewis et al., 2019; Wang et al., 2019; Freitag & Roy, 2018). Contrastive approaches
2
Published as a conference paper at ICLR 2022
include randomly replacing words and distinguishing between real and fake phrases (Collobert et al.,
2011; Mnih & Kavukcuoglu, 2013), random token replacement (Mikolov et al., 2013; Clark et al.,
2020), and adjacent sentences (Joshi et al., 2020; Lan et al., 2019; de Vries et al., 2019).
Within the contrastive learning framework, the choice of loss function is significant. InfoNCE (Gut-
mann & Hyvärinen, 2010; Oord et al., 2018), which can be interpreted as a non-parametric estimation
of the entropy of the representation (Wang & Isola, 2020), is a popular choice. Since then there have
been a number of proposals (Zbontar et al., 2021; Grill et al., 2020; Hjelm et al., 2018); however, we
show that InfoNCE is effective for our framework.
Recently, Yao et al. (2020) adapted the contrastive framework to large-scale recommendation systems
in a way similar to our approach. The key difference is in the way the methods generate multiple
views. Yao et al. (2020) proposes masking random features in a correlated manner and applying a
dropout for categorical features, while our approach involves randomizing random features based on
the features’ respective empirical marginal distribution (in an uncorrelated way). Generating such
views for a task is a difficult problem: there has been much work done in understanding and designing
them (Wu et al., 2018; Purushwalkam & Gupta, 2020; Gontijo-Lopes et al., 2020; Lopes et al., 2019;
Perez & Wang, 2017; Park et al., 2019) and learning them (Ratner et al., 2017; Cubuk et al., 2020;
Ho et al., 2019; Lim et al., 2019; Zhang et al., 2019b; Tran et al., 2017; Tamkin et al., 2020).
Lastly, also similar to our work is VIME (Yoon et al., 2020), which proposes the same corruption
technique for tabular data that we do. They pre-train an encoder network on unlabeled data by
attaching “mask estimator” and “feature estimator” heads on top of the encoder state and teaching the
model to recover both the binary mask that was used for corruption as well as the original uncorrupted
input, given the corrupted input. The pre-trained encoder network is subsequently used for semi-
supervised learning via attachment of a task-specific head and minimization of the supervised loss
as well as an auto-encoder reconstruction loss. VIME was shown to achieve state-of-art results
on genomics and clinical datasets. The key differences with our work is that we pre-train using
a contrastive loss, which we show to be more effective than the denoising auto-encoder loss that
partly constitutes VIME. Furthermore, after pre-training we fine-tune all model weights, including
the encoder (unlike VIME, which only fine-tunes the task head), and we do so using task supervision
only.
3 S CARF
We now describe our proposed method (Algorithm 1), which is also described in Figure 1. For each
mini-batch of examples from the unlabeled training data, we generate a corrupted version x̃(i) for
3
Published as a conference paper at ICLR 2022
each example x(i) as follows. We sample some fraction of the features uniformly at random and
replace each of those features by a random draw from that feature’s empirical marginal distribution,
which is defined as the uniform distribution over the values that feature takes on across the training
dataset. Then, we pass both x(i) and x̃(i) through the encoder network f , whose output we pass
through the pre-train head network g, to get z (i) and z̃ (i) respectively. Note that the pre-train head
network `2 -normalizes the outputs so that they lie on the unit hypersphere – this has been found to be
crucial in practice (Chen et al., 2020; Wang & Isola, 2020). We train on the InfoNCE contrastive
loss, encouraging z (i) and z̃ (i) to be close for all i and z (i) and z̃ (j) to be far apart for i 6= j, and we
optimize over the parameters of f and g via SGD.
Then, to train a classifier for the task via fine-tuning, we take the encoder network f and attach a
classification head h which takes the output of f as its input and predicts the label of the example.
We optimize the cross-entropy classification loss and tune the parameters of both f and h.
While pre-training can be run for a pre-determined number of epochs, much like normal supervised
training, how many is needed largely depends on the model and dataset. To this end, we propose
using early stopping on the validation InfoNCE loss. Given unlabeled validation data, we cycle
through it for some number of epochs, running our proposed method to generate x(i) , x̃(i) pairs.
Once built, the loss on this static set is tracked during pre-training. Prototypical loss curves are shown
in the Appendix.
4 E XPERIMENTS
We evaluate the impact of S CARF pre-training on test accuracy after supervised fine-tuning in three
distinct settings: on the full dataset, on the full dataset but where only 25% of samples have labels and
the remaining 75% do not, and on the full dataset where 30% of samples undergo label corruption.
Datasets. We use 69 datasets from the public OpenML-CC18 benchmark1 under the CC-BY licence.
It consists of 72 real-world classification datasets that have been manually curated for effective
benchmarking. Since we’re concerned with tabular datasets in this work, we remove MNIST, Fashion-
MNIST, and CIFAR10. For each OpenML dataset, we form 70%/10%/20% train/validation/test
splits, where a different split is generated for every trial and all methods use the same splits. The
percentage used for validation and test are never changed and only training labels are corrupted for
the label noise experiments.
Dataset pre-processing. We represent categorical features by a one-hot encoding, and most of the
corruption methods explored in the ablations are on these one-hot encoded representations of the data
(with the exception of S CARF, where the marginal sampling is done before one-hot encoding). We
pre-process missing data as follows: if a feature column is always missing, we drop it. Otherwise,
if the feature is categorical, we fill in missing entries with the mode, or most frequent, category
computed over the full dataset. For numerical features, we impute it with the mean. We explore
scaling numerical features by z-score, min-max, and mean normalization. We find that for a vanilla
network (i.e. control), z-score normalization performed the best for all but three datasets (OpenML
dataset ids 4134, 28, and 1468), for which no scaling was optimal. We thus do not scale these three
datasets and z-score normalize all others.
Model architecture and training. Unless specified otherwise, we use the following settings across
all experiments. As described earlier, we decompose the neural network into an encoder f , a pre-
training head g, and a classification head h, where the inputs to g and h are the outputs of f . We
choose all three component models to be ReLU networks with hidden dimension 256. f consists of 4
layers, whereas both g and h have 2 layers. Both S CARF and the autoencoder baselines use g (for both
pre-training and co-training, described later), but for autoencoders, the output dimensionality is the
input feature dimensionality, and the mean-squared error reconstruction loss is applied. We train all
models and their components with the Adam optimizer using the default learning rate of 0.001. For
both pre-training and fine-tuning we use 128 batch size. Unsupervised pre-training methods all use
early stopping with patience 3 on the validation loss, unless otherwise noted. Supervised fine-tuning
uses this same criterion (and validation split), but classification error is used as the validation metric
for early stopping, as it performs slightly better. We set a max number of fine-tune epochs of 200 and
1
https://docs.openml.org/benchmark/
4
Published as a conference paper at ICLR 2022
Figure 2: Top: Win matrices comparing pre-training methods against each other, and their im-
provement to existing solutions. Bottom: Box plots showing the relative improvement of different
pre-training methods over baselines (y-axis is zoomed in). We see that S CARF pre-training adds value
even when used in conjunction with known techniques.
pre-train epochs of 1000, We use 10 epochs to build the static validation set. Unless otherwise noted,
we use a corruption rate c of 0.6 and a temperature τ of 1, for S CARF-based methods. All runs are
repeated 30 times using different train/validation/test splits. Experiments were run on a cloud cluster
of CPUs, and we used about one million CPU core hours in total for the experiments.
Evaluation methods. We use the following to effectively convey the results across all datasets.
Win matrix. Given M methods, we compute a “win” matrix W of size M × M , where the (i, j) entry
is defined as:
“Beats” and “loses” are only defined when the means are not a statistical tie (using Welch’s t-test
with unequal variance and a p-value of 0.05). A win ratio of 0/1 means that out of the 69 (pairwise)
comparisons, only one was significant and it was a loss. Since 0/1 and 0/69 have the same value but
the latter is more confident indication that i is worse than j, we present the values in fractional form
5
Published as a conference paper at ICLR 2022
and use a heat map. We add an additional column to the matrix that represents the minimum win
ratio across each row.
Box plots. The win matrix effectively conveys how often one method beats another but does not
capture the degree by which. To that end, for each method, we compute the relative percent
improvement over some reference method on each dataset. We then build box-plots depicting the
distribution of the relative improvement across datasets, plotting the observations as small points. We
only consider datasets where the means of the method and the reference are different with p-value
0.20. We use a larger p-value here than when computing the win ratio because otherwise some
methods would not have enough points to make the box-plots meaningful.
Baselines. We use the following baselines:
• Label smoothing (Szegedy et al., 2016), which has proved successful for accuracy (Müller
et al., 2019) and label noise (Lukasik et al., 2020). We use a weight of 0.1 on the smoothing
term.
• Dropout. We use standard dropout (Srivastava et al., 2014) using rate 0.04 on all layers.
Dropout has been shown to improve performance and robustness to label noise (Rusiecki,
2020).
• Mixup (Zhang et al., 2017), using α = 0.2.
• Autoencoders (Rumelhart et al., 1985). We use this as our key ablative pre-training baseline.
We use the classical autoencoder (“no noise AE”), the denoising autoencoder (Vincent
et al., 2008; 2010) using Gaussian additive noise (“add. noise AE”) as well as S CARF’s
corruption method (“S CARF AE”). We use MSE for the reconstruction loss. We try both
pre-training and co-training with the supervised task, and when co-training, we add 0.1 times
the autoencoder loss to the supervised objective. We discuss co-training in the Appendix as
it is less effective than pre-training.
• S CARF data-augmentation. In order to isolate the effect of our proposed feature corruption
technique, we skip pre-training and instead train on the corrupted inputs during supervised
fine-tuning. We discuss results for this baseline in the Appendix as it is less effective than
the others.
• Discriminative S CARF. Here, our pre-training objective is to discriminate between original
input features and their counterparts that have been corrupted using our proposed technique.
To this end, we update our pre-training head network to include a final linear projection and
swap the InfoNCE with a binary logistic loss. We use classification error, not logistic loss,
as the validation metric for early stopping, as we found it to perform slightly better.
• Self-distillation (Hinton et al., 2015; Zhang et al., 2019a). We first train the model on the
labeled data and then train the final model on both the labeled and unlabeled data using the
first models’ predictions as soft labels for both sets.
• Deep k-NN (Bahri et al., 2020), a recently proposed method for label noise. We set k = 50.
• Bi-tempered loss (Amid et al., 2019), a recently proposed method for label noise. We use 5
iterations, t1 = 0.8, and t2 = 1.2.
• Self-training (Yarowsky, 1995; McClosky et al., 2006). A classical semi-supervised method
– each iteration, we train on pseudo-labeled data (initialized to be the original labeled dataset)
and add highly confident predictions to the training set using the prediction as the label. We
then train our final model on the final dataset. We use a softmax prediction threshold of 0.75
and run for 10 iterations.
• Tri-training (Zhou & Li, 2005). Like self-training, but using three models with different
initial labeled data via bootstrap sampling. Each iteration, every model’s training set is
updated by adding only unlabeled points whose predictions made by the other two models
agree. It was shown to be competitive in modern semi-supervised NLP tasks (Ruder &
Plank, 2018). We use same hyperparameters as self-training.
Figure 2 shows our results. From the first win matrix plot, we see that all five pre-training techniques
considered improve over no pre-training (control), and that S CARF outperforms the others and has
6
Published as a conference paper at ICLR 2022
Figure 3: S CARF boosts baseline performance even when 30% of the training labels are corrupted.
Notably, it improves state-of-the-art label noise solutions like Deep k-NN.
more statistically significant wins. The second win matrix shows that S CARF pre-training boosts
the performance of mixup, label smoothing, distillation, and dropout, and it does so better than
alternatives. In other words, pre-training complements existing solutions, suggesting that a distinct
mechanism is at play here. The box plots expand on the second win matrix, showing the relative
improvement that each of the pre-training strategies confers over the baselines. Table 1 summarizes
the box-plots by the average relative gain. We observe that S CARF generally outperforms the
alternatives and adds a 1-2% relative gain across the board.
4.2 S CARF PRE - TRAINING IMPROVES PERFORMANCE IN THE PRESENCE TO LABEL NOISE
To show how pre-training improves model robustness when the data’s labels are unreliable, we do
as follows. Firstly, label noise robustness is often studied in two distinct settings - (1) when some
subset of the training data is guaranteed to be uncorrupted and that set is known in advance, (2)
when the entire dataset is untrustworthy. For simplicity, we consider setting 2 in our experiments.
We corrupt labels as follows: leaving the validation and test splits uncorrupted, we select a random
30% percent of the training data to corrupt and for each datapoint, we replace its label by one of
the classes, uniformly over all classes (including the datapoint’s true class). Results are shown in
Figure 3 and Table 1. Again, we see S CARF outperforms the rest and boosts all baselines by 2-3%.
7
Published as a conference paper at ICLR 2022
100% labeled training S CARF S CARF AE no noise AE add. noise AE S CARF disc.
control 2.352 2.244 1.107 1.559 0.574
dropout 1.609 1.196 0.623 1.228 -1.312
mixup 1.72 1.183 -0.377 0.971 -0.307
label smooth 1.522 0.711 -0.002 1.04 -0.894
distill 2.392 2.186 0.823 1.431 -0.394
25% labeled training
control 3.692 1.702 0.777 1.662 0.233
dropout 2.212 1.848 2.013 1.155 -0.322
mixup 2.809 0.73 0.106 0.439 0.466
label smooth 2.303 0.705 -0.564 0.196 -0.206
distill 3.609 2.441 1.969 2.263 1.795
self-train 3.839 2.753 1.672 2.839 2.559
tri-train 3.549 2.706 1.455 2.526 1.92
30% label noise
control 2.261 1.988 0.914 1.612 -1.408
dropout 2.004 2.058 0.9 1.471 -2.54
mixup 2.739 1.723 0.116 1.409 0.189
label smooth 2.558 1.474 0.703 1.395 -1.337
distill 2.881 2.296 -0.239 1.659 -0.226
deep knn 2.001 1.281 0.814 1.348 0.088
bitempered 2.68 2.915 0.435 1.387 -1.147
Table 1: Results using the fully labeled training data, only 25% of the labeled training data, and the
full training data subject to 30% label noise. Shown is the average relative gain in accuracy when
adding the pre-training methods (columns) to the reference methods (rows). Like the box-plots, we
filter out datasets using p-value 0.20. We see that S CARF consistently outperforms alternatives, not
only in improving control but also in improving methods designed specifically for the setting.
4.3 S CARF PRE - TRAINING IMPROVES PERFORMANCE WHEN LABELED DATA IS LIMITED
To show how pre-training helps when there is more unlabeled data than labeled ones, we remove
labels in the training split so that only 25% of the original split remains labeled. Autoencoders,
S CARF, self-training and tri-training all leverage the unlabeled remainder. Results are shown in
Figure 4 (Appendix) and Table 1. S CARF outperforms the rest, adding a very impressive 2-4% to all
baselines.
4.4 A BLATIONS
We now detail the importance of every factor in S CARF. We show only some of the results here; the
rest are in the Appendix.
Other corruption strategies are less effective and are more sensitive to feature scaling. Here,
we ablate the marginal sampling corruption technique we proposed, replacing it with the following
other promising strategies, while keeping all else fixed.
1. No corruption. We do not apply any corruption - i.e. x̃(i) = x(i) , in Algorithm 1. In this
case, the cosine similarity between positive pairs is always one and the model is learning to
make negative pairs as orthogonal as possible. Under the recent perspective (Wang & Isola,
2020) that the contrastive loss comprises two terms – one that encourages alignment between
views of the same example – and one that encourages the hypersphere embeddings to be
uniformly spread out – we see that with no corruption, pre-training may just be learning to
embed input examples uniformly on the hypersphere.
2. Mean corruption. After determining which features to corrupt, we replace their entries with
the empirical marginal distribution’s mean.
8
Published as a conference paper at ICLR 2022
We also examine the corruption strategies under the following ways of scaling the input features.
Figure 6 (Appendix) shows the results for z-score and min-max scaling. S CARF marginal sampling
generally outperforms the other corruption strategies for different types of feature scaling. Marginal
sampling is neat in that in addition to not having hyperparameters, it is invariant to scaling and
preserves the “units” of each feature. In contrast, even a simple multiplicative scaling requires the
additive noise to be scaled in the same way.
S CARF is not sensitive to batch size. Contrastive methods like SimCLR (Chen et al., 2020) have
shown consistent improvements upon increasing the batch size, N . There is a tight coupling between
the batch size and how hard the contrastive learning task is, since, in our case, the loss term for each
example i involves 1 positive and N − 1 negatives. The need for large (e.g. 5000) batch sizes has
motivated engineering solutions to support them (Gao & Zhang, 2021) and have been seen as grounds
for adopting other loss functions (Zbontar et al., 2021). Figure 5 (Appendix) compares a range of
batch sizes. We see that increasing the batch size past 128 did not result in significant improvements.
A reasonable hypothesis here is that higher capacity models and harder tasks benefit more from
negatives.
S CARF is fairly insensitive to the corruption rate and temperature. We study the effect of the
corruption rate c in Figure 5 (Appendix). We see that performance is stable when the rate is in
the range 50% − 80%. We thus recommend a default setting of 60%. We see a similar stability
with respect to the temperature hyperparameter (see Appendix). We recommend using a default
temperature of 1.
Tweaks to the corruption do not work any better. The Appendix shows the effect of four distinct
tweaks to the corruption method. We do not see any reason to use any of them.
Alternatives to InfoNCE do not work any better. We investigate the importance of our choice
of InfoNCE loss function and see the effects of swapping it with recently proposed alternatives
Alignment and Uniformity (Wang & Isola, 2020) and Barlow Twins (Zbontar et al., 2021). We found
that these alternative losses almost match or perform worse than the original and popular InfoNCE in
our setting. See the Appendix for details.
5 C ONCLUSION
Self-supervised learning has seen profound success in important domains including computer vision
and natural language processing, but little progress has been made for the general tabular setting.
We propose a self-supervised learning method that’s simple and versatile and learns representations
that are effective in downstream classification tasks, even in the presence of limited labeled data
or label noise. Potential negative side effects of this method may include learning representations
that reinforce biases that appear in the input data. Finding ways to mitigate this during the training
process is a potential direction for future research.
9
Published as a conference paper at ICLR 2022
R EFERENCES
Ehsan Amid, Manfred K Warmuth, Rohan Anil, and Tomer Koren. Robust bi-tempered logistic loss
based on bregman divergences. arXiv preprint arXiv:1906.03361, 2019.
Dara Bahri, Heinrich Jiang, and Maya Gupta. Deep k-nn for noisy labels. In International Conference
on Machine Learning, pp. 540–550. PMLR, 2020.
Hangbo Bao, Li Dong, Furu Wei, Wenhui Wang, Nan Yang, Xiaodong Liu, Yu Wang, Jianfeng Gao,
Songhao Piao, Ming Zhou, et al. Unilmv2: Pseudo-masked language models for unified language
model pre-training. In International Conference on Machine Learning, pp. 642–652. PMLR, 2020.
Bernd Bischl, Giuseppe Casalicchio, Matthias Feurer, Frank Hutter, Michel Lang, Rafael G Man-
tovani, Jan N van Rijn, and Joaquin Vanschoren. Openml benchmarking suites. arXiv preprint
arXiv:1708.03731, 2017.
Piotr Bojanowski and Armand Joulin. Unsupervised learning by predicting noise. In International
Conference on Machine Learning, pp. 517–526. PMLR, 2017.
Tom B Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal,
Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are
few-shot learners. arXiv preprint arXiv:2005.14165, 2020.
Ting Chen, Xiaohua Zhai, and Neil Houlsby. Self-supervised gan to counter forgetting. arXiv preprint
arXiv:1810.11598, 2018.
Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for
contrastive learning of visual representations. In International conference on machine learning, pp.
1597–1607. PMLR, 2020.
Kevin Clark, Minh-Thang Luong, Quoc V Le, and Christopher D Manning. Electra: Pre-training text
encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555, 2020.
Ronan Collobert, Jason Weston, Léon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa.
Natural language processing (almost) from scratch. Journal of machine learning research, 12
(ARTICLE):2493–2537, 2011.
Ekin D Cubuk, Barret Zoph, Jonathon Shlens, and Quoc V Le. Randaugment: Practical automated
data augmentation with a reduced search space. In Proceedings of the IEEE/CVF Conference on
Computer Vision and Pattern Recognition Workshops, pp. 702–703, 2020.
Wietse de Vries, Andreas van Cranenburgh, Arianna Bisazza, Tommaso Caselli, Gertjan van Noord,
and Malvina Nissim. Bertje: A dutch bert model. arXiv preprint arXiv:1912.09582, 2019.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep
bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
Jeff Donahue, Philipp Krähenbühl, and Trevor Darrell. Adversarial feature learning. arXiv preprint
arXiv:1605.09782, 2016.
Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou,
and Hsiao-Wuen Hon. Unified language model pre-training for natural language understanding
and generation. arXiv preprint arXiv:1905.03197, 2019.
Matthias Feurer, Jan N. van Rijn, Arlind Kadra, Pieter Gijsbers, Neeratyoy Mallik, Sahithya Ravi,
Andreas Mueller, Joaquin Vanschoren, and Frank Hutter. Openml-python: an extensible python
api for openml. arXiv, 1911.02490, 2019. URL https://arxiv.org/pdf/1911.02490.
pdf.
Markus Freitag and Scott Roy. Unsupervised natural language generation with denoising autoencoders.
arXiv preprint arXiv:1804.07899, 2018.
Luyu Gao and Yunyi Zhang. Scaling deep contrastive learning batch size with almost constant peak
memory usage. arXiv preprint arXiv:2101.06983, 2021.
10
Published as a conference paper at ICLR 2022
Spyros Gidaris, Praveer Singh, and Nikos Komodakis. Unsupervised representation learning by
predicting image rotations. arXiv preprint arXiv:1803.07728, 2018.
Raphael Gontijo-Lopes, Sylvia J Smullin, Ekin D Cubuk, and Ethan Dyer. Affinity and diversity:
Quantifying mechanisms of data augmentation. arXiv preprint arXiv:2002.08973, 2020.
Ian J Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil
Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial networks. arXiv preprint
arXiv:1406.2661, 2014.
Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre H Richemond, Elena
Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi
Azar, et al. Bootstrap your own latent: A new approach to self-supervised learning. arXiv preprint
arXiv:2006.07733, 2020.
Michael Gutmann and Aapo Hyvärinen. Noise-contrastive estimation: A new estimation principle
for unnormalized statistical models. In Proceedings of the Thirteenth International Conference on
Artificial Intelligence and Statistics, pp. 297–304. JMLR Workshop and Conference Proceedings,
2010.
Kaveh Hassani and Amir Hosein Khasahmadi. Contrastive multi-view representation learning on
graphs. In International Conference on Machine Learning, pp. 4116–4126. PMLR, 2020.
Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for
unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on
Computer Vision and Pattern Recognition, pp. 9729–9738, 2020.
Olivier Henaff. Data-efficient image recognition with contrastive predictive coding. In International
Conference on Machine Learning, pp. 4182–4192. PMLR, 2020.
Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. arXiv
preprint arXiv:1503.02531, 2015.
R Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bachman, Adam
Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation
and maximization. arXiv preprint arXiv:1808.06670, 2018.
Daniel Ho, Eric Liang, Xi Chen, Ion Stoica, and Pieter Abbeel. Population based augmentation:
Efficient learning of augmentation policy schedules. In International Conference on Machine
Learning, pp. 2731–2741. PMLR, 2019.
Satoshi Iizuka, Edgar Simo-Serra, and Hiroshi Ishikawa. Globally and locally consistent image
completion. ACM Transactions on Graphics (ToG), 36(4):1–14, 2017.
Simon Jenni and Paolo Favaro. Self-supervised feature learning by learning to spot artifacts. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2733–2742,
2018.
Longlong Jing and Yingli Tian. Self-supervised visual feature learning with deep neural networks: A
survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
Mandar Joshi, Danqi Chen, Yinhan Liu, Daniel S Weld, Luke Zettlemoyer, and Omer Levy. Spanbert:
Improving pre-training by representing and predicting spans. Transactions of the Association for
Computational Linguistics, 8:64–77, 2020.
Guillaume Lample and Alexis Conneau. Cross-lingual language model pretraining. arXiv preprint
arXiv:1901.07291, 2019.
Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu
Soricut. Albert: A lite bert for self-supervised learning of language representations. arXiv preprint
arXiv:1909.11942, 2019.
Gustav Larsson, Michael Maire, and Gregory Shakhnarovich. Learning representations for automatic
colorization. In European conference on computer vision, pp. 577–593. Springer, 2016.
11
Published as a conference paper at ICLR 2022
Gustav Larsson, Michael Maire, and Gregory Shakhnarovich. Colorization as a proxy task for
visual understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 6874–6883, 2017.
Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer
Levy, Ves Stoyanov, and Luke Zettlemoyer. Bart: Denoising sequence-to-sequence pre-training for
natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461,
2019.
Yin Li, Manohar Paluri, James M Rehg, and Piotr Dollár. Unsupervised learning of edges. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1619–1627,
2016.
Sungbin Lim, Ildoo Kim, Taesup Kim, Chiheon Kim, and Sungwoong Kim. Fast autoaugment. arXiv
preprint arXiv:1905.00397, 2019.
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike
Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining
approach. arXiv preprint arXiv:1907.11692, 2019.
Raphael Gontijo Lopes, Dong Yin, Ben Poole, Justin Gilmer, and Ekin D Cubuk. Improving
robustness without sacrificing accuracy with patch gaussian augmentation. arXiv preprint
arXiv:1906.02611, 2019.
Michal Lukasik, Srinadh Bhojanapalli, Aditya Menon, and Sanjiv Kumar. Does label smoothing
mitigate label noise? In International Conference on Machine Learning, pp. 6448–6458. PMLR,
2020.
David McClosky, Eugene Charniak, and Mark Johnson. Effective self-training for parsing. In
Proceedings of the Human Language Technology Conference of the NAACL, Main Conference, pp.
152–159, 2006.
Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeffrey Dean. Distributed repre-
sentations of words and phrases and their compositionality. arXiv preprint arXiv:1310.4546,
2013.
Ishan Misra and Laurens van der Maaten. Self-supervised learning of pretext-invariant representations.
In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.
6707–6717, 2020.
Andriy Mnih and Koray Kavukcuoglu. Learning word embeddings efficiently with noise-contrastive
estimation. Advances in neural information processing systems, 26:2265–2273, 2013.
Rafael Müller, Simon Kornblith, and Geoffrey Hinton. When does label smoothing help? arXiv
preprint arXiv:1906.02629, 2019.
Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive
coding. arXiv preprint arXiv:1807.03748, 2018.
Daniel S Park, William Chan, Yu Zhang, Chung-Cheng Chiu, Barret Zoph, Ekin D Cubuk, and
Quoc V Le. Specaugment: A simple data augmentation method for automatic speech recognition.
arXiv preprint arXiv:1904.08779, 2019.
Luis Perez and Jason Wang. The effectiveness of data augmentation in image classification using
deep learning. arXiv preprint arXiv:1712.04621, 2017.
Xipeng Qiu, Tianxiang Sun, Yige Xu, Yunfan Shao, Ning Dai, and Xuanjing Huang. Pre-trained
models for natural language processing: A survey. Science China Technological Sciences, pp.
1–26, 2020.
12
Published as a conference paper at ICLR 2022
Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep
convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi
Zhou, Wei Li, and Peter J Liu. Exploring the limits of transfer learning with a unified text-to-text
transformer. arXiv preprint arXiv:1910.10683, 2019.
Alexander J Ratner, Henry R Ehrenberg, Zeshan Hussain, Jared Dunnmon, and Christopher Ré.
Learning to compose domain-specific transformations for data augmentation. Advances in neural
information processing systems, 30:3239, 2017.
Sebastian Ruder and Barbara Plank. Strong baselines for neural semi-supervised learning under
domain shift. arXiv preprint arXiv:1804.09530, 2018.
David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. Learning internal representations by
error propagation. Technical report, California Univ San Diego La Jolla Inst for Cognitive Science,
1985.
Andrzej Rusiecki. Standard dropout as remedy for training deep neural networks with label noise. In
International Conference on Dependability and Complex Systems, pp. 534–542. Springer, 2020.
Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, and Tie-Yan Liu. Mass: Masked sequence to sequence
pre-training for language generation. arXiv preprint arXiv:1905.02450, 2019.
Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, and Tie-Yan Liu. Mpnet: Masked and permuted
pre-training for language understanding. arXiv preprint arXiv:2004.09297, 2020.
Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov.
Dropout: a simple way to prevent neural networks from overfitting. The journal of machine
learning research, 15(1):1929–1958, 2014.
Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking
the inception architecture for computer vision. In Proceedings of the IEEE conference on computer
vision and pattern recognition, pp. 2818–2826, 2016.
Alex Tamkin, Mike Wu, and Noah Goodman. Viewmaker networks: Learning views for unsupervised
representation learning. arXiv preprint arXiv:2010.07432, 2020.
Yonglong Tian, Dilip Krishnan, and Phillip Isola. Contrastive multiview coding. arXiv preprint
arXiv:1906.05849, 2019.
Toan Tran, Trung Pham, Gustavo Carneiro, Lyle Palmer, and Ian Reid. A bayesian data augmentation
approach for learning deep models. arXiv preprint arXiv:1710.10564, 2017.
Hsiao-Yu Fish Tung, Hsiao-Wei Tung, Ersin Yumer, and Katerina Fragkiadaki. Self-supervised
learning of motion capture. arXiv preprint arXiv:1712.01337, 2017.
Joaquin Vanschoren, Jan N. van Rijn, Bernd Bischl, and Luis Torgo. Openml: Networked science in
machine learning. SIGKDD Explorations, 15(2):49–60, 2013. doi: 10.1145/2641190.2641198.
URL http://doi.acm.org/10.1145/2641190.2641198.
Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. Extracting and
composing robust features with denoising autoencoders. In Proceedings of the 25th international
conference on Machine learning, pp. 1096–1103, 2008.
Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, Pierre-Antoine Manzagol, and
Léon Bottou. Stacked denoising autoencoders: Learning useful representations in a deep network
with a local denoising criterion. Journal of machine learning research, 11(12), 2010.
Liang Wang, Wei Zhao, Ruoyu Jia, Sujian Li, and Jingming Liu. Denoising based sequence-to-
sequence pre-training for text generation. arXiv preprint arXiv:1908.08206, 2019.
Tongzhou Wang and Phillip Isola. Understanding contrastive representation learning through align-
ment and uniformity on the hypersphere. In International Conference on Machine Learning, pp.
9929–9939. PMLR, 2020.
13
Published as a conference paper at ICLR 2022
Xiaolong Wang and Abhinav Gupta. Unsupervised learning of visual representations using videos.
In Proceedings of the IEEE international conference on computer vision, pp. 2794–2802, 2015.
Zhirong Wu, Yuanjun Xiong, Stella Yu, and Dahua Lin. Unsupervised feature learning via non-
parametric instance-level discrimination. arXiv preprint arXiv:1805.01978, 2018.
Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Ruslan Salakhutdinov, and Quoc V
Le. Xlnet: Generalized autoregressive pretraining for language understanding. arXiv preprint
arXiv:1906.08237, 2019.
Tiansheng Yao, Xinyang Yi, Derek Zhiyuan Cheng, Felix Yu, Ting Chen, Aditya Menon, Lichan
Hong, Ed H Chi, Steve Tjoa, Jieqi Kang, et al. Self-supervised learning for large-scale item
recommendations. arXiv preprint arXiv:2007.12865, 2020.
David Yarowsky. Unsupervised word sense disambiguation rivaling supervised methods. In 33rd
annual meeting of the association for computational linguistics, pp. 189–196, 1995.
Jinsung Yoon, Yao Zhang, James Jordon, and Mihaela van der Schaar. Vime: Extending the success of
self-and semi-supervised learning to tabular domain. Advances in Neural Information Processing
Systems, 33:11033–11043, 2020.
Jure Zbontar, Li Jing, Ishan Misra, Yann LeCun, and Stéphane Deny. Barlow twins: Self-supervised
learning via redundancy reduction. arXiv preprint arXiv:2103.03230, 2021.
Xiaohua Zhai, Avital Oliver, Alexander Kolesnikov, and Lucas Beyer. S4l: Self-supervised semi-
supervised learning. In Proceedings of the IEEE/CVF International Conference on Computer
Vision, pp. 1476–1485, 2019.
Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. mixup: Beyond empirical
risk minimization. arXiv preprint arXiv:1710.09412, 2017.
Linfeng Zhang, Jiebo Song, Anni Gao, Jingwei Chen, Chenglong Bao, and Kaisheng Ma. Be your
own teacher: Improve the performance of convolutional neural networks via self distillation. In
Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3713–3722,
2019a.
Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful image colorization. In European
conference on computer vision, pp. 649–666. Springer, 2016.
Xinyu Zhang, Qiang Wang, Jian Zhang, and Zhao Zhong. Adversarial autoaugment. arXiv preprint
arXiv:1912.11188, 2019b.
Zhi-Hua Zhou and Ming Li. Tri-training: Exploiting unlabeled data using three classifiers. IEEE
Transactions on knowledge and Data Engineering, 17(11):1529–1541, 2005.
14
Published as a conference paper at ICLR 2022
A A PPENDIX
We now present findings held out from the main text. Unless noted otherwise, we use the same
hyperparameters (i.e. c = 0.6, temperature τ = 1, etc.).
Figure 4 presents the results for the case when only 25% of the training labels are available. We see
that S CARF outperforms the rest here as well.
Figure 7 shows prototypical pre-training training and validation loss curves for S CARF. We see that
both losses tend to decrease rapidly at the start of training and then diminish slowly until the early
stopping point. We also notice the noisy (high variance) nature of the training loss - this is due to the
stochasticity of our corruption method.
We present more ablation results, where the metric is accuracy using 100% of the labeled data.
Figure 5 examines the impact of the batch size and corruption rate for the fully labeled, noiseless
setting.
Figure 6 shows the performance of a variety of alternative corruption strategies for SCARF under
z-score and min-max feature normalization. Results for mean scaling are shown in Figure 12.
I MPACT OF TEMPERATURE
Figure 8 shows the impact of the temperature term. While prior work considers temperature an
important hyperparameter that needs to be tuned, we see that a default of 1 (i.e. just softmax) works
the best in our setting.
• Corrupting one view is better than corrupting both the views. The likely explanation for
this is that at a corruption rate of 60%, corrupting both views i.i.d means that the fraction of
feature indices that contain the same value for both views is small - in other words, there is
less information between the two views and the contrastive task is harder.
• Corrupting the same set of feature indices within the mini-batch performs slightly worse
than random sampling for every example in the mini-batch.
• An alternative way to select feature indices to corrupt is to use Bernoulli’s: for each feature
index, corrupt it with probability (corruption rate) c, ensuring that at least one index is
corrupted. In the selection method we describe in Algorithm 1, a constant number of indices
are corrupted, whereas here it is variable. This alternative way of selecting indices performs
roughly the same.
• Once the feature indices to corrupt are determined, rather than corrupting them by drawing
from the empirical marginal distribution, we can instead draw a random example from the
training set and use its feature entries for corrupting all views for the mini-batch. This
performs worse.
15
Published as a conference paper at ICLR 2022
Figure 4: S CARF shows even more significant gain in the semi-supervised setting where 25% of
the data is labeled and the remaining 75% is not. Strikingly, pre-training with S CARF boosts the
performance of self-training and tri-training by several percent.
Figure 5: Win matrix for various batch sizes (Left) and corruption rates (Right) for the fully labeled,
noiseless setting.
16
Published as a conference paper at ICLR 2022
Figure 6: Left: Win matrix comparing different corruption strategies when z-score feature normaliza-
tion is used in the fully labeled, noiseless setting. Right: The same matrix but when min-max feature
scaling is used. We see that S CARF is better than alternative corruption strategies for different types
of feature scaling.
Figure 7: Training and validation loss curves for S CARF pre-training on Phonemes (dataset id 1489).
We observe that both losses drop rapidly initially and then taper off. The training curve is jittery
because of the random corruptions, but the validation curve isn’t because the validation dataset is
built once before training and is static throughout training.
A LTERNATIVE LOSSES
Figure 8 shows the impact of using two recently-proposed alternatives to the InfoNCE loss: Uniform
and Align (Wang & Isola, 2020) and Barlow Twins (Zbontar et al., 2021). We use 5e-3 for the
hyperparameter in Barlow, and equal weighting between the align and uniform loss terms. We find
no benefit in using these other losses.
Here we address the question: is it better to pre-train using S CARF or to apply it as a term in the
supervised training loss? In particular, Lco-train = Lsupervised + λcont Lcont , where Lcont is as described
in Algorithm 1. Figure 9 shows that pre-training outperforms co-training for a range of different λcont .
We see this is also the case for additive noise autoencoders.
Figure 10 shows that using S CARF only for data augmentation during supervised training performs
far worse than using it for pre-training.
17
Published as a conference paper at ICLR 2022
Figure 8: Left: Barlow Twins loss performs similar to InfoNCE while Uniform-Align performs
worse. Right: InfoNCE softmax temperature 1 performs well.
Figure 9: Left: Pre-training with S CARF beats co-training for a range of weights on the co-training
term. Right: The same is true for additive noise autoencoders.
Figure 10: Left: Using S CARF only for data augmentation during supervised training performs worse
than using it for pre-training. Right: Using InfoNCE error instead of InfoNCE loss as the validation
metric for early stopping degrades performance.
U SING I NFO NCE ERROR AS THE PRE - TRAINING VALIDATION METRIC IS WORSE
S CARF uses the same loss function (InfoNCE) for training and validation. If we instead use the
InfoNCE error for validation - where an error occurs when the model predicts an off-diagonal entry
of the batch-size by batch-size similarity matrix - downstream performance degrades, as shown in
Figure 10.
18
Published as a conference paper at ICLR 2022
Figure 11: Top left: Corrupting one view is better than corrupting both. Top right: Using different
random feature indices for each example in the mini-batch is better than using a same set across the
batch. Bottom left: Selecting a variable number of feature indices via coin flips performs similar
to the method described in Algorithm 1. Bottom right: Corrupting by replacing the features by the
features of a single drawn example for the whole mini-batch performs worse.
Figure 12: Left: S CARF’s corruption strategy remains competitive for “mean” feature scaling, as was
the case for both z-score and min-max scaling. Right: Comparison of different hyperparameters for
the uniform-align loss. a and u are the weights on the align and uniform terms respectively.
Figure 12 compares S CARF using the uniform-align loss, for different weights between the align
and uniform loss terms. The best performance is achieved with equal weight between the two, and
Figure 8 shows that this underperforms vanilla InfoNCE. We thus recommend using the vanilla
InfoNCE loss.
19
Published as a conference paper at ICLR 2022
A BSOLUTE PERFORMANCE
We have thus far presented results in terms of relative improvement over control since relative gain is
easier to summarize across the many datasets we consider. In this section, we present the absolute test
accuracy for all methods and datasets, for the keen reader. We also add a gradient-boosted decision
tree baseline using the XGBoost package, since decision trees are common among practitioners
working with tabular data. We use the Python API 2 , version 0.6, and choose the default settings for
XGBClassifier (max depth of 3, 100 estimators, learning rate of 0.1).
Tables 2, 3, 4 show the absolute test accuracies (averaged over the 30 trials) for each dataset and
baseline for the 100% training data, the 30% label noise, and the 25% (semi-supervised) training data
settings respectively.
T RAINING S PEED
We use an early stopping criteria when fine-tuning S CARF and baselines. Table 5 shows the number
of actual training epochs used (averaged over the 30 trials) for each dataset and baseline for the 100%
training data setting.
2
https://xgboost.readthedocs.io/en/latest/python/python_api.html
20
Published as a conference paper at ICLR 2022
dataset id control SCARF SCARF AE add. noise AE no noise AE SCARF disc. XGBoost
6 94.69 95.75 94.6 94.56 94.36 95.1 88.01
6332 74.24 75.4 71.76 69.35 66.6 69.34 77.59
54 75.45 76.75 75.14 75.83 75.88 78.44 76.16
50 90.54 97.53 98.68 97.33 95.91 96.66 90.61
46 93.47 94.59 94.63 94.96 94.33 93.36 95.72
469 18.54 19.08 19.89 19.14 19.79 19.17 21.21
458 98.96 99.51 99.34 99.2 99.11 99.19 98.46
4538 56.28 60.32 56.78 56.89 56.84 57.06 56.31
4534 96.08 95.85 95.91 95.8 95.9 96.46 94.66
44 93.4 94.46 93.53 93.45 93.49 93.55 94.54
4134 75.53 75.51 76.25 75.52 75.77 75.47 78.31
41027 90.96 91.13 91.73 92.15 92.32 91.68 81.75
40994 90.97 90.91 91.02 91.5 90.99 90.86 94.29
40984 90.56 91.97 90.32 90.46 90.53 90.69 92.66
40983 98.59 98.7 98.64 98.66 98.66 98.77 98.6
40982 73.48 73.27 73.47 73.7 73.56 72.99 78.07
40979 96.68 97.45 96.36 96.39 96.6 96.59 95.91
40978 94.84 95.13 95.36 95.81 94.75 94.74 97.33
40975 95.27 97.26 96.86 97.02 96.76 96.7 95.87
40966 98.44 98.77 98.21 98.46 98.68 97.87 97.05
40923 2.17 2.19 2.19 2.19 2.19 2.19 77.49
40701 91.73 92.36 92.36 92.27 92.14 92.29 95.31
40670 92.37 92.88 92.88 93.27 92.25 92.53 96.27
40668 83.38 82.97 83.29 83.17 83.26 83.17 74.11
40499 99.05 98.9 98.65 98.64 98.95 98.63 97.58
3 97.55 98.53 97.72 97.65 97.94 98.06 97.6
38 97.14 97.75 97.09 97.19 97.18 97.45 98.42
37 75.21 75.36 75.15 75.58 75.57 74.58 76.47
32 98.95 99.25 98.81 98.82 98.89 98.9 98.39
31 73.55 73.86 73.3 72.68 72.92 71.92 76.38
307 94.89 96.45 94.29 95.17 94.17 94.44 86.48
300 94.46 94.37 94.3 94.44 94.18 94.47 94.73
29 85.82 86.35 85.62 85.57 85.14 83.7 86.16
28 97.56 97.88 97.63 97.76 97.44 97.55 97.29
23 54.64 54.25 53.22 53.27 53.03 53.41 55.77
23517 51.59 51.55 51.53 51.51 51.69 51.57 52
23381 60.59 59.07 56.78 56.68 53.22 54.62 59.7
22 82.05 80.86 81.18 81.29 81.48 82.05 77.43
18 72.59 74.06 73.51 72.94 73.69 73.06 72.12
188 64.24 65.02 67.34 66.09 64.95 64.7 64.89
182 89.21 90.43 89.15 89.17 88.98 89.42 89.63
16 95.68 96.55 95.51 95.53 95.47 95.14 94.85
15 96.45 97.15 97.06 96.39 96.04 96.07 95.83
1590 85.55 85.6 85.51 85.51 85.54 85.5 86.45
151 82.77 83.55 82.69 82.67 82.34 83.18 83.27
1510 96.94 96.49 96.27 96.54 96.21 96.58 96.26
1501 91.41 93.76 91.55 91.6 92.49 92.35 91.33
14 79.27 80.99 81.05 80.67 80.89 78.82 82.34
1497 88.73 93.65 89.19 88.51 88.94 89.21 99.69
1494 86.76 86.71 86.96 86.59 87.29 86.44 86.32
1489 85.25 87.55 85.84 85.86 85.03 87.15 85.63
1487 93.47 94.25 94.27 94.31 94.27 93.96 94.34
1486 95.74 95.66 95.74 95.79 95.76 95.74 95.63
1485 54.2 70.16 68.76 68.24 64.12 56.73 71.6
1480 70.16 69.74 69.49 69.19 69.23 69.64 68.01
1478 96.85 97.16 97.08 97.12 97.07 94.98 98.38
1475 54.23 54.54 54.02 53.85 53.7 53.97 55.72
1468 93.45 94.04 92.95 91.68 92.72 85.79 91.27
1464 79.29 78.53 79.31 78.6 79.13 78.04 77.58
1462 99.5 99.94 99.75 99.72 99.77 99.92 99.49
1461 90.42 90.46 90.43 90.43 90.41 90.45 90.36
12 96.82 96.46 95.96 95.99 95.97 96.42 96.17
11 95.98 97.27 96.87 97.15 97.05 97.16 87.95
1068 93.04 93.18 93.23 93.06 93.12 93.18 93.78
1067 85.6 85.69 85.66 85.57 85.45 85.46 85.36
1063 84.17 84.06 84.06 83.71 82.16 83.7 83.27
1053 81.47 81.19 81.28 81.21 81.27 81.19 81.44
1050 89.84 89.49 89.24 89.27 89.4 89.7 88.9
1049 90.65 89.97 90.2 90.38 90.74 90.17 91.2
Table 2: Absolute test accuracy (percent; averaged over 30 trials) for the 100% training data setting,
for each dataset and every baseline, including a gradient-boosted decision tree (XGBoost) baseline
(max depth of 3, 100 estimators).
21
Published as a conference paper at ICLR 2022
dataset id control SCARF SCARF AE add. noise AE no noise AE SCARF disc. XGBoost
6 90.45 90.57 90.02 90.01 89.65 90.7 85.53
6332 69.39 70.99 63.41 62.78 60.25 62.96 72.75
54 72.21 71.94 70.84 71.11 71.93 72.76 72.92
50 79.64 87.66 90.52 86.7 85.14 88.12 87.4
46 84.43 92.1 92.76 93.05 91.89 85.77 94
469 19.47 19.02 20.08 19.89 19.41 18.14 19.52
458 96.65 99.27 98.06 97.97 97.1 95.47 96.07
4538 52.24 54.25 52.07 52.15 51.94 52.35 54.88
4534 94.13 94.38 94.2 94.09 94 94.31 93.66
44 91.37 93.21 91.67 91.34 91.31 91.76 93.13
4134 72.09 71.17 71.99 72.13 72.2 72.31 76.95
41027 85.54 85.31 85.43 85.42 85.46 85.53 81.54
40994 91.33 91.37 90.74 91.16 91.53 91.91 90.12
40984 88.63 89.99 88.44 88.51 88.08 88.68 90.71
40983 97.87 98.5 98.3 98.14 98.07 98.57 97.86
40982 70.54 70.92 71.26 70.97 71.11 70.19 73.69
40979 92.68 96.31 92.62 92.7 92.57 93.76 92.43
40978 91.87 92.77 93.81 94.12 93.64 92.52 96.48
40975 88.64 90.26 91.42 90.91 91.27 88.54 91.36
40966 90.78 88.52 88.65 90.12 91.86 87.37 89.58
40923 2.16 2.17 2.17 2.17 2.17 2.17 75.47
40701 88.77 89.34 88.87 88.54 88.92 89.41 93.7
40670 84.41 90.34 89.6 90.39 87.63 87.66 93.94
40668 80.62 80.27 80.66 80.7 80.71 80.55 73.67
40499 98.18 97.85 97.82 98.11 98.16 98.29 95.69
3 94.07 94.77 94.93 95.14 95.21 93.59 96.07
38 96.2 96.88 96.1 96.19 95.9 96.3 97.81
37 75.68 74.6 73.99 73.8 74.27 73.17 72.71
32 98.47 98.6 98.43 98.51 98.43 98.56 97.12
31 71.15 72.12 70.66 71.04 70.63 66.98 73
307 78.65 81.37 80.8 79.83 78.7 78.07 74.95
300 89.13 87.21 89.1 89.21 88.9 89.2 90.89
29 84.12 83.21 84.28 83.86 84.18 81.53 82.44
28 95.77 96.54 96.1 95.94 95.77 95.45 95.22
23 51 51.89 50.29 50.03 50.21 51.37 53.5
23517 51.33 51.12 51.29 51.25 51.38 51.26 51.7
23381 58.07 57.88 55.73 55.73 54.7 53.25 55.47
22 78.75 78.05 78.56 78.86 79.15 78.8 74.94
18 70.8 73.25 72.63 71.93 73.1 72.72 71.23
188 58.3 53.19 58.02 57.64 57.09 55.3 60.29
182 88.31 88.88 87.82 87.83 87.36 87.59 88.25
16 88.19 94.62 92.02 91.56 89.75 87.96 89.57
15 96.31 97.2 96.77 96.17 95.56 95.87 93.6
1590 84.81 85.19 85.07 85.06 85.06 85.05 86.23
151 80.08 81.15 80.06 80.14 80.02 80.52 82.97
1510 94.58 94.37 92.51 92.24 92.18 93.83 92.13
1501 78.73 89.29 81.16 81.73 82.16 84.43 85.6
14 72.34 78.57 77.87 77.52 74.92 71.53 77.64
1497 82.85 90.2 82.64 82.24 82.82 82.72 98.75
1494 84.18 84.97 85.03 85.06 85.18 84.72 83.4
1489 83.29 85.39 84.16 84.11 83.48 84.62 84.26
1487 92.61 92.97 93.27 93.41 92.9 93.16 93.55
1486 94.89 94.71 94.91 94.95 94.91 94.96 95.12
1485 53.97 67.03 66.74 66.62 62.55 54.21 65.67
1480 71.63 69.72 68.86 68.95 68.9 69.53 68.06
1478 94.13 94.42 94.75 94.71 94.71 94.74 96.37
1475 51.47 51.9 51.19 51.29 51.12 51.28 53.8
1468 83.07 90.76 83.39 79.58 78.46 64.89 87.3
1464 76.65 76.91 77.69 76.9 78.28 76.63 75.71
1462 98.97 99.55 99.14 99.21 99.52 99.69 98.07
1461 90.06 90.21 90.11 90.11 90.11 90.05 90.37
12 93.99 95.18 93.93 93.77 93.8 94.2 92.98
11 90.25 88.77 89.77 89.11 89.88 89.68 82.43
1068 92.45 92.44 92.32 92.24 92.1 92.73 92.82
1067 84.74 84.87 84.96 84.53 84.62 84.33 85.33
1063 83.81 83.76 83.59 83.63 82.95 83.95 80.95
1053 81.05 81.05 81.06 81.04 81.09 81.04 81.07
1050 89.08 88.94 88.96 89.11 89.46 89.32 88.81
1049 88.95 87.48 88.53 88.69 88.93 88.58 89.2
Table 3: Absolute test accuracy (percent; averaged over 30 trials) for the 30% label noise setting, for
each dataset and every baseline, including a gradient-boosted decision tree (XGBoost) baseline (max
depth of 3, 100 estimators).
22
Published as a conference paper at ICLR 2022
dataset id control SCARF SCARF AE add. noise AE no noise AE SCARF disc. XGBoost
6 87.22 89.71 87.12 87.38 87.01 88.87 85.02
6332 63.67 66.71 58.53 57.82 57.25 61.5 69.41
54 69.27 66.59 68.31 68.48 69.5 68.85 69.04
50 72.63 79.2 86.85 82.9 75.78 88.62 81.34
46 87.14 93.1 93.2 93.86 93.28 89.35 94.6
469 18.02 18.53 19.31 19.16 19.11 17.85 18.81
458 98.03 99.48 98.6 98.21 97.54 98.16 95.58
4538 50.65 53.75 50.79 50.32 50.45 50.03 54.05
4534 94.1 94.31 93.89 94.06 93.93 94.3 94.22
44 91.6 93.6 92 91.95 91.63 92.5 93.41
4134 68.31 68.44 69.65 69 68.78 68.45 74.58
41027 84.3 84.63 84.7 84.66 84.68 85.24 80.84
40994 90.63 90.96 90.96 90.93 90.97 90.96 92.69
40984 86.48 88.05 86.12 85.83 86.34 86.49 89.96
40983 94.72 97.65 94.85 94.65 97.72 98.58 97.92
40982 68.25 67.99 67.62 68.08 68.62 67.11 70.14
40979 92.81 96.21 92.58 92.15 92.56 93.55 91.33
40978 86.05 89.95 91.12 91.49 88.9 85.88 96.32
40975 72.47 90.55 91.14 90.53 90.02 71.68 91.98
40966 86.53 84.11 83.98 86.01 88.3 83.48 80.65
40923 2.16 2.16 2.16 2.16 2.16 2.16 74.35
40701 87.44 89.59 88.76 88.49 88.44 88.34 93.46
40670 81.67 90.97 89.69 90.27 87.58 88.22 94.37
40668 79.64 79.28 79.41 79.72 79.74 79.37 74.05
40499 97.88 97.65 97.13 97.59 98.13 97.78 94.59
3 93.04 95.44 94.02 94.84 94.34 94.59 96.16
38 95.13 96.92 95.39 95.17 94.53 95.39 97.7
37 70.37 72.24 71.15 71.2 70.21 71.58 72.94
32 98.39 98.61 98.19 98.15 97.94 98.4 96.9
31 70.53 71.66 68.48 68.54 65.82 69.33 73.12
307 69.68 72.47 73.49 72.6 72.31 70.88 63.97
300 91.66 87.01 89.98 90.42 90.58 91.06 90.3
29 83.6 83.57 80.7 81.46 81.93 82.28 84.59
28 96.43 96.56 96.38 96.08 96.23 96.12 95.07
23 49.56 51.45 45.64 46.48 44.84 49.95 51.48
23517 51.17 51.04 51.03 50.98 51.03 51.23 51.52
23381 58.83 58.7 55.62 55.37 53.37 56.43 55.5
22 77.72 77.03 76.27 76.8 77.8 76.34 73.72
18 70.71 71.37 71.05 70.53 72.39 71.11 70.11
188 55.18 46.41 50.56 53.42 53.06 51.72 57.82
182 86.98 88.36 87.37 87.34 86.53 86.62 87.89
16 88.25 93.91 90.7 90.44 89.6 89.97 87.54
15 96.65 97.01 96.7 96.45 94.54 95.54 95.19
1590 84.9 85.01 84.89 84.89 84.92 84.86 86.34
151 79.49 80.7 79.48 79.61 79.4 80.13 82.68
1510 94.63 94.36 93.8 93.87 92.73 94.23 94.12
1501 81.01 88.29 79.66 80.16 81.53 83.97 81.09
14 73.13 76.71 76.51 76.3 75.2 72.6 76.36
1497 80.19 90.15 80.88 80.14 80.18 82.09 98.96
1494 84.13 84.11 84.46 84.65 84.39 83.67 83.11
1489 81.95 84.81 83.27 82.72 82.04 84.18 83.97
1487 93.71 93.72 93.59 93.41 93.68 93.7 94.02
1486 94.75 94.64 94.75 94.84 94.83 94.81 95.38
1485 53.69 66.82 67.1 66.82 62.54 53.86 61.14
1480 70.5 69.57 68.72 69.42 69.62 70.63 68.52
1478 93.82 94.06 93.92 94.41 93.68 93.45 96.9
1475 49.24 50.63 49.67 49.35 49.65 49.06 52.08
1468 85.01 88.74 80.14 74.45 77.73 71.39 81.67
1464 75.26 75.37 75.7 75.36 75.4 75.17 75.82
1462 98.48 99.13 98.9 98.78 99.39 99.58 97.36
1461 89.87 89.94 90.04 89.94 89.92 89.81 90.22
12 91.83 94.38 92.78 92.68 92.95 92.86 91.49
11 88.38 87.99 85.48 85.84 89.13 84.27 82.43
1068 92.97 92.75 92.97 92.91 92.93 93 92.39
1067 84.88 84.87 84.77 84.7 84.27 84.06 84.24
1063 82.22 82.78 81.81 81.3 80.33 82.38 81.21
1053 80.82 80.91 80.87 80.89 80.95 80.94 80.98
1050 90.2 89.3 89.46 89.17 89.06 89.73 88.43
1049 87.59 87.14 87.33 87.47 88.11 87.95 89.18
Table 4: Absolute test accuracy (percent; averaged over 30 trials) for the 25% training data (semi-
supervised) setting, for each dataset and every baseline, including a gradient-boosted decision tree
(XGBoost) baseline (max depth of 3, 100 estimators).
23
Published as a conference paper at ICLR 2022
Table 5: Number of actual training epochs used (averaged over 30 trials) for the 100% training data
setting, for each dataset and every baseline.
24