BEIT: BERT Pre-Training of Image Transformers
BEIT: BERT Pre-Training of Image Transformers
Abstract
We introduce a self-supervised vision representation model BE I T, which stands
for Bidirectional Encoder representation from Image Transformers. Following
BERT [DCLT19] developed in the natural language processing area, we propose a
masked image modeling task to pretrain vision Transformers. Specifically, each
image has two views in our pre-training, i.e., image patches (such as 16×16 pixels),
and visual tokens (i.e., discrete tokens). We first “tokenize” the original image into
visual tokens. Then we randomly mask some image patches and fed them into
the backbone Transformer. The pre-training objective is to recover the original
visual tokens based on the corrupted image patches. After pre-training BE I T, we
directly fine-tune the model parameters on downstream tasks by appending task
layers upon the pretrained encoder. Experimental results on image classification
and semantic segmentation show that our model achieves competitive results with
previous pre-training methods.
1 Introduction
Transformer [VSP+ 17] has achieved promising performance in computer vision [DBK+ 20,
TCD+ 20]. However, empirical studies show that vision Transformers require more training data than
convolutional neural networks. In order to solve the data-hungry issue [LSB+ 21], self-supervised
pre-training is a promising solution to leverage large-scale image data. Several strands of methods
have been explored for vision Transformers, such as contrastive learning [CXH21, XLY+ 21], and
self-distillation [CTM+ 21].
Concurrently, BERT [DCLT19] has achieved great success in natural language processing. Its
masked language modeling task first randomly masks some proportion of tokens within a text, and
then recovers the masked tokens based on the Transformer encoding results of the corrupted text.
Motivated by BERT, we turn to the denoising auto-encoding idea to pretrain vision Transformers,
which has not been well studied by the vision community. It is challenging to directly apply BERT-
style pre-training for image data. First of all, there is no pre-exist vocabulary for vision Transformer’s
input unit, i.e., image patches. So we cannot simply employ a softmax classifier to predict over all
possible candidates for masked patches. In contrast, the language vocabulary, such as words and
BPE [SHB16], is well-defined and eases auto-encoding prediction. A straightforward alternative
is regarding the task as a regression problem, which predicts the raw pixels of masked patches.
However, such pixel-level recovery task tends to waste modeling capability on pre-training short-
range dependencies and high-frequency details [RPG+ 21]. Our goal is to overcome the above issues
for pre-training of vision Transformers.
In this work, we introduce a self-supervised vision representation model BE I T, which stands for
Bidirectional Encoder representation from Image Transformers. Inspired by BERT, we propose a
pre-training task, namely, masked image modeling (MIM). As shown in Figure 1, MIM uses two
∗
Contribution during internship at Microsoft. Correspondence to: Li Dong<[email protected]>, Furu
Wei<[email protected]>
Unused During Reconstructed
Visual Tokens
Pre-Training Image
123 234 456 567
987 876 765 543
Original Decoder
Tokenizer
Image 112 223 334 445
211 322 433 544
Blockwise
Masking BEIT Encoder
Position
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Embedding
+ + + + + + + + + + + + + + + + +
Flatten Patch
[S] [M] [M] [M] [M] [M]
Embedding
views for each images, i.e., image patches, and visual tokens. We split the image into a grid of patches
that are the input representation of backbone Transformer. Moreover, we “tokenize” the image to
discrete visual tokens, which is obtained by the latent codes of discrete VAE [RPG+ 21]. During
pre-training, we randomly mask some proportion of image patches, and feed the corrupted input to
Transformer. The model learns to recover the visual tokens of the original image, instead of the raw
pixels of masked patches.
We perform self-supervised learning and then fine-tune the pretrained BE I T on two downstream
tasks, i.e., image classification, and semantic segmentation. Experimental results indicate that BE I T
outperforms both from-scratch training and previous strong self-supervised models. Moreover, BE I T
is complementary to supervised pre-training. Performance of BE I T can be further improved by
intermediate fine-tuning with ImageNet labels. Ablation studies show that our proposed techniques
are critical to the effectiveness of BERT-style pre-training for image data. Apart from performance,
the improvements of convergence speed and stability of fine-tuning reduce training costs on end tasks.
In addition, we demonstrate that self-supervised BE I T can learn reasonable semantic regions via
pre-training, unleashing the rich supervision signals contained in images.
Our contributions are summarized as follows:
2 Methods
Given an input image x, BE I T encodes it to contextualized vector representations. As shown
in Figure 1, BE I T is pretrained by the masked image modeling (MIM) task in a self-supervised
2
learning manner. MIM aims at recovering the masked image patches based on encoding vectors. For
downstream tasks (such as image classification, and semantic segmentation), we append task layers
upon pretrained BE I T and fine-tune the parameters on the specific datasets.
The images have two views of representations in our method, namely, image patch, and visual tokens.
The two types serve as input and output representations during pre-training, respectively.
Following ViT [DBK+ 20], we use the standard Transformer [VSP+ 17] as the backbone network. So
the results can be directly compared with previous work in terms of the network architecture.
The input of Transformer is a sequence of image patches {xpi }N i=1 . The patches are then linearly
2
projected to obtain patch embeddings Expi , where E ∈ R(P C)×D . Moreover, we prepend a
special token [S] to the input sequence. We also add standard learnable 1D position embeddings
Epos ∈ RN ×D to patch embeddings. The input vectors H0 = [e[S] , Expi , . . . , ExpN ] + Epos is fed
into Transformer. The encoder contains L layers of Transformer blocks H l = Transformer(H l−1 ),
where l = 1, . . . , L. The output vectors of the last layer H L = [hL L L
[S] , h1 , . . . , hN ] are used as the
L
encoded representations for the image patches, where hi is the vector of the i-th image patch.
We propose a masked image modeling (MIM) task. We randomly mask some percentage of image
patches, and then predict the visual tokens that are corresponding to the masked patches.
2
https://github.com/openai/DALL-E
3
Figure 1 shows the overview of our method. As presented in Section 2.1, given an input image
x, we split it into N image patches ({xpi }N N
i=1 ), and tokenize it to N visual tokens ({zi }i=1 ). We
randomly mask approximately 40% image patches, where the masked positions are denoted as
M ∈ {1, . . . , N }0.4N . Next we replace the masked patchesS with a learnableN embedding e[M] ∈ R .
D
M p N
The corrupted image patches x = {xi : i ∈ / M}i=1 {e[M] : i ∈ M}i=1 are then fed into the
L-layer Transformer as described in Section 2.2. The final hidden vectors {hL N
i }i=1 are regarded as
encoded representations of the input patches. For each masked position {hi : i ∈ M}N
L
i=1 , we use a
softmax classifier to predict the corresponding visual tokens pMIM (z 0 |xM ) = softmaxz0 (Wc hL i +bc ),
where xM is the corrupted image, Wc ∈ R|V|×D , and bc ∈ R|V| . The pre-training objective is to
maximize the log-likelihood of the correct visual tokens zi given the corrupted image:
" #
X X
M
max EM log pMIM (zi |x ) (1)
x∈D i∈M
where D is the training corpus, M represents randomly masked positions, and xM is the corrupted
image that is masked according to M.
Rather than randomly choosing patches Algorithm 1 Blockwise Masking
for the masked positions M, we employ
Input: N (= h × w) image patches
blockwise masking in our work. As sum- Output: Masked positions M
marized in Algorithm 1, a block of image M ← {}
patches is masked each time. For each repeat
block, we set the minimum number of s ← Rand(16, 0.4N − |M|) . Block size
patches to 16. Then we randomly choose 1
r ← Rand(0.3, 0.3 ) . Aspect ratio of block
an aspect ratio for the masking block. We √ p
a ← s · r; b ← s/r
repeat the above two steps until obtaining t ← Rand(0,
S h − a) ; l ← Rand(0, w − b)
enough masked patches, i.e., 0.4N , where M ← M {(i, j) : i ∈ [t, t + a), j ∈ [l, l + b)}
N is the total number of image patches, until |M| > 0.4N . Masking ratio is 40%
and 0.4 is masking ratio. return M
The MIM task is greatly inspired by masked language modeling [DCLT19], which is one of the most
successful pre-training objective in natural language processing. Moreover, blockwise (or n-gram)
masking is also widely applied in BERT-like models [JCL+ 20, BDW+ 20, RSR+ 20]. However,
directly using pixel-level auto-encoding (i.e., recovering the pixels of masked patches) for vision pre-
training pushes the model to focus on short-range dependencies and high-frequency details [RPG+ 21].
BE I T overcomes the above issue by predicting discrete visual tokens, which summarizes the details
to high-level abstractions. Ablation studies in Section 3.3 show that our proposed method significantly
outperforms pixel-level auto-encoding.
The BE I T pre-training can be viewed as variational autoencoder [KW14] training. Let x denote the
original image, x̃ the masked image, and z the visual tokens. Considering the evidence lower bound
(ELBO) of the log-likelihood p(x|x̃), i.e., recovering the original image from its corrupted version:
X X
log p(xi |x̃i ) ≥ Ezi ∼qφ (z|xi ) [log pψ (xi |zi )] −DKL [qφ (z|xi ), pθ (z|x̃i )] (2)
(x ,x̃ )∈D (x ,x̃ )∈D
| {z }
i i i i
Visual Token Reconstruction
where (1) qφ (z|x) denotes the image tokenizer that obtains visual tokens; (2) pψ (x|z) decodes the
original image given input visual tokens; (3) pθ (z|x̃) recovers the visual tokens based on the masked
image, which is our MIM pre-training task.
We learn the model following a two-stage procedure similar to [vdOVK17, RvdOV19]. In the first
stage, we obtain the image tokenizer as a discrete variational autoencoder [RPG+ 21]. Specifically,
the first stage minimizes the reconstruction loss −Ezi ∼qφ (z|xi ) [log pψ (xi |zi )] with an uniform prior
as described in Equation (2). In the second stage, we learn the prior pθ while keeping qφ and
pψ fixed. We simplify qφ (z|xi ) to a one-point distribution with the most likely visual tokens
ẑi = arg maxz qφ (z|xi ). Then Equation (2) can be rewritten as:
X
Ezi ∼qφ (z|xi ) [log pψ (xi |zi )] + log pθ (ẑi |x̃i ) (3)
| {z }
(xi ,x̃i )∈D
| {z }
Stage 1: Visual Token Reconstruction Stage 2: Masked Image Modeling
4
where the second term is our BE I T pre-training objective.
The network architecture of BE I T follows that of ViT-Base [DBK+ 20] for a fair comparison. We
use a 12-layer Transformer with 768 hidden size, and 12 attention heads. The intermediate size of
feed-forward networks is 3072. We employ the default 16 × 16 input patch size. We directly borrow
the image tokenizer trained by [RPG+ 21]. The vocabulary size of visual tokens is 8192.
We pretrain BE I T on the training set of ImageNet-1K [RDS+ 15], which contains about 1.2M
images. Our augmentation policy includes random resized cropping, horizontal flipping, color
jittering [WXYL18]. Notice that we do not use the labels for self-supervised learning. We use the
224 × 224 resolution in our experiments. So the input is split to 14 × 14 image patches, and the same
amount of visual tokens. We randomly mask at most 75 patches (i.e., roughly 40% of total image
patches).
The pre-training runs for about 500k steps (i.e., 800 epochs) with 2k batch size. Adam [LH19] with
β1 = 0.9, β2 = 0.999 is employed for optimization. The learning rate is set to 1.5e-3, with a warmup
of 10 epochs, and cosine learning rate decay. The weight decay is 0.05. We employ stochastic
depth [HSL+ 16] with a 0.1 rate, and disable dropout. The 500k training steps take about five days
using 16 Nvidia Telsa V100 32GB GPU cards.
We find that proper initialization is important to stabilize Transformer, especially for large-scale pre-
training. We first randomly initialize all the parameters within a small range, such as [−0.02, 0.02].
Then, for the l-th Transformer layer, we rescale the output matrices (i.e., the last linear projection
within each sub-layer) of the self-attention module and the feed-forward network by √12l .
After pre-training BE I T, we append a task layer upon the Transformer, and fine-tune the parameters
on downstream tasks, like BERT. We take image classification and semantic segmentation as examples
in our work. It is straightforward to leverage the pre-training-then-fine-tuning paradigm on other
vision tasks with BE I T.
Image classification. For image classification tasks, we directly employ a simple linear clas-
sifier as the task layer. Specifically, we use average pooling to aggregate the representa-
tions, and feed the global to a softmax classifier. The category probabilities are computed
as softmax(avg({hL N L
i }i=1 Wc )), where hi is the final encoding vector of the i-th image patch,
D×C
Wc ∈ R is a parameter matrix, and C is the number of labels. We maximize the likelihood of
labeled data by updating the parameters of BE I T and the softmax classifier.
Semantic segmentation. For semantic segmentation, we follow the task layer used in SETR-
PUP [ZLZ+ 20]. To be specific, we use pretrained BE I T as a backbone encoder, and incorporate
several deconvolution layers as decoder to produce segmentation. The model is also end-to-end
fine-tuned similar to image classification.
3 Experiments
We conduct full fine-tuning experiments on image classification and semantic segmentation. Moreover,
we present various ablation studies for pre-training and analyze the representations learned by BE I T.
We also report linear probes on ImageNet in Appendix D.
5
3.1 Image Classification
The image classification task classifies input images to various categories. We evaluate BE I T on the
ILSVRC-2012 ImageNet dataset [RDS+ 15] with 1k classes and 1.3M images. We directly follow
the most of hyperparameters of DeiT [TCD+ 20] in our fine-tuning experiments for a fair comparison.
We reduce fine-tuning epochs compared with training from scratch, as BE I T has been pre-trained.
Accordingly, we use a larger learning rate with layer-wise decay. The detailed hyperparameters are
summarized in Appendix H.
Table 1 reports top-1 accuracy on image classification. We compare BE I T with vision Transformers
trained by random initialization, supervised pre-training, and previous self-supervised learning
methods. All the compared models are base-size, except iGPT has 1.36B parameters. Pre-training is
conducted on ImageNet for the comparison purpose, except ViT-JFT300M is pretrained on Google’s
in-house 300M images.
Compared with the models trained by random initialization, we find that pre-trained BE I T signifi-
cantly improves performance on both datasets. BE I T improves the performance on ImageNet, which
shows the effectiveness under the rich-resource setting.
Moreover, we compare BE I T with previous state-of-the-art self-supervised methods for Transformer,
such as DINO [CTM+ 21], and MoCo v3 [CXH21]. Our proposed method outperforms previous
models on ImageNet fine-tuning. Among them, iGPT-1.36B [CRC+ 20] uses much more parameters
(i.e., 1.36B vs 86M), and ViT-JFT300M [DBK+ 20] is pretrained on larger corpus (i.e., 300M vs
1.3M), while others pretrain ViT-Base on ImageNet-1K. iGPT-1.36B and ViT-JFT300M are the
most comparable methods, which also follows auto-encoding pre-training for vision Transformer.
Specifically, iGPT uses clustered image tokens as both input and output for image GPT or image
BERT. In contrast, we use image patches as input to preserve raw pixels, and employ discrete visual
tokens as a prediction bottleneck. ViT-JFT300 predicts the mean, 3-bit color of each masked patch,
rather than visual tokens learned by discrete VAE. We also pretrain the self-supervised tasks of BE I T
and DINO in a multi-task learning manner, which is presented in Appendix E.
In addition, we evaluate our proposed method with intermediate fine-tuning. In other words, we first
pretrain BE I T in a self-supervised manner, and then fine-tune the pretrained model on ImageNet with
labeled data. The results show that BE I T is complementary to supervised pre-training, achieving
additional gain after intermediate fine-tuning on ImageNet.
Fine-tuning to 384 × 384 resolution. After fine-tuning with resolution 224 × 224, we additionally
fine-tune the model on 384×384 images by 10 more epochs. We follow the standard higher-resolution
setting of DeiT [TCD+ 20], except using fewer epochs. Notice that we keep patch size the same for
both 224 × 224 and 384 × 384 images. So the input sequence length of Transformers becomes longer
for higher resolutions. Table 1 shows that higher resolution improves the BE I T results by 1+ points
on ImageNet. More importantly, BE I T384 pretrained on ImageNet-1K even outperforms supervised
pre-training ViT384 that uses ImageNet-22K, when they use the same input resolution.
Scaling up to larger size. We further scale up BE I T to the large size (same as ViT-L). As shown in
Table 1, ViT384 -L is worse than ViT384 on ImageNet, when training from scratch. The results verifies
the data-hungry issue of vision Transformers. Supervised pre-training on ImageNet-22K partially
relieves the issue, where ViT384 -L finally outperforms ViT384 by 1.2. In comparison, BE I T-L is
better than BE I T by 2.0, and BE I T384 -L outperforms BE I T384 by 1.7. In other words, the benefits
of scaling up BE I T from base to large are greater than supervised pre-training with ImageNet-22K.
More importantly, comparing between BE I T384 with ViT384 that conducts supervised pre-training
on ImageNet-22K, the improvements of BE I T become greater along with scaling the size from base
(i.e., 0.6) to large (i.e., 1.1). The results suggest that BE I T tends to help more for extremely larger
models (such as 1B, or 10B), especially when labeled data are insufficient3 to conduct supervised
pre-training4 for such large models.
3
[ZKHB21] report that supervised pre-training of a 1.8B-size vision Transformer requires billions of labeled
images.
4
Appendix B shows that BE I T fine-tuned on ImageNet-22K (14M) can match the performance of supervised
pre-training on Google’s in-house JFT-3B [ZKHB21], while using 214x less labels. We also demonstrate that
large-size BE I T fine-tuned on 70M labeled images can achieve 89.5% top-1 accuracy on ImageNet and 58.4%
mIoU on ADE20K, creating new state-of-the-art results for large-size vision Transformers.
6
Models Model Size Resolution ImageNet
Training from scratch (i.e., random initialization)
ViT384 -B [DBK+ 20] 86M 3842 77.9
ViT384 -L [DBK+ 20] 307M 3842 76.5
DeiT-B [TCD+ 20] 86M 2242 81.8
DeiT384 -B [TCD+ 20] 86M 3842 83.1
Supervised Pre-Training on ImageNet-22K (using labeled data)
ViT384 -B [DBK+ 20] 86M 3842 84.0
ViT384 -L [DBK+ 20] 307M 3842 85.2
Self-Supervised Pre-Training on ImageNet-1K (without labeled data)
iGPT-1.36B† [CRC+ 20] 1.36B 2242 66.5
ViT384 -B-JFT300M‡ [DBK+ 20] 86M 3842 79.9
MoCo v3-B [CXH21] 86M 2242 83.2
MoCo v3-L [CXH21] 307M 2242 84.1
DINO-B [CTM+ 21] 86M 2242 82.8
BE I T-B (ours) 86M 2242 83.2
BE I T384 -B (ours) 86M 3842 84.6
BE I T-L (ours) 307M 2242 85.2
BE I T384 -L (ours) 307M 3842 86.3
Table 1: Top-1 accuracy on ImageNet-1K. We evaluate base- (“-B”) and large-size (“-L”) models at
resolutions 224 × 224 and 384 × 384. † : iGPT-1.36B contains 1.36 billion parameters, while others
are base-size models. ‡ : ViT384 -B-JFT300M is pretrained with the “masked patch prediction” task
on Google’s in-house 300M images, while others use ImageNet.
80
Models ADE20K
Top-1 Acc.
75
Supervised Pre-Training on ImageNet 45.3
70
+
DeiT (Training from scratch) DINO [CTM 21] 44.1
65 BE I T (ours) 45.6
BEiT (Fine-tuning)
60 BE I T + Intermediate Fine-Tuning (ours) 47.7
50 100 150 200 250 300
Epochs
Table 3: Results of semantic segmentation on
Table 2: Convergence curves of training ADE20K. We use SETR-PUP [ZLZ+ 20] as the task
DeiT from scratch and fine-tuning BE I T on layer and report results of single-scale inference.
ImageNet-1K.
Convergence curves. Figure 2 compares the convergence curves of the training-from-scratch and
pre-training-then-fine-tuning paradigms. We find that fine-tuning BE I T not only achieves better
performance, but also converging much faster than training DeiT from scratch. Moreover, fine-tuning
BE I T can reach reasonable numbers within very few epochs.
Semantic segmentation aims to predict a corresponding class for each pixel of the input image. We
evaluate BE I T on the ADE20K benchmark [ZZP+ 19] with 25K images and 150 semantic categories.
We report the metric of mean Intersection of Union (mIoU) averaged over all semantic categories. As
presented in Section 2.6, we directly follow the task layer and the most of hyperparameters described
in SETR-PUP [ZLZ+ 20]. On ADE20K, we use Adam [LH19] as the optimizer. The learning rate is
set to 1e-3 with layer-wise decay similar to image classification. We conduct fine-tuning for 160K
steps. The batch size is 16. The detailed hyperparameters are described in Appendix I.
As shown in Table 3, we compare BE I T with supervised pre-training that relies on labeled data
of ImageNet. We find that our proposed method achieves better performance than supervised pre-
training, although BE I T does not require manual annotations for pre-training. Moreover, we employ
7
Models ImageNet ADE20K
BE I T (300 Epochs) 82.86 44.65
− Blockwise masking 82.77 42.93
− Visual tokens (i.e., recover masked pixels) 81.04 41.38
− Visual tokens − Blockwise masking 80.50 37.09
+ Recover 100% visual tokens 82.59 40.93
− Masking + Recover 100% visual tokens 81.67 36.73
Pretrain longer (800 epochs) 83.19 45.58
Table 4: Ablation studies for BE I T pre-training on image classification and semantic segmentation.
We conduct ablation studies to analyze the contributions of each component in BE I T. The models
are evaluated on image classification (i.e., ImageNet) and semantic segmentation (i.e., ADE20K). We
set the default pre-training steps to 300 epochs for the ablation studies, which is 37.5% of the total
steps used in the previous experiments.
Table 4 reports the results of various model variants. First, we ablate blockwise masking by randomly
sample masked positions. We find that blockwise masking is beneficial on both tasks, especially on
semantic segmentation. Second, we ablate the usage of visual tokens by predicting the raw pixels of
masked patches, i.e., the pre-training task becomes a pixel regression problem to recover masked
patches. Our proposed masked image modeling task significantly outperforms naive pixel-level
auto-encoding. Compared with the results in Table 1, the ablation result is worse than training vision
Transformer from scratch on two tasks. The results indicate that the prediction of visual tokens is the
key ingredient of BE I T. Third, we ablate the usage of visual tokens and blockwise masking together.
We find that blockwise masking is even more helpful for pixel-level auto-encoding, which relieves
the suffering of short-distance dependency. Forth, recovering all the visual tokens harms performance
on downstream tasks. Fifth, we compare BE I T with different training steps. Pre-training the model
longer can further improve performance on downstream tasks.
We show that the self-attention mechanism in BE I T can separate objects, even though our pre-training
does not rely on any manual annotation at all. Similar properties are also observed by [CTM+ 21].
The probing images are taken from the MS COCO [LMB+ 14] corpus to avoid appearing in the
pre-training data.
As shown in Figure 2, we plot the self-attention map for different reference points within an image.
The visualizations are produced by attention scores computed via query-key product in the last layer.
For each reference point, we use the corresponding patch as query, and show which patch it attends
to. After pre-training, BE I T learns to distinguish semantic regions using self-attention heads, without
any task-specific supervision. The property partially indicates the reason why BE I T is able to help
downstream tasks. Such knowledge acquired by BE I T potentially improves the generalization ability
of fine-tuned models, especially on small-scale datasets.
4 Related Work
Self-supervised visual representation learning. Various methods have been introduced over the
years to pretrain vision models in a self-supervised manner. Pioneering works design clever pretext
tasks, such as predicting the patch orderings [NF16], colorization [ZIE16], and predicting rotation
angles [KG18]. In addition, [TLL19] propose to mask some patches within an image, and classify
whether the masked patches are real or fake for each masked position. The method is similar to the
8
Figure 2: Self-attention map for different reference points. The self-attention mechanism in BE I T is
able to separate objects, although self-supervised pre-training does not use manual annotations.
masked version of Jigsaw pre-training [NF16]. The recent strand of research follows contrastive
paradigm [WXYL18, OLV18, HFLM+ 19, BHB19, HFW+ 20, CKNH20, CFGH20]. The models
typically regard various data augmentations as different views of an image, and then make the
representations of positive pairs similar while pushing negative pairs away. In order to obtain enough
informative negative samples in contrastive learning, the methods usually rely on large memory
banks [WXYL18, HFW+ 20] or large batch size [CKNH20]. BYOL [GSA+ 20] and SimSiam [CH20]
further eliminate the requirement of negative samples, using various techniques to avoid representation
collapse. Another strand of methods use clustering to organize image examples [CBJD18, ARV20,
CMM+ 20, LZXH21].
5 Conclusion
9
size. Moreover, we will conduct multimodal pre-training in a more unified way, using the similar
objectives and the shared architecture for texts and images.
Acknowledgement We would like to acknowledge Yue Cao, Han Hu, Hang Hua, Jingdong
Wang, Zheng Zhang for the helpful discussions, and Yaru Hao for some analysis experiments
using [HDWX20].
References
[ARV20] Yuki M. Asano, Christian Rupprecht, and Andrea Vedaldi. Self-labelling via simulta-
neous clustering and representation learning. In International Conference on Learning
Representations (ICLR), 2020.
[BDW+ 20] Hangbo Bao, Li Dong, Furu Wei, Wenhui Wang, Nan Yang, Xiaodong Liu, Yu Wang,
Jianfeng Gao, Songhao Piao, Ming Zhou, and Hsiao-Wuen Hon. UniLMv2: Pseudo-
masked language models for unified language model pre-training. In Proceedings of
the 37th International Conference on Machine Learning, ICML 2020, volume 119 of
Proceedings of Machine Learning Research, pages 642–652. PMLR, 2020.
[BHB19] Philip Bachman, R Devon Hjelm, and William Buchwalter. Learning representations
by maximizing mutual information across views. In Advances in Neural Information
Processing Systems, volume 32. Curran Associates, Inc., 2019.
[CBJD18] Mathilde Caron, Piotr Bojanowski, Armand Joulin, and Matthijs Douze. Deep clus-
tering for unsupervised learning of visual features. In Proceedings of the European
Conference on Computer Vision (ECCV), pages 132–149, 2018.
[CFGH20] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with
momentum contrastive learning. preprint arXiv:2003.04297, 2020.
[CH20] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning.
preprint arXiv:2011.10566, 2020.
[CKNH20] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A
simple framework for contrastive learning of visual representations. preprint
arXiv:2002.05709, 2020.
[CMM+ 20] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and
Armand Joulin. Unsupervised learning of visual features by contrasting cluster as-
signments. In Advances in Neural Information Processing Systems, volume 33, pages
9912–9924. Curran Associates, Inc., 2020.
[CRC+ 20] Mark Chen, Alec Radford, Rewon Child, Jeffrey Wu, Heewoo Jun, David Luan, and
Ilya Sutskever. Generative pretraining from pixels. In Hal Daumé III and Aarti
Singh, editors, Proceedings of the 37th International Conference on Machine Learning,
volume 119 of Proceedings of Machine Learning Research, pages 1691–1703. PMLR,
13–18 Jul 2020.
[CTM+ 21] Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bo-
janowski, and Armand Joulin. Emerging properties in self-supervised vision transform-
ers. arXiv preprint arXiv:2104.14294, 2021.
[CXH21] Xinlei Chen, Saining Xie, and Kaiming He. An empirical study of training self-
supervised vision transformers. ArXiv, abs/2104.02057, 2021.
[DBK+ 20] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua
Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Syl-
vain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition
at scale. preprint arXiv:2010.11929, 2020.
10
[DCLT19] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: pre-
training of deep bidirectional transformers for language understanding. In Proceedings
of the 2019 Conference of the North American Chapter of the Association for Compu-
tational Linguistics: Human Language Technologies, pages 4171–4186. Association
for Computational Linguistics, 2019.
[GSA+ 20] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre H Richemond,
Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Moham-
mad Gheshlaghi Azar, Bilal Piot, Koray Kavukcuoglu, Rémi Munos, and Michal Valko.
Bootstrap your own latent: A new approach to self-supervised learning. In NeurIPS,
2020.
[HDWX20] Yaru Hao, Li Dong, Furu Wei, and Ke Xu. Self-attention attribution: Interpreting
information interactions inside Transformer. arXiv preprint arXiv:2004.11207, 2020.
[HFLM+ 19] R Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bach-
man, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual
information estimation and maximization. In International Conference on Learning
Representations, 2019.
[HFW+ 20] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum
contrast for unsupervised visual representation learning. In CVPR, 2020.
[HSL+ 16] Gao Huang, Yu Sun, Zhuang Liu, Daniel Sedra, and Kilian Q. Weinberger. Deep
networks with stochastic depth. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max
Welling, editors, Computer Vision – ECCV 2016, pages 646–661, Cham, 2016. Springer
International Publishing.
[JCL+ 20] Mandar Joshi, Danqi Chen, Yinhan Liu, Daniel S. Weld, Luke Zettlemoyer, and
Omer Levy. SpanBERT: Improving pre-training by representing and predicting spans.
Transactions of the Association for Computational Linguistics, 8:64–77, 2020.
[JGP17] Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparameterization with gumbel-
softmax. In 5th International Conference on Learning Representations, ICLR 2017,
Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net,
2017.
[KG18] Nikos Komodakis and Spyros Gidaris. Unsupervised representation learning by pre-
dicting image rotations. In International Conference on Learning Representations
(ICLR), 2018.
[KH09] A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images.
Master’s thesis, Department of Computer Science, University of Toronto, 2009.
[KW14] Diederik P. Kingma and Max Welling. Auto-Encoding Variational Bayes. In 2nd
International Conference on Learning Representations, ICLR 2014, 2014.
[LH19] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In Interna-
tional Conference on Learning Representations, 2019.
[LLC+ 21] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin,
and Baining Guo. Swin Transformer: Hierarchical vision transformer using shifted
windows. arXiv preprint arXiv:2103.14030, 2021.
[LMB+ 14] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva
Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in
context. In European conference on computer vision, pages 740–755. Springer, 2014.
[LSB+ 21] Yahui Liu, Enver Sangineto, Wei Bi, Nicu Sebe, Bruno Lepri, and Marco De Nadai.
Efficient training of visual transformers with small datasets. In Thirty-Fifth Conference
on Neural Information Processing Systems, 2021.
11
[LZXH21] Junnan Li, Pan Zhou, Caiming Xiong, and Steven Hoi. Prototypical contrastive
learning of unsupervised representations. In International Conference on Learning
Representations, 2021.
[MMT17] Chris J. Maddison, Andriy Mnih, and Yee Whye Teh. The Concrete Distribution: A
Continuous Relaxation of Discrete Random Variables. In International Conference on
Learning Representations, 2017.
[NF16] Mehdi Noroozi and Paolo Favaro. Unsupervised learning of visual representations
by solving jigsaw puzzles. In European conference on computer vision, pages 69–84.
Springer, 2016.
[OLV18] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with
contrastive predictive coding. preprint arXiv:1807.03748, 2018.
[PPL+ 20] Yada Pruksachatkun, Jason Phang, Haokun Liu, Phu Mon Htut, Xiaoyi Zhang,
Richard Yuanzhe Pang, Clara Vania, Katharina Kann, and Samuel R. Bowman.
Intermediate-task transfer learning with pretrained language models: When and why
does it work? In Proceedings of the 58th Annual Meeting of the Association for
Computational Linguistics. Association for Computational Linguistics, July 2020.
[RDS+ 15] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma,
Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C
Berg, and Li Fei-Fei. Imagenet large scale visual recognition challenge. IJCV, 2015.
[RPG+ 21] A. Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Rad-
ford, Mark Chen, and Ilya Sutskever. Zero-shot text-to-image generation. ArXiv,
abs/2102.12092, 2021.
[RSR+ 20] Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael
Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. Exploring the limits of transfer learning
with a unified text-to-text transformer. J. Mach. Learn. Res., 21:140:1–140:67, 2020.
[RvdOV19] Ali Razavi, Aaron van den Oord, and Oriol Vinyals. Generating diverse high-fidelity
images with VQ-VAE-2. In Advances in Neural Information Processing Systems,
volume 32. Curran Associates, Inc., 2019.
[SHB16] Rico Sennrich, Barry Haddow, and Alexandra Birch. Neural machine translation of
rare words with subword units. In Proceedings of the 54th Annual Meeting of the
Association for Computational Linguistics (Volume 1: Long Papers), pages 1715–1725,
Berlin, Germany, August 2016. Association for Computational Linguistics.
[TCD+ 20] Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablay-
rolles, and Hervé Jégou. Training data-efficient image transformers & distillation
through attention. preprint arXiv:2012.12877, 2020.
[TCS+ 21] Hugo Touvron, Matthieu Cord, Alexandre Sablayrolles, Gabriel Synnaeve, and Hervé
Jégou. Going deeper with image transformers. arXiv preprint arXiv:2103.17239, 2021.
[TLL19] Trieu H Trinh, Minh-Thang Luong, and Quoc V Le. Selfie: Self-supervised pretraining
for image embedding. arXiv preprint arXiv:1906.02940, 2019.
[vdOVK17] Aaron van den Oord, Oriol Vinyals, and Koray Kavukcuoglu. Neural discrete repre-
sentation learning. In Proceedings of the 31st International Conference on Neural
Information Processing Systems, NIPS’17, page 6309–6318, Red Hook, NY, USA,
2017. Curran Associates Inc.
[VSP+ 17] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N.
Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Isabelle
Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N.
Vishwanathan, and Roman Garnett, editors, Advances in Neural Information Processing
Systems 30: Annual Conference on Neural Information Processing Systems 2017,
December 4-9, 2017, Long Beach, CA, USA, pages 5998–6008, 2017.
12
[WXYL18] Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. Unsupervised feature
learning via non-parametric instance discrimination. In CVPR, 2018.
[XLY+ 21] Zhenda Xie, Yutong Lin, Zhuliang Yao, Zheng Zhang, Qi Dai, Yue Cao, and Han Hu.
Self-supervised learning with swin transformers. arXiv preprint arXiv:2105.04553,
2021.
[XLZ+ 18] Tete Xiao, Yingcheng Liu, Bolei Zhou, Yuning Jiang, and Jian Sun. Unified perceptual
parsing for scene understanding. In ECCV, 2018.
[ZIE16] Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful image colorization. In
ECCV, 2016.
[ZKHB21] Xiaohua Zhai, Alexander Kolesnikov, Neil Houlsby, and Lucas Beyer. Scaling vision
transformers. arXiv preprint arXiv:2106.04560, 2021.
[ZLZ+ 20] Sixiao Zheng, Jiachen Lu, Hengshuang Zhao, Xiatian Zhu, Zekun Luo, Yabiao Wang,
Yanwei Fu, Jianfeng Feng, Tao Xiang, Philip H. S. Torr, and Li Zhang. Rethinking
semantic segmentation from a sequence-to-sequence perspective with transformers.
CoRR, abs/2012.15840, 2020.
[ZZP+ 19] Bolei Zhou, Hang Zhao, Xavier Puig, Tete Xiao, Sanja Fidler, Adela Barriuso, and
Antonio Torralba. Semantic understanding of scenes through the ADE20K dataset. Int.
J. Comput. Vis., 127(3):302–321, 2019.
13
A Architecture Variants of Vision Transformer
We use the standard vision Transformer (ViT) in the experiments for fair comparisons. In addition,
we find that LayerScale [TCS+ 21] and relative position bias [BDW+ 20, RSR+ 20] improve ViTs on
downstream tasks. We employ the same setting as in Section 3.3 for ablation studies, which pretrains
base-size models for 300 epochs on ImageNet-1K.
As shown in Table 5, both LayerScale and relative position bias improve performance on ImageNet
classification and ADE20K semantic segmentation. We denote the improved architecture as BE I T+
and use it for the experiments in Appendix B. We empirically notice that vanilla Transformer is the
most stable when scaling up the model to billions of parameters, so we do not use LayerScale for
extra-large models.
14
Table 6 compares BE I T with previous state-of-the-art supervised pre-training [DBK+ 20, ZKHB21]
on ImageNet fine-tuning. Rather than heavily relying on extremely large-size labeled data (such as
Google’s in-house JFT-300M and JFT-3B), we demonstrate that BE I T pre-training can catch up with
only ImageNet-22k (14M). Specifically, BE I T-L fine-tuned on ImageNet-22K achieves comparable
performance with ViT-L trained on Google JFT-3B. Moreover, BE I T-L obtains 89.5% top-1 accuracy
on ImageNet after intermediate fine-tuning on an in-house 70M dataset. The results indicate that
BE I T pre-training greatly reduces the required labeling efforts and advances the new state of the art
for large-size vision Transformers.
As shown in Table 7, we report the fine-tuning results on the ADE20K semantic segmentation
benchmark. Following Swin [LLC+ 21], we use the same task layer (i.e., UperNet) and evaluate the
models at the resolution 640 × 640. The BE I T-L model obtains state-of-the-art performance on
ADE20K.
15
layer. Similarly, we find that the best layer lies in 9-th layer for BE I T-B, and 14-th layer for BE I T-L.
To be specific, we use AdamW [LH19] to update the linear probe layer for 50 epochs. The learning
rate is 4e-3 with cosine decay. The batch size is 1024. The weight decay is set to 1e-4. We follow
data augmentation used in DINO [CTM+ 21], which uses random resize crops and horizontal flips
augmentation during training and evaluates on central crops.
As shown in Table 9, we evaluate linear probes on ImageNet-1K for self-supervised learning. Overall,
discriminative methods perform better than generative pre-training on linear probing. Linear probes
keep the Transformer parameters fixed and only update the linear layer. So the pre-training of global
aggregation of image-level features is beneficial to linear probing in DINO and MoCo v3, although
full fine-tuning eliminates the gap. Moreover, the results indicate that increasing the model size from
base (86M) to large (304M) significantly improves accuracy for our proposed method. In contrast,
the gap between base- and large-size MoCo v3 is smaller. We also find that BE I T outperforms iGPT
by a large margin even using much fewer parameters.
We train the pre-training tasks of BE I T and DINO [CTM+ 21] together in a multi-task manner. As
shown in Table 10, augmenting masked image modeling with DINO improves semantic segmentation
on ADE20K, and obtains comparable results on ImageNet classification. Moreover, BE I T is more
efficient in terms of pre-training speed, as DINO has two copies of Transformer parameters for
self-distillation and multi-crop augmentation [CMM+ 20]. For the throughput comparisons between
BE I T and BE I T+DINO, we set batch size to the same. Because BE I T is also more memory-efficient,
we can use larger batch size to fully utilize GPU cards, which obtains greater speedup in practice
than the reported numbers.
16
F Image Classification on CIFAR-100
Models CIFAR-100
Training from scratch (i.e., random initialization)
ViT384 [DBK+ 20] 48.5*
Supervised Pre-Training on ImageNet-1K (using labeled data)
ViT384 [DBK+ 20] 87.1
DeiT [TCD+ 20] 90.8
Self-Supervised Pre-Training on ImageNet-1K (without labeled data)
DINO [CTM+ 21] 91.7
MoCo v3 [CXH21] 87.1
BE I T (ours) 90.1
Self-Supervised Pre-Training, and Intermediate Fine-Tuning on ImageNet-1K
BE I T (ours) 91.8
Table 11: Top-1 accuracy of image classification on CIFAR-100. The models are at resolution
224 × 224, except ViT384 uses 384 × 384. The results, unless otherwise indicated, are all obtained
by base-size models. *: result is taken from [CXH21].
17
H Hyperparameters for Image Classification Fine-Tuning
CIFAR-100 ImageNet-1K
Hyperparameters
Base Size Base Size Large Size
Peak learning rate {2e-3, 3e-3, 4e-3, 5e-3}
Fine-tuning epochs 150 100 50
Batch size 512 1024 1024
Warmup epochs 20 20 5
Layer-wise learning rate decay 0.65 0.65 0.75
Adam 1e-8
Adam β (0.9, 0.999)
Minimal learning rate 1e-6
Learning rate schedule Cosine
Repeated Aug 3 3 7
Weight decay 0.3 0.05 0.05
Label smoothing ε 0.1
Stoch. depth 0.1
Dropout 7
Gradient clipping 7
Erasing prob. 7 0.25 0.25
Input resolution 224 × 224
Rand Augment 9/0.5
Mixup prob. 0.8
Cutmix prob. 1.0
Table 13: Hyperparameters for fine-tuning BE I T on ImageNet-1K and CIFAR-100.
18