Write and Paint
Write and Paint
A BSTRACT
Recent advances in vision-language pre-training have pushed the state-of-the-art on
various vision-language tasks, making machines more capable of multi-modal writ-
ing (image-to-text generation) and painting (text-to-image generation). However,
few studies investigate if these two essential capabilities can be learned together and
boost each other, making a versatile and powerful multi-modal foundation model.
In this work, we disclose the potential of symmetric generative vision-language
pre-training in learning to write and paint concurrently, and propose a new unified
modal model, named DAV INCI, trained with prefix language modeling and prefix
image modeling, a simple generative self-supervised objective on image-text pairs.
Thanks to the proposed prefix multi-modal modeling framework, DAV INCI is sim-
ple to train, scalable to huge data, adaptable to both writing and painting tasks, and
also strong on other vision, text, and multi-modal understanding tasks. DAV INCI
achieves competitive performance on a wide range of 27 generation/understanding
tasks and demonstrates the superiority of combining vision/language generative
pre-training. Furthermore, we carefully benchmark the performance of different
vision-language pre-training objectives on different scales of pre-training datasets
on a heterogeneous and broad distribution coverage. Our results demonstrate the
potential of exploiting self-supervision in both language and vision inputs, and
establish new, stronger baselines for future comparisons at different data scales.1
1 I NTRODUCTION
Self-supervised language model pre-training (Peters et al., 2018; Radford et al., 2018; Devlin et al.,
2019; Liu et al., 2019; Lewis et al., 2020; Raffel et al., 2020; Brown et al., 2020; Fu et al., 2022; Zhou
et al., 2021b; Diao et al., 2020; 2021; Zhou et al., 2021a; Xu et al., 2020; Zhou et al., 2020; 2022a;
Pan et al., 2022; Diao et al., 2023) has reshaped the landscape of modern natural language processing
(NLP) research, pushing the state-of-the-art of a wide range of NLP tasks. Recently, this success
has been transferred to the multi-modal context and resulted in a number of vision-language pre-
trained models (VLMs) (Lu et al., 2019; Tan & Bansal, 2019a), achieving state-of-the-art results on
various vision-language tasks. Most existing VLMs are BERT-like Transformer (Vaswani et al., 2017)
encoders pre-trained with a combination of different vision-language pre-training (VLP) objectives:
masked multi-modal modeling (Lu et al., 2019; Tan & Bansal, 2019b; Chen et al., 2020; Li et al.,
2020), multi-modal alignment prediction (Lu et al., 2019; Tan & Bansal, 2019b; Chen et al., 2020; Li
et al., 2020), region of interest feature regression (Tan & Bansal, 2019b), image-text matching (Li
et al., 2021; Zeng et al., 2021), to name a few. However, the roadmap towards large language models
reveals a transition pattern from encoder-only models like BERT (Devlin et al., 2019) / RoBERTa (Liu
et al., 2019) to sequence-to-sequence models like T5 (Raffel et al., 2020) / BART (Lewis et al., 2020)
and autoregressive models like GPT-3 (Brown et al., 2020) / PaLM (Chowdhery et al., 2022) to tackle
∗
Work done during the internship at ByteDance AI Lab.
†
Corresponding author
1
The code and pre-trained models are available at https://github.com/shizhediao/DaVinci.
1
Published as a conference paper at ICLR 2023
more tasks in a unified way, and from complicated objectives like masked language modeling / next
sentence prediction / replace token detection to a simple language modeling objective to improve the
scalability of pre-training. This suggests that the generative pre-training paradigm with simple targets
shows great potential for pre-training more scalable and general VLMs.
To this end, several recent studies (Cho et al., 2021; Zhang et al., 2021a; Wang et al., 2021b; 2022)
investigated sequence-to-sequence (seq2seq) vision-language pre-training and achieved state-of-
the-art results on a range of vision-language understanding and generation tasks. For example,
VL-T5 (Cho et al., 2021), OFA (Wang et al., 2022) and PaLI (Chen et al., 2022) formulate various
vision-and-language problems into seq2seq tasks and pre-train a seq2seq VLM by multi-tasking
on these tasks. In addition, ERNIE-ViLG (Zhang et al., 2021a) and SimVLM (Wang et al., 2021b)
pre-train seq2seq VLMs with a simple language modeling or prefix language modeling objective on
a large number of image-caption pairs. While achieving promising results, these objectives are not
versatile enough, resulting in VLMs that are only capable of a subset of tasks in image-text modalities.
On the other hand, the recent success of generative language pre-training (Brown et al., 2020) and
generative vision pre-training (He et al., 2022; Bao et al., 2021) motivates us to explore generative
vision-language pre-training to learn more versatile and scalable vision-language models.
In this work, we introduce prefix multi-modal modeling, a unified generative pre-training framework
that extends prefix language modeling to the multi-modal context and learns a multi-modal foundation
model by learning to write and paint simultaneously. As illustrated in Figure 1, given an image-
caption pair, we split the image and caption into two parts denoted as prefix and suffix. To make prefix
image modeling compatible with the seq2seq formulation of conventional prefix language modeling,
we follow DALLE (Ramesh et al., 2021) and convert images into discrete sequences of image
tokens (van den Oord et al., 2017). We then train the model to generate the suffix in one modality
based on the prefix in the same modality and the complete input in the other modality. In this way,
prefix multi-modal modeling can fully exploit self-supervision from large-scale image-caption pairs
by learning to write and paint simultaneously. We pre-train DAV INCI 2 , a vision-language foundation
model, with the proposed prefix multi-modal modeling framework on large-scale image-text pairs.
DAV INCI is the first self-supervised vision-language foundation model that is versatile for all kinds of
tasks in vision-and-language modalities, including image-to-text generation, text-to-image generation,
vision-language understanding, and single-modal language / vision tasks. DAV INCI consistently
outperforms FLAVA (Singh et al., 2021), an existing vision-language foundation model, on both
language, vision, and multi-modal tasks, and performs competitively with state-of-the-art models
across a wide range of tasks and modalities. Moreover, DAV INCI also shows strong few-shot and
zero-shot image/text generation capability.
In addition, most existing VLMs are pre-trained with mixed pre-training objectives and different data
sources varying in size, making it difficult to disentangle the impact of pre-training objectives and data
sources on the downstream tasks. To this end, we conduct a systematic analysis of the performance
of generative vision-language pre-training by carefully ablating different pre-training objectives, such
as prefix language / image modeling, and the amount of pre-training data with different qualities,
revealing the impact of different objectives and data sources to facilitating future research.
To summarize, our contribution is three-fold: (1) We introduce prefix multi-modal modeling, a simple
unified generative vision-language pre-training framework that is scalable for large-scale pre-training
and versatile for image-to-text generation, text-to-image generation and various multi-modal / single-
modal understanding tasks. (2) We pre-train DAV INCI, a vision-language foundation model, with the
proposed approach, demonstrating competitive performance on a wide range of 27 downstream tasks
and the superiority of combining vision/language generative pre-training. (3) We conduct an analysis
about the impact of different pre-training data sources and pre-training objectives on the performance
of seq2seq VLMs.
2 R ELATED W ORK
Inspired by the success of language model pre-training, several studies investigated vision-language
pre-training on large-scale image-caption pairs. ViLBERT (Lu et al., 2019) and LXMERT (Tan
& Bansal, 2019b) first propose to extract visual object features with an external object detection
model like Fast-RCNN (Girshick, 2015), feed the image features together with texts into Transformer
2
Named after the Italian polymath Leonardo da Vinci, who displayed infinite grace in everything. We noticed
that this name is used in GPT-3 versioning. However, we think there is no conflict because it is only a suffix for
a specific checkpoint of the GPT-3 family.
2
Published as a conference paper at ICLR 2023
t1
ℒ!"# ℒ!$#
The Last Supper of The Last Supper of
Jesus with the Twelve Embed. t2
Apostles, painting by v1 h1
…
Leonardo da Vinci.
tn
v2 h2
… …
✚
Cross-
✚
vm Cross-
or + t1
modal modal
Encoder Decoder
t2
24
go r
S
un a
…
1
2
3
S
zz
g
de
in
O
EO
10
B
v1 … …
Conv. v2 tn hl
…
vm Image
Decoder
Encoder input Decoder input
Figure 1: Illustration of the overall architecture and pre-training procedures of DAV INCI, a Transformer-based
sequence-to-sequence model. Given an image-text pair, DAV INCI first splits either the word sequence or image
token sequence into prefix and suffix. It then concatenates the prefix with the complete sequence in the other
modality as input. DAV INCI is trained to recover the suffix with maximum likelihood estimation.
models, and train the model to align vision and language representations with masked multi-modal
modeling and multi-modal alignment prediction objectives. Many following works (Li et al., 2020;
Zhang et al., 2021b; Chen et al., 2020; Li et al., 2022a; 2021; Zeng et al., 2021; Wang et al., 2021a)
propose several new objectives to improve object detection based VLP and explored using vision
Transformer (Dosovitskiy et al., 2021; Touvron et al., 2021) as visual feature extractor.
More recently, FLAVA (Singh et al., 2021), a new vision-language foundation model, is pre-trained
with a masked multi-modal modeling objective. Performing competitively on language, vision, and
vision-language understanding tasks, FLAVA is designed for understanding tasks without text and
image generation abilities.
While achieving promising results on multi-modal understanding tasks, most VLMs are based on
encoder-only architectures with bidirectional attention, making them non-trivial to adapt to multi-
modal generation tasks such as image captioning and text-to-image generation. Inspired by the
success of seq2seq pre-trained language models such as T5 (Raffel et al., 2020) and BART (Lewis
et al., 2020), VL-T5 (Cho et al., 2021) and OFA (Wang et al., 2022) propose to formulate both
vision-language pre-training objectives and various downstream vision-language tasks as seq2seq
tasks and pre-train a seq2seq VLM by multi-tasking on these tasks. However, the scalability and
the zero-shot transfer capability of this approach are limited by the availability of large-scale and
diverse vision-language tasks. To this end, SimVLM (Wang et al., 2021b), the most related work to
our approach, instead pre-trains a seq2seq VLM with a simple prefix language modeling objective
on text generation. It easily scales to very large and potentially noisy pre-training data and achieves
competitive results. However, SimVLM only exploits language self-supervision, and thus it does not
perform well on image understanding tasks and is unable to tackle image generation tasks. Another
recent study is CM3 (Aghajanyan et al., 2022), which proposes a causal masked multi-modal model
learned from large web data and differs from our work in pre-training objectives and target tasks.
As for the text-to-image generation task, Ramesh et al. (2021); Ding et al. (2021); Yu et al. (2022)
achieved promising performance by learning an auto-regressive target with Transformer and VQ-VAE
/ VQ-GAN tokenizer. Most recently, Ramesh et al. (2022); Saharia et al. (2022) advanced the image
generation capability by using diffusion models and high-quality text embeddings (e.g., CLIP, T5).
Therefore, it is natural to explore boosting image generation via stronger multi-modal understanding.
Previous studies are good at either image-to-text or text-to-image generation, but few studies investi-
gate whether these two important capabilities can be learned together and boost each other. In this
paper, we explore making a versatile and powerful multi-modal foundation model that is good at
text-to-image generation, image-to-text generation, and multi-modal understanding tasks.
3 DAV INCI
Given the superior performance of auto-regressive language models (LM) (Brown et al., 2020;
Chowdhery et al., 2022; Rae et al., 2021) on zero-shot and few-shot transfer abilities, we decided to
3
Published as a conference paper at ICLR 2023
adopt a decoder optimized by language modeling loss to retain the generalization capabilities, and an
encoder to represent the prefix input. Unlike using a causal mask in the decoder, the encoder employs
fully-visible attention for the prefix input. This architecture resembles prefix language modeling,
which shows effectiveness in a wide range of language tasks (Dong et al., 2019; Raffel et al., 2020)
and enables zero-shot generalization abilities. Contrary to the previous multi-stage approaches (Wang
et al., 2021a; Singh et al., 2021), our model is trained from scratch in an end-to-end manner thanks
to the model’s simplicity. In this section, we introduce the proposed prefix multi-modal modeling
framework and the DAV INCI model. The overall architecture of DAV INCI is depicted in Figure 1.
We first explain our model architecture in detail in §3.1 and then introduce pre-training objectives
and procedures in §3.2.
where I and S are images and captions from the pre-training corpus D.
4
Published as a conference paper at ICLR 2023
Because of the lack of textual information, recovering the suffix caption requires the model to
understand both the image and prefix caption. The full image is rich in semantic information that
would help language modeling. The prefix length is randomly decided during training, and especially
when prefix caption is none, this task will degenerate into “image captioning” task, which forces the
model to generate a caption with the input image.
X
L0PLM = − log p(Ytext |Ximage ) (2)
(I,S)∈D
Prefix Image Modeling (PIM) The core idea of prefix image modeling is “given a full caption and
a corrupted image (we call it prefix image hereafter), recover the masked visual tokens”. Given an
input image, we first randomly mask some continuous image patches at the end (we call it suffix
image hereafter). The prefix image and full caption will be fed into the model and try to recover the
original visual tokens obtained from the image tokenizer by optimizing the cross-entropy loss.
X
LPIM = − log p(Yimage |Xtext , X̃image ) (3)
(I,S)∈D
Similar to PLM, when prefix image is none, this task will degenerate into “text-to-image generation”
task, forcing the model to generate an image with the input caption:
X
L0PIM = − log p(Yimage |Xtext ) (4)
(I,S)∈D
Unified Learning Objective Our model is learned by optimizing the combination of PLM and PIM.
L = LPLM + LPIM (5)
4 E XPERIMENTS
4.1 P RE - TRAINING DATASETS
Since existing studies pre-trained their models on different corpora, making the fair comparison
difficult. Considering results only on state-of-the-art performance would underestimate the potential
of this line of research. Therefore, we propose several practical settings including small-scale and
large-scale, and then conduct detailed comparisons on them in Section 5.1. More details about the
datasets are shown in Appendix A.3.
Table 1: Statistics of the pre-training datasets. #Total denotes the total number of image-text pairs.
5
Published as a conference paper at ICLR 2023
BERT RoBERTa ViT MLM MIM FLAVA CLIP SimVLM DAV INCI SimVLM DAV INCI
1 2 3 4 5 6 7 8
Task Eval. 16GB 160GB 13.2M 70M 70M 70M 70M 46.4M 46.4M 647.7M 647.7M
MNLI FT 84.20 87.60 – 73.23 – 80.33 32.85 82.13 82.25 83.27 83.13
CoLA FT 54.60 63.60 – 39.55 – 50.65 11.02 52.47 52.10 54.22 54.75
MRPC FT 84.75 90.20 – 73.24 – 84.16 68.74 82.70 83.14 84.26 84.54
QQP FT 89.00 91.90 – 86.68 – 88.74 59.17 88.39 88.15 89.05 88.92
SST-2 FT 92.50 94.80 – 87.96 – 90.94 83.49 90.65 90.48 91.12 91.37
QNLI FT 91.00 92.80 – 82.32 – 87.31 49.46 87.55 87.21 88.28 87.90
RTE FT 62.50 78.70 – 50.54 – 57.76 53.07 59.80 60.72 63.34 64.22
STS-B FT 88.20 91.20 – 78.89 – 85.67 13.70 86.62 86.27 87.24 87.05
NLP Avg. 80.84 86.35 – 71.55 – 78.19 46.44 78.79 78.79 80.10 80.23
Vision Avg. – – 77.68 – 49.84 78.12 74.56 74.14 78.56 77.34 79.80
Table 2: Experimental results on vision, language and multi-modal downstream tasks. @B4, @C denote
BLEU@4, CIDEr, respectively. I2T and T2I denote image-to-text and text-to-image tasks. Multi-modal Avg. is
the average score of all multi-modal tasks. FT: fine-tuning, LE: linear evaluation, FS: few-shot, ZS: zero-shot.
Under few-shot setting, we fine-tune a pre-trained model for 3 epochs on 1% training data. Results for BERT are
obtained from Iki & Aizawa (2021). Results for RoBERTa are from its corresponding paper (Liu et al., 2019)
and they use the mid-training (Phang et al., 2018) on MNLI for RTE, MRPC and STS-B while other models
(e.g., BERT, SimVLM, DAV INCI) do not apply this trick. Results for ViT are from ViT-Base/16 model (Radford
et al., 2021). We list the reported performance of text-only and image-only models in grey for reference.
a warm start provides a reliable visual representation and helps the convergence. All pre-training
experiments are conducted on 32GB NVIDIA V100 GPUs. The model trained on the largest data
takes around 10 days on 1024 V100 GPUs. We adopt dynamic masking in our experiments, where
the masking ratio is randomly sampled from a uniform distribution U(0, 1). More details of the
fine-tuning, network architectures, and hyper-parameters setups are given in Appendix A.1.
6
Published as a conference paper at ICLR 2023
Table 3: Comparison with state-of-the-art vision-language models on vision, language, and multi-modal
downstream tasks. All results are from base-size models. LE and FT denote linear evaluation and fine-tuning
performance, respectively. Image2Text results are reported without CIDEr optimization. † are our reproduced
models. ∗ are the results after fine-tuning. SimVLM (1.8B) and OFA are pre-trained with much larger corpus
or human-labeled data of many downstream tasks, and thus they are not comparable and are labeled in gray.
Florence (Yuan et al., 2021) is pre-trained with much larger model size (Florence-CoSwin-H, 637M) and more
pre-training data (900M), so the numbers are in grey. bold denotes the best across unified models.
foundation models. We also include comparisons with some baseline models (e.g., MIM, MLM,
CLIP). There are several observations. First, DAV INCI (column 8) outperforms FLAVA (column 3)
and SimVLM (column 7) across almost all tasks, providing a new and stronger unified foundation
model. Compared with FLAVA, DAV INCI improves an average of 2.04%, 1.68% on language and
vision tasks, respectively. Compared with SimVLM, DAV INCI achieves comparable results on
language tasks (+0.13%) while performing much better on vision tasks (+2.46%). To make a fair
comparison in terms of similar data size, we compare FLAVA (70M data, column 3) with DAV INCI
(46.4M data, column 6). It is observed that DAV INCI still outperforms FLAVA even with much less
data. Considering the multi-modal tasks, DAV INCI consistently outperforms FLAVA and SimVLM
on VQA and VE. Note that FLAVA is incapable of generation and SimVLM cannot generate images;
only DAV INCI is competent to all tasks and demonstrates a stronger capability of unifying vision and
language tasks.
Zero-shot and Few-shot Transfer One of the critical benefits of generative pre-trained vision-
language models is the good generalization ability on zero-shot and few-shot tasks. For zero-shot
transfer, two out-of-domain distribution datasets are considered (NoCaps and VLUE), with results
shown in Table 2. First, DAV INCI outperforms SimVLM on both zero-shot and few-shot settings,
demonstrating its better transfer capabilities. It also shows the effectiveness and robustness of the
synergy of our proposed language supervision and image supervision. Second, it is observed that
the performance improvement is bigger on 647.7M data (column 7 v.s. column 8) than 46.4M data
(column 5 v.s. column 6). This shows DAV INCI generalizes well with the increase of large-scale
data. We even observe some performance drops on small data (46.4M) but excellent performance
improvements on large data (647.7M). It is consistent with the recent observation that zero-shot
ability could only be triggered with large pre-training data (Wei et al., 2022) and scaling to large data
and keeping simple training objectives benefit generalization performance (Wang et al., 2021b).
Comparison with state-of-the-art vision-language models In addition to unified vision-language
foundation models, we compare DAV INCI with state-of-the-art vision-language models as well. The
results are shown in Table 2. DAV INCI demonstrates its superiority in vision understanding and
text-to-image generation. Compared with current popular auto-regressive image generation models
like DALLE and CogView, our model achieves comparable IS and better FID scores with significantly
fewer model parameters than DALLE and CogView. Note that the original DALLE is implemented
based on VQVAE, so here, we compare our model with reproduced VQGAN-based DALLE with
7
Published as a conference paper at ICLR 2023
Settings Pre-training Data #Image #Caption Models COCO Captions VQA SNLI-VE NLVR2
ID SWD ORD VD LWD B@4 / C Acc Acc Acc
SimVLM 35.2 / 115.06 68.89 76.10 71.21
1 X 0.2M 1.3M
DAV INCI 35.8 / 117.30 69.25 76.22 72.55
SimVLM 37.0 / 122.63 71.54 78.36 75.50
2 X X 15.1M 16.2M
DAV INCI 37.4 / 123.11 71.88 78.62 77.46
SimVLM 38.2 / 123.85 69.57 76.65 70.50
3 X X 2.7M 18.3M
DAV INCI 38.0 / 124.20 70.02 76.92 72.01
SimVLM 36.2 / 119.73 70.53 76.90 73.25
4 X X 13.4M 14.5M
DAV INCI 36.6 / 121.27 71.23 77.40 74.62
SimVLM 38.5 / 128.12 71.84 78.81 76.75
5 X X X X 30.5M 46.4M
DAV INCI 38.6 / 128.73 73.53 79.24 77.55
SimVLM 37.3 / 123.81 73.73 78.79 77.69
6 X 601.3M 601.3M
DAV INCI 37.6 / 124.42 73.95 79.29 78.54
SimVLM 37.9 / 125.50 74.64 79.05 77.68
7 X X 601.5M 602.6M
DAV INCI 38.1 / 125.91 74.91 79.22 78.12
SimVLM 38.5 / 128.25 75.04 79.32 78.82
8 X X X X X 631.8M 647.7M
DAV INCI 39.1 / 130.21 76.32 80.04 80.03
Table 4: Evaluation on downstream tasks using COCO Captions, VQA, SNLI-VE, and NLVR2. #Image
and #Caption denote the numbers of images and image-text pairs that are used in the pre-training.
similar model sizes, and find DAV INCI still achieves a significant improvement over it. Generated
images are presented in Appendix A.11 for further qualitative comparison.
On multi-modal tasks such as VQA, DAV INCI not only outperforms unified models (e.g., SimVLM
(640M)) and other encoder-decoder multi-modal models (e.g., E2E-VLP, VL-T5), but also achieves
competitive performance with many conventional encoder-only multi-model models (e.g., VinVL,
ALBEF, VLMO). Note that SimVLM (1.8B) and OFA are not directly comparable because SimVLM
uses 1.8B in-house image-text pairs, and OFA uses human-labeled data of many downstream tasks
during pre-training. Even though, we still report their results for reference and observe a better
performance on ImageNet fine-tuning and text-to-image generation than OFA.
The advantages of image generation over DALLE / CogView, the superiority of image-to-text over
SimVLM, and the competitive performance with conventional multi-modal models demonstrate the
synergistic effect of our proposed PLM (language supervision) and PIM (image supervision).
5 A NALYSIS
5.1 I MPACT OF P RE - TRAINING DATASETS
In this section, we disclose the impact of various multi-modal data sources for VLMs. We choose
SimVLM and DAV INCI as our baseline models for their competitive performance, the capability
of training from scratch, and the scalability of extending to the noisy large-scale corpus. We use
the same text corpus, C4, for all the variations. The results are shown in Table 4. In general, the
performance is increased along with the data size, and DAV INCI consistently outperforms SimVLM
on almost all the data settings and all the downstream tasks. Both object-region data and vision data
are clearly helpful in vision language pre-training (refer to settings 3 and 4). We surprisingly observe
that models pre-trained on object-region data with much fewer images performs even better than
models pre-trained with small-scale web data on the COCO Caption task (refer to settings 2 and 3).
Although large-scale web data is usually noisier than small datasets (e.g., ID, ORD, VD, and SWD),
it is powerful for multi-modal pre-training (refer to settings 5 and 8). We believe our analysis has
broader impacts on the research of VLMs in the community. First, this enables fair comparisons
for pre-trained models in the same data settings. Second, one can focus on the model designs at
part or all of the data settings according to available computation resources. Third, we reveal that
object-region and vision data, normally overlooked in VLM pre-training, also play a significant role.
5.2 A BLATION S TUDY
To verify the contributions of different modules in our framework, we ablate them and evaluate
DAV INCI on five kinds of downstream tasks: language understanding (MNLI, SST-2), vision under-
standing (ImageNet, Food101, CIFAR10), multi-modal understanding (VQAv2, SNLI-VE, NLVR2),
image-to-text generation (COCO Captions), and text-to-image generation. Experiments are conducted
with the same model architecture on in-domain data (ID). The results are shown in Table 5.
Effects of Objectives First, all three objectives (PLM, PIM, and Text2Text) bring improvement
and the combination confirms a synergistic effect. Second, it is observed that without PLM, the
performance decreases significantly on multi-modal understanding and image-to-text generation,
8
Published as a conference paper at ICLR 2023
COCO VQA SNLI-VE NLVR2 ImageNet Food101 CIFAR10 MNLI SST-2 T2I
Method
B@4 / C Acc Acc Acc Acc Acc Acc Acc Acc IS / FID
DAV INCI 35.8 / 117.30 69.25 76.22 72.55 48.88 75.32 73.82 81.76 90.25 12.35 / 53.14
– PLM 33.6 / 111.17 65.15 73.91 53.28 48.05 74.17 72.98 81.42 89.97 10.26 / 59.64
– PIM 34.3 / 116.58 68.89 75.79 69.78 45.54 71.18 70.11 81.94 90.53 –∗
– Text2Text 34.1 / 115.21 68.14 75.38 70.34 48.67 74.26 73.23 76.48 88.14 12.07 / 54.77
PL=0 35.4 / 117.00 66.90 75.52 71.05 48.45 68.18 73.73 78.69 89.00 11.76 / 55.38
PL=15% 35.7 / 116.53 69.16 75.09 70.44 41.58 52.15 68.55 79.02 89.46 –∗
PL=50% 35.1 / 115.53 68.55 74.54 56.92 37.69 49.16 70.15 78.59 89.69 –∗
MIM 34.7 / 113.4 68.18 75.34 69.66 48.46 56.95 72.79 81.72 89.84 9.50 / 74.13
In-painting 34.5 / 112.5 67.46 75.41 68.66 47.50 54.38 71.20 81.55 89.84 9.97 / 68.15
Token Projection 17.7 / 49.2 52.13 71.11 52.01 15.11 25.62 61.01 82.01 90.25 11.89 / 60.96
Patch Projection 25.7 / 79.5 57.69 71.92 57.45 36.23 44.31 69.40 81.73 90.05 11.41 / 61.87
Table 5: Ablation study on COCO Captions, VQA, SNLI-VE, NLVR2, ImageNet, Food101, CIFAR10,
MNLI, SST-2, and text-to-image (T2I) generation. “–” denotes removing the corresponding objective. PL
denotes the prefix length under fixed masking ratio settings. Because the linear probe requires a pre-trained
model to be frozen, “No Pre-training” results on ImageNet, Food101, and CIFAR10 are not reported and labeled
by ∗ . For T2I, we report the zero-shot results. Note that the following four variants cannot perform zero-shot
text-to-image generation (labeled by ∗ ): (1) No Pre-training, (2) DAV INCI – PIM, (3) PL=15%, and (4) PL=50%.
indicating the importance of language supervision. Third, PIM brings more gains than PLM and
text2text on vision understanding, which is expected because it enhances the vision encoding ability
with image supervision. In addition, the text2text objective is important to text understanding. Last,
on the text-to-image generation task, it is observed that PLM is also helpful, confirming the synergistic
effect of PIM and PLM again. Intuitively, PIM and PLM can help each other learn the alignments of
visual and textual features, which will benefit both image generation and other multi-modal tasks.
Effects of Masking Ratios Our model adopts dynamic masking ratios as described in Section 3.2.
We also conduct experiments with static masking ratios with the prefix length fixed to 0, 15%, and
50%. The comparison between dynamic masking ratios and static masking ratios (PL=0, 15%, and
50%) reveals that dynamic masking is better. We attribute this improvement to the smoothing effects
of dynamic masking ratios. We also find that the standard language model (PL=0) performs worse on
VQA, Food101, and text-to-image generation, which is consistent with the observation in SimVLM.
In our experiments, the masking ratio is sampled from a uniform distribution U(0, 1).
Effects of Masking Strategies Here we also compared three different masking strategies: 1) masked
image modeling (randomly masking some patches), 2) in-painting (randomly masking some continu-
ous spans in the middle of the image), and 3) suffix-painting (ours). The results are shown in Table 5.
Both masked image modeling and in-painting are effective and competitive. It is observed that
suffix-painting is better than masked image modeling and in-painting across all tasks, demonstrating
that suffix-painting works well.
Effects of Image Feature Extraction There are several different ways to extract image features.
We compare three different image representation methods: 1) token projection (projecting the prefix
tokens to the hidden dimension of the backbone network on the token-level), 2) patch projection
(similar to ViT embedding, we split an image into fixed-size patches, embed each of them by a
trainable linear projection on the pixel-level), and 3) ResNet feature extraction (ours). From the
results in Table 5, we observed that ResNet feature extraction outperforms token projection and patch
projection by a large margin. Therefore, we decided to adopt ResNet to extract image features.
We provide more details and discussions about the effects of compute (A.5), masking strategies (A.6),
image feature extraction methods (A.7), and scaling effects of data size ( A.8) in the Appendix.
9
Published as a conference paper at ICLR 2023
ACKNOWLEDGMENTS
We thank the anonymous reviewers for their valuable suggestions. We would like to acknowledge
Yan Zeng, Wenguan Huang, and Zhi Zhang at ByteDance, and Zhiling Zhang at Shanghai Jiao Tong
University for their generous assistance in data collection and helpful discussions. We also wish to
thank Hang Li at ByteDance, and Tong Zhang at HKUST for inspiring feedback, valuable comments,
and great support to this work.
R EFERENCES
Armen Aghajanyan, Bernie Huang, Candace Ross, Vladimir Karpukhin, Hu Xu, Naman Goyal,
Dmytro Okhonko, Mandar Joshi, Gargi Ghosh, Mike Lewis, et al. Cm3: A causal masked
multimodal model of the internet. arXiv preprint arXiv:2201.07520, 2022.
Eneko Agirre, Lluís Màrquez, and Richard Wicentowski (eds.). Proceedings of the Fourth Inter-
national Workshop on Semantic Evaluations (SemEval-2007), Prague, Czech Republic, 2007.
Association for Computational Linguistics. URL https://aclanthology.org/S07-1000.
Harsh Agrawal, Peter Anderson, Karan Desai, Yufei Wang, Xinlei Chen, Rishabh Jain, Mark Johnson,
Dhruv Batra, Devi Parikh, and Stefan Lee. nocaps: novel object captioning at scale. In 2019
IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South),
October 27 - November 2, 2019, pp. 8947–8956. IEEE, 2019. doi: 10.1109/ICCV.2019.00904.
URL https://doi.org/10.1109/ICCV.2019.00904.
Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. ArXiv preprint,
abs/1607.06450, 2016. URL https://arxiv.org/abs/1607.06450.
Hangbo Bao, Li Dong, and Furu Wei. BEiT: Bert pre-training of image transformers. arXiv preprint,
2021.
Luisa Bentivogli, Peter Clark, Ido Dagan, and Danilo Giampiccolo. The fifth pascal recognizing
textual entailment challenge. In TAC, 2009.
Lukas Bossard, Matthieu Guillaumin, and Luc Van Gool. Food-101–mining discriminative compo-
nents with random forests. In European conference on computer vision, pp. 446–461. Springer,
2014.
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhari-
wal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agar-
wal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh,
Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler,
Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCan-
dlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language models are few-shot
learners. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan,
and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual
Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-
12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/
1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html.
Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and
Armand Joulin. Emerging properties in self-supervised vision transformers. In Proceedings of the
IEEE/CVF International Conference on Computer Vision, pp. 9650–9660, 2021.
Xi Chen, Xiao Wang, Soravit Changpinyo, AJ Piergiovanni, Piotr Padlewski, Daniel Salz, Sebastian
Goodman, Adam Grycner, Basil Mustafa, Lucas Beyer, et al. Pali: A jointly-scaled multilingual
language-image model. arXiv preprint arXiv:2209.06794, 2022.
Xinlei Chen, Hao Fang, Tsung-Yi Lin, Ramakrishna Vedantam, Saurabh Gupta, Piotr Dollár, and
C Lawrence Zitnick. Microsoft COCO Captions: Data collection and evaluation server. arXiv
preprint, 2015.
10
Published as a conference paper at ICLR 2023
Yen-Chun Chen, Linjie Li, Licheng Yu, Ahmed El Kholy, Faisal Ahmed, Zhe Gan, Yu Cheng, and
Jingjing Liu. UNITER: Universal image-text representation learning. In European Conference on
Computer Vision (ECCV), 2020.
Jaemin Cho, Jie Lei, Hao Tan, and Mohit Bansal. Unifying vision-and-language tasks via text
generation. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International
Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139
of Proceedings of Machine Learning Research, pp. 1931–1942. PMLR, 2021. URL http:
//proceedings.mlr.press/v139/cho21a.html.
Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam
Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, et al. Palm:
Scaling language modeling with pathways. ArXiv preprint, abs/2204.02311, 2022. URL https:
//arxiv.org/abs/2204.02311.
Mircea Cimpoi, Subhransu Maji, Iasonas Kokkinos, Sammy Mohamed, and Andrea Vedaldi. Describ-
ing textures in the wild. In 2014 IEEE Conference on Computer Vision and Pattern Recognition,
CVPR 2014, Columbus, OH, USA, June 23-28, 2014, pp. 3606–3613. IEEE Computer Society,
2014. doi: 10.1109/CVPR.2014.461. URL https://doi.org/10.1109/CVPR.2014.461.
Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised
feature learning. In Proceedings of the fourteenth international conference on artificial intelligence
and statistics, pp. 215–223. JMLR Workshop and Conference Proceedings, 2011.
Ido Dagan, Oren Glickman, and Bernardo Magnini. The pascal recognising textual entailment
challenge. In Machine Learning Challenges Workshop, pp. 177–190. Springer, 2005.
Zihang Dai, Hanxiao Liu, Quoc V Le, and Mingxing Tan. Coatnet: Marrying convolution and
attention for all data sizes. Advances in Neural Information Processing Systems, 34:3965–3977,
2021.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of
deep bidirectional transformers for language understanding. In Proceedings of the 2019 Confer-
ence of the North American Chapter of the Association for Computational Linguistics: Human
Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186, Minneapolis, Min-
nesota, 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1423. URL
https://aclanthology.org/N19-1423.
Shizhe Diao, Jiaxin Bai, Yan Song, Tong Zhang, and Yonggang Wang. Zen: Pre-training chinese text
encoder enhanced by n-gram representations. In Findings of the Association for Computational
Linguistics: EMNLP 2020, pp. 4729–4740, 2020.
Shizhe Diao, Ruijia Xu, Hongjin Su, Yilei Jiang, Yan Song, and Tong Zhang. Taming pre-trained
language models with n-gram representations for low-resource domain adaptation. In Proceed-
ings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th
International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp.
3336–3349, 2021.
Shizhe Diao, Zhichao Huang, Ruijia Xu, Xuechun Li, Yong Lin, and Tong Zhang. Black-box prompt
learning for pre-trained language models. Transactions on Machine Learning Research, 2023.
URL https://openreview.net/forum?id=IvsGP7xRvm.
Ming Ding, Zhuoyi Yang, Wenyi Hong, Wendi Zheng, Chang Zhou, Da Yin, Junyang Lin, Xu Zou,
Zhou Shao, Hongxia Yang, et al. Cogview: Mastering text-to-image generation via transformers.
Advances in Neural Information Processing Systems, 34, 2021.
William B. Dolan and Chris Brockett. Automatically constructing a corpus of sentential paraphrases.
In Proceedings of the Third International Workshop on Paraphrasing (IWP2005), 2005. URL
https://aclanthology.org/I05-5002.
Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou,
and Hsiao-Wuen Hon. Unified language model pre-training for natural language understanding
and generation. In NeurIPS, pp. 13042–13054, 2019.
11
Published as a conference paper at ICLR 2023
Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas
Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit,
and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In
9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May
3-7, 2021. OpenReview.net, 2021. URL https://openreview.net/forum?id=YicbFdNTTy.
Patrick Esser, Robin Rombach, and Bjorn Ommer. Taming transformers for high-resolution im-
age synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern
Recognition, pp. 12873–12883, 2021.
Zhiyi Fu, Wangchunshu Zhou, Jingjing Xu, Hao Zhou, and Lei Li. Contextual representation learning
beyond masked language modeling. In ACL (1), pp. 2701–2714. Association for Computational
Linguistics, 2022.
Danilo Giampiccolo, Bernardo Magnini, Ido Dagan, and Bill Dolan. The third PASCAL recognizing
textual entailment challenge. In Proceedings of the ACL-PASCAL Workshop on Textual Entailment
and Paraphrasing, pp. 1–9, Prague, 2007. Association for Computational Linguistics. URL
https://aclanthology.org/W07-1401.
Ross B. Girshick. Fast R-CNN. In 2015 IEEE International Conference on Computer Vision, ICCV
2015, Santiago, Chile, December 7-13, 2015, pp. 1440–1448. IEEE Computer Society, 2015. doi:
10.1109/ICCV.2015.169. URL https://doi.org/10.1109/ICCV.2015.169.
Priya Goyal, Piotr Dollár, Ross Girshick, Pieter Noordhuis, Lukasz Wesolowski, Aapo Kyrola,
Andrew Tulloch, Yangqing Jia, and Kaiming He. Accurate, large minibatch sgd: Training imagenet
in 1 hour. ArXiv preprint, abs/1706.02677, 2017a. URL https://arxiv.org/abs/1706.
02677.
Yash Goyal, Tejas Khot, Douglas Summers-Stay, Dhruv Batra, and Devi Parikh. Making the V in
VQA matter: Elevating the role of image understanding in visual question answering. In 2017
IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA,
July 21-26, 2017, pp. 6325–6334. IEEE Computer Society, 2017b. doi: 10.1109/CVPR.2017.670.
URL https://doi.org/10.1109/CVPR.2017.670.
R Bar Haim, Ido Dagan, Bill Dolan, Lisa Ferro, Danilo Giampiccolo, Bernardo Magnini, and Idan
Szpektor. The second pascal recognising textual entailment challenge. In Proceedings of the
Second PASCAL Challenges Workshop on Recognising Textual Entailment, volume 7, 2006.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image
recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR
2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 770–778. IEEE Computer Society, 2016. doi:
10.1109/CVPR.2016.90. URL https://doi.org/10.1109/CVPR.2016.90.
Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross B. Girshick. Momentum contrast for
unsupervised visual representation learning. In 2020 IEEE/CVF Conference on Computer Vision
and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pp. 9726–9735. IEEE,
2020. doi: 10.1109/CVPR42600.2020.00975. URL https://doi.org/10.1109/CVPR42600.
2020.00975.
Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked
autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on
Computer Vision and Pattern Recognition, pp. 16000–16009, 2022.
Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans
trained by a two time-scale update rule converge to a local nash equilibrium. In Isabelle Guyon,
Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan,
and Roman Garnett (eds.), Advances in Neural Information Processing Systems 30: Annual
Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach,
CA, USA, pp. 6626–6637, 2017. URL https://proceedings.neurips.cc/paper/2017/
hash/8a1d694707eb0fefe65871369074926d-Abstract.html.
Kashmir Hill and Jeremy White. Designed to deceive: Do these people look real to you. The New
York Times, 11, 2020.
12
Published as a conference paper at ICLR 2023
Yupan Huang, Hongwei Xue, Bei Liu, and Yutong Lu. Unifying multimodal transformer for bi-
directional image and text generation. In Proceedings of the 29th ACM International Conference
on Multimedia, pp. 1138–1147, 2021.
Zhicheng Huang, Zhaoyang Zeng, Bei Liu, Dongmei Fu, and Jianlong Fu. Pixel-BERT: Aligning
image pixels with text by deep multi-modal transformers. arXiv preprint, 2020.
Taichi Iki and Akiko Aizawa. Effect of visual extensions on natural language understanding in
vision-and-language models. In Proceedings of the 2021 Conference on Empirical Methods in
Natural Language Processing, pp. 2189–2196, 2021.
Shankar Iyer, Nikhil Dandekar, Kornél Csernai, et al. First quora dataset release: Question pairs.
data. quora. com, 2017.
Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yun-Hsuan
Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning
with noisy text supervision. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th
International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event,
volume 139 of Proceedings of Machine Learning Research, pp. 4904–4916. PMLR, 2021. URL
http://proceedings.mlr.press/v139/jia21b.html.
Wonjae Kim, Bokyung Son, and Ildoo Kim. Vilt: Vision-and-language transformer without convo-
lution or region supervision. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th
International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event,
volume 139 of Proceedings of Machine Learning Research, pp. 5583–5594. PMLR, 2021. URL
http://proceedings.mlr.press/v139/kim21k.html.
Jonathan Krause, Michael Stark, Jia Deng, and Li Fei-Fei. 3d object representations for fine-grained
categorization. In Proceedings of the IEEE international conference on computer vision workshops,
pp. 554–561, 2013.
Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009.
Yann LeCun and Corinna Cortes. MNIST handwritten digit database. 2010.
Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer
Levy, Veselin Stoyanov, and Luke Zettlemoyer. BART: Denoising sequence-to-sequence pre-
training for natural language generation, translation, and comprehension. In Proceedings of the
58th Annual Meeting of the Association for Computational Linguistics, pp. 7871–7880, Online,
2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.acl-main.703. URL
https://aclanthology.org/2020.acl-main.703.
Junnan Li, Ramprasaath R Selvaraju, Akhilesh Deepak Gotmare, Shafiq Joty, Caiming Xiong, and
Steven Hoi. Align before fuse: Vision and language representation learning with momentum
distillation. In Conference on Neural Information Processing Systems (NeurIPS), 2021.
Liunian Harold Li, Mark Yatskar, Da Yin, Cho-Jui Hsieh, and Kai-Wei Chang. VisualBERT: A
simple and performant baseline for vision and language. arXiv preprint, 2019.
Liunian Harold Li, Pengchuan Zhang, Haotian Zhang, Jianwei Yang, Chunyuan Li, Yiwu Zhong,
Lijuan Wang, Lu Yuan, Lei Zhang, Jenq-Neng Hwang, et al. Grounded language-image pre-training.
In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.
10965–10975, 2022a.
Xiujun Li, Xi Yin, Chunyuan Li, Pengchuan Zhang, Xiaowei Hu, Lei Zhang, Lijuan Wang, Houdong
Hu, Li Dong, Furu Wei, et al. Oscar: Object-semantics aligned pre-training for vision-language
tasks. In European Conference on Computer Vision (ECCV), 2020.
Yehao Li, Jiahao Fan, Yingwei Pan, Ting Yao, Weiyao Lin, and Tao Mei. Uni-eden: Universal encoder-
decoder network by multi-granular vision-language pre-training. ArXiv preprint, abs/2201.04026,
2022b. URL https://arxiv.org/abs/2201.04026.
13
Published as a conference paper at ICLR 2023
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike
Lewis, Luke Zettlemoyer, and Veselin Stoyanov. RoBERTa: A robustly optimized bert pretraining
approach. arXiv preprint, 2019.
Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In 7th International
Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019.
OpenReview.net, 2019. URL https://openreview.net/forum?id=Bkg6RiCqY7.
Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. Vilbert: Pretraining task-agnostic vi-
siolinguistic representations for vision-and-language tasks. In Hanna M. Wallach, Hugo
Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett
(eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neu-
ral Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC,
Canada, pp. 13–23, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/
c74d97b01eae257e44aa9d5bade97baf-Abstract.html.
Subhransu Maji, Esa Rahtu, Juho Kannala, Matthew Blaschko, and Andrea Vedaldi. Fine-grained
visual classification of aircraft. arXiv preprint arXiv:1306.5151, 2013.
Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory F. Diamos, Erich Elsen, David García,
Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, and Hao Wu. Mixed
precision training. In 6th International Conference on Learning Representations, ICLR 2018,
Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenReview.net,
2018. URL https://openreview.net/forum?id=r1gs9JgRZ.
Maria-Elena Nilsback and Andrew Zisserman. Automated flower classification over a large number
of classes. In 2008 Sixth Indian Conference on Computer Vision, Graphics & Image Processing,
pp. 722–729. IEEE, 2008.
Rui Pan, Shizhe Diao, Jianlin Chen, and Tong Zhang. Extremebert: A toolkit for accelerating
pretraining of customized bert. arXiv preprint arXiv:2211.17201, 2022.
Omkar M. Parkhi, Andrea Vedaldi, Andrew Zisserman, and C. V. Jawahar. Cats and dogs. In 2012
IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, June 16-21,
2012, pp. 3498–3505. IEEE Computer Society, 2012. doi: 10.1109/CVPR.2012.6248092. URL
https://doi.org/10.1109/CVPR.2012.6248092.
Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee,
and Luke Zettlemoyer. Deep contextualized word representations. In Proceedings of the 2018
Conference of the North American Chapter of the Association for Computational Linguistics:
Human Language Technologies, Volume 1 (Long Papers), pp. 2227–2237, New Orleans, Louisiana,
2018. Association for Computational Linguistics. doi: 10.18653/v1/N18-1202. URL https:
//aclanthology.org/N18-1202.
Jason Phang, Thibault Févry, and Samuel R. Bowman. Sentence encoders on stilts: Supplementary
training on intermediate labeled-data tasks. ArXiv, abs/1811.01088, 2018.
Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. Improving language under-
standing by generative pre-training. 2018.
Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal,
Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever.
Learning transferable visual models from natural language supervision. In Marina Meila and
Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML
2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research,
pp. 8748–8763. PMLR, 2021. URL http://proceedings.mlr.press/v139/radford21a.
html.
Jack W Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John
Aslanides, Sarah Henderson, Roman Ring, Susannah Young, et al. Scaling language models:
Methods, analysis & insights from training gopher. arXiv preprint arXiv:2112.11446, 2021.
14
Published as a conference paper at ICLR 2023
Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi
Zhou, Wei Li, and Peter J Liu. Exploring the limits of transfer learning with a unified text-to-text
transformer. Journal of Machine Learning Research (JMLR), 2020.
Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. SQuAD: 100,000+ questions for
machine comprehension of text. In Proceedings of the 2016 Conference on Empirical Methods in
Natural Language Processing, pp. 2383–2392, Austin, Texas, 2016. Association for Computational
Linguistics. doi: 10.18653/v1/D16-1264. URL https://aclanthology.org/D16-1264.
Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen,
and Ilya Sutskever. Zero-shot text-to-image generation. In Marina Meila and Tong Zhang (eds.),
Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July
2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 8821–8831.
PMLR, 2021. URL http://proceedings.mlr.press/v139/ramesh21a.html.
Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-
conditional image generation with clip latents. arXiv preprint arXiv:2204.06125, 2022.
Steven J. Rennie, Etienne Marcheret, Youssef Mroueh, Jerret Ross, and Vaibhava Goel. Self-critical
sequence training for image captioning. In 2017 IEEE Conference on Computer Vision and Pattern
Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 1179–1195. IEEE Computer
Society, 2017. doi: 10.1109/CVPR.2017.131. URL https://doi.org/10.1109/CVPR.2017.
131.
Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang,
Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual recognition
challenge. International journal of computer vision, 115(3):211–252, 2015.
Andrei A Rusu, Neil C Rabinowitz, Guillaume Desjardins, Hubert Soyer, James Kirkpatrick, Koray
Kavukcuoglu, Razvan Pascanu, and Raia Hadsell. Progressive neural networks. arXiv preprint
arXiv:1606.04671, 2016.
Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed
Kamyar Seyed Ghasemipour, Burcu Karagol Ayan, S Sara Mahdavi, Rapha Gontijo Lopes, et al.
Photorealistic text-to-image diffusion models with deep language understanding. arXiv preprint
arXiv:2205.11487, 2022.
Tim Salimans, Ian J. Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen.
Improved techniques for training gans. In Daniel D. Lee, Masashi Sugiyama, Ulrike von Luxburg,
Isabelle Guyon, and Roman Garnett (eds.), Advances in Neural Information Processing Systems
29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016,
Barcelona, Spain, pp. 2226–2234, 2016. URL https://proceedings.neurips.cc/paper/
2016/hash/8a3363abe792db2d8761d6403605aeb7-Abstract.html.
Roy Schwartz, Jesse Dodge, Noah A Smith, and Oren Etzioni. Green ai. Communications of the
ACM, 63(12):54–63, 2020.
Amanpreet Singh, Ronghang Hu, Vedanuj Goswami, Guillaume Couairon, Wojciech Galuba, Marcus
Rohrbach, and Douwe Kiela. Flava: A foundational language and vision alignment model. ArXiv
preprint, abs/2112.04482, 2021. URL https://arxiv.org/abs/2112.04482.
Richard Socher, Alex Perelygin, Jean Wu, Jason Chuang, Christopher D. Manning, Andrew Ng, and
Christopher Potts. Recursive deep models for semantic compositionality over a sentiment treebank.
In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing,
pp. 1631–1642, Seattle, Washington, USA, 2013. Association for Computational Linguistics. URL
https://aclanthology.org/D13-1170.
Weijie Su, Xizhou Zhu, Yue Cao, Bin Li, Lewei Lu, Furu Wei, and Jifeng Dai. VL-BERT: pre-
training of generic visual-linguistic representations. In 8th International Conference on Learning
Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020.
URL https://openreview.net/forum?id=SygXPaEYvH.
15
Published as a conference paper at ICLR 2023
Alane Suhr, Stephanie Zhou, Ally Zhang, Iris Zhang, Huajun Bai, and Yoav Artzi. A corpus
for reasoning about natural language grounded in photographs. In Proceedings of the 57th
Annual Meeting of the Association for Computational Linguistics, pp. 6418–6428, Florence,
Italy, 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1644. URL
https://aclanthology.org/P19-1644.
Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jonathon Shlens, and Zbigniew Wojna.
Rethinking the inception architecture for computer vision. In 2016 IEEE Conference on
Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30,
2016, pp. 2818–2826. IEEE Computer Society, 2016. doi: 10.1109/CVPR.2016.308. URL
https://doi.org/10.1109/CVPR.2016.308.
Hao Tan and Mohit Bansal. LXMERT: Learning cross-modality encoder representations from
transformers. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language
Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-
IJCNLP), pp. 5100–5111, Hong Kong, China, 2019a. Association for Computational Linguistics.
doi: 10.18653/v1/D19-1514. URL https://aclanthology.org/D19-1514.
Hao Tan and Mohit Bansal. LXMERT: Learning cross-modality encoder representations from
transformers. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language
Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-
IJCNLP), pp. 5100–5111, Hong Kong, China, 2019b. Association for Computational Linguistics.
doi: 10.18653/v1/D19-1514. URL https://aclanthology.org/D19-1514.
Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé
Jégou. Training data-efficient image transformers & distillation through attention. In Marina Meila
and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning,
ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning
Research, pp. 10347–10357. PMLR, 2021. URL http://proceedings.mlr.press/v139/
touvron21a.html.
Aäron van den Oord, Oriol Vinyals, and Koray Kavukcuoglu. Neural discrete representation learning.
In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N.
Vishwanathan, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 30:
Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long
Beach, CA, USA, pp. 6306–6315, 2017. URL https://proceedings.neurips.cc/paper/
2017/hash/7a98af17e63a0ac09ce2e96d03992fbc-Abstract.html.
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez,
Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Isabelle Guyon, Ulrike
von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Ro-
man Garnett (eds.), Advances in Neural Information Processing Systems 30: Annual Confer-
ence on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA,
USA, pp. 5998–6008, 2017. URL https://proceedings.neurips.cc/paper/2017/hash/
3f5ee243547dee91fbd053c1c4a845aa-Abstract.html.
Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman.
GLUE: A multi-task benchmark and analysis platform for natural language understanding. In 7th
International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May
6-9, 2019. OpenReview.net, 2019. URL https://openreview.net/forum?id=rJ4km2R5t7.
Peng Wang, An Yang, Rui Men, Junyang Lin, Shuai Bai, Zhikang Li, Jianxin Ma, Chang Zhou,
Jingren Zhou, and Hongxia Yang. Unifying architectures, tasks, and modalities through a simple
sequence-to-sequence learning framework. ArXiv preprint, abs/2202.03052, 2022. URL https:
//arxiv.org/abs/2202.03052.
Wenhui Wang, Hangbo Bao, Li Dong, and Furu Wei. Vlmo: Unified vision-language pre-training
with mixture-of-modality-experts. ArXiv preprint, abs/2111.02358, 2021a. URL https://arxiv.
org/abs/2111.02358.
Zirui Wang, Jiahui Yu, Adams Wei Yu, Zihang Dai, Yulia Tsvetkov, and Yuan Cao. Simvlm: Simple
visual language model pretraining with weak supervision. arXiv preprint, 2021b.
16
Published as a conference paper at ICLR 2023
Alex Warstadt, Amanpreet Singh, and Samuel R. Bowman. Neural network acceptability judgments.
Transactions of the Association for Computational Linguistics, 7:625–641, 2019. doi: 10.1162/
tacl_a_00290. URL https://aclanthology.org/Q19-1040.
Jason Wei, Yi Tay, Rishi Bommasani, Colin Raffel, Barret Zoph, Sebastian Borgeaud, Dani Yo-
gatama, Maarten Bosma, Denny Zhou, Donald Metzler, Ed H. Chi, Tatsunori Hashimoto, Oriol
Vinyals, Percy Liang, Jeff Dean, and William Fedus. Emergent abilities of large language models.
Transactions on Machine Learning Research, 2022. URL https://openreview.net/forum?
id=yzkSU5zdwD. Survey Certification.
Adina Williams, Nikita Nangia, and Samuel Bowman. A broad-coverage challenge corpus for
sentence understanding through inference. In Proceedings of the 2018 Conference of the North
American Chapter of the Association for Computational Linguistics: Human Language Technolo-
gies, Volume 1 (Long Papers), pp. 1112–1122, New Orleans, Louisiana, 2018. Association for
Computational Linguistics. doi: 10.18653/v1/N18-1101. URL https://aclanthology.org/
N18-1101.
Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V Le, Mohammad Norouzi, Wolfgang Macherey,
Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, et al. Google’s neural machine translation
system: Bridging the gap between human and machine translation. ArXiv preprint, abs/1609.08144,
2016. URL https://arxiv.org/abs/1609.08144.
Ning Xie, Farley Lai, Derek Doran, and Asim Kadav. Visual entailment: A novel task for fine-grained
image understanding. arXiv preprint, 2019.
Canwen Xu, Wangchunshu Zhou, Tao Ge, Furu Wei, and Ming Zhou. BERT-of-theseus: Com-
pressing BERT by progressive module replacing. In Proceedings of the 2020 Conference on
Empirical Methods in Natural Language Processing (EMNLP), pp. 7859–7869, Online, 2020.
Association for Computational Linguistics. doi: 10.18653/v1/2020.emnlp-main.633. URL
https://aclanthology.org/2020.emnlp-main.633.
Haiyang Xu, Ming Yan, Chenliang Li, Bin Bi, Songfang Huang, Wenming Xiao, and Fei Huang. E2E-
VLP: End-to-end vision-language pre-training enhanced by visual learning. In Proceedings of the
59th Annual Meeting of the Association for Computational Linguistics and the 11th International
Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 503–513, Online,
2021a. Association for Computational Linguistics. doi: 10.18653/v1/2021.acl-long.42. URL
https://aclanthology.org/2021.acl-long.42.
Jingjing Xu, Wangchunshu Zhou, Zhiyi Fu, Hao Zhou, and Lei Li. A survey on green deep learning.
ArXiv preprint, abs/2111.05193, 2021b. URL https://arxiv.org/abs/2111.05193.
Tao Xu, Pengchuan Zhang, Qiuyuan Huang, Han Zhang, Zhe Gan, Xiaolei Huang, and Xiaodong
He. Attngan: Fine-grained text to image generation with attentional generative adversarial
networks. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018,
Salt Lake City, UT, USA, June 18-22, 2018, pp. 1316–1324. IEEE Computer Society, 2018.
doi: 10.1109/CVPR.2018.00143. URL http://openaccess.thecvf.com/content_cvpr_
2018/html/Xu_AttnGAN_Fine-Grained_Text_CVPR_2018_paper.html.
Zhengyuan Yang, Zhe Gan, Jianfeng Wang, Xiaowei Hu, Faisal Ahmed, Zicheng Liu, Yumao Lu, and
Lijuan Wang. Crossing the format boundary of text and boxes: Towards unified vision-language
modeling. ArXiv, abs/2111.12085, 2021.
Yang You, Igor Gitman, and Boris Ginsburg. Large batch training of convolutional networks. ArXiv
preprint, abs/1708.03888, 2017. URL https://arxiv.org/abs/1708.03888.
Jiahui Yu, Yuanzhong Xu, Jing Yu Koh, Thang Luong, Gunjan Baid, Zirui Wang, Vijay Vasudevan,
Alexander Ku, Yinfei Yang, Burcu Karagol Ayan, et al. Scaling autoregressive models for content-
rich text-to-image generation. arXiv preprint arXiv:2206.10789, 2022.
Ning Yu, Vladislav Skripniuk, Sahar Abdelnabi, and Mario Fritz. Artificial fingerprinting for
generative models: Rooting deepfake attribution in training data. In Proceedings of the IEEE/CVF
International Conference on Computer Vision, pp. 14448–14457, 2021.
17
Published as a conference paper at ICLR 2023
Lu Yuan, Dongdong Chen, Yi-Ling Chen, Noel Codella, Xiyang Dai, Jianfeng Gao, Houdong Hu,
Xuedong Huang, Boxin Li, Chunyuan Li, Ce Liu, Mengchen Liu, Zicheng Liu, Yumao Lu, Yu Shi,
Lijuan Wang, Jianfeng Wang, Bin Xiao, Zhen Xiao, Jianwei Yang, Michael Zeng, Luowei Zhou,
and Pengchuan Zhang. Florence: A new foundation model for computer vision. arXiv preprint,
2021.
Yan Zeng, Xinsong Zhang, and Hang Li. Multi-grained vision language pre-training: Aligning texts
with visual concepts. ArXiv preprint, abs/2111.08276, 2021. URL https://arxiv.org/abs/
2111.08276.
Han Zhang, Weichong Yin, Yewei Fang, Lanxin Li, Boqiang Duan, Zhihua Wu, Yu Sun, Hao Tian,
Hua Wu, and Haifeng Wang. Ernie-vilg: Unified generative pre-training for bidirectional vision-
language generation. ArXiv preprint, abs/2112.15283, 2021a. URL https://arxiv.org/abs/
2112.15283.
Pengchuan Zhang, Xiujun Li, Xiaowei Hu, Jianwei Yang, Lei Zhang, Lijuan Wang, Yejin Choi, and
Jianfeng Gao. VinVL: Revisiting visual representations in vision-language models. In Conference
on Computer Vision and Pattern Recognition (CVPR), 2021b.
Wangchunshu Zhou, Canwen Xu, Tao Ge, Julian J. McAuley, Ke Xu, and Furu Wei. BERT loses
patience: Fast and robust inference with early exit. In Hugo Larochelle, Marc’Aurelio Ranzato,
Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information
Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020,
NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.
cc/paper/2020/hash/d4dd111a4fd973394238aca5c05bebe3-Abstract.html.
Wangchunshu Zhou, Tao Ge, Canwen Xu, Ke Xu, and Furu Wei. Improving sequence-to-sequence pre-
training via sequence span rewriting. In Proceedings of the 2021 Conference on Empirical Methods
in Natural Language Processing, pp. 571–582, Online and Punta Cana, Dominican Republic,
2021a. Association for Computational Linguistics. doi: 10.18653/v1/2021.emnlp-main.45. URL
https://aclanthology.org/2021.emnlp-main.45.
Wangchunshu Zhou, Dong-Ho Lee, Ravi Kiran Selvam, Seyeon Lee, and Xiang Ren. Pre-training
text-to-text transformers for concept-centric common sense. In 9th International Conference on
Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net,
2021b. URL https://openreview.net/forum?id=3k20LAiHYL2.
Wangchunshu Zhou, Canwen Xu, and Julian McAuley. BERT learns to teach: Knowledge distillation
with meta learning. In ACL (1), pp. 7037–7049. Association for Computational Linguistics, 2022a.
Wangchunshu Zhou, Yan Zeng, Shizhe Diao, and Xinsong Zhang. Vlue: A multi-task benchmark for
evaluating vision-language models. CoRR, abs/2205.15237, 2022b.
Xiao Zhou, Weizhong Zhang, Zonghao Chen, Shizhe Diao, and Tong Zhang. Efficient neural network
training via forward and backward propagation sparsification. Advances in Neural Information
Processing Systems, 34, 2021c.
Xiao Zhou, Weizhong Zhang, Hang Xu, and Tong Zhang. Effective sparsification of neural networks
with global sparsity constraint. In Proceedings of the IEEE/CVF Conference on Computer Vision
and Pattern Recognition, pp. 3599–3608, 2021d.
Xiao Zhou, Renjie Pi, Weizhong Zhang, Yong Lin, and Tong Zhang. Probabilistic bilevel coreset
selection. In International Conference on Machine Learning. PMLR, 2022c.
Minfeng Zhu, Pingbo Pan, Wei Chen, and Yi Yang. DM-GAN: dynamic memory generative adver-
sarial networks for text-to-image synthesis. In IEEE Conference on Computer Vision and Pattern
Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pp. 5802–5810. Computer
Vision Foundation / IEEE, 2019. doi: 10.1109/CVPR.2019.00595. URL http://openaccess.
thecvf.com/content_CVPR_2019/html/Zhu_DM-GAN_Dynamic_Memory_Generative_
Adversarial_Networks_for_Text-To-Image_Synthesis_CVPR_2019_paper.html.
18
Published as a conference paper at ICLR 2023
A A PPENDIX
A.1 D ETAILS OF H YPER - PARAMETERS
Pre-training Our model is a base-size Transformer implemented with a 6-layer encoder and a
6-layer decoder, 768 dimensions for hidden states, 512 for maximum input length, and 3072 for
intermediate size. We train our model from scratch without initializing the Transformer encoder
and decoder. The image encoder is initialized from ResNet-101 (He et al., 2016) with ImageNet
weights since we find a warm start provides a reliable visual representation and helps the convergence.
For models pre-training on large-scale data, we optimize 10 epochs while for other small-scale
datasets, we optimize 40 epochs with the AdamW optimizer. The weight decay is set to 0.01 with
β1 = 0.9, β2 = 0.999. The learning rate is 2e-4 with a warm-up period for the first 2% steps and
linearly decayed to 0 after 2% of the total training steps. In each batch, there are 8,192 image-text
pairs for text-to-image generation and image-to-text generation with 8,192 text-only documents for
text-to-text generation. We use center-crop to resize each image to the size of 256×256, which is the
only data augmentation used during training. All pre-training experiments are conducted on 32GB
NVIDIA V100 GPUs. We adopt mixed-precision (Micikevicius et al., 2018) to accelerate training
and save memory. The model trained on the largest data takes around 10 days on 1024 V100 GPUs.
The default settings are shown in Table 6. We adopt dynamic masking in our experiments, where the
masking ratio is randomly sampled from a uniform distribution U(0, 1).
Fine-tuning The learning rate is ∈ [1e-5, 5e-5] and our model is optimized by AdamW. Because the
image resolution differs between pre-training and fine-tuning, the position parameters are adapted
using linear interpolation. For all downstream tasks, we apply random resize crops and horizontal flips
augmentation during training. All fine-tuning experiments are conducted on 32GB NVIDIA V100
GPUs. The default settings for text classification, image classification, multi-modal understanding
and image-to-text generation are shown in Tables 7, 8, and 9, respectively.
config value
optimizer AdamW (Loshchilov & Hutter, 2019)
learning rate 2e-4
weight decay 0.01
optimizer momentum β1 , β2 =0.9, 0.999
batch size 8192
learning rate schedule linear decay
warmup ratio (Goyal et al., 2017a) 0.02
training epochs {10, 40}
augmentation RandomResizedCrop
config value
optimizer AdamW
learning rate {1e-5, 2e-5, 5e-5}
weight decay 0.01
optimizer momentum β1 , β2 =0.9, 0.999
batch size {16, 32, 64}
learning rate schedule linear decay
warmup ratio 0.1
training epochs {5, 10}
19
Published as a conference paper at ICLR 2023
config value
optimizer LARS (You et al., 2017)
base learning rate 0.1
weight decay 0
optimizer momentum 0.9
batch size 16384
learning rate schedule cosine decay
warmup epochs 10
training epochs 90
augmentation RandomResizedCrop
config value
optimizer AdamW
learning rate [1e-5, 5e-5]
weight decay 0.02
optimizer momentum β1 , β2 =0.9, 0.999
batch size 1024
learning rate schedule linear decay
warmup epochs [2, 5]
training epochs [5, 15]
label smoothing (Szegedy et al., 2016) 0.1
augmentation RandomResizedCrop, HorizontalFlips
RTE (Dagan et al., 2005; Haim et al., 2006; Giampiccolo et al., 2007; Bentivogli et al., 2009), and
STS-B (Agirre et al., 2007). We follow the practice of BART (Lewis et al., 2020) and feed the
same input to the encoder and decoder, and the hidden state of the final decoder token is fed into a
new multi-class linear classifier or regression head. MNLI results are an average of MNLI-m and
MNLI-mm. MRPC and QQP results are average of accuracy and F1. Matthews correlation coefficient
(MCC) is reported for CoLA and Pearson correlation coefficient (PCC) is reported for STS-B.
Vision Understanding We conduct vision experiments in both fine-tuning and linear evaluation
(linear eval). The linear evaluation follows a common practice (Caron et al., 2021; He et al.,
2020; Singh et al., 2021) in self-supervised learning to evaluate the representation quality, where
the pre-trained backbone model is frozen and a new linear classifier is appended on top of it.
We choose 12 popular datasets: ImageNet (Russakovsky et al., 2015), Food101 (Bossard et al.,
2014), CIFAR10 (Krizhevsky et al., 2009), CIFAR100 (Krizhevsky et al., 2009), Cars (Krause
et al., 2013), Aircraft (Maji et al., 2013), DTD (Cimpoi et al., 2014), Pets (Parkhi et al., 2012),
Flowers102 (Nilsback & Zisserman, 2008), MNIST (LeCun & Cortes, 2010), STL10 (Coates et al.,
2011), and Country211 (Radford et al., 2021).
Multi-modal Understanding We consider three popular multi-modal tasks: VQAv2 (Goyal et al.,
2017b), SNLI-VE (Xie et al., 2019) and NLVR2 (Suhr et al., 2019) to evaluate our model’s multi-
modal understanding ability. For VQAv2, following ALBEF (Li et al., 2021), the image and question
are fed to the encoder and the decoder generates answers based on the multi-modal embeddings. For
SNLI-VE, we follow SimVLM (Wang et al., 2021b) to feed the image to the encoder and the text to
the decoder. A classifier is appended on top of our pre-trained model, and it is trained to predict the
result based on the last hidden states of the decoder. For NLVR2, two input pairs are constructed,
each of them including one image and the textual description. The prediction is made based on the
concatenation of these two embeddings following SimVLM (Wang et al., 2021b). The resolutions for
VQAv2, SNLI-VE, NLVR2 are 480, 384, 384, respectively.
Text-to-Image Generation The text-to-image task requires the model to understand the textual
instruction first and then draw the image according to the input’s intention. The input text is fed to
our encoder, and our decoder will generate visual tokens one by one. After obtaining visual tokens,
they are decoded into a raw image by an image decoder. We directly use an off-the-shelf image
decoder from VQGAN (Esser et al., 2021). Following (Ramesh et al., 2021) we directly evaluate our
20
Published as a conference paper at ICLR 2023
Table 10: Statistics of the pre-training datasets. #Images, #Captions, and #Total denote the number of images,
the number of image-text pairs, and the total number of image-text pairs, respectively.
pre-trained model on 30, 000 images randomly sampled from COCO (Chen et al., 2015) validation
split. Both Fréchet Inception Distance (FID) (Heusel et al., 2017) and Inception Score (IS) (Salimans
et al., 2016) are reported. The image resolution is 256.
Image-to-Text Generation For image-to-text generation (also called image captioning), the image
is given to encoder and the decoder will generate the corresponding caption. Our experiments are
conducted on COCO dataset (Chen et al., 2015) with cross-entropy optimization. Other task-specific
techniques such as CIDEr optimization (Rennie et al., 2017) are not introduced. The image resolution
is 480. We also conduct zero-shot captioning experiments on NoCaps (Agrawal et al., 2019) and
VLUE (Zhou et al., 2022b).
21
Published as a conference paper at ICLR 2023
layers, 512 embedding dimensions, and trained on about 200M web data. To demonstrate the success
of our replication, we train a SimVLMsmall model with the exact same configurations on about 200M
web data. We obtain a VQA score of 68.50, surpassing the reported score of 67.43 in the original
paper. We argue this result verifies our successful replication.
Figure 2: The effects of compute. X-axis is the percentage of compute and Y-axis is the CIDEr score on
COCO captioning task.
COCO VQA SNLI-VE NLVR2 ImageNet Food101 CIFAR10 MNLI SST-2 Text2Image
Method
B@4 / C Acc Acc Acc Acc Acc Acc Acc Acc IS / FID
MIM 34.7 / 113.4 68.18 75.34 69.66 48.46 56.95 72.79 81.72 89.84 9.50 / 74.13
In-painting 34.5 / 112.5 67.46 75.41 68.66 47.50 54.38 71.20 81.55 89.84 9.97 / 68.15
Suffix-painting (ours) 35.8 / 117.3 69.25 76.22 72.55 48.88 75.32 73.82 81.76 90.25 12.35 / 53.14
Token Projection 17.7 / 49.2 52.13 71.11 52.01 15.11 25.62 61.01 82.01 90.25 11.89 / 60.96
Patch Projection 25.7 / 79.5 57.69 71.92 57.45 36.23 44.31 69.40 81.73 90.05 11.41 / 61.87
ResNet Feature (ours) 35.8 / 117.3 69.25 76.22 72.55 48.88 75.32 73.82 81.76 90.25 12.35 / 53.14
Table 11: The effects of masking strategies and image feature extraction on COCO Captions, VQA,
SNLI-VE, NLVR2, ImageNet, Food101, CIFAR10, MNLI, SST-2, and text-to-image generation. MIM
denotes masked image modeling, where some patches are randomly sampled and masked. Because linear probe
and zero-shot text-to-image generation require a pre-trained model to be frozen, the “No Pre-training” results on
ImageNet, Food101, CIFAR10, and Text2Image are not reported and labeled by ∗ .
22
Published as a conference paper at ICLR 2023
23
Published as a conference paper at ICLR 2023
Table 12: Comparison with state-of-the-art vision-language models on vision, language, and multi-modal
downstream tasks. All results are from base-size models. LE and FT denote linear evaluation and fine-tuning
performance, respectively. Image2Text results are reported without CIDEr optimization. † are our reproduced
models. ∗ are the results after fine-tuning. SimVLM (1.8B) and OFA are pre-trained with much larger corpus
or human-labeled data of many downstream tasks, and thus they are not comparable and are labeled in gray.
Florence (Yuan et al., 2021) is pre-trained with a much larger model size (Florence-CoSwin-H, 637M) and more
pre-training data (900M), so the numbers are in grey. bold denotes the best across unified models.
DALLE
OFA
Ours
a large flower is sitting in the a park with flowers on a meat sitting on a plate with green a fire hydrant sitting in a
vase on the shelf sunny day vegetables on the side front yard next to a sign
24
Published as a conference paper at ICLR 2023
a decorative flower vase full of a vase full of flowers on table a park with flowers on a sunny day a fire hydrant sitting in a front
purple and yellow flowers yard next to a sign
a beach on a sunny day a one cart train coming down a red and white boat docked on shore a picture of a snowy mountain
the railroad tracks
a red stop sign on the side a building in front of a roundabout bathroom with marble walls and trees by the river in the mountains
of the road with a tree in the center. counter surrounds a large mirror
many fruits on the plate on the table a bunch of fruit in a fruit shop a table set with a sandwich and a drink noodles and broccoli on a plate
25