NeurIPS 2024 Understanding Transformer Reasoning Capabilities via Graph Algorithms Paper Conference
NeurIPS 2024 Understanding Transformer Reasoning Capabilities via Graph Algorithms Paper Conference
Abstract
Which transformer scaling regimes are able to perfectly solve different classes
of algorithmic problems? While tremendous empirical advances have been at-
tained by transformer-based neural networks, a theoretical understanding of their
algorithmic reasoning capabilities in realistic parameter regimes is lacking. We
investigate this question in terms of the network’s depth, width, and number of extra
tokens for algorithm execution. Our novel representational hierarchy separates 9
algorithmic reasoning problems into classes solvable by transformers in different
realistic parameter scaling regimes. We prove that logarithmic depth is necessary
and sufficient for tasks like graph connectivity, while single-layer transformers
with small embedding dimensions can solve contextual retrieval tasks. We also
support our theoretical analysis with ample empirical evidence using the GraphQA
benchmark. These results show that transformers excel at many graph reasoning
tasks, even outperforming specialized graph neural networks.
1 Introduction
The transformer neural network architecture, which was initially introduced for neural machine
translation [6, 77], quickly became the standard neural network architecture across many fields,
powering recent breakthroughs in language modeling [20, 66, 12, 76, 3, 75, 67], computer vision [21,
49], and natural sciences [36, 53]. Across fields, transformers have superseded other architectures,
surpassing them in downstream performance while maintaining reasonable computational footprint.
How can we analyze reasoning capabilities of neural networks? One approach is to study algorithms
executable with their internal representations. Neural algorithmic reasoning [30, 89, 34, 37, 78] is a
field of research dedicated to exploring such capabilities. Algorithmic execution is desirable because
models use it to generalize out-of-distribution [92, 46] and scale to larger problem sizes [93].
In this work, we focus on classes of transformers solving algorithmic reasoning problems on graphs.
Why graph problems in particular? Recent research by Besta et al. [10] suggests that graphs
are an ideal abstraction of complex reasoning with dependencies, including chain-of-thought [82]
and its generalizations [86, 17, 41, 40]. Furthermore, we investigate graph reasoning in order
to evaluate transformer capabilities compared to specialized models that explicitly capture the
structure of data. Graph neural networks (GNNs) [72, 44, 28, 7, 15] offer strong baselines for
algorithmic reasoning [79, 13] and comprehensive theoretical frameworks for their capabilities and
limitations [51, 84]. The complexity of graph reasoning tasks varies greatly; some are easily solved
by non-graph architectures and others require sophistication beyond standard GNNs [90]. This makes
graph tasks a compelling testbed for both theoretical and empirical evaluations.
While transformer-based models are commonly optimized for downstream performance [38, 32, 64],
theoretical investigations of transformer capabilities in realistic parameter regimes have been limited.
∗
Correspondence: [email protected] or [email protected].
Vertex tokens
∅
Graph G
v4 Self-Attention ∅
v4 v5 ∅
v2 v6 MLP ∅
v3 v7 ∅
v5
v1 v8 Self-Attention ∅
v6 v8 v1 , v2 ∅
v2 , v3 MLP ∅
Edge tokens
v7
v3 , v4 ∅
Task: Are v2 and v4 connected? v6 , v7 Self-Attention ∅
v6 , v8 ... ∅
v7 , v8 ∅
Task tokens
v2 , v4 Yes
Figure 1: The graph encoding scheme employed in our theoretical and empirical analysis that presents
a graph reasoning task (e.g. connectivity) as a tokenized input to a standard transformer model.
We analyze the capabilities of standard transformer models to solve graph reasoning tasks by
employing a graph tokenization scheme similar to [42]; see Figure 1 for a graph encoding example.
We generalize the insights of [71]—which established that graph connectivity can be computed in a
more depth-efficient manner with transformers than GNNs—to broader families of graph reasoning
tasks. We also study the representational impacts of blank tokens, which can be thought of as
a theoretically-tractable version of either chain-of-thought prompting [82], scratch space [61], or
pause/filler tokens [29, 65].
We conduct an extensive study of graph reasoning problems with transformers from both theoretical
and empirical perspectives. We summarize our principal contributions as follows:
1. We introduce a novel representational hierarchy of graph reasoning tasks that formalizes reasoning
capabilities of transformers in several realistic parameter scaling regimes. This includes two graph
reasoning task classes—which we refer to as the parallelizable and search tasks and include
well-known algorithmic tasks like connectivity and shortest path respectively.
2. We prove that logarithmic-depth transformers are necessary and sufficient to solve parallelizable
tasks in a highly parameter-efficient manner; similar constructions for search tasks employ much
larger networks. We distinguish both of these classes from an easier family of problems—retrieval
tasks, such as node count and edge existence—by showing that retrieval tasks can be efficiently
computed by single-layer transformers, while parallelizable and search tasks cannot.
3. We empirically validate our representational contrast by showing that transformers outperform
GNNs on tasks that require the analysis of long-range dependencies, including parallelizable tasks
like connectivity. Our experiments suggest that GNNs perform better on simpler tasks that require
only local analysis of neighboring nodes in low sample-complexity regimes, benefiting from their
well-aligned inductive bias.
2 Related work
Fundamental capabilities of transformers. Early representational research [88, 62, 81] estab-
lished the universality of transformers by simulating Turing machines step-by-step, albeit in an
unrealistic polynomial-depth scaling regime. These universality results were extended to bounded-
depth transformers with chain-of-thought tokens [56, 52], but with a persistent gap between positive
and negative results. A more fine-grained theoretical approach established the limitations of bounded-
size transformers by relating them to threshold circuits, implying that L-complete tasks like graph
connectivity are unsolvable by constant-depth transformers [55, 54, 57]. However, these circuit
complexity reductions are not bidirectional and their positive results pertaining to classes of regular
languages [31] reveal little about when tasks like connectivity can be expressed.
2
The closest analytical framework to ours characterizes transformers as distributed computing protocols
by quantifying the informational bandwidth of their self-attention units. This line of work sharply
separates transformers from other architectures and provides width and depth separations [70, 71].
The ability of transformers to simulate finite-state automata in logarithmic depth provides a further
roadmap for how transformers efficiently leverage parallel computation [48].
Recent work [18] similarly focuses on understanding the capabilities of transformers to solve graph
algorithms, but they focus instead on looped transformers, where a constant-size transformer model is
iterated up to a polynomial number of times in the size of the input graph. While their constructions for
tasks like graph connectivity are more parameter-efficient than our own, this comes at the cost of a very
high depth, which presents both run-time and learnability concerns. Our depth-efficient constructions
emerge by simulating parallel graph algorithms, rather than well-known serial algorithms, such as
Dijkstra’s.
Transformers and graph neural networks. GNNs provide a useful benchmark for model expres-
sivity on graph inputs. For instance, the inability of GNNs to efficiently solve “global structure” tasks
like subgraph connectivity made evident by a connection to the C ONGEST distributed computing
model [51]. A further connection [59] to the Weisfeiler-Lehman (WL) isomorphism heuristic [83]
captures the inability of feature-less GNNs to distinguish certain graph instances [84]. The feature-
less assumption is burdensome and representationally costly; GNNs with randomly intialized node
features are known to be universal [2]. See Appendix D for a further discussion of the representational
limitations of GNNs.
Graph transformers [22, 68, 58] integrate the transformer architecture with graph-structured data.
While GNNs can simulate the WL test [1], transformers can simulate GNNs and exceed their WL
limitations [90]; they are strictly more expressive than 2-WL [42]. Despite our focus on standard
transformers, our empirical results in Section 4 include comparisons to a wide range of GNNs.
3
Task class Example tasks Complexity
Parallelizable Retrieval (§3.3) Node count D1
L=1 Edge count D1
LDP LD LDW m = O(log N ) Edge existence D1
Search Node degree D1
Parallelizable (§3.1) Connectivity LD
D1 L = O(log N ) Cycle check LDP∩ LDW
Retrieval m = O(N ) Bipartiteness LDP∩ LDW
Search (§3.2) Shortest path LDW
L = O(log N ) Diameter LDW
m = O(N 1/2+ )
(a) The complexity hierarchy (b) Example tasks and their complexity.
Figure 2: A summary of the theoretical hierarchy of Section 3 that visualizes which type of graph
reasoning tasks can be solved in which transformer scaling regime (Depth1 (D1), LogDepth (LD),
LogDepthWide (LDW) and LogDepthPause (LDP)).
These results divide the tasks that are we empirically investigate in Section 4 into three difficulty
families based on the hardness of solving the task with parallel computation.
1. Retrieval tasks—including node count, edge count, edge existence, and node degree—are tasks
that can intuitively be solved by a single lookup step or global aggregation. These are the easiest
tasks in our framework, and we show in Section 3.3 that these retrieval tasks can be computed by
a single-layer transformer with small embedding dimension. (In contrast, all other examined tasks
cannot be solved in that regime.)
2. Parallelizable tasks—including connectivity, connected nodes, and cycle check from the exper-
imental results and a host of other graph reasoning tasks, such as bipartiteness, planarity, and
minimum spanning forest—are non-trivial tasks that can be solved efficiently in a parallel compu-
tation setting. Section 3.1 establishes that these tasks can be solved by bounded-size transformers
with logarithmic depth.
3. Search tasks—including shortest path and other tasks like diameter and directed reachability—
comprise a harder family of tasks that are less easily solved by parallel algorithms. In Section 3.2,
we prove that these tasks belong in an equivalence class and exhibit large-model scaling regimes
where they can be computed.
The representational hardness of these classes is quantified by several results that determine whether
transformers that obey different parameter scaling rules can compute them. We define the following
scaling regimes for the depth L, embedding dimension m, and number of “pause” tokens N 0 of a
family of transformers as a function of the size of the input graph, N = O(|V | + |E|).
• LogDepth (LD): Transformers with depth L = O(log N ), embedding dimension m = O(N ) for
any fixed > 0, and no pause tokens.
• LogDepthPause (LDP): Transformers with the same depth and width constraints as LogDepth,
with at most N 0 = poly(N ) blank “pause” tokens appended to the input sequence.
Our positive and negative results that relate scaling regimes and graph reasoning tasks are displayed in
Figure 2. We present high-level result summaries in the following sections with proofs in Appendix.
The most technically significant results concern LogDepth models and are provided in Appendix B.
These bounds are consequences of Theorem 1, an improved analysis of the relationship between
4
transformers and the massively parallel computation (MPC) distributed computing model of [39]2 .
This connection between MPC and transformers is a sharp improvement of a similar result by [71].
Theorem 1 (Simplified version of Theorem 8). For constant δ, > 0, any R-round MPC protocol
with N machines with O(N δ ) bits of local memory each can be simulated by a transformer of depth
L = O(R) and embedding dimension m = O(N δ+ ).
Results pertaining to Depth1 are stated in detail and proved in Appendix C. In addition, we discuss
the triangle counting task (and the more general clique counting task) in Section 3.4, where we show
a distinct result for much smaller depth (L = O(log log N )) that includes pause tokens.
We define the family of parallelizable tasks to consist of graph reasoning tasks that are L-complete
and are equivalent to graph connectivity in O(1)-rounds, as proved by [60]. This family includes
(but is not limited to): graph connectivity, minimum spanning forest, cycle detection, st-connectivity,
number of connected components, graph bipartiteness, planarity testing, and one-cycle vs two-cycles
testing. While graph connectivity and minimum spanning forest were shown to be computable by
logarithmic depth transformers with small polynomial width (LogDepth) by previous work [71], this
poses broader questions: Are all parallelizable graph reasoning tasks computable by a transformer of
logarithmic depth and small embedding dimension? And do sub-logarithmic-depth transformers exist
that solve any parallelizable tasks?
We first show that all parallelizable tasks can be solved in two logarithmic-depth scaling settings.
Theorem 2. For any parallelizable task, there exists transformers in LogDepthPause and
LogDepthPause that solve the task.
Theorem 2 is stated formally as Theorem 18 and proved in Appendix B.2.1. Both components of the
theorem are a consequence of a novel relationship between the MPC model and transformers (Theo-
rem 1) and the analysis of MPC protocols for graph reasoning tasks by [60]. The LogDepthPause
result is the direct implication of an O(1)-round MPC equivalence (Theorem 9) between all paralleliz-
able tasks and an MPC protocol that solves the connectivity task (Theorem 11). The LogDepthWide
bound is a consequence of Theorem 15, which shows that all languages in NC2 (including all lan-
guages in L and NL) can be evaluated by an MPC protocol with O(log N ) rounds and with local
memory O(N 1/2+ ) [60].
We further prove the conditional optimality of logarithmic depth.
Theorem 3. Conditional on Conjecture 13, any transformer that solves some parallelizable task
with width mH = O(N ) and pause tokens N 0 = poly(N ) must have depth L = Ω(log N ).
This result (stated formally as Theorem 19 is a combination of the conditional depth lower bound on
graph connectivity by [71] and the O(1)-round equivalence between all parallel tasks.
Search tasks are similarly defined to be those that are NL-complete and equivalent to shortest path in
O(1)-rounds of MPC and include shortest path, strong connectivity, st-reachability, radius, diameter,
and median. Like before, the O(1)-round MPC equivalence translates to an O(1)-depth equivalence
in transformers. We give a similar positive result for LogDepthWide transformers; whether these
tasks can be solved by LogDepthPause transformers is unknown.
Theorem 4. For any search task, there exists a transformer in LogDepthWide that sovles the task.
This theorem, which is restated in Appendix B.2.2 as Theorem 21, is also an immediate consequence
of Theorem 15.
While the minimum depth of a transformer with small embedding dimension that solves a search
task is not identified, we prove that the minimum depth needed to solve some all search task is
approximately equivalent in Theorem 22.
2
MPC is a theoretical model of the MapReduce computational paradigm that distributed computation among
a large number of machines with restricted local memory size and communication bandwidth. We formally
introduce the MPC model in Appendix B.1.
5
3.3 Retrieval tasks and Depth1 transformers
Graph tasks whose algorithms consist of a single look-up or aggregation step can be efficiently solved
by single-layer transformers. This result assumes that the graph G = (V , E) is encoded as some
input sequence X of length N = O(|V | + |E|) that expresses each edge and vertex a single token.
This encoding scheme is detailed in Appendix C.
Theorem 5. For any retrieval task (including node count, edge count, edge existence, and node
degree) there exists a transformer in Depth1 that solves the task.
We formalize this statement in Theorem 36 and prove it their in Appendix C.1. These rely on proving
the existence of a useful input MLP φ that precomputes embeddings with useful structure for all
retrieval tasks.
In contrast, we show that a collection of parallelizable and search tasks cannot be efficiently solved
by transformers in Depth1.
Theorem 6. Any single-layer transformer that solves the graph connectivity, shortest path, or cycle
detection task has width satisfying mH = Ω̃(N ).
The proof of the formal counterpart of this statement (Theorem 38) appears in Appendix C.2 and is
a consequence of a standard communication complexity argument. A more generalized result than
Theorem 6 was proved by [57], which establishes that all problems outside of NC1 —which include
all L-complete and NL-complete languages, and hence, all parallelizable and search tasks—cannot be
solved by constant-depth transformers with polynomial width because they cannot be computed by
TC0 (constant-depth threshold circuits). We nonetheless include this theorem to demonstrate a clean
lower bound that applies to very simple input graph instances.
We finally construct depth-efficient transformers for triangle counting due to the MPC algorithms
of [11]. Unlike previous positive results, which applied uniformly across all graphs instances of
bounded size, the complexity of the corresponding transformers for triangle counting is a function
of the arboricity3 of the input graph. When the arboricity grows sub-polynomially with N —as is
the case for bounded-degree graphs—no pause tokens are necessary. Unlike the parallelizable and
search classes of problems, strictly sub-logarithmic depth is attainable with pause tokens, even for
worst-case graphs.
Theorem 7. There exists a transformer that computes the number of triangles in any input graph of
arboricity α and has embedding dimension m = O(N ), depth L = O(log log N ), and pause tokens
O(αN 1− ) if α = Ω(N )
N0 =
0 otherwise.
This result is a special case of Theorem 23, a more general result about clique counting that appears
in Appendix B.2.3.
Theoretical conclusions. These results provide a tight characterization of the the reasoning capa-
bilities of transformers whose depth, width, and input padding conform to different scaling regimes.
They strengthen the established connection between transformers and massively parallel computation
(MPC) [71] and generalize the resulting representational bounds to broader categories of graph tasks.
We conclude that the logarithmic-depth regime is apt for for considering tasks in L and NL, which
had previous illuminated the limitations of transformers with constant depth and a limited number
of chain-of-thought tokens [56]. While expressivity does not imply learnability, these theoretical
benchmarks sharply characterize the fundamental limitations of transformers and coincide with
experimental results conveyed in the subsequent section.
Our theoretical model of a transformer has certain limitations, namely the universality of the multi-
layer perceptron units. When maximally exploited (e.g. by solving NP-hard problems within each
MLP), this assumption exaggerates the computational capabilities of transformers. However, the
3
The arboricity is the minimum number of spanning forests needed to cover the graph, which grows at most
linearly with the degree of the graph.
6
local computational complexity of the MPC algorithms employed in this paper tend to be strongly
sublinear in the size of the graph input (see, e.g. the connectivity algorithms of Andoni et al. [4]),
which implies the existence of compact ReLU circuits for each of the model’s MLPs. Furthermore,
the ratio of model parameters that belong to MLPs in state-of-the-art language modelsis known to
be high, which suggests that it is reasonable to assume that MLP units in our theoretical model are
algorithmically rich. We would be interested in pursuing further theoretical research that combine
the computational model [54] and the communication model of [71] to assess the capabilities of
transformers whose MLPs are represented as bounded-size boolean circuits.
1. Transformers excel at global reasoning tasks. Transformers outperform GNNs on tasks that
require efficiently aggregating information about distant nodes in a graph, such as connectivity
and shortest path.
2. GNNs uncover local graph structure with few samples. While transformers are capable of
efficiently expressing all graph learning tasks under investigating, the structural limitations of
GNNs provide them with favorable inductive biases for intrinsically local tasks, such as cycle
check and node degree, and permit them to outperform transformers in a low-sample regime.
3. Trained transformers outperform LLM prompting. Transformers trained to explicitly solve
graph reasoning tasks consistently attain greater accuracy across tasks than a variety of prompting
strategies applied to more recent larger LMs.
A comprehensive evaluation of each GraphQA task on every training setting appears in Appendix E,
in addition details about transformer training, the GraphQA benchmark, and alternative GNN and
prompting approaches.
As indicated in Section 3, graph reasoning algorithms can be categorized based on the extent to which
they entail aggregating “local” information about nodes and their immediate neighbors or modeling
“global” connections between nodes separated by a long distances. This section investigates the
following question about transformers and long-distance reasoning:
We consider two tasks that require reasoning across long distances in a graph instance: evaluating
connectivity and computing the shortest path between a pair of nodes. Neither of these tasks can
be solved by only investigating the neighbors of the source and sink node, which therefore implies
that some analysis of global graph structure is necessary.
Figure 3a displays the accuracy of a variety of trained transformers and GNNs on the connectivity task
contrastsing the performance of all such models when trained on 1,000 and 100,000 graph connectivity
instances. In the most restricted sample complexity regime, trained GNNs are consistently more
accurate than the small transformer; however, increasing the number of training samples yields a far
more substantial improvement in the performance of the small transformer, which outperforms all
GNNs trained on 100,000 samples. Notably, the pre-trained transformer, fine-tuned on just 1000
training instances, nearly solves the connectivity task. This suggests significant enhancements due to
the larger model size and the data-rich fine-tuning phase. Figure 3b plots the training and test error
of ten small transformers trained on connectivity datasets of increasing size and reveals a sharp and
7
# of training samples 1
Model 1K 100K
Accuracy
0.95
GCN [44] 83.8 83.8
MPNN [28] 94.0 94.4 Train
GIN [84] 93.8 94.0 0.9 Test
60M transformer 92.9 98.0 102 103 104 105
11B transformer (FT) 98.4 —
Number of Training Samples (log scale)
(a) Connectivity classification accuracy for trained (b) Connectivity classification accuracy of 60M
transformers and GNNs. transformers by training set size.
Figure 3: Accuracy of a variety of trained transformers and GNNs on the connectivity task.
continual improvement in accuracy. The fine-tuned T5 transformer has similar accuracy to the most
sample-rich small transformer and exceeds that of all GNNs.
On the other hand, Table 1 demonstrates that the
MPNN GNN models outperform small trans- # of training samples
formers when trained to compute the shortest Model 1K 100K
path, even on larger training datasets. However,
the fine-tuned T5 model has far higher accu- GCN [44] 50.2 55.0
racy than all alternatives, even when trained only MPNN [28] 66.8 72.6
1000 samples. GIN [84] 54.0 58.6
60M transformer 57.4 97.1
Theoretical interpretation: Because connec- 11B transformer (FT) 92.8 —
tivity is the prototypical example of a task in the
parallelizable class and can thus be efficiently Table 1: Transformers vs GNNs on shortest path:
implemented by LogDepth transformers with Fine-tuned large transformers outperform other
very small width (Theorem 2), the fact that small transformers and GNNs, even the alternatives are
transformers succeed in solving nearly all con- trained on much larger training sets.
nectivity instances is illuminating but not sur-
prising. In contrast, message-passing GNNs are
unable to solve connectivity in a similarly depth- and width-efficient manner due to fundamental
capacity limitations.4
Shortest path belongs to the search class of graph reasoning tasks and is NL-complete. Theorem 4
implies that shortest path is computable by LogDepthWide transformers, which are likely to require
very large embedding dimension to be learnable by finite samples. This task can only be computed in
a depth- and parameter-efficient manner if a variety of search tasks, including all-pairs shortest-path
and diameter, are as well (Theorem 22). The fact that only the pre-trained model has nearly optimal
performance on shortest path reinforces the theoretical intuition that solving shortest path requires a
very large number of model parameters.
While small transformers outperform GNNs on graph reasoning algorithms that entail analysis of
long-range graph structure, their empirical successes are not uniform. Here, we investigate the
following question:
Figure 3b and Table 1 demonstrate that GNNs outperform small transformers in the low-sample
regime, despite the sufficient expressivity of transformers. This gap in performance, which is
reinforced for the node degree and cycle check tasks in Table 2, suggests that GNNs have a beneficial
4
See Appendix A.2 for a discussion of prior work on theoretical relationships between GNNs and both
the Weisfeiler-Leman graph isomorphism test [84] and the C ONGEST distributed computing model and their
implication that GNNs cannot solve connectivity in a parameter-efficient manner.
8
Retrieval tasks Parallelizable Tasks Search Tasks Subgraph Counting
Method Node count Edge count Edge existence Node degree Connectivity Cycle check Shortest path Triangle counting
ZERO - SHOT [24] 21.7 12.4 44.5 14.0 84.9 76.0 11.5 1.5
Prompting
ZERO - COT [24] 14.6 9.4 33.5 10.4 73.5 32.3 33.6 12.7
FEW- SHOT [24] 25.3 12.0 36.8 17.4 79.4 37.4 22.7 3.0
COT [24] 27.6 12.8 42.8 29.2 45.2 58.0 38.6 8.1
COT- BAG [24] 26.9 12.5 37.3 28.0 45.2 52.1 40.4 8.1
60M transformer-1K 100.0 100.0 67.6 31.5 92.9 97.1 57.4 33.4
Ours
60M transformer-100K 100.0 100.0 96.1 91.7 98.0 98.0 97.2 40.5
11B transformer (FT)-1K 100.0 45.0 100.0 68.8 98.4 98.0 92.8 26.0
Table 3: Comparison of GraphQA task accuracies of transformers explicitly trained for graph
reasoning and LLMs with a variety of prompting strategies.
inductive bias for learning graph reasoning tasks that can be solved by attending exclusively to local
heuristics.
Just as the bounded kernel size of con-
Node Degree Cycle Check
volutional neural networks (CNNs)
enables the sample-efficient learning Model 1K 100K 1K 100K
of relevant textures and gradients GCN [44] 9.8 9.4 83.2 83.2
for image analysis, message-passing MPNN [28] 99.4 99.8 99.0 100.0
GNNs are unable to send messages GIN [84] 36.2 37.8 98.8 83.2
instantaneously across multiple edges,
which simplifies a search of the space 60M transformer 31.6 91.7 97.1 98.0
of “one-hop” graph algorithms and 11B transformer (FT) 68.8 — 98.0 —
represents a positive inductive bias. In
contrast, the ability of transformers to Table 2: Transformers vs GNNs on cycle check and node
send information between any pair of degree: GNNs are favorably biased for local structure.
input tokens—and the alternative inductive biases suggested by the input positional encoding—likely
induces a steeper sample complexity to learn node degree.
Theoretical interpretation: While model expressivity is necessary for learnability, it is not suffi-
cient. The locality constraints of message-passing GNNs likely provides a favorable inductive bias
for learning tasks like node degree with an exclusively on local structure that makes learning these
tasks possible in a sample-efficient manner. While cycle check is more representationally difficult
for GNNs than transformers in the worst case (see Appendix A.2), the random graphs sampled for
the GraphQA benchmark have very small cycles (Figure 7) and do not resemble the large-diameter
worst-case instance.
Large language models (LLMs) are regularly evaluated by their reasoning abilities, and it remains
an open research question to determine what kinds of training data best teaches models to solve
logical problems. We investigate the extent to which LLMs can already perform graph reasoning
tasks without being trained explicitly to do so.
In Table 3, we contrast the capabilities of trained transformer models with several prompt-based ap-
proaches to querying LLMs. Task-specific transformers—including the fine-tuned 11B transformers—
consistently dominated the prompt-based approaches, despite the vast difference in parameter count
and the almost certain presence of graph reasoning in the LLM’s corpus.
9
5 Conclusion
This paper provides a comprehensive evaluation of transformer models’ graph reasoning capabilities,
shedding light on their effectiveness across diverse graph reasoning tasks. By introducing a novel
representational hierarchy, the study distinguishes between retrieval, parallelizable, and search
reasoning tasks and offers insights into the performance of transformers at varying levels of granularity.
The empirical investigation reveals that transformers exhibit strong performance in graph-based
reasoning problems, often matching or surpassing specialized graph models. Furthermore, the study
highlights transformers’ exceptional ability to capture global graph patterns effectively, showcasing
their capability in understanding long-range dependencies, a critical factor in solving tasks involving
global graph structures. Overall, this work crystallizes precise representational trade-offs that reflect
the fundamental reasoning capabilities of transformers and demonstrates that the tasks used to
quantify those capabilities are indeed learnable in a sample-efficient and parameter-efficient manner.
While the hierarchy introduced by this work effectively separates graph algorithmic tasks into distinct
equivalence classes with significant implications for their computability by transformers, several
questions remain for future research. We focused on graph reasoning tasks due to their relevance
to the broader context of transformers, GNNs, and parallel algorithms. However, the complexity
classes presented here could potentially be extended to a wider range of algorithmic problems.
While our assumption of unbounded-size MLPs provides strong lower bounds, further research into
whether parallelizable tasks can be represented by transformers with bounded-size MLP units would
complement this existing work. Broader experimental results that empirically evaluate the scaling
laws would more directly assess the relevance of representational theoretical results to learnability.
References
[1] Anders Aamand, Justin Y. Chen, Piotr Indyk, Shyam Narayanan, Ronitt Rubinfeld, Nicholas
Schiefer, Sandeep Silwal, and Tal Wagner. Exponentially improving the complexity of simulat-
ing the weisfeiler-lehman test with graph neural networks, 2022. Cited on page 3.
[2] Ralph Abboud, Ismail Ilkan Ceylan, Martin Grohe, and Thomas Lukasiewicz. The surprising
power of graph neural networks with random node initialization, 2021. URL https://arxiv.
org/abs/2010.01179. Cited on page 3.
[3] Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni
Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. GPT-4
technical report. arXiv preprint arXiv:2303.08774, 2023. Cited on page 1.
[4] Alexandr Andoni, Zhao Song, Clifford Stein, Zhengyu Wang, and Peilin Zhong. Parallel graph
connectivity in log diameter rounds. In FOCS, October 2018. Cited on pages 7, 17, and 18.
[5] Rohan Anil, Andrew M Dai, Orhan Firat, Melvin Johnson, Dmitry Lepikhin, Alexandre Passos,
Siamak Shakeri, Emanuel Taropa, Paige Bailey, Zhifeng Chen, et al. Palm 2 technical report.
arXiv preprint arXiv:2305.10403, 2023. Cited on page 42.
[6] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointly
learning to align and translate. In ICLR, 2014. Cited on page 1.
[7] Peter W Battaglia, Jessica B Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius
Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan
Faulkner, et al. Relational inductive biases, deep learning, and graph networks. arXiv preprint
arXiv:1806.01261, 2018. Cited on page 1.
[8] Paul Beame, Paraschos Koutris, and Dan Suciu. Communication steps for parallel query
processing. JACM, 64(6), oct 2017. Cited on page 19.
[9] James Bergstra and Yoshua Bengio. Random search for hyper-parameter optimization. JMLR,
13(2), 2012. Cited on page 41.
[10] Maciej Besta, Nils Blach, Ales Kubicek, Robert Gerstenberger, Michal Podstawski, Lukas
Gianinazzi, Joanna Gajda, Tomasz Lehmann, Hubert Niewiadomski, Piotr Nyczyk, and Torsten
Hoefler. Graph of thoughts: Solving elaborate problems with large language models. AAAI, 38
(16):17682–17690, March 2024. Cited on page 1.
10
[11] Amartya Shankha Biswas, Talya Eden, Quanquan C. Liu, Slobodan Mitrović, and Ronitt
Rubinfeld. Massively parallel algorithms for small subgraph counting, 2022. Cited on pages 6
and 20.
[12] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal,
Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are
few-shot learners. NeurIPS, 2020. Cited on pages 1 and 42.
[13] Quentin Cappart, Didier Chételat, Elias B Khalil, Andrea Lodi, Christopher Morris, and Petar
Veličković. Combinatorial optimization and reasoning with graph neural networks. JMLR, 24
(130):1–61, 2023. Cited on page 1.
[14] Ziwei Chai, Tianjie Zhang, Liang Wu, Kaiqiao Han, Xiaohai Hu, Xuanwen Huang, and Yang
Yang. GraphLLM: Boosting graph reasoning ability of large language model, 2023. Cited on
page 3.
[15] Ines Chami, Sami Abu-El-Haija, Bryan Perozzi, Christopher Ré, and Kevin Murphy. Machine
learning on graphs: A model and comprehensive taxonomy. JMLR, 23(89):1–64, 2022. Cited
on page 1.
[16] Sam Coy and Artur Czumaj. Deterministic massively parallel connectivity. In STOC, STOC
’22, June 2022. Cited on page 19.
[17] Antonia Creswell, Murray Shanahan, and Irina Higgins. Selection-inference: Exploiting large
language models for interpretable logical reasoning. In ICLR, 2023. Cited on page 1.
[18] Artur Back de Luca and Kimon Fountoulakis. Simulation of graph algorithms with looped
transformers, 2024. URL https://arxiv.org/abs/2402.01107. Cited on page 3.
[19] Jeffrey Dean and Sanjay Ghemawat. Mapreduce: Simplified data processing on large clusters. In
OSDI’04: Sixth Symposium on Operating System Design and Implementation, pages 137–150,
San Francisco, CA, 2004. Cited on page 18.
[20] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of
deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805,
2018. Cited on page 1.
[21] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai,
Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al.
An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021.
Cited on page 1.
[22] Vijay Prakash Dwivedi and Xavier Bresson. A generalization of transformer networks to graphs.
arXiv preprint arXiv:2012.09699, 2020. Cited on page 3.
[23] Paul Erdős and Alfred Rényi. On random graphs. Publicationes Mathematicae Debrecen, 6:
290–297, 1959. Cited on page 41.
[24] Bahare Fatemi, Jonathan Halcrow, and Bryan Perozzi. Talk like a graph: Encoding graphs for
large language models. In ICLR, 2024. Cited on pages 3, 7, 9, 40, 41, 42, and 43.
[25] Fabian Frei and Koichi Wada. Efficient circuit simulation in mapreduce. ArXiv, abs/1907.01624,
2019. Cited on page 20.
[26] Roy Frostig, Matthew James Johnson, and Chris Leary. Compiling machine learning programs
via high-level tracing. Systems for Machine Learning, 4(9), 2018. Cited on page 41.
[27] Mohsen Ghaffari, Fabian Kuhn, and Jara Uitto. Conditional hardness results for massively
parallel computation from distributed lower bounds. In FOCS, 2019. Cited on page 19.
[28] Justin Gilmer, Samuel S Schoenholz, Patrick F Riley, Oriol Vinyals, and George E Dahl. Neural
message passing for quantum chemistry. In ICML, 2017. Cited on pages 1, 8, 9, 42, and 43.
11
[29] Sachin Goyal, Ziwei Ji, Ankit Singh Rawat, Aditya Krishna Menon, Sanjiv Kumar, and
Vaishnavh Nagarajan. Think before you speak: Training language models with pause tokens. In
ICLR, 2024. Cited on pages 2, 3, and 16.
[30] Alex Graves, Greg Wayne, and Ivo Danihelka. Neural turing machines. arXiv preprint
arXiv:1410.5401, 2014. Cited on page 1.
[31] Yiding Hao, Dana Angluin, and Robert Frank. Formal language recognition by hard attention
transformers: Perspectives from circuit complexity. TACL, 10:800–810, 2022. Cited on page 2.
[32] Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Trevor Cai, Eliza
Rutherford, Diego de Las Casas, Lisa Anne Hendricks, Johannes Welbl, Aidan Clark, et al.
Training compute-optimal large language models. arXiv preprint arXiv:2203.15556, 2022.
Cited on page 1.
[33] Bowen Jin, Gang Liu, Chi Han, Meng Jiang, Heng Ji, and Jiawei Han. Large language models
on graphs: A comprehensive survey. arXiv preprint arXiv:2312.02783, 2023. Cited on page 3.
[34] Armand Joulin and Tomas Mikolov. Inferring algorithmic patterns with stack-augmented
recurrent nets. NIPS, 2015. Cited on page 1.
[35] Norman P. Jouppi, Cliff Young, Nishant Patil, David Patterson, Gaurav Agrawal, Raminder
Bajwa, Sarah Bates, Suresh Bhatia, Nan Boden, Al Borchers, Rick Boyle, Pierre-luc Cantin,
Clifford Chao, Chris Clark, Jeremy Coriell, Mike Daley, Matt Dau, Jeffrey Dean, Ben Gelb,
Tara Vazir Ghaemmaghami, Rajendra Gottipati, William Gulland, Robert Hagmann, C. Richard
Ho, Doug Hogberg, John Hu, Robert Hundt, Dan Hurt, Julian Ibarz, Aaron Jaffey, Alek Jaworski,
Alexander Kaplan, Harshit Khaitan, Daniel Killebrew, Andy Koch, Naveen Kumar, Steve Lacy,
James Laudon, James Law, Diemthu Le, Chris Leary, Zhuyuan Liu, Kyle Lucke, Alan Lundin,
Gordon MacKean, Adriana Maggiore, Maire Mahony, Kieran Miller, Rahul Nagarajan, Ravi
Narayanaswami, Ray Ni, Kathy Nix, Thomas Norrie, Mark Omernick, Narayana Penukonda,
Andy Phelps, Jonathan Ross, Matt Ross, Amir Salek, Emad Samadiani, Chris Severn, Gregory
Sizikov, Matthew Snelham, Jed Souter, Dan Steinberg, Andy Swing, Mercedes Tan, Gregory
Thorson, Bo Tian, Horia Toma, Erick Tuttle, Vijay Vasudevan, Richard Walter, Walter Wang,
Eric Wilcox, and Doe Hyun Yoon. In-datacenter performance analysis of a tensor processing
unit. SIGARCH Comput. Archit. News, 2017. Cited on page 41.
[36] John Jumper, Richard Evans, Alexander Pritzel, Tim Green, Michael Figurnov, Olaf Ron-
neberger, Kathryn Tunyasuvunakool, Russ Bates, Augustin Žídek, Anna Potapenko, et al.
Highly accurate protein structure prediction with alphafold. Nature, 596(7873):583–589, 2021.
Cited on page 1.
[37] Lukasz Kaiser and Ilya Sutskever. Neural GPUs learn algorithms. In ICLR, 2016. Cited on
page 1.
[38] Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B Brown, Benjamin Chess, Rewon Child,
Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. Scaling laws for neural language
models. arXiv preprint arXiv:2001.08361, 2020. Cited on page 1.
[39] Howard Karloff, Siddharth Suri, and Sergei Vassilvitskii. A Model of Computation for MapRe-
duce, pages 938–948. Cited on pages 5 and 18.
[40] Nora Kassner, Oyvind Tafjord, Ashish Sabharwal, Kyle Richardson, Hinrich Schuetze, and
Peter Clark. Language models with rationality. In Houda Bouamor, Juan Pino, and Kalika Bali,
editors, EMNLP, December 2023. Cited on page 1.
[41] Mehran Kazemi, Najoung Kim, Deepti Bhatia, Xin Xu, and Deepak Ramachandran. LAM-
BADA: Backward chaining for automated reasoning in natural language. In ACL, 2023. Cited
on page 1.
[42] Jinwoo Kim, Tien Dat Nguyen, Seonwoo Min, Sungjun Cho, Moontae Lee, Honglak Lee, and
Seunghoon Hong. Pure transformers are powerful graph learners, 2022. Cited on pages 2, 3,
and 31.
12
[43] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint
arXiv:1412.6980, 2014. Cited on page 41.
[44] Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional
networks. arXiv preprint arXiv:1609.02907, 2016. Cited on pages 1, 8, 9, 42, and 43.
[45] Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa. Large
language models are zero-shot reasoners. NeurIPS, 35:22199–22213, 2022. Cited on page 42.
[46] Nayoung Lee, Kartik Sreenivasan, Jason D Lee, Kangwook Lee, and Dimitris Papailiopoulos.
Teaching arithmetic to small transformers. In ICLR, 2024. Cited on pages 1 and 3.
[47] Yuxuan Li and James L. McClelland. Systematic generalization and emergent structures in
transformers trained on structured tasks, 2022. Cited on page 3.
[48] Bingbin Liu, Jordan T. Ash, Surbhi Goel, Akshay Krishnamurthy, and Cyril Zhang. Transform-
ers learn shortcuts to automata, 2022. Cited on page 3.
[49] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining
Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In CVPR, 2021.
Cited on page 1.
[50] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. arXiv preprint
arXiv:1711.05101, 2017. Cited on page 41.
[51] Andreas Loukas. What graph neural networks cannot learn: depth vs width. In ICLR, 2020.
Cited on pages 1, 3, and 39.
[52] Eran Malach. Auto-regressive next-token predictors are universal learners, 2023. Cited on page
2.
[53] Amil Merchant, Simon Batzner, Samuel S Schoenholz, Muratahan Aykol, Gowoon Cheon, and
Ekin Dogus Cubuk. Scaling deep learning for materials discovery. Nature, 624(7990):80–85,
2023. Cited on page 1.
[54] William Merrill and Ashish Sabharwal. A logic for expressing log-precision transformers, 2022.
Cited on pages 2 and 7.
[55] William Merrill and Ashish Sabharwal. The parallelism tradeoff: Limitations of log-precision
transformers. TACL, 11:531–545, 2023. Cited on pages 2 and 32.
[56] William Merrill and Ashish Sabharwal. The expressive power of transformers with chain of
thought, 2023. Cited on pages 2 and 6.
[57] William Merrill, Ashish Sabharwal, and Noah A. Smith. Saturated transformers are constant-
depth threshold circuits, 2021. Cited on pages 2 and 6.
[58] Erxue Min, Runfa Chen, Yatao Bian, Tingyang Xu, Kangfei Zhao, Wenbing Huang, Peilin
Zhao, Junzhou Huang, Sophia Ananiadou, and Yu Rong. Transformer for graphs: An overview
from architecture perspective, 2022. Cited on page 3.
[59] Christopher Morris, Yaron Lipman, Haggai Maron, Bastian Rieck, Nils M Kriege, Martin Grohe,
Matthias Fey, and Karsten Borgwardt. Weisfeiler and Leman go machine learning: The story so
far. JMLR, 24(1):15865–15923, 2023. Cited on pages 3 and 40.
[60] Danupon Nanongkai and Michele Scquizzato. Equivalence classes and conditional hardness in
massively parallel computations. Distributed Computing, 35(2):165–183, January 2022. Cited
on pages 5, 18, 19, and 20.
[61] Maxwell Nye, Anders Johan Andreassen, Guy Gur-Ari, Henryk Michalewski, Jacob Austin,
David Bieber, David Dohan, Aitor Lewkowycz, Maarten Bosma, David Luan, et al. Show
your work: Scratchpads for intermediate computation with language models. arXiv preprint
arXiv:2112.00114, 2021. Cited on page 2.
13
[62] Jorge Pérez, Pablo Barceló, and Javier Marinkovic. Attention is turing complete. J. Mach.
Learn. Res., 22(1), jan 2021. ISSN 1532-4435. Cited on pages 2 and 3.
[63] Bryan Perozzi, Bahare Fatemi, Dustin Zelle, Anton Tsitsulin, Mehran Kazemi, Rami Al-Rfou,
and Jonathan Halcrow. Let your graph do the talking: Encoding structured data for LLMs, 2024.
Cited on pages 3, 42, and 43.
[64] Jackson Petty, Sjoerd van Steenkiste, Ishita Dasgupta, Fei Sha, Dan Garrette, and Tal Linzen.
The impact of depth and width on transformer language model generalization. NAACL, 2024.
Cited on page 1.
[65] Jacob Pfau, William Merrill, and Samuel R Bowman. Let’s think dot by dot: Hidden computation
in transformer language models. arXiv preprint arXiv:2404.15758, 2024. Cited on pages 2
and 16.
[66] Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena,
Yanqi Zhou, Wei Li, and Peter J Liu. Exploring the limits of transfer learning with a unified
text-to-text transformer. JMLR, 21(1):5485–5551, 2020. Cited on pages 1, 7, and 42.
[67] Machel Reid, Nikolay Savinov, Denis Teplyashin, Dmitry Lepikhin, Timothy Lillicrap, Jean-
baptiste Alayrac, Radu Soricut, Angeliki Lazaridou, Orhan Firat, Julian Schrittwieser, et al.
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv
preprint arXiv:2403.05530, 2024. Cited on page 1.
[68] Yu Rong, Yatao Bian, Tingyang Xu, Weiyang Xie, Ying Wei, Wenbing Huang, and Junzhou
Huang. Self-supervised graph transformer on large-scale molecular data. NeurIPS, 33, 2020.
Cited on page 3.
[69] Tim Roughgarden, Sergei Vassilvitskii, and Joshua R. Wang. Shuffles and circuits (on lower
bounds for modern parallel computation). J. ACM, 65(6), nov 2018. Cited on page 19.
[70] Clayton Sanford, Daniel Hsu, and Matus Telgarsky. Representational strengths and limitations
of transformers, 2023. Cited on pages 3, 31, 32, and 36.
[71] Clayton Sanford, Daniel Hsu, and Matus Telgarsky. Transformers, parallel computation, and
logarithmic depth, 2024. Cited on pages 2, 3, 5, 6, 7, 16, 18, 21, 22, 23, 31, 32, and 39.
[72] Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini.
The graph neural network model. IEEE transactions on neural networks, 20(1):61–80, 2008.
Cited on page 1.
[73] Noam Shazeer. GLU variants improve transformer. arXiv preprint arXiv:2002.05202, 2020.
Cited on page 41.
[74] Kaya Stechly, Matthew Marquez, and Subbarao Kambhampati. GPT-4 doesn’t know it’s wrong:
An analysis of iterative prompting for reasoning problems, 2023. Cited on page 3.
[75] Gemini Team, Rohan Anil, Sebastian Borgeaud, Yonghui Wu, Jean-Baptiste Alayrac, Jiahui Yu,
Radu Soricut, Johan Schalkwyk, Andrew M Dai, Anja Hauth, et al. Gemini: a family of highly
capable multimodal models. arXiv preprint arXiv:2312.11805, 2023. Cited on page 1.
[76] Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei,
Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, et al. Llama 2: Open
foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023. Cited on page
1.
[77] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez,
Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In NeurIPS, 2017. Cited on
pages 1, 16, and 42.
[78] Petar Veličković and Charles Blundell. Neural algorithmic reasoning. Patterns, 2(7), 2021.
Cited on page 1.
14
[79] Petar Veličković, Adrià Puigdomènech Badia, David Budden, Razvan Pascanu, Andrea Banino,
Misha Dashevskiy, Raia Hadsell, and Charles Blundell. The CLRS algorithmic reasoning
benchmark. In ICML, 2022. Cited on page 1.
[80] Heng Wang, Shangbin Feng, Tianxing He, Zhaoxuan Tan, Xiaochuang Han, and Yulia Tsvetkov.
Can language models solve graph problems in natural language? In NeurIPS, 2023. Cited on
pages 3 and 42.
[81] Colin Wei, Yining Chen, and Tengyu Ma. Statistically meaningful approximation: a case study
on approximating turing machines with transformers, 2021. Cited on page 2.
[82] Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le,
Denny Zhou, et al. Chain-of-thought prompting elicits reasoning in large language models.
NeurIPS, 2022. Cited on pages 1, 2, 16, and 42.
[83] Boris Weisfeiler and Andrei Leman. The reduction of a graph to canonical form and the algebra
which appears therein. nti, Series, 2(9):12–16, 1968. Cited on pages 3 and 40.
[84] Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. How powerful are graph neural
networks?, 2018. Cited on pages 1, 3, 8, 9, 40, 42, and 43.
[85] Andrew Chi-Chih Yao. Some complexity questions related to distributive computing(preliminary
report). In STOC, page 209–213, New York, NY, USA, 1979. Cited on page 36.
[86] Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Tom Griffiths, Yuan Cao, and Karthik
Narasimhan. Tree of thoughts: Deliberate problem solving with large language models. NeurIPS,
2024. Cited on page 1.
[87] Ruosong Ye, Caiqi Zhang, Runhui Wang, Shuyuan Xu, and Yongfeng Zhang. Language is all a
graph needs, 2023. Cited on page 3.
[88] Chulhee Yun, Srinadh Bhojanapalli, Ankit Singh Rawat, Sashank J. Reddi, and Sanjiv Kumar.
Are transformers universal approximators of sequence-to-sequence functions?, 2019. Cited on
page 2.
[89] Wojciech Zaremba and Ilya Sutskever. Learning to execute. In ICLR, 2015. Cited on page 1.
[90] Bohang Zhang, Shengjie Luo, Liwei Wang, and Di He. Rethinking the expressive power of
GNNs via graph biconnectivity. In ICLR, 2023. Cited on pages 1 and 3.
[91] Yi Zhang, Arturs Backurs, Sébastien Bubeck, Ronen Eldan, Suriya Gunasekar, and Tal Wagner.
Unveiling transformers with lego: a synthetic reasoning task, 2022. Cited on page 3.
[92] Hattie Zhou, Azade Nova, Hugo Larochelle, Aaron Courville, Behnam Neyshabur, and
Hanie Sedghi. Teaching algorithmic reasoning via in-context learning. arXiv preprint
arXiv:2211.09066, 2022. Cited on page 1.
[93] Hattie Zhou, Arwen Bradley, Etai Littwin, Noam Razin, Omid Saremi, Josh Susskind, Samy
Bengio, and Preetum Nakkiran. What algorithms can transformers learn? a study in length
generalization. In ICLR, 2024. Cited on page 1.
A Theoretical preliminaries
For some vector v ∈ RN , let
1
softmax(v) = PN (exp(v1 ), . . . , exp(vN )) ∈ RN .
i=1 exp(vi )
Let ei ∈ Rm denote the ith elementary vector, ~1 = (1, . . . , 1) ∈ Rm and ~0 = (0, . . . , 0) ∈ Rm . Let
[n] = {1, . . . , n}.
15
A.1 Transformer models
We use a similar theoretical definition of a multi-layer transformer model to the one employed by
[71]. We use a bounded communication theoretical model of the standard bidirectional transformer
of [77], which assumes that the principal limitation of the transformer’s representational power is
the low rank of the self-attention matrix. Several theoretical assumptions are necessary to model
transformers in this regime:
• the interleaved element-wise multi-layer perceptron (MLP) units compute arbitrary func-
tions;
• the embedding dimension m and number of heads per layer H are much smaller than the
sequence length N ; and
• all model parameters, inputs, and intermediate products can be represented with O(log N )-
bit numbers.
• The initial embedding is computed by applying the first MLP φ0 to all elements of the input
X, along with N 0 “blank” inputs:
Y 0 = φ0 (X) = (φ0 (x1 , 1), . . . , φ0 (xN , N ), φ0 (0, N + 1), . . . , φ0 (0, N + N 0 )).
16
• The output f (X) ∈ RN is computed to be the first N outputs of ψ(Y L ). That is,
While our theoretical results pertain to the bidirectional regime (where no masking is applied to
the self-attention layer), all results in Appendix C and all lower bounds in Appendix B apply to
autoregressive transformers as well. Our empirical results utilize causal masking.
Furthermore, an L-layer bidirectional transformer that operates on sequences of length N with N 0
pause tokens can be simulated by and L-layer autoregressive transformer with O(L · (N + N 0 )) pause
tokens. In the regime where N 0 = poly(N ) and L = O(log N ), any positive results for bidirectional
models likewise apply to autoregressive models.
We exclusively consider graph reasoning tasks that apply to undirected and unweighted graphs
G = (V , E) of bounded size, i.e. |V | + |E| = O(N ) for some size parameter N . We define the
tasks used for experimentation as follows:
These tasks are foundational to computer science and are solved by famous polynomial-time algo-
rithms, such as Dijkstra’s algorithm for shortest path and Kruskal’s for spanning trees. Several of
these tasks reap significant algorithmic benefits from parallel computing models, including graph
connectivity, which can be evaluated in a logarithmic number of rounds by employing an “iterative
merging” strategy [4].
The amenability of some of these tasks to parallel computing is the core principle underlying our
theoretical results on transformer model capacity and the resulting contrasts with GNNs. For instance,
the iterative merging algorithm for connectivity can be simulated by a logarithmic-depth transformer,
but message-passing GNNs cannot do without a substantial width scaling (see Appendix D). Further-
more, this parallel computing algorithm is widely believed to be optimal, and a crystallization of this
as Conjecture 13 is the bedrock of our transformer optimality conjectures.
The multi-layer results of Appendix B are representation-agnostic; the bounds apply to any fixed
encoding of vertices and edges as a transformer input sequence X ∈ RN ×d . The input representation
for Appendix C must satisfy the more specified node/edge encoding scheme, which represents
each vertex and edge as a single embedding, followed by a query embedding (with optional blank
embeddings as needed). This encoding scheme is reflected by Figure 1 and closely resembles the
graph instance encoding used for the experimental results6 .
Throughout the paper, we partition these and other graph reasoning tasks into three representational
categories: retrieval, parallelizable, and search tasks. The retrieval category contains only four of the
tasks used for experiments:
6
The empirical encoding scheme uses two tokens—not one—to encode each edge and has fixed size N .
For all graphs whose encodings do not require all N tokens, the remaining tokens are blank placeholders that
precede the query tokens.
17
• Retrieval tasks include node count, edge count, edge existence, and node degree.
On the other hand, the other categories reflect two equivalence classes introduce by [60]:
• Parallelizable tasks are defined to be L-complete and equivalent to connectivity in O(1)
rounds of MPC. These tasks include connectivity, cycle check, planarity testing, minimum
cut, bipartiteness testing, minimum spanning forest, connected components, one-cycle
versus two-cycles (see Conjecture 13), and # connected components.
• Search tasks are defined to be NL-complete and equivalent to shortest path in O(1) rounds
of MPC. These tasks include shortest path, strong connectivity (for directed graphs), all-pairs
shortest path, median, diameter, radius, directed cycle detection, and st-reachability.
Triangle counting does not appear in any of these classes and is separately analyzed in Section 3.4.
The massively parallel computation (MPC) model of [39] formalizes distributed computing frame-
works such as MapReduce [19] as theoretical models that are amenable to rigorous analysis. MPC
pertains to a regime where an input that consists of n “words” is distributed across a very large number
of machines q (e.g. q ≈ n0.95 ), each of which contains a bounded local memory s (e.g. s ≈ n0.1 )
where computation on individual machines is inexpensive but communication between machines is
costly. We use the definition of MPC of [4], which quantifies the complexity of a protocol by the
local memory s = O(nδ ), the global memory sq = O(n1+γ ), and the number of communication
rounds R.
Definition 2. For global and local memory constants γ, δ > 0 and input size n, an R-round (γ, δ)-
MPC protocol specifies a distributed computing protocol over q = Θ(n1+γ−δ ) machines, each of
which contains a local memory of size s = O(nδ ).
• An input to the protocol, which is encoded as a length-n sequence of p = Θ(log n)-bit
words, is distributed across the local memories of the first dn/se machines.
• In each of the R rounds, each machine computes an arbitrary function its local memory at
the time, which specifies at most s words to be transmitted to other machines.
• Messages are simultaneously transmitted (subject to the constraint that each machine sends
and receives at most s words of information), and each machine’s new local memory is set
equal to the messages received.
18
• After the final round, the output of the protocol is a concatenation of the local memories of
the first dn/se machines.
We say that an MPC protocol computes some function f : Zn2p → Zn2p if for any X ∈ Zn2p encoded
as input to the protocol, the output of the protocol is f (X).
In this section, we introduce previously known results about the abilities of Massively Parallel
Computation protocols to solve graph reasoning tasks. The novel results presented in the subsequent
section are the primarily the immediate implications of these results and Theorem 8.
MPC round equivalence for parallelizable and search tasks These results largely reflect the
contributions of [60], which establish a hierarchy of graph reasoning tasks that is reflected in our
distinction between parallelizable and search tasks. They show that all parallelizable tasks are
equivalent to connectivity in O(1) rounds of MPC and that all search tasks are likewise equivalent to
shortest path. (See their Figure 1.) Concretely, they prove the following for all graphs G = (V , E)
with |V | + |E| = O(n).
Theorem 9 (Theorem 4 of [60]; round-equivalence of parallelizable tasks). If there exists an R-
round (γ, δ)-MPC protocol for any parallelizable task—including graph connectivity, st-connectivity,
cycle detection, formula evaluation, planarity testing, graph bipartiteness, connected components,
minimum spanning forest, and minimum cut, one-cycle versus two-cycle testing—for some γ > 0 and
δ ∈ (0, 1), then there exists an R0 -round (γ 0 , δ)-MPC protocol for any other other parallelizable task
where R0 = R + O(1) and γ 0 = max(γ, 3).
Theorem 10 (Theorem 5 of [60]; round-equivalence of search tasks). If there exists an R-round
(γ, δ)-MPC protocol for any search task—including st-reachability (for directed graphs), strong
connectivitity, directed cycle detection, shortest path, all-pairs shortest path, diameter, radius, and
median—for some γ > 0 and δ ∈ (0, 1), then there exists an R0 -round (γ 0 , δ)-MPC protocol for any
other other search task where R0 = R + O(1) and γ 0 = max(γ, 2).
The equivalance of Theorem 9 has more immediate implications for the round complexity of paralleliz-
able tasks than Theorem 10 does for search tasks because the round complexity of graph connectivity
is well-understood to be O(log n) and thought to be optimal. We first present a deterministic MPC
positive result for graph connectivity.
Theorem 11 (Theorem 6.2 of [16]; log-round connectivity MPC algorithm). For any γ > 0 and
δ ∈ (0, 1), there exists a deterministic O(log n)-round (γ, δ)-MPC protocol that solves graph
connectivity on any graph G = (V , E) of size |V | + |E| = O(n).
Hence, all parallelizable tasks can be solved with a logarithmic number of MPC rounds with arbitrarily
small polynomial local memory.
Corollary 12 (Log-round parallelizable MPC algorithms). For any γ > 0 and δ ∈ (0, 1) and any
parallelizable task, there exists a deterministic O(log n)-round (γ 0 , δ)-MPC protocol that solves the
task on any graph G = (V , E) of size |V | + |E| = O(n) for γ 0 = max(γ, 3).
The optimality of these logarithmic-round protocols is suggested by a conjecture about the hardness
of distinguishing between two graphs with n vertices and n edges, one of whose edges are arranged
in a single cycle of length n and the other in two disjoint cycles of length n2 . This is the well-known
“one-cycle versus two-cycle conjecture,” which is widely employed as a condition for distributed
computing hardness [see e.g. 8, 69, 27].
Conjecture 13 (One-cycle versus two-cycle conjecture, see e.g. [27]). For any γ > 0 and δ ∈ (0, 1),
any (γ, δ)-MPC protocol that distinguishes a single length-n cycle from two disjoint length- n2 cycles
uses R = Ω(log n) rounds.
Under this conjecture, Theorem 9 immediately implies the optimality of Corollary 12.
Corollary 14 (Optimality of log-round parallelizable MPC algorithms). Conditional on Conjec-
ture 13, for any γ > 0 and δ ∈ (0, 1), any (γ, δ)-MPC protocol that solves a parallelizable graph
task on all graphs G = (V , E) of size |V | + |E| = O(n) uses R = Ω(log n) rounds.
The round complexity of search tasks is less understood, and it is unknown if search tasks can be
solved by O(log n)-round (γ, δ)-MPC protocols if δ ∈ (0, 12 ].
19
MPC protocols for problems with bounded circuit size More general MPC constructions are
known for problems that solved by bounded-size boolean circuits, which include both parallelizable
and search tasks. The well-known NCi classes of Boolean circuits that take N inputs and have poly(n)
gates and depth O(logi n) have been shown to be computable by bounded-round MapReduce-like
computational protocols [25] and by MPC protocols in particular [60].
Theorem 15 (Theorem 1 of [60]; log-round circuit MPC algorithms). For any problem in NCi+1
and any γ > 0 and δ ∈ ( 12 , 1), there exists a deterministic O(logi n)-round (γ, δ)-MPC protocol that
solves the problem.
Since L and NL are known to belong to NC2 , the following corollary is an immediate consequence.
Corollary 16 (Log-round parallelizable and search MPC algorithms with high local memory). For
any parallelizable or search graph reasoning task and any γ > 0 and δ ∈ ( 21 , 1), there exists a
deterministic O(log n)-round (γ, δ)-MPC protocol that solves the task on all graphs G = (V , E) of
size |V | + |E| = O(N ).
MPC protocol for triangle counting Finally, the triangle counting task can be solved in the MPC
framework by utilizing a special case of a parallel algorithm for clique counting. These pertain to
graphs with bounded arboricity α, a quantity that corresponds to the branching factor of a node that
is bounded by the degree of the graph; these apply to arbitrary graphs by noting that α ≤ |V |.
Theorem 17 (Theorem 1.5 of [11]; loglog-round triangle counting MPC algorithm). For any k ≥ 3,
δ ∈ (0, 1), and γ > 0, there exists an O(log log n)-round (γ, δ)-MPC protocol that computes
the number of k-cliques in any graph G = (V , E) with |V | + |E| = O(n) and arboricity α =
O(nγ/(k−2) ).
B.2 Positive and negative graph reasoning results for multi-layer transformers
We state formalized versions of the statements of Sections 3.1, 3.2 and 3.4 and prove them by invoking
Theorem 8 jointly with the relationships between MPC and graph reasoning of Appendix B.1.1.
For the duration of this section, the class of parallelizable tasks includes all of those that are deemed
equivalent to graph connectivity in O(1) rounds of MPC by [60], as stated in Theorem 9.
We first present a statement that formalizes the existence of logarithmic-depth transformer construc-
tions for solving parallelizable tasks.
Theorem 18 (Formal version of Theorem 2; log-depth transformers compute parallelizable tasks).
,N 0
For any ∈ (0, 1) and any parallelizable task, there exists a transformer f ∈ TransformerN m,H,L
such that f (X) computes the task where X ∈ RN is some encoding of any graph G = (V , E) with
|V | + |E| = O(N ) and f has depth L = O(log N ) and heads H = O(1) and embedding dimension
m and pause tokens N 0 satisfying either
0
• LogDepthPause: m = O(N ) and N 0 = O(N 4− ), where 0 ∈ (0, ); or
• LogDepthWide: m = O(N ) and N 0 = 0, if > 21 .
Proof. The first bullet is an immediate implication of Corollary 12 and Theorem 8 with γ = 3,
δ 0 = , and δ = 0 . The second bullet follows from Corollary 16 and Theorem 8 with γ = δ = 2 and
δ 0 = .
We then establish that sub-logarithmic-depth solutions to any parallelizable task are impossible
without having linear embedding dimension m or super-polynomial number of pause tokens N 0 ,
under the assumption that the one-cycle versus two-cycle conjecture holds.
Theorem 19 (Formal version of Theorem 3; log-depth optimality for parallelizable tasks). Condi-
tional on Conjecture 13, for any ∈ (0, 1) and γ > 0 and any parallelizable task, if there exists a
20
0
transformer f ∈ TransformerN ,N
m,H,L that solves the task and has width mH = O(N ) and pause
tokens N + N 0 = O(N 1+γ ), then its depth satisfies L = Ω(log N ).
Proof. The proof is a consequence of Corollary 14 and a result of [71] that proves the to simulate
transformers with MPC protocols, an inversion of Theorem 8. We restate that result as follows.
Theorem 20 (Theorem 3.4 of [71]; MPC simulates transformers). Fix any transformer f ∈
,N 0
TransformerN δ
m,H,L with width mH = O(N ) for some δ ∈ (0, 1) and total sequence length
0
N + N = O(N 1+γ
) for some γ ≥ 0. Then, for any δ 0 ∈ (δ, 1) and γ 0 = 1 + 2γ + δ 0 , there
L(1+γ)
exists an O( δ0 −δ )-round (γ 0 , δ 0 )-MPC protocol that simulates f .
0
Consider a transformer f ∈ TransformerN ,N
m,L,H that solves the task with width mH = O(N ) and
total sequence length N + N 0 = O(N 1+γ
√ √ for some constant ∈ (0, 1) and γ > 1. Then, there
exists an O,γ (L)-round (1 + 2γ + , )-MPC protocol that solves the parallelizable task. If
Conjecture 13 is true, then L = Ω(log N ).
We present the main result of Section 3.2 and show an equivalence between the minimum depth
transformer needed to solve search tasks in a width-efficient manner. As before, the class of searchable
tasks includes tasks that are equivalent to shortest path in O(1) MPC rounds (Theorem 10).
Theorem 21 (Formal version of Theorem 4; LogDepthWide computes search tasks). For any
,N 0
∈ ( 12 , 1) and any search task, there exists a transformer f ∈ TransformerN m,H,L such that f (X)
computes the task where X ∈ RN is some encoding of any graph G = (V , E) with |V |+|E| = O(N ),
and f has depth L = O(log N ), heads H = O(1), embedding dimension m = O(N , and no pause
tokens (N 0 = 0).
While both search and parallelizable tasks are computable logarithmic depth of considerable width,
the minimum depth transformer of width O(N ) for small that computes search tasks is unknown.
Despite this deficit, we can still establish the representation similarity of all search tasks by showing
that their minimum depths vary by at most a constant additive factor.
Theorem 22 (Depth-equivalence of search tasks). Suppose for some ∈ (0, 1) and γ > 0 there
,N 0
exists a transformer f ∈ TransformerN m,H,L with embedding dimension m = O(N ) and total
sequence length N + N 0 = O(N 1+γ ) that computes the some search task on all graphs G = (V , E)
of size |V | + |E| = O(N ). Then, for any other search task, there exists some transformer f 0 ∈
,N̄ 0
TransformerN m̄,H,L̄
with embedding dimension m̄ = O(m), depth L̄ = L + O(1), and pause tokens
0 0 3
N̄ = N + O(N ) that computes that search task.
We prove the existing of doubly-logarithmic-depth transformer that solves the triangle counting task
giving a more general result that counts the number of k-cliques in any graph of bounded arboricity.
Theorem 23 (Generalization of Theorem 7; loglog-depth computes clique counting). For any fixed
,N 0
∈ (0, 1) and k ≥ 3, there exists a transformer f ∈ TransformerN m,H,L with embedding dimension
m = O(n ), heads H = O(1), depth L = O(log log n), and chain-of-thought tokens
O(αk−2 N 1− ) if α = Ω(N /(k−2) ),
0
N =
0 otherwise.
that counts the number of k-cliques in all graphs G = (V , E) of arboricity α and size |V | + |E| =
O(N ).
21
Theorem 7 follows immediately from taking k = 3.
Our principal theoretical result establishes that any MPC protocol with sublinear local memory can
be simulated by a transformer with sublinear embedding dimension.
Theorem 8 (Formal version of Theorem 1; transformers simulate MPC). For constants 0 < δ <
δ 0 < 1 and γ > 0, an R-round deterministic (γ, δ)-MPC protocol with n inputs can be sim-
,N 0
ulated by a transformer f ∈ TransformerN m,H,L with depth L = O(R), single heads H = 1,
0
embedding dimension m = O(nδ ), context length N = n, and blank chain-of-thought tokens
N 0 = max(0, O(n1+γ−δ ) − n).
This offers an improvement on Theorem 3.1 of [71], which only guarantees that MPC protocols
with local memory s = O(n1/4− ) (or with δ < 14 ) can be simulated by transformers with sublinear
embedding dimension. This distinction is significant because several positive results for the MPC
protocol (e.g. the ability to solve all problems in L and NL in a logarithmic number of MPC rounds)
require s = Ω(N 1/2 ) local memory, and hence could not be shown to be simulated by transformers
of sublinear embedding dimension previously7 .
Including an allowance of N 0 blank pause tokens permits the simulation of MPC protocol with γ ≥ δ,
i.e. where number of machines q grows super-linearly with n. If γ < δ, then the protocol can be
simulated without pause tokens (i.e. N 0 = 0) for sufficiently large input sizes n.
To prove Theorem 8, we define a restriction of the MPC computational model that disallows each
machine from communicating with more than k other machines in each round of the protocol.
Definition 3. For constants γ, δ, ρ > 0, an R-round (γ, δ, ρ)-MPC protocol on input sequences of
length n is a (γ, δ)-MPC protocol (Definition 2) that obeys an additional constraint on the number
of outgoing and incoming messages. Namely, for some capacity k = O(nρ ), in each round, every
machine can send its local memory to and receive information from at most distinct k machines.
Theorem 8 is an immediate consequence of Proposition 24—which establishes that any (γ, δ)-MPC
protocol can be simulated by a (γ, δ, ρ)-MPC protocol—and Corollary 25—which shows that any
(γ, δ, ρ)-MPC protocol can be simulated by a transformer.
Proposition 24 ((γ, δ, ρ)-MPC simulates (γ, δ)-MPC). For constants γ, δ > 0 and ρ ∈ (0, 2δ ), if
2
f can be computed by an R-round (γ, δ)-MPC protocol, then there exists a O( R(1+γ)
ρ2 )-round
(γ, δ, ρ)-MPC protocol that computes f as well.
Corollary 25 (Transformers simulate (γ, δ, ρ)-MPC). For constants δ ∈ (0, 1) and γ, ρ > 0, an
R-round deterministic (γ, δ, ρ)-MPC protocol with n inputs can be simulated by a transformer f ∈
,N 0
TransformerN m,H,L with depth L = R + 1, heads H = 1, embedding dimension m = O(n
δ+4ρ
log n),
0 1+γ−δ
context length N = n, and blank chain-of-thought tokens N = max(0, O(n ) − n).
0 0
Proof of Theorem 8. Let ρ := min( 2δ , δ 4−δ ) − for some small constant (e.g. := min( 4δ , δ 8−δ )).
By Proposition 24, we can simulate the target R-round (γ, δ)-MPC protocol with an R0 -round
(γ, δ, ρ)-MPC protocol for
R(1 + γ 2 )
0
R =O .
min((δ 0 − δ)2 , δ 2 )
We apply Corollary 25 to conclude that this latter protocol can be simulated by a transformer of depth
L = R0 + 1 and embedding dimension
0
m = O(nδ+4ρ log n) = O(nδ ).
7
For this theoretical model of the transformer to be non-vacuous, the transformer must have embedding
dimension m = o(N d). Without this, any function over X ∈ RN ×d could be solved by a single-layer
transformer that concatenates all inputs into a single vector in an attention unit and passes that as input to an
MLP that solves the problem.
22
We prove Proposition 24 in Appendix B.3.1 by simulating a single round of a standard MPC with
multiple rounds of restricted MPC, where messages are sent to intermediate “neighborhoods”—which
contain collections of machines with similar destinations—of increasing fine granularity. We prove
Corollary 25 in Appendix B.3.2 by a minor adaptation of the proof of Theorem 3.1 of [71].
Fix an R-round (γ, δ)-MPC protocol π with input size n, q = Θ(n1+γ−δ ) machines, and s = O(nδ )
local memory. To prove Proposition 24, it suffices to show that a single round of MPC communication
2
can be simulated an O( (1+γ)
ρ2 )-round (γ, δ, ρ)-MPC protocol.
Let Inbox = (Inbox1 , . . . , Inboxq ) denote the same collection of message packets, organized by
their destination machines.
X
Inboxi = {(Msg, Src, Dst) ∈ OutboxSrc : Dst = i} s.t. |Msg| ≤ s.
Msg∈Inboxi
Proof. We define r = O( 1+γ ρ ) intermediate steps and prove that each of those can be simulated.
The intuition is that each an intermediate step routes each packet (Msg, Src, Dst) to a machine that
belongs to the same “neighborhood” as Dst. Each step maps each packet to a neighborhood of
smaller radius than the step before, until all packets have been transmitted to their proper destination
location.
ρ/2
m as follows. Fix some branching factor ` = Θ(n ) and number
We define a tree of neighborhoods
l
of intermediate steps r = log q 1+γ
log ` = O( ρ ). For any step t = 0, . . . , r and neighborhood index
q
j = 1, . . . , bt := `r−t , we define
Nbhdtj = (j − 1)`r−t + 1, (j − 1)`r−t + 2, . . . , j · `r−t ∩ [q]
and observe that it satisfies the size constraint |Nbhdtj | ≤ `r−t . We let Nbhdt (Dst) denote the
unique neighborhood Nbhdtj satisfying Dst ∈ Nbhdtj . We say that Nbhdtj < Nbhdtj 0 if j < j 0 (and
equivalently, for all i ∈ Nbhdtj and i0 ∈ Nbhdtj 0 , i < i0 ). Let
n o
Children(Nbhdtj ) = Nbhdt+1 t+1
(j−1)`+1 , . . . , Nbhdj` ,
τ t τ t
Descendants (Nbhdj ) = Nbhd (i) : i ∈ Nbhdj ,
Parent(Nbhdtj ) = Nbhdt−1
dj/`e ,
• The initial “step 0” neighborhood contains all machines (i.e. Nbhd01 = [q]), and the final
“step r” neighborhoods contain a single machine (i.e. Nbhdrj = {j} for all j ∈ [q]).
• For any t < r and j ∈ [bt ], Nbhdtj is the disjoint union of all sets in Children(Nbhdtj ) and
|Children(Nbhdtj )| ≤ `.
23
Step 0 Step 1 Step 2 Step r
1 Nbhd ` Nbhds `2 Nbhds q Nbhds
q machines per Nbhd q/` machines per Nbhd q/`2 machines per Nbhd 1 machine per Nbhd
Machine 2
Outbox02
Nbhd2`2 Nbhdrq
Machine q Machine q Machine q
Outbox0q Outbox1q Outboxrq
• Nbhdt1 , . . . , Nbhdtbt comprise a disjoint union of [q].
• For any Dst ∈ [q], there exist unique sets Nbhd0 (Dst), . . . , Nbhdr (Dst) that satisfy
8
For the sake of simplicity, we assume that the encoding Msg also contains all relevant metadata about the
packet Src and Dst. That is, we only track the message size |Msg|.
24
Note that Outbox satisfies the conditions for Outbox0 , and the only sequence of embeddings that
satisfies the conditions for Outboxr is Inbox. An inductive argument that applies Lemma 27 r times
completes the proof.
Lemma 27 ((γ, δ, ρ)-MPC simulates one intermediate step). For any t ∈ {0, . . . , r − 1} and ρ < 2δ ,
there exists an O( 1+γ
ρ )-round (γ, δ, ρ)-MPC protocol that takes as input any satisfactory Outbox
t
1. The pre-computation phase, where relevant message size metadata is computed using
O(r − t) rounds of communication.
2. The message-passing phase, where all messages are propagated in r − t rounds of commu-
nication in order to convert Outboxt to Outboxt+1 .
Since r = O( 1+γ
ρ ), the bound on total number of rounds is satisfied.
The algorithm maintains the invariant that all communication occurs within t-level neighborhoods
Nbhdtj without any mixing between neighborhoods; this is possible because of the assumed validity of
Outboxt , which implies that all messages whose packets appear in Nbhdtj of Outboxt have ultimate
destinations in Nbhdtj . Concretely, if (Msg, Src, Dst) ∈ Outboxti and i ∈ Nbhdtj , then Dst ∈ Nbhdtj .
We first explain the routing strategy for the message-passing phase by describing an itinerary of
machines that each message packet will be routed to and proving that this itinerary meets the
conditions needed to be executed by an (r − t)-round (γ, δ, ρ)-MPC protocol. We then describe how
the metadata necessary for the itinerary computations can be obtained during the pre-computation
phase.
Message-passing phase. We first introduce some packet notation to track all relevant metadata
about any message in the tth intermediate step. Fix some particular input Outboxt .
• Let P = (Msg, Src, Dst, Srct ) denote a packet that contains a message Msg and meta-
data concerning its “global” source Src and destination Dst (i.e. (Msg, Src, Dst) ∈
OutboxSrc , InboxDst ) and its “local” source Srct from within Outboxt (i.e.
(Msg, Src, Dst) ∈ OutboxtSrct ).
• Let |P| = |Msg| be the total number of bits needed to encode the packet.
• We establish a lexicographical ordering that depends first on Srct and next on Dst. That
is, we say that P 0 < P for some P 0 = (Msg0 , Src0 , Dst0 , Srct0 ) if Srct0 < Srct , or if
Srct0 = Srct and Dst0 < Dst.
We develop a packet scoring function z t,τ (P), which in turn induces an itinerary function bt,τ (P) ∈ [q]
that assigns an intermediate machine to hold packet P after r−τ communication steps. These functions
are carefully chosen in order to ensure that the local memories of each individual machine are bounded
by O(s) and that the number of machine each sends messages to and receives messages from is
bounded by O(`2 ). Critically, we ensure that bt,τ (P) ∈ SrcNbhdt,τ (P); that is, we initially rearrange
packets within solely within the smallest neighborhoods of Outboxt and gradually propagate packets
more widely. That way, each packet P obeys a trajectory of the following form over r − t MPC
25
rounds:
To define these functions, we first define partial sums of message sizes in order to later determine
an itinerary of machines that P will be routed to in between Srct and some destination machine in
DstNbhdt (P). The first term, EqualDstSumt,τ (P), sums the size of all “lesser” packets that share a
destination neighborhood DstNbhdt (P) and a τ th level input neighborhood SrcNbhdt,τ (P):
X
EqualDstSumt,τ (P) = |P0 |.
P0 ∈SrcNbhdt,τ (P)
DstNbhdt (P0 )=DstNbhdt (P)
P0 <P
The second term, LessDstSumt,τ (P), sums the size of all packets that share an input neighborhood
but have a “smaller” destination neighborhood than P:
LessDstSumt,τ (P) = LessDstSumt (SrcNbhdt,τ (P), DstNbhdt (P)),
where
X
LessDstSumt (SrcNbhd, DstNbhd) = |P0 |.
P0 ∈SrcNbhd
DstNbhdt (P0 )<DstNbhd
We now define the packet scoring and itinerary functions for any τ ∈ {t, . . . , r}:
2s · min SrcNbhdt,τ (P) + LessDstSumt,τ (P) + EqualDstSumt,τ (P) τ ≥ t + 1,
t,τ
z (P) =
2s · min DstNbhdt (P) + 2 · EqualDstSumt,τ (P) τ = t.
t,τ
z (P)
bt,τ (P) = .
2s
We prove a series of claims to establish that the packet scoring and itinerary functions are properly
defined.
Claim 28 (Itinerary range). For any τ ∈ {t, . . . , r}, the itinerary function satisfies
SrcNbhdt,τ (P) τ ≥ t + 1
t,τ
b (P) ∈
DstNbhdt (P) τ = t.
As an immediate consequence, bt,r (P) = Srct for P = (Msg, Src, Dst, Srct ).
26
Claim 29 (Gaps between scores). If P1 6= P2 and z t,τ (P1 ) ≤ z t,τ (P2 ), then
zt,τ (P1 ) + |P1 | ≤ z t,τ (P2 ).
Proof. First, let τ > t. Consider the case where SrcNbhdt,τ (P1 ) 6= SrcNbhdt,τ (P2 ). By
Claim 28 and our assumption that z t,τ (P1 ) ≤ z t,τ (P2 ), it must be the case that SrcNbhdt,τ (P1 ) <
SrcNbhdt,τ (P2 ). Hence, we have the following by applying Equation (2).
z t,τ (P2 ) − zt,τ (P1 )
≥ 2s · (min SrcNbhdt,τ (P2 ) − (min SrcNbhdt,τ (P1 ) + |SrcNbhdt,τ (P1 )| − |P1 |))
≥ |P1 |.
Otherwise, if SrcNbhdt,τ (P1 ) = SrcNbhdt,τ (P2 ), then
z t,τ (P2 ) − zt,τ (P1 ) = (LessDstSumt,τ (P2 ) + EqualDstSumt,τ (P2 )) − (LessDstSumt,τ (P1 ) + EqualDstSumt,τ (P1 ))
X X
= |P0 | − |P0 |,
P0 ∈S2 P0 ∈S1
for some packet subsets S1 , S2 ⊂ SrcNbhdt,τ (P1 ). By further inspecting the respective
LessDstSumt,τ and EqualDstSumt,τ terms, we observe that S1 ⊂ S2 and P1 ∈ S2 \ S1 . The
claim immediately follows.
The argument for the case τ = t is nearly identical.
Claim 30 (Local memory bound). For any b ∈ N,
X 2s τ ∈ {t, r}
|P| ≤
t,τ
3s τ ∈ {t + 1, . . . , r} .
P:b (P)=b
Proof. The case τ = r is an immediate consequence of the inductive assumption that Outboxt
satisfies the desired intermediate properties.
For all other cases, let {P1 , . . . , Pn } denote all packets with bt,τ (Pi ) = b and let z t,τ (P1 ) ≤ · · · ≤
z t,τ (Pn ) without loss of generality. We use Claim 29, the assumption that all |Pi | ≤ s, and the
boundedness of z t,τ (Pi ) from Claim 28 to conclude the proof.
n
X n−1
X
|Pi | ≤ (z t,τ (Pi+1 ) − z t,τ (Pi )) + |Pn |
i=1 i=1
≤ z (Pn ) − z t,τ (P1 ) + s
t,τ
2s τ = t,
≤
3s τ > t.
Claim 31 (Intra-class distance preservation). If P1 and P2 satisfy SrcNbhdt,τ +1 (P1 ) =
SrcNbhdt,τ +1 (P2 ) and DstNbhdt (P1 ) = DstNbhdt (P2 ), then
z t,τ (P1 ) − z t,τ (P2 ) = z t,τ +1 (P1 ) − z t,τ +1 (P2 ).
Proof. Since SrcNbhdt,τ +1 (P1 ) = SrcNbhdt,τ +1 (P2 ), it follows that SrcNbhdt,τ (P1 ) =
SrcNbhdt,τ (P2 ) and therefore,
z t,τ (P1 ) − z t,τ (P2 ) = EqualDstSumt,τ (P1 ) − EqualDstSumt,τ (P2 )
X
= |P0 | (3)
0 t,τ
P ∈SrcNbhd (P1 )
DstNbhdt (P0 )=DstNbhdt (P1 )
P1 ≤P0 <P2
27
Because P1 , P2 ∈ SrcNbhdt,τ +1 (P1 ), the defined packet ordering implies that any P0 ∈ [P1 , P2 ) must
satisfy P0 ∈ SrcNbhdt,τ +1 (P1 ). Therefore, Equations (3) and (4) are equal and the claim holds.
Claim 32 (Distinct recipients bound). For any b,
| bt,τ (P) : bt,τ +1 (P) = b | ≤ 3`.
Proof. Within each root neighborhood Nbhdτj , there exist at most ` destination neighborhoods in
DstNbhdt (P) for P ∈ Nbhdτj .
Fix some such DstNbhd and let P1 , . . . , Pn denote all packets with bt,τ +1 (P) = b and
DstNbhdt (Pi ) = DstNbhd. Without loss of generality, assume that z t,τ (P1 ) ≤ · · · ≤ z t,τ (Pn ).
Because all such packets belong to the same machine in step r − τ − 1 (i.e. bt,τ +1 (Pi ) = b for all
i ∈ [n]), they belong share the same source neighborhood of size `r−τ −1 (i.e. SrcNbhdt,τ +1 (Pi ) =
SrcNbhdt,τ +1 (P1 )). By Claim 31 and the definition of bt,τ ,
1
bt,τ (Pi ) − bt,τ (P1 ) ≤ 1 + (z t,τ (Pi ) − z t,τ (P1 ))
2s
1
= 1 + (z t,τ +1 (Pi ) − z t,τ +1 (P1 ))
2s
≤ 2 + bt,τ +1 (Pi ) − bt,τ +1 (P1 ) = 2.
Therefore, there are at most three possible values of bt,τ (Pi ).
The claim follows by considering each of the ` destination neighborhoods separately:
t,τ X
b (P) : bt,τ +1 (P) = b = bt,τ (P) : bt,τ +1 (P) = b, DstNbhdt (P) = DstNbhd
DstNbhd
≤ 3`.
Claim 33 (Distinct senders bound). For any b,
| bt,τ +1 (P) : bt,τ (P) = b | ≤ 3`2 .
Proof. Within each Nbhdτj , there exist at most `2 distinct pairs of destination neighborhoods
DstNbhdt (P) and source neighborhoods Nbhdt,τ (P) for P ∈ Nbhdτj .
As before, we fix some DstNbhd and SrcNbhd and let P1 , . . . , Pn all satisfy bt,τ (Pi ) = b,
DstNbhdt (Pi ) = DstNbhd, and SrcNbhdt,τ +1 = SrcNbhd. Using the same argument, we show that
t,τ +1
b (Pi ) : i ∈ [n] ≤ 3.
We conclude by considering all such pairs.
As a result of Claims 30, 32, and 33, we conclude that each packet P = (Msg, Src, Dst, Srct ) can be
equipped with some itinerary
Srct = bt,r (P), bt,r−1 (P), . . . , bt,t+1 (P), bt,t (P) ∈ DstNbhdt (P)
that properly translates an instances of Outboxt to Outboxt+1 and does so without ever requiring
local memory more than 3s = O(N δ ) on any intermediate step or any machine to send or receive
messages from more than 3`2 = O(N ρ ) other machines. This itinerary can be executed using an
(r − t)-round (γ, δ, ρ)-MPC protocol.
Pre-computation phase. It remains to show that each bt,τ can be computed for each packet. To do
so, we prove that there exists an O(r − t)-round (γ, δ, ρ)-MPC protocol that ends with each machine
i knowing EqualDstSumt,τ (P) and LessDstSumt,τ (P) for each P ∈ Outboxti and τ ∈ {t, . . . , r}.
We design an MPC protocol that uses the tree structure to propagate information about individual
child neighborhoods to their parents and vice versa. We describe recursive relationships that elucidate
how to compute the two salient quantities.
First, we introduce a useful intermediate term. Let
X X
EqualDstSumt,τ (i, DstNbhd) = |P0 |,
i0 ∈Nbhdτ (i) P0 ∈Outboxti0
i0 <i DstNbhdt (P0 )=DstNbhd
28
denote the sum of all packet that are contained by “lesser” machines that share source and destination
neighborhoods. Note that EqualDstSumt,τ (P) can be computed for any P ∈ Outboxti locally by
machine i given prior knowledge of EqualDstSumt,τ (i, DstNbhdt (P)). Thus, the pre-computation
phase need only compute the latter term.
We also introduce a term that represents sum of the sizes of all packets that share source and
destination neighborhoods:
X
NbhdSumt (SrcNbhd, DstNbhd) = |P0 |.
P0 ∈SrcNbhd
DstNbhdt (P0 )=DstNbhd
Now, we provide the recurrences for any τ < r (or for any SrcNbhd satisfying |SrcNbhd| > 1:
EqualDstSumt,τ (i, DstNbhd)
= EqualDstSumt,τ +1 (i, DstNbhd) + EqualDstSumt,τ (min Nbhdτ +1 (i), DstNbhd)
X
= EqualDstSumt,τ +1 (i, DstNbhd) + NbhdSumt (SrcNbhd, DstNbhd),
τ
SrcNbhd∈Children(Nbhd (i))
SrcNbhd<Nbhdτ +1 (i)
X
NbhdSumt (SrcNbhd, DstNbhd) = NbhdSumt (SrcNbhd, DstNbhd), (5)
SrcNbhd0 ∈Children(SrcNbhd)
X
LessDstSumt (SrcNbhd, DstNbhd) = LessDstSumt (SrcNbhd, DstNbhd).
SrcNbhd0 ∈Children(SrcNbhd)
(6)
When τ = r, the terms EqualDstSumt,τ (i, DstNbhd), NbhdSumt (Nbhdr (i), DstNbhd), and
LessDstSumt (Nbhdr (i), DstNbhd) can be computed locally within machine i.
We follow a tree-like communication pattern to compute all relevant sums. Each machine i ∈ [q]
computes
EqualDstSumt,τ (i, DstNbhd) : τ ≥ t, DstNbhd ∈ Children(Nbhdt (i))
and
LessDstSumt (Nbhdτ (i), DstNbhd) : τ ≥ t, DstNbhd ∈ Children(Nbhdt (i))
Propagate-up rounds: Fix some neighborhood Nbhdtj . After r − τ propagate-up rounds, the
goal is to compute the terms NbhdSumt (SrcNbhd, DstNbhd) and LessDstSumt (SrcNbhd, DstNbhd)
for each relevant destination neighborhood DstNbhd ∈ Children(Nbhdtj ) and source neighbor-
hood SrcNbhd ∈ Descendantsτ (Nbhdtj ) within a single machine in each source neighborhood,
min SrcNbhd. We argue that this is possible inductively.
Before performing any computation (that is, after “round 0”), each machine i individually com-
putes NbhdSumt (Nbhdr (i), DstNbhd) and LessDstSumt (Nbhdr (i), DstNbhd) by aggregating the
messages encoded in its own representation of Outboxti .
29
We assume that the procedure works as specified for r − τ rounds of communication. Fix some
SrcNbhdτ −1 ∈ Descendantsτ −1 (Nbhdtj ). Then, for every SrcNbhdτ ∈ ChildrenSrcNbhdτ −1 ,
the quantities
{NbhdSumt (SrcNbhdτ , DstNbhd) : DstNbhd ∈ Children(Nbhdtj )}
∪ {LessDstSumt (SrcNbhdτ , DstNbhd) : DstNbhd ∈ Children(Nbhdtj )}
have already been computed and stored in min SrcNbhdτ . By the recurrence relations of Equa-
tions (5) and (6), NbhdSumt (SrcNbhdτ −1 , DstNbhd) and LessDstSumt (SrcNbhdτ −1 , DstNbhd) are
functions of those quantities. Thus, it suffices to have each machine min SrcNbhdτ machine transmit
its relevant terms to min SrcNbhdτ −1 . A round of MPC communication that transmits such messages
involves each machine sending most one message of size ` and receiving at most ` messages, each of
size O(`).
Inductively, we ensure that all neighborhood sums are computed after r − t propagate-up rounds.
Because each machine handles at most ` distinct messages having total size O(`2 ) per MPC round,
this protocol does not violate the bounded message size and bounded distinct message constraints (so
long as `2 s), which can be guaranteed for sufficiently large n, so long as ` = O(nρ/2 ).
Aggregation rounds: After the completion of the aggregation rounds, each machine i computes terms
of the form
EqualDstSumt,τ (min Nbhdτ +1 (i), DstNbhd) : DstNbhd ∈ Children(Nbhdtj )
from relevant NbhdSumt terms if i = min Nbhdτ +1 (i). By the recurrence, it is sufficient for machine
i to following `2 distinct terms
NbhdSumt (SrcNbhd, DstNbhd) : SrcNbhd ∈ Children(Nbhdτ (i)), DstNbhd ∈ Children(Nbhdtj ) .
Since all machines i already knows such terms for SrcNbhd = Nbhdτ +1 (i), it can obtain the
remaining NbhdSumt terms by simultaneously sharing information with its “cousin machines:”
min SrcNbhd, for all SrcNbhd ∈ Children(Nbhdτ (i)). This can be handled by a single round
of communication where each “(τ + 1)th-level neighborhood representative” machine forwards its
sums sums to up to ` other first representatives, for a total messaging cost of O(`2 ).
We use r − t separate rounds to repeat this process for each τ ≥ t.
Propagate-down rounds: It remains to compute each EqualDstSumt,τ (i, ·) and
LessDstSumt (Nbhdτ (i), ·) term at each machine i. The relevant LessDstSumt terms have
already been computed by each respective min Nbhdτ (i) machine and can be propagated to machine
0
i by using r − t rounds of tree-propagation through intermediate min Nbhdτ (i) machines.
The same is possible for EqualDstSumt,τ terms, although the individual terms need to be added in
order to follow the recurrence. We propagate the terms in the same way as the LessDstSumt terms,
but we take special care to carry out the extra additions. This can be accomplished simultaneously
to the other propagate-down rounds. This protocol involves each first-child node sharing at most `
distinct messages, each of size at most O((r − t)`). As before, every node sends and receives at most
O((r − t)`2 ) s words.
After these 3(r − t) rounds have elapsed, each machine i computes bt,τ (P) for each P ∈ Outboxti .
Using this itinerary, the machines routes tuples of the form
(P, bt,r (P), . . . , bt,t (P))
to the respective machine bt,τ (P) in round r − τ . Due to the arguments presented at the start of the
section, this procedure terminates with each (Msg, Src, Dst) tuple being held by some machine i
such that the resulting Outboxt+1
i is valid, and the procedure involves at most O(r − t) = O( 1+γ ρ )
rounds of (γ, δ, ρ)-MPC computation.
Corollary 25 (Transformers simulate (γ, δ, ρ)-MPC). For constants δ ∈ (0, 1) and γ, ρ > 0, an
R-round deterministic (γ, δ, ρ)-MPC protocol with n inputs can be simulated by a transformer f ∈
,N 0
TransformerN m,H,L with depth L = R + 1, heads H = 1, embedding dimension m = O(n
δ+4ρ
log n),
0 1+γ−δ
context length N = n, and blank chain-of-thought tokens N = max(0, O(n ) − n).
30
The proof of Corollary 25 involves an adaptation to the proof of Theorem 3.1 of [71]. To avoid
restating the proof in its entirety, we provide a brief outline of the proof of Theorem 3.1 and explain
which modification is necessary.
Theorem 3.1 is a consequence of Lemmas B.4, B.5, and B.6 of [71], which establish that there exist
single-layer transformers that simulate the initialization, a round of computation and communication,
and the output formatting of any fixed (γ, δ)-MPC protocol. The input and output steps of (γ, δ), and
(γ, δ, ρ)-MPC protocols are identical, only Lemma B.5 needs to be examined.
To simulate a single round of an MPC protocol with a transformer, all local computations are
simulated in the element-wise multi-layer perceptron (MLP) units, and all communication is handled
in a single multi-headed self-attention layer (Lemma B.7) Since (γ, δ, ρ)-MPC protocols add no
restrictions related to local computation, the former can be simulated in exactly the same manner with
identical MLPs, and it remains to analyze the construction of Lemma B.7. We restate Lemma B.7
and provide an replacement lemma that suffices to prove Corollary 25.
Lemma 34 (Lemma B.7 of [71]; multi-headed attention simulates MPC communication). For any
R-round MPC protocol with local memory s and q machines and any round r ∈ [R − 1], there exists
a single-layer transformer f ∈ Transformerq,0 4
m,H,1 with H = O(log log q) and m = O(s log q),
which, given as input a length-q encoding of each machine’s outgoing messages in round r, returns
an encoding of each machine’s incoming messages in round r + 1.
Lemma 35 (Single-headed attention simulates bounded-message MPC communication). For any
R-round MPC protocol with local memory s, q machines, and a k-machine communication limit and
any round r ∈ [R − 1], there exists a single-layer single-headed transformer f ∈ Transformerq,0 m,H,1
with H = 1 and m = O(k 4 s log q), which, given as input a length-q encoding of each machine’s
outgoing messages in round r, returns an encoding of each machine’s incoming messages in round
r + 1.
Lemma 35 is an immediate consequence of Lemma 3.2 of [71], their main technical result, which
already applies to the regime with limits on the number of machines in communication.
By replacing Lemma 34 with Lemma 35, applying the remainder of the proof of Theorem 3.1 of [71],
and letting k = O(nρ ), the proof of Corollary 25 is complete.
31
• For each v ∈ V , there exists exactly one i ∈ [N − 1] with xi = (v, 0, 1, 0, 0).
• For each (u, v) ∈ E, there exists exactly i ∈ [N − 1] with xi = (u, v, 0, 1, 0).
• The token xN encodes a particular instance of the task P , by encoding isTaskN = 1 with
an optional edge or node encoding. That is, for tasks without arguments (such as triangle
count), xN = (0, 0, 0, 0, 1). For tasks with a single node argument v ∈ [n] (such as node
degree), xN = (v, 0, 1, 0, 1). For tasks with a pair of node arguments u, v ∈ [n] (such as
shortest path and connectivity), xN = (u, v, 0, 1, 1).
• All other tokens satisfy xi = (0, 0, 0, 0, 0).
We say that a single-layer transformer f ∈ TransformerNm,H,1 solves task P on graph G if the output
corresponding to the task description f (X(G, P ))N encodes the output of the task. Since f is a
single-layer transformer, we can write this output as
H
!
X
T T T
f (X)N = ψ softmax φ(xN ) Qh Kh φ(X) φ(X)Vh (7)
h=1
φ : R d → R m , ψ : Rm → R
broadcasted across each input (i.e. φ(X) = (φ(x1 ), . . . , φ(xN )) ∈ RN ×m ) and weight matrices
Q1 , . . . , QH , K1 , . . . , KH , V1 , . . . , VH ∈ Rm×m .
Throughout, we assume that all parameters in the transformer model and intermediate numerical
quantities can be written using O(log N )-bit floating point numbers. This assumption can be satisfied
for the positive results of Appendix C.1 and is necessary to obtain the negative results of Appendix C.2.
We permit the element-wise MLPs φ and ψ to be arbitrary functions for the negative results, while
restricting them to be MLPs that can be approximated using bounded-size multi-layer ReLU networks
for the positive results. While we do not make these ReLU networks explicit, we restrict ourselves to
simple operations that can be computed using linear transformations and the application of smooth
univariate functions.
Finally, we acknowledge that the negative results in are largely superseded by those of [55], which
establishes that L-complete languages (including graph connectivity and cycle check) cannot be
efficiently solved by constant-depth transformers, let alone single-layer transformers. We include
these bounds anyway to mark a contrast with our positive results; draw further connections to the
communication complexity lens on transformers; and exhibit simple task instances (G, P ) that require
greater depth, including constant diameter and constant degree graphs.
Theorem 36 (Formal version of Theorem 5; Depth1 computes retrieval tasks). Fix any graph
reasoning task among node count, edge count, edge existence, and node degree and any graph size N .
Then, there exists a single-layer single-headed transformer f ∈ TransformerNm,1,1 with embedding
dimension m = O(log N ) that solves the task on all graphs G = (V , E) of size |V | + |E| ≤ N − 1
formatted as node/edge embedding sequences.
Proof. We prove that a single head of self-attention with input and output MLPs can solve these
retrieval and aggregation tasks in a parameter-efficient manner by first carefully designing a universal
input MLP φ : Rd → Rm for some m = O(log N ) to produce embeddings that encode useful graph
properties. Then, we define task-specific query, key, and value matrices Q, K, V :∈ Rm×m and
output MLPs ψ : Rm → R that produce the correct answers. While we do not explicitly account for
finite-precision computations, all of the operations utilized can be carefully implemented to respect
O(log N )-bit floating-point numerical representations using the technical approaches of [70, 71].
32
Shared sinusoidal embedding MLP. We denote the embedding output of the input MLP as
φ(xi ) = (isTaski , isVertexi , isEdgei , φ0 (xi )) ∈ Rm
0
for some φ0 (xi ) : Rd → R2m for some m0 = O(log N ) and m = 2m0 + 3. For some fixed
a1 , . . . , am0 ∈ [0, 1] to be determined, let
1
η(xi ) if isVertexi = 1,
φ0 (xi ) = ξ(x1i , x2i ) if isEdgei = 1,
~
0 otherwise,
where η is a sinusoidal embedding MLP for nodes with
0
η(v) = (sin(2πa1 v), cos(2πa1 v), . . . , sin(2πam0 v), cos(2πam0 v)) ∈ R2m , (8)
and ξ is an edge embedding satisfying
m0
ξ(u, v) = (η(u) + η(v)).
m0 + η(u)T η(v)
We first show that the node embeddings are approximately orthogonal. By employing standard
trigonometric identities, we obtain
0
m
X
T
η(u) η(v) = cos(2πaj (u − v)),
j=1
2
and note that kη(v)k2 = m0 . We use a standard concentration argument to prove that |η(u)T η(v)|
m0 if u 6= v with high probability.
Claim 37 (Near-orthogonality of embeddings η). There exist coefficients a1 , . . . , am0 ∈ [0, 1] that
comprise the embedding η : [n] → Rm of Equation (8) such that
p
η(u)T η(v) ≤ 2 m0 log n, if u 6= v.
Proof. We employ the probabilistic method. Consider m0 iid random variables a1 , . . . , am0 ∼
Unif([0, 1]) and let η represent the respective node embedding. Fix some arbitrary u, v ∈ [n] with
u 6= v and note that
Xm0
T
η(u) η(v) = cos(2πaj (u − v)).
j=1
h p i n(n − 1)
Pr ∀u, v ∈ [n], u 6= v : η(u)T η(v) ≥ 2 m0 log n ≤ < 1.
2n2
Hence, there exists a satisfactory set of coefficients a1 , . . . , am0 .
√
For some η satisfying Claim 37, let χ = maxu,v∈[n],u6=v |η(u)T η(v)| ≤ 2 m0 log n. By taking
0
m0 = O(log n) be sufficiently large, we guarantee that χ ≤ mc for any constant c ≥ 2.
We then bound all relevant inner products between vertex and edge embeddings for sufficently
large m0 (and sufficiently small χ). We assume throughout that u, u0 , v, v 0 are distinct and note that
ξ(u, v) = ξ(v, u) (and omit symmetric inequalities).
33
2
kη(v)k2 = m0 . (9)
m0
η(u)T η(v) ≤ χ ≤ . (10)
4
m0 (m0 + η(u)T η(v))
η(u)T ξ(u, v) − m0 = − m0 = 0. (11)
m0 + η(u)T η(v)
m0 (η(u)T η(u0 ) + η(u)T η(v 0 )) 2m0 χ m0
η(u)T ξ(u0 , v 0 ) = ≤ ≤ 4χ ≤ . (12)
m0 + η(u0 )T η(v 0 ) m0 − χ 4
2 m02 · (2m0 + 2η(u)T η(v)) − 2m0 (m0 + η(u)T η(v))2
kξ(u, v)k2 − 2m0 =
(m0 + η(u)T η(v))2
2m0 η(u)T η(v) 2m0 χ m0
≤ ≤ ≤ 4χ ≤ . (13)
m0 + η(u)T η(v) m0 − χ 4
m02 (m0 + η(u)T η(v 0 ) + η(u)T η(v) + η(v)T η(v 0 ))
ξ(u, v)T ξ(u, v 0 ) − m0 = − m0
(m0 + η(u)T η(v))(m0 + η(u)T η(v 0 ))
m02 η(v)T η(v 0 ) − m0 η(u)T η(v)η(u)T η(v 0 )
=
(m0 + η(u)T η(v))(m0 + η(u)T η(v 0 ))
m02 χ + m0 χ2 χ2 m0
≤ 0 2
≤ 4χ + 4 0 ≤ . (14)
(m − χ) m 4
m02 (η(u)T η(u0 ) + η(u)T η(v 0 ) + η(v)T η(u0 ) + η(v)T η(v 0 )
|ξ(u, v)T ξ(u0 , v 0 )| =
(m0 + η(u)T η(v))(m0 + η(u0 )T η(v 0 ))
4m02 χ m0
≤ ≤ 16χ ≤ . (15)
(m0 − χ)2 4
m0 m0
η(u)T η(v), η(u)T ξ(u0 , v 0 ), ξ(u, v)T ξ(u0 , v 0 ) ∈ − , ,
4 4
0
3m 5m0
2
kη(v)k2 , η(u)T ξ(u, v), ξ(u, v)T ξ(u, v 0 ) ∈ , ,
4 4
0
7m 9m0
2
kξ(u, v)k2 ∈ , .
4 4
We now apply the above bounds to prove that each task can be solved by a single-headed transformer
with weights Q, K, V ∈ Rm×m , output MLP ψ : Rm → R, and shared input MLP φ discussed
previously.
Node count task. If P encodes the node count task, we specify weights matrices Q, K, V in order
to ensure that QT φ(xN ) = m0 e1 9 ; K T φ(xi ) = (isVertexi + isTaski ) · e1 ; and V T φ(xi ) =
isTaski · e1 10 . Then, the following is true of exponentiated key/query inner products and scalings
products of value vectors:
0
em
T T if isVertexi = 1 or i = N ,
exp(φ(xN ) QK φ(xi )) =
1 otherwise.
0
em · e1
T T if i = N ,
exp(φ(xN ) QK φ(xi ))V φ(xi ) = ~
0 otherwise.
9
Note that QT φ(xN ) is the only relevant query embedding to the output f (X)N ; see Equation (7).
10
While we do not specify the entries of Q, K, V , they are clearly linear transformations.
34
The output of the softmax is then
PN
exp(φ(xN )T QK T φ(xi ))V φ(xi )
softmax φ(xN )T QK T φ(X)T φ(X)V = i=1
PN T T
i=1 exp(φ(xN ) QK φ(xi ))
0
em
= · e1 .
em0 · (1 + |V |) + 1 · (N − |V | − 1)
Let
yN := (softmax φ(xN )T QK T φ(X) φ(X)V )1 ∈ R
denote the first coordinate of the softmax output. By taking m0 ≥ log(4N ), we guarantee that
1 1 1 1
yN ∈ , ⊆ , .
1 + |V | + N/em0 1 + |V | 5
4 + |V |
1 + |V |
j k
By letting the output MLP ψ approximate the function ψ(z) = z11 − 1 for z1 ∈ [1, N + 2], we can
ensure that f (X)N = |V |.
Edge count task. We similarly let QT φ(xN ) = m0 e1 ; K T φ(xi ) = (isEdgei + isTaski ) · e1 ; and
V T φ(xi ) = isTaski · e1 . By an identical analysis of the attention matrix and ψ construction, we
ensure that f (X)N = |E|.
Edge existence task. Under the node/edge encoding, we assume that the edge existence task
isencoded as xN = (x1N , x2N , 0, 1, 1) for some x1N , x2N ∈ V and should return f (X)N =
1 (x1N , x2N ) ∈ E . We choose our weight matrices to ensure that QT φ(xN ) = φ0 (xN ) =
ξ(x1N , x2N ); K T φ(xi ) = φ0 (xi ), and V T φ(xi ) = 2(1 − isTaski )e1 . By applying Claim 37 and
letting m0 = O(log N ) to be sufficiently large, the following is true the query/key inner products:
2 0
exp(φ(xN )T QK T φ(xi )) = exp( ξ(x1N , x1N ) 2 ) ≥ e7m /4 if x1i , x2i = x1N , x2N ,
0 1 7m0 /4
exp(φ(xN )T QK T φ(xi )) ≤ e5m /4 ≤ e otherwise.
8N
We can therefore analyze the softmax output yN to obtain the following bound:
2 1 7m0 /4
2 exp( ξ(x1N , x1N ) 2 )1 (x1N , x2N ) ∈ E + 2N · 8N e
yN ≤ 2
1 1 1 2
exp(kξ(xN , xN )k2 )(1 + 1 {(xN , xN ) ∈ E})
21 (xN , xN ) ∈ E + 41
1 2
1
≤ 1 (x1N , x2N ) ∈ E + ,
≤ 1 2
1 + 1 {(xN , xN ) ∈ E} 4
1 1 2 1 2
2 exp( ξ(xN , xN ) 2 )1 (xN , xN ) ∈ E
yN ≥ 2 1 7m0 /4
exp(kξ(xN , x1N )k2 )(1 + 1 {(x1N , x2N ) ∈ E}) + N · 8N
1 e
1 2
21 (xN , xN ) ∈ E 1 2 1
≥ 1 2 1 ≥ 1 (xN , xN ) ∈ E − 4 .
1 + 1 {(xN , xN ) ∈ E} + 8
Hence,
1 1
− , if (x1N , x2N ) 6∈ E,
yN ∈ 3 45 4
4, 4 if (x1N , x2N ) ∈ E.
Therefore, it suffices to design a threshold output MLP ψ that satisfies
1 z1 ≥ 34 ,
ψ(z) =
0 z1 ≤ 14 ,
in order to distinguish between the two possible output ranges. This can be easily constructed by
taking a linear combination of two ReLU neurons.
35
1
Node degree task. We assume 1 that the task is encoded as xN = (xN , 0, 1, 0, 1) and should return
f (X)N = deg(xN ) := (xN , v) ∈ E . We use weight matrices with QT φ(xN ) = φ0 (xN ) =
1
η(xN ); K T φ(xi ) = φ0 (xi ); and V T φ(xi ) = isTaski . This ensure that the following is true about
the query/key inner products for sufficiently large m0 :
0
exp(φ(xN )T QK T φ(xi )) = em if i = N or x1N ∈ x1i , x2i ,
0 1 m0
exp(φ(xN )T QK T φ(xi )) ≤ em /4 ≤ e otherwise.
4N
We similarly ensure that the following holds about the softmax output yN .
0
em 1
yN ≤ = ,
em 0 · (deg(x1N ) + 2) deg(x1N ) + 2
0
em
yN ≥ 1 m0
(deg(xN ) + 2)em0 + N ·
1
4N e
1
≥ .
deg(xN ) + 49
1
Theorem 38 (Formal version of Theorem 6; Depth1 cannot compute search or retrieval tasks).
Fix any graph reasoning task among graph connectivity, shortest path, and cycle detection. Any
single-layer transformer f ∈ TransformerN m,H,1 with O(log N )-bit precision that solves the task on
all graphs G = (V , E) of size |V | + |E| ≤ N − 1 formatted as node/edge embedding sequences has
width satisfying mH = Ω(N/ log N ).
Our negative results generalizes and applies the approach of [70] to prove negative results for single-
layer transformers by communication complexity reductions. The bounds hinge on the following
fundamental fact about the hardness of a two-player game where two agents jointly attempt to
compute a set disjointness quantity that depends on both of their inputs with bounded communication.
Fact 1 (Disjointness communication complexity lower bound [85]). Suppose Alice and Bob are given
inputs a, b ∈ {0, 1}r and wish to jointly compute D ISJ(a, b) = maxi ai bi by alternately sending
single-bit messages to one another. Any deterministic protocol that computes D ISJ(a, b) requires at
least r rounds of communication, or r bits of information.
We first generalize Theorem 7 of [70] to show that no transformer can efficiently solve an embedded
set disjointness problem without having a width that scales linearly in r. We prove the theorem and
later apply to it prove negative results about relevant graph reasoning tasks. To demonstrate that the
graph reasoning tasks do not require pathological input graphs G to be hard, we exhibit particular
input graph instances with constant graph diameter or constant degree where the task cannot be
efficiently solved.
Lemma 39 (Generic Depth1 communication complexity negative result). For some sequence
length, fix two disjoint subsets A, B ⊂ [N − 1], and consider a single-layer transformer f ∈
TransformerN m,H,1 with O(log N )-bit precision that solves set disjointness, i.e. f (X)N = D ISJ (a, b)
for any input X where XA is a function of Alice’s input a ∈ {0, 1}r , XB is a function of Bob’s
input b ∈ {0, 1}r , and X[N ]\(A∪B) is fixed regardless of a and b. Then, f has width satisfying
mH = Ω(r/ log N ).
Proof. Consider any transformer f that solves D ISJ as specified in the theorem statement. We
show that this implies the existence of a O(mH log N )-round communication protocol that solves
D ISJ(a, b) for any inputs a, b ∈ {0, 1}r . An application of Fact 1 immediately proves the theorem
statement.
36
If such a transformer f exists, then the following is true for some φ, ψ, and Q, K, V :
H
!
X
softmax φ(xN )T Qh KhT φ(X) φ(X)Vh
D ISJ(a, b) = ψ
h=1
H
!
X Zh,A exp(Lh,A ) + Zh,B exp(Lh,B ) + Zh,[N ]\(A∪B) exp(Lh,[N ]\(A∪B) )
=ψ ,
exp(Lh,A ) + exp(Lh,B ) + exp(Lh,[N ]\(A∪B) )
h=1
for partial softmax and normalization terms11 defined as follows for S ⊂ [N ] and h ∈ [H]:
exp(φ(xN )T Qh KhT φ(xi ))Vh φ(xi )
P
Zh,S = i∈S P T Q K T φ(x ))
∈ Rm ,
i∈S exp(φ(x N ) h h i
!
X
T T
Lh,S = log exp(φ(xN ) Qh Kh φ(xi )) ∈ R+ .
i∈S
The definition of input instances X implies that Alice can compute Zh,A and Lh,A as a function of
her input a; Bob can similarly compute Zh,B and Lh,B from b; and Zh,[N ]\(A∪B) , Lh,[N ]\(A∪B) , and
all implementation details of f are known by all players. Therefore, Bob can compute D ISJ(a, b) by
an O(mH log N )-round protocol where Alice sends him (Zh,A , Lh,A ) bit-by-bit.
It remains to apply Lemma 39 to each graph reasoning task by defining graph instance encodings X
that encode D ISJ(a, b) in the solution to the task for some r = Θ(N ).
Proof of Theorem 38. For each task, we provide a pair of “hard instances:” one with constant degree
and one with constant diameter, in order to obey different notions of graph simplicity.
Graph connectivity task. For both instances, on any disjointness input a, b ∈ {0, 1}r , we define
a graph G = (V , E) with |V | + |E| = O(r) whose edges encode the input, i.e. E = E(a, b). We
define three fixed disjoint sets of potential edges ĒA , ĒB , Ē∗ with |ĒA | + |ĒB | + |Ē∗ | = O(r)
such that E = EA (a) ∪ EB (b) ∪ Ē∗ , where EA (a) ⊂ ĒA is a function of only Alice’s input a and
EB (b) ⊂ ĒB is a function of Bob’s b.
We define an enumeration of all vertices in V and potential edges in ĒA , ĒB , Ē∗ . We first fix their
sizes |V |, |ĒA |, |ĒB |, |Ē∗ | and then index the vertices and edges in the order V , ĒA , ĒB , Ē∗ . That
is,
V = {1, . . . , |V |} ,
Ē∗ = {(ui , vi ) : i ∈ S∗ } , for S∗ = |V | + 1, . . . , |V | + |Ē∗ | ,
ĒA = {(ui , vi ) : i ∈ A} , for A = |V | + |Ē∗ | + 1, . . . , |V | + |Ē∗ | + |ĒA | ,
ĒB = {(ui , vi ) : i ∈ B} , for B = |V | + |Ē∗ | + |ĒA | + 1, . . . , |V | + |Ē∗ | + |ĒA | + |ĒB | .
This implies that the following transformer input X ∈ RN ×d for N = |V | + |S∗ | + |A| + |B| + 1
is a valid node/edge encoding such that inputs XA is a function of a, XB is a function of b, and
X[N ]\(A∪B) is constant:
37
• For i ∈ B, xi represents edge (ui , vi ) if it exists:
(ui , vi , 0, 1, 0) if (ui , vi ) ∈ EB (b),
xi =
(0, 0, 0, 0, 0) otherwise.
• Let |V | = 3r+2, and let nodes 1 and 3r+2 denote the “source” and “sink” node respectively.
That is, xN = (1, 3r + 2, 0, 1, 1).
• Edges between the source node 1 and nodes {2, . . . , r + 1} and between the end node 3r + 2
and {2r + 2, . . . , 3r + 1} always are included. That is,
also solves disjointness. Since N = |V | + |S∗ | + |A| + |B| + 1 = Θ(r), we conclude the proof of
hardness of solving graph connectivity on this instance by applying Lemma 39.
a1 b1
2 5 8
b2
1 3 6 9 11
a3
4 7 10
Figure 5: The constant diameter graph construction for r = 3, A = (1, 0, 1) and B = (1, 1, 0). The
source 1 and sink 11 are connected which is equivalent to D ISJ(A, B) = 1 by construction.
Constant-degree instance: We can define a degree-3 graph G that similarly encodes a disjointness
instance a, b. To do so, we modify the previous construction by replacing the fixed edges E∗ with two
binary trees, each of depth O(log r), between the source and end nodes as roots and Alice and Bob’s
nodes incident to ĒA and ĒB respectively as leaves. The remainder of the construction—including
the encoding of EA (a) and EB (b) and the connectivity analysis—is identical. Since a binary tree with
r leaves has O(r) nodes and edges, the graph G can be similarly encoded as a length-N transformer
input for N = O(r). See Figure 6 for a visualization.
38
b1
a2 b2
S X
a4
Figure 6: The degree 3 graph construction for r = 4, A = (0, 1, 0, 1), B = (1, 1, 0, 0). The source
node S and sink node X are connected, which is equivalent to D ISJ(A, B) = 1 by construction.
Shortest path task. We can modify the graph connectivity constructions G to create a decision
variant of shortest path, rather than graph connectivity. Let D(G) be the length of a path between the
source and sink node in each connectivity construction, and note that D(G) = 4 for the constant-
diameter instance and D(G) = O(log r) for constant-degree. We design a new graph G0 of size O(r)
then appends a path of D(G) + 1 vertices to G. Then, the shortest path between the source and sink
node is D(G) if they are connected and D(G) + 1 if not.
If there exists a transformer f 0 solves shortest path on graphs of the form G0 , then there exists a minor
modification f that solves graph connectivity on graph G.
Cycle check task. We again modify the graph connectivity constructions. We produce G0 by
adding a single edge to G between the source node 1 and the sink node 3r + 2, which ensures that G0
has a cycle if and only if G is connected between the source and sink nodes. Therefore, a transformer
f 0 can only solve check check if there exists a transformer f that solves connectivity.
The bidirectional relationship between transformers and the massively parallel computation (MPC)
distributed computing model of [71] was partially inspired by a similar analogy between GNNs and
the C ONGEST model by [51].
The MPC and C ONGEST distributed computing protocols differ in their models of how messages
are passed between machines. While MPC protocols permit any machine to send a message to any
other machine subject to capacity constraints, C ONGEST protocols operate on a graph topology that
restricts machines to send messages exclusively to and from neighbors. As a consequence, two nodes
in a C ONGEST graph that are separated by a path of length p cannot share information with one
another without Ω(p) rounds of communication; no similar notion of “long-range communication”
exists for MPC.
All message-passing GNNs where each node initially knows only its own features can be simulated
by a C ONGEST protocol whose rounds scales linearly in the depth of the GNN [51]. This reduction
implies a lower bound on solving several parallelizable and search tasks discussed in Appendix A.2.
Theorem 40. Any graph neural network f with width (message size) m and depth L that computes
any of the following tasks on n-node graphs has the following size lower bounds:
39
√
• Subgraph connectivity, minimum spanning forest, minumum cut, shortest path: L m =
√
Ω̃( n).
• Cycle detection, diameter computation: Lm = Ω̃(n).
If L = O(log N ), then for all tasks, m = Ω̃(n).
In contrast, the quantitative bounds in Section 3 establish sharp trade-offs between transformers and
GNNs for parallelizable tasks and suggest a possible equivalence for search tasks.
All parallelizable tasks—including (subgraph) connectivity, minimum spanning forest, minimum cut,
and cycle detection—can be solved by transformers of depth L = O(log n) and width m = O(n )
for any constant ∈ (0, 1) due to Theorem 2. In contrast, Theorem 40 requires that a similar-depth
GNNs have width m = Ω̃(n), and a GNN of comparable width requires depth L = Ω̃(n1− ).
On the other hand, search tasks, such as shortest path and diamter, are only guaranteed to be solvable
by transformers of depth O(log n) and width O(n1+ ) (for graphs with |E| = n2 ) by Theorem 4.
This statement compares to the GNN negative results of Theorem 40.
A relationship between GNNs and the Weisfeiler-Leman heuristic graph isomorphism test [83] (WL
test) further establishes representational limitations of message-passing GNNs. This connection and
the rich literature surrounding it is presented in greater detail by [59].
The 1-WL test is a permutation-invariant test for predicting whether two candidate graphs are
isomorphic that works by first labeling each node with the empty set ∅ and then repeatedly replacing
each label with a multiset of its neighbors’ labels. A hierarchy of WL test variants exists where
the k-WL test maintains a label for each k-tuple of vertices. The failure models of these heuristic
solutions are well-understood; critically, the 1-WL test cannot distinguish between connected and
disconnected graphs.
The abilities of message-passing GNNs without unique node identifiers to determine whether two
graphs are isomorphic are limited by the graphs distinguishable by the 1-WL test [84, 59]. As an
immediate consequence, such GNNs cannot solve graph connectivity at all, unlike transformers,
which can do so with a logarithmic-depth parameter-efficient representation. The relationship is
bidirectional; message-passing GNNs (and transformers as well) admit an efficient approximation of
the 1-WL test.
The effectiveness of these bounds is tempered by the assumption that no node includes identifying
features, which is easily overcome by standard GNN models. The analogy is further limited by the
fact that graph embeddings of transformers must have some (possibly arbitrary) node identifier as
input in order to tokenize a graph using the node/edge encoding without losing the ability to associate
each node with its incident edges. However, the juxtaposition of the 1-WL and C ONGEST-based
limitations on the abilities of GNNs to solve connectivity-based tasks suggests a fundamental gap in
capabilities between models that is apparent in multiple theoretical lenses.
E Experimental Details
E.1 Datasets
We evaluate our model on the diverse graph reasoning tasks presented in GraphQA [24]. We
used the public code of the dataset available at https://github.com/google-research/
google-research/tree/master/graphqa. The code to generate the datasets is licensed un-
der the Apache License, Version 2.0. The tasks in the dataset range in difficulty and encompass the
following categories:
• Graph-level: node counting (counting the number of nodes in a graph), edge counting (counting
the number of edges in a graph), cycle check (determining whether a graph contains a cycle), and
triangle counting (counting the number of triangles in a graph).
• Node-level: node degree (calculating the degree of a given node in a graph).
40
• Edge-level: connectivity (finding if there is a path from one node to another), edge exis-
tence (whether a given edge exists in a graph, and shortest path (finding the length of the shortest
path from one node to another).
The graphs used in the experiments in this paper and the corresponding graph reasoning tasks are
taken from [24]. There are 1, 000 graphs in the original train set, 500 graphs in the dev set, and 500
graphs in the test set. The graphs are generated randomly using Erdős-Rényi (ER) random graph
model [23]. Graph size ranges from 5 to 20 nodes.
Train set statistics. Average number of nodes: 11.90; average number of edges: 37.01; average node
degree: 5.43.
Test set statistics. Average number of nodes: 12.37; average number of edges: 39.79; average node
degree: 5.70.
While random instances of graph reasoning
tasks provide a valuable assessment of the task
complexity on realistic graphs, they do not nec- 800
essarily reflect the “worst case” graph inputs
that convey negative results like Theorem 6 and
Theorem 3. For example, the reduction that es- 600
tablishes that cycle check is “as hard as” graph
connectivity and the consequential logarithmic- 400
depth hardness results hinge on the considera-
tion of graph instances with n nodes and poly-
nomial cycle length. However, as witnessed by 200
Figure 7, the shortest cycles observed in 1000
instances of GraphQA cycle check is almost al- 0
ways of length three, and only 3.2% of instances no cycle 3 4 5 6
are larger. As a consequence, identifying the
existence of a cycle on the GraphQA dataset is Figure 7: Histogram of minimum cycle lengths for
inherently local, which is reflected by a strong cycle check instances.
performance by heuristic-based GNN solutions
(Table 2)—despite the fact that efficient GNNs for worst-case cycle check do not exist (Theorem 40).
For our experiments on the effect of the scale of the number of training data points in the final results
we obtain, we use the open-source code of GraphQA available to generate a larger training dataset
of 100K examples. We follow the original instructions and parameters to create this larger training
dataset.
Model Hyperparameters. We fixed the number of iterations as 1,000,000 and train standard decoder-
only transformers with L = 12 layers, m = 768 embedding dimension, H = 12 heads, learning rate
5 · 10−4 , and dropout 0.1. These models have an approximate parameter count of 60,000,000.
We used random search [9] over the following set of hyperparameters to select a universal architecture
for all tasks: The range provided for the learning rate and dropout rate are [10−4 , 10−1 ] and [0, 0.5].
The number of layers L and embedding dimension m is selected from L ∈ {4, 6, 8, 10, 12, 14, 16}
and m ∈ {192, 384, 576, 768, 960, 1152, 1344, 1536}. We employed the GLU [73] activation as a
non-linearity.
Model Selection. We implemented our model in JAX [26] and used AdamW [43, 50] as the optimizer.
Optimal hyperparameters for each task and model were determined by training on the GraphQAT rain
dataset and evaluating performance on the GraphQADev dataset. The results presented in the paper
are based on the held-out GraphQAT est dataset.
Hardware Acceleration. All experiments were conducted using Google’s TPUv3 and TPUv5e
accelerators [35].
41
E.3 Baseline Results
To rigorously evaluate the performance of transformers on graph reasoning tasks, we compare them
against three established categories of baselines:
1. Prompting-based methods. These methods provide the LLM with a textual descriptions of the
graph and question within the prompt. We consider the following variations and copy the results
from the original papers:
• ZERO - SHOT. In this approach, the model is given a task description and immediately asked to
produce the desired output. No additional examples or demonstrations are provided.
• FEW- SHOT. This approach provides the model with a few examples of the task and their desired
outputs [12]. Unlike traditional training, these examples are included directly in the prompt,
allowing the model to learn and adapt during the inference.
• C OT. Chain-of-thought (CoT) prompting [82] provides examples each showing step-by-step
reasoning, teaching the LLM to generate its own thought processes for tackling new tasks.
• ZERO - COT. Zero-shot CoT [45] builds upon Chain-of-Thought (CoT) prompting by eliminating
the need for training examples. The LLM generates its own step-by-step reasoning process
using a simple trigger phrase like “Let’s think step by step”.
• COT- BAG. BAG prompting [80] extends COT to improve the performance of LLMs on graph-
related tasks by appending “Let’s construct a graph with the nodes and edges first” to the
prompt.
2. Graph-based methods. These models are specifically designed to process graphs as input and
are trained task-specific. They leverage the connections between nodes to learn patterns and
make predictions, making them ideal for tasks where a graph is involved. We use GCN [44],
MPNN [28], and GIN [84] from this category. GraphToken [63] is a GNN-based model that
processes the graph and feed the output of the GNN as soft-tokens to an LLM. We consider both
GNN models with and without identity tokenizations, as the former is known to be theoretically
limited by WL tests, while the latter poses a “fairer” comparison with the positional encodings of
transformers.
3. Transformer models (Ours). The last class of model are task-specific vanilla transformer
models [77]. The 60M transformer-1K model is the one described above trained on 1, 000 training
examples from the GraphQA training set. To investigate the impact of training data scale, we
generated a larger dataset containing 100, 000 examples, ensuring the same distribution as the
original training set by using the official GraphQA code and trained 60M transformer-100K on
that. The 11B transformer (FT)-1K is a vanilla transformer model that is started with a pre-trained
checkpoint of T5 [66] and is fine-tuned on the 1K training dataset. We also include two fine-tuned
PaLM 2 [5] transformers of size XXS and XS. Similar to prompting baselines, this model receives
a textual description of the graph as input to leverage its textual reasoning capabilities.
The results for ZERO - SHOT, ZERO - COT, FEW- SHOT, COT, and COT- BAG are taken from Fatemi et al.
[24]. Results for SOFT- PROMPT and GraphToken are sourced from Perozzi et al. [63].
We independently evaluated GCN, MPNN, and GIN models on these tasks. We used the original
architectures proposed in their respective papers and performed hyperparameter tuning on the
GraphQADev dataset.
Table 4 presents a comprehensive comparison of the graph reasoning capabilities across various
baseline models and our proposed transformer architectures. The results highlight several key findings,
which we summarize below:
Transformers Exhibit Strong Performance on Graph-based Reasoning Problems. While trans-
formers are not explicitly designed for graph reasoning tasks like graph-based models, they demon-
strate surprisingly strong performance in this domain. The results of this study indicate that trans-
formers, despite their versatility as a general architecture, can often match or even surpass specialized
graph models on a variety of graph reasoning benchmarks.
Transformers Excel at Retrieval Tasks. As proved in Theorem 36, retrieval tasks can be solved by
transformers. The obtained results confirm that such tasks are relatively easy for transformers as they
42
Retrieval Tasks Parallelizable Tasks Search Tasks Subgraph Counting
Method Samples Node count Edge count Edge existence Node degree Connectivity Cycle check Shortest path Triangle counting
ZERO - SHOT [24] 21.7 12.4 44.5 14.0 84.9 76.0 11.5 1.5
Prompting
ZERO - COT [24] 14.6 9.4 33.5 10.4 73.5 32.3 33.6 12.7
FEW- SHOT [24] 25.3 12.0 36.8 17.4 79.4 37.4 22.7 3.0
COT [24] 27.6 12.8 42.8 29.2 45.2 58.0 38.6 8.1
COT- BAG [24] 26.9 12.5 37.3 28.0 45.2 52.1 40.4 8.1
Graph-based
GCN (equivariant) [44] 1K 6.4 1.2 47.0 9.8 83.8 83.2 50.2 4.0
GCN (identity tokenization) [44] 1K 8.8 83.8 83.2
GCN (identity tokenization) [44] 100K 42.2 93.2 91.2
MPNN (equivariant) [28] 1K 19.4 16.2 69.2 99.4 94.0 99.0 66.8 30.6
MPNN (identity tokenization) [28] 1K 38.8 94.0 98.6
MPNN (identity tokenization) [28] 100K 100.0 93.8 99.8
GIN (equivariant) [84] 1K 71.2 4.4 71.2 36.2 93.8 98.8 54.0 30.4
GIN (identity tokenization) [84] 1K 37.8 93.4 97.6
GIN (identity tokenization) [84] 100K 100.0 94.0 98.6
GraphToken [63] 1K 99.6 42.6 73.8 96.2 93.2 95.6 63.8 34.8
60M transformer 1K 100.0 100.0 67.6 31.5 92.9 97.1 57.4 33.4
60M transformer 100K 100.0 100.0 96.1 91.7 98.0 98.0 97.2 40.5
Ours
XXS transformer (FT) 1K 100.0 70.6 73.0 31.0 93.6 98.0 60.4 29.0
XS transformer (FT) 1K 100.0 73.2 98.6 50.6 96.6 96.8 60.0 28.6
12B transformer (FT) 1K 100.0 45.0 100.0 68.8 98.4 98.0 92.8 26.0
Table 4: Comparison of various methods in different categories on graph reasoning tasks of GraphQA.
Here, we categorize the tasks using the taxonomy proposed in Section 3.
Train Test
0.6 0.6
0.8 0.6
0.4 0.4
0.7 0.4
0.2 0.2
0.6
0.2
0 2 3 4 5
0 2 3 4 5
10 10 10 10 10 10 10 10 102 103 104 105 102 103 104 105
Connectivity Cycle check Shortest path Triangle counting
1 1 1 1
0.8 0.8
Accuracy
0.6 0.8
0.6
0.4
0.4
0.2 0.6
0.2
0 0.9
102 10 3
10 4
10 102
5
103
10 4
10 5
10 2
10 3
10 4
10 5
102 103 104 105
Training Samples Training Samples Training Samples Training Samples
obtained the full accuracy on most of such tasks. One exception here is the node degree task that
GNNs outperform transformers but still transformers perform relatively well. We discuss why GNNs
outperform transformers well for this task.
Larger Transformers Excel at Solving Search Tasks. As discussed in Theorem 4, transformers
are effective for search tasks, albeit requiring a larger number of parameters compared to retrieval
tasks. This is empirically evident in the comparison between Transformer-1K and Transformer-
1K (pretrained). It’s worth noting that the pretrained transformer used here has 11 billion parameters,
a significant increase from the 1 million parameters in Transformer-1K.
Transformers Excel at Capturing Global Patterns. An interesting observation here is the perfor-
mance gap between Transformers and GNNs across tasks with varying emphasis on local versus
global graph structure. Notably:
43
Retrieval Tasks Parallelizable Tasks Search Tasks Subgraph Counting
Method Samples Node count Edge count Edge existence Node degree Connectivity Cycle check Shortest path Triangle counting
60M transformer 1K 100.0 (0.00) 100.0 (0.00) 67.6 (0.94) 31.2 (1.49) 92.7 (1.01) 96.2 (0.64) 56.7 (1.60) 26.6 (0.73)
60M transformer 100K 100.0 (0.00) 100.0 (0.00) 100.0 (0.00) 93.3 (11.80) 97.4 (0.33) 99.0 (0.14) 97.7 (0.66) 39.1 (0.72)
Table 5: Mean accuracy (and standard deviation) of trained 60M-parameter transformers over five
random seeds.
1. Local Structure: The node degree task, which relies heavily on local node information, is best
handled by MPNN, a GNN-based model.
2. Global Structure: In contrast, tasks like connectivity, triangle counting and shortest path, which
require understanding global graph patterns, are dominated by Transformer models. Notably,
vanilla Transformers achieve a remarkable 45% relative improvement over even much larger
LLMs augmented with GNN-generated soft prompts (GraphToken) on the shortest path task. This
showcases the exceptional ability of Transformers to capture long-range dependencies, a critical
factor in understanding global graph structure.
• Edge count and node count are “easy” retrieval tasks that can be solved perfectly with as
few as 100 samples.
• Edge existence attains near-perfect train and test classification accuracy as the number of
training samples approaches 100,000.
• Connectivity, shortest path, and node degree demonstrate a sharp improvement in evaluation
error as a function of the sample size. These models perfectly fit the training set in most
sample size regimes, but yield a closer correspondence between training and testing error
when trained on 100,000 samples.
• Cycle check and triangle count have persistent gaps between training and testing error and
overfit even in the large sample setting.
While we lack computational resources to train all models multiple times to obtain error bars,
we repeated the full experiments on 60M-parameter transformers with 1000 and 100,000 training
samples on each task with five different random seeds in order to estimate the standard deviation of
the resulting task accuracy and displayed those results in Table 5. With the exception of the node
degree experiments trained on 100,000 samples (which included a single far-outlier), all tasks and
sample regimes had no more than two percentage points of standard deviation. Critically, this implies
the robustness of the superiority of trained transformers over GNNs and prompting methods on
“intrinsically global” tasks, such as graph connectivity.
44
NeurIPS Paper Checklist
1. Claims
Question: Do the main claims made in the abstract and introduction accurately reflect the
paper’s contributions and scope?
Answer: [Yes]
Justification: The empirical and theoretical claims discussed accurately represent the detailed
theoretical analysis and the wide-ranging experimentation on graph reasoning benchmarks.
Guidelines:
• The answer NA means that the abstract and introduction do not include the claims
made in the paper.
• The abstract and/or introduction should clearly state the claims made, including the
contributions made in the paper and important assumptions and limitations. A No or
NA answer to this question will not be perceived well by the reviewers.
• The claims made should match theoretical and experimental results, and reflect how
much the results can be expected to generalize to other settings.
• It is fine to include aspirational goals as motivation as long as it is clear that these goals
are not attained by the paper.
2. Limitations
Question: Does the paper discuss the limitations of the work performed by the authors?
Answer: [Yes]
Justification: Theoretical limitations are discussed in the presentation of the results in
Section 3 and detailed more comprehensively in Appendix A. We discuss limitations of
experimental results in Section 4.
Guidelines:
• The answer NA means that the paper has no limitation while the answer No means that
the paper has limitations, but those are not discussed in the paper.
• The authors are encouraged to create a separate "Limitations" section in their paper.
• The paper should point out any strong assumptions and how robust the results are to
violations of these assumptions (e.g., independence assumptions, noiseless settings,
model well-specification, asymptotic approximations only holding locally). The authors
should reflect on how these assumptions might be violated in practice and what the
implications would be.
• The authors should reflect on the scope of the claims made, e.g., if the approach was
only tested on a few datasets or with a few runs. In general, empirical results often
depend on implicit assumptions, which should be articulated.
• The authors should reflect on the factors that influence the performance of the approach.
For example, a facial recognition algorithm may perform poorly when image resolution
is low or images are taken in low lighting. Or a speech-to-text system might not be
used reliably to provide closed captions for online lectures because it fails to handle
technical jargon.
• The authors should discuss the computational efficiency of the proposed algorithms
and how they scale with dataset size.
• If applicable, the authors should discuss possible limitations of their approach to
address problems of privacy and fairness.
• While the authors might fear that complete honesty about limitations might be used by
reviewers as grounds for rejection, a worse outcome might be that reviewers discover
limitations that aren’t acknowledged in the paper. The authors should use their best
judgment and recognize that individual actions in favor of transparency play an impor-
tant role in developing norms that preserve the integrity of the community. Reviewers
will be specifically instructed to not penalize honesty concerning limitations.
3. Theory Assumptions and Proofs
45
Question: For each theoretical result, does the paper provide the full set of assumptions and
a complete (and correct) proof?
Answer: [Yes]
Justification: The theoretical results follow all of the criteria. The proofs are in Appendix.
Guidelines:
• The answer NA means that the paper does not include theoretical results.
• All the theorems, formulas, and proofs in the paper should be numbered and cross-
referenced.
• All assumptions should be clearly stated or referenced in the statement of any theorems.
• The proofs can either appear in the main paper or the supplemental material, but if
they appear in the supplemental material, the authors are encouraged to provide a short
proof sketch to provide intuition.
• Inversely, any informal proof provided in the core of the paper should be complemented
by formal proofs provided in appendix or supplemental material.
• Theorems and Lemmas that the proof relies upon should be properly referenced.
4. Experimental Result Reproducibility
Question: Does the paper fully disclose all the information needed to reproduce the main ex-
perimental results of the paper to the extent that it affects the main claims and/or conclusions
of the paper (regardless of whether the code and data are provided or not)?
Answer: [Yes]
Justification: We comprehensively detail the models used and the training hyper-parameters
in the appendix. The training dataset is generated by a publicly available source code.
Guidelines:
• The answer NA means that the paper does not include experiments.
• If the paper includes experiments, a No answer to this question will not be perceived
well by the reviewers: Making the paper reproducible is important, regardless of
whether the code and data are provided or not.
• If the contribution is a dataset and/or model, the authors should describe the steps taken
to make their results reproducible or verifiable.
• Depending on the contribution, reproducibility can be accomplished in various ways.
For example, if the contribution is a novel architecture, describing the architecture fully
might suffice, or if the contribution is a specific model and empirical evaluation, it may
be necessary to either make it possible for others to replicate the model with the same
dataset, or provide access to the model. In general. releasing code and data is often
one good way to accomplish this, but reproducibility can also be provided via detailed
instructions for how to replicate the results, access to a hosted model (e.g., in the case
of a large language model), releasing of a model checkpoint, or other means that are
appropriate to the research performed.
• While NeurIPS does not require releasing code, the conference does require all submis-
sions to provide some reasonable avenue for reproducibility, which may depend on the
nature of the contribution. For example
(a) If the contribution is primarily a new algorithm, the paper should make it clear how
to reproduce that algorithm.
(b) If the contribution is primarily a new model architecture, the paper should describe
the architecture clearly and fully.
(c) If the contribution is a new model (e.g., a large language model), then there should
either be a way to access this model for reproducing the results or a way to reproduce
the model (e.g., with an open-source dataset or instructions for how to construct
the dataset).
(d) We recognize that reproducibility may be tricky in some cases, in which case
authors are welcome to describe the particular way they provide for reproducibility.
In the case of closed-source models, it may be that access to the model is limited in
some way (e.g., to registered users), but it should be possible for other researchers
to have some path to reproducing or verifying the results.
46
5. Open access to data and code
Question: Does the paper provide open access to the data and code, with sufficient instruc-
tions to faithfully reproduce the main experimental results, as described in supplemental
material?
Answer: [No]
Justification: While code is not made available, the data can be generated using the GraphQA
dataset linked in the appendix. Full training details are made available, which makes the
code easily reproducible. The code will be open sourced upon acceptance of the paper.
Guidelines:
• The answer NA means that paper does not include experiments requiring code.
• Please see the NeurIPS code and data submission guidelines (https://nips.cc/
public/guides/CodeSubmissionPolicy) for more details.
• While we encourage the release of code and data, we understand that this might not be
possible, so “No” is an acceptable answer. Papers cannot be rejected simply for not
including code, unless this is central to the contribution (e.g., for a new open-source
benchmark).
• The instructions should contain the exact command and environment needed to run to
reproduce the results. See the NeurIPS code and data submission guidelines (https:
//nips.cc/public/guides/CodeSubmissionPolicy) for more details.
• The authors should provide instructions on data access and preparation, including how
to access the raw data, preprocessed data, intermediate data, and generated data, etc.
• The authors should provide scripts to reproduce all experimental results for the new
proposed method and baselines. If only a subset of experiments are reproducible, they
should state which ones are omitted from the script and why.
• At submission time, to preserve anonymity, the authors should release anonymized
versions (if applicable).
• Providing as much information as possible in supplemental material (appended to the
paper) is recommended, but including URLs to data and code is permitted.
6. Experimental Setting/Details
Question: Does the paper specify all the training and test details (e.g., data splits, hyper-
parameters, how they were chosen, type of optimizer, etc.) necessary to understand the
results?
Answer: [Yes]
Justification: All training and evaluation details and model hyper-parameters are included in
the paper appendix.
Guidelines:
• The answer NA means that the paper does not include experiments.
• The experimental setting should be presented in the core of the paper to a level of detail
that is necessary to appreciate the results and make sense of them.
• The full details can be provided either with the code, in appendix, or as supplemental
material.
7. Experiment Statistical Significance
Question: Does the paper report error bars suitably and correctly defined or other appropriate
information about the statistical significance of the experiments?
Answer: [Yes]
Justification: The main claims of the paper are supported by error bars and analysis present
in Table 5.
Guidelines:
• The answer NA means that the paper does not include experiments.
• The authors should answer "Yes" if the results are accompanied by error bars, confi-
dence intervals, or statistical significance tests, at least for the experiments that support
the main claims of the paper.
47
• The factors of variability that the error bars are capturing should be clearly stated (for
example, train/test split, initialization, random drawing of some parameter, or overall
run with given experimental conditions).
• The method for calculating the error bars should be explained (closed form formula,
call to a library function, bootstrap, etc.)
• The assumptions made should be given (e.g., Normally distributed errors).
• It should be clear whether the error bar is the standard deviation or the standard error
of the mean.
• It is OK to report 1-sigma error bars, but one should state it. The authors should
preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis
of Normality of errors is not verified.
• For asymmetric distributions, the authors should be careful not to show in tables or
figures symmetric error bars that would yield results that are out of range (e.g. negative
error rates).
• If error bars are reported in tables or plots, The authors should explain in the text how
they were calculated and reference the corresponding figures or tables in the text.
8. Experiments Compute Resources
Question: For each experiment, does the paper provide sufficient information on the com-
puter resources (type of compute workers, memory, time of execution) needed to reproduce
the experiments?
Answer: [Yes]
Justification: The type of the accelerator used in the experiments is in Appendix.
Guidelines:
• The answer NA means that the paper does not include experiments.
• The paper should indicate the type of compute workers CPU or GPU, internal cluster,
or cloud provider, including relevant memory and storage.
• The paper should provide the amount of compute required for each of the individual
experimental runs as well as estimate the total compute.
• The paper should disclose whether the full research project required more compute
than the experiments reported in the paper (e.g., preliminary or failed experiments that
didn’t make it into the paper).
9. Code Of Ethics
Question: Does the research conducted in the paper conform, in every respect, with the
NeurIPS Code of Ethics https://neurips.cc/public/EthicsGuidelines?
Answer: [Yes]
Justification: We reviewed the guidelines and found no relevant violations to our work.
Since the experimental datasets are synthetically generated, privacy concerns are moot.
Guidelines:
• The answer NA means that the authors have not reviewed the NeurIPS Code of Ethics.
• If the authors answer No, they should explain the special circumstances that require a
deviation from the Code of Ethics.
• The authors should make sure to preserve anonymity (e.g., if there is a special consid-
eration due to laws or regulations in their jurisdiction).
10. Broader Impacts
Question: Does the paper discuss both potential positive societal impacts and negative
societal impacts of the work performed?
Answer: [NA]
Justification: The paper is focused on reasoning capabilities of transformers using theoretical
and empirical evaluation. To the best of our knowledge, there is no societal impact for the
work.
Guidelines:
48
• The answer NA means that there is no societal impact of the work performed.
• If the authors answer NA or No, they should explain why their work has no societal
impact or why the paper does not address societal impact.
• Examples of negative societal impacts include potential malicious or unintended uses
(e.g., disinformation, generating fake profiles, surveillance), fairness considerations
(e.g., deployment of technologies that could make decisions that unfairly impact specific
groups), privacy considerations, and security considerations.
• The conference expects that many papers will be foundational research and not tied
to particular applications, let alone deployments. However, if there is a direct path to
any negative applications, the authors should point it out. For example, it is legitimate
to point out that an improvement in the quality of generative models could be used to
generate deepfakes for disinformation. On the other hand, it is not needed to point out
that a generic algorithm for optimizing neural networks could enable people to train
models that generate Deepfakes faster.
• The authors should consider possible harms that could arise when the technology is
being used as intended and functioning correctly, harms that could arise when the
technology is being used as intended but gives incorrect results, and harms following
from (intentional or unintentional) misuse of the technology.
• If there are negative societal impacts, the authors could also discuss possible mitigation
strategies (e.g., gated release of models, providing defenses in addition to attacks,
mechanisms for monitoring misuse, mechanisms to monitor how a system learns from
feedback over time, improving the efficiency and accessibility of ML).
11. Safeguards
Question: Does the paper describe safeguards that have been put in place for responsible
release of data or models that have a high risk for misuse (e.g., pretrained language models,
image generators, or scraped datasets)?
Answer: [NA]
Justification: This is because the datasets used are synthetically generated and reflect random
graphs, no risks of data leakage are present.
Guidelines:
• The answer NA means that the paper poses no such risks.
• Released models that have a high risk for misuse or dual-use should be released with
necessary safeguards to allow for controlled use of the model, for example by requiring
that users adhere to usage guidelines or restrictions to access the model or implementing
safety filters.
• Datasets that have been scraped from the Internet could pose safety risks. The authors
should describe how they avoided releasing unsafe images.
• We recognize that providing effective safeguards is challenging, and many papers do
not require this, but we encourage authors to take this into account and make a best
faith effort.
12. Licenses for existing assets
Question: Are the creators or original owners of assets (e.g., code, data, models), used in
the paper, properly credited and are the license and terms of use explicitly mentioned and
properly respected?
Answer: [Yes]
Justification: The papers are cited, the link to the source codes used is states along with their
license.
Guidelines:
• The answer NA means that the paper does not use existing assets.
• The authors should cite the original paper that produced the code package or dataset.
• The authors should state which version of the asset is used and, if possible, include a
URL.
• The name of the license (e.g., CC-BY 4.0) should be included for each asset.
49
• For scraped data from a particular source (e.g., website), the copyright and terms of
service of that source should be provided.
• If assets are released, the license, copyright information, and terms of use in the
package should be provided. For popular datasets, paperswithcode.com/datasets
has curated licenses for some datasets. Their licensing guide can help determine the
license of a dataset.
• For existing datasets that are re-packaged, both the original license and the license of
the derived asset (if it has changed) should be provided.
• If this information is not available online, the authors are encouraged to reach out to
the asset’s creators.
13. New Assets
Question: Are new assets introduced in the paper well documented and is the documentation
provided alongside the assets?
Answer: [NA]
Justification: No new assets is introduced in this paper. We have used public data and no
code has been released yet.
Guidelines:
• The answer NA means that the paper does not release new assets.
• Researchers should communicate the details of the dataset/code/model as part of their
submissions via structured templates. This includes details about training, license,
limitations, etc.
• The paper should discuss whether and how consent was obtained from people whose
asset is used.
• At submission time, remember to anonymize your assets (if applicable). You can either
create an anonymized URL or include an anonymized zip file.
14. Crowdsourcing and Research with Human Subjects
Question: For crowdsourcing experiments and research with human subjects, does the paper
include the full text of instructions given to participants and screenshots, if applicable, as
well as details about compensation (if any)?
Answer: [NA]
Justification: No human subjects were involved.
Guidelines:
• The answer NA means that the paper does not involve crowdsourcing nor research with
human subjects.
• Including this information in the supplemental material is fine, but if the main contribu-
tion of the paper involves human subjects, then as much detail as possible should be
included in the main paper.
• According to the NeurIPS Code of Ethics, workers involved in data collection, curation,
or other labor should be paid at least the minimum wage in the country of the data
collector.
15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human
Subjects
Question: Does the paper describe potential risks incurred by study participants, whether
such risks were disclosed to the subjects, and whether Institutional Review Board (IRB)
approvals (or an equivalent approval/review based on the requirements of your country or
institution) were obtained?
Answer: [NA]
Justification: No human subjects were involved.
Guidelines:
• The answer NA means that the paper does not involve crowdsourcing nor research with
human subjects.
50
• Depending on the country in which research is conducted, IRB approval (or equivalent)
may be required for any human subjects research. If you obtained IRB approval, you
should clearly state this in the paper.
• We recognize that the procedures for this may vary significantly between institutions
and locations, and we expect authors to adhere to the NeurIPS Code of Ethics and the
guidelines for their institution.
• For initial submissions, do not include any information that would break anonymity (if
applicable), such as the institution conducting the review.
51