0% found this document useful (0 votes)
46 views106 pages

The Tensor Networks Anthology

Uploaded by

Pietro Silvi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
46 views106 pages

The Tensor Networks Anthology

Uploaded by

Pietro Silvi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 106

SciPost Phys. Lect.

Notes 8 (2019)

The Tensor Networks Anthology: Simulation techniques


for many-body quantum lattice systems
Pietro Silvi1,2? , Ferdinand Tschirsich1 , Matthias Gerster1 , Johannes Jünemann4,5 ,
Daniel Jaschke1,3 , Matteo Rizzi4 and Simone Montangero1,6,7

1 Institute for Complex Quantum Systems and Center for Integrated


Quantum Science and Technologies, Universität Ulm, D-89069 Ulm, Germany.
2 Institute for Theoretical Physics, Innsbruck University, A-6020 Innsbruck, Austria.
3 Department of Physics, Colorado School of Mines, Golden, Colorado 80401, USA.
4 Institut für Physik, Johannes Gutenberg-Universität, D-55128 Mainz, Germany.
5 Graduate School Materials Science in Mainz, D-55128 Mainz, Germany.
6 Theoretische Physik, Universität des Saarlandes, D-66123 Saarbrücken, Germany.
7 Dipartimento di Fisica e Astronomia, Università degli Studi di Padova, I-35131 Italy.

? [email protected]

Abstract
We present a compendium of numerical simulation techniques, based on tensor network
methods, aiming to address problems of many-body quantum mechanics on a classical
computer. The core setting of this anthology are lattice problems in low spatial dimen-
sion at finite size, a physical scenario where tensor network methods, both Density Ma-
trix Renormalization Group and beyond, have long proven to be winning strategies. Here
we explore in detail the numerical frameworks and methods employed to deal with low-
dimensional physical setups, from a computational physics perspective. We focus on
symmetries and closed-system simulations in arbitrary boundary conditions, while dis-
cussing the numerical data structures and linear algebra manipulation routines involved,
which form the core libraries of any tensor network code. At a higher level, we put the
spotlight on loop-free network geometries, discussing their advantages, and presenting
in detail algorithms to simulate low-energy equilibrium states. Accompanied by discus-
sions of data structures, numerical techniques and performance, this anthology serves as
a programmer’s companion, as well as a self-contained introduction and review of the
basic and selected advanced concepts in tensor networks, including examples of their
applications.

Copyright P. Silvi et al. Received 03-11-2017


This work is licensed under the Creative Commons Accepted 04-03-2019
Check for
Attribution 4.0 International License. Published 18-03-2019 updates

Published by the SciPost Foundation. doi:10.21468/SciPostPhysLectNotes.8

Contents
1 Introduction 3
1.1 Structure of the Anthology 4
1.2 Short History of Tensor Networks for Quantum Many-Body Problems 4
1.3 Tensor Network States in a Nutshell 6
1.4 Many-Body Statistics 7

1
SciPost Phys. Lect. Notes 8 (2019)

1.5 Typically Addressed Problems 8


1.6 Accessible Information in Tensor Networks 8
1.7 The Tensor Network Gauge 10

2 Tensors and Tensor Operations 11


2.1 Tensors 11
2.2 Basic Tensor Operations 13
2.2.1 Tensor Manipulations 13
2.2.2 Tensor Linear Algebra Operations 15

3 Abelian Symmetric Tensors and Operations 22


3.1 Quantum Symmetries in a Nutshell: Abelian Groups 23
3.2 Encodable Symmetries in Tensor Networks 25
3.3 Constructing Symmetric Tensors 25
3.3.1 Abelian Symmetric Links 26
3.3.2 Abelian Symmetric Tensors 26
3.3.3 Inner Structure of Abelian Symmetric Tensors 27
3.3.4 Symmetric Link- and Tensor Data Handling 28
3.4 Symmetric Tensor Networks 30
3.5 Symmetric Tensor Operations 31
3.5.1 Link and Tensor Manipulations 31
3.5.2 Symmetric Tensor Linear Algebra Operations 41
3.5.3 Computational Cost Estimation 45

4 Loop-Free Tensor Networks 48


4.1 General Aspects of Loop-Free Tensor Networks 48
4.2 Gauges in Loop-Free Tensor Networks 50
4.2.1 Gauge Freedom 51
4.2.2 Unitary Gauge 52
4.2.3 Canonical Gauge 54
4.2.4 Symmetry Gauge 58
4.2.5 Real Gauge 59
4.3 Preparation of Symmetric TN Ansatz States 59
4.3.1 Randomized Symmetric Ansatz States 60

5 Ground State Search with Loop-Free Tensor Networks 61


5.1 Ground State Search Algorithm for the Adaptive-Gauge Loop-Free TN 62
5.1.1 High-Level Operations 62
5.1.2 Ground State Search Algorithm in Detail 71
5.1.3 Optimizer – Strategies 75
5.1.4 Optimizer – Execution 77
5.1.5 Low Excited States 79
5.1.6 Variational Compression, Addition and Orthogonalization 80
5.2 A Practical Application: Binary Tree Tensor Networks 81
5.3 Binary Tree Tensor Networks – Examples 85

6 Conclusions 91
6.1 Outlook and Perspectives 91

Appendix 92
A Anti-Unitary Symmetries and Time-Reversal 92

2
SciPost Phys. Lect. Notes 8 (2019)

B Graphical Notation Legend 92

References 94

1 Introduction
The understanding of quantum many-body (QMB) mechanics [1] is undoubtedly one of the
main scientific targets pursued by modern physics. Discovering, characterizing, and finally
engineering novel, exotic phases of quantum matter [2] will be the key for the development of
quantum technologies [3], quantum computation and information platforms [4], and even lead
to an enhanced, non-perturbative understanding of the building blocks of reality, including high
energy settings [5–7]. The research development is hindered by the inherent difficulty of such
problems, by the extremely small amount of exactly solvable models [8] and the limitations
posed by the various approximate approaches that have been proposed so far [9].
From the computational point of view, the obstacle for addressing quantum many-body
problems is the exponential growth in the dimension of the configurations space with the
number of elementary constituents of the system (number of quantum particles, lattice sites,
and so forth). Exact diagonalization techniques [10] are thus restricted to small system sizes,
and inapt to capture the thermodynamical limit of QMB systems, which characterizes the
macroscopic phases of matter.
Approaches relying on a semi-classical treatment of the quantum degrees of freedom [11],
such as mean field techniques e.g. in the form of the Hartree–Fock method [12], have been
applied with various degrees of success. It is well known, however, that while they are an
accurate treatment in high spatial dimensions, they suffer in low dimensions [13], especially in
1D where entanglement and quantum fluctuations are so important that it is not even possible
to establish true long-range order through continuous symmetry breaking [14].
On the other hand, stochastic methods for simulating QMB systems at equilibrium, such as
Monte Carlo techniques, have been extremely successful [15, 16]. They have the advantage of
capturing both quantum and statistical content of the system by reconstructing the partition
functions of a model, and have been adopted on a wide scale. Difficulties arise when the
stochastic quasi-probability distribution to be reconstructed is non-positive, or even complex
(sign problem) [17, 18].
An alternative approach to address these intrinsic quantum many-body problems is the
construction of a computationally manageable variational ansatz, which is capable of capturing
the many-body properties of a system while being as unbiased as possible. Tensor Network (TN)
states fulfill this criterion [19–24]: By definition, their computational complexity is directly
controlled by (or controls) their entanglement. In turn, TN states can span the whole quantum
state manifold constrained by such entanglement bounds [25–27]. Consequently, TNs are
an efficient representation in those cases where the target QMB state exhibits sufficiently
low entanglement. Fortunately, tight constraints on the entanglement content of the low-
energy states, the so-called area laws of entanglement have been proven for many physical
systems of interest, making them addressable with this ansatz [28–37]. TN states have the
advantage of being extremely flexible and numerically efficient. Several classes of tensor
networks, displaying various geometries and topologies, have been introduced over the years,
and despite their differences, they share a vast common ground. Above all, the fact that any
type of QMB simulation relies heavily on linear algebra operations and manipulation at the
level of tensors [38].

3
SciPost Phys. Lect. Notes 8 (2019)

Accordingly, in this anthology we will provide a detailed insight into the numerical algebraic
operations shared by most tensor network methods, highlighting useful data structures as well
as common techniques for computational speed-up and information storage, manipulation
and compression. As the focus of the manuscript is on numerical encoding and computational
aspects of TN structures, it does not contain an extended review of physical results or a
pedagogical introduction to the field of TNs in general. For those two aspects, we refer the
reader to the excellent reviews in Refs. [19, 39].
The anthology is thus intended both as a review, as it collects known methods and strategies
for many-body simulations with TNs, and as a programmer’s manual, since it discusses the
necessary constructs and routines in a programming language-independent fashion. The
anthology can be also used as a basis for teaching a technical introductory course on TNs. We
focus mostly on those tensor network techniques which have been previously published by some
of the authors. However, we also introduce some novel strategies for improved simulation: As
a highlight, we discuss ground state search featuring an innovative single-tensor update with
subspace expansion, which exhibits high performance and is fully compatible with symmetric
TN architectures.

1.1 Structure of the Anthology


The anthology is organized as follows: In Sec. 1 we review shortly the historical background and
some general theoretical concepts related to tensor network states. In Sec. 2 we characterize
the data structures which build the TN state, and we list the standard linear algebra tools
employed for common tasks. A brief review on embedding symmetries into the TN design is
provided in Sec. 3, alongside numerical techniques to exploit the (Abelian) symmetry content
for computational speed-up and canonical targeting. In Sec. 4 we introduce tensor network
structures without loops (closed cycles in the network geometry) and explain how we can
explore possible gauges to give these states favorable features. A generalization of the Density
Matrix Renormalization Group (DMRG) applying to tensor networks of this type is detailed in
Sec. 5, as well as instructions for practical realizations of the algorithm. We draw our conclusions
and sketch the outlook for further research in Sec. 6. A diagrammatic representation of the
arrangement of the anthology contents is shown in Fig. 1, where the logical dependencies
among the topics are highlighted, and depicted as embedded boxes. This diagram, in turn,
reflects the hierarchy among the modules of an ideal TN programming library: Using this
hierarchy, the outmost component (ground state search) can be easily replaced with several
other physically relevant tasks, such as real-time dynamics (see also Sec. 1.5).

1.2 Short History of Tensor Networks for Quantum Many-Body Problems


The Density Matrix Renormalization Group (DMRG), introduced in the early ’90s [40, 41],
was developed as a reformulation of the numerical renormalization group [42, 43], where
now the renormalized, effective, degrees of freedom were identified in the density matrix
eigenspaces instead of in the lowest energy manifold of real space subsystems [44]. Eventually
it was realized that DMRG can be recast as a minimization algorithm over the class of finitely-
correlated states, which admit a formulation as Matrix Product States (MPS) [25, 45–47]. In
time it became clear that the reason for the success of DMRG is the limited entanglement
present in ground states of short-range interacting Hamiltonians, as it obeys the area laws
of entanglement [28–33, 37], and such scaling is well captured by MPS for one-dimensional
systems. On the one hand, such reformulation into MPS language [27, 39] opened a clear path
for substantial improvements and expansions of the DMRG technique. Prominent examples are:
concepts for out-of-equilibrium evolution [26, 48–51], direct addressing of the thermodynamic
limit [52–56], exploration of finite temperatures [57, 58], and direct access to the density

4
SciPost Phys. Lect. Notes 8 (2019)

Introduction
(Sec. 1)

Numerical linear
algebra kernel

Tensors and Tensor Operations


(Sec. 2)

Symmetric Tensors (Abelian)


(Sec. 3)

Tensor Networks (Loop-Free)


(Sec. 4)

TN Algorithms:
Ground States with Loop-Free Tensor Networks
(Sec. 5)

Conclusion & Outlook


(Sec. 6)

Figure 1: Organization of the anthology. In an embedded-boxes structure, the


introduction is followed by a self-consistent treatment of elementary tensors and
tensor-operations plus their extension in presence of symmetries. On this foundation
we build up and investigate (loop-free) tensor networks and algorithms. The parts
about symmetries are optional and can be skipped. As this manuscript is intended
as a practical anthology, the structure is chosen such that it can be directly mapped
into numerical libraries, organized in the same hierarchy, starting from a kernel of
numerical linear-algebra routines which are widely available and thus not covered in
this anthology.

5
SciPost Phys. Lect. Notes 8 (2019)

(a) (b) i5
i4
i5 i6 α4 [3]
T
T [4]
i4
i6 α3 α2
T = α5

i3
i1 T [1] T [2]
i2 α1
i1 i3
i2

Figure 2: (a) The coefficients of a QMB wave function (here with N = 6 constituents)
can be understood as the elements of a tensor T with N links. Graphically, we
represent a tensor by a circle-like object (node) with an open link for every physical
site s. Fixing the indices is for all links s then identifies the element Ti1 ...iN . (b) A
possible TN decomposition of the system: The state is now represented by Q = 4
smaller tensors T [q] , which are connected through L = 5 auxiliary links. The original
tensor object is obtained once again by contracting tensors over shared indices, which
are represented by the links connecting them. For a detailed introduction to the
graphical notation, see Sec. 2.

of states and spectral properties [59, 60]. On the other hand, the discovery encouraged the
engineering of new classes of tailored variational wave functions with the common desirable
feature of storing the QMB information in compact tensors, i.e. easy to handle numerical objects,
and controlling the entanglement through a network of virtual connections [19]. Noteworthy
classes of tensor networks that have been proposed towards the quantum many-body problem
are: Projected Entangled Pair States (PEPS) [61,62], weighted graph states [63], the Multiscale
Entanglement Renormalization Ansatz (MERA) [64–67], branching MERA [68,69], Tree Tensor
Networks (TTN) [70–75], entangled-plaquette states [76], string-bond states [77], tensor
networks with graph enhancement [78, 79] and continuous MPS [80]. Additionally, various
proposals have been put forward to embed known stochastic variational techniques, such as
Monte Carlo, into the tensor network framework [76, 81–86], to connect tensor networks to
quantum error correction techniques [87], and to construct direct relations between few-body
and many-body quantum systems [88]. The numerical tools we will introduce in the next two
sections indeed apply to most of these TN classes.

1.3 Tensor Network States in a Nutshell


A tensor network state is a tailored quantum many-body wave function ansatz. For a discrete
system with N components, the QMB state can be written
X
|ΨQMB 〉 = Ti1 ,...,iN |i1 , . . . , iN 〉, (1)
i1 ,...,iN

where {|is 〉}s is a canonical basis of the subsystem s. The complex coefficients of this Hilbert
space wave function, Ti1 ...iN , determine the state uniquely.
The TN prescription expresses the amplitudes tensor as the contraction of a set of smaller
tensors T [q] over auxiliary indices (sometimes also called virtual indices), namely
[1] [2] [Q]
X
Ti1 ...iN = T{i} ,{α} T{i} ,{α} · · · T{i} ,{α} , (2)
1 1 2 2 Q Q
α1 ...α L

6
SciPost Phys. Lect. Notes 8 (2019)

for a network with Q nodes and L links. Each link connects and is shared by two ten-
sors only, see Fig. 2(b). Here a tensor T [q] can possess any number of physical indices
{i}q = {is : s is physical link at node q} as well as any number of auxiliary indices

{α}q = αη : η is auxiliary link at node q . However, to make the ansatz computationally effi-
cient for QMB states, the number of links (indices) connected to it should not scale with the
system size N . Similarly, the auxiliary dimensions Dη = #{αη } (where the symbol # denotes
the cardinality of a set) should be non-scaling quantities. Conversely, the number of tensors Q
will scale with N .
Throughout this anthology, we use a graphical, compact notation to express tensors and
tensor networks, and the typical operations performed on them. These diagrams allow us to
recast most of the equations in a form which is less verbose, often clearer, and mnemonically
convenient, such as in Fig. 2 which represents Eq. (2). More details about the diagrammatic
representation of TNs will be provided in Sec. 2 and Sec. 3.
It is easy to show [89] that the entanglement entropy E of the TN state |ΨQMB 〉 under system
bipartitions satisfies rigorous bounds, which depend on the TN geometry. In fact, the TN state is
a wave function ansatz whose only constraint is a direct control of its entanglement properties,
and is thus considered an only slightly biased ansatz. For the same reason, it is typical to
express TN states in real space, where the N components are the lattice sites: it is within
this setup that one can exploit the area laws and successfully work with TN states having low
entanglement content [21, 23, 30, 31, 37, 90]. Tensor network state descriptions in momentum
space representation have also been proposed and implemented in simulations [91–95].

1.4 Many-Body Statistics


One of the major features of the tensor network ansatz is its flexibility: In fact, it can be adapted
for several types of local degrees of freedom and quantum statistics. The typical scenarios dealt
with are the following:

(i) Spin systems - Spin lattices are the natural setting for tensor networks. The quantum
spins sit on a discrete lattice and their local Hilbert space {|is 〉} has a finite dimension d
fixed by the problem under study; e.g. for lattice sites with spin l: d = 2l + 1.

(ii) Bosonic lattice - Boson commutation rules are analogous to the ones for spins, but
normally, bosons do not have a compact representation space for their local degree of
freedom. Unless some inherent mechanism of the model makes the bosons d-hardcore,
an artificial cutoff of the local dimension d must be introduced. This is, however, usually
not dramatic, since typically the repulsive interactions make highly occupied states very
improbable and the consistency of results can be checked by enlarging d slightly.

(iii) Fermionic lattice - The standard way of dealing with fermionic models is mapping them
into a spin system via a Jordan–Wigner (JW) transformation [96, 97]. Doing so is very
convenient for short-range interacting (parity-preserving) 1D models, as the JW-mapping
preserves the short-range nature. Other techniques to encode the fermionic exchange
statistics directly in the tensor network contraction rules are known [98, 99], but it
is argued that these methods are more useful in 2D, where the JW-mapping does not
preserve interaction range.

(iv) Hybrid lattices - Tensor networks can also handle hybrid theories which include both
bosons and fermions. In this scenario, the two species usually live in two distinct
sublattices. Lattice gauge systems such as the lattice Yang–Mills theories are a textbook
example of this setting [5]: In these models, matter can be fermionic, while the gauge
field is typically bosonic [100–107].

7
SciPost Phys. Lect. Notes 8 (2019)

(v) Distinguishable particles - When quantum particles are distinguishable and discrete,
they can be considered as the lattice sites, and both their motional and internal degrees
of freedom may be described in a first quantization fashion. This scenario appears, for
instance, when studying phononic non-linearities in trapped ion systems [108]. In this
representation, the motional modes of each particle are often truncated to a compact
dimension by selecting d orbitals {|i〉} to act as local canonical basis, e.g. those with
lowest local energy, or the solutions of a mean field approach [109, 110].

It is important to mention that the TN ansatz can be extended to encompass variational


states for quantum field theories, for instance via the continuous-MPS formalism [80, 111].
However, this scenario lies beyond the purpose of this paper and will not be discussed here.

1.5 Typically Addressed Problems


Tensor network states are routinely employed to tackle the following physical problems on 1D
lattices:

(i) Ground states of interacting Hamiltonians, usually (but not necessarily) built on two-
body interactions. For the model to be manifestly 1D, the interaction has to be either finite-
range or decrease with range sufficiently fast [112]. Ground state search algorithms based
on TN technology can usually be extended to target other low energy eigenstates [113].

(ii) Non-equilibrium dynamics of arbitrary QMB states under unitary Hamiltonian evolution.
The Hamiltonian itself can be time-dependent, thus encompassing quenches [114, 115],
quasi-adiabatic regimes [116, 117], and even optimal control [118]. The most com-
mon strategies to perform such dynamics on the TN states rely either on a Suzuki–
Trotter decomposition of the unitary evolution (TEBD, tDMRG) [48, 49] or on a Time-
Dependent Variational Principle (TDVP) on the TN class [51], however, alternative routes
are known [119–122]. Clearly, these techniques perform best when describing pro-
cesses with small entanglement growth: Either short time scale quenches [114, 115],
quasi-adiabatic quenches [116, 117, 123], or controlled dynamics [124]. Long time scale
quenches are also accessible for small system sizes [125].

(iii) Annealing and Gibbs states. Strategies to achieve imaginary-time evolution are analo-
gous to real-time unitary evolution, but in order to reconstruct the Gibbs ensemble at
finite temperatures, one must either perform several pure state simulations or extend
tensor networks to mixed states [57, 58, 126].

(iv) Open system dynamics, usually undergoing a Lindblad master equation evolution
(Markovian scenario), with local dissipative quantum channels. Besides the need to
reconstruct mixed states, this setup has the additional requirement of designing a scheme
for implementing a master equation dynamics into the TN ansatz, either by direct inte-
gration [57, 58, 126] or by stochastic unravelling [127].

Depending on the type of problem one aims to address and on the boundary condition
under consideration (open or periodic), some TN designs are preferable to others. In Sec. 5 we
will describe in detail an algorithm for ground state search based on loop-free TNs.

1.6 Accessible Information in Tensor Networks


Being able to effectively probe and characterize a relevant quantum many-body state is just as
important as determining the state itself to understand the underlying physics and the emergent
properties of the system. The TN architectures that we discuss here are meant to describe

8
SciPost Phys. Lect. Notes 8 (2019)

finite systems, while the thermodynamical limit properties are obtained via extrapolation. The
thermodynamical limit can also be directly addressed, e.g. by means of infinite homogeneous
tensor networks (such as i-MPS/i-PEPS [52, 53, 128] or i-TTN/i-MERA [75, 129, 130]), but
these methods will not be discussed here.
Typical quantities of interest for tensor network ansatz states are:
(i) State norm 〈ΨQMB |ΨQMB 〉 − The capability of extracting the norm of a tensor network
in polynomial computational time as a function of the number of tensors in the network
itself is often simply referred to as efficient contractibility.
Loop-free tensor networks, such as MPS and TTN, have this property built-in (assuming
that the link dimensions are bounded from above), as we will remark in Sec. 4. Some
tensor networks with loops, such as MERA, also have this property as a result of their
specific geometry and algebraic constraints. In general, however, this is not true —
the PEPS being the textbook example of a non-efficiently contractible tensor network
(although approximate, stochastic methods for PEPS contraction are known [81]). In
loop-free TNs and holographic TNs (MERA, branching MERA) the calculation of the state
norm can even be made computationally cheaper by imposing specific TN gauge rules,
as we will see in Sec. 4.
0 0
(ii) State overlap 〈ΨQMB |ΨQMB 〉 − Provided that the two states |ΨQMB 〉 and |ΨQMB 〉 are given
in the same TN geometry and that the resulting network is efficiently contractible (which
is always the case for loop-free TNs), this quantity can be easily computed and the
numerical cost is the same as for the state norm. Here, however, the TN gauge invariance
cannot be exploited to further reduce the costs.

(iii) Expectation values of tensor product observables 〈ΨQMB | s O[s] |ΨQMB 〉 − Efficient
N

calculation of these quantities relies on the fact that every single-site operator can be
effectively absorbed into the TN state without altering the network geometry, and thus
performing local operations preserves the efficient contractibility property. This statement
is true whether the single-site operator O[s] is homogeneous or site-dependent. Useful
subclasses of this observable type contain:

• Local observables 〈O[s] 〉, useful for reconstructing reduced density matrices and
identifying local order.
• Correlation functions 〈O[s] O[s+`] 〉, important for detecting long-range order. This
can, of course, be extended to n-points correlators.
• String observables 〈O[s] O[s+1] . . . O[s+`−1] O[s+`] 〉, useful for pinpointing topological
properties (e.g. the Haldane order [131]) and boundary order, as well as to imple-
ment long-range correlations in Jordan–Wigner frameworks.

(iv) Expectation values of operators in a tensor network form − When the observable
itself can be expressed in a TN formalism, it is natural to extend the notion of efficient TN
contractibility to the 〈 state | operator | state 〉 expectation value. This is, for instance,
the case when measuring the expectation value of a Matrix Product Operator (MPO) over
a matrix product state, which is indeed an efficiently contractible operation [132].

(v) Entanglement properties − Any bipartition of the network tensors corresponds to a


bipartition of the physical system. Via contraction of either of the two sub-networks
it is then possible to reconstruct the spectrum of the reduced density matrix of either
subsystem, and thus extract the entanglement entropy, i.e. the von Neumann entropy
of the reduced density matrix, and all the Rényi entropies, ultimately enabling the
study of critical phenomena by means of bipartite [73, 133], and even multipartite
entanglement [134].

9
SciPost Phys. Lect. Notes 8 (2019)

(a) (c)
αη 0
T [q] T [q ] T [q] Y Y −1
0
T [q ]

=
(b) αη αη 0
T̃ [q] T̃ [q ]
=
βη αη βη0
Y −1
Y

Figure 3: Gauge transformation: For a given tensor network (a), the link η between
0
tensors T [q] and T [q ] is replaced by a matrix Y and its inverse (b). Contracting Y and
Y −1 into their respective neighboring tensors (c) gives an alternative description of
the tensor network without changing the physical content of the stored information.

1.7 The Tensor Network Gauge


Tensor network states contain some form of information redundancy with respect to the quantum
state they describe. This redundancy translates conceptually into a set of linear transformations
of the tensors in the network which leave the quantum state, and thus all the physically relevant
quantities, unchanged [26, 27, 39, 70]. Since these operations have no impact at the physical
level, they are often referred to as gauge transformations, in analogy to field theories. The gauge
invariance relies on the fact that performing local invertible manipulations on the auxiliary TN
spaces does not influence the physical degrees of freedom. We will therefore introduce the
concept of gauge transformation here and use it extensively in the rest of the anthology.
Specifically, let η be a given virtual link of the tensor network state from Eqs. (1) and (2),
connecting node q to q0 , with auxiliary space P dimension Dη (see Fig. 3(a)). Then, given a
Dη0 × Dη (left-) invertible matrix Yαη ,α0η (i.e. α0 Yα,α
−1
0 Yα0 ,α00 = δα,α00 ), the tensor network state
0
is unchanged if T [q] and T [q ] transform as
[q] [q] [q]
X
T{i},{α} → T̃{i},{α} = Yαη ,βη T{i},{α} , (3a)
\η ,αη \η ,αη \η ,βη
βη
[q0 ] [q0 ] [q0 ]
X
T{i 0 },{α0 } → T̃{i 0 },{α0 } = Yβ−1,α T{i 0 },{α0 } , (3b)
\η ,αη \η ,αη η η \η ,βη
βη

where {α}\η collects the indices of all the virtual links of the tensor, except η. This operation
is simply the contraction of Y into T [q] through link η and simultaneous contraction of its
0 0
inverse matrix Y −1 into T [q ] through link η: it leaves the composite tensor T [q+q ] , obtained
0
by the contraction of T [q] and T [q ] over link η, invariant. Since the QMB state |Ψ〉 depends on
0 0
T [q] and T [q ] only via T [q+q ] , it is insensitive to this transformation. Similarly, the network
geometry is unchanged (as long as Y acts on a single auxiliary link). Such an operation is
sketched using the diagrammatic language in Fig. 3.
Combining multiple (single-link) gauge transformations as in Eq. (3), even on different
links, leads to a larger class of transformations, which still leave the quantum many-body state
invariant, and can be performed efficiently numerically. In fact, these algebraic manipulations
turn out to be very useful tools when designing computational algorithms which operate on
TN states. In many architectures (most notably in DMRG/MPS) adapting the gauge during
runtime is a fundamental step for achieving speed-up and enhanced precision.
We will review linear algebra methods which implement the gauge transformations in

10
SciPost Phys. Lect. Notes 8 (2019)

Sec. 2, and their symmetric tensor counterparts in Sec. 3. We will explore the advantages of
employing gauge conditions in Sec. 4 and Sec. 5.

2 Tensors and Tensor Operations


Tensor networks organize numerical data in the form of interconnected tensors. In practice,
these tensors are represented by floating-point numerical data in the computer memory (i.e.
RAM or hard drive); they should be stored and organized in a way as to efficiently perform
linear algebra operations on them. This section formalizes the notion of a tensor and introduces
a set of basic operations serving as the toolbox for designing various tensor network geometries
and respective algorithms.
In the following, we first introduce tensors formally and graphically (Sec. 2.1). We then
present operations acting on these tensors with technical details and extend the graphical
notation (Sec. 2.2).

2.1 Tensors
We define a tensor T with n links (n ∈ N0 ) as an n-dimensional array Ti1 i2 ...in of complex valued
elements. It can be seen as a mapping
T : I1 × I2 × · · · × In → C : (i1 , i2 , . . . , in ) 7→ Ti1 i2 ...in , (4)
where each complex scalar element Ti1 i2 ...in is accessed by a tuple (an ordered integer list) of n
indices i1 , i2 , . . . , in . The index i r at position r takes values in the index set I r = {1, . . . , d r } ⊂ N
and is referred to as the r-th link of the tensor. For numerical purposes, the link dimension
d r := #I r is always finite. Tensors and links allow intuitive graphical representations as network
nodes and network edges respectively (see Fig. 4). We do not need to formally distinguish
virtual and physical links at the level of basic operations for links and tensors. The distinction
between virtual and physical links is instead treated at the level of the network, as we will see
later in Sec. 4.
A zero-link tensor is a complex scalar. A one-link tensor of dimension d1 is a d1 -dimensional
complex vector. A two-link tensor of dimensions d1 × d2 is a (rectangular) complex matrix
with d1 rows and d2 columns, and so forth. A detailed example of a three-link tensor is given
d1 ×···×dn
in Fig. 5. In general, Q the n-link tensor is an element of the complex space C with
#T := dim(T ) = r d r independent components. We notice that in certain cases, e.g. for
problems with time-reversal symmetric Hamiltonians, the tensor elements can be restricted to
real values (see also Appendix A).
For our following discussion, we assume, in accordance with standard programming, that
tensor elements are stored in a contiguous section of linearly addressed computer memory
(array). Specifically, each element Ti1 i2 ...in can be stored and addressed in such a linear array
by assigning a unique integer index (address offset) to it (see an example in Fig. 5). A simple
way to implement this is by the assignment
n
X r−1
Y
offset(i1 , i2 , . . . , in ) := (i r − 1) dk , (5)
r=1 k=1
which is known as “column-major” ordering. There are, indeed, other models for storing
elements. For efficient coding, one should consider the respective advantages of the ordering
method (many programming languages and compilers use an intrinsic row major ordering,
after all). In Sec. 3, we extend the definition of the tensor object: We will also consider tensors
with additional inner structure (given by symmetry), generalizing diagonal or block-diagonal
matrices.

11
SciPost Phys. Lect. Notes 8 (2019)

(a) (b) (c) n (d)


in−1 ...
in
A B C i3
D
i2
a m i1

Figure 4: Graphical representation of (a) a zero-link tensor (i.e. a scalar), (b) a


one-link tensor (a vector), (c) a two-link tensor (a matrix) and (d) a tensor with an
arbitrary number n of links. Indices of a tensor (i.e. its links) can be denoted with
arbitrary symbols; the ordering of the open link-ends reflects the order of the indices
in the tensor.

 p 
offset T i1 i2 i3 i1 i2 i3 a b j 1/ 2
 0 
p  
0 1/ 2 1 1 1 1 1 1  0 
 
1 0 2 1 1 2 1 2
 
 0 
 
2 0 3 1 1 3 1 3  0 
p p 
 2/ 3
 
3 0 1 2 1 1 2 4 Vj =  
4 0 2 2 1 2 2 5  0 
p p 
 0 

5 2/ 3 3 2 1 3 2 6  p 
 1/ 3 
6 0 1 1 2 1 3 7
 p 
 −1/ 2 
 
7 0 2 1 2 2 3 8  
p  1 
8 1/ 3 3 1 2 3 3 9
0
p
9 −1/ 2 1 2 2 1 4 10 (c)
10 1 2 2 2 2 4 11
11 0 3 2 2 3 4 12
i1 a
(a)

 p p  T ⇒ M ⇒ V
1/ 2 0 0 −1/ 2
Mab = 0 0 0 1 
p p p i3
0 2/ 3 1/ 3 0 i2 b j
(b) (d)

Figure 5: Example of a real-valued three-link tensor Ti1 i2 i3 ∈ R of dimensions d1 = 3,


d2 = d3 = 2. (a) In the table we list all tensor elements and their indices (in column-
major order) using the ordering (offset) from Eq. (5). In the additional columns
to the right we also added the respective indices after the tensor is reshaped into a
matrix and a vector (see fusing in Sec. 2.2.1). (b) The matrix Mab = Ti1 (i2 i3 ) , has
row- and column indices a = i1 and b = fuse(i2 , i3 ) of dimensions da = d1 = 3 and
d b = d2 d3 = 4. (c) The vector V j = T(i1 i2 i3 ) has a combined index j = fuse(i1 , i2 , i3 ),
according to Eq. (8), of total dimension d j = #T = 12. (d) A graphical represen-
P tensor. The tensor T could
tation of the three-link p e.g. encode the three normal-
ized states |Ψi1 〉 = i2 ,i3 Ti1 i2 i3 |i2 i3 〉, i.e. |Ψ1 〉 = 1/ 2 (|11〉 − |22〉), |Ψ2 〉 = |22〉 and
p p 
|Ψ3 〉 = 1/ 3 |21〉 + 2|12〉 , spanning a subspace in the two-qubit Hilbert space
{|i2 , i3 〉}. Their expansions can be read from the rows of M, with columns corre-
sponding to (i2 , i3 ) = (1, 1), (2, 1), (1, 2), (2, 2) from left to right.

12
SciPost Phys. Lect. Notes 8 (2019)

2.2 Basic Tensor Operations


All tensor network algorithms involve certain basic operations on the tensors, ranging from
initialization routines, element-wise operations such as multiplication by a scalar, complex
conjugation T ∗ , or tensor addition and subtraction, to very general unary or binary opera-
tions [135]. This section presents selected examples from two important categories of tensor
operations, namely basic manipulations, and more involved linear algebra operations (see
e.g. Fig. 6). Tensor manipulations usually operate on the elements or the shape of a single
tensor, including position, number and dimensions of the links. They often reallocate and
reshuffle data within the computational memory, but generally require no FLoating point
OPerations (FLOPs). On the other hand, linear algebra operations such as contractions and
decompositions require FLOPs, such as complex addition and multiplication.
Tensor network algorithms, operating extensively on the level of tensors, will spend the
majority of their runtime with basic tensor operations. In numerical implementations, it is
therefore important that these operations are highly optimized and/or rely on robust and
efficient linear algebra libraries such as the Linear Algebra PACKage (LAPACK) [136], and the
Basic Linear Algebra Subprograms (BLAS) [137–139]. Here we will not enter such level of
detail and instead outline the algorithms and give estimates on overall computational cost and
scaling behavior. Nevertheless, our advice, for readers who want to program tensor network
algorithms, is to dedicate specific functions and subroutines to each of the following operations.
It is worth mentioning that there are various open-source numerical libraries which handle
tensor manipulations with similar interfaces, such as the iTensor [140], Uni10 [141] or the
TNT library [142].

2.2.1 Tensor Manipulations


Link Permutation. Reordering the link positions i1 , . . . , in of a tensor T by moving a link at
position l in tensor T to a position k = σ(l) in the new (permuted) tensor T p :
p
Ti = Tiσ(1) ,...,iσ(n) . (6)
1 ...in

As an example, imagine a four-link tensor and a permutation σ(1) = 2, σ(2) = 3, σ(3) = 1,


p
σ(4) = 4. The tensor after permutation then has elements Tc,a,b,d = Ta,b,c,d (see Fig. 6(a)).
Note that our definition describes an active permutation, while other conventions prefer passive
permutations where the roles of input- and output tensors are reversed. For tensors with
two links, permutations correspond to matrix transpositions, Mab T
= M ba . Carrying out a
permutation has a computational complexity bounded by O(#T ), i.e. linear in the number of
elements that must be rearranged in memory.

Link Fusion. Fusing links is an operation that combines m − k + 1 adjacent links at positions
k, . . . , m into a single new fused link j ∼ (ik , ik+1 , . . . , im ) which replaces the original links in
the fused tensor (see Fig. 6(b)):

Ti10 ...ik−1 j im+1 ...in = Ti1 ...ik−1 (ik ...im ) im+1 ...in . (7)

The combined index j takes values in the Cartesian product of all fused links J = Ik × · · · × Im .
Typically we use an integer enumeration

j = fuse(ik , ik+1 , . . . , im ) := 1 + offset(ik , ik+1 , . . . , im ) (8)

within all possible combinations of the fused link indices. Fusing can be realized in negligible
computational cost O(1) as with the proper enumeration no elements have to be moved in
memory (cf. Eq. 5). As an important example, a tensor can be converted into a matrix by

13
SciPost Phys. Lect. Notes 8 (2019)

(a)
d d (d)
c c
T = T p b e
A B
a b b
a d
c a
σ

=
C e
d d
(b)
c
F a d
T = Tf
e
a b e a
(e)

T c
a c
b B
d d =
(c)
a b Ts
T S = b
d
c c A
a
a

Figure 6: Basic tensor operations. Common TN operations can be separated into


elemental steps including: (a) Link permutation. Graphically, the permutation of
links of a tensor T is shown by intertwining the links (σ, shaded area) to obtain the
new order in T p . (b) Link fusion: Two or more links of a tensor T are joined to form
a single one in the new tensor T f . This can be seen as a contraction with a fuse-node
F (see Sec. 2.2.2). (c) Link splitting: a single link of a tensor T is split into two or
more resulting in a tensor T s . The splitting is equivalent to the contraction with a
split-node S . (d) Tensor contractions and (e) tensor decompositions.

simultaneously fusing both sides of the link-bipartition 1, . . . , k − 1 and k, . . . , n into row and
column indices a and b respectively (see also Fig. 5):
Mab = T(i1 ...ik−1 ) (ik ...in ) . (9)
Through link fusion, matrix operations become directly applicable to tensors with more than
two links. Important examples include multiplication and factorization, which are discussed
in Sec. 2.2.2. Fusing arbitrary combinations of non-adjacent tensor links can be handled by
permuting the selected links into the required positions k, . . . , m first.

Link Splitting. Splitting of a link at position k of the original tensor T is the inverse operation
to fusing. It produces a tensor T 0 with one original link jk replaced (via inversion of Eq. (8))
by m − k + 1 new links ik , . . . , im of dimensions dk0 , . . . , dm
0
, their product being equal to the
dimension dk of the original link,
Ti 0 ...i = Ti1 ...ik−1 jk im+1 ...in , (10)
1 k−1 (ik ...im ) im+1 ...in

14
SciPost Phys. Lect. Notes 8 (2019)

so that jk = fuse(ik , . . . , im ) (Fig. 6(c)). This operation is uniquely defined once the split
dimensions dk0 , . . . , dm
0
are set. Note that similar to fusing, splitting operations can also be
performed in non-scaling time O(1), but may require a link permutation after execution. A
special case of splitting is the insertion of an extra one-dimensional virtual link (dummy link) at
any position, which may be useful when a certain tensor shape must be matched. Imagine, for
example, that some TN-algorithm interface formally asks for a matrix-product operator (see
also Sec. 5.1), but all we have is a simple product of non-interacting local operators. Joining
those by dummy links allows one to match the generic interface in an elegant manner. A similar
situation arises when computing a direct product of two tensors, for which we could simply
employ the contraction operation, given we join both tensors by a dummy link, as proposed in
Sec. 2.2.2.

Link Index Remapping: Compression, Inflation, and Reorder. A typical class of shape
operations on a tensor are those that rearrange the index values from one (or more) of the
connected links. These link remapping manipulations either reorder the index values, or
increase/decrease the link dimension itself. Compressing one or more links connected to a
tensor Ti1 ...in means selecting a subset of their allowed indices and discarding the rest, thus
actually reducing the link dimensions. Link compression is carried out extracting a subset of
the linked tensor elements by means of subtensor readout

S j1 ... jn := Ti1 ...in for jr ∈ J r ⊆ I r , r ∈ {1, . . . , n} (11)

and we call S a subtensor of T . The subtensor still has n links, and each may have smaller
dimensions, depending on the number #J r of indices kept per link. Subtensors arise naturally
as a generalization of block-diagonal matrices for instance due to the consideration of symmetry
constraints, as we will see in Sec. 3.5.1. More specifically, let a link index i r undergo a mapping

M r : J r → I r : jr 7→ i r . (12)
If M r is injective, the result is a subtensor of reduced link dimension(s). Furthermore, if M r is
invertible, it is a permutation, which allows one to access elements in a different order. The
latter is useful for sorting or aligning elements (e.g. after the fuse-operation of 2.2.1) into
contiguous blocks for fast memory access. Finally, the inflation (padding) of a link corresponds
to an increase of its dimension, and compression is its (left-) inverse operation. Inflation is
performed by overwriting a subtensor S in a larger tensor holding the padded elements (see
Sec. 5.1.3), by means of subtensor assignment

Ti1 ...in := S j1 ... jn for jr ∈ J r ⊆ I r , r ∈ {1, . . . , n} . (13)

The remaining elements in the larger tensor T must then be initialized by the user, e.g. set to
zeros.
Manipulating, overwriting or extracting subtensors from or within larger tensors is an
integral part of variational subspace expansion techniques [143, 144]. Further use cases for
subtensor assignment and readout arise in the construction of MPOs [132] and other operations
listed at the end of Sec. 2.2.2.

2.2.2 Tensor Linear Algebra Operations


Tensor Contraction. Tensor contractions are a generalization of matrix multiplications to
arbitrary numbers  b Whenever two tensors A and B with links {a1...na} and {b1...m
 a of indices: } of
dimensions d1...n and d1...n have a subset of k links in common — say dn−k+ j = d j =: d sj b

for j = 1 . . . k — we can contract over these shared links s j . The result is a contracted tensor

15
SciPost Phys. Lect. Notes 8 (2019)

j
k
A i k
l
contract
m C
o n
r
p B r intermediate q m
q steps

a
a
m
k
m i
k l
i
(a) l (b) j
permute j fuse
s ⇒ s
links I n links
⇒ o ⇒
p r o n
q p r
q
b
b
k m
a
a k
(c) (d) m (e)
multiply split permute

matrices links links II
⇒ ⇒ r ⇒
b q

b
q r

Figure 7: Intermediate steps in the contraction of two tensors, realized through


matrix-multiplication. First, the input tensors A and B are brought into correct
matrix-shape, which in general involves (a) permuting and (b) fusing their links. After
their matrix-multiplication (c) is carried out, the remaining original links are restored
into the expected order by a final (d) splitting and (e) permutation operation. Those
permutation, fuse- and split-steps are very common among tensor-operations that
rely on matrix-operations internally.

16
SciPost Phys. Lect. Notes 8 (2019)

(a) i m i m i m
(b) (c)
k
A B = A B = C
T
j n j n j n
T i1 i2 i3
a
fuse T∗
⇒ C = M

b partial trace norm


direct product

Figure 8: Variants of contractions. (a) Direct product of two matrices, or two-link


tensors Ai j and Bmn . By attaching a one-dimensional “dummy” link to both tensors,
we can indicate the contraction Ci jmn = Ai j · Bmn . Formally, a subsequent (column-
major, Eq. (8)) fusion over the permuted link-bipartition Mab = C(mi)(n j) gives us the
Kronecker product M = A ⊗ B . (b) The partial trace. (c) The Hilbert–Schmidt norm
obtained by full contraction with the complex conjugate T ∗ .

C spanned by the remaining n + m − 2k links {a1...n−k , bk+1...m }. Graphically we represent a


contraction by simply connecting the links to be contracted over (Fig. 6(d)). The elements of C
are obtained by multiplying the respective elements of A and B and summing over the shared
links X
Ca1 ...an−k bk+1 ...bm = Aa1 ...an−k s1 ...sk Bs1 ...sk bk+1 ...bm . (14)
s1 ...sk

Equation (14) can be cast into a matrix multiplication after fusing the shared links into a
single link s = fuse(s1 , . . . , sk ) of dimension ds = d1s · · · dks and the remaining links of tensors A
and B into row- and column indices a and b of dimensions da = d1a · · · dn−k a
and d b = dk+1
b
· · · dmb ,
respectively: X
Cab = Aas Bsb . (15)
s

In the most common scenario for a tensor contraction, the links are not ordered as in
Eq. (14). We then proceed as follows (see also Fig. 7): We first permute the links, so that
tensors A and B appear as in Eq. (14). Then the link fusion is carried out, followed by the
matrix multiplication as in Eq. (15). Finally we split the row and column links again and
possibly permute a second time to obtain the desired link positions at C .
The contraction can be a costly operation in terms of computational complexity. Carried
out straightforwardly as written in Eq. (14), it scales with the product of the dimensions of all
links involved (considering shared links only once) as O(da ds d b ).
Implementing the tensor contraction via the matrix multiplication of Eq. (15) has a twofold
advantage: It enables efficient parallelization [138, 139, 145] and it allows us to exploit
optimized matrix multiplication methods [136, 137, 146]. As an example for the latter, the
complexity of a contraction of square matrices can be improved from O(d 3 ) to O(d 2.376 )
when using optimized linear algebra methods [147]. However, an overhead for the required
permutations must be considered. A careful initial analysis of the arguments can often help to
avoid some of the permutations, or to replace a permutation on a large input or output tensor
at the expense of one or two additional permutations on smaller ones. Such optimization is
possible by exploiting the full freedom we have in rearranging the internal order in which links

17
SciPost Phys. Lect. Notes 8 (2019)

are fused into column- and row indices and also by utilizing implicitly transposed arguments
available with many matrix-multiplication implementations.

Special Cases of Contractions. Some operations which are very common to tensor net-
work architectures and algorithms can be regarded as special instances of tensor contractions.
We list the most relevant ones in this paragraph.

(i) Fusing k links of a tensor T can be achieved by contracting a sparse k + 1 link tensor, the
fuse-node F ≡ Fi1 ...ik , j defined by
¨
1 if j = fuse(i1 , . . . , ik )
Fi1 ...ik , j = (16)
0 otherwise

over those links i1 . . . ik of T that shall be fused into j. Analogously, the split-operation
can be carried out by contracting a corresponding split-node S := F † over j, with

F j,(i = F(i1 ...ik ), j in a matrix notation. In this formalism, it is clear that fusing and
1 ...ik )
splitting are mutually inverse operations, i.e. F † F ≡ δ j, j 0 and FF † ≡ r δir ,ir0 , since F
Q

is unitary by construction.
While possible, fusing or splitting by such a contraction is less efficient than the enumer-
ation procedure of Sec. 2.2.1. Nevertheless, fuse-nodes play an important role in the
context of symmetric tensors where fusing and splitting is carried out according to group
representation theory (see Sec. 3). We also employ them to encode the fusion graphically,
as in Fig. 6(b).

(ii) The Kronecker-product or matrix-direct product

Ca1 ...an−k bk+1 ...bm = Aa1 ...an−k B bk+1 ...bm (17)

corresponds to the case when no links are shared in the contraction of Eq. (14). This
operation can be achieved (up to a permutation) by attaching a shared dummy link on
both tensors and contracting over it, as visualized in Fig. 8(a) with a practical application
in Fig. 11(b).

(iii) In analogy to the multiplication with a diagonal matrix, we can contract a “diagonal”
two-link tensor Di j = δi, j λi over a single link into some other tensor T . This operation
can be carried out in time O(#T ) scaling linearly with the number of tensor elements.
We represent those (often real) diagonal matrices or weights λi over a link graphically by
a thin bar as depicted in Fig. 10, emphasizing their sparsity.

(iv) The (partial) trace is a variant of the contraction, involving a single tensor only. It is
depicted in Fig. 8(b). This operation sums up all tensor elements that share similar
indices on k specified link-pairs at the same tensor
X
tr[T ]i2k+1 ...in = T(s1 ...sk ) (s1 ...sk ) i2k+1 ...in , (18)
s1 ...sk

where we contracted the first k links at positions 1 . . . k with the following k links at
positions 2k + 1 . . . 2k. If the tensor T has n links, the resulting tensor tr[T ] has n − 2k
links. If k = n/2, the tensor trace is equivalent to a matrix trace over the given bipartition
of links. The complexity of the operation equals the product of dimensions of all paired
links and all remaining links (if present), thus O(ds1 · · · dsk di2k+1 · · · din ).

18
SciPost Phys. Lect. Notes 8 (2019)

(v) A tensor norm, defined as a generalization of the Hilbert–Schmidt- or Frobenius-norm by


vX
u
kT k = t T T∗ (19)
i1 ...in i1 ...in
i1 ...in

can be computed in O(#T ). As shown in Fig. 8(c), this is equivalent to the contraction
of T with its element-wise complex conjugate T ∗ over all links.

(vi) In principle, we can also regard permutation as a contraction


p
Tj = Ti1 ...in · σi1 ...in , j1 ... jn , (20)
1 ... jn

where σi1 ...in , j1 ... jn = δi1 jσ(1) · . . . · δin jσ(n) (see Fig. 6(a)). However, this is not the most
efficient numerical approach.
As a general remark, we stress that finding the optimal strategy to contract multiple
tensors efficiently is usually a difficult problem [148]; nevertheless it is a fundamental issue,
as contractions are often the bottleneck of the computational costs. This question ultimately
encourages the usage of specific network geometries whose optimal contraction strategies are
known (see e.g. Ref. [149]). Such geometries will be further discussed in Sec. 5.

Tensor Decompositions. A tensor decomposition takes a single tensor as input, and replaces
it with two or more output tensors (see Fig. 6(e)). For every link of the input tensor, there will
be a similar link attached to one of the output tensors. Additional “internal” links may connect
the output tensors among each other.
For a tensor decomposition to be meaningful, we demand that the contraction of those
output tensors over their shared internal links restores the input tensor within a certain precision.
The simplest example for a tensor decomposition is the inversion of a contraction: we take
the contracted tensor as input and replace it by the tensors that we originally contracted. In
practice, however, we are more interested in special decompositions that equip the output
tensors with favorable properties, such as an advantageously small size or certain isometries
that come in handy when working with larger tensor networks (see Sec. 4).
The most notable decompositions are the QR- and the singular value decomposition (SVD),
both instances of so-called matrix factorizations [150]. As a recipe for such an (exact) tensor
decomposition, first choose a bipartition of the tensor links i1 , . . . , i r and i r+1 , . . . , in of an n-link
tensor and fuse them into single row and column indices a and b respectively. If a different
combination of fused links is required, we apply a permutation beforehand. We then decompose
the matrix
T(i1 ...ir ) (ir+1 ...in ) = Tab (21)
into factor matrices Aak , Bk b , and possibly a diagonal matrix λk , such that the matrix multipli-
cation yields
dk
X
Aak λk Bkb = Tab . (22)
k=1
Finally, we split the links we initially fused to restore the outer network geometry. Thus we
obtained tensors that contract back to T .

QR Decomposition. Following the procedure of Eqs. (21) and (22), the QR decomposition
splits the tensor T into two tensors Q and R such that
dk dk
fuse QR split
X X
T(i1 ...ir ) (ir+1 ...in ) = Ta b = Qak Rkb = Q(i1 ...ir )k Rk(ir+1 ...in ) , (23)
k=1 k=1

19
SciPost Phys. Lect. Notes 8 (2019)

(a) QR (b) k
i4
i4 R Q
T = i2
k i3 i1
i3 =
i1 δk,k0
Q Q∗
i2

i1 k0
i2

Figure 9: QR decomposition and isometry. (a) We split T at the dashed line into R
and an isometry Q via QR-decomposition. (b) The isometry is depicted by a gray-
white filling: the gray partition indicates the links over which we have unitarity. In
other words, the link  index k at the opposite partition enumerates an orthonormal
set of dk vectors q (i ...i ) = Qk(i1 ...ir ) .
k
1 r

SVD i3
i3 λ W
T k
=
i2
i2
V
i1

i1

Figure 10: The singular value decomposition of a tensor produces an intermediate


link k and two semi-unitary tensors W and V (depicted by shade over the orthonormal
link bipartition, cf. Eq. (25)). The singular values appear on the intermediate link
as a non-negative diagonal matrix or weights λk , represented by a bar to highlight
the sparsity. Removing singular values from the link (thereby compressing it into a
smaller dimension dk ) does not disturb the tensor isometries.

where Rk b is an upper-trapezoidal matrix andP Qak is∗ a (semi-)unitary matrix with respect to k
(or left isometry), meaning that it satisfies a Qak Qak0 = δk,k0 (as shown in Fig. 9). Finally, T
is equal to the contraction of Q and R over the intermediate link k, which has the minimal
dimension
dk = min {da , d b } = min {d1 · . . . · d r , d r+1 · . . . · dn }. (24)

Singular Value Decomposition. The SVD is defined as


dk dk
fuse SVD split
X X
T(i1 ...ir )(ir+1 ...in ) = Ta b = Vak λk Wkb = V(i1 ...ir )k λk Wk(ir+1 ...in ) , (25)
k=1 k=1

see also Fig. 10. It yields a set of non-negative, real valued singular values {λk } and (contrary
to the QR) two semi-unitary (or isometric) matrices V and W . Semi-unitarity means that the
tensors fulfill V † V = 1 and WW † = 1, while this does not necessarily hold for VV † and W † W .
The singular values can be organized in a diagonal matrix (as we saw in Sec. 2.2.2) and are

20
SciPost Phys. Lect. Notes 8 (2019)

(a)

exp A B = exp A⊗B = exp M = exp M = exp A ⊗ B

(b)

1
exp A B ≈ 1 1 + A B +··· + n! × An Bn = A0 B0

k k
Subtensors: A0 := p1 × Ak−1 B k−1 × p 1 =: B0
k−1! k−1!

Figure 11: Two strategies for exponentiating a Kronecker product of two tensors A ⊗ B
(i.e. hosting the matrix elements of two local operations on a lattice). Both demonstrate
the power of combining several basic tensor operations with a unified graphical
notation. (a) The exact exponential exp {A ⊗ B } is formed from first expanding the
Kronecker product to a full four-link tensor (by contracting over an inserted dummy
link, cf. Fig. 8 (a)). Fused into a square matrix M, this tensor can be exponentiated
with matrix numerical routines. Finally, we split the fused links into the original
configuration. (b) The Taylor-series expansion up to (small) order n can be rewritten
as a contraction of smaller tensors A0 and B 0 over a third, shared link k of dimension
dk = n + 1. To this end, scaled powers of A, B are inscribed as subtensors into A0 , B 0
at the respective index values of k, as depicted. Note that in practice, conditions on
numerical stability apply.

basically pure weight factors on the intermediate link. Conventional algorithms


 on QR- and
singular value decomposition display an algorithm complexity of O dmax dk [151], with
2

dmax = max {d1 · . . . · d r , d r+1 · . . . · dn } (26)

being the maximum dimension of the fused links.


Even though the SVD carries a computational cost overhead with respect to the QR decom-
position (a direct comparison is shown in Fig. 48), it is often useful to have both available,
since the SVD has several advantages:

(i) In common tensor network algorithms, the singular values reveal entanglement properties.
This is true, in particular, for loop-free TNs (see Sec. 4).

(ii) Singular values equal to numerical zeros can be removed from the intermediate link.
The dimension dk of the latter is then reduced to the actual matrix rank of Tab , and the
factors Vak , Wk b become full rank matrices which can be extracted via subtensor readout
(c.f. Eq. (11)) from the retained intermediate link indices.

(iii) Small singular values can be approximated with zeros, effectively further reducing the
link dimension dk . This truncation, basically equivalent to a link compression as we
discussed in Sec. 2.2.1, usually generates an error on the “compressed” state, but in most

21
SciPost Phys. Lect. Notes 8 (2019)

tensor network geometries the discarded values can be used to estimate an upper error
bound (see e.g. Ref. [39]).

(iv) Both output tensors are semi-unitary (isometries) with respect to k, regardless of the
amount of singular values kept. This may prove advantageous in establishing specific
tensor network gauges, as we will see in Sec. 4.

By discarding negligible singular values (compressing the tensor network) we relax Eq. (25) to
an approximated equality, but it can be motivated physically and has a significant impact on
performance of TN algorithms (see Sec. 5). If the number of singular values hereby removed
from the intermediate link becomes large compared to dk , one can in addition take advantage of
efficient approximate algorithms for matrix-factorizations [152–156], some of which offer good
control over the errors arising both from the stochastic method itself and from the discarded
singular values.

Other Operations. Other relevant linear algebra operations are those restricted to square
matrices, such as diagonalization, which is typically carried out either exactly [136] or via
approximate methods such as Lanczos [157], Jacobi–Davidson [158], or Arnoldi [159], as well
as exponentiation, usually performed via Taylor or Padé expansion [160] or using the EXPOKIT
package [161]. Note that tensor operations, especially when implemented with unified and
compatible interfaces, are a powerful and flexible numerical toolbox. For example, in Fig. 11
we combine several basic operations to exponentiate local interaction operators, a central task
in time-evolution algorithms like TEBD [48].

3 Abelian Symmetric Tensors and Operations


Symmetries are ubiquitous in nature, and both in classical and quantum mechanics they
provide an essential advantage by reducing the full problem into a set of decoupled, simpler
sub-problems: Symmetries are generally expressed in terms of conserved quantities (Noether’s
theorem); in quantum systems, such conserved quantities are identified by operators commuting
with the Hamiltonian. In forming a multiplicative group, all these operators can then be block-
diagonalized simultaneously, along with the Hamiltonian.
In numerical contexts, the role of symmetries is twofold: they provide a substantial compu-
tational speed-up, and they allow for precise targeting of symmetry sectors. More precisely,
one can study any variational problem by restricting only to those states having a well-defined
quantum number. This is a valid, robust, more direct approach than using a “grand canonical”
framework (e.g. selecting the average number of particles by tuning the chemical potential),
which is computationally more expensive as it requires to consider multiple symmetry sectors
at once.
Here, we first review the quantum framework for Abelian symmetries and provide examples
of the Abelian groups appearing in typical problems (Sec. 3.1). We then discuss symmetries
that can be encoded in tensor networks (Sec. 3.2). In the rest of this section, we upgrade the
tensor data structures and operations introduced in the previous section to include Abelian
symmetry content. We discuss, in particular, how to organize the symmetry information in
the tensors (Sec. 3.3) and how networks built from such tensors display the same symmetry
properties (Sec. 3.4). Finally, we review how tensor operations benefit from the inclusion of
symmetry information (Sec. 3.5).

22
SciPost Phys. Lect. Notes 8 (2019)

3.1 Quantum Symmetries in a Nutshell: Abelian Groups


In quantum mechanics, a symmetry is a typically unitary representation U(g) of a group
G of transformations protected by the dynamics, i.e. commuting with the Hamiltonian H:
[U(g), H] = 0 ∀ g ∈ G (we briefly discuss anti-unitary representations in Appendix A). As a
consequence, there exists an eigenbasis of the Hamiltonian such that each energy eigenstate
belongs to a single irreducible representation (irrep) subspace.
Specifically, symmetric eigenstates will be of the form |Ψ`,m,∂ 〉 and satisfy the relation
H|Ψ`,m,∂ 〉 = E`,∂ |Ψ`,m,∂ 〉 where ` labels the irrep subspace, and is often referred to as sector, or
charge, or (primary) quantum number. The index m distinguishes states within the same irrep
subspace (secondary quantum number), and the energy level E`,∂ is independent of m as a
consequence of Schur’s lemma. Finally, ∂ contains all the residual information not classified
by the symmetry (symmetry degeneracy label). In this sense, the symmetry group acts as
[`]
U(g)|Ψ`,m,∂ 〉 = m0 |Ψ`,m0 ,∂ 〉Wm,m0 (g) where W [`] (g) is the `-sector irrep matrix associated
P

to the group  element g. Symmetries in quantum mechanics can be either Abelian (when
[U(g), U g 0 ] = 0 ∀ g, g 0 ) or non-Abelian (otherwise). From a computational perspective, the
former are typically easier to handle as the related Clebsch–Gordan coefficients (fusion rules of
the group representation theory) become simple Kronecker deltas [162]. In tensor network
contexts, handling fusion rules is fundamental (see Sec. 3.5), thus addressing non-Abelian
symmetries is somehow cumbersome (although definitely possible [163–166]), requiring the
user to keep track of Clebsch–Gordan coefficients for arbitrary groups. For this reason, in this
anthology we focus on addressing Abelian symmetry groups: Even when simulating a non-
Abelian symmetric model one can always choose to address an Abelian subgroup [167], gaining
partial speed-up (in this case, unpredicted degeneracies in the energy levels are expected to
occur). Some examples of Abelian reduction strategies are mentioned later in this section.
Abelian symmetries have one-dimensional unitary (complex) irreps. Such irreps take the
form of phase factors W [`] (g) = e iϕ` (g) , where the phase ϕ depends on the quantum number
` and group element g, while no secondary quantum numbers m arise. Then, the symmetric
Hamiltonian is simply block-diagonal in the sectors ` and has no additional constraints.
In the following we present the Abelian symmetry groups commonly encountered in quan-
tum mechanical problems, and we list irreps, group operations and fusion rules (see Eq. (32))
for each of them. Additionally, for later use, we introduce the irrep inversion ` → `† as the
mapping to the inverse (or adjoint) irrep, defined by ϕ`† (g) = −ϕ` (g) ∀g∈G . It will prove
†
useful that inversion is distributive under fusion, ` ⊕ `0 = `† ⊕`0† . We will also generally write
for the identical irrep `ident = 0. This irrep exists for every group, it fulfills ϕ0 (g) = 0 ∀g∈G
and is self-adjoint.

(i) Zn discrete planar rotation group, also known as group of the oriented regular polygon
with n edges. This is the cyclic Abelian group with n elements, and satisfies g n = 0 (the
identity group element) for any element g ∈ Zn . The parity group Z2 is a particularly
relevant case.
Group elements: g ∈ {0 . . . n − 1} 
Group operation: g ◦ g 0 = g + g 0 mod n
Irrep labels: ` ∈ {0 . . . n − 1}
Irrep phases: ϕ` (g) = 2πg`/n
Inverse irrep: `† = (−`) mod n := (n − `) mod n
Fusion rule: ` ⊕ `0 = (` + `0 ) mod n
Examples: Ising model (Z2 ); q-states chiral Potts model (Zq ) [168].

(ii) U(1) continuous planar rotation group, or group of the oriented circle. This is the
symmetry group related to the conservation laws of integer quantities, such as total

23
SciPost Phys. Lect. Notes 8 (2019)

particle number (` is this number), an extremely common scenario in many-body physics.


Group element parameter: g ∈ [0, 2π)
Group operation: g ◦ g 0 = (g + g 0 ) mod 2π
Irrep labels: ` ∈ Z
Irrep phases: ϕ` (g) = g`
Inverse irrep: `† = −`
Fusion rule: ` ⊕ `0 = ` + `0
Examples: XXZ model with field along z-axis (symmetry under any z-axis rotation, applied
to every site); spinless Bose–Hubbard model (total boson number conservation).

(iii) Z infinite cyclic group, or the group of integers under addition. This group is connected
to the conservation law of a quantity within a real interval, usually [0, 2π) (such as
the Bloch wave-vector conservation under translation of a lattice constant in a periodic
potential problem).
Group element parameter: g ∈ Z
Group operation: g ◦ g 0 = g + g 0
Irrep labels: ` ∈ [0, 2π)
Irrep phases: ϕ` (g) = g`
Inverse irrep: `† = (2π − `) mod 2π
Fusion rule: ` ⊕ `0 = (` + `0 ) mod 2π
Examples: Discrete translations on an infinite 1D lattice.

(iv) Multiple independent symmetry groups. For a model exhibiting multiple individual
Abelian symmetries, which are independent (i.e. mutually commuting and having only the
identity element of the composite symmetry in common), the composite symmetry group
is given by the direct product of the individual symmetry groups, and is Abelian. Typical
composite groups are combinations of the aforementioned basic groups (Zn , U(1), Z).
Here we provide a symmetry combination example built from two individual symmetries,
generalization to more symmetries being straightforward. In this scenario, one can
select the irrep label `k for each elementary subgroup k (or subsymmetry), making the
composite irrep label actually an ordered tuple of scalar labels, which fuse independently.
Group elements parameter:
[1] [2]
 g1 g2 = g2 g1 , with g1 ∈ G1 and g2 ∈ G2 .
Irrep labels: ` = ` , ` , every element in this ordered tuple is an allowed scalar irrep
label of the corresponding elementary symmetry group, with `ident = (0, 0).
Irrep phases: ϕ` (g) = ϕ`[1] (g)+ ϕ`[2] (g)
Inverse irrep: `† = `[1]† , `[2]†
Fusion rule: ` ⊕ `0 = `[1] ⊕1 `[1]0 , `[2] ⊕2 `[2]0 , where ⊕k is the fusion rule of subgroup k.


Examples: XYZ model for an even number of sites: Z2 × Z2 = D2 (π-rotation along any
of the three Cartesian axes, applied at every site). D2 is the simplest noncyclic Abelian
group, also known as the group of the rectangle, or Klein four-group.

Regarding the reduction of non-Abelian symmetry groups into their largest Abelian compo-
nent, typical examples follow. The full regular polygon group Dn (present in the standard Potts
model) is often reduced to Zn : the rotations are kept, the reflections are discarded. The unitary
group of degree N , U(N ), is often reduced to U(1)×N by considering only the diagonal unitary
matrices, which are defined by N independent angles. Similarly, SU(N ) is usually reduced to
U(1)×(N −1) (N angles with 1 constraint). Abelian reduction can be often performed naively,
even for more complex symmetry groups: For instance, the symmetry group of the spin-1/2
Fermi–Hubbard model, which is SO(4) (∼ = SU(2) × SU(2)/Z2 ) [169], is often treated just as a
U(1) × U(1) symmetry (particle conservation for each spin species).

24
SciPost Phys. Lect. Notes 8 (2019)

3.2 Encodable Symmetries in Tensor Networks


Not every symmetry (regardless of its group) can be efficiently encoded in the tensor network
formalism: Depending on how a symmetry transformation acts on the QMB system it may
be suitable or not for TN embedding. Symmetries that are known to be particularly apt for
implementation in tensor networks [135,165,170,171] are those which transform independently
each degree of freedom without making them interact, namely those satisfying U(g) = j Vj (g),
N

where Vj (g) is the local representation of group element g at site j. Vj (g) acts only on the
degrees of freedom at site j, and it may depend explicitly on j. We refer to this symmetry class
as “pointwise” symmetries, because they act separately on every point (site) of the real space.
There are two main sub-types of this symmetry class which have a major impact on physical
modeling:

(i) Global pointwise symmetries: These symmetries have support on the full system, i.e.
the Vj (g) are non-trivial (different from the identity) for every site j. Usually these
symmetries also have a homogeneous pointwise representation, that is, Vj (g) does not
explicitly depend on j. Typical examples are conservation of the total particle number or
total spin magnetization.

(ii) Lattice gauge symmetries [104, 172–174]: These pointwise symmetries have restricted
supports, which usually extend just to a pair or plaquette of neighboring sites (Vj (g) is
the identity elsewhere). In turn, lattice gauge models often exhibit an extensive number
of such symmetries (e.g. one for every pair of adjacent sites), which makes them again a
powerful computational tool.

Most condensed matter systems also include symmetric content which is not of the pointwise
form. For instance, this is the case for lattice geometry symmetries, such as translational
invariance, lattice rotation/reflection and chiral symmetry. These symmetries play often an
important role, and characterize some non-local order properties of various lattice systems
(e.g. topological order) [175, 176]. Although there are indeed proposals on how to address
explicitly these other symmetry classes in specific tensor network architectures [93, 177, 178],
they cannot be properly treated on a general ground, and will not be discussed here. Symmetric
tensor network structures have also been proposed as analytical tools in the characterization of
topological phases of matter; we refer the reader to [179] and references therein.
Having reviewed which classes of Hamiltonian symmetries we can efficiently use, we
now provide in-detail documentation on how to exploit pointwise symmetries to achieve a
computational advantage on TNs. Our presentation primarily draws upon Ref. [135], which
specializes the underlying construction introduced in Ref. [170] to the case of an Abelian group.
The construction relies on precise targeting of quantum many-body states having a well-defined
quantum number: In this sense we will talk about invariant and covariant TN states, as we
will see soon in Sec. 3.4. Since these TNs are built of symmetry-invariant tensor objects and
symmetric link objects, we first introduce those in the next section.

3.3 Constructing Symmetric Tensors


In the construction of symmetric tensors, links (both physical and auxiliary) become associated
with group representations and tensors become invariant objects with respect to the linked
representations. This allows us to upgrade the data structures of links and tensors and their
manipulation procedures to accommodate the symmetry content. Ultimately, symmetric links
and tensors are the building blocks to construct symmetry-invariant and symmetry-covariant
many-body TN states, as we will see in Sec. 3.4.

25
SciPost Phys. Lect. Notes 8 (2019)

Wn† (g)
... ...

= W4 (g)
T W1† (g) T

W2† (g) W3 (g)

Figure 12: Graphical representation of an Abelian symmetric tensor (left) with n


directed links. Incoming links are drawn as arrows directed towards the tensor. By
definition, the tensor is invariant under application of the representations Wrout (g) of
G associated with all outgoing links rout (e.g. 3 and 4) and inverse representations
Wr† (g) associated with the incoming links rin (e.g. 1 and 2).
in

3.3.1 Abelian Symmetric Links


A tensor link is upgraded to be symmetric by associating a quantum number ` to each link index
i ∈ I. The number of different quantum numbers ¯` can be smaller than the link dimension
d. If we then distinguish indices having the same quantum number by a degeneracy count
∂ ∈ {1, . . . , ∂¯` }, we identify
i ≡ (`, ∂ ) , (27)
and call the tuple (`, ∂ ) the (Abelian) symmetric link index. It is useful to organize the symmetric
link indices with the same quantum number ` in a ∂¯` -dimensional degeneracy space or sector
D` such that the link index set I decomposes into their disjoint union
]
I ' D` . (28)
`

From a physical perspective, we associate a specific unitary representation of the group G


with the link. Namely, each tensor-link index i identifies a specific irrep of group G , labeled by
the quantum number (or charge) ` and acting as a phase multiplication e iϕ` (g) . In this sense,
one can also interpret a symmetric link as a link equipped with a diagonal, unitary matrices
group W (g) ∈ Cd×d (or representation of G ) in which the diagonal elements are irreps e iϕ` (g)
of G , each appearing ∂¯` times in accordance with the index decomposition of Eq. (28):

[W (g)] j, j 0 ≡ [W (g)](`,∂ ),(`0 ,∂ 0 ) = e iϕ` (g) δ`,`0 δ∂ ,∂ 0 ∀g ∈ G , j, j 0 ∈ I. (29)

Thanks to the group representation structure embedded in the links, tensors can share informa-
tion related to the symmetry.

3.3.2 Abelian Symmetric Tensors


A tensor with all n links symmetric is called symmetric (invariant) tensor if the tensor is left
unchanged by the simultaneous application of the n representations Wr (g) of G associated
with each tensor link r:
n X
!
Y  † 
Ti1 ...in = Wr r (g) i j T j1 ... jn ∀g ∈ G , i r , jr ∈ I r . (30)
r r
r=1 jr

26
SciPost Phys. Lect. Notes 8 (2019)

With the superscripts † r ∈ {1, †} we specify on which links the acting representations are meant
to be inverted. That is, we distinguish between incoming links, where the representations
are inverted, and outgoing links, where they are not inverted. This formalism allows for a
convenient graphical representation of links as arrows directed away from the tensor or into the
tensor (see Fig. 12). In a network of symmetric tensors, every internal link must be going out of
one tensor and into another: it is then guaranteed that the group acts as a gauge transformation
on internal links of a tensor network, W (g)W † (g) = 1.

The product nr=1 Wr r (g) in Eq. (30) represents a symmetry of the tensor and is a proto-
N

typical global pointwise symmetry, because it acts on all the links of the tensor (global), but as
a tensor product of local operations (pointwise).

3.3.3 Inner Structure of Abelian Symmetric Tensors


With the link indices i r replaced by the associated quantum numbers ` r and degeneracy indices
∂ r , the invariance Eq. (30) reads
n
` ...` ! ` ...` † r
Y
e iϕ`r (g)

T∂ 1...∂ n = T∂ 1...∂ n ∀g ∈ G , (31)
1 n 1 n
r=1
∂ ...∂
and we conclude that a tensor T is symmetric if it has nonzero elements T` 1...` n only where
1 n
Qn  iϕ (g) †r
r=1 e
` r = 1. For our convenience, we introduce the Abelian irrep- or quantum number
fusion rule
`1 ⊕ `2 = `0 :⇔ ϕ`1 (g) + ϕ`2 (g) = ϕ`0 (g) (mod 2π) ∀g ∈ G , (32)
and require that the quantum numbers from all the links fuse to the identical irrep `ident = 0:
† † !
`11 ⊕ `22 ⊕ · · · ⊕ `†nn = `ident (33a)
⇔ ±ϕ`1 (g) ± ϕ`2 (g) ± · · · ± ϕ`n (g) = 0 (mod 2π) ∀g ∈ G , (33b)
where the sign of the individual phases is to be taken opposite for incoming and outgoing links
(due to the inversion ϕ`† = −ϕ` ).
On this basis we define the structural tensor
† † †
¨
1 if `11 ⊕ `22 ⊕ · · · ⊕ `nn = `ident
δ̃`1 ...`n = (34)
0 otherwise.
We refer to any set of quantum numbers `1 , . . . , `n that fuse to the identical one as a matching
set of quantum numbers, or match for short. In addition to the structural  tensor, according
to Eq. (31), a symmetric tensor is identified by degeneracy (sub-)tensors R`1 ...`n ∂ ...∂ . The
1 n
latter are defined over the degeneracy subspaces D`1 ⊗ D`2 ⊗ · · · ⊗ D`n of all possible matches
`1 , . . . , `n :
` ...`
T∂ 1...∂ n = R`1 ...`n ∂ ...∂ × δ̃`1 ...`n .
 
(35)
1 n 1 n

The maximal number of (nonzero) elements of a symmetric tensor T hence reduces to the
number of elements inside the degeneracy tensors
X X Y
#T = #{R`1 ...`n } × δ̃`1 ...`n = δ̃`1 ...`n ∂¯`r (36)
`1 ...`n `1 ...`n r

which is often much smaller than the number of elements of its non-symmetric counterpart
(see later in this section for a more detailed discussion).
In Fig. 13 we give a graphical representation of a symmetry-invariant tensor. Attaching a
dummy selector link (holding a single, non-degenerate irrep `select ) to a symmetry-invariant
tensor is a practical and yet rigorous way to build a covariant tensor (Fig. 14).

27
SciPost Phys. Lect. Notes 8 (2019)

∂n
(`n , ∂n ) `n
...
... (`3 , ∂3 ) `3
∂3

T =
R`1 ...`n
(`1 , ∂1 ) (`2 , ∂2 )
`2
∂1 δ̃`1 ...`n ∂2
`1

Figure 13: Inner structure of an Abelian symmetric tensor: The underly-


ing cyan shape on the right represents the structural tensor δ̃. Here it reads
δ̃`1 ...`n = δ`† ⊕` ⊕`† ⊕···⊕` ,` with inversion (†) of the quantum numbers on incoming
1 2 3 n ident
links (such as links 1 and 3). Only sets `1 . . . `n which are matches contribute a
degeneracy tensor R`1 ...`n , depicted as non-symmetric tensor (gray shape) defined
over degeneracy indices ∂1 . . . ∂n (black lines).

`select `select

T =
R`1 ...`3 ,`select
(`3 , ∂3 )
`3
(`1 , ∂1 ) δ̃`1 ...`n
(`2 , ∂2 ) ∂3

`1
∂1
∂2 `2

Figure 14: A covariant symmetric tensor is a symmetric (invariant) tensor with one
link being a quantum-number selector link. This special link holds only a single,
non-degenerate quantum number `select . With this artifice, the structural tensor takes
the usual form δ̃`1 `2 `3 `select = δ` ⊕` ⊕` ⊕`† ,` .
1 2 3 select ident

3.3.4 Symmetric Links and Tensor Data Handling


We have shown in Eq. (35) that symmetric tensors are characterized by an inner block structure.
This block structure emerged from the links being organized in quantum numbers and corre-
sponding degeneracy spaces of Eq. (28), together with an assumed tensor invariance under
symmetry transformations associated with the quantum numbers (Eq. (30)). In this section, we
define all the necessary symmetric tensor operations to run symmetric TN algorithms on a com-
puter. But first of all, we define the following three numerical objects, suited for object-oriented
programming.

Abelian Symmetry Group Object. This object is a composition of a set of allowed quantum
numbers ` : e iϕ` (g) is irrep of G , the quantum numbers fusion rule ⊕ and the quantum


number inversion operation ` 7→ `† .

28
SciPost Phys. Lect. Notes 8 (2019)

L R`1 `2 `3
`†1 , `2 , `3 ∂¯`1 × ∂¯`2 × ∂¯`3 R`1 `2 `3
  
∂1 ∂2 ∂3 ∂1 ∂2 ∂3
p
0† , 0, 0

2×1×1 1 1 1 1/ 2
2 1 1 0
p p
1† , 1, 0

1×1×1 1 1 1 2/ 3
i1 `1 ∂¯`1 ∂1 p
1† , 0, 1

1×1×1 1 1 1 1/ 3
p
0† , 1, 1

1 1 2×1×1 1 1 1 −1/ 2
0 2
2 2 2 1 1 1
3 1 1 1
(b)

i2 `2 ∂¯`2 ∂2 i1 ≡ (`1 , ∂1 )
1 0 1 1
2 1 1 1

i3 `3 ∂¯`3 ∂3

1 0 1 1 T
2 1 1 1

(a)

i3 ≡ (`3 , ∂3 )
i2 ≡ (`2 , ∂2 )
(c)

Figure 15: Symmetric tensor version of the same tensor Ti1 i2 i3 as introduced in
Fig. 5. To better see the symmetry of this tensor, imagine that the three-link tensor
P dimensions d1 = 3 and d2 = d3 = 2 encodes two-qubit superpositions
with link
Ψi1 = i2 ,i3 Ti1 i2 i3 |i2 i3 〉 of well-defined parity i2 + i3 (mod) 2. It is therefore invariant
under a G = Z2 group of parity symmetry operations: Take Wr (g=0) = 1 and
Wr (g=1) = σz , e.g. [Wr P (g=1)] j, j = e i( j−1)π , on the single-qubit
Plinks r = 2, 3. Then
the tensor transforms as j2 , j3 [W2 (g)]i2 j2 [W3 (g)]i3 j3 Ti1 j2 j3 = j1 [W1 (g)]i1 j1 T j1 i2 i3 ,
hence it is invariant under the global pointwise representation W1† ⊗ W2 ⊗ W3 with
[W1 (g)] j, j = e iϕ`1 (g) . Since the tensor is symmetry invariant, we can recast it as
a symmetric tensor using the following data structures: (a) The three symmetric
tensor links r = 1, 2, 3 as i r = (` r , ∂ r ). (b) The symmetric tensor itself, defined by
its set of matches L (in  this example all four possible matches are present), plus a
degeneracy subtensor R`1 `2 `3 ∂ ∂ ∂ × δ̃`1 `2 `3 for each match in L. (c) The graphical

1 2 3
representation, where we consider the two qubit links outgoing and the first link
incoming in accordance with the inversion on W1† in the defining invariance. Although
the Z2 symmetry is self-adjoint and link directions could be neglected in this example,
this is not true in general.

29
SciPost Phys. Lect. Notes 8 (2019)

For practical purposes, it is often sufficient to represent (a subset of) the group’s quantum
numbers as integers. In the presence of multiple symmetries, acting as a direct-product
group G = G1 ×· · ·× GK of K independent symmetry groups Gk , k = 1 . . . K, it is convenient
to operate with ordered tuples of individual quantum numbers instead: ` = `[1] , . . . , `[K] .


These tuples obey element-wise fusion rules, as described in item (iv) of Sec. 3.1.

Symmetric Link Object. The symmetric link object encompasses a subset of ¯` different quan-
tum numbers ` of the group G , and their corresponding degeneracies ∂¯` ≥ 1.

Symmetric Tensor Object. A symmetric tensor object T consists of:

(i) n symmetric link objects.


(ii) n directions † r ∈ {1, †} (boolean variables), specifying for each link whether it is
incoming (inverted representation) or outgoing (non-inverted representation).
(iii) A set of matching quantum numbers
¦ ©
† †
L = (`1 , . . . , `n ) : `11 ⊕ `22 ⊕ · · · ⊕ `†nn = 0 , (37)

also simply called matches, where each quantum number ` r is taken from its respec-
tive link object.
(iv) For each match (`1 , . . . , `n ) within L, a degeneracy tensor R`1 ...`n . This is defined
as an ordinary, non-symmetric tensor (defined in Sec. 2.1) over the corresponding
degeneracy spaces with index dimensions ∂¯`1 , . . . , ∂¯`n , which can be queried from
the links.

Fig. 15 shows a practical example of this data structure. An important optimization is


that only matches that come with a non-vanishing degeneracy tensor R`1 ...`n 6= 0 have to
be listed in L. We say that these matches are “present” in T .
In order to unify the  handling of multiple symmetries in direct product groups with tuples
` = `[1] , . . . , `[K] of K individual quantum numbers, one can define simple integer
sectors s and use them to index quantum numbers ` ≡ ` (s) within a given link. In this
way, the elements of L can be replaced by matching sectors (s1 , . . . , sn ).

3.4 Symmetric Tensor Networks


For a symmetric Hamiltonian H, different symmetry sectors ` 6= `0 are dynamically decoupled.
Therefore, we can always restrict each TN simulation to span a single sector `select of the
symmetry, without loss of generality. In particular, we will construct, via a symmetric tensor
network ansatz, quantum many-body states which are either invariant or covariant under the
action of the (pointwise) symmetry. This construction will rely, of course, on symmetric tensor
objects.

(i) Invariant states are those which satisfy U(g)|Ψ〉 = |Ψ〉, i.e. they belong by definition to
the sector ` = 0, the identical irrep. Let now the QMB state |Ψ〉 be a tensor network state,
given by Eqs. (1) and (2). Let us also associate a representation Wη (g) of the symmetry
group G to every auxiliary space (or nonphysical link) η. Then it is easy to see that if
every tensor is invariant under the action of the group, i.e. it fulfills Eq. (30), the state |Ψ〉
is symmetry invariant (see Fig. 16) [170]: We name this construct a symmetric (invariant)
tensor network. The reverse is also true: it is possible to decompose an arbitrary invariant
state into the symmetric tensor network formulation, but link dimensions might grow if
the TN is not loop-free [180]. We will investigate this construction in detail in Sec. 4.2.4.

30
SciPost Phys. Lect. Notes 8 (2019)

Rη̄ (g)

(a) Wη (g) (b)


= = =
Wη† (g)

Vj (g)

Figure 16: Symmetric tensor network, composed of symmetric tensors. A global


pointwise operation on the physical links (left) is absorbed by the network as follows:
(a) We apply a gauge transformation with the representations on the virtual links. (b)
Now we use the invariance property of each tensor to absorb all surrounding transfor-
mations at each tensor. We see that the application of the global pointwise symmetry
iϕ` (g)
j Vj (g) is equivalent to the application of the phase shift R η̄ (g) = e select
N
. The
representation at the selector link thus determines the sector `select of the physical
state.

(ii) Covariant states are states |Ψ` 〉 that belong to a single, selected irrep `select of the
symmetry (thus actually extending the concept of invariance to other irreps than the
identical one, `ident = 0). For Abelian symmetries they are the eigenstates of the whole
symmetry group U(G ), and transform as U(g)|Ψ` 〉 = e iϕ` (g) |Ψ` 〉. They are necessary to
address a variational problem for a given (conserved) number of particles `.
Encoding covariant states in TN language is a simple extension of the invariant case [170]:
Specifically, one can set every tensor to be invariant, as in Eq. (31), but then introduce
an additional, special virtual link η̄. This selector link is attached to a single tensor (any
tensor fits, see e.g. Fig. 16), has dimension Dη̄ = 1 and its symmetry representation
Rη̄ (g) is fixed to the `select irrep Rη̄ (g) = W [`select ] (g). Equivalently, one can think of a
covariant symmetric tensor network as a TN made of a single covariant tensor (i.e. an
invariant tensor equipped with a selector link), as in Fig. 14, while all the other tensors
are invariant, as in Fig. 13. This is an effective way to target the ` covariant subspaces
while using the same data structure as used for invariant states: the symmetric tensor.

3.5 Symmetric Tensor Operations


So far, we focused on the upgraded data structures and fundamental group operations. Now,
we will equip symmetric tensors with the same basic functionality that we provided for ordinary
tensors in Sec. 2.2. Again, element-wise operations (such as scalar multiplication, complex
conjugation or addition and subtraction) are not covered in detail. It is easy to cast them into a
symmetric form by carrying out the respective ordinary operation on all degeneracy tensors
independently. This strategy extends naturally to the linear algebra operations, discussed in
Sec. 3.5.2.
In the modular programming of tensor network algorithms, these symmetric operations
constitute an outer layer with respect to the basic tensor operations, as we sketched in Fig. 1.

3.5.1 Link and Tensor Manipulations


As for the non-symmetric tensors, we again first discuss operations that involve a single
symmetric link or tensor.

31
SciPost Phys. Lect. Notes 8 (2019)

(A) Input: Links (`1¦, ∂1 ), (`2 , ∂2 ), (`3 , ∂3 ) with


© directions (†, 1, 1) and a set of ele-
` ` ` 0,0,0 1,0,1
ments T∂ 1∂ 2∂ 3 : T2,1,1 = π, T1,1,1 = iπ .
1 2 3
(`1 , ∂1 )
Output:
L R`1 `2 `3
`†1 , `2 , `3 ∂¯`1 × ∂¯`2 × ∂¯`3 R`1 `2 `3
  
∂1 ∂2 ∂3 ∂1 ∂2 ∂3
T


0 , 0, 0 2×1×1 1 1 1 0
2 1 1 π

1† , 0, 1

1×1×1 1 1 1 iπ (`2 , ∂2 ) (`3 , ∂3 )
Matches L = {(0, 0, 0), (1, 0, 1)} from provided nonzero elements.

(B) Input: Links (`1 , ∂1 ), (`2 , ∂2 ), (`3 , ∂3 ) with directions (†, 1, 1). (`1 , ∂1 )
Output:
L R`1 `2 `3
`†1 , `2 , `3 ∂¯`1 × ∂¯`2 × ∂¯`3 R`1 `2 `3
  
∂1 ∂2 ∂3 ∂1 ∂2 ∂3 0
empty

Matches L = {} all non-present.


(`2 , ∂2 ) (`3 , ∂3 )

(C) Input: Links (`1 , ∂1 ), (`2 , ∂2 ), (`3 , ∂3 ) with directions (†, 1, 1) and a random
number generator.
Output:
L R`1 `2 `3
(`1 , ∂1 )
`†1 , `2 , `3 ∂¯`1 × ∂¯`2 × ∂¯`3 R`1 `2 `3
  
∂1 ∂2 ∂3 ∂1 ∂2 ∂3

0† , 0, 0

2×1×1 1 1 1 1.217 − 0.419i
2 1 1 −0.311 + 1.520i

1† , 1, 0

1×1×1 1 1 1 1.731 + 0.048i

1† , 0, 1

1×1×1 1 1 1 1.096 + 0.654i

0† , 1, 1 (`2 , ∂2 )

2×1×1 1 1 1 0.512 − 0.291i (`3 , ∂3 )
2 1 1 −0.442 − 0.673i

Matches L = {(0, 0, 0), (1, 1, 0), (1, 0, 1), (0, 1, 1)} all identified and present.

(D) Input: Link (`a , ∂a ), will be mirrored to (` b , ∂ b ) ≡ (`a , ∂a ).


Output:
L R`a ` b (`a , ∂a )
`†a , ` b ∂¯`a × ∂¯` b R`a ` b
  
∂1 ∂2 ∂a ∂ b

0† , 0

2×2 1 1 1.0
2 1 0.0 1
1 2 0.0
2 2 1.0

1† , 1

1×1 1 1 1.0 (` b , ∂ b ) ≡ (`a , ∂a )

Matches L = {(0, 0), (1, 1)}, one pair of each quantum number.

Figure 17: Symmetric tensor initialization examples for tensors with the same G = Z2
symmetry and links (`1 , ∂1 ), (`2 , ∂2 ), (`3 , ∂3 ) as in Fig. 15. (A) from user provided
elements, (B) filled with zeros, (C) from random elements, e.g. a Gaussian distribution,
(D) an “identity matrix” on a given link (`a , ∂a ) := (`1 , ∂1 ).

32
SciPost Phys. Lect. Notes 8 (2019)

Symmetric Tensor Initialization. Four notable ways to create a symmetric tensor object are
(see Fig. 17):

(A) Initialize from user-provided elements. Create the respective degeneracy tensors containing
the provided elements and check the fusion rule of the corresponding matches.

(B) Initialize with zeros. No match will be present, hence no memory allocated for degeneracy
tensors.

(C) Fill all degeneracy tensors with random elements. First, we search for all possible matches
given the directed links. We make all of them present, and we initialize the respective
degeneracy tensors with elements provided from a random number generator.

(D) Set up an “identity matrix”. this requires a duplicate symmetric link. We then define
a two-link tensor connected to the two links, with opposite directions. Then from all
quantum numbers in the links we form matches  (`, `), and create the corresponding
degeneracy tensors, each equal to the identity: R`,` ∂ ∂ = δ∂1 ,∂2 . Note that it might be
1 2
advantageous to use sparse diagonal matrices for degeneracy tensors (see Sec. 2.2.2).

Symmetric Link Inversion. This operation inverts a symmetric link: it maps all quantum num-
bers ` from the link into their inverses ` → `† . Degeneracies remain unchanged, ∂¯`(new)
† = ∂¯`(old) .

Inversion of a Link in a Symmetric Tensor. This operation inverts a link r connected to a


tensor while preserving the invariance of the tensor. It acts as follows: (i) The corresponding
symmetric link r is inverted (see above). (ii) The direction of link r is flipped. (iii) The set of
matches is preserved, however every match is manipulated so that ` r → `†r at link r. Notice
that this preserves the fusion rule. (iv) The degeneracy tensors are unaltered. Multiple links
connected to a tensor can be inverted simultaneously or sequentially.

Hermitian Conjugate of a Symmetric Tensor. The Hermitian conjugation is an operation


well distinct from link inversions, because it flips link directions without inverting the link
representations. This operation becomes fundamental in e.g. calculating expectation values
from symmetric tensor networks. It acts as follows: (i) The links are left unchanged. (ii) All the
link directions are flipped. (iii) The set of matches is unchanged (thanks to ` = `† for ` = 0).
(iv) The degeneracy tensors are complex conjugated.

Symmetric Link Index Manipulations. Reducing, enlarging, and rearranging the set of
indices of a symmetric link all belong to this class of transformations. For every irrep ` present
in link r, its degeneracy indices ∂` ∈ {1, . . . , ∂¯` } can be compressed, inflated, or rearranged as
in Sec. 2.2.1. Additionally, quantum numbers can be entirely removed from a link, such as
when ∂¯` is compressed to zero (in which case ` should be explicitly removed from the internal
set), or added anew to the link, in which case a nonzero degeneracy ∂¯`new should be assigned
to it.
Two special cases of symmetric link index manipulations should be mentioned here, because
they play an important role in algorithms:

(i) The intersection (`, ∂ ) ∩ `0 , ∂ 0 is the symmetric link obtained from the set intersection


of the quantum numbers carried in the links ` and `0 with the minimal degeneracy
dimensions min ∂¯` , ∂¯` .
0


(ii) Padding two links (`, ∂ )+ `0 , ∂ 0 is defined as the disjoint union of their symmetric indices.


The padded link has quantum numbers from the union of ` and `0 with corresponding

33
SciPost Phys. Lect. Notes 8 (2019)

degeneracy dimensions ∂¯` + ∂¯`0 added up (taking e.g. ∂¯` = 0 if ` is not a quantum number
in the respective link).

Index Manipulations in a Symmetric Tensor. Index manipulation within a link of a sym-


metric tensor impacts the degeneracy tensors separately. On each degeneracy tensor, it acts
as ordinary index inflation, reordering or restriction (as described in Sec. 2.2.1). Two special
situations may arise: (i) When quantum numbers get entirely removed from a link (i.e. de-
generacy drops to zero), corresponding matches and degeneracy tensors must be removed.
(ii) When new quantum numbers appear on a link, new matches may become possible and
can be assigned nonzero degeneracy tensors. For the sake of efficiency, creating new matches
should be avoided unless the related degeneracy tensors are to be filled with nonzero elements.

Symmetric Tensor Element Access. An individual symmetric tensor element can be accessed
from its quantum numbers and degeneracy indices. Similarly, an entire degeneracy tensor can
be addressed by providing the quantum numbers (match) alone. A more general interface to
address tensor elements is provided by subtensors (see Sec. 2.2.1). Also in the symmetric case,
a subtensor S is obtained from T by reducing the set of degeneracy indices in all links (see
Sec. 3.5.1). As a consequence, S is a subtensor in T if all its degeneracy tensors are ordinary
subtensors in the corresponding degeneracy tensors of T .

Subtensor Readout. This operation reads out a subset of elements at specified indices
from the tensor T and stores them in a new tensor S , according to a link-index reduction
given for every link r. Matches in S are the subset of all matches from T whose quantum
numbers ` r have not been deleted from the respective link. Since each degeneracy tensor of S
is a subtensor in the respective degeneracy tensor of T , this operation performs a block-wise
ordinary subtensor extraction.

Subtensor Assignment. This operation overwrites elements at specified indices in a


tensor T with the content of a subtensor S . Again, the respective element indices are specified
by a link-index reduction for every link r, such that we can perform block-wise ordinary
subtensor assignment on degeneracy tensors of S and T . It is however important to keep in
mind that degeneracy tensors associated with non-present matches implicitly contain zeros.
In order to perform the subtensor assignment, we compare the sets of matches LT and LS of
both tensors T and S , respectively. In detail, we distinguish three cases (see Fig. 18 for an
example):

(i) Matches in LT ∩ LS , present in T and S : We perform ordinary subtensor assignment


between the respective degeneracy tensors of similar matches.

(ii) Matches in LT \ LS , present in T but not S : We set corresponding degeneracy tensor


elements in T to zero, equivalent to ordinary subtensor assignment where the degeneracy
tensor of S contains only zeros. In the special case where the respective degeneracy
tensors of T and S are of same dimensions, we can instead remove the former from LT
as it becomes entirely zero.

(iii) Matches in LS \ LT , present in S but not T : We add these matches to the set LT , initialize
the corresponding degeneracy tensors to zero, and finally inscribe the subtensor from S
via ordinary subtensor assignment.

34
SciPost Phys. Lect. Notes 8 (2019)

Input:
Tensor T : Tensor links:

L R`1 `2 `3 `1 ∂¯`1 `2 ∂¯`2 `3 ∂¯`3


`†1 , `2 , `3 ∂¯`1 × ∂¯`2 × ∂¯`3 R`1 `2 `3
  
∂1 ∂2 ∂3 ∂1 ∂2 ∂3 0 2 0 1 0 1
1† , 1, 0

1×1×1 1 1 1 1.0 1 1 1 1 1 1

1† , 0, 1

1×1×1 1 1 1 2.0

0† , 1, 1

2×1×1 1 1 1 3.0
2 1 1 4.0

Matches LT = {(1, 1, 0), (1, 0, 1), (0, 1, 1)} are present in T .


Subtensor S : Subtensor links:

L R`1 `2 `3 `1 ∂¯`1 `2 ∂¯`2 `3 ∂¯`3


`†1 , `2 , `3 ∂¯`1 × ∂¯`2 × ∂¯`3 R`1 `2 `3
  
∂1 ∂2 ∂3 ∂1 ∂2 ∂3 0 1 0 1 0 1
0† , 0, 0

1×1×1 1 1 1 5.0 1 1 1 1 1 1

1† , 1, 0

1×1×1 1 1 1 6.0

Matches LS = {(0, 0, 0), (1, 1, 0)} are present in S .

Output:
Tensor T with subtensor S inscribed:

L R`1 `2 `3 (`1 , ∂1 ) (`01 , ∂10 )


`†1 , `2 , `3 ∂¯`1 × ∂¯`2 × ∂¯`3 R`1 `2 `3
  
∂1 ∂2 ∂3 ∂1 ∂2 ∂3

0† , 0, 0

2×1×1 1 1 1 5.0
2 1 1 0.0 T S
1† , 1, 0

1×1×1 1 1 1 6.0

0† , 1, 1

2×1×1 1 1 1 0.0
0 0
2 1 1 4.0 (`2 , ∂2 ) (`3 , ∂3 ) (`2 , ∂2 ) (`03 , ∂30 )

Matches LT ∩ LS = {(1, 1, 0)} present in both T and S : Degeneracy tensors have been partially or entirely overwritten with
elements from S .
Matches LT \ LS = {(1, 0, 1), (0, 1, 1)} present only in T : Degeneracy tensors have been partially filled with zeros or entirely
removed.
Matches LS \ LT = {(0, 0, 0)} present only in S : Degeneracy tensors have been initialized with zeros and partially or entirely
overwritten with elements from S .

Figure 18: Example of a symmetric subtensor assignment within a three-link tensor


T with the same G = Z2 symmetry and links as in Fig. 15. Top: T and the subtensor
S come with similar symmetry and number of links, but the second degeneracy
index ∂1 = 2 has been removed from the degeneracy subspace D`1 =0 of the invariant
quantum number in the first subtensor link. This is an exemplary link-index reduction,
mapping all the lower degeneracy indices (here ∂ ≡ 1) from S onto the same indices
(∂ ≡ 1) in T . Bottom: Output tensor T after assignment, with all elements from S
inscribed into respective subtensors of T .

35
SciPost Phys. Lect. Notes 8 (2019)

Permutation of Symmetric  Links at a Tensor. The reordering of symmetric tensor links from
(` r , ∂ r ) into `σ(r) , ∂σ(r) , given a permutation σ, is equal to a permuting all present matching
quantum numbers and their associated degeneracy tensors according to σ, and finally keeping
track of the symmetric tensor’s link directions †σ(r) . The computational cost is dictated by the
ordinary permutations on the degeneracy tensors, which is at most linear in the number of
elements.

Fusion and Splitting of Symmetric Links. The fuse- and split-operations in Sec. 2.2.1 can be
conveniently adapted to symmetric tensors. However, taking into account the special structure
of symmetric links, we cannot rely on the simple enumeration scheme of Eq. (8) anymore.
Instead, we must resort to a more general permutation which is still computationally cheap but
also respects the extra symmetry structure of the links.
We organize this paragraph as follows: First we define the symmetric link fusion as a
bijection between multiple directed symmetric links and the combined index of the fused link.
This mapping is then summarized in the form of a (symmetric) fuse-node. Finally, we describe
how to obtain the corresponding fused or split tensor in full symmetric representation.

Link Fusion. When fusing symmetric links, the goal is to replace multiple links of an
n-link tensor T with given directions {† r } at positions r = k, . . . , m by a single, again symmetric
fused link whose associated symmetric index (`, ∂ ) is a bijection from the original link indices

(`, ∂ ) = fuse(`k , . . . , `m ; ∂k , . . . , ∂m ). (38)

We want to choose this mapping such that the group representation W (g) associated with the
fused link and each group element g ∈ G is the matrix direct product (or Kronecker product)
of the representations on the original links
m
O
W (g) := Wr†r (g) ∀g ∈ G , (39)
r=k

i.e. such that in agreement with the defining Eq. (30) the fused tensor T 0 remains symmetric.
This requirement fixes

` = fuse(`k , . . . , `m ) := `kk ⊕ · · · ⊕ `†mm (40)
and we recover the quantum-numbers fusion rule. Note that Eq. (40) holds for an outgoing
fused link. To obtain an incoming fused link, ` has to be inverted.
For the degeneracy indices ∂ , any enumeration over fused link indices with the same
quantum number ` can be adopted. Such degeneracies arise from two sources: the degeneracies
already present in the original links, and “collisions” of similar  quantum numbers ` obtained
from different combinations fuse(`k , . . . , `m ) = fuse `0k , . . . , `0m , as Eq. (40) is not bijective for
itself. We adopt the following strategy: On each combination `k , . . . , `m we use the ordinary
` ...`
fusion D`k × · · · × D`m → D`k m (Eq. (8)), mapping the original degeneracies into a partial
` ...`
degeneracy index ∂` k m := fuse(∂k , . . . , ∂m ). The partial degeneracy indices of colliding
combinations {`k , . . . , `m : fuse(`k , . . . , `m ) = `} are then concatenated (or stacked) into the
U ` ...`
fused degeneracy D` ' D`k m = α Dα` by adding the appropriate degeneracy offset ∆α` for
U

each collision:
∂` = ∆α` + fuse(∂k , . . . , ∂m ). (41)
0
Here α ↔ (`k , . . . , `m ) uniquely enumerates all collisions into a given `, and ∆α` := α0 <α ∂¯`α
P

amounts to the cumulated partial degeneracy obtained from fusing all preceding colliding
degeneracy subspaces of dimensions ∂¯`α := #Dα` = ∂¯`k × · · · × ∂¯`m .

36
SciPost Phys. Lect. Notes 8 (2019)

This choice of concatenated partial degeneracies makes fusing and splitting of consecutive
links at symmetric tensors computationally cheap, because none of the original degeneracy
tensor elements needs to be reordered in memory. It only involves moving contiguous sections
of memory holding complete degeneracy tensors into, or from, the offset positions Eq. (41)
within the fused degeneracy tensors.
Examples are given in Figs. (19) and (20), where we iterate over all possible combinations
`k , . . . , `m in column-major ordering (fast increment in `k ) and obtain α = 0, 1, . . . as collision
counter for each fused quantum number `. Simultaneously, we Qmincrement the cumulated
α+1 α
degeneracy offset with every collision according to ∆` = ∆` + r=k ∂¯`r , such that we finally
obtain the fused link’s total degeneracy ∂¯` , provided that we started from ∆0` = 0.

Fuse-Node. The mapping of the symmetric link fusion in Eq. (38) is well-defined through
Eqs. (40) and (41). Consequently, it is practical to store the relevant information of this
mapping in a separate numerical object, the (symmetric) fuse-node F . The fuse-node is a
special symmetric tensor defined over the links k, . . . , m from T but with inverted directions,
plus the fused link (`, ∂ ) with the same direction as it appears on the fused tensor T 0 :
¨
`k ...`m ,` 1 if (`, ∂ ) = fuse(`k , . . . , `m ; ∂k , . . . , ∂m ),
F∂ ...∂ ,∂ = (42)
k m 0 otherwise.

The fused tensor T 0 is then obtained by simply contracting F into T over the original links
which shall be fused, as shown in Fig. 21. Splitting up the fused link into the original links
restores T and can be achieved by a second contraction of T 0 with the split-node S = F † over
the fused link. S equals the fuse-node but has all link directions inverted. Since splitting a link
is the inversion of fusing, their contraction over the fused link is again the identity F † F = 1.
However, owing to its sparse structure, storing F in the form of a symmetric tensor and
carrying out fuse- and split-operations by actually contracting such a fuse- or split-node is
computationally inefficient. Therefore, we avoid expanding Eq. (42) into tensor elements and
instead store the relation (`, α) ↔ (`k , . . . , `m ) between the fused quantum number ` with
corresponding collision index α, and each possible combination of quantum numbers `k , . . . , `m
from the original links. One possibility to achieve this is to enumerate the latter by a scalar
index similar to Eq. (5), which can also easily be inverted.
Finally, we store the degeneracy offsets ∆α` in the fuse-node, along with the original
symmetric links and the fused link, including all link directions. Such a fuse-node can be
viewed as a fully equipped network node. One can then define further basic operations for
these fuse-nodes — just like we did for tensors. The permutation is a common example which is
frequently combined with link fusion: Permuting a fuse-node merely involves rearranging some
link positions, collision indices and degeneracy offsets, which is much faster than permuting
the fused links in a tensor before fusing or after splitting.
We now outline two algorithms that, given one such fuse-node, create the fused- or split
tensor. For efficiency, simultaneous fusions of arbitrary (not necessarily adjacent) disjoint
groups of links of a tensor can be handled with minor adaptations, allowing the algorithm to
process multiple fuse-nodes simultaneously. The total computational cost of both algorithms is
composed of ordinary non-symmetric fuse- and split-operations on all degeneracy tensors, and
thus remains at most linear in the number of symmetric tensor elements.

Fused Tensor. The complete link fusion at a symmetric tensor T could proceed as follows:

(i) Initialize the fused tensor T 0 with n0 = n − m + k links similar to T . The links from
positions r = k, . . . , m at T are replaced with the fused link at T 0 .

37
SciPost Phys. Lect. Notes 8 (2019)

b `2 `3 ∂2 ∂3 `b ∂b α ∆α` `b ∂¯` b ∂b
b

1 0 0 1 1 0 1 0 0 1
0 2
2 1 0 1 1 1 1 0 0 2
3 0 1 1 1 1 2 1 1 1
1 2
4 1 1 1 1 0 2 1 1 2

(a) (b)

L R`a ` b a ≡ (`a , ∂a )
 ` ` 
`†a , ` b ∂¯`a × ∂¯` b

∂a ∂b Ra b ∂ ∂
a b
p
0† , 0

2×2 1 1 1/ 2
2 1 0
p
1 2 −1/ 2
2 2 1
p p M
1† , 1

1×2 1 1 2/ 3
p
1 2 1/ 3

(c)
p  p 
1/ 2 −1/ 2 0 0
` `
M∂a ∂b = 0 1 0 0 
a b p p p
0 0 2/ 3 1/ 3 b ≡ (` b , ∂ b )

(d) (e)

Figure 19: Symmetric link fusion (` b , ∂ b ) = fuse((`2 , ∂2 ) , (`3 , ∂3 )) of the last two
(` ` )`
links in the tensor T(∂ 1∂ 2)∂ 3 from the example in Fig. 15: In (a), we show the fuse
1 2 3
mapping in detail; b is the linear index in column-major enumeration of `2 , `3 , ∂2 , ∂3 .
The collision index α and the corresponding degeneracy offset ∆α` help us in the
b
algorithm of Sec. 3.5.1 to place degeneracy indices of pairs `2 , `3 that fuse to the
same ` b into successive intervals ∆α` + 1, . . . , ∆α+1
`b
up to ∆2` = ∂¯` b of the degeneracy
b b
space D b . Here, the complete degeneracy ∂¯` is induced by those collisions, but in
b
Fig 20 we will encounter additional degeneracy already present in the links to be
fused. The resulting matrix M shown in (e) has links (`a , ∂a ) = (`1 , ∂1 ) (cf. Fig. 15)
and the fused (` b , ∂ b ) given in (b). The elements are detailed in (c). We have only
two matching quantum numbers and hence two degeneracy tensors, which appear as
blocks in the block-diagonal matrix-representation M shown in (d). Here we present
rows and columns in ascending order (degeneracies being the fast index) according
to (`a , ∂a ) = (0, 1) , (0, 2) , (1, 1) and (` b , ∂ b ) = (0, 1) , (0, 2) , (1, 1) , (1, 2).
Note that in this example, b also corresponds to the non-symmetric link index of
Fig. 5, but here we have reordered the fused symmetric link indices ` b , ∂ b according
to quantum numbers ` b .

(ii) In all tuples of matching quantum numbers present in T , replace `k , . . . , `m by the fused
` and remove possible duplicates (“collisions”). This yields the set of matches L0 of T 0 .

(iii) Set the corresponding fused degeneracy tensors R0`1 ...`k−1 ,`,`m+1 ...`n in T 0 to zero, with
degeneracy dimensions ∂¯` queried from the fused link.

(iv) Go through the set L of matching quantum numbers of T again, but this time fuse the
links (k, . . . , m) at each degeneracy tensor R`1 ...`n and place these as subtensors into the
respective degeneracy tensors of T 0 , which might have a larger degeneracy dimension

38
SciPost Phys. Lect. Notes 8 (2019)

j `†1 `2 `3 ∂1 ∂2 ∂3 `j ∂j α ∆α`
j
 p 
1 0 0 0 1 1 1 0 1 0 0 1/ 2
2 2 1 1 2
 0 
 p p 
2/ 3
 
3 1 0 0 1 1 1 1 1 0 0 p
 
 
 1/ 3 
 p 
4 0 1 0 1 1 1 1 2 1 1  −1/ 2 
 
5 2 1 1 3 `j
 1 
V∂ =
 

6 1 1 0 1 1 1 0 3 1 2
j  0 
 
 0 
7 0 0 1 1 1 1 1 4 2 3
 

 0 

8 2 1 1 5
0
 
 
 
9 1 0 1 1 1 1 0 4 2 3  0 
0
10 0 1 1 1 1 1 0 5 3 4
11 2 1 1 6 (d)
12 1 1 1 1 1 1 1 6 3 5

(a)

L R` j V
∂¯` j R` j
  
`j ∂j ∂j
p
`j ∂¯` j ∂j (0) 6 1 1/ 2
2 0
0 6 1...6 p p
3 2/ 3
1 6 1...6 p
4 1/ 3 (` j , ∂ j )
p
(b) 5 −1/ 2
(e)
6 1

(c)

`j
Figure 20: Fusing T from Fig. 15 into a vector V∂ : The three-link fusion
j

` j , ∂ j = fuse((`1 , ∂1 ) , (`2 , ∂2 ) , (`3 , ∂3 )) is detailed in (a). The fused link is sum-
marized in (b) and (c), where the degeneracy ∂¯` j originates from two sources: First,
the original link (`1 , ∂1 ) already comes with degeneracy ∂¯` =0 > 1. Furthermore, the 1
collisions of combinations `1 , `2 , `3 into similar quantum numbers ` j are cumulated in
∆α` (see Sec. 3.5.1). The resulting vector elements vanish in all but the identical sector
j
` j = 0, as the vector must be invariant. If we again order the indices ascendingly in
` j , ∂ j , with fast increment in the degeneracy, we obtain the vector representation in
(d). Graphically, the vector can be represented as shown in (e). Note that we would
have obtained the same result by fusing links (`a , ∂a ) and (` b , ∂ b ) at the matrix M
from Fig. 19, as link fusions can be nested.

39
SciPost Phys. Lect. Notes 8 (2019)

fuse

S
T T0
split

Figure 21: Fusion and splitting of a symmetric tensor. From the fuse-operation acting
on a tensor T (left-to-right) we obtain the fused tensor T 0 together with a split-node
S = F † which encodes the symmetry fusion rule for the irreps (cyan filled circle) on
the fused links and the index enumeration in the degeneracy space (black filled circle
on top). The inversion of the process, splitting the fused link into the three original
links (right-to-left), is equivalent to a contraction of the split-node over the fused link.

Qm
∂¯` ≥ r=k ∂¯`r due to collisions. The subtensor’s exact position in the degeneracy subspace
D` is determined by the interval of degeneracy offsets ∆α` + 1, . . . , ∆α+1
`
, where α is the
enumeration index of collision `k , . . . , `m into ` as specified with the fuse-node F .
Note that if the original tensor had non-present matches, the fused tensor also does not
necessarily have all matching quantum numbers present in the fused set L0 . This must be
∂ ∂
considered when e.g. fusing a tensor into a block-diagonal matrix M` a ` b in order to perform
a b
matrix operations. For example, in the exponentiation of a square matrix with one incoming and
one outgoing link (such as a time-evolution operation), non-present matches (`a , ` b = `a ) ∈ / L0
`a ` b
must be identified and explicitly provided with identity degeneracy tensors R = 1, resulting
from the exponentiation of the implicit zero blocks.

Split Tensor. Restoring the original links (`k , ∂k ) , . . . , (`m , ∂m ) in place of a fused link
(`, ∂ ) at some splitting-position k in the fused tensor T 0 yields the split tensor T 00 . Step-by-step,
the split tensor can be obtained from the n0 -link tensor T 0 and the respective fuse-node F as
follows:
(i) Initialize the split tensor T 00 with n = n0 + m − k links, copying links and directions from
T 0 , but replacing the link at position k of T 0 with the “original links” (`k , ∂k ) , . . . , (`m , ∂m )
including their directions as stored in the fuse-node.

(ii) In every tuple of matching quantum numbers present in T 0 , expand ` at the splitting-
position k to all possible combinations `k , . . . , `m of quantum numbers in the “original
links” (now at positions k, . . . , m in T 00 ) that fuse to ` (both the links and the respective
combinations are stored as collisions in F ). This yields the set of matches L00 of T 00 .
` ...` ,` ...` ,` ...`
(iii) For each extended match, create the split degeneracy tensor R00 1 k−1 k m m+1 n by
` ...` ,`,` ...`
ordinarily splitting the k-th link of the subtensor R0 1 k−1 m+1 n of T 0 . All degeneracy
dimensions can be queried from the links in the fuse-node. Again, the subtensor’s exact
position in the degeneracy subspace D` is determined by the interval of degeneracy offsets
∆α` + 1, . . . , ∆`α+1 , available with the collision index α from F .
Note that the tensor T 0 , at which we split a link, does not necessarily have to be the result of a
fuse-operation. The required fuse-node could also be obtained from the fusion of similar links
at another tensor, or it might have been created manually. If however we consecutively fuse
and split the same links, with no further operations performed in between, the original tensor
is restored: T 00 = T . However, degeneracy tensors with all elements equal to zero may appear
in such a process. For the sake of efficiency, one can remove those from T 00 in a post-processing
step after splitting.

40
SciPost Phys. Lect. Notes 8 (2019)

Fused Links and Symmetric Tensor Invariance.

(i) The definition of the fused quantum number ` in Eq. (40) can be invoked to give an
alternative definition of a symmetric tensor T : After fusing all links on T into a single link
`

(`, ∂ ), the result is a symmetric vector V ≡ V∂ × δ̃` which has non-vanishing elements
in the identical sector only, i.e. there is only one possible degeneracy tensor V∂0 (see
Fig. 20 for an example).

(ii) Similarly, fusing a bipartition of links into symmetric row- and column-indices (`1 , ∂1 )
and (`”2 , ∂2 ) yields, in— accordance with Schur’s Lemma, a block-diagonal matrix
` `
M ≡ M∂1 ∂2 × δ̃`1 `2 (see Fig. 19 for an example).
1 2

(iii) For numerical purposes, it may be advantageous to avoid fusions and to keep as many
links as possible, since that gives us the freedom not to store certain matching sectors
(`1 , . . . , `n ) ∈ L with vanishing degeneracy tensors R`1 ...`n that would otherwise appear
` ...` ,`,` ...`
explicitly as blocks of zeros within the fused degeneracy tensors R0 1 k−1 m+1 n of the
fused tensor.

(iv) The tensor symmetry invariance is not broken by adding a one-dimensional “dummy” link
(in either direction) which carries only the identical quantum number ` = 0 (cf. Fig. 22).

Downgrade of a Symmetric Tensor to an Ordinary Tensor. The operation of recasting a


symmetric tensor (and tensor network) into an ordinary one can be useful for debugging and
interfacing symmetric to non-symmetric tensor objects. For example, it can be used to measure
a non-symmetric observable O for a symmetric state. The links are first downgraded from
symmetric links to ordinary links, where each downgraded link dimension D = ` ∂¯` is the
P

sum of its previous degeneracy space dimensions ∂¯` , and an index mapping is chosen. The
degeneracy subtensors R are written into the new ordinary tensor at the right positions, and
the remaining tensor elements are filled with zeros.

Link Clean-up. This is a special operation upon a symmetric link, which employs information
from the surrounding network links to wipe out redundant data on the link itself. Its goal is to
remove indices while preserving the TN state manifold, and it finds extensive use in the random
symmetric TN initialization (see Sec. 4.3.1). Namely, assume we want to clean-up the link
connecting the network nodes q and q0 (tensors at these nodes do not have to be defined). We
then intersect the link representation with the fusion of all the other (directed) links connected
to q and again, with the fusion of all the other (directed) links connected to q0 . The original
link is then replaced by the resulting intersection.

3.5.2 Symmetric Tensor Linear Algebra Operations


Symmetric Tensor Contraction. In the contraction of two symmetric n- and m-link input
tenors A and B over any k shared symmetric links (s, ∂s ), the (m + n − 2k)-link output tensor
C is again symmetric. It is however mandatory that the shared links are similar in quantum
numbers and respective degeneracy dimensions ∂¯s , and that each of them is either directed
from A to B or vice-versa. For example, the symmetric version of Eq. (14) reads
X X
a ...a b ...bm a ...a s ...s s ...s b ...bm
C∂ 1 ...∂n−k k+1
∂ ...∂
= A∂ 1 ...∂n−k 1 ∂ k...∂ B∂1 ...∂k k+1
∂ ...∂
, (43)
a1 an−k bk+1 bm a1 an−k s1 sk s1 sk bk+1 bm
s1 ...sk ∂s ...∂s
1 k

where (a, ∂a ) and (b, ∂ b ) are now the symmetric indices of the remaining links not contracted
over, attached to A and B , respectively. The degeneracy tensors Ra1 ...an−k bk+1 ...bm in C are

41
SciPost Phys. Lect. Notes 8 (2019)

(a) i m i m
(b) (c)
k
A B = A B
T
j n j n
i m T

= C T∗

j n
partial trace norm
direct product

Figure 22: Variants of contractions (compare Sec. 2.2.2 and Fig. 8), now cast into a
symmetric form: (a) Direct product. Between any two tensors, a dummy link can be
inserted to indicate a contraction. With symmetric tensors, such intermediate product
links may acquire a deeper meaning as quantum-number “transfer links” if they carry
a non-degenerate quantum number ` 6= 0. A typical application is the correlation
function c † ⊗ c of particle creation and annihilation operators c † and c under a U(1)
particle-conservation symmetry, which might, for example, occur in a Hubbard model.
In this context, the transfer link carries a single quantum number ` = +1 or ` = −1
(depending on the direction of the link) which accounts for the particle hopping
associated with the operation c † ⊗ c. Contracting the (covariant) tensors representing
the two operators over the intermediate link yields again a symmetric invariant tensor.
(b) Partial trace, which can be carried out as a contraction over pairs of similar,
oppositely directed symmetric links at the same tensor. (c) Hilbert–Schmidt norm,
performed as contraction with the hermitian conjugate T † which extends the complex
conjugation by the simultaneous inversion of all links.

obtained by adding up (outer sum in Eq. (43)) the standard contractions (inner sum) of
degeneracy tensors from the input tensors, directly exposing the block-wise structure of such a
procedure.
It seems natural to transform all involved tensors into (block-diagonal) symmetric matrices
and carry out the contraction as a (block-wise) matrix-matrix multiplication, similar to ordinary
contraction (Sec. 2.2.2). Indeed, symmetric fuse-, split- and permute-operations allow us to
implement this sequence of operations in a simple and elegant manner. However, this may
not always be the most efficient strategy, as it does not fully exploit the possible absence of
matches within the fused tensor.
In the following, we present a specialized algorithm, and afterwards discuss the possible
advantages over the “fuse-multiply-split” strategy: Namely, one first identifies pairs of subsets
s ...s s ...s
LA1 k ⊆ LA , LB1 k ⊆ LB holding matches from A and B that share similar combinations of
quantum numbers s1 , . . . , sk on the contraction links. Next, the degeneracy tensors associated
a ...a s ...s s ...s b ...b
with these subsets, given by RA1 n−k 1 k and RB1 k k+1 m , are contracted for all available
combinations a1 , . . . , an−k and bk+1 , . . . , bm of the quantum numbers not contracted over. The
resulting contracted tensors are then stored together with an intermediate list L0C of possibly
non-unique tuples (a1 , . . . , an−k , bk+1 , . . . , bm ). As we required all contracted links to have
pairwise opposite directions on tensors A and B , these tuples are already matching quantum
numbers obeying
† †n−k †k+1 †m
a11 ⊕ · · · ⊕ an−k ⊕ bk+1 ⊕ · · · ⊕ bm = 0, (44)

42
SciPost Phys. Lect. Notes 8 (2019)

where the †-superscripts yield the inverse quantum numbers only on the incoming links of
A and B , respectively. A final reduction step completes the contracted and again symmetric
tensor C : The set of matches LC is obtained by identifying and removing all recurrences of
matches from L0C while adding up the associated degeneracy tensors into the final degeneracy
tensors Ra1 ...an−k bk+1 ...bm .
Even though our example Eq. (43) again resembles a matrix-multiplication (after fusing
row- and column-links), it should be noted that the set of matches L obtained by our specialized
procedure usually contains only a subset of all possible matches (a1 , . . . , an−k , bk+1 , . . . , bm )
that are allowed by Eq. (44). It is therefore not advisable to first permute the full symmetric
input tensors A and B into (block-diagonal) matrices as in Sec. 2.2.2, since the symmetric link
fusion might force us to represent some of the previously non-present degeneracy tensors as
blocks of zeros in memory — thereby giving up computational speed1 and precision.
The computational complexity Otot of the symmetric tensor contraction is governed by
the degeneracy tensor contractions, each of which can be carried out independently and is
bounded linearly in all involved degeneracy dimensions by
k n−k m
‚ Œ
Y Y Y
Odeg = O ∂¯sr ∂¯ar ∂¯br . (45)
r=1 r=1 r=k+1

Consequently, if the degeneracy dimensions of all involved links are peaked around cer-
tain quantum numbers, the total cost is governed by the largest contraction of degeneracy
tensors involved. If, instead, degeneracies come in rather flat distributions over the active
number
€Q of sectors, the number of Š contractionsS between degeneracy tensors may grow to
k Qn−k Qm
S≤ r=1 s̄ r r=1 ā r r=k+1 b̄ r / āmax · b̄max , where āmax and b̄max denote the maximum
number of symmetry sectors among uncontracted links at A and B (compare also Sec. 3.5.3).
The maximal speedup that can be achieved over the full, non-symmetric contraction therefore
scales with āmax · b̄max .
In a similar way, we can also cast all contraction variants discussed in Sec. 2.2.2 into fully
symmetric operations, as depicted in Fig. (22).

Symmetric Tensor Decompositions. Decompositions of symmetric tensors are in principle


similar to decompositions of ordinary non-symmetric tensors (discussed in Sec. 2.2.2). We
require that the output tensors of such symmetric decompositions have to be element-wise
equal to the outcome of the corresponding symmetry-unaware operations. However, we must
also require that these output tensors are all symmetric tensors again, and that they are mutually
connected by directed symmetric links.
The usual QR- and singular-value matrix decompositions indeed fulfill these requirements.
Focusing on the SVD, we proceed by restating the procedure of Sec. 2.2.2 for symmetric tensors.
Specifically, we want to decompose an arbitrary n-link symmetric input tensor T over a certain
link-bipartition (1, . . . , r),(r + 1, . . . , n) into two (semi-)unitary, symmetric output tensors V and
W (holding the singular vectors) and a non-negative diagonal matrix λ (holding the singular
values). Both output tensors shall be connected to λ through an intermediate symmetric link
(k, ∂k ),
` ...` ` ...` k k ` ...`
X
T∂ 1...∂ n = V∂ 1...∂ r ∂ λ∂k W∂ ∂r+1 ...∂n . (46)
1 n 1 r k k k r+1 n
k,∂k

In order to achieve this result, we do the following (see Fig. 23):


1
Even though optimized matrix-matrix multiplications exhibit advantageous performance on those larger blocks,
in practice the performance gain by not having to contract non-present degeneracy tensors can be significantly
larger.

43
SciPost Phys. Lect. Notes 8 (2019)

⇓ fuse

a b

block-wise ⇓ tensor decomposition

symmetric ⇓ intermediate link

⇓ split

λ
V W

Figure 23: Steps in symmetric tensor decomposition (e.g. QR or singular value


decomposition), from top to bottom: First the input tensor T is fused into a matrix
over the indicated link bipartition a,b. The resulting degeneracy tensors form a
block-wise matrix and can be decomposed independently, often at greatly reduced
computational cost. Their factors are separated into two symmetric tensors V and W ,
connected through a new intermediate symmetric link k, and weighted by singular
values λ in the case of an SVD. Finally, the original links are restored with the help of
fuse-nodes that were obtained in the initial step.

(i) Fuse T into a matrix with incoming row-link (a, ∂a ) = fuse(`1 , . . . , ` r ; ∂1 , . . . , ∂ r ) and
outgoing column-link (b, ∂ b ) = fuse(` r+1 , . . . , `n ; ∂ r+1 , . . . , ∂n ). Keep the fuse-nodes. Due
to the symmetry invariance, we obtain a block-diagonal matrix
 
T∂ab∂ = Rab ∂ ∂ × δab , (47)
a b a b

in which the degeneracy tensors Rab play the role of blocks.

(ii) Create the intermediate symmetric link from the intersection (k, ∂k ) = (a, ∂a ) ∩ (b, ∂ b )

44
SciPost Phys. Lect. Notes 8 (2019)

(see Sec. 3.5.1).

(iii) Carry out the ordinary matrix decomposition on all present matrix blocks Rab indepen-
dently. In the case of an SVD, we obtain block-wise matrix factors V ak and W kb holding
the left- and right singular vectors, and vectors (or diagonal matrices) λk holding their
respective singular values:

∂¯k
X
 ab
      
R ∂a ∂ b
= V ak ∂a ∂k
λk ∂k
W kb ∂k ∂ b
, (48)
∂k =1

where k = a = b are the quantum numbers associated with each block in  the interme-
diate link, directed from V to W . The degeneracy dimensions ∂k = min ∂¯a , ∂¯b of the
¯
factorization can be truncated, e.g. by the exclusion of small singular values. To this end
it may be necessary to first compute λk for all blocks k and then globally identify the
singular values to be discarded among them. This aspect is particularly relevant when
the implemented Abelian symmetry is a subgroup of a larger, possibly non-Abelian one
(as we discussed in Sec. 3.1): singular values which are quasi-degenerate should either
be all kept, or all discarded, even if they are distributed across different sectors, because
they might form a multiplet of an underlying non-Abelian symmetry.

(iv) Compose the symmetric output tensors V∂ak∂ , W∂kb∂ and λ∂k from all obtained degeneracy
a k k b k
tensors (or subtensors therein if selected singular values shall be discarded):
 
V∂ak∂ := V ak ∂ ∂ × δak , (49a)
a k a k
 
W∂kb∂ := W kb ∂ ∂ × δkb , (49b)
k b k b
 
λ∂k := λk ∂ . (49c)
k k

If desired, the direction of the intermediate link can be inverted afterwards.

(v) With the help of the fuse-nodes obtained in step (i), split the links a and b into the
original links `1 . . . ` r on V and ` r+1 . . . `n on W .

We note that a QR decomposition follows the exact same procedure, yielding two output
tensors, one of which is unitary with respect to the intermediate link. Since the computational
costs of QR- and singular value decomposition behave highly nonlinearly in their operands’
dimensions, their symmetric versions can gain a lot in efficiency by operating independently on
multiple smaller blocks instead of one large matrix. For example,
 a square matrix defined over
3
two similar links comes at a full decomposition cost of O da in the absence of symmetric inner
structure. On a symmetric tensor one can gain a factor ā2 , quadratic in the number of active
sectors of a flat or only slightly peaked distribution of degeneracy dimensions. Furthermore,
these block-wise operations allow for a natural parallelization of the algorithm implementation.

3.5.3 Computational Cost Estimation


The resources spent in a symmetric tensor operation depend on the number and dimensions
of the degeneracy tensors, and are thus highly problem-dependent. Nevertheless, we can
isolate common scenarios and give cost estimates for the respective operations. As it will turn
out, the symmetric operations presented above often display a highly favorable computational
complexity compared to equivalent ordinary tensor operations. The latter usually constitute
the worst-case scenario for symmetric tensors. Therefore, provided that data structures and

45
SciPost Phys. Lect. Notes 8 (2019)

algorithms with only a negligible overhead are used for the internal quantum-number book-
keeping (as outlined in Sec. 3.3), one will always benefit from exploiting the additional inner
structure.
We stress that all basic tensor operation listed in Sec. 3.5, on the level of degeneracy tensors,
make use of the ordinary procedures presented for non-symmetric tensors. Therefore, we
can easily estimate the cost of symmetric operations. Namely, if the complexity of a certain
operation on degeneracy tensors scales like
€ Š
p p
Odeg,`1 ...`n = O ∂¯` 1 · · · ∂¯` n , (50)
1 n

where p r are the maximal power coefficients for the degeneracy dimensions ∂¯`r and, in total, n
links are involved, then its cost on a symmetric tensor is bounded by
X
Osym = Odeg,`1 ...`n × δ̂`1 ...`n , (51)
`1 ...`n

where δ̂`1 ...`n ∈ {0, 1} is an operation-specific symmetry constraint on the involved quantum
numbers. Depending on the operation under consideration (we saw complexity examples
in Sec. 3.5.2), this constraint can be even stricter than the “structural tensor”-constraint δ̃
formulated in Eq. (34) (which, for example, would be valid for a block-wise operation on a
single tensor). In any case, we have δ̂`1 ...`n ≤ δ̃`1 ...`n .
We can therefore provide an upper bound for the computational cost of a symmetric
operation:
! p1 ! pn !
X € Š X X
p p
Osym ≤ δ̃`1 ...`n × O ∂¯` 1 · · · ∂¯` n ≤ O ∂¯`1 ··· ∂¯`n = Ofull , (52)
1 n
`1 ...`n `1 `n

where Ofull is the complexity of the ordinary, non-symmetric operation. We observe that the
symmetric operations outperform the ordinary ones for two reasons:

(i) Thanks to the constraints δ̂ posed by symmetry P (invariance), the number of block-wise
operations between degeneracy tensors, S = `1 ...`n δ̂`1 ...`n , is rigorously restricted. Due
to the sparsity of the allowed block-wise operations on degeneracy tensors, we can naively
upper-bound their number by S ≤ ¯`1 · · · ¯`n . However, since those operations are only
performed over symmetry-invariant quantum numbers, they at least fulfill the constraint
δ̃. This allows us to further tighten down the number of combinations contributing to S:

S ≤ ¯`1 · · · ¯`n /¯`max , (53)

where ¯`max = max r ¯` r is the largest number of sectors found among all links r = 1, . . . , n.
When specifying an operation, one can possibly formulate even tighter restrictions, as we
encountered e.g. for the symmetric contraction (Sec. 3.5.2). In practice, one often finds
even smaller S due to non-present matches.

(ii) The most costly operations are those that scale nonlinearly (p r > 1) with the dimensions.
Those operations are also the ones which benefit most from operating on smaller blocks.

In order to provide a more concrete estimate of computational costs, we need to account for
the actual degeneracy dimensions ∂¯`r . To this end, we sketch two extremal and one common
degeneracy distributions of symmetric links and study their cost estimates for all basic symmetric
tensor operations. In particular, we demonstrate that we can benefit the more from symmetric
tensors operations, the broader the indices of links are spread over different quantum numbers.

46
SciPost Phys. Lect. Notes 8 (2019)

Flat distribution. We define a flat distribution as a distribution characterized by constant


degeneracy dimensions ∂¯`r = d r /¯` r which only depend on the total dimension d r and
the number of sectors (different quantum numbers) ¯` r of a given link r, but not on the
quantum numbers ` r itself. Consequently, all degeneracy tensor operations come at a
similar cost, totaling to
€ Š
p p
Osym = O ∂¯` 1 · · · ∂¯` n · S (54a)
1 n
p −p
= O d1 1 · · · dnpn · ¯`1 1 · · · ¯`−p

n ·S (54b)
n
= Ofull · Ogain . (54c)

Ofull is again the complexity of the equivalent non-symmetric operation, and we separate
−p −p
all improvement from using symmetric tensors into Ogain = O ¯`1 1 · · · ¯`n n · S . Using
Eq. (53), which was derived for unary operations on a single symmetric tensor, results in
a conservative upper bound of
€ Š
1−p
Ogain ≤ O ¯`1 1 · · · ¯`1−p
n
n /¯
`max , (55)

i.e. the computational cost scales at least reciprocally in the maximal number of sec-
tors ¯`max , but possibly significantly better depending on the operation (p r , S) and the
number of non-present matches in the tensor(s).
Consider for example the memory consumption of a symmetric tensor in a flat degeneracy
distribution. The number of elements in Eq. (36) is linear in the degeneracy dimensions
(meaning p r = 1), hence according to Eq. (55) this number is reduced by a factor of ¯`max
compared to the number of elements in an ordinary tensor.
Among the degeneracy distributions discussed here, the flat distribution is optimal for
symmetric tensor operations, because it allows computations to benefit from the additional
inner structure to the greatest possible extent. Tensors with this distribution usually
contain many small degeneracy tensors, which significantly reduces the cost in operations.
The flat distribution is occasionally realized in symmetries with small finite numbers
of available quantum numbers (e.g. Zn with small n), such as in the Ising model (see
Sec. 5.3).

Highly peaked distribution. In the extremal case of a narrowly peaked distribution, we can
expect one degeneracy tensor to have dimensions much larger than all the others. The
computational cost is then dominated by these maximal degeneracy dimensions, ap-
proaching the full dimensions d r of the peaked links. In this case, the cost of performing
the symmetric operation becomes roughly equal to performing the equivalent operation
on non-symmetric tensors: Osym = Odeg = Ofull . In this worst-case scenario, symmetric
tensors cannot distinguish themselves by any relevant inner structure any more, and
basic operations cannot benefit from symmetry invariance in a noticeable way.

Slightly peaked distribution. The most common scenario in groups with more than a few
accessible quantum numbers is a distribution of degeneracy over a wider, but effectively
limited, range of quantum numbers ¯`eff . If the degeneracy dimensions are more or
less evenly distributed within such a range, we can resort to the results obtained for
the flat distribution, replacing ¯` by the distribution width ¯`eff . Sometimes, degeneracy
concentrates around certain quantum numbers, with a distribution resembling e.g. a
Gaussian. The narrower such a distribution becomes, the more the cost will be dominated
by the largest degeneracy dimensions in the involved links. Then, the effective number
of active quantum numbers ¯`eff on those links becomes very small, resulting in less
advantageous cost estimates. We will encounter such scenarios in Sec. 5.3.

47
SciPost Phys. Lect. Notes 8 (2019)

(a)
(b)

(c) (d)

Figure 24: Loop-free TN geometries: (a) a generic tree tensor network, (b) an MPS,
(c) a binary tree tensor network (bTTN), (d) a molecular TTN, shaped according to
an arbitrary loop-free graph geometry, finding application in physical chemistry.

4 Loop-Free Tensor Networks


In this section, we address general properties of loop-free tensor networks, i.e. TN states
whose network graphs contain no cycles. In Fig. 24 some examples are given, including the
linear MPS [25, 27], the hierarchical binary Tree Tensor Network (bTTN, discussed in detail in
Sec. 5.2) [70, 73, 75], and the flat tensor product state, which finds applications in quantum
chemistry [181].
Tensor networks without cycles can exploit the full power of the TN gauge transformations:
Smart gauging establishes rules on the network which can tremendously reduce the number
of necessary contractions, thereby enhancing computational speed (see Fig. 25). It has been
argued that the computational advantage of working in such a loop-free framework in fact
enables the use of larger link dimensions, and, in turn, may partially compensate for the lower
scaling of entanglement captured with respect to a “loopy” tensor network [182]. The textbook
example is the comparison between a tree tensor network [70, 73–75, 183], which allows for
efficient algorithms, but suffers from the fact that the maximal entanglement between two
subsystems depends on the chosen bipartition of the system (entanglement clustering) [184],
and a MERA [64–67] where the entanglement distribution is “smooth”, but calculations are
expensive.
Loop-free TNs exhibit advantages also in relation to symmetries. In fact, while upgrading
a generic TN (describing a covariant state) to its symmetric version may require an overall
increase in link dimensions [180], for loop-free TNs the upgrade can always be performed in a
link dimension-preserving way (see a proof in Sec. 4.2.4).
In the following, we first introduce some general aspects of the structure of loop-free TNs
(Sec. 4.1). We then discuss different possible gauges in loop-free TNs and their application
(Sec. 4.2). Finally, we show how to randomly initialize symmetric ansatz states for loop-free
TNs (Sec. 4.3.1).

4.1 General Aspects of Loop-Free Tensor Networks


The underlying graph structure of a loop-free TN defines a distance dist(a, b) between any
two nodes a and b of the network, which is equal to the number of links encountered on the

48
SciPost Phys. Lect. Notes 8 (2019)

(a) (b) (c)


|Ψ〉 = with =

(d)
〈Ψ|O[s] Ψ〉
|Ψ〉 (e)

O[s] = = =

〈Ψ|

Figure 25: (a) A binary tree TN as an example of a loop-free TN describing some QMB
state |Ψ〉: Writing it in a certain gauge (b) imposes certain rules (c) on the network.
Knowledge about the gauge then allows us to simplify the numerical calculations,
here the calculation of the expectation value of some local observable 〈Ψ|O[s] |Ψ〉. The
bra TN is drawn as the vertically flipped version of the ket TN, indicating hermitian
conjugation of all its tensors (d): By applying the rules, we see that most contractions
(which would be necessary in a network without a gauge) do not have to be performed
explicitly numerically. We are left with a much simpler network (e) where only the
contractions that cannot be “annihilated” using gauge rules have to be executed.

shortest path between them. In particular, for loop-free TNs such a path is always unique. It is
practical to call all nodes which share the same distance d to some specified center node c a
level, L(d; c). 
We further introduce the maximal link dimension D := maxη Dη among all virtual
links, which usually exceeds the physical link dimensions {ds }. As it will become apparent in
the following, the maximal link dimension plays a fundamental role in any TN ansatz. The
computational complexity of algorithms operating on loop-free TNs is a smooth polynomial of
D. This can be seen, for instance, in the calculation of the scalar product between two QMB
states represented in the same loop-free TN geometry. Following a level-structure towards
some center node, one can successively contract all tensors at the currently outermost level
of the ket-state with their bra-state counterparts into temporary nodes, while absorbing any
temporary  nodes on links in between. The cost of these contractions is then upper-bounded by
O D Z+1 , where Z is the maximal number of links of a tensor in the network. Tighter bounds
can be identified in specific situations, e.g. when certain link dimensions are known to differ
substantially from D (such as in an MPS, where one link at each tensor carries the physical
dimension).
Another immediate consequence of the absence of loops is that every link between two
nodes q and q0 induces a bipartition of the QMB system. The Hilbert spaces of the two partitions
are spanned by the canonical basis states |is 〉 : dist(s, q) < dist s, q0 s and
 

|is 〉 : dist(s, q) > dist s, q0 s , respectively, i.e. by fusing all physical links s ultimately connected
 

to either q or q0 after removing the bipartition link between the two nodes from the network.

49
SciPost Phys. Lect. Notes 8 (2019)

At any virtual link η, we can therefore rewrite the loop-free TN state by means of a Schmidt
decomposition

X
|Ψ〉 = |Ak 〉|Bk 〉λk , (56)
k=1

where |Ak 〉 and |Bk 〉 are states described in the (contracted) TNs of the two partitions and we
assume, without loss of generality, that the Schmidt values λk are sorted descendingly. On this
basis, we make the following observations:

• A loop-free TN can be viewed as a simultaneous Schmidt decomposition over all virtual


bonds. This point of view is emphasized in the canonical gauge of Sec. 4.2.3.

• At each network bipartition, the bond dimension D provides an upper bound to the
Schmidt rank, which measures bipartite entanglement.

• Discarding the smallest singular values achieves the optimal single link compression.
More precisely, suppose that given a virtual bond η, we wish to reduce its prior dimension
Dη to some smaller dimension χη . We achieve this compression by simply truncating the
summation at k = χη in Eq. (56) and discarding (setting to zero) the smallest Schmidt
values in the decomposition. Thus, we obtain the following truncated state
χη
X
|Ψ trunc 〉 := |Ak 〉|Bk 〉λk /Nkept , (57)
k=1

with normalization constant v


u χη
uX
Nkept := t λ2 . k
(58)
k=1

Any other way of discarding λk ’s leads to a larger error in quantum state fidelity, according
to the Eckart–Young–Mirsky theorem [185]. Note that while being optimal for a single
link, compressing a quantum state by discarding Schmidt values on multiple bonds
simultaneously does not lead to the optimal fidelity (as each compression is a local
non-unitary operation; see techniques for an enhanced compression approach, e.g. in
Ref. [39] and Sec. 5.1.6). The most common purpose for link compression in loop-free
TN algorithms, such as the ground state search outlined in Sec. 5.1, is to preserve the
bond dimension D throughout the simulation: This is often ensured by re-compressing
bond dimensions after temporary growths.

• Similar to performing a bipartition through link-cutting, detaching an n-link network


node from a loop-free TN induces a partition into n disjunct subspaces. Each virtual index
then contributes variational degrees of freedom in form of associated tensor elements.

4.2 Gauges in Loop-Free Tensor Networks


In Sec. 4.2.1 we review the freedom introduced by gauges specifically in the framework of loop-
free TNs [26, 27, 39, 70]. We present the specific gauges that are related to the orthogonality
properties of the tensors in the network. We dub these gauges center gauges and discuss in
particular the unitary gauge (Sec. 4.2.2) and its refinement, the canonical gauge (Sec. 4.2.3).
Both are of great practical and conceptual importance. They install isometries in each tensor,
which strongly simplify contractions of the network with its conjugate (see again Fig. 25). They
also allow for efficient local error estimation and efficient application of local operators in time-
evolution algorithms and measurements. In variational algorithms, they maintain the network

50
SciPost Phys. Lect. Notes 8 (2019)

state’s normalization during local optimization (see Sec. 5). Conceptually, the canonical gauge
provides an understanding of the loop-free TN ansatz as simultaneous, independent, truncated
Schmidt decompositions over all lattice bipartitions induced by the network geometry.
Then, we introduce the symmetry gauge (Sec. 4.2.4): We demonstrate that any loop-free TN
ansatz state that transforms in- or covariantly under a global pointwise symmetry (as introduced
in Sec. 3.4) can be recast into a symmetric TN, i.e. made entirely of symmetric tensors, through
gauge transformations. Remarkably, this is compatible with both unitary and canonical gauge,
meaning that all those gauges can be installed at the same time, and without increasing the
bond dimension. Once installed, the symmetry gauge is strictly protected for any algorithm
formulated via the symmetric tensor operations that we described in Sec 3.5. This last property
is shared by the real gauge (Sec. 4.2.5), which can be useful in the presence of time-reversal
symmetry.

4.2.1 Gauge Freedom


When searching for the optimal gauge, it is helpful to keep in mind that the whole freedom
we have in choosing the tensor elements to encode a specific ansatz state |Ψ〉 in a loop-free
network with given link dimensions is entirely determined by link-local gauge transformations
as introduced in Sec. 1.7. This means that the only possible difference in the choice of tensors
are matrices X , Y inserted as products X · Y on virtual links and contracted into the adjacent
tensors, as we demonstrate in the following.

Gauges on Virtual Links and Minimal Bond Dimension. Let the TN have a virtual link η
of dimension Dη . We can perform a Schmidt decomposition over the original state |Ψ〉 between
the two subnetworks connected by η, and possibly obtain a lower link dimension χη ≤ Dη .
We now discuss how to gauge-transform the tensor network via local operations to obtain the
explicit decomposition. The reduction of dimension is non-trivial as it requires the insertion of
a pair of matrices X and Y of dimensions (Dη × χη ) and (χη × Dη ) on the link. Theoretically,
the existence of the gauge matrices can be shown as follows:

(i) In the given TN, we contract over all links except for the selected bond η and obtain two
tensors à and B̃ for the respective bipartitions. We can rewrite these tensors as matrices
à and B̃. As every physical site of the system belongs to one of the two partitions, we
collect the physical degrees of freedom of the respective subspacesP in indices a and b.
We can then write the coefficient tensor of the TN state as Ψa,b = k Ãa,k B̃k,b , where

k ∈ 1, . . . , Dη is the index of bond η.

(ii) StartingPfrom the original state |Ψ〉, we can also compute an exact Schmidt decomposition
χη
Ψa,b = k=1 Aa,k Bk,b , where we already absorbed the Schmidt values into the matrices A
or B and the indices a and b label the physical degrees of freedom as above. χη denotes
the Schmidt rank of Ψ and we now assume that χη < Dη , i.e. for the description of Ψ a
lower bond dimension than given in the TN above suffices.

(iii) We will now determine two gauge matrices X and Y which can be inserted on bond η
and then be absorbed into adjacent nodes, thereby reducing the original link dimension
from Dη to χη , without modifying the encoded state. By exploiting à · B̃ = A · B we arrive
at:
A = Ã B̃B −1R =: ÃX
 
(59)
and
B = A−1L Ã B̃ =: Y B̃ ,
 
(60)

51
SciPost Phys. Lect. Notes 8 (2019)

(a) (b) (c)

1 1 2 1 1 2 1 1 2
⇒ ⇒
c c c
2 2 2
1 1 1

Figure 26: Example for a loop-free TN gauging procedure. The numbers indicate
the distance of each node from the center c, determining the order of isometriza-
tion: Starting from the un-isometrized network (a), tensors with largest distance are
isometrized first, and the next tensor towards the center is updated (b). Then the
isometrization continues on the next level (c). This procedure is repeated until the
center (red) is reached.

where the expressions in square brackets now define the anticipated gauges X and Y .
The right-inverse (left-inverse) B −1R (A−1L ) exists due to the full row-rank (column-rank)
χη of the matrix B (A) resulting from the Schmidt decomposition.
While X · Y 6= 1 in general, these gauges have the property Y · X = 1:

Y X = A−1L ÃB̃B −1R = A−1L ABB −1R = 1. (61)

(iv) By repeating the procedure on all virtual links of the network, we obtain the gauges that
transform an arbitrary given TN into an equivalent TN of minimal bond dimensions equal
to the Schmidt rank on the respective bond.

Consequently, two loop-free TN representations TN1 and TN2 of the same state |Ψ〉, with
the same geometry, are connected by a link-local gauge transformation: We can explicitly
construct two sets of local gauges {X 1 , Y1 } and {X 2 , Y2 } which transform either one into the
(same) TN with minimal bond dimensions.

Gauges on Physical Links. On physical links, gauge transformations shall be restricted to


unitary transformations to keep the associated basis states orthonormal. Such a change of basis
can e.g. be helpful for entirely real Hamiltonians, as these allow us to solve eigenproblems with
entirely real TNs, therefore consuming less computational resources. Gauge transformations
on physical links can also be used to install the symmetry gauge (more on this in Sec. 4.2.4).
Note that within symmetric TNs, we are restricted to symmetric gauge matrices, and the local
basis can only be changed within degenerate manifolds associated to the quantum numbers.

4.2.2 Unitary Gauge


The unitary gauge is always defined with respect to a single selected node c of the TN and
therefore also referred to as “central gauge” (some literature refers to this gauge as the canonical
gauge, while we prefer to adopt this term for the gauge defined in Sec. 4.2.3). Starting with
nodes of maximal distance d = dmax from c, the unitary gauge can be installed on this level-
structure {L(d; c) : d = 0, . . . , dmax } as follows (see also Fig. 26):

(i) On each node q ∈ L(d; c) perform the QR decomposition

D̃η
[q] [q] [q]
X
T{s},{α},α = Q{s},{α},β Rβ (62)
η η η ,αη
βη =1

52
SciPost Phys. Lect. Notes 8 (2019)

(a) (b)

C [η]
T [c]

QR contract

(c) (d)

QR 0
C [η ]
0
T [c ]

contract
(e)
00
T [c ]

Figure 27: Moving the center in unitary gauge: For an excerpt from a possibly
larger TN, we show how the gauge center (red) is moved from network node c to
c 00 . Network center-tensors on the path of links η, η0 between the old and the new
0
designated center node are QR-decomposed (a,c) and the non-unitary parts C [η] , C [η ]
are subsequently contracted into the respective next node on the path (b,d). Tensor
isometrization (gray-white shading) is always directed to the current gauge center.

(a) (b)
[η] T̃ [q] T̃ [q]
C C [η] =
= ⇔

T [q] C [η̃] T [q]


(C [η̃] )−1

Figure 28: (a) Re-isometrization of a node in the unitary gauge (essentially steps
(b)→(c) and (c)→(d) from Fig. 27), achieved by QR decomposition. (b) Formulation
of the same transformation in terms of a link-local gauge transformation (see Sec. 1.7),
by contracting the respective non-unitary center matrices C [η] and (C [η̃] )−1 (both red)
into their adjacent node. However, this is only possible if their (single-sided) inverses
exist, i.e. if the bond dimensions equal the matrix ranks.

53
SciPost Phys. Lect. Notes 8 (2019)

on the bipartition between the virtual link η that leads towards c and all remaining
(physical and virtual) tensor links of q, which provide the remaining indices {s} and {α}.

(ii) Update T [q] → T̃ [q] := Q[q] with the semi-unitary isometry of the QR decomposition.

(iii) Contract the (upper trapezoidal) matrix C [η] := R[q] over the corresponding virtual link
η into the adjacent tensor from the inner layer L(d − 1; c).

(iv) Advance inwards to d → d − 1 and repeat from step (i) until d = 0 is reached.

After this procedure all tensors have been updated to T̃ [q] . Also, virtual links might
turn out reduced in bond dimension from Dη to D̃η , since the QR decomposition applied in
step (i) automatically selects the minimum between Dη and the product of the remaining link
dimensions.
Note that the iteration presented above indeed implements a global gauge transformation,
because it does not change the physical state of the network: Neither QR decompositions nor the
propagation of R-transformations over virtual bonds alter the contraction of the affected tensors
over shared virtual links. Computationalcomplexity is dominated by the QR factorizations (see
Sec. 2.2.2), upper-bounded by O D Z+1 , where D is the maximal bond dimension and Z the
maximal number of links of a tensor.
For algorithms operating in unitary gauge, a common task is to move the center from one
node c1 to another node c2 (as we will see in the variational search for ground states in Sec. 5).
This can be achieved by a sequence of QR decompositions T [q] = T̃ [q] C [η] of all nodes q on the
path between c1 and c2 , scaling linearly with the distance, O(dist(c1 , c2 )). In every step, we
obtain a possibly different non-unitary part C [η] which is passed over the shared bond η from
the re-isometrized node T̃ [q] to the next tensor on the path towards c2 . This process is shown
in Fig. 27, and yields the isometry relation of the unitary gauge, shown in Fig. 28(a),
[q] [η] [q] [η̃]
X X
T{s},{α},β Cβ ,α = T̃{s},{α},β Cβ ,α , (63)
η η η η̃ η̃ η̃
βη βη̃

where C [η] and C [η̃] are the (non-unitary) center matrices defined on any two links η and η̃ of
q respectively, as they might be encountered on a path moving the center. This relation holds
for every network node q and transforms the tensor T [q] , which is isometric over η, into the
tensor T̃ , which is isometric over η̃.
If we had access to the single-sided inverses (C [η] )−1 , the re-isometrization in Eq. (63)
could be performed by local contractions avoiding costly matrix factorizations (see Fig. 28(b))
and the unitary gauge transformations could be understood in terms of link-local gauges as
discussed in Sec. 1.7. However, the inversion requires the matrix C to be of full rank, and we
need to find a way to reduce the possibly oversized matrix dimensions first: There exists a
very elegant gauge that realizes such minimal bond dimensions together with a very efficient
re-isometrization rule. We will discuss this gauge in the following section.

4.2.3 Canonical Gauge


A canonical tensor network [26] is obtained from singular-value decomposing the unitary-gauge
center matrices C [η] on all bonds η, as in
χη
[η] SVD [η]
X
Cα = Vα[η]
η ,γη
λ[η]
γη Wγ . (64)
η ,βη η ,βη
γη =1

The isometries V [η] and W [η] merely act as link-local gauge transformations on η (and pairs of
the form V [η] V [η]† or W [η]† W [η] can therefore be introduced on the respective links at will, as

54
SciPost Phys. Lect. Notes 8 (2019)

(a) (c)
V [η] W [η]
SVD unitary
=
=
C [η] λ[η]
with
[η]† [η]
V V
=


(1)
=
W [η] W [η]†
=

(b)


(2)

T̃ [q]
=
canonical gauge
:=

(3)
[q]
T

=
unitary
gauge
canonical
Figure 29: Transformation from unitary- to canonical gauge: (a) We define isometries
W [η] , V [η] (black triangles, pointing towards the center singular values) and their
one-sided inverses (white) from left- and right singular vectors of the unitary-gauge
center matrices C [η] (red) on all virtual links η in a TN. (b) The canonical-gauge
tensor (double outline) is obtained from the unitary-gauge tensor by absorbing, from
all virtual links, those isometries that point towards the current TN gauge center C [η]
sitting on some network link η. Inverses are taken if tensor links lead away from
the center. When repeated on each tensor, this procedure establishes the canonical
gauge by insertion of link-local gauges VV † or W † W on each non-center virtual link
(cf. Fig. 30), and V [η] λ[η] W [η] on the center link η. (c) The isometry relation of the
canonical gauge (bottom) follows from the isometry relation in the unitary gauge
(top): We (1) perform SVDs on the center matrices and (2) multiply all (virtual) links
on both sides with the respective inverse isometries pointing towards the tensor node.
(3) By exploiting (a) and (b), we recover an identity and the definition of a canonical
tensor. The result (bottom) is the isometry relation in the canonical gauge picture. In
order to visualize the compression to minimal bond dimension, entailed by the SVD,
we assign different line widths to links of different dimension.

55
SciPost Phys. Lect. Notes 8 (2019)

= =
move SVD
center

=
= = =
insert contract move
identity center

Figure 30: Proof that pairs of isometries W † W act indeed as link-local gauges in
the transformation from unitary- to canonical gauge (see also Fig. 29(b)): In unitary
gauge, we first move the center (red) onto the selected virtual link (here depicted with
two adjacent nodes as part of a larger TN) and compute its SVD (top left to right). The
SVD yields a left-unitary term V and a right-unitary term W with V † V = WW † = 1,
but the commuted products do not yield identities in general. We then insert such
an identity (here WW † ), restore the center matrix, and move it back to its original
position (bottom left to right). This leaves the gauges W † W inserted on the link,
while none of the steps changed the network state. The same is true for an insertion
of VV † .

shown in Fig. 30). They are absorbed into the adjacent nodes, thereby updating the network
tensors T [q] → T̃ [q] to the canonical tensors as illustrated in Fig. 29. Singular values that are
zero or fall below a numerical threshold are discarded, so that the virtual links η attain their
minimal bond dimensions χη ≤ Dη equal to the respective center-matrix ranks.
A practical approach to installing the canonical gauge does not have to find the center
matrices C [η] on all bonds explicitly. It is sufficient to install the unitary gauge with a single
center-bond η towards which all tensors are isometries. Then a series of SVDs (instead of
simple QRs) for moving the unitary center, performed in reverse order from the center outwards,
establishes the center matrix successively on all bonds (as in Fig. 31).
Due to the orthonormality induced by the isometries, every SVD of the center matrix, which
can be considered as a reduced amplitudes tensor, also achieves a Schmidt decomposition
of the complete TN state, with Schmidt values λ[η] and Schmidt rank χη on every virtual
link [52]. Hence, the canonical gauge can be viewed as the explicit and simultaneous Schmidt
decomposition of the QMB state on all lattice bipartitions induced by the loop-free network
geometry.
By uniqueness of the Schmidt decomposition (up to phases and degeneracies) when ordering
the Schmidt values descendingly, the canonical gauge becomes similarly unique. This gauge
implements minimal bond dimensions simultaneously on all network links, which cannot be
reduced further without altering the state. The main reason for actually installing the more
costly canonical gauge is that the isometry relation Eq. (63) in this case translates into a mere
multiplication with diagonal matrices (which can be implemented as a scaling of tensors with
link-weights, cf. Sec. 2.2.2), as shown in Fig. 32:
[q] [q]
T{s},{α},α λ[η] [η̃]
αη = T̃{s},{α},α λαη̃ . (65)
η η̃

Again, the tensor T [q] , which originally is isometric over η, turns into an isometry T̃ [q] over
another of its virtual links η̃. As we can easily afford to store and invert the Schmidt values

56
SciPost Phys. Lect. Notes 8 (2019)

(a)

=
(1)

= = = = ··· =
(2) (3) (4) (5)

(b)
= ⇔ = =

Figure 31: (a) Practical installation of the canonical gauge. (1) The gauge center
(always red) of a TN in unitary gauge is SVD-decomposed. Starting with the nodes
adjacent to the center, we perform the following steps (resembling a higher-order
SVD [186]): (2) Absorb unitary matrix and singular values from parent link, (3)
perform SVD with respect to a first child link, (4) re-absorb singular values, (5) repeat
previous two steps on remaining child links. The final tensor (without the unitaries
on the child links) is now in canonical gauge. (b) Comparison of the final and the
initial tensor allows us to recover the definition of the canonical gauge.

λ[η]
=

(λ[η̃] )−1
Figure 32: Re-isometrization of a node in the canonical gauge: If we store the Schmidt
values λ[η] (red) for all links of a network, we can perform a re-isometrization of a node
(and consequently of the full network) using only numerically efficient contractions
with these “link-weights” (see Sec. 2.2.2) and their inverses. This approach is much
faster than the iterative sequence of QR decompositions in Fig. 27. Due to the minimal
bond dimension in canonical gauge, the inversion is always possible. Moreover, the
Schmidt values can be efficiently stored.

57
SciPost Phys. Lect. Notes 8 (2019)

λ[η] of all bonds, we are able to quickly re-isometrize the TN with only local operations. A
possible danger when implementing this step is that the inversion of small Schmidt values (if
not discarded in a compression) can result in numerical instabilities. However, workarounds
for this issue have been discussed [187].
The re-isometrization relation of the canonical gauge allows for parallelized algorithms
(see Ref. [52]) and eliminates directional dependency of errors, arising for instance in the
compression of virtual links performed by many TN algorithms. Therefore, the canonical gauge
is a favorable choice in algorithms driven by (local) unitary evolution, e.g. real-time evolution
algorithms, because a unitary operation does not alter the Schmidt decomposition except for
the bonds it acts on [52]. Hence, (costly) re-installation can be avoided. To a certain extent,
this can also be true for operations which are only approximately unitaries or identities. This
includes the aforementioned compression of virtual links, if it is carried out by truncated SVDs
which locally restore the gauge.

4.2.4 Symmetry Gauge


In the symmetric gauge, all tensors in the TN representing a covariant state become symmetric
tensors as discussed in Sec. 3.3. Contrary to the gauges discussed so far, its installation may
require a (fixed) transformation in the physical basis. While in practice one usually never
explicitly transforms into symmetry gauge, and no simple recipes exist for this operation, we
show that such a transformation is theoretically possible for any TN state:

(i) Identify the group of unitary, global pointwise symmetry operations Vg : g ∈ G and the
global quantum number `select of the encoded QMB state.
[ j]
(ii) The global pointwise operations Vg define on every physical site j a representation of
[ j]
G itself. Focusing again on Abelian groups for simplicity, we can diagonalize Vg and
obtain a new basis of possibly degenerate irrep states |s̃〉 j . We choose these as our new
physical basis states.

(iii) Contract the whole TN into a single tensor Ψ. This tensor is now by definition invariant or
covariant and can directly be written as a symmetric tensor as in Eq. (31). In the covariant
case, i.e. in the case `select 6= 0, we need to attach a corresponding quantum-number
selector link to this tensor.

(iv) Decompose this tensor again into a TN of the original geometry, resorting to exact symmet-
ric tensor decompositions as described in Sec. 3.5.2. The details of these decompositions
(bond dimensions etc.) are irrelevant. We can always obtain the original (loop-free)
geometry, representing the same state (because basic decompositions do not alter the
state); hence we have already shown that it is a gauge transformation.

(v) Install the canonical gauge in the symmetric TN. As this task requires only basic QR,
SVD, fusing (and subsequent splitting) operations, and contractions — all of which are
well-defined symmetric tensor operations — we can always install the canonical gauge
in a symmetric TN. By uniqueness of the Schmidt decomposition (up to degeneracy and
ordering of singular values), we can obtain the same minimal bond dimension as in a
non-symmetric TN. Note that the local basis transformations in step (ii) explicitly do
not alter the bond dimensions because, being local operations, they cannot increase the
entanglement (Schmidt rank).

Note that such a construction is also possible in loopy geometries. However, the canonical
gauge of Sec. 4.2.3 does not exist in such networks and the loops introduce further constraints

58
SciPost Phys. Lect. Notes 8 (2019)

on symmetric links that may lead to larger bond dimensions. The symmetric tensor ansatz can
still be beneficial for a loopy network, but a trade-off between the speed-up through the use of
symmetric tensors and the unfavorable increase of bond dimensions must be made [180].

4.2.5 Real Gauge


In the real gauge all tensor elements are in the real domain. As such, it is not restricted to loop-
free networks. If it can be installed in the tensors of a TN state and the operators (Hamiltonian)
acting on it, the real gauge can increase computational speed and reduce the variational space
by 1/2. This is, e.g., the case in diagonalization problems, where it can usually be installed in
the presence of a time-reversal symmetry (see Appendix A). As for the symmetry gauge, the TN
operations we are considering can always be implemented such that they yield a real output if
the input is real.
When time-reversal and global pointwise Abelian symmetries are present, one can often find
a local basis transformation that simultaneously installs the real gauge on top of the symmetry
gauge, by exploiting the freedom of phases, or more generally unitary transformations in
degeneracy spaces, in the local physical basis.

4.3 Preparation of Symmetric TN Ansatz States


The initialization of symmetric tensor network states presents an additional level of difficulty
with respect to initializing ordinary TNs: This is due to the fact that we need to assign a
symmetry group representation to every virtual link, and these representations must be such
that the symmetric tensors contain actual matches. In the pathological scenario of a tensor
not containing any match due to the link representations, the TN is representing no state at all
(zero vector). As as rule of thumb, the more matches are allowed, the less information in the
TN is wasted. Constructing meaningful initial representations for virtual links is in general a
non-trivial, non-local problem that involves fusion rules of the symmetry, but can nevertheless
be treated on a general ground.
The simplest approach is to initialize the symmetric TN state as a product state, i.e. in any
configuration which respects the symmetry and for which every virtual link holds only a single
sector of degeneracy one. Consider, for instance, a system with N physical sites, each of which
can hold up to n particles, and a U(1)-symmetry, which fixes the total particle number Np . We
could assign these particles to physical sites and, following the selection rules of the symmetry,
assign a single sector of degeneracy one to every virtual link of the network.
In a more sophisticated approach, we could use knowledge about the expected spatial
particle distribution in the state of interest (usually the ground state). As discussed earlier, any
link η of a loop-free tensor network introduces a bipartition of the underlying physical system
into e.g. x and N − x sites. Typically, the expected value of particles in the subsystems then
establishes the most relevant sector of η with the largest degeneracy. Around that sector, and
in the range of possible quantum numbers, one can now distribute degeneracy dimensions
with a certain variance until the maximal bond dimension is exhausted. Suppose we have no
prior knowledge about the distribution of the particles and their correlations. For fermions or
hard-core bosons we could then make a binomial ansatz and consider initializing the sector `
(equal to the number of particles in the subsystem of length x) with degeneracy
 ‹  ‹
¯ x N−x
∂` ∝ min , . (66)
` Np − `

Additional information, such as a harmonic confinement of the system, can be taken into account
by centering the largest degeneracies around particle expectations from e.g. a Thomas-Fermi
density distribution [188]. If we expect an insulating behavior where the quantum numbers

59
SciPost Phys. Lect. Notes 8 (2019)

(a) (b)
2 2 2
1 1 1

1 0 1 0 1 0
1 1 1

fuse

(c)
2 2
1 1

1 0 (1) 1 0
1 1

clean-up

2 2
1 1
⇒ ⇒
(2) 1 0 (3) 1 0
1 1

Figure 33: First steps in the preparation of a randomized symmetric ansatz state: (a)
Set up a level-structure, assign directions for symmetric links and equip the TN with
a selector link (blue). (b) Build up virtual links by means of fusing: Starting from
the physical links, fuse links at nodes of the highest level to obtain the links towards
the next lower level; successively repeat this procedure until reaching the center. (c)
Clean-up round: (1) Intersection (green circle) with fused link (red dashed) yields
cleaned-up link (green). (2) Clean up remaining virtual links at the center node. (3)
Propagate the clean-up outwards.

are spatially locked, a narrowly peaked degeneracy distribution with small variance might be a
good ansatz. Obviously, such initialization schemes are highly problem-dependent.
As a very general initialization method, we now introduce a scheme based on randomization,
which makes no a priori assumption and thus can be always adopted.

4.3.1 Randomized Symmetric Ansatz States


In this construction, we start with a loop-free TN where the geometry, the physical links and
the selector link are defined and a maximal bond dimension D is given, while other virtual
links and symmetric tensors are to be constructed. The recipe we present guarantees that
the network is non-trivial in the end, and that all the degeneracy tensors in the network may
indeed contribute to the state (and do not drop out in a contraction to the complete amplitudes
tensor). This recipe can be applied to an arbitrary symmetry group. It makes extensive use of
the symmetric link clean-up operation defined in Sec. 3.5.1, and it is sketched in Fig. 33.

(i) We choose a center node c and assign a level-structure to the network, as we discussed
in Sec. 4.1.

60
SciPost Phys. Lect. Notes 8 (2019)

(ii) Starting from the outer levels (branches) and moving inwards towards the center, we
fuse the outer links (already defined) of a tensor and assign the inner link to be the fusion
of the outer links. We repeat this step towards the center node, so that the outer links
are always well-defined (they possess a representation), by recursion.
Option: We can limit the dimensions of the inner links by intersecting the fusion of
the outer links with a user-defined “maximal” link (for symmetric link intersection see
Sec. 3.5.1). This option presents a possible bias to the random state, but it can make the
approach numerically more efficient.

(iii) Once we reach the center node and all the links have been defined by the previous step,
we start to clean up the links from the center outwards. First, we clean up all the virtual
links connected to the center node. Then we move outwards into the branches until all
virtual links have been cleaned up. At this point, no (physical) truncation has occurred
yet.

(iv) Select randomly a virtual link whose total dimension is larger than D. Within this link,
remove randomly a single symmetric link index. Then clean up all the other virtual links,
starting from the closest ones in the distance and moving outwards, to update this change.
No link is completely emptied by the clean-up procedure as long as D > 0. Repeat this
step until all virtual links have dimension D or less. This completes the virtual links setup.

(v) Finally, generate the tensors. Within each tensor, include all possible matches given the
links, and assign a random degeneracy tensor to every match (see Sec. 3.5.1). Eventually,
install a gauge.

Note that in the reduction step (iv), at every iteration we reduce a link dimension by one
and then propagate the truncation: This is a conservative choice, which allows every state |Ψ〉
within the given bond dimensions D to be generated. Nevertheless, quicker reduction strategies
can be considered. Furthermore, during clean-up, propagation into a branch can be stopped as
soon as the link leading outwards remains unchanged by the clean-up operation.
Now we have highlighted the main features of loop-free TNs, and identified the complete
freedom in defining the tensors and links therein, exploited in the unitary, canonical and finally
the symmetric gauge. In the next section we take advantage of those features and describe
exemplarily a TN algorithm which searches ground states in a loop-free TN.

5 Ground State Search with Loop-Free Tensor Networks


In this section we discuss a robust scheme for achieving ground states (GS) of Hamiltonians
by means of tensor networks with no loops (or cycles) in their geometry. This is a natural
extension of the DMRG algorithm [39–41] adapted to any tensor network architecture without
cycles [73, 74, 79, 181–183, 189–194]. The absence of cycles leads to several computational
advantages: We have already encountered one of them, namely the TN gauge freedom, in
the previous section and we will exploit it extensively in what follows. Another striking
advantage of a loop-free geometry is the possibility to formulate local optimization problems
(which are at the heart of efficient variational energy minimization procedures) as simple
eigenvalue problems. This feature is not present in TNs with cycles, where often no exact
strategies for the solution of local optimization problems are available, and one has to rely
on approximate approaches (as occurring e.g. for MERA [66], PEPS [128, 195]). Moreover,
loopy tensor networks often imply performing more expensive tensor contractions, and dealing
with symmetries is less straightforward (as mentioned in Sec. 4.2.4): This is, for instance, the

61
SciPost Phys. Lect. Notes 8 (2019)

(a) Tvar (b)

|Ψ〉 Tvar

H Heff


〈Ψ| Tvar


Tvar

Figure 34: Basic idea of the variational ground state search: (a) In order to reduce
the complexity of the minimization problem, only the parameters of one tensor
Tvar (red) are taken as variational while all other tensors are taken as fixed. The
variational tensor forms the “optimization center", while the fixed tensors form an
“environment". The latter can be efficiently contracted with the Hamiltonian H, leading
to a reduced “effective" Hamiltonian Heff . (b) By solving the reduced optimization
problem sequentially (for varying locations of the optimization center), the energy
expectation value is gradually reduced. This iterative minimization is commonly
called “sweeping".

case for periodic boundary DMRG, where specific methods must be engineered to work around
these issues [196, 197].
For these reasons, we choose to focus on ground state search algorithms for loop-free TN
setups instead. This will break down the linear algebra demands to only three basic procedures
(contraction, SVD/QR decomposition, simple eigenvalue problem) and take advantage of
adaptive TN gauges [181, 182].

5.1 Ground State Search Algorithm for the Adaptive-Gauge Loop-Free TN


In this section we describe in detail the algorithm we adopt to achieve the QMB ground state
variationally with loop-free TNs. The basic idea is to break down the diagonalization problem
for the full QMB ansatz into a sequence of smaller, manageable diagonalization problems,
each of which is formulated for a small subset of tensors (in the simplest case for one single
tensor, as illustrated in Fig. 34(a)). Starting from a possibly random initial state in the manifold
parametrized by the TN ansatz, we thus employ an iterative scheme of local tensor optimizations
(“sweeping", see Fig. 34(b)) that converges to the state which best approximates the true QMB
ground state. This approach is akin to what is known in applied mathematics as domain
decomposition [198, 199].
Before listing the algorithm in Sec. 5.1.2, we first describe in detail the core concepts and
high-level operations (based on the elementary ones defined in Sec. 2 and Sec. 3) that it will
rely on. Note that our discussion applies both to symmetric and non-symmetric loop-free TNs,
with the obvious requirement that in the symmetric case all tensors and tensor operations have
to be formulated in a symmetry-aware fashion as detailed in Sec. 3.

5.1.1 High-Level Operations

In what follows, we consider a loop-free TN of tensors T [q] sitting at network nodes q = 1 . . . Q,


mutually joined by virtual links η (with dimension Dη ). We can arbitrarily choose any node to
be the center node: Here, we exclusively work in the unitary (center) gauge (see Sec. 4.2.2),

62
SciPost Phys. Lect. Notes 8 (2019)

a) is b) is is+1

γµ [s]
γµ0 [s] γ1,2 [s+1]
h p,k h p,1 h p,2

γµ00
js js js+1

c) d)
i1 i2 i3 γ L,1 iL

[1] [2] [3] [L]


h p,1 γ h h p,3 γ h
1,2 p,2 γ2,3 3,4 γ L−1,L p,L

j1 j2 j3 jL

Figure 35: The TPO interaction terms H p which sum up to the full Hamiltonian. (a)
[s]
Each interaction p is composed of local operators h p,k where [s] stands for the TN
physical link (or site) at which it is applied. The index k enumerates the TPO operators
from k = 1 . . . K p , with K p being the TPO operator length (in general different from the
interaction range), and p distinguishes different interactions in the full Hamiltonian.
Any two TPO operators k, k0 can be interconnected by a TPO link µ = (k, k0 ), drawn in
gray, but efficiency mandates to keep their overall dimension small. (b)-(d) Examples
of interactions: (b) a nearest-neighbor product interaction (such as the Heisenberg
Hamiltonian), (c) a global TPO and (d) a possible interaction of a two-dimensional
model, where the center-tensor has however four TPO links.

and introduce the notion of an effective Hamiltonian.

Hamiltonians. For the GS algorithm we consider Hamiltonians of the form


P
X
H= Hp , (67)
p=1

and we focus on Tensor Product Operator (TPO) [181] interactions of the form
— { γµ }
[s]
  X Y”
=
k(s)
Hp {i },{ j }
h p,k(s) , (68)
s s is , js
{γµ } s

[s]
where h p,k(s) is the k-th local operator of the TPO acting on site s and the operators are
interconnected by links µ = (k, k0 ), labeled by the two operators k and k0 they connect.
The sum runs over all TPO links {µ}, and we define {µ}k(s) ⊆ {µ} as the set of all links µ
attached to the k-th operator of the TPO. The outer product can be restricted to sites on which
H p acts non-trivially. TPOs can be viewed as a generalization of Matrix Product Operators
(MPOs) [39, 56, 200], widely used in the context of MPS, to arbitrary TN geometries. Some
examples for how a TPO H p can look like are given in Fig. 35.
Computational efficiency is maintained by keeping the number and bond dimensions of
TPO links small. Specifically, we consider interactions to be efficient if they are compatible with

63
SciPost Phys. Lect. Notes 8 (2019)

(a) (b)

η η

Figure 36: Two possible configurations for global TPO interactions (white tensors)
acting on a TN state (gray tensors). Every local operator has two non-trivial TPO
links: While the TPO configuration in (a) is considered efficient, in (b) there are four
TPO links joining the network bipartition over the virtual link η (indicated by the
dashed red line).

the TN geometry, in that every network bipartition over any virtual link η corresponds to a cut
of two non-trivial TPO links (see Fig. 36). More generally, the condition to beQ met in order
to keep computational efforts manageable is that the maximal total dimension Dµ of links
crossing any bipartition should be non-scaling (or at least weakly scaling) with the system
size. Typical 1D lattice Hamiltonians fulfill this criterion, including sums of nearest-neighbor
interactions, long-range two-site interactions, and TPOs in alignment with (i.e. having the same
arrangement as) the physical sites, as exemplified in Fig. 35.
As the representation in Eqs. (67) and (68) is not unique, it has to be decided individually
whether e.g. a large number P of lightweight interactions (i.e. with small bond dimensions
Dµ ) is more efficient than an equivalent representation with smaller P but more or larger TPO
links {µ}. On the one hand, there are certain interaction types (e.g. exponentially decaying
interactions) which are known to have a very efficient global TPO representation, characterized
by a constant, non-scaling link dimension Dµ [39]. On the other hand, it has been noted that a
single global TPO representation can be less efficient than treating the individual interaction
terms separately [181], for example when interactions are long-range and inhomogeneous
(e.g. in spin glass models or quantum chemistry applications). In particular, the algorithm can
easily exploit the locality (i.e. the spatially restricted non-trivial action) of individual interaction
terms, as shown in the next paragraph.

Renormalization. Once a Hamiltonian TPO has been provided, we assign each non-trivial
[s]
local operator h p,k from every “bare” Hamiltonian term H p to the physical link s it acts on. We
now associate with each virtual network link ζ a renormalized Hamiltonian, which describes
the effective action of the physical Hamiltonian onto the subsystem (or bipartition part) Sζ
composed of all lattice sites separated over ζ from the current network center c. A single
[ζ]
renormalized Hamiltonian operator takes the form h p,k , where k is the set of all local operators
of the TPO in the renormalization k(ζ) := ∪s∈Sζ k(s); in particular, k(s) = k(s) on physical sites.
The following recursive relation defines the renormalized Hamiltonian operators associated
with the virtual link ζ that leads towards the current network center c at some node q. It
must be performed for each contribution p separately: We start with collecting all (physical or
renormalized) Hamiltonian terms on all remaining (physical or virtual) links {ξ}q , ξ = 6 ζ, of q.

64
SciPost Phys. Lect. Notes 8 (2019)

(b)
ζ
(a)
[η] T [q]
¦ ©
h p,k(η) ζ
η

p=1
p=2
p=3 =
... ξ00
ξ
ξ0

T [q]∗

(c) η000 η000

P T [c] T [c]
p =

[c]
Heff
η00 η00
η η
0
η η0

Figure 37: Building blocks of the GS algorithm: (a) Local interaction operators are
defined on the (physical and virtual) network links. Tensors inside the blue shaded
[η]
area on a link η represent local operations h p,k(η) , each with a different p. Not for
every p in the Hamiltonian’s sum Eq. (67) there must be a tensor defined at every
link η; the missing ones are implicitly treated as identities. (b) Renormalizing the
local interactions from the links ξ, ξ0 , ξ00 towards the virtual link ζ. Each interaction
p is contracted independently into a renormalized term, and the global sum over p
is not carried out explicitly. An exception are renormalized operators p that have
no TPO links left (for example the two independent interactions indicated in red),
which can already be summed into a single new renormalized tensor. (c) The effective
Hamiltonian involves the contraction of all renormalized Hamiltonian terms over
all tensor links {η}c at the center node c. Often, the final sum over p is postponed,
[c]
e.g. until contracting the individual terms h p,eff into the tensor T [c] as described in
Sec. 5.1.3 (solving the reduced optimization problem). The figure also shows the
center tensor (semi-transparent) to highlight how it is acted upon by the effective
Hamiltonian, which is a key ingredient for the GS algorithm.

65
SciPost Phys. Lect. Notes 8 (2019)

(a) ζ (b) (c)

ξ ξ0
+
[ξ] [ξ0 ] [ξ] [ξ] [ξ0 ]
h1,k(ξ) h1,k(ξ0 ) h2,k(ξ) h3,k(ξ) h3,k(ξ0 )

[ζ] [ζ] [ζ]


h1,k(ζ) h2,k(ζ) h3,k(ζ)

=
=

[ζ] [ζ]
h1+2,k(ζ) h3,k(ζ)

Figure 38: Example for summation over renormalized Hamiltonians without open
TPO links: the resulting TPO tensors on link ζ (dashed) in (a) and (b) have no open
TPO links, they can therefore be summed over, and saved in a single tensor. In (c),
this is not the case.

They are generally joined by some mutual TPO-links {µ}, representing the contraction
{γµ }k(ζ) —{γµ }
[ξ]
X Y”
k(ξ)

h p,k(ζ) = h p,k(ξ) . (69)
{αξ },{βξ } αξ ,βξ
{γµ }paired ξ6=ζ

The left-hand side ultimately combines k(ζ) = ∪ξ6=ζ k(ξ) renormalized physical operators of
the TPO. To obtain the renormalized Hamiltonian over link ζ we must additionally contract the
Hamiltonian operators in Eq. (69) over the network links ξ 6= ζ with the node-tensor T [q] , as
well as with its complex conjugate, as depicted in Fig. 37(b):
—{γµ } {γµ }k(ζ)
[ζ] [q] [q]∗
” X X
k(ζ)

h p,k(ζ) = T h p,k(ζ) T . (70)
αζ ,βζ αζ ,{αξ } {αξ },{βξ } {βξ },βζ
{αξ } {βξ }

With small TPO-link dimensions, an efficient order in which to perform the contractions
in Eqs. (69) and (70) is to generally first absorb the Hamiltonian operators by successively
contracting them over all shared links into T [q] , and finally contract with the conjugate node
tensor.
When working in the unitary gauge, the tensor T [q] appears as an isometry over the links
ξ 6= ζ. Therefore, it is immediately clear that the result of Eq. (70) will be the identity whenever
Eq. (69) produced an identity, i.e. whenever the TPO acts trivially on all links ξ 6= ζ. This feature
can save a lot of computational time on renormalizing interactions with spatially restricted
(local) support. Another optimization which can be applied at this stage is to carry out the sum
P [ζ]
p over all renormalized, non-interacting parts h p,k , i.e. those without any open TPO links
ζ
{µ}. All other renormalized interaction operators are stored in separate tensors on link ζ (see
also Fig. 38).
Given that the bare Hamiltonians are compatible with the TN geometry, the renormalized
Hamiltonians on the left-hand side of Eq. (70) have at most Q two non-trivial
Q  TPO links {µ} by
construction, and all contractions remain efficient in O Dζ 2 Dµ Dξ .

66
SciPost Phys. Lect. Notes 8 (2019)

(a) |Ψ〉

H1
+
H2
H=
+
H3
+ [s]
··· h p,k(s)

(b) e.g. p = 3
c ζ ζ
T [q]
ξ00 T [q] ξ00
ξ 0
ξ 0 ζ
ξ ξ

[ξ0 ] [ξ00 ] [ζ]


= h3,k(ξ0 ) h3,k(ξ00 ) = h3,k(ζ)
[ξ]
h3,k(ξ)
ζ
[q]∗ T [q]∗
T

c ζ ζ

(c) ζ0 ζ
c c

Heff = h
[ζ0 ] [ζ]
h1,k(ζ)
+ h
[ζ0 ] [ζ]
h2,k(ζ)
+ ···
1,k(ζ0 ) 2,k(ζ0 )

c c

Figure 39: Example for the renormalization process that leads to the renormalized
Hamiltonians: (a) The Hamiltonian H acting on the TN representation of some
state |Ψ〉 can be decomposed into TPOs H p . (b) In a bottom-up approach, departing
from the state-tensors closest to the bare Hamiltonian tensors, we can build the
[η]
renormalized terms h p,k(η) on all links η below the chosen center c by iteratively
contracting state tensors with the renormalized Hamiltonian terms of the previous
level (dashed circles). (c) We can then determine the effective Hamiltonian Heff at
the center node c by summing over the renormalized Hamiltonian terms acting on
the links of c.

67
SciPost Phys. Lect. Notes 8 (2019)

Effective Hamiltonian. At any given center node c (a unitary-gauge center), we define an


effective Hamiltonian
[c]
Heff = 〈Ψenv |H|Ψenv 〉 , (71)

where “env” stands for the TN contraction over all tensors except T [c] , i.e. the environment of
the center node. With this definition, the energy expectation value E of the TN state can be
written as (cf. Fig. 34(a))
[c]
E = 〈Ψ|H|Ψ〉 = 〈T [c] |〈Ψenv |H|Ψenv 〉|T [c] 〉 = 〈T [c] |Heff |T [c] 〉 . (72)

[c]
In practice, to obtain Heff , we do not need to contract the entire environment, if we have
already the renormalized Hamiltonians (as defined in the previous paragraph) on all links of
node c available. Then, Eq. (71) reduces to effective Hamiltonian terms (see Fig. 37(c))
—{γµ }
[c] [η]
” — X Y”
k(η)
h p,eff = h p,k(η) , (73)
{αη },{βη } η
αη ,βη
{γµ }
 
defined over all tensor links {η} at node c with corresponding indices αη , βη . The

contraction sum runs over all TPO links {µ} with corresponding indices γµ . To complete the
effective Hamiltonian, we have to sum over all interactions p:
[c] [c]
X
Heff = h p,eff . (74)
p

The whole process of building the effective Hamiltonian is summarized in Fig. 39.

Projectors. Adding projective operators to the Hamiltonian opens the possibility to explore
degenerate GS manifolds and to target excited states (we will comment further on this paradigm
in Sec. 5.1.5 below). More precisely, we apply the ground state search algorithm to the
Hamiltonian H0 = H + H p , where H p = " p |Ψ p 〉〈Ψ p |, to find the first excited state of H,
assuming its ground state |Ψ p 〉 is known, and where " p is much larger than the energy gap.
Here we discuss the explicit construction of projective summands H p = " p |Ψ p 〉〈Ψ p |, where |Ψ p 〉
is expressed as a TN with the same geometry as the variational TN ansatz state. The scaling
factor " p will play the role of an energy penalty for occupying the state |Ψ p 〉.
[q]
The projectors are thus defined via the (not necessarily isometric) tensors T p with possibly
different dimensions in the virtual links (and symmetry sectors if they are symmetric; but
the global symmetry group has to be the same). Projective operators can be implemented
within the same structure and almost the same equations as for the renormalized and effective
Hamiltonians. For a graphical representation, see Fig. 40. In the recursion relation of Eq. (70),
we replace T ∗ with T p∗ , and set identities for h[s]
p on physical sites:
!
[q] [q]∗
” — X X Y” —
h[ζ]
p = T h[ξ]
p T . (75)
αζ ,βζ αζ ,{αξ } αξ ,βξ p,{βξ },βζ
{αξ } {βξ } ξ

Note that here we dropped the internal TPO label k because the h p are defined on every link
P the projective terms H p have
without TPO links. It is important to keep in mind, however, that
global support. Therefore here we can never carry out the sum p , in contrast to renormalized
TPO interactions with local support which have no TPO links left. Furthermore, isometry can
no longer be exploited here, and the resulting renormalized projectors h[ζ]
p will in general be
different from the identity.

68
SciPost Phys. Lect. Notes 8 (2019)

(a) η (b)
ζ
interactions
[η]
¦ ©
= h p,k(η) T [q]
ζ
projectors ... ...
¦
[η]
© ...
= hp ...
=
00
ξ ξ
ξ0
[q]∗
Tp
(c)
T [c] ... ζ
...
... ...

η η000
η0 η00
T [c]
T p[c]∗
"p · =
P
p
[c]
T p[c] η000 Heff
η000
η 0
η00
η η η00

∗ η0

... ...

... ...

Figure 40: Adding projectors to the Hamiltonian: a) (Renormalized) local projection


operators are put on the links. Contrary to interactions which act non-trivially on
physical links, projective terms are always identities on physical links. b) Similarly
to the renormalization of TPO interactions, projectors are renormalized to a virtual
link ζ of a node q by contracting over all local projection operators on the remaining
[q]∗
node links η, η0 , η00 , with the conjugate tensor T p now taken from the projector
[ζ]
TN. The renormalized projectors h p have no TPO links, but in contrast to the rule
for TPO interactions they cannot be summed up over p (every projector results in
a separate renormalized two-link tensor on each link). c) The effective projector
at the center c is given by the contraction of the renormalized projectors over all
links with the conjugated projector-tensor T p[c]∗ . The resulting expression is then
multiplied by its complex conjugate. This contraction has to be repeated for every p in
the global Hamiltonian sum individually, constituting one summand of the complete
[c]
effective Hamiltonian Heff . As before, we show the center tensor (semi-transparent)
to demonstrate how the effective projector acts upon it.

69
SciPost Phys. Lect. Notes 8 (2019)

(a) ζ
c ζ
00
T [q]
ξ T [q] ξ00
|Ψ〉 0
ξ ξ ξ ξ 0

0 00 ζ
h[ξ
p
]
h[ξ
p
]

h[ζ]
p
[q]∗ [q]∗
Tp Tp

ζ ζ
ζ
|Ψ p 〉〈Ψ p | = =
ζ ζ ζ
[q] [q]
Tp Tp

h[ζ]∗
p

0 00
h[ξ
p
]∗
h[ξ
p
]∗
ζ

ξ ξ
〈Ψ| ξ0 ξ0
T [q]∗ 00
ξ T [q]∗
ξ00
c ζ ζ

(b) ζ0 ζ
c c

[ζ] [ζ]
[ζ0 ] h7 [ζ0 ] h8
h7 h8

[q]∗ [q]∗
T7 T8
Heff = ··· + "7 · + "8 · + ···
[q] [q]
T7 T8

[ζ0 ]∗ [ζ]∗ [ζ0 ]∗ [ζ]∗


h7 h7 h8 h8

c c

Figure 41: Example for the renormalization process that leads to the renormalized
[η]
projection operators: (a) Renormalized projector terms h p are built on all links η
below the center node c by iteratively contracting state tensors with the renormalized
projector terms of the previous level (dashed circles). (b) For a given center c, we
can then determine the projector contributions to the effective Hamiltonian Heff by
summing over the renormalized projector terms acting on the links of c.

70
SciPost Phys. Lect. Notes 8 (2019)

The effective projector at a node c becomes " p 〈Ψenv |Ψ p 〉 〈Ψ p |Ψenv 〉, which is analogous to
the definition of the effective Hamiltonian. Written out explicitly it reads
[c] [c] [c]∗
” —
h p,eff = " p T̃ · T̃ , (76)
{αη },{βη } p,{αη } p,{βη }

where ‚ Œ
[c] [c]∗
X Y” —
T̃ = h[η]
p T . (77)
p,{αη } αη ,βη p,{βη }
{ βη } η

Note that this lengthy expression is rather simple to compute, by contracting all renormalized
projectors into T p[c]∗ , and multiplying with the complex conjugate of the outcome. The resulting
[c]
links constitute the bra- and ket- links of the effective projector h p,eff , see Fig. 40(c). The whole
process of building effective projectors is summarized in Fig. 41.
Some specific applications, such as variational addition of TNs in Sec. 5.1.6, employ P projec-
tive operators |Ψ p 〉〈Ψ p | onto superpositions of TN states. Those are of the form |Ψ p 〉 = i " p,i |Ψ p,i 〉
with all |Ψ p,i 〉 given as separate TNs sharing the same geometry. In such a case, one computes
[ζ]
renormalized projectors h p,i for each state in the superposition separately — just as before, by
simply replacing the single projector index p with p, i in Eq. (75). A more relevant adjustment

is to be made in the effective Hamiltonian, which now reads i, j " p,i " p,
P
j 〈Ψenv |Ψ p,i 〉 〈Ψ p, j |Ψenv 〉.
Correspondingly, as a replacement for the single factor " p in Eq. (76), one additionally has to
sum over superposition states in Eq. (77):
‚ Œ
[c] [η] [c]∗
X X Y” —

T̃ = " p,i h p,i T . (78)
p,{αη } αη ,βη p,i,{βη }
i { βη } η

Nevertheless, evaluating such an effective Hamiltonian remains a simple computation by


[c]∗
successive absorption of all effective projectors into T p,i .

5.1.2 Ground State Search Algorithm in Detail


We now describe the algorithm, which is based on the high-level operations introduced in the
previous section, and follows the flowchart shown in Fig. 42. In the following, we list additional
explanations and remarks, useful when implementing such an algorithm.

Setup and Configuration Tasks.

(i) Definition of the TN geometry: From the physical point of view, the TN should be designed
such that it complies best with the system and the Hamiltonian being targeted. On the
one hand, the TN has to be compatible with the interaction terms in the Hamiltonian.
On the other hand, it has to be capable of hosting the required amount and spatial
distribution of entanglement, controlled by its virtual links. Finally, the entanglement
arises from the interactions and (at least for ground states) can often be theoretically
proven to be upper-bounded (e.g. by area laws [37]). Fig. 43 summarizes schematically
these requirements. At the purely technical level, any network fulfilling the following
requirements is allowed:

• The network has N physical links, supporting the system Hamiltonian H (defined
in Eq. (67)).
• The network is acyclic: for any pair of nodes q, q0 there exists a unique path
connecting q and q0 .

71
SciPost Phys. Lect. Notes 8 (2019)

Initialization

• define network geometry


5
4 • define sweep sequence
3 1
2
• initialize TN state

Target 1st opt. center

• install unitary gauge

• renormalize physical Hamiltonian


see Eq. (73) and Fig. 37
to effective Hamiltonian

Optimization loop
s 7→ s + 1 • increment sweep counter

see Sec. 5.1.3 • optimize center node


n+1
n
• target next optimization center

- adjust gauge

see Fig. 44 - adjust eff. Hamiltonian

sweep no
sequence end?
yes
energy no
converged?
yes
Finalization

• measure observables

• compute (bi)partitions entanglement

A B • ...

Figure 42: Flowchart of the GS algorithm. The left column provides illustra-
tions/references for the corresponding items in the bullet list.

72
SciPost Phys. Lect. Notes 8 (2019)

TN geometry
co
mp
s ati
nk ble
al li int
e
virtu rac
tio
ns

entanglement model (Hamiltonian)


area laws, ...
Figure 43: Elements influencing the choice of the TN geometry: The network has to
be compatible with the interactions present in the Hamiltonian, and at the same time
must be able to capture the ground state entanglement, which in turn is determined
by the Hamiltonian itself.

• The dimension of any virtual link in the network does not exceed D (bond dimension
of the TN).
• While in principle it is not prohibited, the employment of tensors with less than three
links is redundant and may deadlock symmetries. Absorbing one-link and two-link
tensors into adjacent tensors is strongly suggested, as it leaves the variational TN
state manifold unchanged while reducing the technical issues with symmetries (see
Sec. 5.1.3).

(ii) Definition of sweep sequence: This is the rule according to which the optimization centers
are chosen sequentially in the optimization loop. In principle, any sequence that visits
every tensor of the TN (and every variational link, in case of multi-tensor optimization)
is allowed. The specific form of the optimal sweep sequence, i.e. the one that achieves
fastest energy convergence within the shortest computation time, depends on the TN
structure and the model under consideration. For regular hierarchical networks (e.g.
binary TTNs, see Sec. 5.2) it is natural to select a sweep sequence based on the distance
from the physical links: Assign an optimization precedence p := mins [dist(c, s)] to every
optimization center c, where s is a physical link. Within a sweep sequence, optimization
centers with smaller p take precedence over centers with larger p. Centers sharing the
same p are chosen in such a way that the distance that has to be covered in order to visit all
of them is minimal. For less regular networks (e.g. TNs for quantum chemistry problems)
other schemes (such as depth-first search with backtracking) have been proposed [181].

(iii) Initialization of TN state: An obvious choice is to start from a random state (see Sec. 4.3.1
on how to construct such a random state for a symmetric TN with a defined global
quantum number). This choice guarantees the least possible bias and therefore bears
the smallest risk of ending up in a local energy minimum state, which would corrupt
the convergence to the GS (by definition the global energy minimum state). Another
valid option is to start from an already computed GS, previously obtained with the same
algorithm on another Hamiltonian, assuming the symmetries and quantum numbers are
the same.

(iv) Targeting of initial optimization center c:

• Transform the network into unitary gauge with respect to c (see Sec. 4.2.2).
• Build up the effective Hamiltonian for c, according to the procedure outlined in
Sec. 5.1.1.

73
SciPost Phys. Lect. Notes 8 (2019)

c0
c

c
c0

Figure 44: Moving the effective Hamiltonian from an optimization center c to another
optimization center c 0 . Only the renormalized Hamiltonian terms (Eq. (70)) defined
on the links located on the path from c to c 0 have to be updated (blue shaded). All
other renormalized Hamiltonian terms (contractions within the red-dashed area) are
left unchanged by this move and therefore do not need to be recomputed.

In case of multi-tensor optimization, we extend the notion of the network center to two
or more network nodes. This can be achieved by temporarily contracting some adjacent
nodes {q} into a single center node c. The renormalized Hamiltonian terms and isometry
properties have to be computed as usual for all remaining links and tensors, respectively.
After the optimization center has been prepared, the effective Hamiltonian is given by
Eq. (73). The energy expectation value E of the TN state is then (see Eq. (72))
[c]
E = 〈T [c] |Heff |T [c] 〉 , (79)

where T [c] is the tensor of the center node. Now, the generic local (reduced) optimization
problem consists in determining the variational tensor T [c] in such a way that it minimizes
E. This optimization problem is solved sequentially for all centers c in the following
optimization loop.

Optimization Loop.

(i) Optimize T [c] , i.e. minimize Eq. (79). The optimal T [c] is given by the eigenvector of
[c]
Heff corresponding to the lowest eigenvalue (see a proof in Ref. [89]). While this basic
observation remains always the same, there are technical differences for different types
of optimization centers (e.g. single-tensor, double-tensor, etc.). Sec. 5.1.3 is dedicated to
the different flavors of the “optimizer routine”.

(ii) Target the next optimization center c 0 according to the sweep sequence by moving
the isometry center to c 0 via gauge transformations and updating the renormalized
Hamiltonian terms between c and c 0 (see Fig. 44).

(iii) Loop back to step (i) and repeat the optimization with the new center node c 0 , until the
end of the sweep sequence is reached.

(iv) Check convergence. If not converged, start a new sweep. Typical convergence criteria
are: change in energy expectation value, variance of energy, distance of TN states from

74
SciPost Phys. Lect. Notes 8 (2019)

subsequent sweeps. Most commonly, one monitors the energy expectation value E(s)
after sweep s: Since E(s) is expected to approach the GS energy asymptotically from
above, a possible convergence criterion is to demand

E(s − 1) − E(s) < ε , (80)

where ε is a constant (positive) threshold. In practice ε has to be adapted according


to the bond dimension D in order to avoid unnecessary sweeps that would push the
convergence too far beyond the accuracy limit imposed by finite bond dimension. Usually,
the bond dimension dependence of this accuracy limit (and hence the form of ε(D)) is
well modeled by some polynomial or exponential decay (or combination thereof), e.g.
one might choose the ansatz ε(D) = a + b D−c exp(−D/d). In practice, the parameters a,
b, c, d need to be determined empirically by some fitting procedure because the form of
the decay highly depends on the model parameters (system size N , proximity to a critical
point, etc.).

5.1.3 Optimizer – Strategies


Here we describe in detail various possible strategies when interfacing with the reduced
optimization problem (while how to practically perform the numerical optimization routine
will be discussed in Sec. 5.1.4). The most commonly employed schemes for this purpose are
the single-tensor update and the double-tensor update, where the optimization centers consist of
one tensor and two tensors, respectively. We additionally introduce a single-tensor update with
subspace expansion.
The three schemes differ in the complexity of the variational problem they solve, and
have different requirements on computational resources. The first two schemes, single- and
double-tensor update, are a generalization of DMRG to arbitrary loop-free TN architectures.
They are directly related to the traditional single-center site and double-center site DMRG [40]
respectively. Finally, the subspace expansion scheme allows us to reconcile the improved
convergence behavior of the double-tensor update with the favorable computational cost
scaling of the single-tensor update.
In the single-tensor update, the optimization center consists of just a single network tensor,
and it is the most straightforward and computationally inexpensive scheme. However, it can
experience serious shortcomings: If the Hamiltonian under consideration exhibits a certain
symmetry, the same symmetry will usually be present in the effective Hamiltonian of the reduced
diagonalization problem. As a consequence, local optimizations are constrained to preserve the
link representations on the individual links of the optimization center. The present quantum
numbers (sectors) and their associated degeneracies can therefore not be handled variationally.
This issue is particularly acute for symmetries that are encoded in the TN ansatz (e.g. by the
methods described in detail in Sec. 3): in that case the mentioned constraints are hard and
will be exactly (by construction of the tensors) respected in each local diagonalization. For TNs
made from tensors that are unaware of symmetry rules the constraints are softer: In this case
they are only metastable configurations which can be violated through numerical fluctuations,
caused e.g. by finite precision floating point operations or by the fact that the environment (i.e.
the tensors surrounding the optimization center) does not obey the symmetry. Nevertheless,
the existence of these metastabilities still often leads to poor convergence behavior (as observed
for example in the standard single-site DMRG algorithm [201]).
The standard way to cure this issue is to formulate the diagonalization problem not for a
single tensor, but for several (usually two) adjacent tensors that previously have been contracted
together. After optimizing this compound tensor, the original tensor structure is restored
by means of SVDs. The accompanying truncation of the state spaces associated with the

75
SciPost Phys. Lect. Notes 8 (2019)

(a)

diag. (1)
T (1) Topt

Heff

(b)

T (1) Topt
(1)
diag.
SVD
T (2) Topt (2)
Topt

Heff

Figure 45: (a) Single-tensor update: the reduced diagonalization problem is formu-
lated for one tensor. The optimized tensor is given by the eigenvector of Heff with
the smallest eigenvalue. Note that link representations are preserved if Heff obeys
some symmetry. (b) Double-tensor update: the reduced diagonalization problem is
formulated for two tensors. The number of links of the optimized tensor is given
by the number of links on which Heff acts. Link representations of these links are
again preserved, in contrast to the one of the link which joins T (1) and T (2) : This
link results from an SVD and can differ from the original link, if the SVD performs
a non-trivial separation of the links of Topt . Note that if one of the two tensors is a
two-link tensor, this link separation is trivial and the link created by the SVD will be
the same as the external link of the two-link tensor.

compound’s internal links provides the variational handle lacking for external links, as illustrated
in Fig. 45. The downside of the multi-tensor optimization strategy is its higher computational
cost resulting from the increased number of links of the involved tensors.
In order to avoid the increase in computational costs, we can employ a so-called subspace
expansion at a single tensor T (1) , by temporarily expanding the state space of one selected
virtual link ηo of T (1) in order to allow for a variation of the symmetry representation on that
link. While the same ideas apply to both ordinary and symmetric TNs, we now describe the
technique explicitly for symmetric tensors since some extra steps are required for this case. As
a first step, an expansion of the selected link’s state space can be achieved in the following way:
Start by fusing all the links of T (1) other than ηo to one large link η1 . Repeat the operation
for tensor T (2) , which is connected to T (1) through ηo , to obtain another large link η2 . Now
determine the intersection ηmax := η1 ∩η2 (see Sec. 3.5.1). Note that the resulting ηmax has the
full link representation accessible to the double-tensor optimization scheme, i.e. it can capture
any link resulting from an SVD of the optimized double-tensor. Since our goal is to reduce the
computational cost with respect to the double-tensor optimization, we introduce a parameter
dpad to truncate all degeneracies ∂¯`ηmax = dpad (for all sectors `ηmax in ηmax ). The resulting
link ηmax
pad
is used to pad (see Sec. 3.5.1) the link ηo , according to η0o := (ηo + ηmaxpad
) ∩ ηmax
(the additional intersection makes sure that no redundant degeneracies are generated by the
padding operation). The parameter dpad now allows to smoothly interpolate between a single-

76
SciPost Phys. Lect. Notes 8 (2019)

tensor update and a double-tensor update. The latter limiting case can be achieved by choosing
dpad sufficiently large such that η0o = ηmax , while the former limiting case is obtained by setting
dpad = 0 which results in η0o = ηo . Note that in case the link ηo exceeds the minimal bond
dimension determined by the Schmidt rank (see Sec. 4.2.1), that is to say, in case there exist
some sectors ` for which ∂¯`η > ∂¯`ηmax , the procedure to obtain η0o may also cause some sectors
o
to have reduced degeneracies, effectively removing redundant indices from ηo .
The actual optimization within the single-tensor update with subspace expansion is carried
out in the following way (see Fig. 46): First the link ηo at the tensors T (1) and T (2) is replaced
by the expanded link η0o . The hereby newly created tensor elements in the two tensors are
filled (via subtensor assignment, see Sec. 3.5.1) with random entries (this step corresponds to
the subspace expansion). Then, the tensor T (1) is made the center node of the TN and it is
optimized as usual (by diagonalizing Heff ). Since the subspace expansion also affects T (2) , this
tensor has to be optimized as well (analogously to T (1) ), keeping the same expanded link η0o .
The two alternating optimizations of T (1) and T (2) are repeated until convergence is achieved
(usually one or two repetitions are enough). Finally, the original bond dimension D is restored
via an SVD, discarding the smallest singular values (see Fig. 46). We will show and discuss
numerical results obtained with this method in Sec. 5.3. A diagrammatic comparison of the
various optimization strategies is presented in Fig. 46.
Subspace expansion techniques in single- and double tensor updates can become an even
more powerful tool when combined with perturbative approaches, such as the procedure
introduced in Refs. [143, 201] for curing the shortcomings of the traditional single-center site
DMRG. These approaches typically differ from the single-tensor update with subspace expansion
presented above in the following three aspects: First, the padded elements are generated from
a contraction of the effective Hamiltonian (or a component thereof) into the center tensor, such
as to introduce first-order contributions to H|Ψ〉 (“enrichment step”). Second, the perturbation
coupling (in the simplest approach, a scalar factor) has to be driven to zero in a controlled
way to achieve convergence of the algorithm. Third, after some form of local bond truncation,
the perturbation introduced by the padded elements persists within the network until the
enriched node is revisited in a later sweep. This strategy enables the subspace expansion to
propagate through the network, making it potentially more powerful than the double-tensor
update discussed above without such perturbative approach [143].

5.1.4 Optimizer – Execution


A crucial step of the optimization routine (common to all update schemes) is the diagonalization
of the effective Hamiltonian Heff as this step usually constitutes the computational bottleneck of
the whole GS algorithm. Thanks to the central gauging we are merely dealing with a standard
eigenvalue problem (as opposed to a generalized eigenvalue problem arising in TNs with loops
in their geometry). While this already provides a computational advantage, the identification
of a fast eigensolver is still of utmost importance. Two observations are helpful in this context:
Heff is usually sparse2 , and only the ground-eigenstate of Heff has to be computed. Therefore,
direct diagonalization algorithms (such as tridiagonal reduction, implemented by e.g. LAPACK’s
zheev [136]) are not convenient since they compute the whole spectrum, for which they usually
require the explicit construction of the full matrix Heff . It is instead advantageous to employ
iterative algorithms based on power methods, e.g. Lanczos/Arnoldi iteration [157, 159]. These
algorithms exploit the sparsity of Heff , only require knowledge of the action Heff |T [c] 〉 on
the center node tensor |T [c] 〉 (see Fig. 37(c)) and, in addition, are specialized to return only
2
More precisely, the two conditions that Heff has to fulfill in order to be sparse are: (i) The dimensions of the
TPO links {µ} should be much smaller than the bond dimensions of the virtual links (Dµ  Dη ), and (ii) the total
number of interactions p should be small (see Eq. (73)). Many important Hamiltonian classes (e.g. short-range
interacting models) meet these requirements.

77
SciPost Phys. Lect. Notes 8 (2019)

single-tensor update (I) single-tensor update with double-tensor update


subspace expansion (II) (III)

(`, ∂ ) (`, ∂ ) (`, ∂ )


(a) T (1) T (2) T (1) T (2) T (1) T (2)

(`exp , ∂exp )

(b) (1)
Texp (2)
Texp T

(`exp , ∂exp )

(c) (1) (2)

swap (1)↔(2)
Teexp Teexp

×n
(`exp , ∂exp )
(d) (`, ∂ )
e (1) (2)
Te (1) T (2) Teexp
Teexp Te

(e) (˜`, ∂˜ ) (˜`, ∂˜ )


Te (1) (2)
Teexp Te (1) Te (2)

(`, ∂ ) (˜`, ∂˜ ) (˜`, ∂˜ )


(f) Te (1) T (2) Te (1) Te (2) Te (1) Te (2)

Figure 46: Comparison of different update schemes. (a) Initial configuration (identical
in all three schemes). Note that scheme I only operates on one tensor, which means
T (2) is left unchanged throughout this scheme. (b) Making the link representation of
the intermediate link variational: scheme I does not have this step; scheme II realizes
it via link expansion (see text), while scheme III achieves it via contraction with T (2) .
(c) Restore gauge and compute effective Hamiltonian for T (1) (only scheme II). (d)
Update center node with lowest eigenvector of Heff . Only scheme II: repeat (c) and
(d) until converged, the center node being alternately T (1) and T (2) . (e) Restoring
the original bond dimension of the intermediate link: scheme I does not have this
step; scheme II and III realize it via an SVD, discarding the smallest singular values.
(f) Final configuration. Note that in scheme II and III the link representation of the
intermediate link may have changed, while in scheme I it is constant.

78
SciPost Phys. Lect. Notes 8 (2019)

a few eigenpairs. The computational costs of the different update schemes are thus given
by the numerical effort of evaluating Heff |T [c] 〉: All schemes share a factor f = Napp P DTPO
3
,
where Napp is the number of operator applications in the iterative eigensolver (see Sec. 5.2
for an analysis of the behavior of Napp in a specific TN architecture) and P is the number of
terms in the effective Hamiltonian (including possible projectors), labeled by the index p in
Eqs. (73) and (76). Here we assumed for simplicity that each TPO tensor (see Eq. (68)) has
two virtual links {µ}, all of uniform dimension DTPO . The leading cost for the diagonalization
in the single-tensor update is then
!
Y
O f · max Dη · Dη , (81)
{η}1
{η}1

where {η}1 is the set of links of the tensor T (1) under optimization. For the single-tensor update
with subspace expansion we get a diagonalization cost of
 
 Y 
O n · f · max max Dη · Dη  , (82)
j∈{1,2} {η} j
{η} j

where n is the number of times that the tensors T (1) and T (2) are alternately optimized, and
the sets {η}1,2 contain one expanded link η0o with bond dimension Dη0o = Dηo + dpad . For the
double-tensor update the cost for the diagonalization scales like
!
Y
O f · max Dη · Dη . (83)
{η}1∪2 \ηo
{η}1∪2 \ηo

The SVD needed for the re-compression of the expanded link at the end of the single-tensor
update with subspace expansion has a cost of (see Sec. 2.2.2)
 
  Y 2  Y 
O  min Dη0o , Dη max Dη0o , Dη  , (84)
{η}1 \η0o {η}1 \η0o

while the final SVD in the double-tensor update has a computational cost scaling of
 
 Y 2 Y
O  min Dη max Dη  . (85)
j∈{1,2} j∈{1,2}
{η} j \ηo {η} j \ηo

In Sec. 5.2 we will see how the general computational cost estimations given here specialize
for specific TN architectures. This will also allow for an easier, more evident comparison of the
various costs for the different update schemes.

5.1.5 Low Excited States


The GS algorithm listed in Sec. 5.1.2 can be extended to determine low excited states, with
very little overhead in terms of both implementation effort and computational cost. The basic
idea for targeting a specific excited state is to penalize overlap with all lower eigenstates by
adding energy penalty terms to the Hamiltonian H. It should be stressed that this method
works iteratively, starting from the bottom end of the spectrum. Hence, the n-th excited state
|Ψn 〉 can only be targeted if all lower eigenstates |Ψ p 〉, p ∈ [0, n − 1] have been determined
previously.

79
SciPost Phys. Lect. Notes 8 (2019)

More in detail, while for obtaining the GS |Ψ0 〉 of the Hamiltonian H we solved the
minimization problem
!
E0 = 〈Ψ0 |H|Ψ0 〉 = min. , (86)
for targeting the n-th excited state we now need to solve the minimization problem
n−1
!
X
En = 〈Ψn |H|Ψn 〉 + ε p 〈Ψn |P p |Ψn 〉 = min. . (87)
p=0

Here, P p = |Ψ p 〉〈Ψ p | is the projector on the p-th eigenstate and ε p is an energy penalty which
has to be set large enough, i.e. at least as large as the energy difference |E p − En | to the
target state. Since the target state is not known at the beginning of the algorithm, this energy
difference is also unknown; in practice, one therefore simply estimates a value for ε p which
is guaranteed to be large enough (e.g. by setting ε p one order of magnitude larger than a
typical energy scale of the system). For solving Eq. (87), one runs again the algorithm listed in
Sec. 5.1.2, additionally taking into account the projective terms defined in Eq. (75), as detailed
in Sec. 5.1.1. Each projective term contributes exactly one additional interaction term p (see
Eq. (76)) to the effective Hamiltonian.

5.1.6 Variational Compression, Addition and Orthogonalization


An interesting feature of the presented variational ground state search algorithm is that we
can readily use it for compressing a generic loop-free tensor network state. This feature then
enables further operations such as the addition and orthogonalization of entire tensor networks.
For the case of MPSs, the procedure has been described in detail in [39].

Compression of a Tensor Network. Let a state |Ψ 0 〉 be given by some loop-free tensor


network of bond dimension D0 . We then seek the best approximation |Ψ〉 encoded in the same
network geometry, but with a smaller bond dimension D < D0 .
As long as the change in bond dimension affects only one virtual link, an efficient and
optimal strategy is to truncate the Schmidt values associated with that bond, as discussed in
Sec. 4.1. In all other cases, the optimal compressed state can be found by running the GS
algorithm of Sec. 5.1 with bond dimension D for the Hamiltonian H ≡ −|Ψ 0 〉〈Ψ 0 |, i.e. with a
single projector as described in Sec. 5.1.1. Ideally, the algorithm then converges to the lowest
2
energy E = − 〈Ψ|Ψ 0 〉 that can be achieved with bond dimension D, maximizing the quantum
fidelity between compressed and uncompressed TN.
In practice, depending on the severity of compression and the chosen initial state, this
approach may improve fidelity but fail to find the optimal solution, even if extended single-
or double-tensor update schemes are used as discussed in Sec. 5.1.3. In order to increase
robustness and efficiency, it has been suggested to run a sequence of local compressions first,
e.g. to successively truncate Schmidt values, bringing all bond dimensions of Ψ 0 down to D [39].
The accumulation of consecutive truncation errors may prevent that state from being optimal,
but it is usually sufficiently well-conditioned to serve as initial state for a few more iterations in
the GS algorithm.
For the sole purpose of compression, the GS algorithm no longer requires the sophisticated
eigensolvers from Sec. 5.1.4: At the level of tensors, projection onto the compressed state
results in effective Hamiltonians taking the form of outer products
[c] [c] [c]∗
” —
heff = −T̃ · T̃ (88)
{αη },{βη } {αη } {βη }
as written in full detail in Eqs. (76) and (77). Up to a normalization factor, the lowest eigenstate
solution for the center tensor is thus given by T [c] = T̃ [c]∗ , which is easily constructed.

80
SciPost Phys. Lect. Notes 8 (2019)

Addition of Tensor Networks. A particular need for compression arises in the addition of
two or more TNs |Ψ (i) 〉, i = 1, . . . , k. We require that these TNs share the same geometry,
including the outcome |Ψ〉, which shall be the closest approximation to the sum that can be
achieved within a fixed maximal bond dimension D. Once more, we can find |Ψ〉 by means of
the variational GS algorithm, which handles projections on superpositions P of TNs efficiently:
After setting up the projector |Ψ 0 〉〈Ψ 0 | defined by the sum of TNs |Ψ 0 〉 = i |Ψ (i) 〉 the algorithm
for compression can be run as outlined before.
Note that alternatively,
P (i) one could also first construct the exact sum as a network of bond
dimension D = i D as follows: Set each resulting tensor T 0[q] to zero and subsequently
0

inscribe the respective tensors T (i)[q] from TNs |Ψ (i) 〉, i = 1, . . . , k, as subtensors (Sec. 3.5.1) into
Pi−1 ( j) Pi ( j)
disjunct index ranges [1 + j=1 Dη , j=1 Dη ] for each virtual link η. The resulting network
has to be compressed to the final bond dimension D — a task that requires computational
resources scaling polynomially with the bond dimension D0 , analogous to computing TN scalar
products. Unless one resorts to specific data structures for the block-diagonal structure in the
contractions of the effective projectors, a drastic compression from dimension D0 is therefore
less efficient compared to the direct variational compression of a sum of multiple states with
smaller bond dimensions each.

Orthogonalization of Tensor Networks. As an example for the addition of TNs, one might
want to orthonormalize a TN state |Ψ2 〉 with respect to another, |Ψ1 〉. This can be achieved (up
to a normalization) by subtracting their common overlap as in |Ψ⊥ 〉 := |Ψ2 〉−|Ψ1 〉 〈Ψ1 |Ψ2 〉. After
evaluating the involved scalar product, we can employ the variational addition to obtain |Ψ⊥ 〉.
Note however that the compression involved in maintaining a fixed bond dimension D usually
prevents this approach from delivering an exact orthogonal solution. However, it returns the
highest possible fidelity to the exact solution |Ψ⊥ 〉 at the expense of strict orthogonality. Of
course, such a result can nevertheless be useful, and one can directly extend this procedure to
more than two states, e.g. in a (modified) Gram–Schmidt process.
Interestingly, the specific task of orthogonalization can be addressed more directly with
variational compression, similar to targeting excited states: For this purpose, we include in the
effective Hamiltonian of Eq. (88) the projectors onto all states that we wish to orthonormalize
against. These additional projectors have to be multiplied by a positive energy penalty " p . With
" p → ∞, orthonormality can be implemented as a hard constraint during compression, and
the optimizer’s eigenproblem takes the form of a simple orthogonalization against the effective
projectors. Conversely to the additive approach, we now expect full orthonormality at the
expense of fidelity to the exact solution.

5.2 A Practical Application: Binary Tree Tensor Networks


In this section, we discuss a specific loop-free tensor network and see it in action: the binary
tree tensor network (bTTN). This network is built on top of a one-dimensional lattice of physical
links, but unlike MPS this type of tensor network has the practical advantage of treating open-
boundary and periodic-boundary one-dimensional systems on equal footing [182]. It can thus
efficiently address periodic systems without the limitations typical of periodic-DMRG [196].
The bTTN is composed entirely of tensors with three links. As we will see, this guarantees
moderate scaling of computational cost with the bond dimension D. The bTTN architecture with
built-in Abelian symmetry handling that we use (as introduced in Sec. 3) is illustrated in Fig. 47.
The essential ingredients for achieving the ground state of a many-body Hamiltonian using
a loop-free TN ansatz have been outlined previously in Refs. [74, 181, 182] and in Sec. 5.1.2.
Here, we summarize some practical technical details that are important when implementing a
bTTN ground state algorithm:

81
SciPost Phys. Lect. Notes 8 (2019)

Figure 47: A binary TTN: The black dots at the bottom represent the physical sites of
local dimension d. Each link in the l-th layer (counting from below towards the top)
l
has dimension min(d 2 , D), where D is a constant that we will refer to as the bond
dimension. The dangling blue link at the top left tensor has dimension one and carries
the global symmetry sector of this symmetric TN state. Its position in the network
can be arbitrarily chosen. All tensors are gauged towards a center node (red).

• Choice of decomposition algorithm for central gauging: In order to transform the bTTN
into a central gauge, tensor decompositions of the type shown in Fig. 48(a) have to be
performed repeatedly (see Sec. 4.2.2). Both the QR decomposition and the SVD are suited
to carry out this task. However, as we show in Fig. 48(b), the QR decomposition can be
performed in roughly two thirds of the time of the SVD, while exhibiting the same scaling
O(D3.8 ) in the bond dimension (at least in the range D < 500 commonly accessible to
bTTN simulations). Therefore, it is advisable to implement any gauging routine in a
bTTN by means of QR decompositions, and employ SVDs solely for compression purposes.

• Benchmark of diagonalization methods: As mentioned in Sec. 5.1.4, the eigensolver is a


component of the GS algorithm which consumes a significant amount of runtime. The
performance of this routine therefore crucially influences the whole algorithm runtime
and should be first looked at when optimizing the algorithm speed. In Tab. 1 we list
three publicly available diagonalization methods, and compare their performances in
Fig. 49. Our benchmarks suggest that ARPACK [202] (an implementation of the Arnoldi
method [159]) is the most suitable eigensolver in the context of a bTTN ground state
algorithm. On a side note, we remark that the setting of a sensible initialization guess can
further reduce runtimes of iterative eigensolvers. For the eigenvalue problems occurring in
the GS algorithm presented in Sec. 5.1.2, an obvious choice for such an initialization guess
is the previous state of the optimization center tensor, before optimization. Especially
during the final sweeps, when the tensor elements are already close to their optimal
configuration, this measure can significantly reduce the number of necessary operator
applications Napp . Conversely, one could reduce Napp in the early sweeps of the algorithm
by selecting looser starting threshold values for convergence (not included in Fig. 49).

• Computational cost of a bTTN GS algorithm: We observe that the total number of tensors
Q scales linearly with the system size N , which is the same scaling as in an MPS:
(N )−1
log2X
2log2 N − 1
Q= 2l = − 1 = N − 2. (89)
l=1
2−1

When using a single-tensor update scheme one deals exclusively with three-link tensors,
while for a double-tensor update scheme intermediate four-link tensors occur. Hence,

82
SciPost Phys. Lect. Notes 8 (2019)

(a) D (b)
100

D 10

t (sec)
D
1
dec. SVD
= slope: 3.8
0.1 QR
slope: 3.8
D D D D 0.01
100 200 300 400 500
D

Figure 48: (a) Tensor decomposition arising when transforming a bTTN into cen-
tral gauge. (b) Corresponding computational performances for QR decomposition
(LAPACK routine zgeqrf ) and SVD (LAPACK routine zgesdd). The graph shows the
computation time, averaged over 50 runs, needed for the decomposition shown in (a)
(performed on a tensor with random entries without any symmetries) as a function of
the bond dimension D.

Table 1: List and characteristics of some numerical packages capable of finding


eigenstates of a hermitian operator Heff

Method Remarks

LAPACK (zheev) [203] Full diagonalization (delivers all eigenpairs of Heff ). Requires
the construction of Heff in full matrix form.
ARPACK [202] Implementation of the Arnoldi algorithm [157, 159]. Iterative
diagonalization, designed to deliver a few eigenpairs of Heff .
Requires only knowledge of the action Heff |v〉 on a given vec-
tor |v〉. The number of necessary operator applications Napp
strongly depends on the size of the Krylov subspace (parameter
nc v in ARPACK). A consequence of choosing nc v too small is a
fast growth of Napp with the problem size dim(|v〉) (where we
empirically find that “fast" usually means super-logarithmic, see
Fig. 49).
Jacobi–Davidson Implementation of the Davidson [158] algorithm. Requirements
(jdqz) [204] and features similar to Arnoldi. Here Napp crucially depends on
the maximum number of matrix-vector multiplications in the
linear equation solver (parameter mx mv in jdqz). We achieve
best performance when choosing mx mv ≈ 5, i.e. at the lower
end of the recommended interval [5 .. 100] (especially when
only one eigenpair is demanded).

83
SciPost Phys. Lect. Notes 8 (2019)

(a) D
D

Heff
= λ ·

D D

D D

(b)

method: (t ∝ Dβ ), β =
10000
zheev, n = al l 8.8
1000
Apack, n = 1 4.3
100 Arpack, n = 3 4.3
t (sec)

10 Arpack, n = 15 4.3
Davidson, n = 1 4.3
1
Davidson, n = 3 4.2
0.1 Davidson, n = 15 4.1
0.01
10 100
D

3000
method:
2500 Apack, n = 1
2000 Arpack, n = 3
Arpack, n = 15
Napp

1500 Davidson, n = 1
1000 Davidson, n = 3
Davidson, n = 15
500
0
10 100
D

Figure 49: (a) Eigenvalue problem arising in a bTTN single-tensor update scheme.
λ is an eigenvalue of the hermitian operator Heff . (b) Comparison of the performances
of the diagonalization algorithms listed in Tab. 1. Upper plot: Computation times
needed to solve the eigenvalue problem defined in (a), using an effective Hamiltonian
with random entries, as a function of the bond dimension D and for different numbers
of calculated eigenpairs n. Computation times are averaged over 50 runs, with the
error bars indicating the standard deviation. Lower plot: Corresponding number of
operator applications Napp . Note how Napp only increases weakly (not faster than
logarithmically) with the problem size (given that the algorithm parameters are
properly configured).

84
SciPost Phys. Lect. Notes 8 (2019)

(a) (b) D
D

D
D D
D
D D
D

Figure 50: Examples for computationally most expensive contractions in a bTTN


GS algorithm (as occurring during the application of the effective Hamiltonian,
cf. Fig. 37(c)) in the case of (a) a single-tensor update scheme (contraction cost:
O(D4 )), and (b) a double-tensor update scheme (contraction cost: O(D5 )).

we face memory scalings of O(N D3 ) and O(N D4 ), respectively. As for the scaling of
the computation time, the most expensive contractions exhibit a cost of O(D4 ) (single-
tensor update) and O(D5 ) (double-tensor update), see Fig. 50 for two corresponding
example contractions. The SVD in the double-tensor update has a cost bounded by
O(D6 ) (see Sec. 5.1.4). Therefore, the algorithm runtime scalings are O(N D4 ) and
O(N D6 ), respectively3 . We stress that for a bTTN all reported scalings are independent
of the chosen boundary conditions, and thus apply both for open and periodic boundary
conditions.

• Evaluation of local and two-point observables: The expectation value of a local observable
O[s] , acting on some site s of a bTTN state |Ψ〉, can be calculated by contracting just three
tensors (independent of the system size N ), if the bTTN is center-gauged towards the
tensor q attached to site s. This is shown in Fig. 51(a), where we illustrate the expectation
value
[s] [q]∗
X
〈Ψ|O[s] |Ψ〉 = Tα[q],α ,α Oα ,β Tβ ,α ,α . (90)
s (s+1) ζ s s s (s+1) ζ
αs ,βs ,α(s+1) ,αζ

Two-point correlation observables


[s] e 0 [q]∗
0
X
〈Ψ|O[s] O0[s ] |Ψ〉 = Tα[q] Oα ,β Oαζ ,βζ Tβ ,α , (91)
s ,α(s+1) ,αζ s s s (s+1) ,βζ
αs ,βs ,α(s+1) ,αζ ,βζ

0
where Oe 0 is the result of a renormalization of the observable O0[s ] (obtained from con-
tractions involving all tensors along the path between sites s and s0 ), can be evaluated
with a very small number of contractions O(log2 N ), owing to the enhanced reachability
of two arbitrary sites s and s0 due to the hierarchical tree structure (see Fig. 51(b)). This
feature not only makes the calculation of observables a computationally cheap task in
a bTTN, but also lies at the heart of their predisposition to accurately describe critical
systems with algebraically decaying correlations [182].

5.3 Binary Tree Tensor Networks – Examples


In this subsection we present bTTN ground state solutions for two different 1D many-body
Hamiltonians, both equipped with Abelian symmetries. Namely, we will consider the quantum
Ising model in a transverse field (as a typical benchmark model with Z2 parity symmetry and an
easy, well-known analytic solution [8]), and a Bose–Hubbard ring with a rotating barrier [205]
3
Note that the SVD in the single-tensor update with subspace expansion has a cost of O(D4 ) (provided that dpad
is sufficiently small) and therefore does not increase the overall runtime scaling of the algorithm.

85
SciPost Phys. Lect. Notes 8 (2019)

(a)

q q

〈Ψ|O[s] |Ψ〉 = O[s] = O[s]

q q

(b)

q q
0
〈Ψ|O[s] O0[s ] |Ψ〉 =
0
O[s] O0[s ] = O[s] e0
O

q q

Figure 51: Evaluation of (a) a local and (b) a two-point observable expectation
value or a binary TTN state |Ψ〉. The bra TN 〈Ψ| is drawn as the vertically flipped
version of the ket TN |Ψ〉, indicating hermitian conjugation of all its tensors. If the
bTTN is properly gauged, the number of necessary contractions (involved tensors are
highlighted) can be reduced to O(1) and O(log2 N ), respectively.

(with U(1) boson number symmetry) as a more complex example. In particular, we provide
a field test of the different update schemes presented in Sec. 5.1.3 by comparing both their
convergence behaviors and computational resource usage.

Spin-1/2 Ising Model. We use the following Hamiltonian


N
X N
X
x [s] x [s+1]
H=− σ σ +λ σz [s] , (92)
s=1 s=1

where σ x,z [s] are Pauli matrices acting on site s. We employ periodic boundary conditions
(N + 1 ≡ 1) and fix the transverse field strength |λ| = 1, so that the model becomes critical in
the thermodynamic limit. At the critical point, the ground-state energy has a simple closed-form
solution [8, 206]
2 h  π i−1
E(N ) = − sin . (93)
N 2N

86
SciPost Phys. Lect. Notes 8 (2019)

(a) N = 1024:
0 0
N = 128:
0 −2 −2 0

−2 −4 −4 −2
−6 −6
log10 (∆E/N )

log10 (∆E/N )
−4 −4
D= 0 4 8 8 4 0
30
−6 40 −6
50
−8 60 −8
70
80
−10 −10
0 1 2 3 4 5 6 6 5 4 3 2 1 0
number of sweeps number of sweeps

(b) N = 128:
10 without sym.
with sym.
(t ∝ Dβ ), β =
t (h)

1 4.1
4.0
×4
0.1

50 100 150
D

Figure 52: (a) Convergence to the ground state energy of the 1D critical Ising model in
a bTTN using a single-tensor update without any inbuilt symmetries (left), compared
to a single-tensor update with inbuilt Z2 symmetry and fixed ∂¯e = ∂¯o = D/2 (right).
Plotted is the error of the ground state energy per site ∆E/N as a function of the
number of optimization sweeps for different bond dimensions D and different system
sizes N . (b) Runtime comparison of the two algorithms from (a), demonstrating the
faster performance of the algorithm with symmetries.

The Hamiltonian defined in Eq. (92) has a global pointwise parity (i.e. group Z2 ) symmetry,
dividing the state space into two sectors, one with even and one with odd parity, corresponding
to the two representations of Z2 (for details, see Sec. 3.1). The parity is therefore a good
quantum number. Specifically, the Hamiltonian commutes with the symmetry group elements
U(g) ∈ {1⊗N , s σz [s] }.
N

For |λ| < 1 (ferromagnetic phase) the ground state of H exhibits symmetry breaking, i.e.
in the thermodynamic limit there are two degenerate ground states (one in the even sector
and one in the odd sector), while for |λ| > 1 (paramagnetic phase) the thermodynamic ground
state is unique and even. For finite system sizes with an even number of sites, the lowest energy
state is always in the even sector (i.e. the trivial irrep of Z2 ); therefore we will target this
global sector in our bTTN simulations. Since there are only two possible symmetry sectors
` ∈ {e = 0, o = 1}, any link representation is defined by two degeneracy subspace dimensions ∂¯e

87
SciPost Phys. Lect. Notes 8 (2019)

double-tens.: single-tens. with subsp. exp.:


0 0
1 1
log10 (∆E/N )

log10 (∆E/N )
〈∂ e / ∂ o 〉links
−2 −2

−4 0.5 0.5 −4

−6 −6
0 0
−8 −8
0 1 2 3 4 5 6 7 7 6 5 4 3 2 1 0
number of sweeps number of sweeps

Figure 53: Comparison of link representation convergence for the double-tensor


update (left) and the single-tensor update with subspace expansion (right). Plotted
is the error of the ground state energy per site ∆E/N (blue) and the over all virtual
links averaged ratio of degeneracy dimensions 〈∂¯e /∂¯o 〉links (red) as a function of the
sweep number. For both plots a system size of N = 128 and a bond dimension of
D = 40 have been used.

and ∂¯o . In this sense, it is easy to perform a convergence analysis for virtual link representations
during the algorithm runtime. While the ability of a bTTN to accurately describe the ground
state of the critical 1D Ising model has been discussed in detail in Ref. [182], here we focus
on a comparison between the different algorithm types (with/without inbuilt Z2 symmetry
handling, single-tensor simple update/single-tensor with subspace expansion update/double-
tensor update). In Fig. 52(a) we show the convergence to the ground state energy as a function
of the number of optimization sweeps. Since the size of the search space is reduced when
exploiting symmetries, the convergence behavior towards the ground state is improved in this
case. This is most clearly seen for large system sizes, e.g. N = 1024. In Fig. 53 we compare
the two update schemes: Single-tensor update with subspace expansion and double-tensor
update, as described in Sec. 5.1.3. We benchmark their capability to converge to the optimal
link representations, which are characterized by ∂¯e = ∂¯o = D/2 for all links. For the sake of
this benchmark, we initialize all virtual links with strongly unbalanced degeneracy dimensions
∂¯e  ∂¯o and check whether the ratio ∂¯e /∂¯o converges to one as a function of the number
of sweeps. Although the double-tensor update converges with slightly fewer sweeps, it is
evident from Fig. 53 that both update schemes eventually manage to reach the optimal link
representations.
A careful consideration of the theoretically possible runtime reduction resulting from
the presence of two symmetry sectors with a flat distribution of degeneracy dimensions (see
Sec. 3.5.3) yields a factor of 1/4 for the specific case of a bTTN single-tensor update ground state
algorithm. Moreover, the memory consumption is expected to be halved in this case (because
half the sectors of each tensor do not couple and therefore do not need to be stored). In practice,
we obtained resource savings very close to these predicted values (see Fig. 52(b)), demonstrating
that symmetric tensor operations can be implemented with very little bookkeeping overhead.

Bose–Hubbard Ring With a Rotating Barrier. A graphical illustration of the system is shown
in Fig. 54. In the rotating frame the Hamiltonian can be written as [205]
N N
X  UX
e−2πi Ω/N b† [s] b[s+1] + h.c. + n[s] n[s] − 1 + β n[s b ] ,

H = −t (94)
s=1
2 s=1

88
SciPost Phys. Lect. Notes 8 (2019)


~
B

rotating
t β frame t β
U U

v
I(Ω)

Figure 54: Sketch of a Bose–Hubbard ring with a barrier that rotates with a constant
velocity v. In the rotating frame the barrier is at rest, but the bosons experience an
~ giving rise to a flux Ω through the ring, which leads to a boson
artificial gauge field B
current I(Ω) ∝ ∂ E/∂ Ω.

where the barrier of height β is now at a fixed lattice site s b and the (artificial) gauge flux Ω
is proportional to the barrier’s rotation velocity, t is the hopping amplitude and U the on-site
repulsion, while b† [s] , b[s] , and n[s] are bosonic creation, annihilation, and number operators,
respectively. The Hamiltonian defined in Eq. (94) has a total boson number conservation symme-
try, which is a global pointwise symmetry with group U(1) (see P Sec. 3.1 for details). Specifically,
H commutes with the total boson number operator ntot = s n[s] , and is thus invariant under
[s]
V (ϕ) = e iϕ ntot = s e iϕ n . Hence, the total boson number Nb (with ntot |Ψ〉 = Nb |Ψ〉) is a
N

good quantum number which we can enforce exactly by targeting a specific boson filling in
a bTTN algorithm with inbuilt U(1) symmetry handling. Since the U(1) group entails many
irreps — potentially all quantum numbers ` = 0, . . . , Nb could be present on some links —
there is no obvious way a priori to assign the sector degeneracies ∂¯` (` = 0, 1, . . .) at each
virtual link. Therefore, we necessarily need to select the ∂¯` via a variational procedure, as
explained in Sec. 5.1.3. We compare again the performances of the two different algorithm
types: Double-tensor update strategy against single-tensor update with subspace expansion
strategy. The results are summarized in Fig. 55, leading to the same conclusions already
obtained from the benchmarks on the Ising model: Both algorithms manage to converge to
the optimal link representations, with the double-tensor update requiring a slightly smaller
number of sweeps. Nevertheless, the overall computational cost of the double-tensor update
is significantly higher (both in terms of runtime and memory consumption), as expected (see
Sec. 5.2). We verified that the obtained link representations are indeed optimal by comparing
the converged ground state energies to the ones resulting from a bTTN simulation of the same
model without inbuilt symmetries. (Without symmetries we have no longer a way of fixing Nb
exactly. Nevertheless a desired global filling can be achieved
P [s] on average by working in a grand
canonical ensemble, i.e. by adding an extra term µ s n to the Hamiltonian and tuning the
chemical potential µ.)
A numerical comparison of the runtime behaviors of the different algorithm types is shown
in Fig. 55(c). It appears that the double-tensor update algorithm with inbuilt U(1) performs
roughly ∼ 4 times faster than the single-tensor update algorithm without symmetries, through-
out the numerically accessible regime of D (albeit exhibiting a slightly steeper power-scaling).
Moreover, the single tensor update with subspace expansion reduces the runtime by another

89
SciPost Phys. Lect. Notes 8 (2019)

(a) 0 0
D=
−1 −1
40 120
−2 60 140 −2
log10 (∆E/N )

log10 (∆E/N )
−3 80 160 −3
−4 100 180 −4
−5 −5
−6 −6
−7 −7
−8 −8
0 2 4 6 8 10 12 10 8 6 4 2 0
sweep number sweep number
(b)
D = 60:

16 sweep: 16 sweep:
0 0
12 1 12 1
2 2
3 3
∂`

∂`

8 4 8 4
5
4 4

0 0
2 4 6 8 10 12 14 16 2 4 6 8 10 12 14 16
` `
(c)
1000
single-tens. (no sym.)
100 double-tens.
single-tens. with subsp. exp.
10 single-tens. (fixed link reps.)
t (h)

×15 (t ∝ Dβ ), β =
×4
1 4.2
4.5
0.1 3.4
3.6
0.01
50 100 200 400
D

Figure 55: Benchmark of different bTTN algorithm types for a Bose–Hubbard ring
with a rotating barrier (N = 128, Nb = 19, β/t = 1, U/t = 10, Ω = 0, local bosonic
state spaces truncated to d = 5). (a) Comparison of ground state energy convergence
between double-tensor update (left) and single-tensor update with subspace expansion
(right). As reference for the calculation of the error ∆E the energy from a simulation
with bond dimension D = 400 has been used. (b) Degeneracy dimensions convergence
at the top link of the bTTN, shown for the same two algorithms as in (a). The solid
lines are fitted Gaussians, the gray dashed line in the right plot is the final configuration
from the left, for ease of comparison. (c) Algorithm runtimes as a function of the bond
dimension. The cyan line corresponds to a single-tensor update with fixed degeneracy
dimensions ∂¯` (equipped with the converged ∂¯` obtained from the single-tensor update
with subspace expansion) and demonstrates that the subspace expansion (purple line)
adds a comparatively small overhead.

90
SciPost Phys. Lect. Notes 8 (2019)

considerable factor (∼ 3 times faster for D = 50), on top of having a favorable scaling in
D, demonstrating the huge potential speed-up accessible by performing TN algorithms with
symmetric tensors, especially when the underlying symmetry exhibits a large number of sectors.

6 Conclusions
In this anthology we reviewed some cutting-edge computational methods, based on tensor
network states, for simulating low-dimensional many-body quantum systems at finite size,
approached from the numerical user’s perspective. After introducing the concepts and abstract
objects which are the building blocks for every tensor network algorithm, we described in
detail how these objects are meant to be embedded in a computer. We introduced which
manipulations, including (multi-)linear algebra routines, are useful for simulation purposes,
stressing the interfaces we suggest for the computation in order to be efficient in terms of
time and memory usage. We showed how the same data structures and routines have to
be adapted in order to encode pointwise Abelian symmetries, stressing how the conserved
quantities help the computation by sensitively reducing the effective variational spaces involved.
We reviewed in detail the geometrical and algebraic properties of loop-free tensor networks,
focusing on the advantages presented by various gauges and canonical forms. Then, we gave a
full description of an algorithm for capturing the properties of ground states and low excited
states of Hamiltonians based on loop-free tensor networks: We presented several optimization
strategies, and discussed their respective advantages and scaling, while showing practical
numerical comparisons on their precision and performance. Altogether, these components
form a useful and versatile toolkit which can deal with many low-entanglement scenarios for
one-dimensional quantum systems, encompassing the most common simulations in finite-size
realizations.

6.1 Outlook and Perspectives


Tensor network methods are still far from having reached their ultimate potential. Despite
intrinsic difficulties, several steps have been made in the last decade in the direction of simulating
two-dimensional quantum systems: 2D quantum physics is inherently fascinating for the non-
trivial topological properties of the phases of matter, and consequent exotic quasi-particle
statistics (anyons), equilibrium properties still ruled by quantum fluctuations and entanglement
(far from mean field description), but it is more computationally challenging. In tensor network
language, the challenge translates into the fact that it is difficult to design a tensor network
geometry which is both efficiently contractible and capable of capturing the 2D area-law
of entanglement [61, 128]. Devising powerful, versatile TN methods able to deal with this
scenario is without doubt one of the most relevant challenges of the present decade [62, 195,
207, 208]. Regardless, we stress that recently the same loop-free tensor network architectures
that we focused on in this work were employed to study topological properties of interacting
two-dimensional models (with toric and open boundary conditions), and showed simulation
capabilities well-beyond exact diagonalization [209]. Moreover, loop-free TNs seem to be a
promising platform also for out-of-equilibrium simulations, a task likely achievable by adapting
the time-dependent variational principle for matrix product states [51, 210, 211], or similar
strategies [119, 122], to the more flexible loop-free network geometries, possibly to include
finite temperatures [57, 120, 121] and open-system dynamics [212].

91
SciPost Phys. Lect. Notes 8 (2019)

Acknowledgments
The authors gratefully acknowledge support from the EU via the projects SIQS, RYSQ and
UQUAM, the German Research Foundation (DFG) via TWITTER, the SFB/TRR21 and OSCAR,
the Baden-Württemberg Stiftung via the Eliteprogramm for Postdocs, the Baden-Württemberg
Ministry of Research via an LGFG scholarship, the Carl-Zeiss-Stiftung via the Nachwuchs-
förderprogramm, and the Studienstiftung des deutschen Volkes via Promotionsförderung. SM
gratefully acknowledges the DFG support via a Heisenberg fellowship. Computational resources
granted by the bwUniCluster project and by the MOGON cluster of the JGU (made available by
the CSM and AHRP) are acknowledged.

Appendix

A Anti-Unitary Symmetries and Time-Reversal


In this section we report a short note on the usage of anti-unitary symmetries for tensor network
architectures, and briefly discuss related challenges and implications. To this end we expand the
formalism of symmetric tensor networks to include anti-unitary representations of symmetry
groups. We remark that anti-unitary symmetries are extremely relevant from a physical point
of view since they include, for instance, time-reversal symmetry. As we review here, this
characteristic allows in some cases to reduce the Hamiltonian problem from complex to real
linear algebra, thus generally providing a computational simplification. Here, we will follow a
formulation analogous to Sec. 3.

Anti-unitary Symmetries. A complementary class of invariance [T, H] = 0 entails the anti-


unitary symmetries [213] of the form T = U K where U is unitary, and K shall denote complex
conjugation of expansion coefficients in a canonical product-basis |S〉 of the lattice. Such sym-
metry arises for instance as a consequence of conventional time-reversal invariance, implying
T 2 = ±1. In the case T 2 = +1 (as observed for integer spin or in certain spin-models where
spin-degrees of freedom are not explicitly present, e.g. by means of exchange-interactions), H
becomes real in the sense that it takes the form of a real matrix when written in an invariant
basis T |R〉 = |R〉. The latter is however easily constructed starting from superpositions in the
form |S〉 + T |S〉 without having to diagonalize H explicitly. Specifically we consider
∗ ∗
〈R|HR0 〉 = 〈T R|T HR0 〉 = 〈R|HR0 〉 (95)

by means of anti-unitarity, T -invariance of |R〉 and |R0 〉, and commutation with the Hamilto-
nian [214].
In the context of lattice models suitable for TNs, our interest lies on cases where we can
transform into the basis |R〉 with a local unitary transformation, i.e. one that acts separately on
the physical sites. The simplest example is a Hamiltonian that is already real in |S〉 to begin
with, e.g. that is invariant under T = K. As a consequence, H becomes real symmetric and all of
its eigenstates have real expansions, too. Then, in TN algorithms, by means of a (generalized)
gauge transformation, acting locally on links (see Sec. 4.2.1), we can encode certain cases of
anti-unitary symmetry by simply restricting tensors to real elements. When realized, we call
this the “real gauge” of a TN (see Sec. 4.2.5).

B Graphical Notation Legend


Fig. 56 contains the graphical notation legend, reporting the pictorial symbols that we use
throughout the manuscript to identify the various TN components (links, tensors, etc.).

92
SciPost Phys. Lect. Notes 8 (2019)

Links: symmetric

i (`, ∂ ) `

`select
Selector Link:

Tensors:

T
T ≡ R δ̃

W (g)
etc. Link Representation:

Fuse-Nodes: Link-Weights:

F λi λ`∂

Special Tensors:

T0
Isometry: Q Padded:

C
T [c]
Gauge Centers:

Figure 56: Graphical notation legend. Pictorial symbols for links, tensors, their
symmetric counterparts, and special properties (such as isometry) are listed.

93
SciPost Phys. Lect. Notes 8 (2019)

References
[1] D. J. Thouless, The quantum mechanics of many-body systems, Courier Dover Publications
(2013).

[2] X. G. Wen, Topological orders in rigid states, Int. J. Mod. Phys. B 04, 239 (1990),
doi:10.1142/S0217979290000139.

[3] J. P. Dowling and G. J. Milburn, Quantum technology: the second quantum revolu-
tion, Philos. Trans. Royal Soc. Lond., Ser. A: Math. Phys. Eng. Sci. 361, 1655 (2003),
doi:10.1098/rsta.2003.1227.

[4] M. A. Nielsen and I. L. Chuang, Quantum computation and quantum information, Cam-
bridge University Press (2009), doi:10.1017/cbo9780511976667.

[5] K. G. Wilson, Confinement of quarks, Phys. Rev. D 10, 2445 (1974),


doi:10.1103/PhysRevD.10.2445.

[6] J. B. Kogut, An introduction to lattice gauge theory and spin systems, Rev. Mod. Phys. 51,
659 (1979), doi:10.1103/revmodphys.51.659.

[7] M. Dalmonte and S. Montangero, Lattice gauge theory simulations in the quantum infor-
mation era, Contemp. Phys. 57, 388 (2016), doi:10.1080/00107514.2016.1151199.

[8] B. Sutherland, Beautiful Models, World Scientific (2004), doi:10.1142/5552.

[9] M. Jarrell and D. Scalapino, Numerical methods for lattice quantum many-body problems,
Addison-Wesley (1997).

[10] J. M. Zhang and R. X. Dong, Exact diagonalization: the Bose–Hubbard model as an


example, Eur. J. Phys. 31, 591 (2010), doi:10.1088/0143-0807/31/3/016.

[11] L. P. Kadanoff, More is the same; Phase transitions and mean field theories, J. Stat. Phys.
137, 777 (2009), doi:10.1007/s10955-009-9814-1.

[12] C. Froese Fischer, General Hartree-Fock program, Comput. Phys. Commun. 43, 355
(1987), doi:10.1016/0010-4655(87)90053-1.

[13] V. L. Ginzburg, Certaintheoretical aspects of radiation due to superlumial motion in


a medium, Fiz. tverdogo tela 2, 2031 [Sov. Phys. Solid State 2, 1824] (1960),
doi:10.1070/PU1960v002n06ABEH003185.

[14] N. D. Mermin and H. Wagner, Absence of ferromagnetism or antiferromagnetism in


one- or two-dimensional isotropic Heisenberg models, Phys. Rev. Lett. 17, 1133 (1966),
doi:10.1103/PhysRevLett.17.1133.

[15] M. H. Kalos and P. A. Whitlock, Monte Carlo methods, John Wiley & Sons (2008),
doi:10.1002/9783527626212.

[16] W. M. C. Foulkes, L. Mitas, R. J. Needs and G. Rajagopal, Quantum Monte Carlo simula-
tions of solids, Rev. Mod. Phys. 73, 33 (2001), doi:10.1103/revmodphys.73.33.

[17] E. Y. Loh, J. E. Gubernatis, R. T. Scalettar, S. R. White, D. J. Scalapino and R. L. Sugar,


Sign problem in the numerical simulation of many-electron systems, Phys. Rev. B 41, 9301
(1990), doi:10.1103/PhysRevB.41.9301.

94
SciPost Phys. Lect. Notes 8 (2019)

[18] M. Schiró and M. Fabrizio, Real-time diagrammatic Monte Carlo for nonequilibrium
quantum transport, Phys. Rev. B 79, 153302 (2009), doi:10.1103/PhysRevB.79.153302.

[19] R. Orús, A practical introduction to tensor networks: Matrix product states and projected
entangled pair states, Ann. Phys. 349, 117 (2014), doi:10.1016/j.aop.2014.06.013.

[20] S. Sachdev, Tensor networks—a new tool for old problems, Physics 2, 90 (2009),
doi:10.1103/Physics.2.90.

[21] J. Eisert, Entanglement and tensor network states (2013), arXiv:1308.3318.

[22] R. Orús, Advances on tensor network theory: symmetries, fermions, entanglement, and
holography, Eur. Phys. J. B 87, 280 (2014), doi:10.1140/epjb/e2014-50502-9.

[23] F. Verstraete, V. Murg and J. I. Cirac, Matrix product states, projected entangled pair states,
and variational renormalization group methods for quantum spin systems, Adv. Phys. 57,
143 (2008), doi:10.1080/14789940801912366.

[24] S.-J. Ran, E. Tirrito, C. Peng, X. Chen, G. Su and M. Lewenstein, Review of tensor network
contraction approaches (2017), arXiv:1708.09213.

[25] S. Östlund and S. Rommer, Thermodynamic limit of density matrix renormalization, Phys.
Rev. Lett. 75, 3537 (1995), doi:10.1103/PhysRevLett.75.3537.

[26] G. Vidal, Efficient classical simulation of slightly entangled quantum computations, Phys.
Rev. Lett. 91, 147902 (2003), doi:10.1103/PhysRevLett.91.147902.

[27] D. Pérez García, F. Verstraete, M. M. Wolf and J. I. Cirac, Matrix product state representa-
tions, Quantum Inf. Comput. 7, 401 (2007), [arXiv:quant-ph/0608197].

[28] M. Srednicki, Entropy and area, Phys. Rev. Lett. 71, 666 (1993),
doi:10.1103/PhysRevLett.71.666.

[29] K. Audenaert, J. Eisert, M. B. Plenio and R. F. Werner, Entanglement properties of the


harmonic chain, Phys. Rev. A 66, 042327 (2002), doi:10.1103/PhysRevA.66.042327.

[30] G. Vidal, J. I. Latorre, E. Rico and A. Kitaev, Entanglement in quantum critical phenomena,
Phys. Rev. Lett. 90, 227902 (2003), doi:10.1103/physrevlett.90.227902.

[31] Ö. Legeza and J. Sólyom, Quantum data compression, quantum information generation,
and the density-matrix renormalization-group method, Phys. Rev. B 70, 205118 (2004),
doi:10.1103/PhysRevB.70.205118.

[32] M. B. Plenio, J. Eisert, J. Dreißig and M. Cramer, Entropy, entanglement, and area:
Analytical results for harmonic lattice systems, Phys. Rev. Lett. 94, 060503 (2005),
doi:10.1103/PhysRevLett.94.060503.

[33] M. M. Wolf, Violation of the entropic area law for fermions, Phys. Rev. Lett. 96, 010404
(2006), doi:10.1103/PhysRevLett.96.010404.

[34] M. B. Hastings and T. Koma, Spectral gap and exponential decay of correlations, Commun.
Math. Phys. 265, 781 (2006), doi:10.1007/s00220-006-0030-4.

[35] B. Bauer and C. Nayak, Area laws in a many-body localized state and its impli-
cations for topological order, J. Stat. Mech. P09005 (2013), doi:10.1088/1742-
5468/2013/09/p09005.

95
SciPost Phys. Lect. Notes 8 (2019)

[36] H. Bernigau, M. J. Kastoryano and J. Eisert, Mutual information area laws for thermal
free fermions, J. Stat. Mech. P02008 (2015), doi:10.1088/1742-5468/2015/02/p02008.

[37] J. Eisert, M. Cramer and M. B. Plenio, Colloquium: Area laws for the entanglement entropy,
Rev. Mod. Phys. 82, 277 (2010), doi:10.1103/RevModPhys.82.277.

[38] T. Huckle, K. Waldherr and T. Schulte-Herbrüggen, Computations in quantum tensor


networks, Linear Algebra Appl. 438, 750 (2013), doi:10.1016/j.laa.2011.12.019.

[39] U. Schollwöck, The density-matrix renormalization group in the age of matrix product
states, Ann. Phys. 326, 96 (2011), doi:10.1016/j.aop.2010.09.012.

[40] S. R. White, Density matrix formulation for quantum renormalization groups, Phys. Rev.
Lett. 69, 2863 (1992), doi:10.1103/PhysRevLett.69.2863.

[41] S. R. White, Density-matrix algorithms for quantum renormalization groups, Phys. Rev. B
48, 10345 (1993), doi:10.1103/PhysRevB.48.10345.

[42] K. G. Wilson, Renormalization group and critical phenomena. I. Renormalization group and
the kadanoff scaling picture, Phys. Rev. B 4, 3174 (1971), doi:10.1103/PhysRevB.4.3174.

[43] K. G. Wilson, The renormalization group: Critical phenomena and the Kondo problem, Rev.
Mod. Phys. 47, 773 (1975), doi:10.1103/RevModPhys.47.773.

[44] U. Schollwöck, The density-matrix renormalization group, Rev. Mod. Phys. 77, 259 (2005),
doi:10.1103/RevModPhys.77.259.

[45] I. Affleck, T. Kennedy, E. H. Lieb and H. Tasaki, Valence bond ground states
in isotropic quantum antiferromagnets, Commun. Math. Phys. 115, 477 (1988),
doi:10.1007/BF01218021.

[46] M. Fannes, B. Nachtergaele and R. F. Werner, Finitely correlated states on quantum spin
chains, Commun. Math. Phys. 144, 443 (1992), doi:10.1007/bf02099178.

[47] F. Verstraete, D. Porras and J. I. Cirac, Density matrix renormalization group and periodic
boundary conditions: A quantum information perspective, Phys. Rev. Lett. 93, 227205
(2004), doi:10.1103/PhysRevLett.93.227205.

[48] G. Vidal, Efficient simulation of one-dimensional quantum many-body systems, Phys. Rev.
Lett. 93, 040502 (2004), doi:10.1103/physrevlett.93.040502.

[49] S. R. White and A. E. Feiguin, Real-time evolution using the density matrix renormalization
group, Phys. Rev. Lett. 93, 076401 (2004), doi:10.1103/physrevlett.93.076401.

[50] A. J. Daley, C. Kollath, U. Schollwöck and G. Vidal, Time-dependent density-matrix


renormalization-group using adaptive effective Hilbert spaces, J. Stat. Mech.: Theor. Exp.
P04005 (2004), doi:10.1088/1742-5468/2004/04/P04005.

[51] J. Haegeman, J. Ignacio Cirac, T. J. Osborne, I. Pižorn, H. Verschelde and F. Verstraete,


Time-dependent variational principle for quantum lattices, Phys. Rev. Lett. 107, 070601
(2011), doi:10.1103/PhysRevLett.107.070601.

[52] G. Vidal, Classical simulation of infinite-size quantum lattice systems in one spatial dimen-
sion, Phys. Rev. Lett. 98, 070201 (2007), doi:10.1103/PhysRevLett.98.070201.

[53] I. P. McCulloch, Infinite size density matrix renormalization group, revisited (2008),
arXiv:0804.2509.

96
SciPost Phys. Lect. Notes 8 (2019)

[54] J. Haegeman, S. Michalakis, B. Nachtergaele, T. J. Osborne, N. Schuch and F. Verstraete,


Elementary excitations in gapped quantum spin systems, Phys. Rev. Lett. 111, 080401
(2013), doi:10.1103/PhysRevLett.111.080401.

[55] R. Orús and G. Vidal, Infinite time-evolving block decimation algorithm beyond unitary
evolution, Phys. Rev. B 78, 155117 (2008), doi:10.1103/PhysRevB.78.155117.

[56] G. M. Crosswhite, A. C. Doherty and G. Vidal, Applying matrix product operators


to model systems with long-range interactions, Phys. Rev. B 78, 035116 (2008),
doi:10.1103/PhysRevB.78.035116.

[57] F. Verstraete, J. J. García-Ripoll and J. I. Cirac, Matrix product density operators: Simu-
lation of finite-temperature and dissipative systems, Phys. Rev. Lett. 93, 207204 (2004),
doi:10.1103/PhysRevLett.93.207204.

[58] M. Zwolak and G. Vidal, Mixed-state dynamics in one-dimensional quantum lattice systems:
A time-dependent superoperator renormalization algorithm, Phys. Rev. Lett. 93, 207205
(2004), doi:10.1103/physrevlett.93.207205.

[59] T. J. Osborne, A renormalisation-group algorithm for eigenvalue density functions of


interacting quantum systems (2006), arXiv:cond-mat/0605194.

[60] F. Schrodi, P. Silvi, F. Tschirsich, R. Fazio and S. Montangero, Density of states of


many-body quantum systems from tensor networks, Phys. Rev. B 96, 094303 (2017),
doi:10.1103/PhysRevB.96.094303.

[61] F. Verstraete, M. M. Wolf, D. Perez-Garcia and J. I. Cirac, Criticality, the area law, and
the computational power of projected entangled pair states, Phys. Rev. Lett. 96, 220601
(2006), doi:10.1103/PhysRevLett.96.220601.

[62] N. Schuch, M. M. Wolf, F. Verstraete and J. Ignacio Cirac, Computational com-


plexity of projected entangled pair states, Phys. Rev. Lett. 98, 140506 (2007),
doi:10.1103/PhysRevLett.98.140506.

[63] S. Anders, H. J. Briegel and W. Dür, A variational method based on weighted graph states,
New J. Phys. 9, 361 (2007), doi:10.1088/1367-2630/9/10/361.

[64] G. Vidal, Entanglement renormalization, Phys. Rev. Lett. 99, 220405 (2007),
doi:10.1103/PhysRevLett.99.220405.

[65] G. Vidal, Class of quantum many-body states that can be efficiently simulated, Phys. Rev.
Lett. 101, 110501 (2008), doi:10.1103/PhysRevLett.101.110501.

[66] G. Evenbly and G. Vidal, Algorithms for entanglement renormalization, Phys. Rev. B 79,
144108 (2009), doi:10.1103/PhysRevB.79.144108.

[67] M. Rizzi, S. Montangero and G. Vidal, Simulation of time evolution with mul-
tiscale entanglement renormalization ansatz, Phys. Rev. A 77, 052328 (2008),
doi:10.1103/PhysRevA.77.052328.

[68] G. Evenbly and G. Vidal, Class of highly entangled many-body states that can be efficiently
simulated, Phys. Rev. Lett. 112, 240502 (2014), doi:10.1103/PhysRevLett.112.240502.

[69] A. J. Ferris and D. Poulin, Branching MERA codes: a natural extension of polar codes
(2013), arXiv:1312.4575.

97
SciPost Phys. Lect. Notes 8 (2019)

[70] Y.-Y. Shi, L.-M. Duan and G. Vidal, Classical simulation of quantum many-
body systems with a tree tensor network, Phys. Rev. A 74, 022320 (2006),
doi:10.1103/PhysRevA.74.022320.

[71] M. Van den Nest, W. Dür, G. Vidal and H. J. Briegel, Classical simulation versus univer-
sality in measurement-based quantum computation, Phys. Rev. A 75, 012337 (2007),
doi:10.1103/physreva.75.012337.

[72] I. L. Markov and Y. Shi, Simulating quantum computation by contracting tensor networks,
SIAM J. Comput. 38, 963 (2008), doi:10.1137/050644756.

[73] L. Tagliacozzo, G. Evenbly and G. Vidal, Simulation of two-dimensional quantum systems


using a tree tensor network that exploits the entropic area law, Phys. Rev. B 80, 235127
(2009), doi:10.1103/PhysRevB.80.235127.

[74] V. Murg, F. Verstraete, Ö. Legeza and R. M. Noack, Simulating strongly corre-


lated quantum systems with tree tensor networks, Phys. Rev. B 82, 205105 (2010),
doi:10.1103/PhysRevB.82.205105.

[75] P. Silvi, V. Giovannetti, S. Montangero, M. Rizzi, J. I. Cirac and R. Fazio, Homogeneous


binary trees as ground states of quantum critical Hamiltonians, Phys. Rev. A 81, 062335
(2010), doi:10.1103/PhysRevA.81.062335.

[76] F. Mezzacapo, N. Schuch, M. Boninsegni and J. I. Cirac, Ground-state properties of


quantum many-body systems: entangled-plaquette states and variational Monte Carlo,
New J. Phys. 11, 083026 (2009), doi:10.1088/1367-2630/11/8/083026.

[77] N. Schuch, M. M. Wolf, F. Verstraete and J. Ignacio Cirac, Simulation of quantum many-
body systems with strings of operators and Monte Carlo tensor contractions, Phys. Rev. Lett.
100, 040501 (2008), doi:10.1103/PhysRevLett.100.040501.

[78] R. Hübener, C. Kruszynska, L. Hartmann, W. Dür, F. Verstraete, J. Eisert and M. B. Plenio,


Renormalization algorithm with graph enhancement, Phys. Rev. A 79, 022317 (2009),
doi:10.1103/PhysRevA.79.022317.

[79] R. Hübener, C. Kruszynska, L. Hartmann, W. Dür, M. B. Plenio and J. Eisert, Ten-


sor network methods with graph enhancement, Phys. Rev. B 84, 125103 (2011),
doi:10.1103/PhysRevB.84.125103.

[80] F. Verstraete and J. I. Cirac, Continuous matrix product states for quantum Fields, Phys.
Rev. Lett. 104, 190405 (2010), doi:10.1103/PhysRevLett.104.190405.

[81] L. Wang, I. Pižorn and F. Verstraete, Monte Carlo simulation with tensor network states,
Phys. Rev. B 83, 134421 (2011), doi:10.1103/physrevb.83.134421.

[82] A. J. Ferris and G. Vidal, Perfect sampling with unitary tensor networks, Phys. Rev. B 85,
165146 (2012), doi:10.1103/PhysRevB.85.165146.

[83] A. J. Ferris and G. Vidal, Variational Monte Carlo with the multiscale entanglement renor-
malization ansatz, Phys. Rev. B 85, 165147 (2012), doi:10.1103/PhysRevB.85.165147.

[84] A. J. Ferris, Unbiased Monte Carlo for the age of tensor networks (2015), arXiv:1507.00767.

[85] A. W. Sandvik and G. Vidal, Variational quantum Monte Carlo simulations with tensor-
network states, Phys. Rev. Lett. 99, 220602 (2007), doi:10.1103/physrevlett.99.220602.

98
SciPost Phys. Lect. Notes 8 (2019)

[86] O. Sikora, H.-W. Chang, C.-P. Chou, F. Pollmann and Y.-J. Kao, Variational Monte Carlo
simulations using tensor-product projected states, Phys. Rev. B 91, 165113 (2015),
doi:10.1103/PhysRevB.91.165113.

[87] A. J. Ferris and D. Poulin, Tensor networks and quantum error correction, Phys. Rev. Lett.
113, 030501 (2014), doi:10.1103/PhysRevLett.113.030501.

[88] S.-J. Ran, A. Piga, C. Peng, G. Su and M. Lewenstein, Few-body systems capture
many-body physics: Tensor network approach, Phys. Rev. B 96, 155120 (2017),
doi:10.1103/PhysRevB.96.155120.

[89] P. Silvi, Tensor networks: a quantum-information perspective on numerical renormalization


groups (2012), arXiv:1205.4198.

[90] T. Xiang, J. Lou and Z. Su, Two-dimensional algorithm of the density-matrix renormaliza-
tion group, Phys. Rev. B 64, 104414 (2001), doi:10.1103/PhysRevB.64.104414.

[91] T. Xiang, Density-matrix renormalization-group method in momentum space, Phys. Rev. B


53, R10445 (1996), doi:10.1103/PhysRevB.53.R10445.

[92] S. Nishimoto, E. Jeckelmann, F. Gebhard and R. M. Noack, Application of the density


matrix renormalization group in momentum space, Phys. Rev. B 65, 165114 (2002),
doi:10.1103/PhysRevB.65.165114.

[93] A. J. Ferris, Fourier transform for fermionic systems and the spectral tensor network, Phys.
Rev. Lett. 113, 010401 (2014), doi:10.1103/PhysRevLett.113.010401.

[94] G. Ehlers, J. Sólyom, Ö. Legeza and R. M. Noack, Entanglement structure


of the Hubbard model in momentum space, Phys. Rev. B 92, 235116 (2015),
doi:10.1103/PhysRevB.92.235116.

[95] G. Ehlers, S. R. White and R. M. Noack, Hybrid-space density matrix renormalization


group study of the doped two-dimensional Hubbard model, Phys. Rev. B 95, 125125 (2017),
doi:10.1103/PhysRevB.95.125125.

[96] P. Jordan and E. Wigner, Über das Paulische Äquivalenzverbot, Z. Physik 47, 631 (1928),
doi:10.1007/bf01331938.

[97] M. A. Nielsen, The fermionic canonical commutation relations and the Jordan-Wigner
transform, School of Physical Sciences, The University of Queensland (2005),
http://michaelnielsen.org/blog/archive/notes/fermions_and_jordan_wigner.pdf.

[98] P. Corboz and G. Vidal, Fermionic multiscale entanglement renormalization ansatz, Phys.
Rev. B 80, 165129 (2009), doi:10.1103/PhysRevB.80.165129.

[99] P. Corboz, R. Orús, B. Bauer and G. Vidal, Simulation of strongly correlated fermions in
two spatial dimensions with fermionic projected entangled-pair states, Phys. Rev. B 81,
165104 (2010), doi:10.1103/PhysRevB.81.165104.

[100] M. Creutz, Quarks, gluons and lattices, Cambridge University Press (1983).

[101] D. Banerjee, M. Bögli, M. Dalmonte, E. Rico, P. Stebler, U.-J. Wiese and P. Zoller, Atomic
quantum simulation of U(N) and SU(N) non-abelian lattice gauge theories, Phys. Rev. Lett.
110, 125303 (2013), doi:10.1103/physrevlett.110.125303.

99
SciPost Phys. Lect. Notes 8 (2019)

[102] M. C. Bañuls, K. Cichy, J. I. Cirac and K. Jansen, The mass spectrum of the
Schwinger model with matrix product states, J. High Energ. Phys. 11, 158 (2013),
doi:10.1007/JHEP11(2013)158.

[103] M. C. Banuls, K. Cichy, J. I. Cirac, K. Jansen and H. Saito, Matrix product states for lattice
field theories, Proc. Sci. 187-LATTICE2013, 0332 (2014), doi:10.22323/1.187.0332.

[104] L. Tagliacozzo, A. Celi and M. Lewenstein, Tensor networks for lattice gauge theories with
continuous groups, Phys. Rev. X 4, 041024 (2014), doi:10.1103/PhysRevX.4.041024.

[105] S. Kühn, E. Zohar, J. I. Cirac and M. C. Bañuls, Non-abelian string breaking


phenomena with matrix product states, J. High Energ. Phys. 07, 130 (2015),
doi:10.1007/JHEP07(2015)130.

[106] T. Pichler, M. Dalmonte, E. Rico, P. Zoller and S. Montangero, Real-time dynamics


in U(1) lattice gauge theories with tensor networks, Phys. Rev. X 6, 011023 (2016),
doi:10.1103/PhysRevX.6.011023.

[107] B. Buyens, S. Montangero, J. Haegeman, F. Verstraete and K. Van Acoleyen, Finite-


representation approximation of lattice gauge theories at the continuum limit with tensor
networks, Phys. Rev. D 95, 094509 (2017), doi:10.1103/physrevd.95.094509.

[108] P. Silvi, G. De Chiara, T. Calarco, G. Morigi and S. Montangero, Full characterization


of the quantum linear-zigzag transition in atomic chains, Ann. Physik 525, 827 (2013),
doi:10.1002/andp.201300090.

[109] S. Iblisdir, R. Orús and J. I. Latorre, Matrix product states algorithms and continuous
systems, Phys. Rev. B 75, 104305 (2007), doi:10.1103/PhysRevB.75.104305.

[110] P. Silvi, T. Calarco, G. Morigi and S. Montangero, Ab initio characterization of the quantum
linear-zigzag transition using density matrix renormalization group calculations, Phys. Rev.
B 89, 094103 (2014), doi:10.1103/PhysRevB.89.094103.

[111] J. Haegeman, J. I. Cirac, T. J. Osborne and F. Verstraete, Calculus of continuous matrix


product states, Phys. Rev. B 88, 085118 (2013), doi:10.1103/PhysRevB.88.085118.

[112] S. A. Cannas, One-dimensional Ising model with long-range interactions: A renormalization-


group treatment, Phys. Rev. B 52, 3034 (1995), doi:10.1103/PhysRevB.52.3034.

[113] H. J. Changlani, S. Ghosh, C. L. Henley and A. M. Läuchli, Heisenberg antiferromagnet on


Cayley trees: Low-energy spectrum and even/odd site imbalance, Phys. Rev. B 87, 085107
(2013), doi:10.1103/PhysRevB.87.085107.

[114] C. Kollath, A. M. Läuchli and E. Altman, Quench dynamics and nonequilibrium


phase diagram of the Bose-Hubbard model, Phys. Rev. Lett. 98, 180601 (2007),
doi:10.1103/physrevlett.98.180601.

[115] P. Barmettler, M. Punk, V. Gritsev, E. Demler and E. Altman, Relaxation of antiferromagnetic


order in spin-1/2 chains following a quantum quench, Phys. Rev. Lett. 102, 130603 (2009),
doi:10.1103/physrevlett.102.130603.

[116] F. Pellegrini, S. Montangero, G. E. Santoro and R. Fazio, Adiabatic quenches


through an extended quantum critical region, Phys. Rev. B 77, 140404 (2008),
doi:10.1103/PhysRevB.77.140404.

100
SciPost Phys. Lect. Notes 8 (2019)

[117] B. Gardas, J. Dziarmaga and W. H. Zurek, Dynamics of the quantum phase transition
in the one-dimensional Bose-Hubbard model: Excitations and correlations induced by a
quench, Phys. Rev. B 95, 104306 (2017), doi:10.1103/physrevb.95.104306.

[118] P. Doria, T. Calarco and S. Montangero, Optimal control technique for many-body quantum
dynamics, Phys. Rev. Lett. 106, 190501 (2011), doi:10.1103/PhysRevLett.106.190501.

[119] Y. Saad, Analysis of some Krylov subspace approximations to the matrix exponential
operator, SIAM J. Numer. Anal. 29, 209 (1992), doi:10.1137/0729014.

[120] C. Karrasch, J. H. Bardarson and J. E. Moore, Finite-temperature dynamical density


matrix renormalization group and the drude weight of spin-1/2 chains, Phys. Rev. Lett.
108, 227206 (2012), doi:10.1103/PhysRevLett.108.227206.

[121] C. Karrasch, J. H. Bardarson and J. E. Moore, Reducing the numerical effort of finite-
temperature density matrix renormalization group calculations, New J. Phys. 15, 083031
(2013), doi:10.1088/1367-2630/15/8/083031.

[122] M. P. Zaletel, R. S. K. Mong, C. Karrasch, J. E. Moore and F. Pollmann, Time-evolving


a matrix product state with long-ranged interactions, Phys. Rev. B 91, 165112 (2015),
doi:10.1103/PhysRevB.91.165112.

[123] P. Silvi, G. Morigi, T. Calarco and S. Montangero, Crossover from classi-


cal to quantum Kibble-Zurek scaling, Phys. Rev. Lett. 116, 225701 (2016),
doi:10.1103/PhysRevLett.116.225701.

[124] S. van Frank et al., Optimal control of complex atomic quantum systems, Sci. Rep. 6,
34187 (2016), doi:10.1038/srep34187.

[125] M. C. Bañuls, M. B. Hastings, F. Verstraete and J. I. Cirac, Matrix product states


for dynamical simulation of infinite chains, Phys. Rev. Lett. 102, 240603 (2009),
doi:10.1103/PhysRevLett.102.240603.

[126] A. H. Werner, D. Jaschke, P. Silvi, M. Kliesch, T. Calarco, J. Eisert and S. Montangero,


Positive tensor network approach for simulating open quantum many-body systems, Phys.
Rev. Lett. 116, 237201 (2016), doi:10.1103/PhysRevLett.116.237201.

[127] A. J. Daley, Quantum trajectories and open many-body quantum systems, Adv. Phys. 63,
77 (2014), doi:10.1080/00018732.2014.933502.

[128] J. Jordan, R. Orús, G. Vidal, F. Verstraete and J. I. Cirac, Classical simulation of infinite-size
quantum lattice systems in two spatial dimensions, Phys. Rev. Lett. 101, 250602 (2008),
doi:10.1103/physrevlett.101.250602.

[129] M. Rizzi, S. Montangero, P. Silvi, V. Giovannetti and R. Fazio, Homogeneous multiscale


entanglement renormalization ansatz tensor networks for quantum critical systems, New J.
Phys. 12, 075018 (2010), doi:10.1088/1367-2630/12/7/075018.

[130] G. Evenbly and G. Vidal, Tensor network states and geometry, J. Stat. Phys. 145, 891
(2011), doi:10.1007/s10955-011-0237-4.

[131] K. Wierschem and P. Sengupta, Characterizing the Haldane phase in quasi-one-


dimensional spin-1 Heisenberg antiferromagnets, Mod. Phys. Lett. B 28, 1430017 (2014),
doi:10.1142/S0217984914300178.

101
SciPost Phys. Lect. Notes 8 (2019)

[132] B. Pirvu, V. Murg, J. I. Cirac and F. Verstraete, Matrix product operator representations,
New J. Phys. 12, 025012 (2010), doi:10.1088/1367-2630/12/2/025012.

[133] P. Calabrese and J. Cardy, Entanglement entropy and quantum field theory, J. Stat. Mech.:
Theor. Exp. P06002 (2004), doi:10.1088/1742-5468/2004/06/p06002.

[134] R. Orús, T.-C. Wei, O. Buerschaper and A. García-Saez, Topological transitions from
multipartite entanglement with tensor networks: a procedure for sharper and faster charac-
terization, Phys. Rev. Lett. 113, 257202 (2014), doi:10.1103/PhysRevLett.113.257202.

[135] S. Singh, R. N. C. Pfeifer and G. Vidal, Tensor network states and algorithms
in the presence of a global U(1) symmetry, Phys. Rev. B 83, 115125 (2011),
doi:10.1103/PhysRevB.83.115125.

[136] E. Anderson et al., LAPACK Users’ Guide, Society for Industrial and Applied Mathematics
(1999), doi:10.1137/1.9780898719604.

[137] A. Petitet, PBLAS (1995), http://www.netlib.org/utk/papers/scalapack/node9.html.

[138] K. Goto, GotoBLAS (2010), https://www.tacc.utexas.edu/research-development/


tacc-software/gotoblas2.
[139] Z. Xianyi, W. Qian and W. Saar, OpenBLAS (2017), http://www.openblas.net/.

[140] E. M. Stoudenmire et al., iTensor – Intelligent Tensor library (2010), http://itensor.org/.

[141] Universal Tensor Network Library – Uni10 version 2.0.0 (2018), https://uni10.gitlab.io/.

[142] S. Al-Assam, S. R. Clark and D. Jaksch, The tensor network theory library, J. Stat. Mech.
093102 (2017), doi:10.1088/1742-5468/aa7df3.

[143] C. Hubig, I. P. McCulloch, U. Schollwöck and F. A. Wolf, Strictly single-site


DMRG algorithm with subspace expansion, Phys. Rev. B 91, 155115 (2015),
doi:10.1103/physrevb.91.155115.

[144] J. Liesen and Z. Strakos, Krylov subspace methods: Principles and analysis, Oxford
University Press (2012).

[145] E. Agullo, J. Dongarra, B. Hadri, J. Kurzak, J. Langou, J. Langou, H. Ltaief, P. Luszczek


and A. YarKhan, PLASMA Users’ Guide, Tech. rep., ICL, University of Tennessee (2009).

[146] L. S. Blackford et al., ScaLAPACK Users’ Guide, Society for Industrial and Applied Mathe-
matics (1997), doi:10.1137/1.9780898719642.

[147] D. Coppersmith and S. Winograd, Matrix multiplication via arithmetic progressions, J.


Symb. Comput. 9, 251 (1990), doi:10.1016/S0747-7171(08)80013-2.

[148] R. N. C. Pfeifer, J. Haegeman and F. Verstraete, Faster identification of opti-


mal contraction sequences for tensor networks, Phys. Rev. E 90, 033315 (2014),
doi:10.1103/physreve.90.033315.

[149] R. N. Pfeifer, G. Evenbly, S. Singh and G. Vidal, NCON: A tensor network contractor for
MATLAB (2014), arXiv:1402.0939.

[150] G. Strang, Introduction to linear algebra, Wellesley-Cambridge Press (2011).

[151] L. N. Trefethen and D. Bau III, Numerical linear algebra, Society for Industrial and Applied
Mathematics (1997).

102
SciPost Phys. Lect. Notes 8 (2019)

[152] R. M. Larsen, PROPACK-software for large and sparse SVD calculations (2004), http:
//sun.stanford.edu/~rmunk/PROPACK (2004).
[153] N. Halko, P. G. Martinsson and J. A. Tropp, Finding structure with randomness: Proba-
bilistic algorithms for constructing approximate matrix decompositions, SIAM Rev. 53, 217
(2011), doi:10.1137/090771806.

[154] P.-G. Martinsson and S. Voronin, A randomized blocked algorithm for efficiently computing
rank-revealing factorizations of matrices (2015), arXiv:1503.07157.

[155] D. Tamascelli, R. Rosenbach and M. B. Plenio, Improved scaling of time-evolving block-


decimation algorithm through reduced-rank randomized singular value decomposition,
Phys. Rev. E 91, 063306 (2015), doi:10.1103/physreve.91.063306.

[156] L. Kohn, F. Tschirsich, M. Keck, M. B. Plenio, D. Tamascelli and S. Montangero, Probabilis-


tic low-rank factorization accelerates tensor network simulations of critical quantum many-
body ground states, Phys. Rev. E 97, 013301 (2018), doi:10.1103/PhysRevE.97.013301.

[157] C. Lanczos, An iteration method for the solution of the eigenvalue problem of lin-
ear differential and integral operators, J. Res. Natl. Bur. Stan. 45, 255 (1950),
doi:10.6028/jres.045.026.

[158] E. R. Davidson, The iterative calculation of a few of the lowest eigenvalues and corre-
sponding eigenvectors of large real-symmetric matrices, J. Comput. Phys. 17, 87 (1975),
doi:10.1016/0021-9991(75)90065-0.

[159] W. E. Arnoldi, The principle of minimized iterations in the solution of the matrix eigenvalue
problem, Quart. Appl. Math. 9, 17 (1951), doi:10.1090/qam/42792.

[160] M. Arioli, B. Codenotti and C. Fassino, The Padé method for computing the matrix expo-
nential, Linear Algebra App. 240, 111 (1996), doi:10.1016/0024-3795(94)00190-1.

[161] R. B. Sidje, Expokit: a software package for computing matrix exponentials, ACM Trans.
Math. Softw. 24, 130 (1998), doi:10.1145/285861.285868.

[162] M. Tinkham, Group theory and quantum mechanics, Courier Corporation (2003).

[163] I. P. McCulloch and M. Gulácsi, The non-Abelian density matrix renormalization group
algorithm, Europhys. Lett. 57, 852 (2002), doi:10.1209/epl/i2002-00393-0.

[164] A. Weichselbaum, Non-abelian symmetries in tensor networks: A quantum symmetry space


approach, Ann. Phys. 327, 2972 (2012), doi:10.1016/j.aop.2012.07.009.

[165] S. Singh and G. Vidal, Tensor network states and algorithms in the presence of a global
SU(2) symmetry, Phys. Rev. B 86, 195114 (2012), doi:10.1103/PhysRevB.86.195114.

[166] P. Schmoll, S. Singh, M. Rizzi and R. Orús, A programming guide for tensor networks with
global SU(2) symmetry (2018), arXiv:1809.08180.

[167] P. Corboz, A. M. Läuchli, K. Penc, M. Troyer and F. Mila, Simultaneous dimerization and
SU(4) symmetry breaking of 4-color fermions on the square lattice, Phys. Rev. Lett. 107,
215301 (2011), doi:10.1103/PhysRevLett.107.215301.

[168] P. Fendley, Parafermionic edge zero modes in Zn -invariant spin chains, J. Stat. Mech.
P11020 (2012), doi:10.1088/1742-5468/2012/11/p11020.

103
SciPost Phys. Lect. Notes 8 (2019)

[169] F. H. L. Essler, H. Frahm, F. Göhmann, A. Klümper and V. E. Korepin, The one-dimensional


Hubbard model, Cambridge University Press (2005), doi:10.1017/cbo9780511534843.

[170] S. Singh, R. N. C. Pfeifer and G. Vidal, Tensor network decompositions in the presence of a
global symmetry, Phys. Rev. A 82, 050301 (2010), doi:10.1103/physreva.82.050301.

[171] B. Bauer, P. Corboz, R. Orús and M. Troyer, Implementing global abelian symme-
tries in projected entangled-pair state algorithms, Phys. Rev. B 83, 125106 (2011),
doi:10.1103/PhysRevB.83.125106.

[172] P. Silvi, E. Rico, T. Calarco and S. Montangero, Lattice gauge tensor networks, New J.
Phys. 16, 103015 (2014), doi:10.1088/1367-2630/16/10/103015.

[173] L. Tagliacozzo and G. Vidal, Entanglement renormalization and gauge symmetry, Phys.
Rev. B 83, 115127 (2011), doi:10.1103/physrevb.83.115127.

[174] L. Tagliacozzo, A. Celi, A. Zamora and M. Lewenstein, Optical abelian lattice gauge
theories, Ann. Phys. 330, 160 (2013), doi:10.1016/j.aop.2012.11.009.

[175] B. A. Bernevig, Topological insulators and topological superconductors, Princeton University


Press (2013), doi:10.1515/9781400846733.

[176] F. Pollmann, A. M. Turner, E. Berg and M. Oshikawa, Entanglement spectrum


of a topological phase in one dimension, Phys. Rev. B 81, 064439 (2010),
doi:10.1103/PhysRevB.81.064439.

[177] V. Nebendahl and W. Dür, Improved numerical methods for infinite spin chains with long-
range interactions, Phys. Rev. B 87, 075413 (2013), doi:10.1103/PhysRevB.87.075413.

[178] B. Pirvu, J. Haegeman and F. Verstraete, Matrix product state based algorithm for deter-
mining dispersion relations of quantum spin chains with periodic boundary conditions,
Phys. Rev. B 85, 035130 (2012), doi:10.1103/PhysRevB.85.035130.

[179] N. Schuch, I. Cirac and D. Pérez-García, PEPS as ground states: Degeneracy and topology,
Ann. Phys. 325, 2153 (2010), doi:10.1016/j.aop.2010.05.008.

[180] S. Singh and G. Vidal, Global symmetries in tensor network states: Symmet-
ric tensors versus minimal bond dimension, Phys. Rev. B 88, 115147 (2013),
doi:10.1103/physrevb.88.115147.

[181] N. Nakatani and G. Kin-Lic Chan, Efficient tree tensor network states (TTNS) for quantum
chemistry: Generalizations of the density matrix renormalization group algorithm, J. Chem.
Phys. 138, 134113 (2013), doi:10.1063/1.4798639.

[182] M. Gerster, P. Silvi, M. Rizzi, R. Fazio, T. Calarco and S. Montangero, Unconstrained tree
tensor network: An adaptive gauge picture for enhanced performance, Phys. Rev. B 90,
125154 (2014), doi:10.1103/PhysRevB.90.125154.

[183] V. Alba, L. Tagliacozzo and P. Calabrese, Entanglement entropy of two disjoint blocks in
critical Ising models, Phys. Rev. B 81, 060411 (2010), doi:10.1103/physrevb.81.060411.

[184] A. J. Ferris, Area law and real-space renormalization, Phys. Rev. B 87, 125139 (2013),
doi:10.1103/physrevb.87.125139.

[185] C. Eckart and G. Young, The approximation of one matrix by another of lower rank,
Psychometrika 1, 211 (1936), doi:10.1007/bf02288367.

104
SciPost Phys. Lect. Notes 8 (2019)

[186] L. De Lathauwer, B. De Moor and J. Vandewalle, A multilinear singular value decomposi-


tion, SIAM J. Matrix Anal. Appl. 21, 1253 (2000), doi:10.1137/S0895479896305696.

[187] M. B. Hastings, Light-cone matrix product, J. Math. Phys. 50, 095207 (2009),
doi:10.1063/1.3149556.

[188] S. Giorgini, L. P. Pitaevskii and S. Stringari, Theory of ultracold atomic Fermi gases, Rev.
Mod. Phys. 80, 1215 (2008), doi:10.1103/revmodphys.80.1215.

[189] J. I. Cirac and F. Verstraete, Renormalization and tensor product states in spin
chains and lattices, J. Phys. A: Math. Theor. 42, 504004 (2009), doi:10.1088/1751-
8113/42/50/504004.

[190] R. Hübener, V. Nebendahl and W. Dür, Concatenated tensor network states, New J. Phys.
12, 025004 (2010), doi:10.1088/1367-2630/12/2/025004.

[191] F. Gliozzi and L. Tagliacozzo, Entanglement entropy and the complex plane of replicas, J.
Stat. Mech. P01002 (2010), doi:10.1088/1742-5468/2010/01/p01002.

[192] V. Alba, L. Tagliacozzo and P. Calabrese, Entanglement entropy of two disjoint intervals in c
= 1 theories, J. Stat. Mech. P06012 (2011), doi:10.1088/1742-5468/2011/06/p06012.

[193] P. Calabrese, L. Tagliacozzo and E. Tonni, Entanglement negativity in the critical Ising
chain, J. Stat. Mech. P05002 (2013), doi:10.1088/1742-5468/2013/05/p05002.

[194] M. Gerster, M. Rizzi, F. Tschirsich, P. Silvi, R. Fazio and S. Montangero, Superfluid density
and quasi-long-range order in the one-dimensional disordered Bose–Hubbard model, New
J. Phys. 18, 015015 (2016), doi:10.1088/1367-2630/18/1/015015.

[195] H. N. Phien, J. A. Bengua, H. D. Tuan, P. Corboz and R. Orús, Infinite projected entangled
pair states algorithm improved: Fast full update and gauge fixing, Phys. Rev. B 92, 035142
(2015), doi:10.1103/PhysRevB.92.035142.

[196] P. Pippan, S. R. White and H. Gerd Evertz, Efficient matrix-product state method for periodic
boundary conditions, Phys. Rev. B 81, 081103 (2010), doi:10.1103/PhysRevB.81.081103.

[197] B. Pirvu, G. Vidal, F. Verstraete and L. Tagliacozzo, Matrix product states for critical spin
chains: Finite-size versus finite-entanglement scaling, Phys. Rev. B 86, 075117 (2012),
doi:10.1103/PhysRevB.86.075117.

[198] A. Toselli and O. B. Widlund, Domain decomposition methods: algorithms and theory,
Springer (2005).

[199] T. P. A. Mathew, Domain decomposition methods for the numerical solution of partial
differential equations, Springer Berlin Heidelberg (2008), doi:10.1007/978-3-540-77209-
5.

[200] I. P. McCulloch, From density-matrix renormalization group to matrix product states, J.


Stat. Mech. P10014 (2007), doi:10.1088/1742-5468/2007/10/p10014.

[201] S. R. White, Density matrix renormalization group algorithms with a single center site,
Phys. Rev. B 72, 180403 (2005), doi:10.1103/PhysRevB.72.180403.

[202] D. Sorensen, R. Lehoucq, C. Yang, K. Maschhoff, S. Ledru and A. Cornet, arpack-ng,


(2017), https://github.com/opencollab/arpack-ng.

[203] Netlib, LAPACK, (2017), http://www.netlib.org/lapack.

105
SciPost Phys. Lect. Notes 8 (2019)

[204] D. R. Fokkema, G. L. G. Sleijpen and H. A. Van der Vorst, Jacobi–Davidson style QR and
QZ algorithms for the reduction of matrix pencils, SIAM J. Sci. Comput. 20, 94 (1998),
doi:10.1137/s1064827596300073.

[205] M. Cominotti, D. Rossini, M. Rizzi, F. Hekking and A. Minguzzi, Optimal persistent


currents for interacting bosons on a ring with a gauge field, Phys. Rev. Lett. 113, 025301
(2014), doi:10.1103/PhysRevLett.113.025301.

[206] L. Onsager, Crystal statistics. I. A two-dimensional model with an order-disorder transition,


Phys. Rev. 65, 117 (1944), doi:10.1103/PhysRev.65.117.

[207] S.-J. Ran, W. Li, B. Xi, Z. Zhang and G. Su, Optimized decimation of tensor networks
with super-orthogonalization for two-dimensional quantum lattice models, Phys. Rev. B
86, 134429 (2012), doi:10.1103/PhysRevB.86.134429.

[208] S.-J. Ran, B. Xi, T. Liu and G. Su, Theory of network contractor dynamics for exploring
thermodynamic properties of two-dimensional quantum lattice models, Phys. Rev. B 88,
064407 (2013), doi:10.1103/PhysRevB.88.064407.

[209] M. Gerster, M. Rizzi, P. Silvi, M. Dalmonte and S. Montangero, Fractional quantum Hall
effect in the interacting Hofstadter model via tensor networks, Phys. Rev. B 96, 195123
(2017), doi:10.1103/PhysRevB.96.195123.

[210] J. Haegeman, C. Lubich, I. Oseledets, B. Vandereycken and F. Verstraete, Unifying time


evolution and optimization with matrix product states, Phys. Rev. B 94, 165116 (2016),
doi:10.1103/PhysRevB.94.165116.

[211] P. Hauke and L. Tagliacozzo, Spread of correlations in long-range interacting quantum


systems, Phys. Rev. Lett. 111, 207202 (2013), doi:10.1103/physrevlett.111.207202.

[212] F. A. Y. N. Schröder and A. W. Chin, Simulating open quantum dynamics with time-
dependent variational matrix product states: Towards microscopic correlation of en-
vironment dynamics and reduced system evolution, Phys. Rev. B 93, 075105 (2016),
doi:10.1103/PhysRevB.93.075105.

[213] E. P. Wigner, Gruppentheorie und ihre Anwendung auf die Quantenmechanik der Atom-
spektren, Springer (1931).

[214] F. Haake, Quantum signatures of chaos, Springer Science & Business Media (2013),
doi:10.1007/978-3-662-04506-0.

106

You might also like