of 7
Tensor Network Computations That Capture Strict Variationality, Volume Law
Behavior, and the Efficient Representation of Neural Network States
Wen-Yuan Liu ,
1
,*
Si-Jing Du ,
2
,*
Ruojing Peng,
1
Johnnie Gray ,
1
and Garnet Kin-Lic Chan
1
,
1
Division of Chemistry and Chemical Engineering,
California Institute of Technology
, Pasadena, California 91125, USA
2
Division of Engineering and Applied Science,
California Institute of Technology
, Pasadena, California 91125, USA
(Received 21 May 2024; accepted 27 November 2024; published 30 December 2024)
We introduce a change of perspective on tensor network states that is defined by the computational graph
of the contraction of an amplitude. The resulting class of states, which we refer to as tensor network
functions, inherit the conceptual advantages of tensor network states while removing computational
restrictions arising from the need to converge approximate contractions. We use tensor network functions to
compute strict variational estimates of the energy on loopy graphs, analyze their expressive power for
ground states, show that we can capture aspects of volume law time evolution, and provide a mapping of
general feed-forward neural nets onto efficient tensor network functions. Our work expands the realm of
computable tensor networks to ones where accurate contraction methods are not available, and opens up
new avenues to use tensor networks.
DOI:
10.1103/PhysRevLett.133.260404
Introduction
Simulating and representing quantum
many-body systems is a central task of physics. For this,
tensor network states have emerged as a fundamental lan-
guage and numerical tool
[1
6]
, with applications across
diverse areas such as condensed matter physics
[7
12]
,
statistical physics
[13
16]
, quantum field theory
[17
19]
,
and quantum information theory
[20
22]
, as well as adjacent
disciplines such as quantum chemistry and machine learning
[23
26]
.
Tensor networks represent the amplitude of a quantum
state on an
L
site lattice, denoted
h
n
j
Ψ
i
Ψ
ð
n
Þ
Ψ
ð
n
1
;n
2
;
...
;n
L
Þ
, as the contraction of a set of tensors. As
a simple example, we consider
L
tensors connected in a graph
[Fig.
1(a)
],eachassociatedwithalatticesiteHilbertspace;this
is known as a projected entangled pair state (PEPS)
[27]
.Each
tensor contains
dD
m
elements (
m
is the number of bonds
which connect tensors to adjacent tensors):
d
is the local
Hilbert space dimension and
D
controls the expressivity. For
tree graphs, the resulting contraction over the shared tensor
bonds can be computed exactly in
O
ð
L
Þ
poly
ð
D
Þ
time
[28]
.
Forgeneral graphs,however, both
h
n
j
Ψ
i
and
h
Ψ
j
ˆ
O
j
Ψ
i
cannot
be exactly contracted efficiently: the cost is exponential in the
tree width of the graph
[21]
. Approximate contraction
algorithms have thus been introduced whereby during the
contraction process, the large intermediate tensors that are
formed are compressed, via singular value decomposition
(SVD), to a controlled size,
O
½
poly
ð
χ
Þ
,where
χ
is termed the
auxiliarybonddimension
[2,29]
.Onlyinthelimitof
χ
is
the contraction exact. The resulting approximate contraction
has a cost
O
ð
L
Þ
poly
ð
D
Þ
poly
ð
χ
Þ
, with an error with respect to
the exact contraction controlled by
χ
.
From the perspective of approximate contraction, tensor
network computations should seek to be converged with
respect to
χ
, otherwise the results are not meaningful. For
example, for a Hamiltonian
ˆ
H
,if
h
Ψ
j
ˆ
H
j
Ψ
i
=
h
Ψ
j
Ψ
i
is com-
puted by approximate contraction, with insufficient
χ
the
energy may violate the variational theorem. The requirement
that computations should use a sufficiently large
χ
(such that
the evaluation of
h
Ψ
j
ˆ
O
j
Ψ
i
or
h
n
j
Ψ
i
is no longer changing
significantly with
χ
) places restrictions on the types of graphs
that can be used to define the tensor network. For example, in
many-body physics simulations, it is believed that only highly
(b)(c)
(a)
(d)
(e)
FIG. 1. (a) PEPS on a general graph with
L
site tensors.
(b) PEPS on a
4
×
4
square lattice. (c) Insertion of a pair of
isometries for approximate tensor contractions; the isometries can
be obtained with standard tensor network techniques by SVD
[2,3]
. (d) Approximate PEPS contractions by inserting isometries
for a given configuration
j
n
i
(gray circles). (e) For a given input
vector
j
n
i¼ð
n
1
;n
2
;
...
;n
L
Þ
, a tensor network function outputs a
unique value
Ψ
ð
n
Þ
.
*
These authors contributed equally to this work.
Contact author: gkc1000@gmail.com
PHYSICAL REVIEW LETTERS
133,
260404 (2024)
0031-9007
=
24
=
133(26)
=
260404(7)
260404-1
© 2024 American Physical Society
structured graphs with limited entanglement, such as (near-)
area law states, support an efficient contraction scheme.
Because of the latter requirement, it has been argued that,
in practical use, tensor networks cannot replicate the expres-
sivity of other flexible types of many-body ansatz, such as
neural network quantum states
[30
36]
.
In the current work, we observe that the above limi-
tations on tensor network computations arise due to an
assumption, namely that the tensor network ansatz is only
meaningfully defined in the exact contraction limit. We can,
however, adopt an alternative perspective, namely, to
consider the approximate contraction scheme for any
χ
as itself part of the tensor network ansatz for the amplitude
Ψ
ð
n
Þ
. This view is natural and in some ways obvious when
using tensor networks in variational Monte Carlo (VMC)
calculations
[37
41]
, but we will argue that this is not a
manifestation of tensor networks in a particular numerical
algorithm, but rather a change of perspective that alters and
broadens the definition and utility of tensor networks. We
refer to the larger set of parameterized many-body states
that arise from
consistent
contraction schemes (to be
defined later) as
tensor network functions
. This wider class
of states removes many restrictions that are assumed to
apply to tensor networks. We demonstrate that tensor
network functions provide strictly variational estimates
of the energy, allow for practical tensor network compu-
tations on graphs that are highly challenging to contract
accurately, including those that describe volume law
physics, and expand the set of states that can be efficiently
represented by tensor networks to cover other classes of
ansatz, such as neural network quantum states.
Tensor network functions
A tensor network function
(TNF) is a consistent contraction of the tensor network
representation of the amplitude
Ψ
ð
n
Þ
. To illustrate the basic
idea, we first consider a standard PEPS for a two-dimensional
lattice,illustratedinFig.
1(b)
.Becausethegraphhasloops,the
PEPS amplitude cannot be computed efficiently according to
Fig.
1(b)
. Instead, in practice, the amplitude is defined by a
different tensor network illustrated in Fig.
1(d)
.Inthisfigure,
isometries (triangles) with bond dimension
χ
have been
inserted, where the values of the isometries are determined
by an SVD of the contraction of neighbouring tensors [see
Fig.
1(c)
]
[2,3]
. The final amplitude is obtained by contracting
the tensors and isometries together. The pattern of isometries
used here corresponds to a specific approximate contraction
method (boundary contraction
[2,27]
) and other choices can
be made
[29,42]
, but the essential purpose is to make the
contraction of the modified graph containing the original
tensors and the isometries efficient.
Note, however, that although the structure of the tensor
networkwith isometries in Fig.
1(d)
nowappearspseudo-one-
dimensional, this does not mean that its entanglement
structure (for
χ
>
1
) is reduced to that of a matrix product
state (MPS). This is because the isometries are themselves
functions of the configuration
j
n
i
, i.e., different values of the
isometriesareobtained,viaSVD,fordifferentconfigurations.
However, we choose to keep the same
positions
of the
isometries in the graph for any configuration (even though
the entries are different); we refer to this as using
fixed
(position) isometries, which defines a consistent contraction.
Because of the configuration dependence of the isometry
entries, the graph with isometries no longer represents a
multilinear ansatz like the original tensor network state; it is
only a multilinear ansatz in the limit that
χ
(and the
isometries become identities). Instead, it defines a
tensor
network function
: for any given
ð
D;
χ
Þ
,any
j
n
i
is mapped to a
single value, the amplitude, through an efficient contraction.
The efficient TNF representation of
Ψ
ð
n
Þ
does not guarantee
deterministic efficiency of computing
h
Ψ
j
ˆ
O
j
Ψ
i
;butsuch
expectations may be evaluated stochastically by sampling
with efficient sample complexity.
We emphasize that herewe aredefining a new classofstate,
i.e., the TNF, whose amplitudes are exactly given by a certain
contraction graph of a tensor network. The essential
differences between TNF and conventional tensor network
computationsarefurtherillustratedinthefollowingexamples,
and discussed also in Sec. 1 in Supplemental Material
[43]
.
Tensor network functions and approximate contraction
VMC algorithms also compute
Ψ
ð
n
Þ
via approximate
contraction (for a brief review of tensor network VMC
(TN-VMC) algorithms, see Supplemental Material
[43]
or
Refs.
[37
41]
) and it is instructive to consider the differences
between the standard TN-VMC approximation of
Ψ
ð
n
Þ
and
the definition of a TNF above. There are two seemingly small,
but significant, differences (i) in TN-VMC,
χ
is viewed as a
parameter (typically
χ
¼
const ×
D
, with const
>
1
)where
χ
is increased to convergence for a fixed
D
,and(ii) the positions
of the isometries are not fixed but dependent on the amplitude
thatisbeingcomputed(werefertothisas
dynamic
isometries),
inordertoallowforthereuseofcomputationswhencomputing
amplitudes of configurations which differ by only a few
substitutions
[41]
. See Supplemental Material for details
[43]
.
The most important distinction from TNF is that the use
of dynamic isometries in TN-VMC
does not lead to a
consistent function
. In other words, given the same con-
figuration
j
n
i
,
Ψ
ð
n
Þ
needs not be the same, as it depends on
the prior configuration and which components of the tensor
network contraction are being reused. Only in the limit of
χ
are the amplitudes consistent and the function
defined, which underlies the usual need to converge
χ
in
TN-VMC calculations. In contrast, it is clear that with
using fixed isometries, the corresponding TNF is consistent
for any
D
and any
χ
. See Supplemental Material for
details
[43]
.
Tensor network functions and variational energies
As
an immediate consequence of the use of fixed isometries to
define a TNF, we can always obtain a variational energy in
Monte Carlo energy evaluation (up to statistical error).
In Fig.
2(a)
we show the energy of an optimized PEPS
(
D
¼
8
, various
χ
) TNF for the ground state of the open-
boundary condition (OBC) spin-
1
=
2
square-lattice
32
×
32
Heisenberg model defined using the boundary contraction
PHYSICAL REVIEW LETTERS
133,
260404 (2024)
260404-2
computational graph, see Supplemental Material
[43]
,
compared to that of a PEPS in standard TN-VMC using
dynamic isometries, where
χ
controls the error of approxi-
mate contraction
[41]
. We see immediately that the PEPS
TNF energy is variational for all
χ
. In contrast, viewing the
PEPS amplitude as an (inconsistent) approximate contrac-
tion as in standard VMC leads to a nonvariational energy,
where the nonvariationality is typically regarded as the
contraction error.
This behavior is even more clear for a
lattice that is usually considered more difficult to contract.
Using the square-lattice
J
1
J
2
model with periodic boun-
dary conditions (PBC) as an example, in Fig.
2(b)
we show
the same comparison between TNF and standard TN-
VMC. In addition to the variational versus nonvariational
nature of the energies, we also see a large variance of the
TN-VMC energy arising from the inconsistent definition of
amplitudes. This leads to substantial difficulties in the
stochastic optimization of the ansatz. Overall, we see that
lack of variationality is not intrinsic to approximate con-
traction: it results from the perspective on the definition of
the ansatz.
Expressivity of tensor network functions with small
χ
In
general, a TNF is parametrized by two bond dimension
parameters,
D
and
χ
. We already know that as
χ
becomes
large, the TNF reduces to the standard tensor network state
from which it is derived, but TNFs are well defined also in
the limit of small
χ
(e.g.,
χ
<D
) and it is important to ask
whether they are expressive in this limit.
In Figs.
3(a)
and
3(b)
, we show results from approxi-
mating the ground-state of a spin-
1
=
2
square-lattice
16
×
16
J
1
J
2
Heisenberg model for
J
2
=J
1
¼
0
.
5
, and an
8
×
8
Hubbard model with
U=t
¼
8
and
1
=
8
hole doping, using
bosonic and fermionic PEPS TNFs for OBC systems
[27,55
57]
, respectively. We minimize the energy for many
different
ð
D;
χ
Þ
combinations using gradient descent, and
we plot the computational cost of the amplitude against the
obtained accuracy. For a specified relative accuracy, we can
then search for the
ð
D;
χ
Þ
pair with the lowest computa-
tional cost. We see that contrary to the standard situation
with tensor network states where one uses
χ
>D
, the most
powerful TNFs for a given computational cost have
χ
<D
.
For example, for
0
.
01
relative accuracy in the Hubbard
model, using
ð
D;
χ
Þ¼ð
12
;
6
Þ
yields the same expressivity
as
ð
D;
χ
Þ¼ð
10
;
10
Þ
, but with lower cost. In no case is it
computationally more efficient to use
χ
>D
. The insets of
Fig.
3
presenting the minimized energy for every ansatz
j
Ψ
ð
D;
χ
Þi
, show that the most rapid improvement in energy
is obtained at small
χ
, rather than saturating
χ
for a given
D
,
confirming the above conclusion.
Tensor network functions and volume laws
As an
example of a new kind of application that we can consider
with TNFs, we consider the case where the tensor network
graph corresponds to chaotic time evolution leading to a
volumelaw state.Thenthestandardview isthatapproximate
contraction should be performed with
χ
exp
ð
const ×
t
Þ
(where
t
is evolution time) to capture the volume law.
However, if we think of the TNF formed by a random
circuit, we might expect that the set of amplitudes, defined
using fixed position but configuration dependent and essen-
tially random isometries, are also scrambled as if by an
approximate unitary design, regardless of
χ
, capturing
(some) elements of the volume law behavior.
To illustrate this numerically, we consider the Floquet
dynamics of an
L
-site 1D kicked Ising chain
[58
60]
.
(a)
(b)
FIG. 3. Time cost (in seconds) versus accuracy of the tensor
network function
j
Ψ
ð
D;
χ
Þi
, for the
J
1
J
2
model on the
16
×
16
lattice at
J
2
=J
1
¼
0
.
5
(a) and Fermi-Hubbard model on the
8
×
8
lattice at
U=t
¼
8
with hole doping
n
h
¼
1
=
8
(b). Refer-
ence energies
E
ref
taken from PEPS
D
¼
10
[
J
1
J
2
,
E
¼
0
.
490 478
ð
4
Þ
] and DMRG [
D
¼
10 000
with SU(2) symmetry,
Hubbard model,
E
¼
0
.
694 391
]. The time cost is for comput-
ing
h
n
j
Ψ
ð
D;
χ
Þi
for a given
j
n
i
on a single core. Highlighted
symbols (connected by a black line) indicate the optimal
ð
D;
χ
Þ
pairs that have smallest time cost for a given accuracy. Here
D
¼
4
8
and
χ
¼
2
12
is used for the
J
1
J
2
model and
D
¼
6
12
and
χ
¼
2
14
for the Hubbard model. The insets present the energy
variation of the TNF
j
Ψ
ð
D;
χ
Þi
with
D
and
χ
.
4
8
12
16
20
24
28
32
-0.68
-0.67
-0.66
-0.65
-0.64
2345678
-0.510
-0.505
-0.500
-0.495
-0.490
32

32 OBC
Energy
E

Fixed-iso
Dynamic-iso
Reference
(a)
(b)
Energy
E

Fixed-iso
Dynamic-iso
Reference
8

8 PBC
FIG. 2. For a given PEPS, energies (per site) obtained from
fixed (tensor network function) and dynamic (standard VMC)
isometry contractions. (a) OBC square-lattice
32
×
32
spin-
1
=
2
Heisenberg model with PEPS
D
¼
8
, and 15000 Monte Carlo
sweeps are used for each data point. The dashed line denotes the
numerically exact energy
[41]
. (b) PBC square-lattice
8
×
8
spin-
1
=
2
frustrated
J
1
J
2
model at
J
2
=J
1
¼
0
.
5
with PEPS
D
¼
5
,
and 5000 Monte Carlo sweeps are used for each data point. The
dashed line denotes the energy from fixed-isometry contraction
with
χ
¼
8
using
D
¼
5
.
PHYSICAL REVIEW LETTERS
133,
260404 (2024)
260404-3
Expressing the Floquet evolution operator
F
as a matrix
productoperator(MPO)(seeSupplementalMaterial
[43]
),the
time evolution corresponds to a
ð
1
þ
1
Þ
Dtensornetwork,
shown in Fig.
4(a)
. The tunable parameters in the model
include the Ising interaction strength
J
, transverse field
g
and
longitudinal field
h
. We focus on the maximally chaotic
regime
[61]
,
ð
J; g; h
Þ¼ð
π
=
4
;
π
=
4
;
0
.
5
Þ
, which exhibits vol-
umelawphysics ina shortevolutiontime.Foreachtime step
t
,
we compute the bipartite von Neumann entropy
S
ð
t
Þ¼
Tr
ð
ρ
A
log
ρ
A
Þ
of the state
j
Ψ
ð
t
Þi ¼
F
t
j
0
i
L
,where
ρ
A
is
thereduceddensitymatrixofthehalf-chain.(Computing
ρ
A
is
not an efficient operation, but we do so to illustrate the physics
being captured). To define a useful TNF, we construct a
contraction path that leads to configuration dependent iso-
metries: we consider two natural ones, namely, (i) contraction
in the inverse time direction, and (ii) (transverse) contraction
along the spatial direction. We then compare against the
standard MPS-MPO compression to a bond dimension
χ
.
Results from (i), as well as from amplitude-independent
MPO-MPO compression in the inverse time direction, are
shown in the Supplemental Material
[43]
.
The evolution of
S
ð
t
Þ
is depicted in Fig.
4(b)
, with exact
resultsincluded.Forthemaximallychaoticdynamics,which
might be thought of as closest to the random circuit intuition
above, the entanglement entropy grows linearly from the
beginning until half-width saturation. The required conven-
tional MPS bond dimension to describe the full dynamics
growsexponentiallywithtime(hereforafinite
L
¼
14
chain
the maximum
χ
needed would be
2
L=
2
¼
128
). However,
volume law physics is already recovered using a TNF with
χ
¼
2
.The
χ
¼
8
TNFaccuratelycapturesthe features ofthe
fullevolution,includingthemaximumpointintheentropyat
t
¼
8
. Similar results are seen for the TNF defined in the
inverse time-direction (see Supplemental Material
[43]
),
although it is quantitatively less accurate. Notably, the
entanglement spectrum (inset of Fig.
4
) for the TNF, even
for small
χ
, has a long tail and captures the broad features of
the spectrum, rather than exhibiting the sharp cutoff of a
conventional truncated MPS. Similar features for the less
chaotic regime
ð
J; g; h
Þ¼ð
0
.
7
;
0
.
5
;
0
.
5
Þ
are also observed;
see Supplemental Material
[43]
.
Tensor network functions can efficiently represent
neural network quantum states
As discussed, TNFs only
require a consistent and efficient contraction of
Ψ
ð
n
Þ
, and
there exist TNFs for which this is the case without any
(amplitude dependent) isometries, but for which the effi-
cient contraction of
h
Ψ
j
ˆ
O
j
Ψ
i
is still not possible. [We note
that tensor networks where computing
h
Ψ
j
ˆ
O
j
Ψ
i
is efficient
but
Ψ
ð
n
Þ
is not are used in the form of the multi-scale
entanglement renormalization ansatz
[62]
]. Here we
describe one such application of TNFs without amplitude
dependent isometries.
It has previously been observed that common tensor
networks that can be efficiently exactly contracted for
Ψ
ð
n
Þ
and
h
Ψ
j
ˆ
O
j
Ψ
i
, i.e., MPS and tree tensor network
states, can be mapped onto efficient neural networks by
constructing polynomial size neural nets that emulate tensor
network contraction, establishing that exactly contractible
tensor networks are a formal subclass of efficient neural
quantum states
[36]
. It has also been suggested that neural
networks may not have an efficient tensor network repre-
sentation, because the well known contractible tensor net-
works satisfy area laws, while neural networks need not do
so
[36]
.
However, if one allows for general tensor network
geometries, it is straightforward to construct a TNF (with-
out using isometries) that represents a neural network
quantum state without the restriction of an area law. We
consider the explicit example of a feed forward neural
network (FNN), a popular universal architecture with
k
-
layers of neurons, where the outputs of each layer
f
y
ð
k
1
Þ
i
g
form the inputs
f
x
ð
k
Þ
i
g
to the next layer, and the
i
th neuron
(activation) function in layer
k
is
f
ð
k
Þ
i
ðf
x
ð
k
Þ
j
g
;
θ
ð
k
Þ
i
Þ
where
θ
ð
k
Þ
i
are the neuron parameters. Assuming that all
f
are
implemented without recursion, then the computational
graph of a FNN is a directed acyclic graph. Then we
observe that (i) if the FNN is efficient, the classical binary
circuit implementation of the neural network function is of
polynomial size in logic and copy gates, where a binary
logic gate is a tensor with two binary input legs and one
binary output leg, while copy takes one binary input and
returns two binary outputs (nonlinearity of the activation
functions enters through the copy operator) (ii) given the
input product state
j
n
i
(i.e., a bitstring), the application of
each gate
G
produces another product state
j
n
0
i
(i.e.,
another bitstring). Thus contracting the resulting tensor
network in the appropriate order from inputs to the outputs
(for example, using sparse tensor contraction, or by
introducing SVDs to discover the product state structure
of
j
n
0
i
, see Supplemental Material for more details
[43]
)
can always be done efficiently.
(a)
(b)
FIG. 4. (a)
ð
1
þ
1
Þ
D TN representation of the Floquet dynam-
ics of an
L
-site kicked Ising chain, starting from a product state
j
Ψ
ð
0
Þi¼j
0
i
L
. TNF contraction direction is either along the
spatial direction or inverse time directon (see inverse direction
results in Supplemental Material
[43]
). Conventional MPS-MPO
contraction is along the time direction. (b) Shows the dynamics
of the half-system entropy
S
ð
t
Þ
of a 14-site chain at
ð
J; h; g
Þ¼
ð
π
=
4
;
π
=
4
;
0
.
5
Þ
. Inset of (b) presents the entanglement spectrum
for the 40 largest eigenvalues of the reduced density matrix
at
t
¼
15
.
PHYSICAL REVIEW LETTERS
133,
260404 (2024)
260404-4
The above can be seen as equating a TNF with a type of
circuit computational model, in this case without a memory.
Although the tensor network contraction is of polynomial
cost, it is not very concise with respect to the number of
tensors, due to representing floating point numbers in
binary. In principle, we can introduce new elements to
the computational model that improve the conciseness. In
the Supplemental Material
[43]
, we show how to use the
arithmetic circuit tensor network construction of Ref.
[63]
to compute with tensor networks with floating point entries,
and the tensor network contraction is efficient if we further
introduce the ability to store results from contractions of
tensor network subgraphs.
Conclusions
In summary, we have introduced tensor
network functions (TNFs), a new perspective on the tensor
network state, as a flexible representation for quantum
many-body systems. The broadened perspective removes
the conventional restrictions of tensor network computa-
tions on the form of the ansatz, and the computational need
to converge the auxiliary bond dimension in an approxi-
mate contraction. On the other hand, TNFs inherit the
conceptual advantages of tensor networks. Unlike generic
function approximators, such as neural networks, the
function architecture is directly suggested by the physics
at hand and one can reuse standard tensor network
algorithms. For example, in time evolution, the TNF
structure can follow the evolution circuit and choice of
contraction path; in variational optimization, existing TN
algorithms provide a deterministic guess for the TNF
tensors, which may then be further relaxed. TNFs derived
from fermionic tensor networks naturally encode fermionic
sign structure
[56]
, avoiding the need for more complicated
constructions
[64
67]
. Finally, TNFs can be further gen-
eralized into a wider class of states, where nonlinearity
other than the singular value decomposition is introduced
into the tensor network computational graph. These pos-
sibilities open up many new avenues to investigate and to
explore with tensor networks for the future.
Acknowledgments
This work was supported by the U.S.
NationalScienceFoundationunderGrantNo.CHE-2102505.
G. K. C. acknowledges additional support from the Simons
Investigator program and the Dreyfus Foundation under the
program Machine Learning in the Chemical Sciences and
Engineering.
[1] S. R. White, Density matrix formulation for quantum
renormalization groups,
Phys. Rev. Lett.
69
, 2863 (1992)
.
[2] F. Verstraete, V. Murg, and J. I. Cirac, Matrix product states,
projected entangled pair states, and variational renormali-
zation group methods for quantum spin systems,
Adv. Phys.
57
, 143 (2008)
.
[3] U. Schollwöck, The density-matrix renormalization group
in the age of matrix product states,
Ann. Phys. (Amsterdam)
326
, 96 (2011)
.
[4] G. K.-L. Chan and S. Sharma, The density matrix renorm-
alization group in quantum chemistry,
Annu. Rev. Phys.
Chem.
62
, 465 (2011)
.
[5] J. I. Cirac, D. P ́
erez-García, N. Schuch, and F. Verstraete,
Matrix product states and projected entangled pair states:
Concepts, symmetries, theorems,
Rev. Mod. Phys.
93
,
045003 (2021)
.
[6] T. Xiang,
Density Matrix and Tensor Network Renormal-
ization
(Cambridge University Press, Cambridge, England,
2023).
[7] Z.-C. Gu, M. Levin, and X.-G. Wen, Tensor-entanglement
renormalization group approach as a unified method for
symmetry breaking and topological phase transitions,
Phys.
Rev. B
78
, 205116 (2008)
.
[8] N. Schuch, D. P ́
erez-García, and I. Cirac, Classifying
quantum phases using matrix product states and projected
entangled pair states,
Phys. Rev. B
84
, 165139 (2011)
.
[9] P. Corboz, T. M. Rice, and M. Troyer, Competing states in
the
t
J
model: Uniform
d
-wave state versus stripe state,
Phys. Rev. Lett.
113
, 046402 (2014)
.
[10] H. J. Liao, Z. Y. Xie, J. Chen, Z. Y. Liu, H. D. Xie, R. Z.
Huang, B. Normand, and T. Xiang, Gapless spin-liquid
ground state in the
S
¼
1
=
2
kagome antiferromagnet,
Phys.
Rev. Lett.
118
, 137202 (2017)
.
[11] B.-X. Zheng, C.-M. Chung, P. Corboz, G. Ehlers, M.-P. Qin,
R. M. Noack, H. Shi, S. R. White, S. Zhang, and G. K.-L.
Chan, Stripe order in the underdoped region of the two-
dimensional Hubbard model,
Science
358
, 1155 (2017)
.
[12] W.-Y. Liu, J. Hasik, S.-S. Gong, D. Poilblanc, W.-Q. Chen,
and Z.-C. Gu, Emergence of gapless quantum spin liquid
from deconfined quantum critical point,
Phys. Rev. X
12
,
031039 (2022)
.
[13] M. Levin and C. P. Nave, Tensor renormalization group
approach to two-dimensional classical lattice models,
Phys.
Rev. Lett.
99
, 120601 (2007)
.
[14] Z. Y. Xie, H. C. Jiang, Q. N. Chen, Z. Y. Weng, and T.
Xiang, Second renormalization of tensor-network states,
Phys. Rev. Lett.
103
, 160601 (2009)
.
[15] G. Evenbly and G. Vidal, Tensor network renormalization,
Phys. Rev. Lett.
115
, 180405 (2015)
.
[16] S. Yang, Z.-C. Gu, and X.-G. Wen, Loop optimization for
tensor network renormalization,
Phys. Rev. Lett.
118
,
110504 (2017)
.
[17] F. Verstraete and J. I. Cirac, Continuous matrix product
states for quantum fields,
Phys. Rev. Lett.
104
, 190405
(2010)
.
[18] J. Haegeman, T. J. Osborne, H. Verschelde, and F.
Verstraete, Entanglement renormalization for quantum
fields in real space,
Phys. Rev. Lett.
110
, 100402 (2013)
.
[19] L. Tagliacozzo, A. Celi, and M. Lewenstein, Tensor net-
works for lattice gauge theories with continuous groups,
Phys. Rev. X
4
, 041024 (2014)
.
[20] G. Vidal, Efficient classical simulation of slightly entangled
quantum computations,
Phys. Rev. Lett.
91
, 147902 (2003)
.
[21] I. L. Markov and Y. Shi, Simulating quantum computation
by contracting tensor networks,
SIAM J. Comput.
38
, 963
(2008)
.
[22] I. Arad and Z. Landau, Quantum computation and the
evaluation of tensor networks,
SIAM J. Comput.
39
, 3089
(2010)
.
PHYSICAL REVIEW LETTERS
133,
260404 (2024)
260404-5
[23] K. Boguslawski, K. H. Marti, O. Legeza, and M. Reiher,
Accurate
ab initio
spin densities,
J. Chem. Theory Comput.
8
, 1970 (2012)
.
[24] N. Nakatani and G. K.-L. Chan, Efficient tree tensor net-
work states for quantum chemistry: Generalizations of the
density matrix renormalization group algorithm,
J. Chem.
Phys.
138
, 134113 (2013)
.
[25] E. M. Stoudenmire and D. J. Schwab, Supervised learning
with tensor networks, Adv. Neural Inf. Process. Syst.
29
,
4799 (2016).
[26] Z.-Y. Han, J. Wang, H. Fan, L. Wang, and P. Zhang,
Unsupervised generative modeling using matrix product
states,
Phys. Rev. X
8
, 031012 (2018)
.
[27] F. Verstraete and J. I. Cirac, Renormalization algorithms for
quantum-many body systems in two and higher dimensions,
arXiv:cond-mat/0407066
.
[28] Y.-Y. Shi, L.-M. Duan, and G. Vidal, Classical simulation of
quantum many-body systems with a tree tensor network,
Phys. Rev. A
74
, 022320 (2006)
.
[29] S.-J. Ran, E. Tirrito, C. Peng, X. Chen, L. Tagliacozzo, G.
Su, and M. Lewenstein,
Tensor Network Contractions:
Methods and Applications to Quantum Many-Body Systems
(Springer Nature, London, England, 2020).
[30] H. J. Changlani, J. M. Kinder, C. J. Umrigar, and Garnet
Kin-Lic Chan, Approximating strongly correlated wave
functions with correlator product states,
Phys. Rev. B
80
,
245116 (2009)
.
[31] D.-L. Deng, X. Li, and S. Das Sarma, Quantum entangle-
ment in neural network states,
Phys. Rev. X
7
, 021021
(2017)
.
[32] J. Chen, S. Cheng, H. Xie, L. Wang, and T. Xiang,
Equivalence of restricted Boltzmann machines and tensor
network states,
Phys. Rev. B
97
, 085104 (2018)
.
[33] I. Glasser, N. Pancotti, M. August, I. D. Rodriguez, and J. I.
Cirac, Neural-network quantum states, string-bond states,
and chiral topological states,
Phys. Rev. X
8
, 011006
(2018)
.
[34] Y. Levine, O. Sharir, N. Cohen, and A. Shashua, Quantum
entanglement in deep learning architectures,
Phys. Rev.
Lett.
122
, 065301 (2019)
.
[35] L. Pastori, R. Kaubruegger, and J. C. Budich, Generalized
transfer matrix states from artificial neural networks,
Phys.
Rev. B
99
, 165123 (2019)
.
[36] O. Sharir, A. Shashua, and G. Carleo, Neural tensor
contractions and the expressive power of deep neural
quantum states,
Phys. Rev. B
106
, 205136 (2022)
.
[37] A. W. Sandvik and G. Vidal, Variational quantum
Monte Carlo simulations with tensor-network states,
Phys.
Rev. Lett.
99
, 220602 (2007)
.
[38] N. Schuch, M. M. Wolf, F. Verstraete, and J. I. Cirac,
Simulation of quantum many-body systems with strings
of operators and Monte Carlo tensor contractions,
Phys.
Rev. Lett.
100
, 040501 (2008)
.
[39] L. Wang, I. Pi
ž
orn, and F. Verstraete, Monte Carlo simu-
lation with tensor network states,
Phys. Rev. B
83
, 134421
(2011)
.
[40] W.-Y. Liu, S.-J. Dong, Y.-J. Han, G.-C. Guo, and L. He,
Gradient optimization of finite projected entangled pair
states,
Phys. Rev. B
95
, 195154 (2017)
.
[41] W.-Y. Liu, Y.-Z. Huang, S.-S. Gong, and Z.-C. Gu, Accurate
simulation for finite projected entangled pair states in two
dimensions,
Phys. Rev. B
103
, 235155 (2021)
.
[42] J. Gray and Garnet Kin-Lic Chan, Hyperoptimized approxi-
mate contraction of tensor networks with arbitrary geom-
etry,
Phys. Rev. X
14
, 011009 (2024)
.
[43] See Supplemental Material at
http://link.aps.org/
supplemental/10.1103/PhysRevLett.133.260404
including
Refs. [44
54], for additional information about the numeri-
cal simulations and tensor network function representation
of neural network computational graphs.
[44] S. Sorella, Green function Monte Carlo with stochastic
reconfiguration,
Phys. Rev. Lett.
80
, 4558 (1998)
.
[45] H. C. Jiang, Z. Y. Weng, and T. Xiang, Accurate determi-
nation of tensor network state of quantum lattice models in
two dimensions,
Phys. Rev. Lett.
101
, 090603 (2008)
.
[46] E. Neuscamman, C. J. Umrigar, and Garnet Kin-Lic Chan,
Optimizing large parameter sets in variational quantum
Monte Carlo,
Phys. Rev. B
85
, 045103 (2012)
.
[47] T. Vieijra, J. Haegeman, F. Verstraete, and L.
Vanderstraeten, Direct sampling of projected entangled-pair
states,
Phys. Rev. B
104
, 235141 (2021)
.
[48] S. Kumar,
Digital Vlsi Design
(Prentice Hall India Pvt.,
Limited, Indore, India, 2010).
[49] G. Lancaster,
Excel HSC Softw Design&Devel + Cards SG
,
Excel HSC (Pascal Press, Leichhardt, Australia, 2001).
[50] M. Mano,
Digital Logic and Computer Design
(Pearson
India, Chennai, India, 2017).
[51] T.-K. Liu, K. Hohulin, L.-E. Shiau, and S. Muroga, Optimal
one-bit full adders with different types of gates,
IEEE Trans.
Comput.
C-23
, 63 (1974)
.
[52] C. Baugh and B. Wooley, A two
s complement parallel array
multiplication algorithm,
IEEE Trans. Comput.
C-22
, 1045
(1973)
.
[53] M. Hatamian and G. Cash, A 70-MHz 8-bit/spl times/8-bit
parallel pipelined multiplier in 2.5-/spl mu/m cmos,
IEEE J.
Solid-State Circuits
21
, 505 (1986)
.
[54] Y. Leong, H. Lo, M. Drieberg, A. B. Sayuti, and P.
Sebastian, Performance comparison review of 8
3 com-
pressor on FPGA, in
TENCON 2017
2017 IEEE Region 10
Conference
(2017), pp. 2462
2465.
[55] W.-Y. Liu, S.-S. Gong, Y.-B. Li, D. Poilblanc, W.-Q. Chen,
and Z.-C. Gu, Gapless quantum spin liquid and global phase
diagram of the spin-
1
=
2
J
1
-
J
2
square antiferromagnetic
Heisenberg model,
Sci. Bull.
67
, 1034 (2022)
.
[56] C. V. Kraus, N. Schuch, F. Verstraete, and J. I. Cirac,
Fermionic projected entangled pair states,
Phys. Rev. A
81
, 052338 (2010)
.
[57] S. Lee, J. Lee, H. Zhai, Y. Tong, A. M. Dalzell, A. Kumar, P.
Helms, J. Gray, Z.-H. Cui, W. Liu
et al.
, Evaluating the
evidence for exponential quantum advantage in ground-state
quantum chemistry,
Nat. Commun.
14
, 1952 (2023)
.
[58] T. Prosen, General relation between quantum ergodicity and
fidelity of quantum dynamics,
Phys. Rev. E
65
, 036208
(2002)
.
[59] T. Prosen, Ruelle resonances in quantum many-body
dynamics,
J. Phys. A
35
, L737 (2002)
.
[60] T. Prosen, Chaos and complexity of quantum motion,
J.
Phys. A
40
, 7881 (2007)
.
PHYSICAL REVIEW LETTERS
133,
260404 (2024)
260404-6
[61] B. Bertini, P. Kos, and T. Prosen, Entanglement spreading in
a minimal model of maximal many-body quantum chaos,
Phys. Rev. X
9
, 021033 (2019)
.
[62] G. Vidal, Entanglement renormalization,
Phys. Rev. Lett.
99
, 220405 (2007)
.
[63] R. Peng, J. Gray, and Garnet Kin-Lic Chan, Arithmetic
circuit tensor networks, multivariable function representa-
tion, and high-dimensional integration,
Phys. Rev. Res.
5
,
013156 (2023)
.
[64] Y. Nomura, A. S. Darmawan, Y. Yamaji, and M. Imada,
Restricted Boltzmann machine learning for solving strongly
correlated quantum systems,
Phys. Rev. B
96
, 205152
(2017)
.
[65] J. R. Moreno, G. Carleo, A. Georges, and J. Stokes, Fermionic
wave functions from neural-network constrained hidden
states,
Proc.Natl.Acad.Sci.U.S.A.
119
,e2122059119(2022)
.
[66] J. Nys and G. Carleo, Variational solutions to fermion-to-qubit
mappings in two spatial dimensions,
Quantum
6
, 833 (2022)
.
[67] Z. Liu and B. K. Clark, A unifying view of fermionic neural
network quantum states: From neural network backflow to
hidden fermion determinant states,
Phys. Rev. B
110
,
115124 (2024)
.
PHYSICAL REVIEW LETTERS
133,
260404 (2024)
260404-7