Chapter 7 Cluster quantum Monte Carlo algorithms for...

22
Chapter 7 Cluster quantum Monte Carlo algorithms for lattice models 7.1 World line representations for quantum lattice models The basic problem for Monte Carlo simulations of quantum systems is that the partition function is not a simple sum over classical configurations but an operator expression Z = Tr exp(-β H), (7.1) where H is the Hamilton operator and the trace Tr goes over all states in the Hilbert space. Similarly the expression for an observable like the magnetization m is an operator expression: m = 1 Z Tr [m exp(-β H)] , (7.2) and the Monte Carlo method cannot directly be applied except in the classical case where the Hamilton operator H is diagonal. The first step of any QMC algorithm is thus the mapping of the quantum system to an equivalent “classical” system Z = Tr exp(-β H)= c w(c) (7.3) m = 1 Z Tr [m exp(-β H)] = 1 Z c m(c)w(c)= c m(c)P (c), (7.4) where P (c)= w(c)/Z and the sum goes over configurations c in an artificial classical system, which here will be a system of world lines. As an introduction to the three common world line representations discrete time path integrals continuous time path integrals stochastic series expansion we will first consider the special case of a single quantum mechanical spin-1/2 in a transverse and longitudinal magnetic field, before discussing the general representations in the following sections. 55

Transcript of Chapter 7 Cluster quantum Monte Carlo algorithms for...

Page 1: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

Chapter 7

Cluster quantum Monte Carloalgorithms for lattice models

7.1 World line representations for quantum latticemodels

The basic problem for Monte Carlo simulations of quantum systems is that the partitionfunction is not a simple sum over classical configurations but an operator expression

Z = Tr exp(!!H), (7.1)

where H is the Hamilton operator and the trace Tr goes over all states in the Hilbertspace. Similarly the expression for an observable like the magnetizationm is an operatorexpression:

"m# = 1

ZTr [m exp(!!H)] , (7.2)

and the Monte Carlo method cannot directly be applied except in the classical casewhere the Hamilton operator H is diagonal. The first step of any QMC algorithm isthus the mapping of the quantum system to an equivalent “classical” system

Z = Tr exp(!!H) =!

c

w(c) (7.3)

"m# =1

ZTr [m exp(!!H)] =

1

Z

!

c

m(c)w(c) =!

c

m(c)P (c), (7.4)

where P (c) = w(c)/Z and the sum goes over configurations c in an artificial classicalsystem, which here will be a system of world lines.

As an introduction to the three common world line representations

• discrete time path integrals

• continuous time path integrals

• stochastic series expansion

we will first consider the special case of a single quantum mechanical spin-1/2 in atransverse and longitudinal magnetic field, before discussing the general representationsin the following sections.

55

Page 2: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

7.2 A Spin-1/2 in a Magnetic Field

The Hamilton operator for a single quantum mechanical spin-1/2 in a longitudinalmagnetic field h and transverse magnetic field ! is given by

H = Hz +Hx = !hSz ! !Sx. (7.5)

Using a basis set in which the z-component of the spin operator is diagonal

{|1/2", |! 1/2"} (7.6)

the spin-1/2 operators Sx and Sz are given by the Pauli matrices

Sx =!

2!x =

!

0 !/2!/2 0

"

(7.7)

Sz =!

2!z =

!

!/2 00 !!/2

"

. (7.8)

Setting ! = 1 from now on, the Hamilton operator in this basis becomes

H =

!

!h/2 !!/2!!/2 h/2

"

. (7.9)

7.2.1 Discrete Time Path Integrals

The first representation is a path integral using discrete imaginary time steps "! . Sincewe will later take the"! # 0 limit to arrive at a continuous time formulation we will justuse a lowest-order approximation that will better illustrate the connection to continuoustime path integrals.

We first divide the inverse temperature " into M time steps "! = "/M , arriving ata discrete time approximation of the quantum transfer matrix exp(!"!H):

exp(!"!H) = U +O("2!), (7.10)

where we have introduced the quantum transfer matrix

U = 1!"!H =

!

1 +"!h/2 "!!/2"!!/2 1!"!h/2

"

. (7.11)

Since the Hamilton matrix H will in general be hard to exponentiate, we have approx-imated the exponential exp(!"!H) by a lowest-order Taylor expansion. In an actualdiscrete time code a Suzuki-Trotter decomposition might be used instead.

In the discrete time path integral formulation, we evaluate the operator exponentialin equation (7.1) by writing the trace as a sum over all basis states |i". Inserting Msums over a complete set of states we obtain

Z = Tr exp(!"H) = Tr exp(!"!H)M = Tr#

U +O("2! )$M

=%

(i1,...,iM )

$i1|U |i2"$i2|U |i3" · · · $iM |U |i1"+O("! ). (7.12)

56

Page 3: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

0βτ1 τ2 τ3 τ4

!"

!"

#"

$"

Figure 7.1: Relationship between discrete and continuous time path integrals illustratedfor a single spin-1/2. From a), over b) to c) the time step !! is reduced, approachingthe continuous time limit in d). While the discrete time representations a)-c) theconfiguration can be stored by specifying the value of the spin at each time step, thecontinuous time representation d) is stored by specifying the locations !i of the domainwalls and the values of the spin in each domain.

This expression is identical to the partition function of a one-dimensional chain ofclassical Ising spins "i = ±1 of length M at an inverse temperature #cl

Hcl = !Jcl

M!

i=1

"i"i+1 ! hcl

M!

i=1

"i + E0 (7.13)

with periodic boundary conditions "M+1 = "1, and the following relation between theclassical and quantum mechanical coupling constants:

#clJcl = !1

2log(!!"/2) (7.14)

#clhcl = log(1 + !!h/2) " !!h/2 (7.15)

#clE0 = M#clJcl. (7.16)

Hence, any Monte Carlo algorithm for the classical Ising model, including clusterupdates, can be used for the simulation. Similarly any d-dimensional quantum Isingmodel

H = !J!

!i,j"

Szi S

zj ! h

!

i

Szi ! "

!

i

Sxi (7.17)

can be mapped to a (d+ 1)-dimensional classical Ising model.

7.2.2 Continuous Time Path Integrals

Instead of using a discrete time path integral formulation, modern world line QMCalgorithms are based on a continuous time representation, which works directly in thelimit !! # 0 (M # $). To arrive at a continuous time representation we considerwhat happens when we reduce the time step !! , as is illustrated in figure 7.1.

The discrete time path integral is represented by a chain of classical Ising spins,shown in Fig. 7.1a). Reducing the time step !! increases the number of spins as

57

Page 4: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

shown in Fig. 7.1b,c). While the number of spins M diverges in the limit !! ! 0, theaverage number of domain walls between parallel spins remains finite since the couplingbetween two neighboring spins (7.14) becomes infinitely strong as !! ! 0. Since theratio exp("2!clJcl) = !!"/2 of the weights of two antiparallel spins and two parallelspins12 probability of having a domain wall, the average number of domain walls is lessthan

M exp("2!clJcl) = M!!"/2 = !"/2. (7.18)

and hence finite in the !! ! 0 limit.

Since the number of domain walls n remains finite while the number of spins Mdiverges, it will be more e#cient to only store the values ij of the spins in the j-thdomain (j = 1, . . . n) and the location of the last spin of the j-th domain ij , insteadof storing the value of each single spin. In such a compacted notation the weight of aconfiguration in the discrete time representation (7.12) becomes

#i1|U |i2$#i2|U |i3$ · · · #iM |U |i1$= (1 + !!hi1)

i1!1#i1|U |i2$(1 + !!hi2)(i2!i1!1) · · · (7.19)

· · · (1 + !!hin)(in!in!1!1)#in|U |i1$(1 + !!hi1)

(M!in).

Identifying "j = ij!! and Ej = "hij, realizing that #ij |U |ij+1$ = "!! #ij |Hx|ij+1$when |ij$ %= |ij+1$, and taking the limit !! ! 0 we end up with the continuous timeexpression

("1)ne!!1E1#i1|Hx|i2$e!(!2!!1)E2 · · · (7.20)

· · · e!(!n!!n!1)En#in|Hx|i1$e!("!!1)E1d"1 · · · d"n.

Integrating this weight over all possible location for the domain walls and summingover all possible number of domain walls n, we finally obtain the continuous timerepresentation

Z =!

i0

#i0|e!"E0 |i0$+

"!

n=1

!

(i1,...in)

("1)n" "

0

d"1 · · ·" "

!n!1

d"n e!!1E1#i1|Hx|i2$e!(!2!!1)E2 · · ·

· · · e!(!n!!n!1)En#in|Hx|i1$e!("!!n)E1 . (7.21)

This is nothing but the result of a time-dependent perturbation expansion in the oper-ators Hx.

1To be replaced with the exact average of domain walls.2Any additional interaction, such as the longitudinal field h will only decrease the number of domain

walls.

58

Page 5: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

0 βτ1 τ2 τ3 τ4

i = 1 2 3 4 5 6 7 8 9 10 11 12

!"

#"

Figure 7.2: Relationship between continuous time path integrals and the stochasticseries expansion (SSE) representation illustrated for a single spin-1/2. a) in continuoustime path integrals a representation is specified by the times !i at which o!-diagonalsoperators act on the configuration. b) in the SSE representation the expansion is interms of both diagonal and o!-diagonal operators. An integer index specifying theordering of operators replaces the continuous time label.

7.2.3 Stochastic Series Expansion

The stochastic series expansion (SSE) representation performs a Taylor expansion ofthe full Hamiltonian:

Z = Tr exp(!"H) =!!

n=0

"n

n!Tr(!H)n

= 1 +!!

n=1

"n

n!

!

{i1,...in}

"i1|!H|i2#"i2|!H|i3# · · · "in|!H|i1#,(7.22)

giving a much simpler expression than discrete or continuous time path integrals.The advantage of the SSE representation over continuous time path integrals is that

instead of associating times !i with the operators there is just an integer index indicatingthe order of operators. This simplifies QMC programs based on an SSE representation,compared to continuous time programs. This advantage is paid for by the need forhigher orders n in the expansion since we perturb in all terms of the Hamiltonian andnot just in the o!-diagonal term Hx. Figure 7.2 illustrates the relationship betweencontinuous time path integrals and the SSE representation.

7.3 More complicated models: the XXZ chain

We will next discuss the loop cluster algorithm for a spin-1/2 quantum XXZ modelwith the Hamiltonian

H = !!

"i,j#

"

JzSzi S

zj + Jxy(S

xi S

xj + Sy

i Syj )#

= !!

"i,j#

$

JzSzi S

zj +

Jxy

2(S+

i S$j + S$

i S+j )

%

. (7.23)

For J $ Jz = Jxy we have the Heisenberg model (J > 0 is ferromagnetic, J < 0antiferromagnetic). Jxy = 0 is the (classical) Ising model and Jz = 0 the quantum XYmodel.

59

Page 6: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

space

imag

inar

y tim

e

0

β

Figure 7.3: Example of a world line configuration for a spin-1/2 quantum Heisenbergmodel. Drawn are the world lines for up-spins only. Down spin world lines occupy therest of the configuration.

Continuous-time world lines

We again start by still discretizing discretize the imaginary time (inverse temperature)direction and subdivide ! = M!" :

e!!H =!

e!!"H"M

= (1! !"H)M +O(!") (7.24)

In the limit M " # (!" " 0) this becomes exact. We will take the limit later, butstay at finite !" for the derivation.

The next step is to insert the identity matrix, represented by a sum over all basisstates

#

i |i$%i| between all operators (1! !"H):

Z = Tre!!H = Tr (1! !"H)M +O(!")

=$

i1,...,iM

%i1|1! !"H|i2$%i2|1! !"H|i3$ · · · %iM |1! !"H|i1$+O(!")

=: Pi1,...,iM (7.25)

and similarly for the measurement, obtaining

%A$ =$

i1,...,iM

%i1|A(1! !"H)|i2$%i1|1! !"H|i2$

Pi1,...,iM + O(!"). (7.26)

If we choose the basis states |i$ to be eigenstates of the local Sz operators we endup with an Ising-like spin system in one higher dimension. Each choice i1, . . . , iMcorresponds to one of the possible configurations of this classical spin system. Thetrace is mapped to periodic boundary conditions in the imaginary time direction of thisclassical spin system. The probabilities are given by matrix elements %in|1!!"H|in+1$.We can now sample this classical system using classical Monte Carlo methods.

However, most of the matrix elements %in|1!!"H|in+1$ are zero, and thus nearly allconfigurations have vanishing weight. The only non-zero configurations are those where

60

Page 7: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

neighboring states |in! and |in+1! are either equal or di!er by one of the o!-diagonalmatrix elements in H , which are nearest neighbor exchanges by two opposite spins. Wecan thus uniquely connect spins on neighboring “time slices” and end up with worldlines of the spins, sketched in Fig. 7.3. Instead of sampling over all configurations oflocal spins we thus have to sample only over all world line configurations (the othershave vanishing weight). Our update moves are not allowed to break world lines buthave to lead to new valid world line configurations.

Finally we take the continuous time limit "! " 0. Instead of storing the configura-tions at all M " # time steps, we will not just store the times ! where a spin flips asthe consequence of an o!-diagonal operator acting at that time. The number of suchevents stays finite as M " #: as can be seen from equation (7.24) the probability ofH acting at a given time is proportional to 1/M , and hence the total number of spinflips will stay finite although M " #

7.4 Cluster updates

Before discussing cluster updates for quantum systems, we will review the cluster al-gorithms for the classical Ising model which should be known from the computationalstatistical physics course.

7.4.1 Kandel-Domany framework

To provide a general framework, which can be extended to quantum systems, we usethe Fortuin-Kastelyn representation of the Ising model, as generalized by Kandel andDomany. The phase space of the Ising model is enlarged by assigning a set G of possible“graphs” to each configuration C in the set of configurations C. We write the partitionfunction as

Z =!

C!C

!

G!G

W (C,G) (7.27)

where the new weights W (C,G) > 0 are chosen such that Z is the partition function ofthe original model by requiring

!

G!G

W (C,G) = W (C) := exp($"E[C]), (7.28)

where E[C] is the energy of the configuration C.The algorithm now proceeds as follows. First we assign a graph G % G to the

configuration C, chosen with the correct probability

PC(G) = W (C,G)/W (C). (7.29)

Then we choose a new configuration C " with probability p[(C,G) " (C ", G)], keepingthe graph G fixed; next a new graph G" is chosen

C " (C,G) " (C ", G) " C " " (C ", G") " . . . (7.30)

61

Page 8: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

What about detailed balance? The procedure for choosing graphs with probabilitiesPG obeys detailed balance trivially. The non-trivial part is the probability of choosinga new configuration C !. There detailed balance requires:

W (C,G)p[(C,G) ! (C !, G)] = W (C !, G)p[(C !, G) ! (C,G)], (7.31)

which can be fulfilled using either the heat bath algorithm

p[(C,G) ! (C !, G)] =W (C !, G)

W (C,G) +W (C !, G)(7.32)

or by again using the Metropolis algorithm:

p[(C,G) ! (C !, G)] = min(W (C !, G)/W (C,G), 1) (7.33)

The algorithm simplifies a lot if we can find a graph mapping such that the graphweights do not depend on the configuration whenever it is nonzero in that configuration.This means, we want the graph weights to be

W (C,G) = !(C,G)V (G), (7.34)

where

!(C,G) :=

!

1 if W (C,G) "= 0,0 otherwise.

. (7.35)

Then equation (7.32) simply becomes p = 1/2 and equation (7.33) reduces to p = 1 forany configuration C ! with W (C !, G) "= 0.

7.4.2 The cluster algorithms for the Ising model

Let us now show how this abstract and general algorithm can be applied to the Isingmodel. Our graphs will be bond-percolation graphs on the lattice. Spins pointinginto the same direction can be connected or disconnected. Spins pointing in oppositedirections will always be disconnected. In the Ising model we can write the weightsW (C) and W (C,G) as products over all bonds b:

W (C) ="

b

w(Cb) (7.36)

W (C,G) ="

b

w(Cb, Gb) ="

b

!(Cb, Gb)V (Gb) (7.37)

where the local bond configurations Cb can be one of {##, $#, #$, $$}and the local graphs can be “connected” or “disconnected”. The graph selection

can thus be done locally on each bond.Table 7.1 shows the local bond weights w(c, g), w(c), !(c, g) and V (g). It can easily

be checked that the sum rule (7.28) is satisfied.The probability of a connected bond is [exp(!J) % exp(%!J)]/ exp(!J) = 1 %

exp(%2!J) if two spins are aligned and zero otherwise. These connected bonds groupthe spins into clusters of aligned spins.

62

Page 9: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

Table 7.1: Local bond weights for the Kandel-Domany representation of the Ising model.

c =!! c ="! c =!" c ="" V(g)!(c, discon.) 1 1 1 1 exp(#!J)!(c, con.) 1 0 0 1 exp(!J)# exp(#!J)w(c) exp(!J) exp(#!J) exp(#!J) exp(!J)

A new configuration C ! with the same graph G can di"er from C only by flippingclusters of connected spins. Thus the name “cluster algorithms”. The clusters can beflipped independently, as the flipping probabilities p[(C,G) $ (C !, G)] are configurationindependent constants.

There are two variants of cluster algorithms that can be constructed using the rulesderived above.

The Swendsen-Wang algorithm

The Swendsen-Wang or multi-cluster algorithm proceeds as follows:

i) Each bond in the lattice is assigned a label “connected” or “disconnected” ac-cording to above rules. Two aligned spins are connected with probability 1 #exp(#2!J). Two antiparallel spins are never connected.

ii) Next a cluster labeling algorithm, like the Hoshen-Kopelman algorithm is used toidentify clusters of connected spins.

iii) Measurements are performed, using improved estimators discussed in the nextsection.

iv) Each cluster of spins is flipped with probability 1/2.

The Wol! single-cluster algorithm

The Swendsen Wang algorithm gets less e#cient in dimensions higher than two as themajority of the clusters will be very small ones, and only a few large clusters exist.The Wol" algorithm is similar to the Swendsen-Wang algorithm but builds only onecluster starting from a randomly chosen point. As the probability of this point beingon a cluster of size s is proportional to s the Wol" algorithm builds preferedly largerclusters. It works in the following way:

i) Choose a random spin as the initial cluster.

ii) If a neighboring spin is parallel to the initial spin it will be added to the clusterwith probability 1# exp(#2!J).

iii) Repeat step ii) for all points newly added to the cluster and repeat this procedureuntil no new points can be added.

63

Page 10: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

iv) Perform measurements using improved estimators.

v) Flip all spins in the cluster.

We will see in the next section that the linear cluster size diverges with the corre-lation length ! and that the average number of spins in a cluster is just "T . Thus thealgorithm adapts optimally to the physics of the system and the dynamical exponentz ! 0, thus solving the problem of critical slowing down. Close to criticality thesealgorithms are many orders of magnitudes (a factor L2) better than the local updatemethods. Away from criticality sometimes a hybrid method, mixing cluster updatesand local updates can be the ideal method.

7.4.3 Improved Estimators

In this section we present a neat trick that can be used in conjunction with clusteralgorithms to reduce the variance, and thus the statistical error of Monte Carlo mea-surements. Not only do these “improved estimators” reduce the variance. They arealso much easier to calculate than the usual “simple estimators”.

To derive them we consider the Swendsen-Wang algorithm. This algorithm dividesthe lattice into Nc clusters, where all spins within a cluster are aligned. The nextpossible configuration is any of the 2Nc configurations that can be reached by flippingany subset of the clusters. The idea behind the “improved estimators” is to measurenot only in the new configuration but in all equally probable 2Nc configurations.

As simplest example we consider the average magnetization "m#. We can measureit as the expectation value "#!i# of a single spin. As the cluster to which the spin belongscan be freely flipped, and the flipped cluster has the same probability as the originalone, the improved estimator is

"m# = "12(#!i $ #!i)# = 0. (7.38)

This result is obvious because of symmetry, but we saw that at low temperatures asingle spin flip algorithm will fail to give this correct result since it takes an enormoustime to flip all spins. Thus it is encouraging that the cluster algorithms automaticallygive the exact result in this case.

Correlation functions are not much harder to measure:

"#!i#!j# =!

1 if $i und $j are on the same cluster0 otherwise

(7.39)

To derive this result consider the two cases and write down the improved estimators byconsidering all possible cluster flips.

Using this simple result for the correlation functions the mean square of the mag-netization is

"m2# = 1

N2

"

!i,!j

"#!i#!j# =1

N2""

cluster

S(cluster)2#, (7.40)

where S(cluster) is the number of spins in a cluster. The susceptibility above Tc issimply given by %"m2# and can also easily be calculated by above sum over the squaresof the cluster sizes.

64

Page 11: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

Table 7.2: The six local configurations for an XXZ model and their weights.

configuration weight

Si(τ)

Si(τ+dτ)

Sj(τ)

Sj(τ+dτ)

, Si(τ)

Si(τ+dτ)

Sj(τ)

Sj(τ+dτ)

1 + Jz4 d!

Si(τ)

Si(τ+dτ)

Sj(τ)

Sj(τ+dτ)

, Si(τ)

Si(τ+dτ)

Sj(τ)

Sj(τ+dτ)

1! Jz4 d!

Si(τ)

Si(τ+dτ)

Sj(τ)

Sj(τ+dτ)

, Si(τ)

Si(τ+dτ)

Sj(τ)

Sj(τ+dτ)

Jxy2 d!

In the Wol! algorithm only a single cluster is built. Above sum (7.40) can berewritten to be useful also in case of the Wol! algorithm:

"m2# =1

N2"!

cluster

S(cluster)2#

=1

N2

!

!i

1

S(cluster containing "i)S(cluster containing "i)2

=1

N2

!

!i

S(cluster containing "i) =1

N"S(cluster)#. (7.41)

The expectation value for m2 is thus simply the mean cluster size. In this derivationwe replaced the sum over all clusters by a sum over all sites and had to divide thecontribution of each cluster by the number of sites in the cluster. Next we can replacethe average over all lattice sites by the expectation value for the cluster on a randomlychosen site, which in the Wol! algorithm will be just the one Wol! cluster we build.

7.4.4 The loop algorithm for quantum spins

We will now generalize these cluster algorithms to quantum systems and present theloop algorithm. 3

This algorithm is best described by first taking the continuous time limit M $ %("! $ d!) and by working with infinitesimals. Similar to the Ising model we look attwo spins on neigboring sites i and j at two neighboring times ! and !+d! , as sketchedin Tab. 7.2. There are a total of six possible configurations, having three di!erentprobabilities. The total probabilities are the products of all local probabilities, like inthe classical case. This is obvious for di!erent time slices. For the same time slice itis also true since, denoting by Hij the term in the Hamiltonian H acting on the bondbetween sites i and j we have

"

!i,j"(1 ! d!Hij) = 1 ! d!#

!i,j"Hij = 1 ! d!H . Inthe following we focus only on such local four-spin plaquettes. Next we again use the

3H. G. Evertz et al., Phys. Rev. Lett. 70, 875 (1993); B. B. Beard and U.-J. Wiese, Phys. Rev.Lett. 77, 5130 (1996); B. Ammon, H. G. Evertz, N. Kawashima, M. Troyer and B. Frischmuth, Phys.Rev. B 58, 4304 (1998).

65

Page 12: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

a) , b) , c) , d)

Figure 7.4: The four local graphs: a) vertical, b) horizontal c) crossing and d) freezing(connects all four corners).

Table 7.3: The graph weights for the quantum-XY model and the ! function specifyingwhether the graph is allowed. The dash – denotes a graph that is not possible for aconfiguration because of spin conservation and has to be zero.

G !( , G) !( , G) !( , G)

= !( , G) = !( , G) = !( , G) graph weight

1 1 – 1! Jxy4 d!

– 1 1 Jxy4 d!

1 – 1 Jxy4 d!

0 0 0 0

total weight 1 1 Jxy2 d!

Kandel-Domany framework and assign graphs. As the updates are not allowed to breakworld lines only four graphs, sketched in Fig. 7.4 are allowed. Finally we have to find! functions and graph weights that give the correct probabilities. The solution for theXY -model, ferromagnetic and antiferromagnetic Heisenberg model and the Ising modelis shown in Tables 7.3 - 7.6.

Let us first look at the special case of the Ising model. As the exchange term isabsent in the Ising model all world lines run straight and can be replaced by classicalspins. The only non-trivial graph is the “freezing”, connecting two neighboring worldlines. Integrating the probability that two neighboring sites are nowhere connectedalong the time direction we obtain: times:

!!

"=0

(1! d!J/2) = limM!"

(1!!!J/2)M = exp(!"J/2) (7.42)

Taking into account that the spin is S = 1/2 and the corresponding classical cou-pling Jcl = S2J = J/4 we find for the probability that two spins are connected:1 ! exp(!2"Jcl). We end up exactly with the cluster algorithm for the classical Isingmodel!

66

Page 13: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

Table 7.4: The graph weights for the ferromagnetic quantum Heisenberg model and the! function specifying whether the graph is allowed. The dash – denotes a graph thatis not possible for a configuration because of spin conservation and has to be zero.

G !( , G) !( , G) !( , G)

= !( , G) = !( , G) = !( , G) graph weight

1 1 – 1! J4d!

– 0 0 0

1 – 1 J2d!

0 0 0 0

total weight 1 + J4d! 1! J

4d!J2d!

Table 7.5: The graph weights for the antiferromagnetic quantum Heisenberg model andthe ! function specifying whether the graph is allowed. The dash – denotes a graphthat is not possible for a configuration because of spin conservation and has to be zero.To avoid the sign problem (see next subsection) we change the sign of Jxy, which isallowed only on bipartite lattices.

G !( , G) !( , G) !( , G)

= !( , G) = !( , G) = !( , G) graph weight

1 1 – 1! |J |4 d!

– 1 1 |J |2 d!

0 – 0 0

0 0 0 0

total weight 1! |J |4 d! 1 + |J |

4 d!|J |2 d!

67

Page 14: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

Table 7.6: The graph weights for the ferromagnetic Ising model and the ! functionspecifying whether the graph is allowed. The dash – denotes a graph that is not possiblefor a configuration because of spin conservation and has to be zero.

G !( , G) !( , G) !( , G)

= !( , G) = !( , G) = !( , G) graph weight

1 1 – 1! Jz4 d!

– 0 0 0

0 – 0 0

1 0 0 Jz2 d!

total weight 1 + Jz4 d! 1! Jz

4 d! 0

The other cases are special. Here each graph connects two spins. As each of thesespins is again connected to only one other, all spins connected by a cluster form aclosed loop, hence the name “loop algorithm”. Only one issue remains to be explained:how do we assign a horizontal or crossing graph with infinitesimal probability, suchas (J/2)d! . This is easily done by comparing the assignment process with radioactivedecay. For each segment the graph runs vertical, except for occasional decay processesoccuring with probability (J/2)d! . Instead of asking at every infinitesimal time stepwhether a decay occurs we simply calculate an exponentially distributed decay time tusing an exponential distribution with decay constant J/2. Looking up the equationin the lecture notes of the winter semester we have t = !(2/J) ln(1 ! u) where u is auniformly distributed random number.

The algorithm now proceeds as follows (see Fig. 7.5): for each bond we start at time0 and calculate a decay time. If the spins at that time are oriented properly and anexchange graph is possible we insert one. Next we advance by another randomly chosendecay time along the same bond and repeat the procedure until we have reached theextent ". This assigns graphs to all infinitesimal time steps where spins do not change.Next we assign a graph to all of the (finite number of) time steps where two spinsare exchanged. In the case of the Heisenberg models there is always only one possiblegraph to assign and this is very easy. In the next step we identify the loop-clusters andthen flip them each with probability 1/2. Alternatively a Wol"-like algorithm can beconstructed that only builds one loop-cluster.

Improved estimators for measurements can be constructed like in classical models.The derivation is similar to the classical models. I will just mention two simple ones

68

Page 15: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

world lines world lines + decay graphs

world linesafter flips of some loop clusters

Figure 7.5: Example of a loop update. In a first step decay paths are inserted wherepossible at positions drawn randomly according to an exponential distribution andgraphs are assigned to all exchange terms (hoppings of world lines). In a second stage(not shown) the loop clusters are identified. Finally each loop cluster is flipped withprobability 1/2 and one ends up with a new configuration.

for the ferromagnetic Heisenberg model. The spin-spin corelation is

Szi (!)S

zj (!

!) =

!

1 if (i, !) und (j, ! !) are on the same cluster0 otherwise

(7.43)

and the uniform susceptibilty is

" =1

N#

"

c

S(c)2, (7.44)

where the sum goes over all loop clusters and S(c) is the length of all the loop segmentsin the loop cluster c.

7.5 The negative sign problem

Now that we have an algorithm with no critical slowing down we could think that wehave completely solved the problem of quantum many body problems. However, in thissection we will show that the sign problem is NP-hard, following the paper M. Troyerand U.J. Wiese, Phys. Rev. Lett. 94, 170201 (2005).

The di!culties in finding polynomial time solutions to the sign problem are rem-iniscent of the apparent impossibility to find polynomial time algorithms for nonde-terministic polynomial (NP)-complete decision problems, which could be solved inpolynomial time on a hypothetical non-deterministic machine, but for which no poly-nomial time algorithm is known for deterministic classical computers. A hypotheticalnon-deterministic machine can always follow both branches of an if-statement simul-taneously, but can never merge the branches again. It can, equivalently, be viewed ashaving exponentially many processors, but without any communication between them.

69

Page 16: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

In addition, it must be possible to check a positive answer to a problem in NP on aclassical computer in polynomial time.

Many important computational problems in the complexity class NP, including thetraveling salesman problem and the problem of finding ground states of spin glasseshave the additional property of being NP-hard, forming the subset of NP-completeproblems, the hardest problems in NP. A problem is called NP-hard if any problem inNP can be mapped onto it with polynomial complexity. Solving an NP-hard problemis thus equivalent to solving any problem in NP, and finding a polynomial time solutionto any of them would have important consequences for all of computing as well as thesecurity of classical encryption schemes. In that case all problems in NP could besolved in polynomial time, and hence NP=P.

As no polynomial solution to any of the NP-complete problems was found despitedecades of intensive research, it is generally believed that NP!=P and no deterministicpolynomial time algorithm exists for these problems. The proof of this conjectureremains as one of the unsolved millennium problems of mathematics for which the ClayMathematics Institute has o!ered a prize of one million US$ . In this section we willshow that the sign problem is NP-hard, implying that unless the NP !=P conjectureis disproved there exists no generic solution of the sign problem.

Before presenting the details of our proof, we will give a short introduction to clas-sical and quantum Monte Carlo simulations and the origin of the sign problem. In thecalculation of the phase space average of a quantity A, instead of directly evaluatingthe sum

"A# = 1

Z

!

c!!

A(c)p(c) , Z =!

c!!

p(c), (7.45)

over a high-dimensional space " of configurations c, a classical Monte Carlo methodchooses a set of M configurations {ci} from ", according to the distribution p(ci). Theaverage is then approximated by the sample mean

"A# $ A =1

M

M!

i=1

A(ci), (7.46)

within a statistical error #A ="

VarA(2!A + 1)/M , where VarA is the variance of Aand the integrated autocorrelation time !A is a measure of the autocorrelations of thesequence {A(ci)}. In typical statistical physics applications, p(c) = exp(%"E(c)) is theBoltzmann weight, " = 1/kBT is the inverse temperature, and E(c) is the energy ofthe configuration c.

Since the dimension of configuration space " grows linearly with the number Nof particles, the computational e!ort for the direct integration Eq. (7.45) scales ex-ponentially with the particle number N . Using the Monte Carlo approach the sameaverage can be estimated to any desired accuracy in polynomial time, as long as theautocorrelation time !A does not increase faster than polynomially with N .

In a quantum system with Hamilton operator H , instead of an integral like Eq.(7.45), an operator expression

"A# = 1

ZTr[A exp(%"H)] , Z = Tr exp(%"H) (7.47)

70

Page 17: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

|i1>

|i2>

|i3>

|i4>

|i1>

Figure 7.6: A configuration of a fermionic lattice model on a 4-site square. Theconfiguration has negative weight, since two fermions are exchanged in the sequence|i1! " |i2! " |i3! " |i4! " |i1!. World lines connecting particles on neighboring slicesare drawn as thick lines.

needs to be evaluated in order to calculate the thermal average of the observable A(represented by a self-adjoint operator). Monte Carlo techniques can again be appliedto reduce the exponential scaling of the problem, but only after mapping the quantummodel to a classical one, for example using world line configurations c with weights p(c).If all the weights p(c) are positive, standard Monte Carlo methods can be applied, asit is the case for non-frustrated quantum magnets and bosonic systems. In fermionicsystems negative weights p(c) < 0 arise from the Pauli exclusion principle, when alongthe sequence |i1! " |i2! " · · · " |in! " |i1! two fermions are exchanged, as shown inFig. 7.6.

The standard way of dealing with the negative weights of the fermionic system is tosample with respect to the bosonic system by using the absolute values of the weights|p(c)| and to assign the sign s(c) # sign p(c) to the quantity being sampled:

$A! =

!

cA(c)p(c)!

c p(c)(7.48)

=

!

cA(c)s(c)|p(c)| /!

c |p(c)|!

c s(c)|p(c)| /!

c |p(c)|# $As!!

$s!! .

While this allows Monte Carlo simulations to be performed, the errors increase ex-ponentially with the particle number N and the inverse temperature !. To see this,consider the mean value of the sign $s! = Z/Z !, which is just the ratio of the partitionfunctions of the fermionic system Z =

!

c p(c) with weights p(c) and the bosonic sys-tem used for sampling with Z ! =

!

c |p(c)|. As the partition functions are exponentialsof the corresponding free energies, this ratio is an exponential of the di!erences "f

71

Page 18: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

in the free energy densities:!s" = Z/Z ! = exp(#!N!f). As a consequence, the rela-tive error !s/!s" increases exponentially with increasing particle number and inversetemperature:

!s

!s" =

!

(!s2" # !s"2) /M!s" =

!

1# !s"2$M!s"

% e!N!f

$M

. (7.49)

Similarly the error for the numerator in Eq. (7) increases exponentially and the timeneeded to achieve a given relative error scales exponentially in N and !.

In order to avoid any misconception about what would constitute a “solution” ofthe sign problem, we start by giving a precise definition:

• A quantum Monte Carlo simulation to calculate a thermal average !A" of anobservable A in a quantum system with Hamilton operator H is defined to su"erfrom a sign problem if there occur negative weights p(c) < 0 in the classicalrepresentation.

• The related bosonic system of a fermionic quantum system is defined as the systemwhere the weights p(c) are replaced by their absolute values |p(c)|, thus ignoringthe minus sign coming from fermion exchanges:

!A"! = 1

Z !

"

c

A(c)|p(c)|. (7.50)

• An algorithm for the stochastic evaluation of a thermal average such as Eq. (7.50)is defined to be of polynomial complexity if the computational time t(", N, !)needed to achieve a relative statistical error " = !A/!A" in the evaluation of theaverage !A" scales polynomially with the system size N and inverse temperature!, i.e. if there exist integers n and m and a constant # < & such that

t(", N, !) < #""2Nn!m. (7.51)

• For a quantum system that su"ers from a sign problem for an observable A, andfor which there exists a polynomial complexity algorithm for the related bosonicsystem Eq. (7.50), we define a solution of the sign problem as an algorithm ofpolynomial complexity to evaluate the thermal average !A".

It is important to note that we only worry about the sign problem if the bosonicproblem is easy (of polynomial complexity) but the fermionic problem hard (of expo-nential complexity) due to the sign problem. If the bosonic problem is already hard,e.g. for spin glasses 4, the sign problem will not increase the complexity of the problem.Also, changing the representation so that all p(c) ' 0 might not su#cient to solve thesign problem if the scaling remains exponential, since then we just map the sign prob-lem to another exponentially hard problem. Only a polynomial complexity algorithmcounts as a solution of the sign problem.

At first sight such a solution seems feasible since the sign problem is not an intrinsicproperty of the quantum model studied but is representation-dependent: it depends on

4F. Barahona, J. Phys. A 15, 3241 (1982)

72

Page 19: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

the choice of basis sets {|i!}, and in some models it can be solved by a simple localbasis change.. Indeed, when using the eigen basis in which the Hamilton operator H isdiagonal, there will be no sign problem. This diagonalization of the Hamilton operatoris, however, no solution of the sign problem since its complexity is exponential in thenumber of particles N .

We now construct a quantum mechanical system for which the calculation of athermal average provides the solution for one and thus all of theNP-complete problems.This system exhibits a sign problem, but the related bosonic problem is easy to solve.Since, for this model, a solution of the sign problem would provide us with a polynomialtime algorithm for anNP-complete problem, the sign problem isNP-hard. Of course, itis expected that the corresponding thermal averages cannot be calculated in polynomialtime and the sign problem thus cannot be solved. Otherwise we would have found apolynomial time algorithm for the NP-complete problems and would have shown thatNP=P.

The specific NP-complete problem we consider is to determine whether a state withenergy less than or equal to a bound E0 exists for a classical three-dimensional Isingspin glass with Hamilton function

H = "!

!j,k"

Jjk!j!k. (7.52)

Here the spins !j take the values ±1, and the couplings Jjk between nearest neighborlattice points j and k are either 0 or ±J .

This problem is in the complexity class NP since the non-deterministic machinecan evaluate the energies of all configurations c in polynomial time and test whetherthere is one with E(c) # E0. In addition, the validity of a positive answer (i.e. thereis a configuration c) can be tested on a deterministic machine by evaluating the energyof that configuration. The evaluation of the partition function Z =

"

c exp(""E(c))is, however, not in NP since the non-deterministic machine cannot perform the sum inpolynomial time.

This question whether there is a state with energy E(c) # E0 can also be answeredin a Monte Carlo simulation by calculating the average energy of the spin glass at alarge enough inverse temperature ". Since the energy levels are discrete with spacing Jit can easily be shown that by choosing an inverse temperature "J $ N ln 2 + ln(12N)the thermal average of the energy will be less than E0+J/2 if at least one configurationwith energy E0 or less exists, and larger than E0 + J otherwise.

In this classical Monte Carlo simulation, the complex energy landscape, created bythe frustration in the spin glass (Fig. 2a), exponentially suppresses the tunneling of theMonte Carlo simulation between local minima at low temperatures. The autocorrelationtimes and hence the time complexity of this Monte Carlo approach are exponentiallylarge # % exp(aN), as expected for this NP-complete problem.

We now map this classical system to a quantum system with a sign problem. We doso by replacing the classical Ising spins by quantum spins. Instead of the common choicein which the classical spin configurations are basis states and the spins are representedby diagonal !z

j Pauli matrices we choose a representation in which the spins point in

73

Page 20: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

?

|i1>

|i2>

|i3>

|i1>

a) b)

Figure 7.7: a) A classically frustrated spin configuration of three antiferromagneticallycoupled spins: no configuration can simultaneously minimize the energy of all threebonds. b) A configuration of a frustrated quantum magnet with negative weights: threeantiferromagnetic exchange terms with negative weights are present in the sequence|i1! " |i2! " |i3! " |i1!. Here up-spins with z-component of spin !z

j = 1 and down-spins with !z

j = #1 are connected with di!erently colored world lines.

the ±x direction and are represented by !xj Pauli matrices:

H = #!

!j,k"

Jjk!xj !

xk , (7.53)

Here the random signs of the couplings are mapped to random signs of the o!-diagonalmatrix elements which cause a sign problem (see Fig. 2b). The related bosonic modelis the ferromagnet with all couplings Jjk $ 0 and e"cient cluster algorithms withpolynomial time complexity are known for this model. Since the bosonic version is easyto simulate, the sign problem is the origin of the NP-hardness of a quantum MonteCarlo simulation of this model. A generic solution of the sign problem would providea polynomial time solution to this, and thus to all, NP-complete problems, and wouldhence imply that NP=P. Since it is generally believed that NP%=P, we expect thatsuch a solution does not exist.

By constructing a concrete model we have shown that the sign problem of quan-tum Monte Carlo simulations is NP-hard. This does not exclude that a specific signproblem can be solved for a restricted subclass of quantum systems. This was indeedpossible using the meron-cluster algorithm 5 for some particular lattice models. Sucha solution must be intimately tied to properties of the physical system and allow anessentially bosonic description of the quantum problem. A generic approach might scalepolynomially for some cases but will in general scale exponentially.

In the case of fermions or frustrated quantum magnets, solving the sign problemrequires a mapping to a bosonic or non-frustrated system – which is, in general, almost

5S. Chandrasekharan and U.-J. Wiese, Phys. Rev. Lett. 83, 3116 (1999)

74

Page 21: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

certainly impossible for physical reasons. The origin of the sign problem is, in fact, thedistinction between bosonic and fermionic systems. The brute-force approach of takingthe absolute values of the probabilities means trying to sample a frustrated or fermionicsystem by simulating a non-frustrated or bosonic one. As for large system sizes N andlow temperatures the relevant configurations for the latter are not the relevant ones forthe former, the errors are exponentially large.

Given the NP-hardness of the sign problem one promising idea for the simula-tion of fermionic systems is to use ultra-cold atoms in optical lattices to constructwell-controlled and tunable implementations of physical systems, such as the Hubbardmodel, and to use these “quantum simulators” to study the phase diagrams of cor-related quantum systems. But even these quantum simulators are most likely not ageneric solution to the sign problem since there exist quantum systems with exponen-tially diverging time scales and it is at present not clear whether a quantum computercould solve the NP-complete problems.

7.6 Worm and directed loop updates

The quantum loop algorithm, like the classical cluster algorithms work only for modelswith spin-inversion symmetry: a configuration of spins and the flipped configurationneed to have the same weight. Applying a magnetic field breaks this spin-inversionsymmetry, and the loop algorithm cannot be applied. For that case, Nikolay Prokof’evand coworkers invented the worm algorithm.6 The worm algorithm is based on a verysimple idea: while local updates on world lines are not ergodic (since knots cannot becreated or undone), the worm algorithm proceeds by cutting a world line, and thenmoves the open ends in local updates until they meet again and the world line is gluedtogether once more.

6N. V. Prokof’ev, B. V. Svistunov and I. S. Tupitsyn, Sov. Phys - JETP 87, 310 (1998).

75

Page 22: Chapter 7 Cluster quantum Monte Carlo algorithms for ...edu.itp.phys.ethz.ch/fs12/cqp/chapter07.pdf · 0 (7.13) with periodic boundary conditions σ M+1 = σ 1, and the following

76