1. Introduction
Since the advent of the theoretical description of classical and quantum phase transitions (QPTs), long-range interactions between degrees of freedom challenged the established concepts and propelled the development of new ideas in the field [
1,
2,
3,
4,
5]. It is remarkable that only a few years after the introduction of the renormalisation group (RG) theory by K. G. Wilson in 1971 as a tool to study phase transitions and as an explanation for universality classes [
6,
7,
8,
9,
10,
11], it was used to investigate ordering phase transitions with long-range interactions. These studies found that the criticality depends on the decay strength of the interaction [
1,
2,
3]. It then took two decades to develop numerical Monte Carlo (MC) tools capable to simulate basic magnetic long-range models with thermal phase transitions following the behaviour predicted by the RG theory [
12,
13]. The results of these simulations sparked a renewed interest in finite-size scaling above the upper critical dimension [
12,
14,
15,
16,
17,
18,
19] since "hyperscaling is violated" [
13] for long-range interactions that decay slowly enough. In this regime, the treatment of dangerous irrelevant variables (DIVs) in the scaling forms is required to extract critical exponents from finite systems.
Meanwhile, a similar historic development took place regarding the study of QPTs under the influence of long-range interactions. By virtue of pioneering RG studies [
20,
21], the numerical investigation of long-range interacing magnetic systems has been triggered [
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38]. In particular, Monte Carlo based techniques became a popular tool to gain quantitative insight into these long-range interacting quantum magnets [
22,
25,
26,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40]. On the one hand this includes high-order series expansion techniques, where classical Monte Carlo integration is applied for the graph embedding scheme, allowing to extract energies and observables in the thermodynamic limit [
25,
29]. On the other hand, there is stochastic series expansion quantum Monte Carlo [
39], which enables calculations on large finite systems. To determine physical properties of the infinite system, finite-size scaling is performed with the results of these computations. Inspired by the recent developments for classical phase transitions [
15,
16,
17,
18,
19,
41], a theory for finite-size scaling above the upper critical dimension for QPTs was introduced [
32,
34].
When investigating algebraically decaying long-range interactions
with the distance
r and the dimension
d of the system, there are two distinct regimes: One for,
(strong long-range interaction) and another one for
(weak long-range interaction) [
5,
42,
43,
44,
45]. In the case of strong long-range interactions, common definitions of internal energy and entropy in the thermodynamic limit are not applicable and standard thermodynamics breaks down [
5,
42,
43,
44,
45]. We will not focus on this regime in this review. For details specific to strong long-range interactions, we refer to other review articles such as Refs. [
5,
42,
43,
44,
45]. For the sake of this work, we restrict the discussion to weak long-range interaction or competing antiferromagnetic strong long-range interactions, for which an extensive ground-state energy can be defined without rescaling of the coupling constant [
5].
The interest in quantum magnets with long-range interactions is further fueled by the relevance of these models in state-of-the-art quantum-optical platforms [
5,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87]. To realise long-range interacting quantum lattice models with a tunable algebraic decay exponent, one can use trapped ions which are coupled off-resonantly to motional degrees of freedom [
5,
81,
82,
83,
84,
85,
88]. Another possibility is to couple trapped neutral atoms to photonic modes of a cavity [
5,
86,
87]. Alternatively, one can realise long-range interactions decaying with a fixed algebraic decay exponent of six or three using Rydberg atom quantum simulators [
46,
47,
48,
49,
50,
51,
52,
53,
54,
55] or ultracold dipolar quantum atomic or molecular gases in optical lattices [
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73]. Note, in many of the above listed cases it is possible to map the long-range interacting atomic degrees of freedom onto quantum spin models [
5,
52,
89]. Therefore, they can be exploited as analogue quantum simulators for long-range interacting quantum magnets and the relevance of the theoretical concepts transcends the boundary between the fields.
From the perspective of condensed matter physics, there are multiple materials with relevant long-range interactions [
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106]. The compound LiHoF
4 in an external field realises an Ising magnet in a transverse magnetic field [
102,
103,
104,
105]. A recent experiment with the two-dimensional Heisenberg ferromagnet Fe
3GeTe
2 demonstrates that phase transitions and continuous symmetry breaking can be implemented by circumventing the Hohenberg-Mermin-Wagner theorem with long-range interactions [
106]. This material is in the recently discovered material class of 2d magnetic van der Waals systems [
107,
108]. Further, dipolar interactions play a crucial role in the spin-ice state in frustrated magnetic pyrochlore materials Ho
2Ti
2O
7 and Dy
2Ti
2O
7 [
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101].
In this review, we are interested in physical systems described by quantum spin models, where the magnetic degrees of freedom are located on the sites of a lattice. We concentrate on the following three paradigmatic types of magnetic interactions between lattice sites: First, Ising interactions where the magnetic interaction is oriented only in the direction of one quantisation axis. Second, XY interactions with a -symmetric magnetic interaction invariant under planar rotations and, third, Heisenberg interactions with a -symmetric magnetic interaction invariant under rotations in 3d spin space. In the microscopic models of interest a competition between magnetic ordering and trivial product states, external fields, or quasi long-range order leads to QPTs.
In this context, the primary research pursuit revolves around how the properties of the QPT depend on the long-range interaction. The upper critical dimension of a QPT in magnetic models with non-competing algebraically decaying long-range interactions is known to depend on the decay exponent of the interaction for a small enough exponent, and decreases as the decay exponent decreases [
20,
21]. If the dimension of a system is equal or exceeds the upper critical dimension, the QPT displays mean-field critical behaviour. At the same time standard finite-size scaling as well as standard hyperscaling relations are no longer applicable. Therefore, these systems are primary workhorse models to study finite-size scaling above the upper critical dimension. In this case the numerical simulation of these systems is crucial in order to gauge novel theoretical developments. Further, systems with competing long-range interactions do not tend to depend on the long-range nature of the interaction [
23,
24,
25,
26,
29,
30,
32]. In several cases long-range interactions then lead to the emergence of ground states and QPT which are not present in the corresponding short-range interacting models [
27,
30,
54,
55,
109,
110,
111,
112].
In this review, we are mainly interested in the description and discussion of two Monte Carlo based numerical techniques, which were successfully used to study the low-energy physics of long-range interacting quantum magnets, in particular with respect to the quantitative investigation of QPTs [
22,
25,
29,
30,
31,
32,
34,
35,
36,
37,
38,
40]. We chose to review this topic due to our personal involvement with the application and development of these methods [
25,
29,
30,
31,
32,
34,
35]. On one hand, we explain in detail how classical Monte Carlo integration can enhance the capabilities of linked-cluster expansions (LCEs) with the pCUT+MC approach (a combination of the perturbative unitary transforms approach (pCUT) and MC embedding). On the other hand, we describe how stochastic series expansion (SSE) quantum Monte Carlo (QMC) is used to directly sample the thermodynamic properties of suitable long-range quantum magnets on finite systems.
This review is structured as follows. In
Section 2 we review the basic concept of a QPT in a condensed way, focusing on the details relevant for this review. We define the quantum critical exponents and the relations between them in
Section 2.1. Here, we also have the first encounter with the generalised hyperscaling relation which is also valid above the upper critical dimension where conventional hyperscaling breaks down. As the SSE QMC discussed in this review is a finite-system simulation, we discuss the conventional finite-size scaling below the upper critical dimension in
Section 2.2 and the peculiarities of finite-size scaling above the upper critical dimension in
Section 2.3. In
Section 3, we summarise the basic concepts of Markov chain Monte Carlo integration: Monte Carlo sampling, Markovian random walks, stationary distributions, the detailed balance condition, and the Metropolis-Hastings algorithm. We continue by introducing the series-expansion Monte Carlo embedding method pCUT+MC in
Section 4. We start with the basic concepts of a graph expansion in
Section 4.1 and introduce the perturbative method of our choice, the perturbative continuous unitary transformation method, in
Section 4.2. We introduce the theoretical concepts for setting up a linked-cluster expansion as a full graph decomposition in
Section 4.3 and subsequently discuss how to practically calculate perturbative contributions in
Section 4.4 and
Section 4.5. We prepare the discussion of the white-graph decomposition in
Section 4.6 with an interlude on the relevant graph theory in
Section 4.6.1 and
Section 4.6.2 and the important concept of white graphs in
Section 4.6.3. Further, in
Section 4.7 we discuss the embedding problem for the white-graph contributions. Starting from the nearest-neighbour embedding problem in
Section 4.7.1, we generalise it to the long-range case in
Section 4.7.2 and then introduce a classical Monte Carlo algorithm to calculate the resulting high-dimensional sums in
Section 4.7.3. This is followed by some technical aspects on series extrapolations in
Section 4.8 and a summary of the entire workflow in
Section 4.9. In the next section the topic changes towards the review of the SSE, which is an approach to simulate thermodynamic properties of suitable quantum many-body systems on finite systems at a finite temperature. First, we discuss the general concepts of the method in
Section 5. We review the algorithm to simulate arbitrary transverse-field Ising models introduced by A. Sandvik [
39] in
Section 5.1. We then review an algorithm used to simulate non-frustrated Heisenberg models in
Section 5.2. After the introduction to the algorithms, we summarise techniques on how to measure common observables in the SSE QMC scheme in
Section 5.3. Since the SSE QMC method is a finite temperature method, we discuss how to rigorously use this scheme to perform simulations at effective zero temperature in
Section 5.4. We conclude this section with a brief summary of path integral Monte Carlo techniques used for systems with long-range interactions (see
Section 5.5). To maintain the balance between algorithmic aspects and their physical relevance, we summarise several theoretical and numerical results for quantum phase transitions in basics long-range interacting quantum spin models, for which the discussed Monte Carlo based techniques provided significant results. First, we discuss long-range interacting transverse-field Ising models in
Section 6. For ferromagnetic interactions this model displays three regimes of universality: A long-range mean-field regime for slowly decaying long-range interactions, an intermediate long-range non-trivial regime, and a regime of short-range universality for strong decaying long-range interactions. We discuss the theoretical origins of this behaviour in
Section 6.1.1 and numerical results for quantum critical exponents in
Section 6.1.2. Since this model is a prime example to study scaling above the upper critical dimension in the long-range mean-field regime, we emphasise these aspects in
Section 6.1.3. Further, we discuss the antiferromagnetic long-range transverse-field Ising model on bipartite lattices in
Section 6.2 and on non-bipartite lattices in
Section 6.3. The next obvious step is to change the symmetry of the magnetic interactions. Therefore, we turn to long-range interacting XY models in
Section 7 and Heisenberg models in
Section 8. We discuss the long-range interacting transverse-field XY chain in
Section 7 starting with the
-symmetric isotropic case in
Section 7.1, followed by the anisotropic case for ferromagnetic (see
Section 7.2) and antiferromagnetic (see
Section 7.3) interactions which display similar behaviour to the long-range transverse-field Ising model on the chain discussed in
Section 6. We conclude the discussion of results with unfrustrated long-range Heisenberg models in
Section 8. We focus on the staggered antiferromagnetic long-range Heisenberg square lattice bilayer model in
Section 8.1 followed by long-range Heisenberg ladders in
Section 8.2 and the long-range Heisenberg chain in
Section 8.3. We conclude in
Section 9 with a brief summary and with some comments on next possible steps in the field.
2. Quantum Phase Transitions
This review is part of the special issue with the topic "Violations of Hyperscaling in Phase Transitions and Critical Phenomena". In this work, we summarise investigations of low-dimensional quantum magnets with long-range interactions targeting in particular quantum phase transitions (QPTs) above the upper critical dimension, where the naive hyperscaling relation is no longer applicable. In this section, we recapitulate the relevant aspects of QPTs needed to discuss the results of the Monte Carlo based numerical approaches. First, we give a general introduction to QPTs. After that, we discuss in detail the definition of critical exponents and relations among them in
Section 2.1 as well as the scaling below (see
Section 2.2) and above (see
Section 2.3) the upper critical dimension.
Any non-analytic point of the ground-state energy of an infinite quantum system as a function of a tuning parameter
is identified with a QPT [
113]. This tuning parameter can, for instance, be a magnetic field or pressure but not the temperature. Quantum phase transitions are a concept of zero temperature as there are no thermal fluctuations and all excited states are supressed infinitely strong such that the system remains in its ground state. There are two scenarios how a non-analytic point in the ground-state energy can emerge [
113]: First, an actual (sharp) level crossing between the ground-state energy and another energy level. Second, the non-analytic point can be considered as a limiting case of an avoided level crossing.
In this review, we are interested in second-order QPTs, which fall into the second scenario. At a second-order QPT, the relevant elementary excitations condense into a novel ground state while the characteristic length and time scales diverge. Apart from topological QPTs involving long-range entangled topological phase, such continuous transitions are described by the concept of spontaneous symmetry breaking. On one side of the QPT the ground state obeys a symmetry of the Hamiltonian, while on the other side this symmetry is broken in the ground state and a ground-state degeneracy arises.
Following the idea of the quantum-to-classical mapping [
113,
114],
d-dimensional quantum systems can be mapped in the vicinity of a second-order QPT to models of statistical mechanics with a classical (thermal) second-order phase transition in
dimensions. In many cases, the models obtained from a quantum-to-classical mapping are rather artificial [
113]. However, such mappings often allow to categorise QPTs in terms of universality classes and associated critical exponents by the non-analytic behaviour of the classical counterparts [
10,
11,
113,
115,
116]. The mapping further illustrates that the renormalisation group (RG) theory is also applicable to describe QPTs [
10,
11,
113,
115,
116].
In the RG theory, each QPT belongs to a non-trivial
1 fixed point of the RG transformation [
10,
11]. Critical exponents are connected to the RG flow in the immediate vicinity of these non-trivial fixed points [
10,
11,
113]. The concept of universality classes arises from the fact that different microscopic Hamiltonians can have a quantum critical point that is attracted by the same non-trivial fixed point under successive application of the RG transformation [
10,
11]. Due to this, the QPTs in these models have the same critical exponents.
Another remarkable result of the RG theory is the scaling of observables in the vicinity of phase transitions. Historically, the theory of scaling at phase transitions was heuristically introduced before the RG approach [
117,
118,
119,
120,
121,
122,
123]. The latter provided the theoretical foundation for the scaling hypothesis [
6,
7]. The main statement of the scaling theory is that the non-analytic contributions to the free energy and correlation functions are mathematically described by generalised homogeneous functions (GHF) [
123]. A function with
n variables
is called a GHF, if there exist
with at least one being non-zero and
such that for
The exponents
are the scaling powers of the variables and
is the scaling power of the function
f itself. An in-depth summary of the mathematical properties of GHFs can be found in
Appendix B. The most important properties of GHFs are that its derivatives, Legendre transforms, and Fourier transforms are also GHFs. As we will outline in
Section 2.2, the theory of finite-size scaling is formulated in terms of GHFs and relates the non-analytic behaviour at QPTs in the thermodynamic limit with the scaling of observables for different finite system sizes. In this, the variables
are related to physical parameters like the temperature
T, control parameter
, symmetry-breaking field
H and also irrelevant, more abstract parameters that parameterise microscopic details of the model like the lattice spacing. Later in this section, we will define irrelevant variables in the context of RG and GHFs.
Another aspect relevant for this work is that quantum fluctuations are the driving factor with QPTs [
113]. In general, fluctuations are more important in low dimensions [
116]. The universality class of QPTs for a certain symmetry breaking depends on the dimensionality of the system.
An important aspect regarding this review is the so-called upper critical dimension
. The upper critical dimension is defined as a dimensional boundary such that for systems with dimension
, the critical exponents are those obtained from mean-field considerations. The upper critical dimension is of particular importance for QPTs in systems with non-competing long-range interactions. For sufficiently small decay exponents of an algebraically decaying long-range interaction
, the upper critical dimension starts to decrease as a function of the decay exponent
[
20,
32,
34]. In the limiting case of a completely flat decay (
) of the long-range interaction resulting in an all-to-all coupling, the model is intrinsically of mean-field type and mean-field considerations become exact. For a certain value of the decay exponent, the upper critical dimension becomes equal to the fixed spatial dimension, and for decay exponents below this value, the dimension of the system is above the upper critical dimension of the transition [
20,
32,
34]. This makes long-range interacting systems an ideal test bed for studying phase transitions above the upper critical dimension in low-dimensional systems. In particular, long-range interactions can make the upper critical dimension accessible in real-world experiments as the upper critical dimension of short-range models is usually not below three.
Although phase transitions above the upper critical dimension display mean-field criticality, they are still a matter worth studying, since naive scaling theory describing the behaviour of finite systems close to a phase transition (see
Section 2.2) is no longer applicable [
16,
19,
124]. Moreover, the naive versions of some relations between critical exponents, as discussed in
Section 2.1, do not hold anymore [
15,
16]. The reason for this issue are the dangerous irrelevant variables (DIVs) in the RG framework [
125,
126,
127]. During the application of the RG transformation, the original Hamiltonian is successively mapped to other Hamiltonians which can have infinitely many couplings. All these couplings, in principle, enter the GHFs. In practice, all but a finite number of these couplings are set to zero since their scaling powers are negative which means they flow to zero under renormalisation. These couplings are therefore called irrelevant. This approach of setting irrelevant couplings to zero can be used to derive the finite-size scaling behaviour as described in
Section 2.2. However, above the upper critical dimension, this approach breaks down because it is only possible to set irrelevant variables to zero if the GHF does not have a singularity in this limit [
125]. Above the upper critical dimension such singularities in irrelevant parameters exist, which makes them DIVs [
126]. We explain the effect of DIVs on scaling in
Section 2.3.
2.1. Critical Exponents in the Thermodynamic Limit
As outlined above, a second-order QPT comes with a singularity in the free energy density. In fact, also other observables experience singular behaviour at the critical point in the form of power law singularities. For instance, the order parameter
m as a function of the control parameter
behaves as
in the ordered phase. Without loss of generality, the system is taken to be in the ordered phase for
and the notation
means that
is approaching
from below, i. e. it is approaching in the ordered phase. In the disordered phase
, the order parameter by definition vanishes such that
. The observables with their respective power-law singular behaviour, that is characterised by the critical exponents
and
z, are summarised in
Table 1 together with how they are commonly defined in terms of the free energy density
f, the symmetry-breaking field
H, that couples to the order parameter, and the reduced control parameter
.
One usually defines reduced parameters like
r that vanish at the critical point not only to shorten the notation but also to express the power-law singularities independent of the microscopic details of the specific model one is looking at. While the value of
depends on these details, the power-law singularities are empirically known to not depend on the microscopic details but only on more general properties like the dimensionality, the symmetry that is being broken and, with particular emphasis due to the focus of this review, on the range of the interaction. It is therefore common to classify continuous phase transitions in terms of universality classes. These universality classes share the same set of critical exponents. In terms of RG, this behaviour is understood as distinct critical points of microscopically different models flowing to the same renormalisation group fixed point, which determines the criticality of the system [
6,
7,
10]. Prominent examples for universality classes of magnets are the 2d- and 3d-Ising (
symmetry), 3d XY (
symmetry), 3d Heisenberg (
symmetry) universality classes [
113]. It is important to mention that the dimension in the classifications are referring to classical and not quantum systems and they should not be confused with each other. In fact, the universality class of a short-range interacting non-frustrated quantum Ising model of dimension
d lies in the universality of the classical
-dimensional Ising model.
There are only a few dimensions for which a separate universality class is defined for the different models. For lower dimension, the fluctuations are too strong in order for a spontaneous symmetry breaking to occur. In case of the classical Ising model, there is no phase transition for 1d, while for the classical XY and Heisenberg models with continuous symmetries there is not even a phase transitions for 2d due to the Hohenberg-Mermin-Wagner (HWM) theorem [
129,
130,
131]. This dimensional boundary is referred to as lower critical dimension
dlc. The lower critical dimension is the highest dimension for which no transition occurs, i. e.
dlc for the Ising model and
dlc for the XY and Heisenberg model. For higher dimensions
, the critical exponents of the mentioned models do not depend on the dimensionality anymore and they take on the mean-field critical exponents in all dimensions. The underlying reason is that with increasing dimensions the local fluctuations become smaller due to the higher connectivity of the system [
132]. This has been also exploited in diagrammatic and series expansions in
[
133,
134,
135]. This dimensional boundary, at which the criticality becomes the mean-field one, is called upper critical dimension
duc. Usually, the upper critical dimension is too large to realise a system above its upper critical dimension in the real world. However, long-rang interactions can increase the connectivity of a system in a similar sense as the dimensionality. Sufficiently long-ranged interaction can therefore lower the upper critical to a value that is accessible in experiments.
Finally, it is worth mentioning that the critical exponents are not independent from each other but obey certain relations [
128], namely
The first relation in Eq. (3) is the so-called hyperscaling relation whose classical analogue (without
z) was introduced by Widom [
10,
136]. The Essam-Fisher relation in Eq. (4) [
137,
138] is reminiscent of a similar inequality that was proven rigorously by Rushbrooke using thermodynamic stability arguments. Eq. (5) is called Widom relation. The last relation in Eq. (6) is the Fisher scaling relation which can be derived using the fluctuation-dissipation theorem [
10,
128,
138]. Those relations were originally obtained from scaling assumptions of observables close to the critical point which were only later derived rigorously when the RG formalism was introduced to critical phenomena [
10,
128]. Due to these relations, it is sufficient to calculate three, albeit not arbitrary, exponents to obtain the full set of critical exponents.
The hyperscaling relation Eq. (3) is the only relation containing the dimension of the system and is therefore often said to break down above the upper critical dimension where one expects the same mean-field critical exponents independent of the dimension
d [
10]. It therefore deserves special focus in this review since the long-range models discussed will be above the upper critical dimension in certain parameter regimes. Personally, we would not agree that the hyperscaling relation breaks down above the upper critical dimension, but we would rather call Eq. (3) a special case of a more general hyperscaling relation
with the pseudo-critical exponent Ϙ ("koppa") [
34]
Below the upper critical dimension, the general hyperscaling relation therefore relaxes to Eq. (3). Above the upper critical dimension the relation becomes
which is independent of the dimension of the system. For the derivation of this generalised version of the hyperscaling relation for QPTs, see
Section 2.3 or Ref. [
34]. The derivation of the classical counterpart can be found in Ref. [
15] and is reviewed in Ref. [
41].
2.2. Finite-Size Scaling below the Upper Critical Dimension
Even though the singular behaviour of observables at the critical point is only present in the thermodynamic limit, it is possible to study the criticality of an infinite system by investigating their finite counterparts. In finite systems, the power-law singularities of the infinite system are rounded and shifted with respect to the critical point, e.g. the susceptibility with its characteristic divergence at the critical point
is deformed to a broadened peak of finite height. The peak’s position
is shifted with respect to the critical point
. A possible definition of a pseudo-critical point of a finite system is the peak position
. As the characteristic length scale of fluctuations
diverges at the critical point, the finite system approaching the critical point will at some point begin to "feel" its finite extent and the observables start to deviate from the ones in the thermodynamic limit. As
diverges with the exponent
like
at the critical point, the extent of rounding in a finite system is related to the value of
. Similarly, the peak magnitude of finite-size observables at the pseudo-critical point will depend on how strong the singularity in the thermodynamic limit is, which means it depends on the respective critical exponents
,
,
, and
. The shifting, rounding, and the varying peak magnitude are shown for the susceptibility of the long-range transverse-field Ising model in
Figure 1. This dependence of observables in finite systems on the criticality of the infinite system is the basis of finite-size scaling.
In a more mathematical sense, the relation between critical exponents and finite-size observables has its origin in the renormalisation group (RG) flow close to the corresponding RG fixed point that determines the criticality [
139]. Close to this fixed point, one can linearise the RG flow so that the free energy density and the characteristic length
become generalised homogeneous functions (GHFs) in their parameters [
123,
127,
128,
140,
141]. For a thorough discussion of the mathematical properties of GHFs we refer to Ref. [
123] and
Appendix B. This means, the free energy density
f and characteristic length scale
as functions of the couplings
and the inverse system length
obey the relations
with the respective scaling dimensions
,
, and
governing the linearised RG flow with spatial rescaling factor
around the RG fixed point, at which all couplings vanish by definition. All of those couplings are relevant except for
u which denotes the leading irrelevant coupling [
10,
113]. Relevant couplings are related to real-world parameters that can be used to tune our system away from the critical point like the temperature
T, a symmetry-breaking field
H, or simply the control parameter
r. The irrelevant couplings
u do not per se vanish at the critical point like the relevant ones do. However, they flow to zero under the RG transformation and are commonly set to zero in the scaling laws
by assuming analyticity in these parameters. The generalised homogeneity of thermodynamic observables can be derived from the one of the free energy density
f. For example, the generalised homogeneity of the magnetisation
can be derived by taking the derivative of
f with respect to the symmetry-breaking field
H.
By investigating the singularity of
in
r via Eq. (13), one can show that the scaling power
of the control parameter
r is related to the critical exponent
by
[
113]. For this, one fixes the value
of the first argument to
in the right-hand side of Eq. (13) by setting
such that
Analogously, further relations between the scaling powers and other critical exponents can be derived by looking at the singular behaviour of the respective observables in the corresponding parameters. Overall, one further gets
From these equations, one can already tell that the critical exponents are not independent from each other. In fact, the scaling relations
,
and
(see Eqs. (3) (5)) can be derived from Eq. (16) and
. By expressing the RG scaling powers
in terms of critical exponents, the homogeneity law for an observable
with a bulk divergence
is given by
In order to investigate the dependence on the linear system size
L, the last argument in the homogeneity law is fixed to
by inserting
. This readily gives the finite-size scaling form
with
being the universal scaling function of the observables
. The scaling function
itself does not depend on
L anymore, but in order to compare different linear system sizes one has to rescale its arguments. To extract the critical exponents from finite systems, the observable
is measured for different system sizes
L and parameters
close to the critical point
. The
L-dependence according to Eq. (19) is then fitted with the critical exponents
,
,
and
z as well as the critical point
, which is hidden in the definition of
r, as free parameters. It is advisable to fix two of the three parameters
to their critical values in order to minimise the amount of free parameters in the fit. For example, with
and only
one can extract the two critical exponents
and
alongside
. For further details on a fitting procedure, we refer to Ref. [
32]. If one knows the critical point, one can also set
and look at the
L-dependent scaling
directly at the critical point to extract the exponent ratio
. There are many more possible approaches to extract critical exponents from the FSS law in Eq. (19) [
142,
143,
144]. For relatively small system sizes, it might be required to take corrections to scaling into account [
143,
144].
2.3. Finite-Size Scaling above the Upper Critical Dimension
In the derivation of finite-size scaling below the upper critical dimension, it was assumed that the free energy density is an analytic function in the leading irrelevant coupling u and therefore one can set it to zero. However, this is not the case above the upper critical dimension anymore and the free energy density f is singular at . Due to this singular behaviour u is referred to as a dangerous irrelevant variable (DIV).
As a consequence, one has to take the scaling of
u close to the RG fixed point into account. This is done by absorbing the scaling of
f in
u for small
u into the scaling of the other variables [
127]
up to a global power
of
u. This leads to a modification of the scaling powers in the homogeneity law for the free energy density [
127]
with the modified scaling powers [
34,
127]
In the classical case [
127],
yL* was commonly set to 1 by choice. This is justified because the scaling powers of a GHF are only determined up to a common non-zero factor [
123]. However, for the quantum case [
34], this was kept general as it has no impact on the FSS.
As the predictions from Gaussian field-theory and mean-field differed for the critical exponents
,
, and
but not for the "correlation" critical exponents
,
z,
, and
[
145], the correlation sector was thought not to be affected by DIVs at first [
127,
142,
145]. Later Q-FSS, another approach to FSS above the upper critical dimension, pioneered by Ralph Kenna and his colleagues, was developed for classical [
15,
19,
124] as well as for quantum systems [
34] which explicitly allowed the correlation sector to also be affected by DIV. In analogy to the free energy density, the homogeneity law of the characteristic length scale is then also modified to
with
in order to reproduce the correct bulk singularity
. A new pseudo-critical exponent Ϙ ("koppa")
is introduced. This exponent describes the scaling of the characteristic length scale with the linear system size. This non-linear scaling of
with
L is one of the key differences to the previous treatments above the upper critical dimension in Ref. [
127].
Analogous to the case below the upper critical dimension, the modified scaling powers
can be related to the critical exponents:
By using the mean-field critical exponents for the
quantum rotor model, one gets restrictions for the ratios of modified scaling powers
Furthermore, one can link the bulk scaling powers
, and
to the scaling power
yL* of the inverse linear system size [
34]
by looking at the scaling of the susceptibility in a finite system [
34,
127]. This relation is crucial for deriving a FSS form above the upper critical dimension as the modified scaling power
yL* or rather its relation to the other scaling powers determines the scaling with the linear system size
L. For details on the derivation, we refer to Ref. [
34]. We want to stress again that the scaling powers of GHFs are only determined up to a common non-zero factor [
123]. Therefore, it is evident that one can only determine the ratios of the modified scaling powers but not their absolute value. The absolute values are subject to choice. Different choices were discussed in Ref. [
34], but these choices rather correspond to taking on different perspectives and have no impact on FSS or the physics.
From Eq. (32) together with Eqs. (26) and (27), a generalised hyperscaling relation
can be derived. This also determines the pseudo-critical exponent
Finally, we can express the modified scaling powers in the FSS law for an observable
with power-law singularity
For
below the upper critical dimension, Eq. (36) relaxes to the standard FSS law Eq. (19). The scaling in temperature has not yet been studied for finite quantum systems above the upper critical dimension. However, in Ref. [
34] it was conjectured that
based on Eq. (32), which is also in agreement with
z being the dynamical critical exponent that determines the space-time anisotropy
as we will shortly see. This means that the finite-size gap scales as
with the system size [
34]. Of particular interest is the scaling of the characteristic length scale above the upper critical dimension, for which the modified scaling law Eq. (36) also holds with
. Hence, the characteristic length scale in dependence of the control parameter
r scales like
with the scaling function
. Directly at the critical point
, this leads to
. Comparing this with the scaling of the inverse finite-size gap
verifies that
z still determines the space-time anisotropy. Prior to Q-FSS [
15,
17], the characteristic length scale was thought to be bound by the linear system size
L [
127]. However, this was shown not to be the case by measurements of the characteristic length scale for the classical five-dimensional Ising model [
14] and for long-range transverse-field Ising models [
34].
For the latter, the data collapse of the correlation length according to Eq. (37) is shown in
Figure 2 as an example.
3. Monte Carlo Integration
In this section we provide a brief introduction to Monte Carlo integration (MCI). We focus on the aspects of Markov chain MCI as the basis to formulate the white-graph Monte Carlo embedding scheme of the pCUT+MC method in
Section 4 and the stochastic series expansion (SSE) quantum Monte Carlo (QMC) algorithm in
Section 5 in a self-contained fashion. MCI is the foundation for countless numerical applications which require the integration over high-dimensional integration spaces. As this review has a focus on "Monte Carlo based techniques for quantum magnets with long-range interactions", we forward readers with a deeper interest in the fundamental aspects of MCI and Markov chains to Refs. [
146,
147,
148].
MCI summarises numerical techniques to find estimators for integrals of functions
over an integration space
using random numbers. The underlying idea behind MCI is to estimate the integral, or the sum in the case of discrete variables, of the function
f over the configuration space by an expectation value
with
sampled according to a probability density function
(PDF) and the function
reweighted by the PDF. A famous direct application of this idea is the calculation of number "pi" which is in great detail discussed in Ref. [
148].
In this review, MCI is used for the embedding of white graphs on a lattice to evaluate high orders of a perturbative series expansion or to calculate thermodynamic observables using SSE. In both cases, non-normalised relative weights
within a configuration space
arise which are used for the sampling of the PDF
P
being oftentimes not directly accessible. In the context of statistical physics,
is often chosen to be the relative Boltzmann weight
of each configuration
. While this relative Boltzmann weight is accessible as long as
is known, the full partition function to normalise the weights is in general not.
In order to efficiently sample the configuration space
according to the relative weights, the methods in this review use a Markovian random walk to generate
. Let
be the random state of a random walk at a discrete step
n. The state
at the next step is randomly determined according to the conditional probabilities
(transition probabilities). This transition probabilities are normalised by
Markovian random walks obey the Markov property, which means the random walk is memory free and the transition probability for multiple steps factorises into a product over all time steps
with
the start configuration. We require the Markovian random walk to fulfil the following conditions: First, the random walk should have a certain PDF
defined by the weights
in Eq. (
39) as a stationary distribution. By definition,
is a stationary distribution of the Markov chain if it satisfies the global balance condition
Second, we require the random walk to be irreducible, which means that the transition graph must be connected and every configuration
can be reached from any configuration
in a finite number of steps. This property is necessary for the uniqueness of the stationary distribution [
147]. Lastly, we require the random walk to be aperiodic (see Ref. [
147] for a rigorous definition). Together with the irreducibility condition, this ensures convergence to the stationary distribution [
147].
There are several possibilities to design a Markov chain with a desired stationary distribution [
146,
147,
148,
149,
150,
151]. Commonly, the Markov chain is constructed to be reversible. This means that it satisfies the detailed balance condition
which is a stronger condition for stationarity of
P than the global balance condition in Eq. (42). One popular choice for the transition probabilities
that satisfies the detailed balance condition is given by the Metropolis-Hastings algorithm. Most applications of MCI reviewed in this work are based on the Metropolis-Hastings algorithm [
149,
150]. In this approach, the transition probabilities
are decomposed into propositions
and acceptance probabilities
as follows
The probabilities to propose a move
can be any random walk satisfying the irreducibility and aperiodicity condition. By inserting the decomposition of the transition probabilities Eq. (
44) into the detailed balance condition Eq. (
43), one obtains for the acceptance probabilities
where in the last step it was used that the unknown normalisation factors (see Eq. (
39)) of the PDF cancel. The condition in Eq. (
45) is fulfilled by the Metropolis-Hastings acceptance probabilities [
150]
For the special case, that the proposition probabilities are symmetric
, Eq. (46) reduces to the Metropolis acceptance probabilities
As an example, we regard a classical thermodynamic system with Boltzmann weights
given by the energies of configurations and the inverse temperature to give an intuitive interpretation of the Metropolis acceptance probabilities in Eq. (
47). The proposition to move from a configuration
to a configuration
with a smaller energy
is always accepted independent of the temperature. On the other hand, the proposition to move to a configuration
with a larger energy than
is only accepted with a probability depending on the ratio of the Boltzmann weights. If the temperature is higher, it is more likely to move to states with a larger energy. This reflects the physics of the system in the algorithm, focusing on the low-energy states at low-temperatures and going to the maximum entropy state at large temperatures.
4. Series-Expansion Monte Carlo Embedding
In this section we provide a self-contained and comprehensive overview of linked-cluster expansions for long-range interacting systems [
25,
29,
30,
31,
34,
35] using white graphs [
152] in combination with Monte Carlo integration for the graph embedding. First, we introduce linked-cluster expansions (LCEs) and discuss perturbative continuous unitary transformations (pCUT) [
153,
154] as a suitable high-order series expansion method. We then establish an adequate formalism for setting up LCEs and discuss the calculation of suitable physical quantities in practice. With the help of white graphs we can employ LCEs for models with long-range interactions and use the resulting contributions in a the Monte Carlo algorithm to deal with the embedding problem posed by long-range interactions. This approach we dub pCUT+MC.
4.1. Motivation and Basic Concepts
The goal of all our efforts is to calculate physical observables in the thermodynamic limit, i.e., for an infinite lattice
, using high-order series expansions. The starting point of every perturbative problem is
where the Hamiltonian describing the full physical problem
can be split up into an unperturbed part
that is readily diagonalisable and a perturbation
associated with a perturbation parameter
that is small compared to the energy scales of
. We aim to obtain a power series up to a maximal reachable order
as an approximation of a desired physical quantity
where the coefficients
are to be determined by series expansion. We want to use the information contained in the power series to infer properties of the approximated function
[
155]. The cost of determining the coefficients is associated with an exponential growth in complexity with increasing order [
155]. Hence, calculations are performed with the help of a computer programme. Obviously, the computer cannot deal with an infinitely large lattice. Instead, we must look at finite cut-outs consisting of a finite set of lattice sites that are connected by bonds (or links) symbolising the interactions of the Hamiltonian on the lattice. We term these cut-outs
clusters. If two clusters
A and
B do not share a common site or conterminously do not have a link that connects any site of A and B with each other (
), then the cluster
is called a
disconnected cluster. Otherwise, if no such partition into disconnected clusters
A and
B exists (
), the cluster
C is called
connected. We can define quantum operators
(e.g., a physical observable) on these clusters just as on the infinite lattice.
There are essentially two ways of performing high-order series expansions. The first one is the naive approach of taking a single finite cluster
[
153,
154,
156,
157,
158] and designing it such that the contribution of
coincides with the contributions on the infinite lattice
up to the considered order in the perturbation parameter. The cluster needs to be chosen large enough such that the perturbative calculations up to the considered order are not affected by the boundaries of the cluster. Another way of performing calculations is to construct the operator contribution on a cluster – coinciding with the infinite lattice contributions up to a given order – by decomposing it into all possible contributions on smaller clusters [
155,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170]. Now the contributions on many but smaller clusters must be determined and added up to obtain the contribution on the infinite lattice
In contrast to the previous approach we willingly accept boundary effects for the many subclusters. Such a cluster decomposition is known to be computationally more efficient because it suffices to calculate the contributions on the Hilbert space of the smaller clusters reducing the overhead of non-contributing processes and it also suffices to do the calculations only on a few clusters as many give identical contributions due to symmetries. This can significantly reduce the overhead.
However, there are subtleties about the validity of performing calculations on finite clusters, e.g., when setting up a linked-cluster expansion (linked-cluster means only connected clusters contribute) the operator
must satisfy a certain property, namely cluster additivity. The quantity
is called
cluster additive if and only if the contribution on disconnected clusters
solely comes from the contributions on its individual connected clusters
A and
B. This means we can simply add the contributions of
A and
B from the smaller connected clusters to obtain the one for the disconnected cluster, i.e.,
as illustrated schematically in
Figure 3.
We can also understand cluster additivity in the language of perturbation theory where acting on bonds in every order forms a cluster of perturbatively active bonds. If such a cluster is connected, we call these processes linked. So cluster additivity simultaneously means that only linked processes will contribute. Cluster additivity is at the heart of the linked-cluster theorem, which states that only linked processes will contribute to the overall contribution in the thermodynamic limit. To set up a linked-cluster expansion we want to exploit cluster additivity and the linked-cluster theorem so that we can "simply" add up the contributions from individual connected clusters to obtain the desired power series in the thermodynamic limit.
An example of a cluster additive quantity is the ground-state energy
. Imagine we want to calculate the ground-state energy of a non-degenerate ground-state subspace, then cluster additivity is naturally fulfilled
and we can calculate the ground-state energy on
from its individual parts
A and
B. However, cluster additivity is not satisfied in general. We can construct a counterexample by considering the first excited state with energy
. For example, the first excitation above the ferromagnetic ground state of the transverse-field Ising model in the low-field limit is a single spin flip dressed with quantum fluctuations induced by the transverse field [
113]. We usually refer to such excitation as
quasiparticles qp. Here, we cannot add the contributions on clusters
A and
B to obtain the excitation energy on cluster
How to set up a linked-cluster expansion for intensive properties is not obvious and it seemed out of reach after the introduction of linked-cluster expansions in the 1980s [
159,
160,
161]. Only several years later, it was noticed by Gelfand [
162] that additivity can be restored for excited states when properly subtracting the ground-state energy. This approach was later generalised to multiparticle excitations [
154,
156,
164,
165] and observables [
154,
163].
In the following, we first introduce a perturbation theory method that maps the original problem in Eq. (
48) to an effective one. We will show that the derived effective Hamiltonian and observables satisfy cluster additivity. In the subsequent section we make use of the property and show how we can set up a linked-cluster expansion for energies of excited states and observables by properly subtracting contributions from lower energy states.
4.2. Perturbation Method—Perturbative Continuous Unitary Transformations
The first step towards setting up a linked-cluster expansion is to find a perturbation method that satisfies cluster additivity, which is generically not given and a non-trivial task [
171]. Here, we use perturbative continuous unitary transformations (pCUT) [
153,
154] that transform the original Hamiltonian perturbatively order by order to a quasiparticle-conserving Hamiltonian, reducing the original many-body problem to an effective few-body problem. We start discussing how to solve the flow equation to obtain the pCUT method and show afterwards how the Hamiltonian decomposes into additive parts that can be used for a linked-cluster expansion.
We strive to solve the usual problem of perturbation theory of Eq. (
48). The unperturbed part
can be easily diagonalised exactly with a spectrum that has to be equidistant and bounded from below. Additionally, the perturbation
must be a sum of operators
containing all processes changing the energy by
n quanta and – if properly rescaled – corresponds to the same number of quasiparticles
n. The goal of the pCUT method is to find an optimal basis in which the many-body problem of the original Hamiltonian reduces to an effective few-body problem. For that, we introduce a unitary transformation depending on a continuous flow parameter
ℓ and define
In the limiting case
we require
to recover the original Hamiltonian and for
we require
so that the unitary transformation maps the original to the desired effective Hamiltonian. We can rewrite the unitary transformation as
where
is the anti-hermitian generator generating the unitary transformation and
the ordering operator for the flow parameter. Taking the derivatives of Eq. (
55) and Eq. (
56) in
ℓ, we eventually arrive at the flow equation
Flow equations have been studied for quite some time in mathematics and physics with a variety of applications [
172,
173,
174,
175,
176,
177,
178,
179,
180]. It was Knetter and Uhrig [
153] who proposed a perturbative ansatz for the generator of continuous unitary transformations along the lines of Mielke [
180], introducing the quasiparticle generator (also known as the "MKU generator") for the pCUT method
where the indices
refer to blocks of the Hamiltonian labelling the quasiparticle number. Diagonal blocks
contain all processes conserving the number of quasiparticles
i, while offdiagonal blocks
contain all processes changing the quasiparticle number from
i to
j. The reasoning behind the ansatz can be explained by looking at
, where processes
in
are assigned the opposite sign of the inverse processes
and therefore the idea is to "rotate away" offdiagonal blocks by the unitary transformation during the flow of
ℓ, while processes that do not change the quasiparticle number are not transformed away due to
but get renormalised during the flow. Consequently, in the limit
we obtain an effective Hamiltonian
that is block diagonal in
n.
This idea is depicted in
Figure 4. Next, we make a perturbative ansatz for the Hamiltonian during the flow
with the notation
and
being undetermined real functions. We introduce
distinct expansion parameters instead of just a single
to keep the notation as general as possible because below in
Section 4.6.3 about white-graphs we will need multiple expansion parameters to encode additional information. Inserting Eq. (
59) and Eq. (
58) into the flow equation (
57), we can solve the equation perturbatively order by order as we get a recursive set of differential equations for
.
To recover the original Hamiltonian
, we have to demand the correct initial conditions
for
and
for
. We can solve the differential equations (c.f. Ref. [
153]) exactly for
, yielding
with
being exact rational coefficients and the restriction
making the products
quasiparticle conserving [
153]. Hence, the commutator of the effective Hamiltonian with the unperturbed diagonal part of the original Hamiltonian vanishes (
). Note, so far the effective Hamiltonian (
64) is model independent. It only depends on the overall structure of Eq. (
54). The generic form of the Hamiltonian comes at the cost of an additional normal ordering usually by applying the Hamiltonian to a cluster. Of course, it could also be done explicitly by using the hard-core bosonic commutation relations but the former approach can be handled much easier by a computer programme. Yet, we achieved our goal of obtaining a block-diagonal Hamiltonian
where
is the effective irreducible Hamiltonian of
n quasiparticle processes (see also
Section 4.3). Let us emphasise again, that this block diagonal structure allows us to solve the
n quasiparticle blocks individually which significantly reduces the complexity of the original many-body problem to an effective one that is block-diagonal in the quasiparticle number
n.
If we want to calculate an effective observable we can make an ansatz along the same lines [
154]. We insert the perturbative ansatz
with undetermined functions
. The operator product is defined as
Inserting exactly the same generator (
58) and the ansatz for the observable in Eq. (
66) instead of the Hamiltonian into the flow equation (
57), we arrive at
with
by solving the resulting set of differential equations for
[
154]. Note that the last sum does not contain a restriction
and therefore – in contrast to the effective Hamiltonian – effective observables are not (necessarily) quasiparticle conserving.
We have just derived the effective form of the Hamiltonian and observables in the pCUT method that have a very generic form depending only on the structure of the perturbation of Eq. (
54). As already stated, the model dependence of our approach comes into play when performing a linked-cluster expansion by applying the effective Hamiltonian or observable to finite clusters. But how do we know if the effective quantities are cluster additive?
We follow the argumentation of Refs. [
181,
182] by looking at the original Hamiltonian (
48) that trivially satisfies cluster additivity as long as all bonds represent a non-vanishing term in
between sites ("bond equals interaction"). Thus, the Hamiltonian on a disconnected cluster
is cluster additive because
and
are non-interacting. Here, we denote the restriction of the Hamiltonian
to a cluster
C as
. We can further insert this property into the flow equation
Here, we used the property of
that it commutes on disconnected clusters and the fact that the Hamiltonian is continuously transformed during the flow starting from
. Therefore, the derivative and the commutator can be split up acting on each cluster individually and preserving cluster additivity during the flow. Consequently, the effective Hamiltonian
in the limit
is cluster additive as well. The same proof holds for effective observables. Another more physical argument is that the effective pCUT Hamiltonian can be written as a sum of nested commutators of
T-operators [
152,
167]. For instance, considering the perturbation
, the effective Hamiltonian looks like
Splitting up
into local operators acting on bonds
l, the nested commutators vanish for processes that are not linked. Hence the linked-cluster theorem is fulfilled and the effective Hamiltonian is cluster additive. To emphasise the linked-cluster property the generic effective Hamiltonian is often written as
where the sum over
C runs over all possible connected clusters with maximal
k bonds (
) [
152]. The notation
will be clarified in
Section 4.6.2 where graphs are formally introduced. In this context it is simply the set of bonds of a connected cluster
C and
the number of bonds in this set. The condition
arising from the linked-cluster theorem ensures that the cluster consisting of active links and sites during a process must match with the bonds and sites of the connected cluster
C. For observables the generalised condition
holds, where the index
x can either refer to a site (local observable) or a link (non-local observable) and we have
Although we showed that the effective Hamiltonian and observables are cluster additive and therefore fulfil the linked-cluster theorem, to set up a linked-cluster expansion, there are important subtleties remaining when we restrict the effective Hamiltonian and observables to the quasiparticle basis that we need to address before we can discuss how to do the calculations in practice.
4.3. Unraveling Cluster Additivity
In this subsection we need to clarify how we can use the cluster additive property of the effective pCUT Hamiltonian and observables to set up a linked-cluster expansion not only for the ground-state energy but also for energies of excited states. Many aspects of this section are based on the original work of Ref. [
154], in which a general formalism was developed how to derive suitable quantities for the calculation of multiparticle excitations and observables. We further develop this formalism by inferring the concept of cluster additivity to the quasiparticle basis, introducing the notion of particle additivity. The term "additivity" in this context was recently introduced by Ref. [
171].
We start by recalling that the effective Hamiltonian is block diagonal and we can write the Hamiltonian operator as a sum of irreducible operators of
n quasiparticle processes
We can express the
n quasiparticle processes in second quantisation in terms of local (hard-core) bosonic operators
creating and
annihilating a quasiparticle at site
i. When considering quantum magnets like we do in this review a hard-core repulsion comes into play allowing only a single quasiparticle at a given site [
154]. For instance, in the ferromagnetic ground-state of the 2D Ising model an elementary excitation is given by a single spin flip that can be interpreted as a quasiparticle excitation [
113]. Obviously, at most one excitation on the same site is allowed. Different particle flavours
can also be accounted for by incorporating an additional index
of the operator
. To keep the notation simple, we will drop this additional index in the following. The irreducible operators in second quantisation and normal-ordered form then read
Written in normal order the meaning of these processes is directly clear when acting on states in the quasiparticle basis. The prefactors
are to be determined by applying the effective pCUT Hamiltonian (
64) to a appropriately designed cluster
C. Let us consider the quasiparticle basis on a connected cluster
C
where the number
n specifies the number of particle excitations and the indices
denote the positions of the
n (local) excitations. The effective pCUT Hamiltonian is quasiparticle conserving, so let us restrict it to
N particle states, like when evaluating its matrix elements in this basis. If we evaluate by acting on a state with fewer particles than particles involved in the process then the irreducible operator annihilates more particles than there are and the contribution is zero, that is
for
. When we determine the action of
for
this allows us to determine all prefactors
defining the action of
on the entire unrestricted Hilbert space. Second quantisation presents a natural generalisation of the Hamiltonian restricted to a finite number of particles to an arbitrary number of particles [
154]. We can construct
for
from
since the latter completely defines the action
on the entire Hilbert space.
Although everything seems fine so far and we cannot wait to set up a linked-cluster expansions, let us tell you that it is not. We finished the motivation in
Section 4.1 with the statement that we cannot simply add up contributions for energies of excited states (c.f Eq. (
53)). The reason is that although we showed that
is cluster additive, the irreducible operators restricted to the
N particle basis
are in fact not. To grasp a better understanding about the abstract concept of cluster additivity and why setting up a linked cluster expansion for higher particle channels usually fails, let us consider the following basis on a disconnected cluster
where
represents all possible
n-particle states living on a cluster
C. While there is only one way to decompose the zero-particle states
on the disconnected cluster
, one-particle states
decompose into two sets of states with a particle on cluster
A (
) and a particle on
B (
). For two-particle excitations
there are three possibilities to distribute the particles. In general,
N-particle states have the form
with
and there are
possibilities to decompose the states.
When restricting Eq. (
51) to these
N-quasiparticle states, a cluster-additive Hamiltonian must decompose as
where we introduce the notation
restricting the Hamiltonian to all
n-quasiparticle states on a cluster
C. The direct sum in Eq. (
79) is introduced to emphasise that for a cluster additive Hamiltonian there must not be any particle processes between the two disconnected clusters
A and
B. The Hilbert space on the disconnected cluster
can be seen as the natural extension of the Hilbert spaces on cluster
A and
B and we can define the operators on the clusters
A and
B in terms of the Hilbert space on
as a tensor product
The issue with Eq. (
79) is that when we restrict the particle basis to
N on the disconnected cluster
there are contributions from lower particles channels coming from the
possibilities to distribute the
N particles on the two clusters. For example, if we look at the one-particle space
we see that besides the one-particle contributions we get additional zero-particle contributions. The left part of the direct sum stems from acting on
and the right side from acting on
. There are always two possibilities to distribute the one-particle excitation on the two clusters where the other cluster is always unoccupied which gives an additional zero-particle contribution. This fact is schematically illustrated in
Figure 5a.
This is not the desired behaviour for a linked-cluster expansion. We would like the general notion of Eq. (
51) to directly translate to the particle-restricted basis which is illustrated in
Figure 5b. In words, our goal is to find a notion of (cluster) additivity in the restricted particle basis such that we can simply add up the
N-particle contributions of individual clusters without caring about lower particle channels. We define
particle additivity as
where we demand the other contributions from Eq. (
79) to vanish. The crucial thing to notice is that irreducible operators (
76) in the particle basis
are in fact particle additive [
154,
171]. In the following we will show that this is indeed the case. First, we remember that for the ground-state energy we can trivially add up the contributions. Starting from the definition for cluster additivity (
79), we have
Second, from restricting the decomposition of Eq. (
75) to the
particle channel, we can express the irreducible one-particle operator as
We recall that by calculating
, we can automatically derive
on the entire Hilbert space which subsequently defines us
. Therefore, by inserting the definition for cluster additivity (
79), we obtain
where we used the definition of Eq. (
82) in the last line. Hence, we have proven Eq. (
83) for
. The above prove can be readily extended to
.
We achieved our goal of finding a notion of cluster additivity in the particle basis which we termed particle additivity. We can determine the desired particle additive quantities by using the subtraction scheme
that comes from Eq. (
75) by restricting it to a
N particle basis. This is an inductive scheme starting from
calculating the irreducible additive quantity
. This result can be used to calculate the subsequent irreducible additive quantity
for
. Then for
we use the results from
and
to calculate
and so on. Again, it is important that
completely defines the operator
and therefore any
.
When considering effective observables, the particle number is no longer conserved and more types of processes are allowed. We need to generalise Eq. (
75) for effective observables by introducing an additional sum over
d that is the change in the quasiparticle number. An effective observable thus decomposes as
where
are irreducible contributions [
154]. When writing them in second quantisation, we have
We can directly see that the
d quasiparticles are created because there are
d additional creation operators. When
d is negative
d quasiparticles are annihilated. We can infer a notion for cluster additivity and particle additivity along the same lines
To determine the additive parts we can use an analog subtraction scheme as described in Eq. (
87), that can be denoted as
If we want to calculate
then we have to inductively apply Eq. (
92).
There are several things to be noted at this point. First, not all perturbation theory methods satisfy cluster additivity (
79) and in this case we cannot write operators as a direct sum anymore. There will be quasiparticle processes between one and the other cluster changing the number of particles on each cluster [
155,
162]. This is sketched in
Figure 6 by the presence of an additional term.
When falsely performing a linked-cluster expansion, it can be noticed immediately that the approach breaks down. A symptom of non-cluster additivity is the presence of contributions of lower orders than expected from the number of edges of the graph. When calculating reduced contributions in a linked-cluster expansion, we subtract only contributions of connected subgraphs which leaves non-zero contributions of disconnected clusters when the perturbation theory method is not cluster additive. However, there are notable exceptions when a linked-cluster expansion for energies of excited states is still correct even though the perturbation theory method is not cluster additive [
163,
164,
165,
183,
184,
185,
186]. This is only possible when the considered excitation does not couple with a lower particle channel, i.e., lower-lying states are described by a distinct set of quantum numbers lying in another symmetry sector [
155,
165,
187]. For instance, consider the elementary quasiparticle excitation in a high-field expansion for the TFIM, then the structure of the perturbation is
and therefore the first excited state does not couple directly with the ground state (there is no
which is due to symmetry). If one wants to draw a comparison, we can think of this to be similar to the case when calculating the excitations in Density Matrix Renormalisation Group. It is no problem to target an excited state if it is in a different symmetry sector than the ground state but if it is in the same symmetry sector described by same set of quantum numbers then the Hamiltonian needs to be modified to project out the ground state [
188].
Recently, a minimal transformation to an effective Hamiltonian was discovered that preserves cluster additivity. This method called "projective cluster additive transformation" [
171] can be used analogously and is even more efficient for the calculation of high-order perturbative series. In this review, however, we stick to the well established pCUT method.
4.4. Calculating Cluster Contributions
At this point, we may ask how to evaluate physical quantities on finite clusters in practice. To evaluate these quantities we must evaluate them in the quasiparticle basis. In general, when setting up a cluster expansion may it be non-linked or linked, it is important to subtract contributions from all possible subclusters to prevent over counting. Mathematically for a quantity
(
or
), this can be written as
where the sum runs over all real subclusters
in cluster
C and we call the resulting quantity
reduced. Starting from the smallest possible cluster (e.g., a single bond between two sites) this formula can be inductively applied to determine the reduced quantity on increasingly big clusters. An essential observation to make is that reduced operators vanish on disconnected clusters if the operator
is additive by construction since we subtract all contributions from individual subclusters. As the linked-cluster theorem applies, we can set up a cluster expansion
of connected clusters but we need to consider reduced quantities
to prevent over counting. For a light notation we will drop the bar in the sections below as we will only consider reduced contributions on graphs anyway.
Now we are ready to look at the problem from a more practical point of view. From the previous subsection we know how cluster additivity translates into the particle basis and how to construct particle additive parts, namely the irreducible quasiparticle contributions. We decompose the effective Hamiltonian into its irreducible contributions by explicitly calculating:
Again, consider the effective Hamiltonian in second quantisation made up of hard-core bosonic operators annihilating (creating) quasiparticles and the quasiparticle counting operator occurring in the unperturbed Hamiltonian . We also consider a connected cluster C and we denote n quasiparticle states on this cluster as with the quasiparticles on the sites to . Note, for multiple quasiparticle flavours or multiple sites within a lattice unit cell this notation can be generalised to by introducing additional indices . To lighten the notation in the following we stick to the former case. Let us consider the three lowest particle channels:
- n = 0
We can directly calculate the ground-state energy
on a cluster
C as it is already additive
as can be seen from Eq. (
95).
-
n = 1
To calculate the irreducible amplitudes
associated with the hopping process
in
, we need to subtract the zero-particle channel as can be seen from Eq. (96). However, we only need to subtract the ground-state energy if the hopping process is local,
since the ground-state energy only contributes for diagonal processes. Thus, we calculate
In the two-particle case we have to distinguish between three processes: pair hoppings (
with four distinct indices), correlated hoppings (
) and density-density interactions (
). The free quasiparticle hopping is already irreducible and nothing has to be done, but for the correlated hopping contribution we have to subtract the free one-particle hopping. In case of the two-particle density-density interactions we need to subtract the local one-particle hoppings as well as the ground-state energy as this process is diagonal (c. f. Eq. (97)). Therefore we calculate
An analog procedure can be applied for effective observables. Here, we need to determine the irreducible contributions for a fixed
d. The subtraction scheme is given by
For
we recover exactly the same subtraction procedure as before. It is straightforward to generalise this procedure for
. Let us specifically consider the example we will encounter in the next section, when calculating correlations for the spectral weight. The effective observable is applied to the unperturbed ground state
. Hence, there are only contributions out of the ground state (
) and the effective observables decomposes into
Since only processes contribute, nothing needs to be subtracted and the effective observable is irreducible and already particle additive.
Although for these types of calculations the lowest orders can be analytically determined by hand, the calculations usually become cumbersome quickly and must be evaluated using a computer programme to push to higher perturbative orders (in most cases a maximal order ranging from 8 to 20 is achievable). Such a programme reads the information of the cluster (site and bond information), the bra and ket states, the coefficient lists
or
from Eqs. (
64) and (
68) up to the desired order, the structure of the Hamiltonian
, the local
-operators from the perturbation
and if necessary the observable. The input states as well as the
-operators should be efficiently implemented in the eigenbasis of
bitwise encoding the information as known for instance from exact diagonalisation. If possible, the calculation should be done with rational coefficients for the exact representation of the perturbative series up to a desired order. The routine of the programme is then to iterate through the operator sequences from the coefficient list
or
and to consecutively apply the
-operators by systematically iterating over all bonds of the cluster and calculating the action of the operator, saving the intermediate states for the action of the next operator in the sequence. As intermediate states are superpositions of basis states, they are saved in associative containers (maps in C++ or dictionaries in python)
where
is the bit representation of a basis state. The key of the associative container is the basis state
and the associated value the prefactor
. The bitwise representation of the basis states
as well as the
-operators allow for a fast access and modification during the calculation.
So far, we have introduced the pCUT method for calculating the perturbative contributions on clusters. We demonstrated that the resulting effective quasiparticle-conserving Hamiltonian is cluster additive and showed how to extract particle-additive irreducible contributions. We introduced a subtraction scheme for the effective Hamiltonian and observables that can be easily applied to calculate additive quantities to set up a linked-cluster expansion. Finally, we briefly discussed an efficient implementation of the pCUT method.
4.5. Energy Spectrum and Observables
Having established the basic theoretical framework for the pCUT method we want to give a short overview over the physical quantities that are most frequently calculated with this approach. To this end we assume that we consider a single suitably designed cluster for the calculations instead of setting up a LCE as full graph decomposition. We will see how we can calculate the desired quantities without thinking about the abstract concepts necessary for a linked-cluster expansions and with the insights from this section it will be easier to recognise what we are aiming at. Here, we first consider the energy spectrum of the Hamiltonian. We derive both the control-parameter susceptibility as the second derivative of the ground-state energy as well as the elementary excitation gap from the effective one-quasiparticle Hamiltonian. Second, we consider observables. In the pCUT approach often spectral weights are calculated that are derived from the dynamic structure factor which is of great importance for inelastic neutron scattering experiments.
4.5.1. Ground-State Energy and Elementary Excitation Gap
Following the above described recursive scheme, we start with the zero-quasiparticle channel assuming a non-degenerate unperturbed ground state which is the situation in all applications discussed in this review. The ground-state energy can be directly calculated from the cluster as in Eq. (
100). We consider a suitable designed cluster that is large enough to accommodate all fluctuations of a given maximal order
and has periodic boundary conditions to correctly account for translational invariance. We calculate
and obtain a high-order perturbative series
in the expansion parameter
, which is valid in the thermodynamic limit up to the given maximal order
. We can extract the ground-state energy per site by dividing
by the number of sites of the cluster
. By taking the second derivative, we obtain the control-parameter susceptibility
We are usually interested in the quantum-critical properties of the model and therefore analyse the behaviour about the quantum-critical point
. The control-parameter susceptibility shows the diverging power-law behaviour
with the associated critical exponent
as we know from
Table 1.
Turning to the one-quasiparticle channel, we calculate the hopping amplitudes following Eq. (
101). Here, we use open boundary conditions, again with a cluster large enough to accommodate all fluctuations contributing to the hopping process. Note, in our notation we denote
for hopping amplitudes on a graph or cluster level and the character
a for processes in the thermodynamic limit. As for the ground-state energy we can directly infer the contribution in the thermodynamic limit if contributing fluctuations do not feel finite-size effects and thus we use
in the following. Further, we can generalise our notation to multiple particle types or larger unit cells containing multiple sites as mentioned earlier by introducing additional indices
,
. We denote a hopping from unit cell
to
with
and within the unit cells from
to
. We calculate
by fixing
, due to translational symmetry. The effective one-quasiparticle Hamiltonian in second quantisation then is
Applying the Fourier transform for a discrete lattice
we can diagonalise the resulting Hamiltonian in momentum space
introducing the operators
that diagonalise the matrix
. The eigenenergies
are the associated bands of the one-quasiparticle dispersion. In case of a trivial unit cell or a single particle flavour, the dispersion matrix becomes a scalar and we can directly express the single-banded dispersion as
where the sum over
is restricted to symmetry representatives and we assumed real hopping amplitudes
. We determine the hopping amplitudes
as a perturbative series by performing the calculations on the properly designed cluster. Note, even for a single cluster we need to perform a subtraction scheme because in order to get the irreducible contribution
explicitly we need to subtract the ground-state energy from
. When we determine the elementary excitation gap at the minimum of the lowest band
we can as well extract the gap directly as such a series. The gap closing shows a power-law behaviour
about the critical point with the critical exponent
.
4.5.2. Spectral Properties
Neutron scattering is a powerful method resolving spatial and dynamical structures in condensed matter physics since thermal neutrons have a de Broglie wavelength of similar length scale as interatomic distances and their energy is of the same order of magnitude as typical excitations [
189]. By measuring the change in momentum and kinetic energy in inelastic neutron scattering experiments determining the dynamic response, not only static properties like magnetic order can be resolved but also dynamic properties like spin-spin correlations [
190]. The dynamic response
can be determined as it is proportional to the cross section
of inelastic neutron scattering [
189,
190]. We follow the derivations in Refs. [
155,
191,
192] and start with the definition of the
dynamic response
which is the space and time Fourier transform of the spin correlation function
with
and
referring to the thermal expectation value. In the limit of vanishing temperature
, the expectation value simplifies to
with
being the ground state. Then we call
the
dynamic structure factor. We introduce a complete set of energy eigenstates
where
denotes a set of quantum numbers, for instance
, where
n is the number of quasiparticles and
the lattice momentum. Writing the dynamic structure factor in terms of these energy eigenstates, this yields the dynamic structure factor in the spectral form as a sum
where
is called
exclusive structure factor or
spectral weight associated with the quantum numbers
. We insert
into the correlation function and switch to the Heisenberg picture, where
. The spectral weight then reads
which after integrating over time t yields
If we consider the case
or some observable with
, that could be a linear combination of spin operators, then the above expression further simplifies to
By the above equations we can identify the spectral weights
2 as
These quantities are usually visualised as a heat map where the dispersion is plotted against the momentum and the value , that is the intensity of the scattering signal associated with , is colour-coded.
In the pCUT approach we want to reformulate the observable in terms of an effective one
. Here, we want to restrict ourselves to the one-quasiparticle spectral weight. In Ref. [
191] you can also find a formulation for the two-quasiparticle case. For the 1qp spectral weight,
are the quantum numbers defining one-quasiparticle states and we denote
in the following. Since the pCUT method is a perturbative approach we want to reformulate the problem in the language of
states
where we introduce the momentum states
with additional index
denoting a quantum number like a flavour of the excitation or denoting a site in a unit cell. The momentum states are defined via Fourier transform
Inserting these identities, we obtain
where we defined
. For a problem with translational invariance, we can fix the site
of the observable and introduce
, which yields
where we used
. Note, the form of the effective observable is
, exactly as denoted above in Eq. (
108). As long as the processes
and
are equivalent by means of the model symmetries we can use
and further simplify
to
So, in the pCUT method, we determine
as a perturbative series with which we can determine
. Note, this observable is already an irreducible contribution according to Eq. (
92) and no subtraction scheme needs to be performed. Because we consider an observable
the initial particle number is
and nothing needs to be subtracted. To extract the quantum-critical behaviour, we need to evaluate the expression at the critical momentum
as
yielding a perturbative series for this quantity as well. It shows a diverging critical behaviour that goes as
with the associated critical exponent
. After determining the three quantities
,
, and
described above, we can use the extrapolation techniques described in
Section 4.8 to extract estimates for the critical point
and the associated critical exponents. However, we continue with the description of a linked-cluster expansion as a full graph decomposition for long-range interacting systems. The next step on the way is to formally introduce graphs, their generation and the associated concept of white graphs.
4.6. White Graph Decomposition
In this section we first give a brief introduction to graph theory which is the basis to understand how linked-cluster expansions as a graph decomposition work. Then, we discuss how to generate all topologically distinct graphs up to a certain number of edges corresponding to the maximal number of active edges at a given perturbative order. We conclude this section by explaining the concept of white graphs where edge colours are ignored as a topological attribute in the classification of graphs. White graphs are essential for tackling long-range interactions and therefore every graph decomposition in this review is in fact a white graph decomposition.
4.6.1. Graph Theory
So far, we already defined clusters as a cut-out of the infinite lattice with a finite set of sites and a set of bonds connecting those sites. More generally, only considering the topology of clusters without restricting them to the geometry of lattices, we can define a
graph as a tuple
consisting of a (finite) set of vertices
and a (finite) set of edges
. An edge
consists of a pair of vertices
and these vertices
are called
adjacent. The
degree of a vertex is the number of edges connecting it to other vertices of the graph. In the following, we only consider
undirected,
simple, and
connected graphs which means there are neither directed edges, multiple edges between two vertices, nor loops (no edge that is joining a vertex to itself) and there always exists a path of edges connecting any two vertices of a graph [
155,
193,
194]. As an example we depict a graph
with
and
in
Figure 7.
A
subgraph of a graph
– we write
– is defined as a subset of
and
[
193,
194,
195]. We call
a proper subgraph if
and
are proper subsets of
and
.
To set up a full graph decomposition it is essential to define how to distinguish different graphs. If there exists an isomorphism between two graphs we call them
topologically equivalent otherwise they are
topologically distinct. A
graph isomorphism is a bijective map
between the vertex sets of two graphs, such that
So, an isomorphism preserves adjacency and non-adjacency, i.e.,
and
are adjacent if and only if the vertices
and
are adjacent for any vertices
[
193,
194,
196]. A special case are
graph automorphisms , which are maps of a graph on itself, so we have
In other words a graph automorphism is a permutation of the vertex set preserving adjacency [
194,
196]. The number of graph automorphisms
of a given graph gives the number of its symmetries and we call it the
symmetry number of a graph [
155,
195]. Examples of a graph isomorphism and automorphism are depicted in
Figure 7. Further, if
the mapping (
134) is injective instead of a bijective such that
and we call it
subgraph isomorphism or
monomorphism [
195,
197,
198].
An example of a monomorphism is depicted in
Figure 8a. Monomorphisms will later become of utmost importance as they give the number of embeddings of the subgraph onto a graph which is the infinite lattice.
To account for hopping processes during the embedding of graphs we need to assign additional attributes to graphs like colouring their vertices [
195]. Then, a coloured graph
is a tuple
, where
is a set of pairs
with
and
a the colour attribute. In
Figure 8b an example is depicted with
. We can extend the above definitions for isomorphisms and automorphisms and say that they must preserve the vertex colour, i.e., only vertices of the same colour can be mapped onto each other. Of course, this reduces the cardinality
and therefore the symmetry number
associated to the coloured graph. We can later exploit the colour information for matching hopping vertices of the graph with the actual hopping of the quasiparticle on the lattice. Note, also colouring edges of graphs with attributes
is useful to distinguish different types of interactions of a Hamiltonian on the graph level and very similar properties as for coloured vertices hold. As stated above two graphs are topologically distinct if there does not exist a graph isomorphism between the two graphs. Thus, for coloured graphs the colour attribute serves as another topological attribute. The importance of this will become apparent later in
Section 4.6.3. For a more elaborate overview of graph theory in the context of linked-cluster expansions, we recommend Ref. [
195].
4.6.2. Graph Generation
For the graph generation of simple connected graphs we need to define an ordering between all graph isomorphs such that it is possible to pick out a unique representative of all graph isomorphs that is called
canonical representative [
199]. One challenge lays in efficiently generating graphs as the number of connected graphs grows exponentially with the number of edges and the idea behind every algorithm must be to restrict the number of possible graph isomorph candidates when generating new edge sets by permutation. Most popular, there is McKay’s algorithm [
199,
200] that exploits the degree information of the vertices and sets up a search tree for vertices with the same degree and uses the ordering to check if the canonical representative of the current graph already exists. We recommend using open-source libraries for the graph generation and calculation of additional symmetries. For instance, there is "nauty" [
201] and "networkX" [
202] or the "Boost Graph Library" [
203].
There are various conventions for how to write the graph information to a file. One could simply save the site and edge set as lists or save its
adjacency matrix, where the rows and columns refer to the sites and a non-zero entry in the corresponding matrix element marks the existence of an edge between the sites [
155,
193,
194]. Here, we suggest to simply use a
bond list. Each entry in the list contains the edge information of the graph, denoted as
, where
with
adjacent to
e and
is just a number associated to the edge
e. The number
can be interpreted as a specific expansion parameter corresponding to this bond. In the simplest case there is just a single expansion parameter and number
for all edges. Assigning multiple expansion parameters becomes especially important in the next
Section 4.6.3. Usually, the symmetry number
of a graph is calculated on the fly when generating a set of graphs and should be saved as well. In our workflow, we save the generated graphs into bond lists and create a list containing all connected graphs (e.g., the filenames) along with their symmetry number
. These types of lists suffice for the calculation of the ground-state energy. When calculating 1qp irreducible processes for the dispersion or for the spectral weight observable, we can think of these processes breaking the graph symmetry. Therefore, after the graph generation, we consider all possible processes on a graph and assign colour attributes to the start and end vertices. Due to the symmetry of the processes we assign the same colour for hoppings and distinct colours for processes of the spectral weight.
3 We calculate the symmetry number
associated with the coloured graph. In the end, we create a list, where each entry contains the graph, its symmetry number
, a representative process (start and end vertex) and the associated symmetry number of the coloured graph
, counting the number of processes that give the same contributions as the representative process due to symmetries.
After generating the graphs and the lists, we can employ perturbation theory calculations on the graph level viewing graphs as abstract clusters with vertices as lattice sites connected by bonds. A programme as described above reads in the graph and process information and repeatedly performs the pCUT calculations on every graph. The resulting graph contributions must be added up in a way such that the information of the lattice geometry is restored by weighting the graph contributions with embedding factors. That is, how many ways a graph can be fitted onto a lattice apart from translational invariance (c.f. Ref. [
155]). For conventional embedding of contributions from models with just nearest-neighbour interactions, the number of graphs can be reduced as the lattice geometry puts a restriction on the graphs. For example, graphs containing cycles of odd length cannot be embedded on the square lattice. In contrast, for long-range interactions no such restriction exists and therefore the lattice can be seen as a single fully connected graph where every site interacts with each other. Hence, we have to generate every simply connected graph up to a given number of edges as all graphs contribute. Before we turn to the embedding problem in more detail, we will first deal with the challenges long-range interactions pose and address the problem by using white graphs.
4.6.3. White-Graphs for Long-Range Interactions
In many cases in perturbation theory there may be more types of expansion parameters, e.g., due to the geometry of the lattice as can be seen in the
n-leg Heisenberg ladder [
152]
with an expansion parameter
associated with the rungs and another one
associated with the legs. The interaction
is given by XY-interactions and the series expansion is done about the Ising limit
. Setting up a graph decomposition for this model, the canonical approach would be to associate each expansion parameter with a distinct edge colour, blue for
and purple for
.
In
Figure 9a all graphs with two edges and two colours (one for
and one for
) are depicted on the left. It is necessary to incorporate the edge colour information as the graphs can only be embedded correctly on the infinite lattice when the edge colour matches the bond type of the lattice. Another common type of perturbation problem is where the perturbation splits into different types of interactions although associated with the same perturbation parameter. For instance, the problem can look like
which is essentially the form of the Hamiltonian when performing a high-field series expansion for Kitaev’s honeycomb model [
205] where each site has one
x-,
y-, and
z-type Ising interaction bond to one of its neighbours on the honeycomb lattice, respectively [
204]. Here, we associate the three different interaction types with three types of edge colours blue, green, and purple for
x-,
y-, and
z-bonds. See
Figure 9b (left) for an illustration of all three graphs with two edges and three colours (there are only three possibilities because colours must alternate due to the constraints posed by Kitaev’s honeycomb model). In both cases the edge colour is an additional topological attribute of the graph leading to exponentially more graphs with the number of colours which becomes relevant when pushing to high orders in perturbation.
In case of long-range interactions such an approach becomes unfeasible. A Hamiltonian with long-range interactions is of the form
where
is the distance between interacting sites,
d the dimension of the lattice, and
the long-range decay exponent and
. Applying the conventional approach from above we would associate each of the infinitely many perturbations
with its own edge colour. The only obvious way to resolve this problem would be to truncate the sum over
only considering very small distances. Instead, the use of
white graphs [
152] can be a solution to problems of this kind [
25,
29]. The idea is to change our view how to tackle these Hamiltonians. We ignore the edge colours of the graph – significantly reducing the number of graphs for a given order – and instead encode the colour information in the expansion parameters on the graph level in a more abstract way. This is done by associating each edge
e of the graph
with a different "abstract" perturbation parameter
such that we can track how often
-operators acted on each edge
e yielding a multivariable polynomial
with the sum over all monomial contributions. The individual monomial contributions consist of a prefactor
and its monomial dependency
.
comprises a product of expansion parameters
associated with edge
e and their respective integer powers
tracking how often each bond was active during the calculation. Let us emphasise, we simply wrote the white graph contribution for
,
, or
explicitly as a polynomial
. It is only later during the embedding of these abstract generalised contributions when the proper link colour is reintroduced by substituting the expansion parameters by the actual expansion parameters for each realisation on the actual lattice. This is the origin of the name "white graphs" because during the calculation on the graph the colour of the links is unspecified (but encoded in the multivariable polynomial) and only during the embedding the colour of the edges is reintroduced. In
Figure 9 we illustrate the white-graph concept in the middle. On the right, we depict how to recover the polynomial of the conventional graph contribution from the abstract white graph contribution for the models in Eq. (
137) and Eq. (
138) by correctly substituting the abstract expansion parameters. The main difference between these two models (
137) and Eq. (
138) is that for different interaction types multiple parameters are associated with each edge. To account for three types of interaction flavours we have to consider three expansion parameters
with
per edge
e.
For long-range interactions we can straight-forwardly apply the exact same white-graph scheme. We substitute the abstract expansion parameter with the correct algebraically decaying interaction strength depending on the distance between the interacting sites on the lattice [
25,
29]. For this, we have to use the substitution
with
and
. It is also possible to incorporate multiple interaction types but the substitution for the different interaction types must then be done before the embedding as it was done in Refs. [
31,
35,
206].
So far, we explained how to resolve the problem of infinitely many perturbation parameters by introducing white graphs. We managed to reduce the number of expansion parameters from infinity to the number of edges, i.e., the order of the perturbation. Yet, the polynomial in Eq. (
140) still grows exponentially with the number of expansion parameters. Following the description in Ref. [
152], we can further mitigate this issue by an efficient representation of white-graph contributions. The abstract concept is to track the relevant information in a quantity
as a generalised monomial with the property
and use
as an abstract parameter that encodes the tracked information. In Eq. (
140) the monomial quantity
is just
, tracking how often each edge was active during the calculation by associating each edge with its own abstract expansion parameter. We can generalise the expression of Eq. (
109) for states comprising additional information
Instead of a simple superposition of states
with associated prefactor
as in Eq. (
109), we have an additional superposition over all monomials
comprising the information of the all distinct processes encoded in
leading to the state
. We can make use of this factorisation property by using nested containers. The key of the outer container is the basis state
and the value contains an inner container with the monomial
as the key and the prefactor
as the value. Thus, the action of a
-operator on a state
can be calculated independently of the action on the monomial
. For a flat container we would have to calculate the action on the same state
multiple times. To further improve the efficiency of the calculation we can directly encode into the programme which edge of a graph was active in a bitwise representation. Using the information of the number of edges of a graph and tracking the applied perturbative order during the calculation we can neglect subcluster contributions on the fly and reduce the computational overhead even further. Therefore, we can directly calculate the reduced contribution on the graph without the need of explicitly subtracting subcluster contributions.
Wrapping things up, we explained how the use of white graphs can be applied to models with long-range interactions resolving the issue of infinitely many graphs or expansion parameters at any finite perturbative order. Instead of associating each perturbation as another topological attribute in the classification of graphs we associate an abstract expansion parameter with each edge of a graph and only during the embedding on the lattice we substitute these expansion parameters with the actual bond information of long-range interacting sites. An efficient representation of white graphs can further help to reduce the computational overhead.
4.7. Monte Carlo Embedding of White Graphs
We discussed ways to set up a linked-cluster expansion, either by designing a single appropriately large cluster hosting all relevant fluctuations up to a given perturbative order or by setting up a linked-cluster expansion as a full graph decomposition, where the latter is the more efficient way to perform high-order series expansions. Of course, a quantity
must be cluster additive in the first place. To decompose
on the lattice
into many smaller subcluster contributions, we can add up the contributions
where
are reduced contributions on a cluster
C to prevent overcounting. It is not necessary to calculate the contribution of every possible cluster since many clusters have the same contribution. It suffices to calculate the contribution of a representative cluster, only containing the relevant topological information, that is a graph defined by its vertex set
and its edge set
. We can see a graph
as representing an equivalence class
, whose elements are all clusters
C realising all possible embeddings of a graph on the lattice [
195]. Figuratively, all elements of the equivalence class are related by translational and rotational symmetry or are different geometric realisations on the lattice. We can now split the sum over all clusters into one sum over all possible graphs on the lattice and another one over all elements in the equivalence class
where it suffices to calculate
once for all
[
195]. We are left counting the number of elements
C in the equivalence class such that we can write
where the embedding factor
is simply a number counting the number of embeddings [
195]. The important point in our line of thought is that we can calculate the quantity
only once on the graph level and multiply the contribution with a weight that is the embedding factor
and sum up the resulting contribution for all graphs to obtain the desired quantity
in the thermodynamic limit. We are essentially left with determining the embedding factors after calculating the graph contributions
. Depending on the topology of the graph, the number of possible embeddings is different and therefore also the embedding factor. When calculating one quasiparticle processes for the 1qp dispersion or the 1qp spectral weight it is important that we account for the graph vertices’ colour attributes for the definition of the equivalence class. If conventional graph contributions are considered we directly obtain the correct contribution
. If there are multiple physical parameters or different interaction flavours the graph edges have to be colour matched with the bonds of the lattice. If white graphs are used we do not have to match any colours and can be ignored but the white graph contributions have to be evaluated appropriated substituting the abstract expansion parameters with the correct physical parameter for each embedding. Regarding continuative reading on the embedding, we want to point out the standard literature for linked-cluster expansions [
152,
155,
187] as well as Ref. [
195]. Nonetheless, we show in the following how the embedding procedure works for models with nearest-neighbour interactions specifically on the example of the ground-state energy and one-quasiparticle processes yielding the 1qp dispersion of the Hamiltonian. Likewise, we derive the 1qp spectral weight. With these findings we can turn to the embedding problem of long-range interacting models and eventually describe the Monte-Carlo algorithm as a solution to this problem.
4.7.1. Conventional Nearest-Neighbour Embedding
Ground-state energy: We start with the simplest case of calculating the ground-state energy on the infinite lattice
. We know that
is cluster additive and that the ground-state energy is an extensive quantity so we can directly calculate
In words this means we have to multiply the ground-state energy contributions on the graphs with the correct embedding factor and add up the resulting weighted contributions for every graph. For the definition of the embedding factor we follow the formalism introduced be Ref. [
195] and write
The embedding factor
is the number of subgraph isomorophism (monomorphism) of the graph
on the lattice
divided by number of graph automorphisms, i.e., the symmetry number of the graph. It is necessary to divide by the symmetry number
because the number of monomorphisms is in general not equal to the number of subgraphs because there may be multiple monomorphisms that map to the identical subgraph [
195]. To properly account for this, we have to divide by the number of automorphisms.
In
Figure 10 you see an example for the embedding problem. We can recognise the fact that there are multiple monomorphisms by the presence of arrows illustrating the ambiguity of mapping onto a specific subgraph embedding. Further, we always have to consider reduced graph contributions subtracting all subgraph contributions as
Here,
is the embedding factor for subgraphs
with respect to the considered graph
. Note again, with an efficient white graph implementation reduced contributions are calculated on the fly without the need of explicit subtraction. We state the subtraction scheme for completeness. Going back to the embedding problem for the infinite lattice
, the embedding factor will be extensive as the ground-state energy is extensive as well. Usually, we calculate the ground-state energy per site
, where
N is the number of sites and fix an arbitrary edge of the graph on the lattice. We then have
where the normalised embedding factor is
denoted with a lower case
w, where
q is the coordination number of the lattice (the number of neighbours of any site) [
155,
195]. One can think about fixing sites as colouring two adjacent vertices on the graph and two adjacent sites on the lattice and considering the monomorphism with respect to the additional colour attributes (for instance
) of the coloured graph
and lattice
. In
Figure 10 we depict the number of embeddings for a given graph on the square lattice and show how the embedding factor is calculated for this example. Note, it would be equally valid to just fix a single site. Then one would not have to account for the coordination number
q and the number of monomorphism would be larger by a factor of
q making it computationally more expensive. In the end we obtain the ground-state energy (per site) as a high-order perturbative series in an expansion parameter
up to a computationally feasible maximal order
as in Eq. (
110).
1qp dispersion: We now turn to the one-quasiparticle channel and calculate the hopping amplitudes. For a quasiparticle hopping
on the lattice with additional hopping within the unit cell from
to
or a quasiparticle changing its flavour from
to
, we need to calculate
where we fix the initial vertex
and end vertex
of the quasiparticle process on the graph to the initial site
and end site
in real space. Due to translational symmetry we can fix site
. This can be formally achieved by assigning colours to the initial and end sites on the graph and on the lattice. The second sum goes over all representative hopping processes
as described above in
Section 4.6.2. The embedding factor here is similar to Eq. (
150) and reads
where the colour attribute comes from fixing the sites to the hopping. We account for the reduced symmetry of coloured graphs stemming from the representative hopping by multiplying with the symmetry number of the coloured graph
.
The embedding factor is then calculated with respect to the coloured graph
and lattice
. Again, we need to be careful as we have to consider reduced and additive contributions, i.e., we have to determine
for each contribution on a graph
. After having determined the ground-state energy per site
and the hopping amplitudes
on the lattice
, we derive the effective one-quasiparticle Hamiltonian (
113). As we have seen in
Section 4.5, we can readily derive the one-quasiparticle gap as a series in the perturbation parameter
as in Eq. (
117).
1qp spectral weight: Lastly, we can do the same for one-quasiparticle spectral weights. We calculate the process amplitudes
Note that the process
does not satisfy
in general as a quasiparticle is created at
and subsequently hops to
. On the other hand 1qp hopping processes of the Hamiltonian fulfil
as long as the hopping amplitudes are real. While for hopping processes of the Hamiltonian we can use the same colour (for start and end vertex) as a topological attribute, here we must use two different colours leading to a smaller symmetry number of the coloured graphs. On the graph level only subgraph contributions must be subtracted
With the contributions
the spectral weight can be determined with Eq. (
131) and evaluating this quantity for example at the critical momentum
, we again obtain a series (
132) in the perturbation parameter
.
For the conventional embedding problem we can also consider several generalisation, which we want to briefly mention. First, we could consider quasiparticle processes between different particle types. In a 1qp process a particle flavour could change from
to
and the graph contribution would be denoted by including the additional indices
and
. The rest of the formalism is identical. Second, we may want to consider different interaction types like in Eq. (
138) or more expansion parameter like in Eq. (
137). In such a case, when considering coloured graphs, we also have to consider their coloured edges that must be matched with the appropriate bonds on the lattice. Then, the coloured graphs
are given by the tuple
where
are the edge colour attributes and
the vertex colour attributes. Now, every embedding factor
w needs to be determined using coloured graphs with respect to
and
. When using white graphs the additional edge colour can be ignored and the embedding is as before but we need to appropriately substitute the abstract expansion parameters with the actual physical expansions parameters.
By now, we have everything together to calculate the ground-state energy, the 1qp dispersion, and the 1qp spectral weight in the thermodynamic limit from a linked-cluster expansion set up as a full-graph decomposition. Therefore, we need to calculate the embedding factor, multiply them with the associated graph contribution and add up the resulting weighted contributions. The embedding factor can be determined using available graph libraries like the Boost Graph library [
203] as only automorphisms and monomorphisms (with colours) need to be determined. For long-range interactions however, we cannot just simply calculate the embedding factor because for every graph there are infinitely many possible embeddings even when accounting for translational invariance.
4.7.2. Embedding for Models with Long-Range Interactions
In this section we only consider a single physical perturbation parameter because there cannot be an additional functional dependency on a parameter when we later perform numerical Monte Carlo summation. If we have more than one physical perturbation parameter we have to sample the additional parameters and perform the Monte Carlo summation for each parameter ratio like it was done in [
31,
35,
206] for anisotropic XY interactions, distinct ladder interactions, and for XXZ interactions. We also need to use white graphs to account for long-range interactions. In the following, we restrict to a single quasiparticle flavour due to simplicity and a trivial unit cell as we have not generalised the algorithm to larger unit cells yet. The starting point to describe the embedding problem for long-range interactions are the embedding formulas given in the previous section in Eqs. (
149), (
151), and (
154) for the ground-state energy per site, 1qp hopping processes and the 1qp spectral weight respectively, which we have to rewrite and adapt for long-range interactions.
Ground-state energy: Starting with the embedding of the ground-state energy per site (
149), we can write
We have replaced
with
in the sum to emphasise that the graphs are not restricted to the nearest-neighbour lattice geometry anymore due to the long-range interactions making the lattice a fully connected graph. From the first to second line, we decomposed the ground-state energy contribution from graphs into contributions of individual orders
. Note, the minimum order of the ground-state energy is
. From the second to third line we introduced a second sum over sites
and restricted the sum over all graphs to a sum over graphs with a fixed number of vertices
. Next, we can reformulate the embedding factor in Eq. (
147) as
where we replace the expression for the number of monomorphisms (divided by
N) with a sum over all possible configurations
. A configuration is nothing else than the current embedding of graphs. But, as we will see later, we can calculate the contributions of multiple graphs simultaneously as the Monte Carlo sum only depends on the number of sites (we use "site" also as a synonym for "vertex position"). When using the word configuration we think about it as the current set of vertex positions on the lattice. The sum over all configurations comprises individual sums for each vertex over all lattice sites excluding configurations where vertices overlap as shown in the next subsection in Eq. (180).
4 At first sight we have not gained anything from rewriting the embedding factor. However, the white-graph contributions
still need to be replaced with the correct colour, i.e., the general expansion parameters need to be substituted with the algebraically decaying long-range interaction depending on the current configuration. In reality, the contribution
depends on the current configuration
c. Thus, replacing the expression with an explicit sum is necessary as the contribution for each configuration is different and
cannot just be a number. The substitution must look like
where the index
m of the sum runs over all monomials of the contribution
and the product is over all edges
of the graph
with the adjacent vertices
. The power law in the product arises from substituting the expansion parameters
on the edges
e with the appropriate algebraically decaying long-range interaction of the current embedding (c.f. Eq. (
140)). The adjacent vertices
are embedded on the lattice sites
and
with the distance
and the multiplicity
comes from the power
of the associated expansion parameter
. This way we can reduce the many expansion parameters from the white-graph contribution to a single physical perturbation parameter
and by reordering the expression (
156) of the ground-state energy, we have
We can define
where
is the integrand function and
denotes the associated sum over all possible embeddings on the lattice that will be evaluated using a classical Monte Carlo algorithm. Since Monte Carlo runs are usually done for a batch of different seeds we introduce an additional sum over seeds averaging the Monte Carlo runs, which we denote as
Eventually, this yields the expression
To express the ground-state energy
as a perturbative series, we write
where we have to sum up the contributions of multiple Monte Carlo runs to obtain the series prefactors
for
. The zeroth prefactor
is simply given by the ground-state energy of
.
1qp dispersion: We now turn to extracting the one-quasiparticle dispersion. In Eq. (
116) we have seen that the dispersion can be analytically determined in terms of the hopping amplitudes of Eq. (
151) up to some perturbative order. For nearest-neighbour interactions it is an analytic function in
, however, for long-range interactions there are infinitely many hoppings possible at any order so we can neither explicitly determine the hopping amplitudes
nor is it possible to have the dispersion as an analytical function in
. This would introduce an functional dependence in the integrand that we cannot sample. Instead, we will have to evaluate the dispersion for certain values
to get an explicit series in the perturbation parameter
. Evaluating Eq. (
116) at
and inserting Eq. (
151), we can write
where we rewrote the sum over
by introducing the function
Again, we can split
into contributions of individual orders
and split the graph set into subsets of fixed number of sites
, yielding
Note, here the second sum goes until while for the ground-state energy it runs only until . The maximal number of sites at a given order is because graphs with edges can maximally have sites. For the ground-state energy fluctuations every quasiparticle that is created has to be annihilated again, so at order a process can only touch maximal edges which restricts the sum to sites.
Now, we argue that we can drop the sum over
by thinking differently about the embedding problem for the dispersion. The information of the start and end vertex of the hopping process is encoded into vertex colours and when finding the subgraph monomorphisms for the embedding on the infinite lattice
the colours of the vertices must match with the coloured sites on the lattice, i.e., the hopping vertices are fixed to the hopping sites on the lattice. Since the long-range interactions allow any hopping – i.e., of any distance – at any order, it is not useful to think in this picture. Instead, we should think about the embedding problem analogous to the one for the ground-state energy, where no such hopping constraint exists and the embedding factor
is simply proportional to a sum over all configurations. This is valid as we let the sum over all lattice sites and account for constraints on
by multiplying with the symmetry number of the coloured graph
. The relevant hopping information of the vertices, that was previously fixed by coloured vertices, is anyway encoded into the cosine terms. Hence, we can make the substitution
where we account for the reduced symmetry of the graph due to the hopping by multiplying with the symmetry number of the coloured graph
. As before, we need to substitute the general white-graph contribution with the actual algebraically decaying long-range interactions of the current embedding
Inserting Eq. (
168) into Eq. (
167) we end up with the expression
For a lighter notation we again define the integrand function and the Monte Carlo sum
Introducing an average over a batch of seeds for the MC sum
, we obtain
The perturbative series of the dispersion evaluated at
can then be expressed as
The sum prefactors for can be determined by summing up the individual contributions from the Monte Carlo runs.
1qp spectral weight: Last, the evaluation for the spectral weight observable is analogous to the 1qp dispersion. The integrand and Monte Carlo sum are defined as
which we use to calculate
We determine by again calculating the series prefactors for and then determine the 1qp spectral weight with .
It should be noted that we have to perform
Monte Carlo runs for a series of order with . This means the number of runs grows quadratically with the maximal order .
So far, we have derived the necessary formalism for how to express the embedding problem of models with long-range interactions but we have not talked about how to evaluate the Monte Carlo sums for the integrand functions f. In the next section we investigate how to evaluate such sums by introducing a suitable Monte Carlo algorithm.
4.7.3. Monte Carlo Algorithm for the Long-Range Embedding Problem
We are left with evaluating the Monte Carlo sum
which runs over all configurations
of graphs. The embeddings on the lattice depend only on the number of vertices of a graph
and there is no constraint by the edge set as in the nearest-neighbour case because every site of the lattice interacts with any other site making it a fully connected graph. The only restriction for models with long-range interactions is that the vertices of the graph are not allowed to overlap as they do not self interact and an overlap would imply infinite interaction strength resulting from the term
. In conclusion, we can write the sum over all possible configurations as number-of-sites-many sums over all possible lattice positions
where the primed sums
for any vertex at position
is the short notation for excluding overlaps with any other vertex position. For example, for contributions with three sites on a one-dimensional chain, this sum would look like
Due to the overlap constraint these are nested high-dimensional summations over the integrand functions which are in general hard to solve. The dimensionality of the summation is given by because higher dimensions introduce additional sums for each component. If we wanted to evaluate the Monte Carlo sum in two dimensions for contributions with 8 sites, which already occurs in eighth order perturbation theory, we would have to determine the integral value of 16 nested sums over the integrand function . This makes it clear, that the evaluation of such sums becomes challenging very quickly. In the following we use the short notation interchangeable for both the ground-state energy integrand and the 1qp process integrands .
The first approach to tackle the problem of evaluating these sums using conventional numerical integration techniques was pioneered by S. Fey and K. P. Schmidt already in 2016 [
25]. They managed to successfully determine the phase diagram and critical exponents from the closing of the gap of the long-range transverse-field Ising model on a one-dimensional chain with ferro- and antiferromagnetic interactions. While they were successful with their approach over a large range of decay exponents in one dimension, the extraction of the critical properties for small decay exponents was challenging. The two-dimensional problem was out of reach with this approach as the number of nested sums doubles and the integral converges significantly slower. Here, Monte Carlo integration came into play as it is known to be a powerful integration technique for high-dimensional problems where conventional integration fails. The reason behind the slow convergence of such high-dimensional sums is often that the configuration space where the integrand mainly contributes to the integral is significantly smaller than the entire configuration space. In 2019 Fey et al. [
29] introduced a Markov-chain Monte Carlo algorithm to efficiently sample the relevant configuration space. They were able to determine the quantum-critical properties of the long-range TFIM on two-dimensional lattices to even higher precision than previously for the one-dimensional chain extending the accessible range of decay exponents without having to forfeit perturbative orders in higher dimensions.
In the following, we describe the Markov-chain Monte Carlo algorithm introduced by Ref. [
29] to evaluate the high-dimensional nested sums. To sample the relevant configuration space efficiently, we use importance sampling with respect to some convenient probability weight
with respect to a configuration
c and the associated partition function
. We can insert an identity to Eq. (
179) or Eq. (180) and rewrite it as
where
can be interpreted as the probability of being in configuration
c. The integrand now reads as the contribution
(we dropped the order
and momentum
as indices to lighten the notation) from configuration
c multiplied with its probability, which allows us to write the sum as the expectation value
of
with respect to the weight
. We later call this sum "target sum". Since the partition function
Z is not known a priori, we introduce also a "reference sum"
over a reference function
. We require this sum to be analytically solvable to avoid introducing an additional source of error. We denote its analytical expression as
. The reference function
should behave similar to the integrand function of interest
. This means that reference sum and target sum should have considerable contributions in the same area of configuration space and their asymptotic behaviour should be similar as well. Although we could make in principle an arbitrary choice for the reference function the latter properties guarantee to lead to good convergence. In one dimension we choose the reference integrand as
with
and the reference integrand exponent
that is a free simulation parameter. We can solve the reference sum as follows
where we accounted for
possibilities to randomly embed the vertices by ordering the indices of the sums and
is the Riemann
function. One major difference between the reference and the target sum is that in the target sum many different graph contributions contribute. In fact the reference sum above is exactly the contribution of order
of a chain graph with
vertices and the contribution from the associated target sum is the same up to a linear factor. In higher dimension we cannot choose a contribution proportional to the one of a chain graph anymore since it cannot be solved analytically. Instead we make simplifications to the reference sum and require that the reference sum is still good enough to capture the same properties as the target sum. We choose to decouple the dimensions in the reference integrand
and explicitly allow overlaps in the reference sum, such that it can be solved analytically as follows
Although the exponent
can be chosen freely, we want to achieve similar asymptotic behaviour as the target integrand, therefore we choose
in one dimension as the reference sum exactly behaves like the target integrand of a chain graph. In two dimension we made some simplifications to the reference sum and we have to adopt the parameter to
for
,
for
, and
for
. This is by no means a rigorous choice but it empirically proofed to produce good convergence [
29,
207]. Solving Eq. (
183) for
Z and inserting it into Eq. (
182), we obtain
We use the analytic expression of Eq. (
185) in 1d or Eq. (
187) in 2d for the reference sum
. We got rid of the partition function
Z and now can use this expression in our Monte Carlo run to determine the sum
using the analytic expression of the reference sum
while tracking the running averages in the numerator and denominator expressions.
We are left with just one missing ingredient that is the choice of the probability weight
. For our choice to be a probability weight, it must fulfil
and we want the weight to be the largest if both the reference and the target integrand contribute the most. An obvious choice may be the quadratic mean
that is always
and rotationally invariant in
and
. However, we also want both quantities to contribute equally to the probability weight on average over all configurations. As the contributions of target and reference sum may differ significantly, we introduce a factor for rescaling
that can be estimated in an in-advance calibration run. We then use an adjusted probability weight
for the actual Monte Carlo run. The weight needs to be evaluated at every Monte Carlo step to track the running averages of the numerator and denominator in Eq. (
188).
Now, we have everything together to describe the actual Monte Carlo algorithm. We employ a Markov-Chain Monte Carlo algorithm where we need to sample the configuration space
according to the probability weight
in Eq. (
191). Each configuration with a non-zero weight must be in principle accessible by the Markov chain. On the one hand, we propose a high acceptance rate of the Monte Carlo steps to sample the large configuration space efficiently not staying in the same configuration for too long. On the other hand, we want to sample closely confined configurations with rather small distances between vertices more often than configurations that are farther apart (configurations with large distances between vertices) such that we capture the asymptotic behaviour of the model. The interaction strength decays algebraically with the distance between vertices leading to smaller contributions for configurations in which sites are far apart. What we call a confined configuration therefore depends on the decay exponent of the long-range interaction
. In the algorithm we have the free exponent parameter
(for the reference sum) and
(for probability distributions) that can be changed to tweak this behaviour but are usually chosen similar or equal to
. In two dimensions we had to adapt the values of
to obtain a similar asymptotic behaviour for reference and target sum due the approximations we made. An optimal choice of these parameters ensures fast convergence of the Monte Carlo sum. The current embedding configuration should be represented as an array container where the entries are the positions of the graph vertices. In one dimension entries are simple integers while in higher dimensions the position needs to be represented as a mathematical vector. For small decay exponents
, very large distances can occur between vertices from time to time which need to be squared when calculating the absolute distance. This can lead to an integer overflow and therefore the use of 128 bit integer may be considered. Further, we define functions in the programme for the target integrand
and for the reference integrand
, where the current configuration is passed as a parameter and the function returns the contribution from the integrand evaluated for the current configuration.
Turning back to the sampling scheme, the idea of the Markov chain is to interpret the vertex positions on the lattices as random walkers. We randomly select a graph vertex and then draw a new position from a probability distribution such that the move fulfils the detailed balance condition. In each Monte Carlo step we perform the following two moves:
-
shift move:
-
This Monte Carlo move is implemented to introduce confined random fluctuations to the current configuration independent of the strength of the algebraically decaying long-range interactions. It is especially important for larger decay exponents
when the configurations are much more likely to be confined. First, we randomly select a vertex
drawn from a discrete uniform distribution with
. Second, for the fluctuation we draw a shift value
from a discrete uniform distribution
. In one dimension we have to draw a single time and in higher dimensions we draw repeatedly for each component. Subsequently, we add the shift to the position of the selected vertex and propose the position
We might have proposed a position that is already occupied by another vertex so we have to check for overlaps. In one dimension we reset the proposed position to the original one if there is an overlap while in higher dimensions we explicitly allow overlaps. As we remember from above, this distinction is also present in the reference sums in one dimension in Eq. (
185) compared to higher dimensions in Eq. (
187). If an overlap occurs in dimensions higher than one then the target summand is explicitly set to zero such that these configurations can not contribute (otherwise the sum would become infinity). Then, we calculate the Metropolis acceptance probability
by determining the probability weights
of the current and the proposed configuration. The result of the target and reference function calls should be saved into variables to prevent redundant and expensive function calls at each Monte Carlo step. Note, the transition weights
cancel out as we draw only from uniform distributions. Last, the minimum function is implemented by drawing a random number
and we accept the proposed move if
and update the current configuration. An example of such a shift move is depicted in
Figure 11a.
-
rift move:
-
In contrast to the previous move that should introduce fluctuations to the configuration independent of the current one and independent of the long-range interaction strength, "rift moves" are introduced to better capture the correct asymptotic behaviour induced by the algebraically decaying interactions. The moves are able to propose very large distances between vertices but are also able to do the opposite closing the "rift" between vertices when the configuration is split in essentially two clusters. At first, we select a site
from the vertex set with discrete uniform probability
, explicitly excluding the last site. In one dimension we can order the vertex set such that the first vertex is the one with the smallest positional value and the last the one with the largest value, so we order by
where
are vertex indices and
,
the associated sites on the lattice. The same ordering was also done when we solved the reference sum in Eq. (
185). In higher dimensions a similar ordering comes at a much higher computational cost so we stick to the vertex numbering given by the array indices, i.e., the order is
. Here, it is also important that the vertex labelling of the reference sum coincides with the labelling of the chain graph. To capture the physical asymptotics of the system, we draw random values from a
-function distribution. In one dimension we draw from
yielding a power-law distribution with
with the free exponent parameter
. We choose
for obvious reasons. The distance to the next vertex is given by
and
is the proposed distance drawn from the
distribution. Since we ordered by the position and only selected sites in
it is sufficient to draw positive values only. We shift all indices
according to
In higher dimensions we have no such ordering and therefore extend such a distribution to negative values – we refer to it as "double-sided"
-function distribution – and draw random values from
for each component. Note, the additional one is introduced to prevent divergence when sites overlap. After drawing the new distance
we shift all vertices componentwise with
according to
The underlying idea is that if there is a large distance between two vertices
and
we can close the "rift" of the entire configuration instead of introducing a new one between
and
. The transition weights for this move are given by
With these we can calculate the Metropolis-Hastings acceptance probability in one dimension
and likewise in higher dimensions
As above, we randomly draw
, accept if
, and update the current configuration if the proposed configuration is accepted. In
Figure 11b you can find a typical rift move illustrated.
Figure 11.
Exemplary Monte Carlo moves for a linear graph on a two-dimensional lattice. a During a shift move a vertex is selected randomly from a uniform distribution. Then a shift vector is drawn (uniformly for each component) that moves the selected site to if accepted. b For rift moves a vertex is selected from . Instead of drawing from uniform distributions, rift moves account for correct asymptotic behaviour of the system by drawing the a new distance to the next vertex from a -function distribution (From a normal function in 1d and from a double-sided function in higher dimension for each component). If accepted, the vertices are shifted to the new position .
Figure 11.
Exemplary Monte Carlo moves for a linear graph on a two-dimensional lattice. a During a shift move a vertex is selected randomly from a uniform distribution. Then a shift vector is drawn (uniformly for each component) that moves the selected site to if accepted. b For rift moves a vertex is selected from . Instead of drawing from uniform distributions, rift moves account for correct asymptotic behaviour of the system by drawing the a new distance to the next vertex from a -function distribution (From a normal function in 1d and from a double-sided function in higher dimension for each component). If accepted, the vertices are shifted to the new position .
To implement the Monte Carlo algorithm we just have to introduce a repeating loop. For each loop iteration we perform a Monte Carlo step consisting of those two moves. From trying out different probabilities we set the probability to perform a shift move to
and for the rift move we set it to
[
29,
207].
5 After performing a move – accepted or not – we update the estimate for the target and reference sum in the numerator and denominator in Eq. (
188), respectively, and keep track of the statistics like the variance of target and reference sum. To determine the value
of the Monte Carlo run, the ratio of both quantities has to be multiplied with the analytical value of the reference sum as in Eq. (
188).
Further improvements to the algorithm can be made by recentering the configuration to the origin since the graphs on the lattice may drift away due to the random walk. Also, for integer powers occurring in the target integrand from powers of the expansion parameters, it may be better to use plain multiplication than using generic integer power functions. Most importantly, lookup tables should be used for the non-integer powers stemming from the algebraically decaying interaction strengths for distances within a cut off occurring most commonly. The integrand function also depends on as cosine terms are present. It is useful to compile the code for a desired value so lookup tables can be defined at compile time due to the periodicity of the cosine. For instance, in the simplest case the cosine is always 1 or for only values 1 and can occur.
Let us emphasise that for the pCUT+MC approach we need to perform individual runs for every perturbative order, for every possible numbers of sites at a given order, for different values of momentum if necessary (e.g., for a dispersion), and for different seeds so that an average value can be calculated. Typical runs are performed with -20 seeds for 12-24 hours. For instance, for the long-range transverse-field Ising model we were able to a extract perturbative series up to order 10 for the 1qp spectral weight, order 11 for the 1qp gap and order 13 for the ground-state energy. Much higher orders in perturbation are likely not feasible with the current implementation as the number of Monte Carlo runs scales quadratically with the order. In the future, to further improve the efficiency of the approach, one may come up with additional Monte Carlo moves for the Markov chain that can change the number of sites. As a result the algorithm would only scale linear with the perturbative order but potentially at the cost of slower convergence. Maybe a compromise would be a favourable option, where the lower orders that converge faster are computed using an algorithm with the additional move and higher orders with the algorithm presented above.
Another important issue worth mentioning is the fact that so far the above algorithm can be only be applied to lattices with a trivial single-site unit cell. To generalise the algorithm to arbitrary lattice geometries in the future, we would need to introduce another Monte Carlo move. We would keep moves introduced above for moving the vertices along the underlying Bravais lattice. A new move then changes the position within the unit cell. However, new subtleties emerge from introducing a larger unit cell, as we would have to fix two vertices within the unit cells for hopping processes while the remaining vertices can be moved freely. A larger unit cell also means that we have to calculate the entire matrix of the dispersion as can be seen in Eq. (
115). Entries of the matrix are in general complex valued which needs to be accounted for as well. Lastly, the Monte Carlo dynamics of the system is changed due to the additional Monte Carlo move which may impact the convergence. So, the choice of the reference sum should probably also be adapted to achieve the desired convergence. See also the discussion in Ref. [
207]
4.8. Series Extrapolation
We are interested in extracting the quantum-critical properties from the perturbative series obtained from the Monte Carlo embedding. DlogPadé extrapolations are an established and powerful method that allows us to extrapolate high-order series even beyond the radius of convergence and determine the quantum-critical point and associated critical exponents. A more elaborate description of DlogPadés and its application to critical phenomena can be found in Refs. [
208,
209].
We have given a high-order perturbative series of a physical quantity
in the perturbation parameter
. A Padé extrapolant is defined as
where
and the degrees
L,
M of the numerator polynomial
and denominator polynomial
are restricted to
, where
is the maximal perturbative order. The coefficients
and
are fixed by as set of linear equations by cross multiplying Eq. (
202) with
and requiring
, i.e., that all higher order terms must vanish on the left side of the equation [
208]. We introduce
as the Padé extrapolant of the logarithmic derivative of
that must satisfy
, as we lose one perturbative order due to differentiation. The DlogPadé extrapolant of
can now be defined as
Given that the quantity of interest
shows a second-order phase transition with a dominant power-law
about the critical point
, we can extract the critical point
and the associated critical exponent
with DlogPadé extrapolants.
6 We can determine estimates for the critical point
by analysing the poles of the extrapolant
. We have to identify the physical pole whose position determines
and exclude spurious extrapolants that have non-physical poles in the complex plane close to the real line with
. If
is known, we can define biased DlogPadés by
Here, also defective extrapolants have to be removed by excluding all extrapolants that have poles in the vicinity of
. We can extract estimates for the critical exponent
by calculating the residua. For an unbiased estimate we calculate
and for the biased estimate we calculate
Biased DlogPadé extrapolants were also used in the past trying to extract the exponent of multiplicative logarithmic corrections at the upper critical dimension [
25,
29,
31,
35]. At the upper critical dimensions there are corrections to the dominant power law behaviour of the form
We can bias the critical point
and the critical exponent
to the known mean-field value and define the extrapolant
so we can determine the estimate
by calculating the residuum of the Padé extrapolant.
In practice we have to calculate several extrapolants to obtain reliable estimates for the critical properties of interest. We calculate estimates that satisfy
. We exclude all defective extrapolants as briefly described above and arrange the remaining DlogPadés in families with
. Usually we only allow
or
since more diagonal families are expected to converge faster to the real physical value as a function of the perturbative order
[
208]. We further take the mean over the highest order extrapolants of each family that has at least two or three extrapolants to obtain an estimate for the critical point and exponents. The uncertainty obtained from averaging over the extrapolants is by no means an rigorous error but is rather a "subjective" measure for the uncertainty obtained from systematically analysing extrapolants [
209].
4.9. Workflow of Series Expansion Monte Carlo Embedding
In the previous subsections we comprehensively described the pCUT+MC method for models with long-range interactions. The approach in its entirety is quite involved. Hence, to conclude this section, we want to give a short overview over the individual steps necessary and the workflow associated. The workflow is sketched in
Figure 12.
The approach starts with the generation of graphs with a software like "nauty" [
201] and saving the graphs to bond files. The symmetry number
of the graph can be obtained on the fly as it is usually a byproduct of searching for further graph isomorphs. For the 1qp dispersion or spectral weight we want to calculate quasiparticle processes on these graphs, however we do not want to calculate every process on a graph since many are related by symmetries and give the exact same contribution. Therefore, we calculate additional symmetry numbers
of coloured graphs. In a programme we iterate over all possible vertex pairs and colours of the associated vertices (two colours for the 1qp spectral weight and one colour for 1qp processes of the Hamiltonian), choose a representative process and save the associated vertices and the symmetry number of the coloured graph to a list. While for the calculation of the ground-state energy, a list containing the graph names suffices, for particle processes we want to have a list that contains the graph name, symmetry number, start and end vertex of the process, and a symmetry number associated to the process counting the number of equivalent processes. On the upside, it suffices to generate these lists and graphs once.
The second step is to iterate over the list entries, read the graph as a cluster, read the associated input states and apply the pCUT method. Of course, such a programme must also read the pCUT coefficients or for observables and the model information, i.e., the -operators. The programme iterates over the different operator sequences, through the operators from each sequence and systematically iterates over all edges of a graph and applies the -operators. We additionally associate each edge of a graph with its own expansion parameter, such that we obtain the white-graph contributions.
Having calculated the white-graph contributions over every representative process for every graph we need to convert the white graph contributions to an integrand function – also embeddable function – that is a callable function in our Monte Carlo algorithm. We have a programme that substitutes the expansion parameter with the algebraical decay expression (). This must be an expression in the programming language of choice, such that the Monte Carlo programme can call this function. Further, we multiply the white-graph contribution with the symmetry factor for 1qp processes and with for the ground-state energy to properly account for the symmetries during the embedding. Also all white-graph contributions of a given order and number of vertices are grouped into one such function. The functions are saved in a file and included in the Monte Carlo algorithm code. If the white graph contributions are associated to models with different interaction flavours, we must sample for certain parameter values during this step.
The physical parameters are also fixed like the momentum
, the decay exponent
and of course the code must be adapted for different lattice geometries. There are also simulation parameters like the number of seeds, the ratio between shift and rift moves, and the simulation exponents
of the reference sum and
for the rift move
-function distribution. The Monte Carlo algorithm is then executed for each embeddable function with fixed order, number of sites and seed, yielding the target values of Eq. (
188).
We average the target values for multiple runs with different seeds and add them up according to Eqs. (
164), (
174) and (
177) and obtain a perturbative series in
. Afterwards we employ DlogPadé extrapolation to extract critical quantities of interest like the critical point
and the associated critical exponent
for the ground-state energy,
for the 1qp gap or
for the 1qp spectral weight. We can also use the extrapolations to construct the dispersion close to the critical point.
5. Stochastic Series Expansion Quantum Monte Carlo
In this section we discuss the method of stochastic series expansion (SSE) QMC. This class of QMC algorithms is closely related to path integral (PI) QMC and samples configurations according to the Boltzmann distribution of a quantum mechanical Hamiltonian. This sampling is achieved by extending the configuration space in imaginary-time direction by operator sequences. The objective is to evaluate thermal expectation values for operators at a finite temperature on a finite system.
The canonical partition function of a system with a quantum mechanical Hamiltonian
can be expressed as
with
being the inverse temperature and the sum over an arbitrary orthonormal basis
. The task is to bring Eq. (
210) into the form of
with all weights
required to be non-negative.
Of course, for a Hamiltonian that is traced over in its eigenbasis (or a classical system), Eq. (
210) is already in the form of Eq. (
211) and the system can be directly sampled by a Metropolis-Hastings algorithm (see Eq. (
46)). For a general quantum mechanical problem, we do not have access to the eigenstates of a system and require a reformulation of Eq. (
210).
The SSE QMC resolves this issue in the following way: Given a Hamiltonian
, a computational orthonormal basis
is chosen in which the trace is evaluated. Further, there should exist a decomposition of the Hamiltonian
into operators
.
and
are chosen such that the following two conditions are met:
No-branching rule:
ensuring that no superpositions of basis states are created by acting with
.
Non-negative real matrix elements in the computational basis:
The second condition is not strictly necessary but makes sure that no sign problem arises which would lead to exponentially hard computational complexity. In general, it is not necessarily possible to find a computational basis in which this condition can be fulfilled for all Hamiltonians. However, if the negative matrix elements contribute in such a way that they always occur in pairs and the minus signs cancel, the condition can be relaxed without inducing a sign problem. We will encounter this case for the antiferromagnetic Heisenberg models in
Section 5.2.
For operators
that are diagonal in the computational basis, the conditions (
213) and (
214) never pose a problem as diagonal operators intrinsically obey the no-branching rule and can always be made non-negative by adding a suitable constant to the Hamiltonian. The main difficulty is to find a computational basis in which the off-diagonal matrix elements are non-negative, which is not necessarily possible as mentioned above. In particular, for fermionic or frustrated systems, negative signs typically occur.
In order to reformulate Eq. (210) in the form of Eq. (211), a high-temperature expansion for the partition function is performed
In general, the evaluation of
is not feasible. The way SSE tackles this expression is by expanding it as
as a sum over all occuring operator sequences
resulting from the exponentiation. The additional dimension created by the operator sequence is usually referred to as imaginary time in analogy to path-integral formulations. Inserting Eq. (
218) into Eq. (217), one obtains
Note that each of the summands in Eq. (
219) is non-negative by design due to condition (
214) and can be interpreted as the relative weight of a configuration. By comparing Eq. (
219) with Eq. (
211), we see that it is of a suitable form for a Markov chain Monte Carlo sampling. We identify the direct product of the set of all basis states
with the set of all sequences as the configuration space
The weight of a configuration is given by
In the next step, we discuss the structure of each configuration consisting of a computational basis state
and an operator sequence
. Regarding Eq. (
219) we stress that the action of the product of operators in the sequence onto the basis state is crucial. Due to the no-branching rule (see Eq. (
213)), the action of
onto the basis state can be interpreted as a discrete propagation of the state
in imaginary time according to the operators in the sequence. We define the state at propagation index
to be
Due to the periodic boundary condition of the trace and the orthogonality of the basis states
, only sequences for which
have a non-zero weight (see Eq. (
221)).
At this point we can demonstrate, why it does not matter if some matrix elements are negative in some instances, e.g., for some antiferromagnetic spin models on bipartite lattices. If a matrix element of an operator is negative but it is assured by the Hamiltonian and the periodicity of the trace, that there is always an even number of negative matrix elements in a sequence with non-vanishing weight, then the definition of weights as in Eq. (
221) is nevertheless possible.
In the discussion so far we considered sequences of all possible lengths. In order to formulate algorithms sampling the configuration space efficiently, a scheme with a fixed sequence length
can be introduced, in which all sequences with
are padded with identity operators
to length
and all sequences
are discarded. The physical justification to discard all sequences above a certain fixed length
is that they are exponentially suppressed for sufficiently large
. In short, this is the case because the mean operator number
is proportional to the mean energy
and the variance is related to the specific heat in the following way
A derivation and discussion of these statements can be found in Refs. [
39,
210,
211,
212,
213]. From Eq. (
223), we can infer that the infinite sum over all sequence lengths can be truncated at a finite
. From rearranging Eq. (
223) and using the extensivity of the mean energy, the mean sequence length scales as
proportional to the inverse temperature and system size
N. As the mean sequence length has a finite value, the idea is to choose an
great enough to be able to sample all but a negligible amount of operator sequences. Further, we will argue that the introduction of a large enough cutoff results in an exponentially small and negligible error. In the limit of
, the specific heat has to vanish. Therefore, the variance of the mean sequence length is proportional to
. From this, it is concluded that the weights of sequences vanish exponentially for large enough sequence lengths
n [
212].
We therefore introduce a large enough cutoff for the sequence length
and consider operator sequences with a fixed length
. The expression for the partition function can then be written as
The new configuration space includes all sequences of length
where the shorter sequences are padded by inserting unity operators. The random insertion of unity operators into a sequence of
non-trivial operators results in
sequences of length
. The modified prefactor in Eq. (
224) accounts for this overcounting. Although Eq. (224) is strictly speaking an approximation to the partition function, we want to note that in practice this does not cause a measurable systematic error as
can be chosen large enough in a dynamical fashion such that during the finite simulation time no sequence of length
would occur. Therefore, we will not consider the fixed-length scheme as an approximation below.
To summarise: Following the argumentation discussed in this section, it is possible to bring any Hamiltonian into the form of Eq. (
224), whereby all the summands are non-negative if a suitable decomposition and computational basis can be found. The next step is to implement a Markov chain MC sampling on the configuration space. As the configuration space largely depends on the model, the sampling is done in a model-dependent way. In
Section 5.1, we will introduce an algorithm to sample the (LR)TFIM. In
Section 5.2, we will further describe an algorithm to sample unfrustrated (long-range) Heisenberg models. In
Section 5.3, the measurement of a variety of operators is discussed. Further, we will discuss how to sample systems at effectively zero temperature in
Section 5.4 as this review considers the zero-temperature physics of quantum phase transitions. In
Section 5.5, we give a short overview of other MC algorithms for long-range models based on path integrals.
5.1. Algorithm for Arbitrary Transverse-Field Ising Models
In this section we describe the SSE algorithm to sample arbitrary TFIMs of the form
as introduced by A. Sandvik in Ref. [
39]. The Pauli matrices
with
describe
N spins 1/2 located on the lattice sites
. The transverse-field strength at a lattice site
i is
and the Ising couplings between sites
i and
j have the strength
. For
, the interaction is antiferromagnetic and it is energetically favourable for the spins to anti-align, while for
the interaction is ferromagnetic and it is energetically favourable for the spins to align.
Choosing the
-eigenbasis
for the SSE formulation avoids the sign problem for arbitrary Ising interactions. The Hamiltonian is decomposed using the following operators
We call
trivial operator,
field operator,
constant operator, and
Ising operator. The operator
is associated with the site
i, even though it is proportional to
. With these operators we can rewrite Eq. (
225) up to an irrelevant constant as
Note, Eq. (
230) is a sum over field, constant, and Ising operators. The constant operators are not part of the original Hamiltonian, but will be relevant for algorithmic purposes. The trivial operators (Eq. (
226)) are not relevant to express the Hamiltonian, but are necessary for the fixed-length sampling scheme. The proposed decomposition fulfils the no-branching property and there are no negative matrix elements of the operators in the computational basis due to the positive constant
that is added to the Ising operators. The possible matrix elements for Ising operators
acting on a pair of spins
i,
j are given by
This implies that only sequences where (anti)ferromagnetic Ising bonds are placed on (anti)aligned spins have a non-vanishing weight.
The partition function (Eq. (
224)) in the fixed-length scheme reads
The propagated states only change in the imaginary-time propagation if a field operator, the only off-diagonal operator in the chosen basis, acts on it. A configuration only has a non-zero weight, if all the Ising operators in the sequence are placed on spins which have the correct alignment. Constant operators are included in the Hamiltonian for algorithmic purposes which will become important in the off-diagonal update described below.
Before going into the description of the Markov chain to sample the configuration space, it is illustrative to visualise a configuration with non-vanishing weight defined by a state
and an operator sequence
. In
Figure 13, an exemplary configuration with non-vanishing weight is illustrated for the one-dimensional TFIM and visualisations of the different operators Eqs. (227)–(229) in such a configuration are shown.
Before going into the details of the algorithm, we introduce the main concept of the update scheme and the crucial obstacles that are encountered in setting up an efficient algorithm [
39]. Each step of the Markov chain sampling of configurations is done by performing a so-called diagonal update followed by an off-diagonal quantum cluster update [
39]. In the diagonal update, one iterates over the sequence and exchanges trivial operators with constant or Ising operators and vice versa while propagating the states along the sequence. In the off-diagonal update, constant operators are exchanged with field operators while preserving the weight of the configuration. The main obstacle that is circumvented with this update procedure is the following: It is a non-trivial and non-local task to insert off-diagonal operators into the sequence without creating a configuration with vanishing weight. The first problem is that one cannot simply insert a single field operator into the sequence as it breaks the periodicity of the propagated states
leading to a vanishing weight due to the orthonormality of the computational basis. Therefore, field operators can only occur in an even number for each site to preserve the periodicity in imaginary time. The second issue involves Ising operators placed on a pair of sites
i and
j. If one places a field operator at one of the sites before and behind the Ising operator in the sequence, this preserves the periodicity in imaginary time but the matrix element of the Ising operator becomes zero as the spins will be misaligned with respect to the sign of the Ising coupling (see Eq. (
231)). These issues can be tackled by a non-local off-diagonal update [
39], which we will thoroughly discuss after the diagonal update.
In the diagonal update, the number of non-trivial operators
n is altered by exchanging trivial operators with constant and Ising operators in the operator sequence and vice versa. This does not change the states
, including the state
. Starting from propagation step
and state
, one iterates over the sequence step-by-step and conducts the exchange of trivial operators with non-trivial diagonal operators. If a field operator is encountered at the current propagation step, the state is propagated by flipping the respective spin and the iteration is proceeded. If a diagonal operator is encountered, a local update following the Metropolis-Hastings algorithm as described in
Section 3 is performed and the total transition probability is made of a proposition probability
and an acceptance probability
.
If a trivial operator is encountered at propagation step
k, a non-trivial diagonal operator gets proposed with the probability
taking into account the weight
of the proposed operator with the normalising constant
C
The weight is essentially given by the matrix elements of the respective operators
with the special case that the Ising operators are a priori handled as if they were allowed. At a later stage, it is checked if the spins are correctly aligned at the considered propagation step
k and state
and the operator gets rejected if this is not the case. This has the benefit that one does not need to check every single bond for correct alignment for the insertion of a single operator, which scales like
for long-range models. This sampling can be done in a constant time complexity in the number of elements in the distribution by using the so-called walker method of aliases [
214]. Instructions on how to set up a walker sampler from a distribution of discrete weights and how to draw from this distribution can be found in
Appendix C or in Ref. [
214]. Drawing from the discrete distribution of weights, an operator
gets proposed to be inserted. On the other side, if a non-trivial diagonal operator is encountered, it is always proposed to be replaced by a trivial operator
The acceptance probabilities are then chosen according to the Metropolis-Hastings algorithm in Eq. (46). For a constant field operator
this gives the acceptance probability
Similarly, for an Ising operator
one has
up to a factor
that is either 1 or 0 depending on if the spins at the current propagation step
k are correctly aligned or misaligned. An Ising bond with misaligned spins would lead to a vanishing weight of the configuration and is therefore not allowed.
Up to this factor, the acceptance probability is the same independent of the type of non-trivial diagonal operator that is proposed to be inserted as the operator weight arising in the weight for the newly proposed configuration cancels with the same factor in the proposition probability. This is because we already chose to propose an operator by considering its respective weight . As the acceptance probabilities do not differ, one can therefore also first check if one accepts to insert any non-trivial diagonal operator and only if this is accepted, draw the precise operator to be inserted. Of course, if the chosen operator is an Ising operator, it still has to be checked if the spins are correctly aligned to prevent a non-vanishing weight.
The cancellation of the operator weights
also makes it easier to do the reverse process and replace a non-trivial diagonal operator with a trivial one in the sense that the acceptance probability for inserting
does not depend on the current non-trivial diagonal operator
.
If a proposition gets rejected, the iteration along the sequence continues and the procedure starts again for the next operator. After each diagonal update sweep (iteration over the whole sequence) during the equilibration, trivial operators are appended to the end of the sequence such that
[
39]. This allows to dynamically adjust the sequence length
for the fixed-length scheme to ensure a sufficiently long sequence.
For an efficient implementation of the off-diagonal quantum cluster update, it is crucial to introduce the concept of operator legs. An operator at position
p in the sequence has legs with the numbers
. For an Ising operator, these legs are associated with two legs per site, one upper and one lower leg (see
Figure 14a).
For a constant or field operator, only two of the four legs are real vertex legs since these operators act only on a single site (see
Figure 14b). The remaining two legs are called ghost legs and are considered solely due to algorithmic reasons as it is numerically beneficial to let every operator have the same number of legs. This allows to calculate the propagation index
p of an operator from the leg numbers using integer division with
. Trivial operators can be described with four ghost legs or can be ignored entirely in the sequence for the off-diagonal update.
For the chosen representation of the Ising model within SSE, one can subdivide the configuration into disjoint clusters that extend in space as well as in imaginary time with Ising operators acting as bridges in real space and with constant and field operators acting as delimiters in imaginary time. These clusters of spins can be flipped by replacing the delimiting constant operators with field operators and vice versa without changing the weight . If the cluster winds around the boundary in imaginary time, the respective spins in state have to be updated as well. In order to flip half of the configuration on average and get a good mixing, all clusters are constructed and the probability to flip a specific cluster is chosen to be for each cluster separately.
For the construction of the clusters, the propagated spin states along imaginary time are not needed. The entire problem can be dealt with in the language of vertices with legs. The relevant information for the cluster formation is which legs are connected. This information about the connection between legs of different operators is stored in a doubly-linked list. The list is set up in the following way: At the index of a vertex leg
i we store the index
j of the leg it is connected to, i. e.
and vice versa
. This segmentation is illustrated for an exemplary configuration in
Figure 15 together with the doubly-linked leg list for the depicted configuration. The doubly-linked list can already be set up during the diagonal update when the sequence is traversed either way. An efficient algorithm to set up the data structure is described in Refs. [
213,
215].
In comparison to general off-diagonal loop updates [
215,
216,
217], the formation of clusters in the presented off-diagonal cluster update for the TFIM is fully deterministic. The whole configuration is divided into disjoint clusters, all of which will be build and flipped with probability
. It is beneficial to already decide whether a cluster is flipped or not before constructing the cluster so the constant and field operators and spin states can already be processed during the construction. A leg that is processed during the construction of a cluster is marked as visited in order to not process the same leg twice. This is also the reason why all the clusters have to be constructed even if they will not be flipped as otherwise the same cluster would get constructed starting from another leg later on. Further, we introduce a stack
7 for the legs which were visited during the construction but whose connections are yet to be processed.
The formation of each cluster starts by choosing a leg which has not yet been visited in the current off-diagonal update. At the beginning of the cluster formation, it is randomly determined if the cluster is flipped or not. If the leg corresponds to an Ising vertex, the cluster branches out to all four legs of the vertex. This means, all four legs of the operator are put on the stack. If the leg corresponds to a constant or field operator, the type of the operator is exchanged if the cluster is flipped and only this primal leg is put on the stack. Next, the following logic is repeated until the stack is empty. We pop a leg l from the stack. If the leg has been visited already, we continue with a new leg from the stack. Else, we determine the new leg using the doubly-linked list. We mark both legs l and as visited. If one passes by the periodic boundary in imaginary time while going from leg l to and the cluster will be flipped, the corresponding site belonging to the legs in the state has to be flipped. If belongs to an Ising operator, we add all legs of the Ising operator which have not been visited yet to the stack. If belongs to a constant or field operator, we exchange the operator if the cluster is said to be flipped. This procedure is repeated until the stack is empty. After that, one proceeds with the next cluster starting from a leg that has not yet been visited in the current cluster update. If no such leg is left, the cluster update is finished.
In addition to this cluster update, spins which have no operators acting on it in the sequence can be flipped with a probability of (thermal spin flip).
In summary, the off-diagonal update exchanges field and constant operators with each other and changes the state . Combining the diagonal update with the off-diagonal cluster update samples the entire configuration space.
5.2. Algorithm for Unfrustrated Heisenberg Models
In this section we describe the algorithm to sample arbitrary unfrustrated spin-1/2 Heisenberg models with SSE [
213]. By unfrustrated Heisenberg models we mean Hamiltonians
written as a sum over three-component interactions between two sites
and
connected by bond
b, where each bond is either ferromagnetic (F) with
or antiferromagnetic (AF) with
, with the property that there is no loop of lattice sites connected by bonds which contains an odd number of antiferromagnetic bonds, as this would lead to frustration. The spin operators are a compact notation for a three-component vector of spin operators
. The coupling strengths
can have a priori arbitrary amplitudes. It is crucial to look at unfrustrated Heisenberg models in order to define non-negative weights for configurations as the off-diagonal components of the antiferromagnetic operators have a negative matrix element. In an unfrustrated model, these matrix elements always occur in an even number of times in the operator sequence of any valid configuration, which makes it possible to construct a non-negative SSE weight [
213].
For the SSE algorithm the
-eigenbasis
is chosen, but the
- or
-eigenbasis would work the same way. The Hamiltonian is decomposed into
using the following operators
The diagonal operators
and
in Eqs. (
245) and (247) are defined in the same fashion as for the TFIM up to the factor of
due to the usage of spin operators
instead of Pauli matrices
with
. The contribution to the weight of a sequence of these operators is either
if the bond fulfils the (anti)ferromagnetic condition and zero else. Although the expressions for the ferromagnetic and antiferromagnetic bonds look the same, we distinguish between these two bonds to highlight that these objects behave differently within the off-diagonal update.
As for antiferromagnetic bonds, the off-diagonal operators do not fulfil the non-negativity of matrix elements (see Eq. (214)) in the computational basis. Therefore, they must always appear in an even number of times in the operator sequence to avoid the sign problem. For the ferromagnetic bonds this restriction is not necessary.
Analogous to the TFIM, the operator sequence additionally contains trivial operators
, which are not part of the Hamiltonian but are used to pad the sequence to a fixed length
. In contrast to the TFIM, there is no need for further artificial operators like the constant field operators in the TFIM used to limit the cluster in the cluster update. In the case of the Heisenberg model, the non-local off-diagonal update is constructed in the form of loops instead of cluster with several branches that need to be limited in imaginary time. An exemplary SSE configuration for a Heisenberg chain using the decomposition from above can be seen in
Figure 16.
Similar to the sampling of the TFIM in
Section 5.1, each step of the Markov chain sampling of configurations is done by performing a diagonal update followed by a non-local off-diagonal update. The diagonal update exchanges trivial operators with diagonal operators while the off-diagonal update exchanges diagonal bond operators with the respective off-diagonal bond operators.
Due to the structure of the diagonal operators (see Eqs. (
245) and (247)) similar to the TFIM, the diagonal update is performed similar to the diagonal update of the TFIM described in
Section 5.1. Nonetheless, to make the description of the algorithm for unfrustrated Heisenberg models self-contained, we recapitulate the diagonal update and adapt it to the Heisenberg case.
The diagonal update exchanges trivial operators with diagonal operators in the sequence and vice versa. It will therefore not change the states
, including the state
. Starting from
at
, the sequence is again iterated over by
k. If the operator at the current position
k in the sequence is an off-diagonal operator operator, the state
is propagated by applying the off-diagonal operator and the iteration is proceeded. If one encounters diagonal or trivial operators in the sequence, the Metropolis-Hastings algorithm (see
Section 3) is performed and the transition probability to an altered configuration is again split into a proposition probability
and an acceptance probability
.
Analogous to the algorithm for the TFIM, if a trivial operator is encountered, a non-trivial diagonal operator gets proposed with the probability
taking into account the weight
of the proposed operator and the normalising constant
C
The diagonal update for the Heisenberg model only differs from the one of the TFIM by the detailed probabilities
and the normalising constant
C. The factor of
in the bond weights in comparison to the Ising model comes from using spin operators in contrast to Pauli matrices in the Hamiltonian. This sampling according to the weight
can be done in a constant time complexity in the number of elements in the distribution by using the walker method of aliases [
214]. An introduction on how to set up a walker sampler from a distribution of discrete weights and how to draw from this distribution can be found in
Appendix C or in Ref. [
214]. Drawing from the discrete distribution of weights, an operator
gets proposed to be inserted. On the other side, if a non-trivial diagonal operator is encountered, it is always proposed to be replaced by a trivial operator
exactly as for the TFIM. The acceptance probabilities are then chosen according to the Metropolis-Hastings algorithm Eq. (46). The respective Metropolis-Hastings probabilities with which the proposed replacements at position
k in the operator sequence are accepted are given by the ratios of the configuration weights before and after the potential replacement and the ratio of the proposition probabilities
As in the case of the TFIM, the specific matrix element for the propagated states and has to be considered, which is either and cancels with this factor in the acceptance probability or is 0 and the acceptance probability vanishes. Intuitively, this factor simply checks if the spins are correctly aligned or misaligned with respect to the sign of the coupling and only allows the operator to be inserted if the spins and of the propagated state are correctly aligned. If the proposition is accepted, the replacement of the operator at index k is performed and one continues with the next propagation index in imaginary time. If the proposition is denied, the current operator remains and one continues with the next propagation index in imaginary time.
As for the TFIM, a fixed-length scheme is used and the sequence length
limiting the amount of non-trivial operators
n has to be dynamically adjusted. During the thermalisation of the sampling procedure, trivial operators are appended to the end of the sequence after each diagonal update (iteration over the whole sequence) such that
[
39].
In order to efficiently implement an off-diagonal loop update, it is crucial to introduce the concept of operator legs. An operator at position p in the sequence has four legs with the numbers , two legs (an upper and a lower leg) for each of the spin sites. Trivial operators can be described with four ghost legs or can be ignored entirely in the sequence for the off-diagonal update.
For the unfrustrated Heisenberg models, the off-diagonal update is done using a loop update. During the update, loops are constructed in the space-time configuration (see
Figure 16 for an example) and the spin states on the loop are flipped along the way, which changes the operator vertex types. Thereby, diagonal operators become off-diagonal operators and vice versa. It is noteworthy, that the weight of a configuration is not modified by this since the weights of diagonal and off-diagonal operators are the same in the isotropic Heisenberg model. Therefore, flipping loops with a fixed probability satisfies detailed balance as long as the probability to construct the reverse loop is the same.
For the Heisenberg model, the creation of the loops is deterministic. Whenever a loop enters an operator during the loop construction, there is only one exit leg that creates a valid operator for the bond type (AF or F). Flipping the spins along the loop, exchanges an diagonal operator
with its off-diagonal counterpart
and vice versa. The entrance leg and exit leg never belong to the same site. For ferromagnetic operators, the propagation direction of the loop in imaginary time remains the same (see
Figure 17). For anti-ferromagnetic operators, the propagation direction of the loop in imaginary time changes (see
Figure 17). When the loop crosses the periodic boundary in imaginary time at a site
i, the computational basis state
has to be updated by flipping the spin at site
i. As each leg only belongs to one loop, the configuration can be split into disjoint loops. Therefore, it is possible to construct all loops during the off-diagonal update and flip each loop independently with a probability of
.
Combining the diagonal and off-diagonal update allows to sample arbitrary unfrustrated Heisenberg models. This includes antiferromangetic Heisenberg ladders and bilayer systems with an unfrustrated long-range interaction.
5.3. Observables
In this section, we introduce some observables that can be easily measured within the SSE formalism introduced. Throughout this section, we do not use the fixed-length scheme but keep the general form with operator sequences of fluctuating length to keep the notation short and simple. The sampling of observables can be done analogously in the fixed-length scheme. When implementing the formulas for the observables in the fixed-length scheme, one just has to keep in mind that the sequences used in the formulas are the same ones but without the padded trivial operators. This means, for instance, that a sum over all propagation steps p becomes a sum over all non-trivial propagation steps in the fixed-length scheme.
Up to this point in
Section 5, we have focused on the partition function of a quantum-mechanical Hamiltonian
expressed as a sum over non-negative weights
of a configuration
. However, we are actually not interested in calculating the partition function itself but rather expectation values of observables.
Analogous to the partition function, quantum-mechanical expectation values can be expanded into a high-temperature series
It is important to realise that configurations with a non-vanishing weight
constituting the partition function do not necessarily contribute to the expectation value
with the same weight or a non-vanishing weight at all [
213]. In general, only for operators
A that are diagonal in the computational basis
, the expectation value can be written as
with
. For instance, the magnetisation in
z-direction or any
n-th moment thereof are such diagonal operators when using the
-eigenbasis
as a computational basis. The statistics of the MC estimates for such observables can be improved by realising that [
218]
with
being the sequence obtained from cyclically permuting
for
p times [
218]. This means that the expectation value Eq. (256) can be expressed as
where
is additionally averaged over imaginary time with
.
For off-diagonal operators one needs to find customised formulas for the expectation value. However, some of these expectation values are accessible in a quite general way. For instance, one can show that the mean energy has a rather simple formula [
210,
211,
218]
Moreover, for the operators
(see Eq. (212)) one finds [
210,
211,
218]
where
is the amount of operators
occurring in the operator sequence. Similarly, one can derive a formula for the heat capacity [
210,
211,
218]
However, for small temperatures, the heat capacity is calculated as the small difference of large numbers.
5.3.1. Linear Response and Correlation Functions
An important class of observables are linear response functions like susceptibilities. The linear response of an observable
A to a perturbation
tuned by perturbation parameter
is given by a Kubo integral [
212,
219]
Once again, if
A or
B are off-diagonal operators, one has to consider case by case and find customised formulas that can be used to sample the specific observables with SSE. We will focus on the case where
A and
B are diagonal in the chosen computational basis. Important examples are the magnetic susceptibility or its local version of spin-spin correlation functions averaged over imaginary time, i. e. the correlation function at zero frequency when regarding Eq. (263) as a Laplace transformation from imaginary time to frequency space. To make this type of observables accessible to SSE, the imaginary-time correlation function
is expanded in temperature analogous to the partition function
where in the last step the operators
A and
B were replaced by their respective eigenvalues
and
of the state
at the propagation steps
.
By replacing the sum over
k by a sum over
and inserting the decomposition of the Hamiltonian, this takes the form
where in the last step the average over imaginary time was taken in order to improve the statistics. From Eq. (271), the connection between the discrete propagation steps of SSE and continuous imaginary time
becomes apparent [
211,
212]. An imaginary time separation
corresponds to a binomial distribution of separations
l of SSE propagation steps
which is peaked around
[
211]. If one is interested in the spectral properties of the system, one can use this formula for sampling imaginary-time correlation functions
. However, there are more efficient ways to calculate imaginary-time correlation functions by embedding the SSE configuration into continuous imaginary time. We refer to Refs. [
212,
220] for details as this is out of the scope of this review.
To get to the linear response function
in Eq. (263), the imaginary time integral of in Eq. (271) has to be calculated. This integral can be analytically calculated
by performing
l partial integrations. Inserting Eq. (271) into the formula (263) and performing the imaginary time integral Eq. (275), yields
One can further separate the
term while using the periodicity
in imaginary time and rewrite the sums over
l and
p as a product of two sums. This eventually yields [
210]
From an algorithmic perspective, the two sums in Eq. (277) need to be calculated by traversing the operator sequence. The effort for measuring therefore scales like the algorithm with complexity when the observables , and are not calculated from scratch at every propagation step p but gradually updated and averaged while propagating the state through the sequence.
For the special case of the susceptibility, one needs to set
and
. The last term
in Eq. (277) can be dropped for simulations of finite systems due to
. Explicitly, this gives
for the susceptibility. Once again, we want to stress that this formula is not formulated in the fixed-length scheme and the average over the propagation index
p therefore refers to propagation steps of non-trivial operators. Averaging over all propagation steps in the fixed-length scheme is incorrect.
For the spin-spin correlation functions
, the operators are set to
and
leading to
where it was already used that
for any finite system. Similar to the zero-frequency spin-spin correlation function, one can also calculate an equal-time spin-spin correlation function
Sampling the correlations among all sites
i and
j at all propagation steps
p, leads to a complexity of
which is worse than the computational complexity
of the SSE updates. Even though one could eliminate the average over imaginary time in the case of Eq. (281) to reduce the complexity to
at the expense of statistical accuracy, this is not possible for Eq. (280) as the sum over imaginary time is intrinsic in the definition of the zero-frequency correlation function. However, by realising that the correlations
between two sites
i and
j are not altered at the order
times in imaginary time but only
8 one can measure the correlation functions Eq. (280) and Eq. (281) in
time. This is achieved by traversing imaginary time while memorizing the propagation step last
of the last preceding spin-flip operator for every site separately. When a spin flip occurs at a site
j at propagation step
p, one needs to update the correlation function
for every site
i before the local magnetisation is propagated with
. One further needs to update last
. In order to avoid that the measurement dominates the simulation for large systems, we only measure the correlations between a fixed site
and
. This finally reduces the complexity of the measurement to
.
Correlation functions are observables that contain a lot of information about a system’s state and in particular its order. Moreover, its decay at the critical point with the distance between two spins is connected to the critical exponent
(anomalous dimension). Away from the critical point, the correlation function usually decays exponentially with long-range models being one exception to this. The length scale of this exponential decay is given by the correlation length
, which itself is an interesting quantity at a continuous phase transition as its divergence is the reason for systems becoming scale-free at the critical point. As the correlation function does not decay exponentially for long-range system, which we focus on in this review, we will use the more general term "characteristic length scale" to refer to the length scale at which the correlations switch to their long-distance behaviour. In particular with respect to Q-FSS for systems above the upper critical dimension, the characteristic length scale is a crucial quantity as its scaling was predicted to be different from standard FSS with
instead of
(see
Section 2.3 or Ref. [
34] for details on scaling above the upper critical dimension in quantum systems). This quantity therefore played a crucial role in confirming the Q-FSS hypothesis. However, the characteristic length scale is a subtle quantity which is hard to extract or even define on a finite lattice. For a finite system, the definition of a characteristic length scale is not unique and there are several definitions for
which will converge to
for
[
221]. For long-range systems finding a suitable definition for the characteristic length scale is even more difficult, as the correlation function decays algebraically even away from the critical point [
222].
Common definitions that are tailored for correlation lengths, which specify the exponential decay of a correlation function at long distances, such as the second moment
therefore might yield
in an infinite system not only at the critical point but also away from the critical point [
223].
One possible definition for long-range quantum systems is [
224]
with
being the Fourier transform of
from real space to momentum space and
the smallest wavevector fitting on the finite lattice. By inserting the Gaussian propagator
for long-range interactions [
20,
21] into Eq. (284), one gets
so that the momentum dependency cancels. Another definition for the same quantity but using the equal-time Gaussian propagator
[
212] is
By inserting the scaling of the gap
for the
n-component quantum rotor model in the long-range mean-field regime [
20,
21], which is relevant for most applications discussed below, one obtains the scaling of
in the thermodynamic limit to be
which is the singularity one expects from the characteristic length scale.
5.4. Sampling at Effectively Zero Temperature
The SSE QMC approach is used to stochastically calculate the thermal average of operators for systems of a finite size at a finite temperature. We discussed in
Section 2 that QPTs are a concept defined in the thermodynamic limit at zero temperature. Further, we presented in
Section 2.2 and
Section 2.3 how the critical point as well as critical exponents can be extracted from certain observables calculated on finite systems at zero temperature. The goal of many works that use SSE is to calculate the expectation value of these observables on finite systems at an effective zero temperature [
32,
34]. We define the notion of effective zero temperature as a temperature at which in practice only the ground state contributes to the thermal average as all excited states are exponentially suppressed [
32,
34]. This definition can be applied only to gapped systems as gapless systems at infinitely small, but finite temperature lead to a thermal average over the infinitesimally low-lying excitations. As the SSE is sampling finite systems, there cannot be a non-analytic point in the ground-state energy associated with a second-order QPT [
113]. Closely related with this statement, the ground state of a finite system can not spontaneously break the symmetry of the Hamiltonian as it is the case at a second-order QPT [
113]. An intuition to this statement can be build from a perturbative point of view. On a finite system, there is always a finite-order perturbative process coupling the hypothetical ground states of a symmetry broken phase. Due to this finite-order coupling there will always be a level repulsion between the states and no true degeneracy occurs. That means on a finite system, there is always a finite-size energy gap between the ground state and the excitations. As discussed in
Section 2.2 on a finite system, there is a pseudo-critical point close to the parameter values of the QPT in the thermodynamic limit. For a system with gapped phases on both sides of the QPT, we expect the relevant finite-size gap to be the smallest in the vicinity of the pseudo-critical point. The finite-size gap at the pseudo-critical point is expected scales as
with the linear system size
L to the power of the dynamical correlation length exponent
z[
34,
113,
128,
212] and
for QPT below the upper critical dimension (see
Section 2.3). This statement can be directly derived from the finite-size scaling form of the energy gap
Note, the scaling dependence for systems above the upper critical dimension has not been confirmed yet by numerical studies. To perform simulations at effective zero temperature, there are two main approaches in the contemporary literature. Firstly, one just scales the simulation temperature with
L for different linear system sizes. This is a valid approach for many systems as
is a common value for the critical exponent indicating a space-time symmetry in the correlations. For long-range interacting systems such as the LRTFIM or unfrustrated long-range Heisenberg models, one has
which makes the scaling sufficient but overly ambitious [
21]. An improved version of this naive approach is to scale the simulation temperature with
for different linear system sizes. This follows the expected scaling of the finite-size gap. The correct prefactor of the scaling is not known and accounted for in the scaling and corrections to scaling for small system sizes can lead to errors. A more sophisticated approach to determine a suitable effective zero temperature for a finite system is discussed in Refs. [
32,
34,
225]. This technique was introduced in Ref. [
225] as the beta-doubling method. The idea is to study the convergence of observables
for a fixed system size and parameter value
r at successively doubled inverse temperatures
of the simulation. The fraction of
is considered with
being the largest considered
value. This quantity is used to probe the convergence of the observable
towards zero temperature. The temperature convergence can be gauged using a plot as demonstrated in
Figure 18. Note, the value of
is always at one. Therefore, as a rule of thumb, one can validate simulation temperatures for which at least the two points prior to the last point also have a value close to one.
Although the beta-doubling scheme has more overhead than a naive scaling with , it provides multiple advantages. First of all, one does not need to know the dynamical critical exponents z beforehand. Secondly, in most cases it is easy to implement: If one uses either way a simulated annealing scheme by successively cooling the simulation temperature during the thermalisation of the algorithm, the beta-doubling scheme can be easily incorporated in this procedure. Thirdly, the beta doubling provides a direct demonstration that a simulation is sampling zero-temperature properties. It captures intrinsically the relevant scaling , as well as potentially large prefactors and corrections to this scaling. Fourthly, it does not need to be applied to all parameter values r as it suffices to perform the beta-doubling procedure only for the parameter values where the relevant finite-size gap is expected to be the smallest and use this temperature for the whole parameter range.
In general, the beta-doubling scheme can help to reduce computational cost by performing simulations at a tailored effective zero temperature. If computational time is not a critical factor, in many practical applications one can use the temperature that is sufficient for the largest system and sample all smaller systems at the same temperature.
5.5. Overview: Path Integral Quantum Monte Carlo Algorithms for Long-Range Models
The QMC methods discussed so far are both based on the SSE formulation [
39,
213]. In parallel to the development of the SSE based QMC methods for long-range systems, path integral (PI) QMC methods operating in continuous imaginary time have been introduced for extended Bose-Hubbard models with long-range density-density interactions [
215,
226,
227] (See.
289). The main application of these methods in the context of long-range interactions is the determination and classification of ground states [
64,
228]. Since we focus on reviewing QPTs in long-range interacting systems which are usually simulated using SSE, we only skim over the basic concepts relevant for the PI QMC methods. One motivation for the development of these methods was the study of the Mott insulator to superfluid transition in the presence of disorder [
229,
230,
231]. Further, long-range density-density interactions were added to the algorithms [
226]. A great application case for these algorithms is the study of extended Bose-Hubbard models [
64,
228] which are an effective description of ultracold atomic or molecular gases trapped in optical lattice potentials [
56,
61,
62,
64,
68,
232,
233,
234,
235,
236,
237]. For a comprehensive explanation on how to derive effective Bose-Hubbard models for these systems see Refs. [
56,
235,
236,
237].
A basic Hamiltonian studied with the world-line QMC methods in this context [
226,
227] reads
with (hardcore) bosonic degrees of freedom on the lattice site
i described by creation
, annihilation
and particle number
operators. The coupling
describes a chemical potential,
the density-density interaction between particles at sites
i and
j and the hopping amplitude
.
PI QMC methods formulated in imaginary time are based on the PI formulation of the partition function
using the Suzuki-Trotter decomposition [
114]. The PI formulation is used to extend the configuration space to all imaginary-time trajectories
of computational basis states [
39,
226] which are periodically closed in
and which are connected by an imaginary-time evolution
. These methods can be implemented in discrete [
238] or continuous [
226,
239] imaginary time.
In the SSE approach, the configuration space is extended using the operator sequences arising from a high-temperature expansion of the exponential in the partition function. These configurations can be regarded as a series of discretely propagated states
with an artificial propagation index
p labeling its position in the sequence. This introduction of an additional dimension to obtain a configuration space which can be sampled using MC in the PI and SSE QMC approach hints that both approaches are closely related. Studies of the close relationships between the SSE and PI representations can be found in Refs. [
211,
240]. Concepts developed in one of the two pictures can be equivalently formulated in the other one. For example, the concept of loops [
215,
216,
241] and directed loops [
215,
217,
241] can be applied in both formulations. The sampling of the Heisenberg model as described in
Section 5.2 can be transferred to the PI QMC picture, as it is a direct application of the directed loop idea [
217,
225].
Since we will focus in this review on quantum critical properties of magnetic systems, we only briefly summarise the major applications of PI QMC methods for extended (hardcore) Bose-Hubbard models with dipolar density-density interactions [
64,
228]. The PI QMC is used in this field to determine ground-state phase diagrams studying the emergence of solid, supersolid, and superfluid phases [
64,
228]. It has beed demonstrated in Ref. [
64] that long-range interactions stabilise more solid phases than the checkerboard solid with a filling of
present in systems with nearest-neighbour interactions. It has also been shown that long-range interactions lead to the emergence of supersolid ground states [
64]. The great benefit of the QMC simulations in comparison to other methods (e. g. mean-field calculations [
111]) is the quantitative nature of these methods. Therefore, the numerical study of these models is highly relevant, since experimental progress in cooling and trapping atoms and molecules with dipolar electric or magnetic moments [
67,
69,
72,
73,
242] enables experiments to realise extended Hubbard models with long-range interactions [
68]. We expect the implementation of SSE QMC techniques for this kind of models to be straightforward using the directed loop approach [
217,
225].
6. Long-Range Transverse-Field Ising Models
In this section we review the ground-state quantum phase diagrams of the long-range transverse-field Ising model (LRTFIM) with algebraically decaying long-range interactions. We emphasise how the Monte Carlo based techniques introduced in this work are a reliable way to obtain critical exponents of quantum phase transitions (QPTs) in this model. The Hamiltonian of the LRTFIM is given by
with Pauli matrices
and
describing spins
located on the lattice sites
. The Ising coupling is tuned by the parameter
J. For
the Ising interaction is antiferromagnetic and for
ferromagnetic. The amplitude of the transverse field is denoted by
h. The positive parameter
governs the algebraic decay of the Ising interaction. Here,
d denotes the spatial dimension of the system and
is a tunable real-valued parameter. The limiting cases of the algebraically decaying long-range interaction are the nearest-neighbour interaction for
and an all-to-all coupling for
. As a side note, there is literature where
is used as a parameter to tune the decay of the long-range interaction. For this review we stay with
in order to treat systems with different spatial dimensions
d on the same footing.
The Hamiltonian (
291) can be treated using pCUT+MC as described in
Section 4. A generic, perturbative starting point for the LRTFIM is the high-field limit. The transverse field term is regarded as the unperturbed Hamiltonian and the Ising interaction as the perturbation. Using the Matsubara-Matsuda transformation [
25,
29,
89]
the Hamiltonian (
291) can be brought into a hardcore bosonic quasi-particle picture
Here,
is a constant,
N is the number of sites, and
is a hardcore bosonic quasiparticle annihilation (creation) operator. The perturbation parameters are given by
In the pCUT language the perturbation associated to the perturbation parameter can be written as , where the hopping processes are contained in and the pair creation (annihilation) processes are in ().
Further, the Hamiltonian (
291) is of the form to straightforwardly apply the SSE QMC algorithm reviewed in
Section 5.1 (see Eq. (
225)). We recall that this QMC algorithm is sign-problem free for arbitrary, potentially frustrated, Ising interactions.
The studied effects of algebraically decaying long-range Ising interactions on the ground-state phase diagram can be categorised into three scenarios:
First, for ferromagnetic interactions, there is a QPT with
symmetry breaking between a ferromagnetic low-field phase and a symmetric
x-polarised high-field phase. The intriguing effect of long-range interaction is the emergence of three regimes with distinct types of universality of the QPT. For large values of the parameter
, the universality class is the same as in the nearest-neighbour model. For intermediate
values of
, the critical exponents change as a function of
and the criticality can be described by a non-trivial long-range theory. For
, the long-range interaction lowers the upper critical dimension below the physical dimension of the model and the transition becomes a long-range mean-field transition. We review the recent studies investigating this model in
Section 6.1 with a particular focus on the
-theory in
Section 6.1.1 and on numerical studies investigating the critical properties in
Section 6.1.2. The aspects regarding FSS above the upper critical dimension are summarised in
Section 6.1.3.
The second relevant category are antiferromagnetic Ising interactions where the nearest-neighbour couplings form a bipartite lattice. In this case, there is a QPT with
symmetry breaking between an antiferromagnetic low-field phase and a symmetric
x-polarised high-field phase. The current state of literature indicates no change in the universality class of the phase transition in dependence of the long-range interaction. We summarise and review the recent progress in studying this case in
Section 6.2.
The last aspect are scenarios where the nearest-neighbour couplings do not form a bipartite lattice. In this case, the long-range interactions may substantially alter the quantum phase diagram of the nearest-neighbour model. We summarise and review the recent progress regarding this case in
Section 6.3.
6.1. Ferromagnetic Long-Range Transverse-Field Ising Models
The ferromagnetic TFIM with nearest-neighbour interactions in dimensions is a paradigmatic model to display a QPT in a D-Ising universality class. This transition describes the non-analytic change in the ground-state between a symmetry-broken low-field phase and a symmetric x-polarised high-field phase. By adding algebraically decaying long-range interactions to this model, one can study how such interactions can alter non-universal as well as universal properties of a QPT and which new features can emerge.
In the case of ferromagnetic interactions, the algebraically decaying long-range interaction stabilises the symmetry-broken phase and the QPT shifts to larger
h values [
25,
29,
32,
34]. An intuitive way to understand this behaviour is that, due to the long-range interaction, more connections are introduced that align the spins in
z-direction and the energy cost for a single spin flip increases. In the limit
, the critical value of the transverse field diverges as the cost for a single spin flip becomes extensive.
In this review, we focus exclusively on the regime of so-called weak long-range interactions with an algebraic decay exponent
[
21]. The research field of strong long-range interactions considers
[
5,
42,
43,
44,
45]. For weak long-range interactions, the Hamiltonian (
291) is well defined in the thermodynamic limit and common properties such as the ground-state additivity and thermodynamic quantities are well defined [
5]. For strong long-range interactions, the energy of the ferromagnetically aligned state is superextensive and, in order to study the model in the thermodynamic limit, an appropriate rescaling of the Ising coupling with the system size is required [
5].
The significance of the ferromagnetic LRTFIM comes from its paradigmatic nature to display changes in the universal behaviour due to the long-range interaction [
20,
21,
25,
29,
32]. It is known since the advent of the theory of classical phase transitions, that algebraically decaying long-range interactions are a potential knob to alter the fixed-point structure of the renormalisation flow and therefore change the universal properties (e.g., critical exponents) [
1,
2,
140]. Coming from the limit
of nearest-neighbour interactions, the QPT of a
d-dimensional model remains in the
D-Ising universality class until
(
is the anomalous dimension critical exponent of the nearest-neighbour universality class). This means, the critical exponents are constant as a function of
and the fixed point associated with the QPT remains the one of the nearest-neighbour model. For
, the RG fixed point associated with the QPT changes to the one of a non-trivial long-range interacting theory and the critical exponents become
-dependent [
21]. The upper critical dimension of this theory with long-range interaction is lowered with respect to the short-range model as
. If the (fixed) spatial system dimension
d is larger or equal to the
-dependent upper critical dimension, the universality class describing the QPT enters a long-range mean-field regime. The three different regimes as well as their boundaries in dependence of the dimensionality of the system are visualised in
Figure 19.
In the following, we recapitulate the basic field-theoretical arguments leading to the distinction between three universality regimes (see
Section 6.1.1). We will further emphasise the most relevant aspects for the two long-range regimes. In the non-trivial, intermediate regime the precise values of the critical exponents depend on
and need to be determined numerically. Therefore, we review the recent numerical studies computing the critical exponents in
Section 6.1.2. To conclude the discussion, we outline how to utilise ferromagnetic long-range Ising models to study QPT above the upper critical dimension (see
Section 6.1.3).
6.1.1. -Theory for Quantum Rotor Models with Long-Range Interactions
Starting with the short-range interacting action (see Eq. (A3)) of the
n-component quantum rotor model, Dutta et al. [
20] introduced an action for long-range interacting rotor models, by adding the algebraic decay between interacting fields as an additional term to the action in real space. Note, the action for the Ising model is equivalent to the one-component (scalar) rotor model [
113]. By performing a Fourier transformation of the resulting action, the authors obtained the action
with
,
from the decay exponent
of the microscopic model and
coupling constants as in the nearest-neighbour case. The
term, which is new compared to the nearest-neighbour theory, results from the Fourier transformation of the added long-range interacting term
of the order-parameter field
[
40]. As the relevant modes for the QPT are the ones with long wavelengths, it is clear that for
the
term gives the leading contribution in
q, the
term can be neglected and therefore the system is in this case described by the short-range interacting theory. In contrast to this, a different critical behaviour distinct to the short-range case is possible for
. With the same argument as above, the
term becomes negligible for
. In the following we will focus on this regime where the behaviour differs from the nearest-neighbour case. In order to gain insight into the QPT affected by long-range interactions, an investigation of the Gaussian theory is a good starting point to enter the framework of perturbative renormalisation group calculations [
20,
21,
113] like the
-expansion in Ref. [
20]. From equation (
295), one can directly read off the propagator of the Gaussian theory as
From field-theoretical power counting arguments, Dutta et al. [
20] derived several properties of the Gaussian theory. Using the divergence of the mass renormalisation term, the lower critical dimension below which no phase transition is occurring can be derived as
[
20]. By regarding the Gaussian propagator, one finds directly that
as
is defined by the deviation from 2 of the leading power by which the momentum enters the propagator (see Eq.
A37). Note that in the long-range case, the Gaussian theory has an
. From further power counting analysis, it is possible to derive more critical exponents from the Gaussian theory [
20],
In the limit
, the short-range mean-field exponents are recovered [
20]. With the argument that the hyperscaling relation still holds directly at the upper critical dimension, Dutta et al. [
20] derived the upper critical dimension by inserting the long-range Gaussian exponents
9 (see Eq. (
298))
In addition to the investigation of the Gaussian part of the action in Eq. (
295), one can derive non-trivial exponents below the upper critical dimension by deriving perturbative corrections to the Gaussian exponents in an
-expansion around the upper critical dimension [
113]. Dutta et al. [
20] established with a one-loop renormalisation group expansion of flow equations that the Gaussian fixed point is stable for
. Therefore, the long-range Gaussian exponents (see Eq. (
298)) are valid for
. For
, the correction indicates a flow to a non-trivial fixed point. The first-order corrections of the
-expansion for
and
are provided [
20] and the first-order corrections to
and
z are argued to be zero [
20]. It should be noted, that the dynamic correlation exponent
for
and therefore correlations are not isotropic in space and imaginary-time direction [
20,
21]. "For any value of
,
sticks to its mean-field value
, because the renormalization does not generate new
terms" [
20]. This also fits well with the claim that
, because then there is no discontinuity at the boundary [
1,
2,
3,
21,
243].
A recent study by Defenu et al. [
21] with a functional renormalisation group approach investigated the flow of couplings in an effective action with anomalous dimension effects. This analysis showed that the boundary between the short-range and the non-trivial long-range regime does not occur at
but at
. So it is shifted towards lower
by the anomalous dimension of the short-range criticality. This anomalous dimension consideration in [
21] is mainly driven by the following argument: From the functional renormalisation group ansatz, the authors derived two flow equations for the coupling of the long-range coupling
(connected to
a in Eq. (
295)) and the short-range coupling
(connected to
b in Eq. (
295))
The first flow equation (
300) has fixed points for
or
. The first fixed point results in long-range exponents whereas the second fixed point means that the long-range coupling becomes irrelevant.
Additional to this anomalous dimension effect, the authors in [
21] also calculate critical exponents in the non-trivial intermediate phase with their functional renormalisation group technique. They also compared their results with Monte Carlo simulations of the dissipative non-Ohmic spin chains, which can be mapped to the same behaviour as the long-range interacting quantum Ising model [
244], showing reasonable agreement.
6.1.2. Critical Exponents and Critical Points for One- and Two-Dimensional Systems
The criticality of the ferromagnetic LRTFIM has been studied by a wide variety of methods with different strengths and weaknesses. Overall, all methods qualitatively agree with each other and confirm the three expected regimes from the field-theoretical analysis [
20,
21], namely long-range mean field, short-range regime and an intermediate non-trivial long-range regime which connects the two limiting regimes with monotonously changing exponents. In the limit
, the value of the critical field
hc diverges as the energy cost for a single spin flip in the low-field phase becomes extensive and the quantum fluctuations introduced by the transverse field fail to compete with the Ising interaction. In the limit
, the value of
hc converges towards the exact value for the nearest-neighbour transverse-field Ising model [
25,
29,
32,
245].
In the case of the LRTFIM on the one-dimensional linear chain, the model has been studied by perturbative continuous unitary transformation with Monte Carlo embedding (pCUT + MC) [
29,
31,
34,
246], functional renormalisation group (FRG) [
21], density matrix renormalisation group (DMRG) [
28] as well as Stochastic Series expansion (SSE) [
32,
34,
246,
247], pathintegral (PI) [
33] QMC, and QMC with stochastic parameter optimisation (QMC+SPO) [
248].
The critical value of the transverse field was calculated by pCUT [
29,
31,
34] by estimating the gap closing using DlogPadé extrapolation of the perturbative gap series coming from the high-field phase, by DMRG using finite-size scaling of the fidelity susceptibility [
28] and by SSE using finite-size scaling of the magnetisation [
32,
247]. The finite-temperature transition points were studied by PI QMC [
33] for
in the long-range mean field regime and
in the strong long-range regime.
Depending on the method, different critical exponents were extracted. Only in the case of pCUT + MC and SSE QMC [
34,
246], all three regimes were investigated. By extracting three independent critical exponents the authors were able to provide the full set of critical exponents for each method individually. In the case of pCUT + MC [
34,
246], the exponents were computed by (biased) DlogPadé extrapolants of high-order series of the gap (
), the one-quasiparticle static spectral weight (
) and the control-parameter susceptibility (
). In SSE QMC [
34,
246] the exponents were calculated using data collapses of the magnetisation (
and
) and the order-parameter susceptibility (
). The DMRG study [
28] applied the method of data collapse to the fidelity susceptibility to extract the single exponent
in the non-trivial long-range regime. Using FRG [
21], the exponents
and
z were calculated, from which one can additionally calculate the exponent
. In the PI QMC study [
33], the finite-temperature criticality of the LRTFIM was investigated for
in the long-range mean field regime and
in the strong long-range regime. The study includes the extraction of an exponent
, that can be related to
, from the Binder cumulant and classical order-parameter susceptibility via finite-size scaling as well as the exponent
from finite-size scaling of the classical order-parameter susceptibility. In
Figure 20 the critical field
hc and the canonical critical exponents are plotted for the discussed studies.
The two-dimensional case of the square lattice is less studied. The critical values of the transverse field were only calculated by pCUT + MC [
29] and SSE QMC [
32,
247]. There is so far no study that extracted the full set of critical exponents altogether. With pCUT [
29], the gap exponents
was also extracted by DlogPadé extrapolants of the gap series. In FRG [
21], the critical exponents
z and
were calculated and in SSE [
32] the critical exponents
and
were extracted from data collapses of the magnetisation. These results are summarised in
Figure 21. Note, we present the results from pCUT+MC [
29] with one additional order to the maximal order in the perturbation parameter.
Overall, the different methods all perform well even for relatively small
. For pCUT and SSE, where all three regimes were simulated, the limiting cases of long-range mean-field criticality and nearest-neighbour criticality were correctly reproduced which underlines the reliability of the presented results in the intermediate regime. The main challenge that the methods are facing is to sharply resolve the regime boundaries due to rounding effects which are probably due to the limited length scales in the simulation. Additionally, the boundary to the long-range mean field regime is spoiled by logarithmic corrections to scaling at the upper critical dimension. Even in the case of pCUT + MC, which is a method operating in the thermodynamic limit, there is an intrinsic limit of length scale due to the finite order of the series. In particular, the rounding makes it hard to verify (or falsify) the claim that the boundary between the short-range and non-trivial long-range regime is shifted
. For 1d, the boundary is shifted to
, which could be resolvable in extensive simulation. However, in 2d the boundary is only marginally shifted to
which is probably not resolvable by the reviewed methods. The best boundary resolution is probably given by SSE QMC from Ref. [
34]. If one is interested to study the shift of the boundary, one could therefore push the SSE QMC simulations of the one-dimensional chain to larger systems for
. Another difficulty for the methods operating on finite systems is the breakdown of the common hyperscaling relation and standard FSS above the upper critical dimension which affects the processing of data in the long-range mean-field regime for
. The scaling above the upper critical dimension was already outlined in
Section 2.3. However, we will recapitulate the most important points in the following
Section 6.1.3 not only because this modified scaling was used in several of the above mentioned methods to extract critical exponents [
32,
33,
34] but also because the LRTFIM constitutes a testing ground for quantum Q-FSS scaling and the modified scaling of the correlation length was confirmed in a study of the LRTFIM [
34].
6.1.3. Scaling above the Upper Critical Dimension
As discussed in
Section 6.1.1, for
the action describing the QPT and its universality changes from short range to long range. In this long-range regime, the upper critical dimension
was found to depend on the value of
[
20,
21]. This means, the connectivity of the system increases for decreasing
and a smaller spatial dimension
is already sufficient to reach or even exceed the upper critical dimension. The sigma below which the model is above its upper critical dimension is
. In this regime, it is not possible to correctly extract all critical exponents from standard FSS and the common hyperscaling relation
becomes invalid as the critical exponents become independent of the dimension
d. The origin of this discrepancy comes from the effect of dangerous irrelevant variables that one has to take into account in the derivation of scaling above the upper critical dimension. When doing so, one gets a modified hyperscaling relation
with the exponent
. This relation can be used for conversion of different critical exponents [
34,
35].
Additionally, the standard FSS scaling form of an observable
with power-law singularity
is modified to [
34]
following the formulation of Q-FSS for classical [
15,
16,
17,
18,
19,
41] and quantum [
34] systems. Finite-size methods that rely on the method of data collapse or other FSS-based techniques have to use this adapted formula to extract all the critical exponents successfully [
34]. For a more elaborate description of quantum Q-FSS, please refer to
Section 2.3 or to Ref. [
34].
On the other side, the LRTFIM can also be used as a testing ground for quantum Q-FSS. One key difference that distinguishes Q-FSS from standard FSS is that the correlation sector is also affected by DIV in the case of Q-FSS. This leads to a modified scaling of the characteristic length scale
with the system size instead of
as it is for standard FSS. The characteristic length scale was measured by SSE QMC [
34] in the long-range mean-field regime and the exponent Ϙ was extracted by a data collapse with fixed mean field value
. The extracted values coming from Ref. [
34] are presented in
Figure 22 and are clearly in line with the predictions made by Q-FSS.
6.2. Antiferromagnetic Long-Range Transverse-Field Ising Models on Bipartite Lattices
In this section we review results for the antiferromagnetic LRTFIM on lattices where the coupling structure of the nearest-neighbour couplings forms a bipartite lattice. A lattice – or more generally a graph – is called bipartite if the sites can be split into two disjoint sets A and B so that there are no edges between sites within each of the two sets A and B. In the case of a lattice A and B are called bipartite sublattices. Note, that only in the nearest-neighbour limit on bipartite lattices, there is an exact duality between the ferromagnetic and antiferromagnetic TFIM, and the quantum-critical properties coincide. This duality comes from a sublattice rotation of along the x-axis for the spins in one of the two sublattices. In the nearest-neighbour limit there is a x-polarised symmetric high-field phase and a symmetry-broken low-field phase. The low-field phase is adiabatically connected to the zero-field limit. In the zero-field limit, there are two ground states associated with the states where the spins on the sites in one set of the bipartite sublattices are pointing in one while the others point in the other direction. Note, these states can be directly mapped onto the two ferromagnetic ground states by the above mentioned sublattice rotation. As for the ferromagnetic nearest-neighbour TFIM, the QPT between these two phases is of D-Ising universality.
The antiferromagnetic long-range interaction beyond the nearest-neighbours induces a hierarchy of competing interactions. Due to the long-range interactions, there is no longer a bipartite coupling structure as the spins cannot be aligned such that the energy of all interactions is minimised. Nevertheless, there is evidence that the two ground states in the low-field phase of the nearest-neighbour limit are adiabatically connected to the ground states in the low-field phase for all
[
23,
24,
25,
26,
29,
31,
32,
111].
For the all-to-all connected case
, every state with zero
z-magnetisation is a ground state of the system [
30,
40] at
. The Hamiltonian (
291) can then be written in terms of total spin
with
as
with the total number of spins
[
30,
40]. For all finite transverse fields
, the ground state of Eq. (
305) is directly located in the
x-polarised phase [
40,
252].
The natural questions arising for the QPT in the antiferromagnetic LRTFIM on the bipartite lattices are similar to the ferromagnetic LRTFIM: How does the algebraically decaying long-range interaction influence the critical field value of the QPT? Is there a change in the universality class of the QPT depending on the precise values of the decay exponent ?
Regarding the critical field values in one-dimensional chains, there are MC based studies using the SSE QMC (see
Section 5) [
32] and pCUT+MC (see
Section 4) [
25,
31], as well as DMRG [
24,
26] and matrix product states [
23]. We summarise the results regarding critical field values and selected critical exponents of the studies mentioned above in
Figure 23. In two-dimensional systems there are MC based studies for the square lattice using SSE QMC (see
Section 5) [
32] and pCUT+MC (see
Section 4) [
29] (for selected results see
Figure 24). Variational methods in two dimensions such as projected entangled pair states [
253,
254] or DMRG calculations [
255] have not been applied for antiferromagnetic systems on bipartite lattices. Entanglement-based methods become more challenging from the increased entanglement due to the area law in two dimensions [
256]. Further there are reports about the violation of the area law for one-dimensional systems with long-range interactions [
23,
24]. On general grounds, one cannot efficiently represent algebraically decaying long-range interactions in these techniques [
257,
258]. However, there exist DMRG calculations for antiferromagnetic LRTFIMs on non-bipartite lattices [
27,
54,
55,
109].
In general, it has been numerically demonstrated that the critical field values decrease monotonically as a function of
from the nearest-neighbour value at
to zero at
[
23,
24,
25,
26,
29,
31,
32]. Regarding the QPT on the antiferromagnetic chain, the established picture is that it is of
D Ising type for all
[
23,
24,
25,
26,
31,
32]. In the regime of ultra-long-range couplings, recent finite-size DMRG findings by G. Sun [
26] suggest that the
D Ising universality even holds for any
. On the other hand, earlier results using matrix product states generalizing the time-dependent variational principle [
23] indicate that critical exponents may vary for
. However, the results from Ref. [
26] convincingly challenge these findings. The studies using MC based techniques [
25,
31,
32] cannot extract reliable critical exponents in this regime. In SSE QMC [
32], the increasing competition of interactions of the long-range interactions is the major problem making the numerical simulations in this regime impractical. If the long-range interaction becomes more prominent, this leads to similar algorithmic challenges as in frustrated systems [
32,
40,
259,
260]. The autocorrelation times of the SSE Monte Carlo dynamics increase as more and more bond operators are present in the operator sequence and the field operators diffuse only slowly [
32]. Nevertheless, following the trend of the results in Ref. [
32] the scenario of a single universality class across the entire
range is plausible. Regarding the pCUT+MC studies [
25,
31], the main obstacle to extract information about the quantum criticality in the regime
is the fact that the critical field value is decreasing. Refs. [
25,
31] use a perturbative expansion around the high-field limit, therefore, if the critical field value increases, then the extrapolation of the gap series becomes increasingly challenging.
To summarise, there are studies indicating a change from the
D Ising criticality on the chain for small
values [
23,
24]. These studies also report a possible breakdown on the area law of the entanglement entropy for the ground state in this regime [
23,
24] which is a vital prerequisite for the methods they implemented. A more recent study [
26] demonstrates no change in the universality class which is backed by [
25,
30,
31]. For two-dimensional systems, the universality class is demonstrated to be
D Ising for
[
29,
32] and no trend is reported that the universality class should change for smaller values of
.
In general, a great algorithmic challenge that remains for the MC-based methods with regard to the antiferromagnetic LRTFIM on bipartite lattices, is a reliable study of the regime of small values . Here, as discussed above, all of the commonly applied methods have a handicap in some way.
6.3. Antiferromagnetic Long-Range Transverse-Field Ising Models on Non-Bipartite Lattices
We start with an antiferromagnetic nearest-neighbour interacting TFIM on a non-bipartite lattice (e.g., the sawtooth chain, triangular lattice, Kagome lattice, pyrochlore lattice) at zero field. In the case of non-bipartite lattices, one can always find an odd number of Ising bonds (edges) that form a closed loop. The presence of a loop of odd length means it is impossible to satisfy all antiferromagnetic couplings simultaneously. The phenomenon of not being able to minimise all interactions due to geometrical lattice constraints is called geometrical frustration [
262]. A notion for the strength of geometric frustration can be defined using the resulting ground-state degeneracy [
263]. A theoretical tool to access the ground-state degeneracy is through Maxwell counting arguments [
263,
264,
265]. For systems with a large degree of frustration, e. g. the antiferromagnetic nearest-neighbour TFIM on the triangular or Kagome lattice, there is a finite residual entropy per site in the thermodynamic limit [
168,
266,
267,
268,
269,
270].
In general, extensively degenerate ground-state spaces due to frustration pose a formidable resource for emergent exotic quantum phenomena. Quantum fluctuations introduced, e. g. by a transverse field result in a breakdown of the extensive ground-state degeneracy and the potential emergence of non-trivial ground states [
269,
270]. In general, it is useful to think about the low-field physics as a degenerate perturbation theory problem on the zero-field ground-state space [
30,
32,
269,
270]. We will later review, that this is a reasonable framework in order to treat the breakdown of the degenerate subspace due to fluctuations and algebraically decaying long-range interactions on the same footing [
30,
32,
112].
Perturbing extensively degenerate ground-state spaces with fluctuations may result in several distinct scenarios. First, a distinct symmetry-broken order can emerge for infinitesimal perturbations (order-by-disorder) [
168,
269,
270,
271,
272]. Further, a direct realisation of a symmetry-unbroken phase may occur (disorder-by-disorder). This phase can either be trivial [
168,
273] or exotic, e.g., quantum spin liquids [
99,
100,
101,
274,
275].
For the first part of this section we review the antiferromagnetic LRTFIM on the triangular lattice [
27,
29,
30,
32,
40,
110,
276]. In the nearest-neighbour limit of that model, the zero-field case has an extensively degenerate ground state [
266,
268,
269,
270,
272]. When adding an infinitesimal transverse field, this ground-state degeneracy breaks down and a
-symmetry-broken clock order emerges from an order-by-disorder mechanism [
269,
270,
272]. The effective Hamiltonian describing the breakdown of the degeneracy in leading order can be expressed as a quantum dimer model [
269,
270,
272]. By investigating the degenerate ground-state space, it was observed that there are local spin configurations in which spins can be flipped without leaving the ground-state space. These configurations, called flippable plaquettes, consist of a spin surrounded by six spins with alternating orientation. It can be easily seen, that flipping the spin in the middle of the flippable plaquette results in turning the plaquette by
(see
Figure 25).
The effective low-field quantum dimer model [
269,
270,
272], consisting of the plaquette rotation term, is therefore the hexagonal-lattice version of the Rokhsar-Kivelson quantum dimer model with
and
[
269,
270,
277]
From this mapping, the nature of the state at
was predicted, as well as the breakdown of the order with a 3D-XY QPT [
269,
270]. Further numerical studies extended the insights into this system [
168,
260,
272]. The critical field value at which the phase transition between the symmetric
x-polarised paramagnetic high-field phase and the symmetry broken gapped clock-ordered low-field phase occurs is at
[
168,
272].
To infer the nature of the full phase diagram including the transverse field and the long-range interaction, the next step is to discuss the zero-field limit of the Ising model with long-range interactions. As the antiferromagnetic long-range interactions introduce a hierarchy of constraints due to the interaction between sites further apart, the long-range interaction breaks the ground-state degeneracy of the nearest-neighbour zero-field as well [
30,
110,
276,
278]. It is important to emphasise that the long-range interaction lifts the degeneracy in a different way and stabilises a gapped stripe phase breaking the translational symmetry in a different way than the clock order promoted by the transverse field [
30,
110,
276,
278,
279]. This plain stripe state is sixfold degenerate by rotations around
as well as spin-flips [
30,
110,
276,
278]. The spins are aligned in straight lines with alternating
z-orientation [
30,
110,
276,
278]. An important aspect is that these stripe states are gapped and therefore stable against finite transverse fields [
30,
110,
276].
Regarding the limit of large transverse fields, pCUT+MC was used to investigate the QPT between the
x-polarised high-field phase and the clock-ordered phase [
29,
30]. Similar to the antiferromagnetic models on the bipartite lattices, the critical field value is reduced by the long-range interaction towards smaller field strengths [
29,
30] (see
Figure 26). On the triangular lattice, the universality class of the transition from the high-field to a clock-ordered phase is of (2+1)D-XY type [
269,
270,
272]. It is reported to remain in this category for all
values investigated (
) [
29]. For small
values, the extrapolation of high-field series expansion is inconclusive [
29,
30]. Infinite DMRG (iDMRG) investigations on triangular lattice cylinders suggest that the clock order vanishes for small values of
and a direct transition to a low-field stripe phase occurs [
27]. A first-order phase transition between the
x-polarised high-field phase and a low-field stripe phase would be consistent with the incapability of the high-field gap to track the transition [
27,
29,
30,
276]. Note that the stripe phase determined from the iDMRG study [
27] does not agree with subsequent studies focusing on the low-field ground states of the model [
30,
32,
276,
278,
279].
Complementary to the high-field analysis and numerical iDMRG studies, approaches in the low-field limit have also been conducted [
30,
110,
276] (see
Figure 26). As already mentioned, it was discussed in the literature for some time [
30,
278,
279] and recently demonstrated using a unit-cell based optimization technique [
110] that gapped plain stripes are the zero-field ground state as soon as long-range interactions are present. Therefore, in a phase diagram with a
and a transverse-field axis, the clock order phase is wedged between the high-field phase from above and the plain stripe phase from below [
27,
29,
30,
276] (see
Figure 26). The extent of both, the stripe phase and the high-field phase, increases with stronger long-range interactions [
27,
29,
30,
276]. This behaviour was also demonstrated perturbatively from the nearest-neighbour zero-field limit by treating the long-range interaction as well as the transverse field as perturbations on the degenerate subspace [
30,
276].
We sketch a qualitative picture of the quantum phase diagram in
Figure 27. The precise ground state at small values of
is still an open research question. We believe that there is some evidence that after a certain value of
, only a transition between the high-field phase and the low-field stripes remains which is expected to be of first order. However there is, by now, no numerical technique to make reliable statements in this regime. In this context, the high-field series expansions cannot detect a first-order phase transition from the elementary excitation gap. There is a SSE QMC study of the system by S. Humeniuk [
40] applying the algorithm discussed in
Section 5.1. The direct QMC simulation was used to determine the critical point between the high-field phase and the clock-ordered phase for
[
40]. For smaller transverse fields, the author identified a "region dominated by classical ground states [
40]" which are adiabatically connected to the zero-field states. Therefore, there is a qualitatively sketch of a phase diagram in Ref. [
40] similar to
Figure 27. Nevertheless, the naive application of the SSE QMC as discussed in
Section 5.1 is not as efficient as for models with interactions which are not competing [
40,
259,
260]. However, there are ideas on how to implement efficient quantum cluster algorithms for frustrated Ising models in a transverse field [
259,
260] which need to be adjusted for long-range interacting Ising models.
The antiferromagnetic LRTFIM on the triangular lattice is a well studied example of the interplay between long-range interactions and order-by-disorder. Recently, several more examples arise from the field of Rydberg atom quantum simulators [
54,
55,
109,
110,
112]. In Rydberg atom simulators, atoms are positioned in a desired configuration using optical tweezers [
52] and are laser-driven to a Rydberg state [
52]. The Rydberg blockade mechanism leads to an algebraically decaying interaction between Rydberg excitations which decays with
[
52]. Using the Matsubara-Matsuda transformation [
89], the excitations/non-excitations of a Rydberg atoms are associated with spin degrees of freedom [
52]. The density-density interaction transforms into an Ising interaction [
110,
112]. Rydberg atom quantum simulators are capable of simulating the antiferromagnetic LRTFIM with
and a longitudinal field [
52,
54,
55,
109]. Order-by-disorder scenarios are described for the Kagome lattice at a filling of
Rydberg excitations [
54] and the Ruby lattice at a filling of
Rydberg excitations [
55,
109] and at vanishing longitudinal field [
112]. Regarding these examples, the order-by-disorder was studied for a long-range interaction truncated after the third-nearest neighbours [
54,
55,
109,
112]. Considering the remaining long-range interaction as a perturbation to the degenerate ground-state space for most of these mechanisms, a similar scenario as for the triangular lattice is expected [
54,
55,
109,
110]. Using the same unit-cell based optimisation method as for the triangular lattice, Koziol et al. [
110] have calculated ground states of the zero-field model using the full untruncated long-range interaction. For the Kagome lattice at a filling of
Rydberg excitations [
54] and the Ruby lattice at a filling of
Rydberg excitations [
54,
55,
109] gapped ground states were determined [
110] which do not coincide with the order arising from the order-by-disorder mechanism. Interestingly, in all these examples the zero-field ground state determined by the long-range interaction possess none of the motifs relevant for the leading-order order-by-disorder mechanism [
110]. For example, on the triangular lattice, the plain stripe states contain zero flippable plaquettes which are the relevant motif to lower the energy at a finite transverse field in leading order. Recently, an order-by-disorder mechanism for the antiferromagnetic
-
-
TFIM on the Ruby lattice was introduced by A. Duft et al. [
112]. The remarkable aspect of the mechanism is that adding long-range interactions does stabilise the same order which is also selected by the quantum fluctuations [
112]. A general theory, under which conditions the long-range interactions stabilise the same order as quantum fluctuations or stabilise a completely disjoint order, is yet to be found.
Conclusively, the breakdown of extensively degenerate ground-state spaces due to an interplay of long-range interactions and quantum fluctuations is a highly vibrant and relevant field. Long-range interactions are relevant for a wide range of quantum-optical quantum simulation platforms, including cold atoms [
52,
70,
72] and ions [
74,
75,
76,
77,
79,
80,
81,
82,
84]. These platforms can be used to realise exotic phases of matter, e. g. quantum spin liquids [
54,
55,
109], glassy behaviour [
282] or clock-ordered states [
29,
30,
112,
276]. To understand the emergence of these exotic states of matter, simplified models are oftentimes considered [
54,
55,
109,
112,
282] which truncate the long-range interaction. Therefore, it is imminent to understand the effects of the full long-range interactions on the mechanisms driving the emergence of these exotic phases. At the moment the commonly used tools to investigate the phase diagrams of these frustrated systems consist of:
the derivation of effective models [
30,
54,
55,
109,
112]. These can be used in order to predict exotic emergent phases and the nature of QPTs [
30,
54,
55,
109,
112].
DMRG calculations [
27,
54,
55,
109], which are used to obtain a numerical insight into the ground-state phase diagrams.
QMC calculations, in particular the SSE as discussed in
Section 5.1 can be used for an unbiased sampling of ground-state properties [
40]. In order to omit the slowdown of the algorithm due to the geometric frustration, algorithmic improvements are required [
259,
260]. It is still an open research question how to set up an algorithm that samples long-range interaction and frustration efficiently.
high-order high-field series expansions using a graph decomposition. This is also a very capable tool to track the first QPT coming from the high-field limit. With this method, the critical field value as well as critical exponents can be determined [
29,
30,
112]. It is also possible to infer information about the phase on the other side of the phase transition by studying the momentum at which the elementary excitation gap is located [
29,
30,
112]. A MC embedding of white graphs can be performed to study the entire algebraically decaying long-range interaction [
29,
30] while an ordinary graph decomposition can be used for systems with truncated interactions [
112].
7. Long-Range Transverse-Field XY Chain
In this section we review results on the XY chain with long-range interactions in a transverse field from Ref. [
31]. The Hamiltonian of the long-range transverse-field anisotropic XY model (LRTFAXYM) on a one-dimensional chain is given by
with Pauli matrices
and
describing spin-
degrees of freedom on the
i-th lattice site of the chain. The transverse-field strength is given by
and the strength of the coupling between sites
i and
j is given by
. The coupling is (anti-) ferromagnetic for
(
). We include a continuous interpolation parameter
to tune the system from the XY chain with isotropic interactions (
) to Ising interactions (
). The exponent
determines the decay of the long-range interaction. The limit of short-range interactions is recovered for
and the limit of all-to-all couplings for
.
Similar to the discussion of the transverse-field Ising model, it is useful to express Eq. (
307) in terms of hard-core bosonic operators
,
using the Matsubara-Matsuda transformation [
89]
We obtain
in units of
with
being the unperturbed ground-state energy per site,
N the number of sites, and
the perturbation parameter associated with the perturbation
(second sum) taking the form
in the pCUT language, so that a high-order series expansion about the high-field limit can be performed.
In the limit the ground state is a nondegenerate z-polarised state that serves as an unperturbed reference state. Elementary excitations are called local spin flips at an arbitrary site i that can be annihilated (created) by the hard-core bosonic operator . Upon increasing , these quasiparticles (qps) become spin flips dressed by quantum fluctuations induced by the perturbation . In the limit of the system exhibits ferromagnetic or antiferromagnetic magnetic order depending on the sign of J in the x-direction (y-direction) for (). When tuning the interpolation parameter from pure Ising to isotropic XY interactions, ordering in x-direction starts to compete with ordering in y-direction.
In order to better appreciate the results for
, we briefly discuss the quantum phase diagram in
Figure 28 of the model (
307) with nearest-neighbour interactions [
283,
284,
285,
286,
287,
288,
289,
290,
291,
292]. For any
the system with nearest-neighbour interactions undergoes a (1+1)D Ising QPT at
from the paramagnetic symmetric high-field phase to the (anti)ferromagnetically ordered low-field phase where the ground state spontaneously breaks the
symmetry of the Hamiltonian [
283]. The transition along the
line between the two distinct
symmetry broken phases is of (Ising)
type which is conformally and
invariant [
283]. These critical lines meet at
and
in multicritical points
with critical exponents
and
[
283].
For
, the Hamiltonian (
309) becomes particle conserving as the terms annihilating (creating) two quasiparticles
cancel each other making the isotropic XY Hamiltonian
symmetric. Due to this particle conserving nature of Eq. (
309), there are no quantum fluctuations dressing the symmetric
ground state for
and the perturbative treatment of the high-field dispersion becomes exact in first order of the perturbation theory [
31]. This distinguished symmetry property of the
case motivates that we are reviewing the properties of the isotropic transverse-field XY chain in the ferromangetic and antiferromagnetic case first in
Section 7.1. Subsequently, we discuss the anisotropic XY chain for ferromagnetic interactions in
Section 7.2 and for antiferromagnetic interactions in
Section 7.3. Here, we show improved results of the ones shown in Ref. [
31] since we noticed right after their publication that the Monte Carlo runs were performed with a suboptimal choice of simulation parameters.
7.1. Isotropic Long-Range XY Chain in a Transverse-Field
To consider the isotropic XY chain,
is set in the Hamiltonians Eq. (
307) and Eq. (
309). The quantum criticality of this model was studied in Ref. [
31] in an analytical fashion, by evaluating the ground-state energy and the dispersion of the elementary excitations analytically. Setting
, Eq. (
309) reads
since the pair creation and annihilation terms
cancel as stated above. The perturbation
acting on the unperturbed
z-polarised ground state does not introduce any quantum fluctuations. Therefore the
z-polarised state becomes an exact eigenstate for arbitrary
and stays the ground state until a QPT occurs [
31]. Analogously, the one quasiparticle dispersion is exact in first-order perturbation theory. The one-particle dispersion in the symmetric high-field phase reads [
31]
The critical value
can be determined from the dispersion (
311) as the
value at which the quasiparticle gap
closes. The critical exponent
associated with the gap closing can be extracted from the dominant power-law behaviour of the gap near
Further, it is possible to separate the dynamic
z and static
exponent by evaluating the 1qp dispersion
at the quantum-critical point such that
which allows the extraction of the critical dynamic exponent
z. For ferromagnetic XY interactions, the minimum of the dispersion (
311) is located at the momentum
. Therefore the gap series is given by
and
. The Riemann-zeta function
is convergent for all
and diverges for
. This mathematical observation coincides with the qualitative behaviour found in ferromagnetic Ising systems discussed in
Section 6.1 in the sense that
. Since the expression for the gap
is linear in
, the critical exponent is
, independent of the
value. Using the expression for the dispersion (
311), the knowledge about
and the definition of the dynamic critical exponent (
313), Adelhardt et al. [
31] evaluated
z and therefore also
as a function of
. The results of their calculations are presented in
Figure 29. The exponents resulting from the consideration of the dispersion can be categorised in two distinct regimes: For
, the critical exponents of the nearest-neighbour transition are found (
and
). For
, the exponent
z decreases linearly from 2 to 0 and
increases like
from
to
∞. Note, that this linear decrease of
z and the divergence of
appears similar to the behaviour of the ferromagnetic LRTFIM at small
values (see
Section 6.1)[
20,
21]. To explain this behaviour, a quantum field theory was proposed in Ref. [
31]. The idea was to add a
term (analogous to [
20,
21]) to the well-studied bosonic action used for the study of the isotropic short-range transition [
113,
229,
283,
293]. The suggested action reads
with
being a complex
c-number order-parameter field of the transition,
and the real constants
and
r [
31]. The notation of Eq. (
314) is taken from Ref. [
31,
229]. From powercounting one obtains the critical exponents
from the Gaussian part of the action in Eq. (
314) [
31]. Following the arguments of Ref. [
229], these exponents hold "presumably" [
229] for
below the upper critical dimension
. Refs. [
229,
293] provide arguments that the self-energy vanishes for every order in
u and the renormalisation of
u can be performed in all orders via ladder diagrams [
31]. The vanishing self-energy is explained by the fact that, in each diagram, every pole in
lies in the complex upper-half plane. The frequency integral can be deformed into the lower half-plane to give zero [
293]. The fact that the free propagator of the field theory is not changed by a self energy is a manifestation of the absence of fluctuations that do not preserve the particle number [
31]. The predictions of the field theory (
314) for
z in Eq. (
315) and
in Eq. (316) are in perfect agreement with the results of the high-field excitation gap (see
Figure 29) [
31].
Note in the nearest-neighbour case, the low-field ground state does not break the
symmetry of the Hamiltonian due to the Hohenberg-Mermin-Wagner (HMW) theorem ruling out continuous symmetry breaking [
113,
129,
130,
131,
283,
294,
295]. Long-range interactions are a known mechanism to circumvent the HMW theorem [
22,
35,
296,
297,
298,
299,
300,
301,
302,
303,
304,
305]. With the high-field approach discussed above, it is not possible to determine the nature of the low-field ground state and if a continuous symmetry breaking occurs for sufficiently small decay exponents. We are not aware of studies addressing this topic.
For the remainder of this section, we discuss the antiferrromagnetic isotropic XY chain in a transverse field along the same lines as for the ferromagnetic case. The minimum of the dispersion is at momentum
and the gap can be expressed as
with
being the Dirichlet eta function (also known as the alternating
function) [
31]. The Dirichlet eta function
is convergent for all
, therefore the definition of an excitation gap is also possible in the regime of strong long-range interactions (
). Analogously to the discussion of the Ising interaction in
Section 6.2, the extent of the low-field phase is diminishing for decreasing
values. For
the energy gap closes at diverging
[
31]. With a similar analysis to the ferromagnetic case, the same critical exponents
and
are found along the entire
line [
31]. This is also analogous to the antiferromagnetic LRTFIM discussed in
Section 6.2 where the nearest-neighbour criticality is believed to be the correct universality class for all values of
.
7.2. Ferromagnetic Anisotropic Long-Range XY Chain in a Transverse Field
The fundamental difference between the isotropic
and anisotropic
XY chain in a transverse field is that for
the Hamiltonian is no longer
invariant (particle conserving), see Eq. (
309). Therefore, a perturbative treatment of the anisotropic model with long-range interaction from the high-field limit requires the entire procedure described in
Section 4 to perform pCUT+MC. In Ref. [
31] this procedure was applied to obtain the critical values
and critical exponents
from the gap. We review these results in this section, followed by the results for the antiferromagnetic case in
Section 7.3.
Critical values
and critical exponents
for several
values as a function of
are depicted in
Figure 30. The overall
dependence of
for
is similar to the
case (see inset of
Figure 30). However, the low-field ferromagnetic phase becomes more stable for increasing
values [
31]. Note, the perturbative parameter is
, therefore, a QPT at smaller
values means a larger extent of the phase in
. In Ref. [
306] the QPT of the anisotropic LRTFAXYM was studied with exact diagonalisation and the critical point was determined for
, which coincides perfectly with the value from pCUT+MC (see
Figure 30). For all
, Adelhardt et al. [
31] identified the same three regimes for the universality of the QPT as for the ferromagnetic LRTFIM. The gap exponent behaves analogously to the LRTFIM (
) as depicted in
Figure 30. Solely the
exponent behaves special, being
for all
values in full agreement with the existence of a multicritical point in the nearest-neighbour model. For
, the first domain identified in Ref. [
31] is the long-range mean-field regime for
with
. For large values of
the gap exponent is
and the transition belongs to the 2D-Ising universality class [
31]. Third, in the intermediate regime the critical exponents vary continuously between the two values in a non-trivial fashion [
31].
The criticality regimes for
can be understood using the same field-theoretical argument (see
Section 6.1.1) as for the ferromagnetic LRTFIM, since the underlying symmetry that is spontaneously broken is in both cases the same
symmetry of the Hamiltonian. Therefore, at
the upper critical dimension as a function of
becomes smaller than one and the transition enters the regime above the upper critical dimension. The boundary between nearest-neighbour criticality and the intermediate long-range non-trivial regime is expected to be at
[
21]. The data from Ref. [
31] presented in
Figure 30 is in good agreement with these three regimes. Nevertheless, the changes between the regimes cannot be determined accurately by the pCUT+MC approach. In
Figure 30 we can observe that the interfaces between the nearest-neighbour and intermediate long-range regime becomes more pronounced for increasing
values while at the interface between long-range mean-field and intermediate regime the deviation is the same for all values of
. The reason for the distinct deviations at the nearest-neighbour interface may be due to the finite nature of the perturbative series or may also be indicative of different corrections to scaling. At
, where the system dimension equals the upper critical dimension of the transition, there are multiplicative logarithmic corrections to the critical exponents [
124,
135,
307,
308,
309]. From the biased DlogPadé extrapolations at
, Adelhardt et al. [
31] determined these corrections to scaling at the upper critical dimension with a qualitatively good agreement for different
values.
To summarise the findings: Similar to the ferromagnetic LRTFIM, the ferromagnetic LRTFAXYM is a paradigmatic toy model extending the LRTFIM. With the pCUT+MC method described in
Section 4, the critical exponents as a function of the decay exponent
can be determined as demonstrated in Ref. [
31]. As the LRTFIM displays three distinct regimes of quantum criticality including a mean-field regime above the upper critical dimension for
for all
. The sensitivity of the series expansion method suffices to even study corrections to scaling at the upper critical dimension. The limiting case
is especially interesting as in the high-field limit the ground state contains no quantum fluctuations, the dispersion is exactly solvable in first-order perturbation theory, the underlying QFT is different to the case
, and two instead of three critical regimes were found. We would expect that the setup of a SSE QMC algorithm following the directed-loop idea [
215,
217,
310] is conceptually possible for the model.
7.3. Antiferromagnetic Anisotropic Long-Range XY Chain in a Transverse-Field
For frustrated antiferromagnetic interactions, we begin our discussion with the Ising case at
. Here, we have already discussed in
Section 6.2 that for the antiferromagnetic LRTFIM there is strong numerical evidence that it remains in the 2D-Ising universality class for all
. It was also demonstrated that it is possible to determine critical gap exponents up to
with the pCUT+MC method [
29,
31,
35].
Following the same procedure as for the ferromagnetic model, the critical values
and the critical gap exponents
were studied in Ref. [
31] and we summarise these results in
Figure 31. The critical values
shift towards larger values when decreasing
, eventually diverging for
. It is possible to reliably determine the critical exponents for the anisotropic XY chain until
, but it becomes increasingly challenging to extract the exponents for smaller
due to the shifting of
[
31]. For
, the critical values
behave qualitatively in a similar way to
. However, the larger
the more the high-field phase is stabilised (see inset of
Figure 31). As a consequence, we observe more reliable exponent estimates in the regime
for
values closer to zero. In general, the observations in Ref. [
31] are in line with a constant gap exponent
within the limitations of the method.
This promotes the scenario that the nearest-neighbour 2D-Ising universality class remains the correct universality class for all considered
values and
. For
, the critical exponents
and critical values
were determined analytically in
Section 7.1 and it was shown that the universality class of the nearest-neighbour isotropic chain remains for all
[
31]. A recent study [
311] investigated the multicritical point using exact diagonalisation for comparably small system sizes. The critical points for
and
agree well with the exact limit and the pCUT+MC values. The study further confirms 2D Ising universality up to
, however, in contrast to Ref. [
35], signatures of a different crossover criticality at
is claimed.
8. Long-Range Heisenberg Models
We review the results obtained from SSE and pCUT+MC for Heisenberg models with long-range interactions. The starting point is the nearest-neighbour Heisenberg model that can be written as
with the spin-1/2 operators
and
. Instead of a transverse field that induces quantum fluctuations in the (anti-)ferromagnetic phase of the (LR)TFIM or (LR)TFAXYM, a magnetic field would immediately break the
-symmetry of the Heisenberg Hamiltonian. Here, we introduce a dimerisation limit. We consider two layers of lattices with Heisenberg spins on each site stacked on top of each other where each spin couples to its nearest-neighbour. The nearest-neighbour coupling between the layers gives rise to interlayer dimers also referred to as rungs. In one dimension the layers are just two chains resulting in a ladder system while in two dimensions we consider two square lattice layers resulting in a square lattice bilayer model. Generalising this model for long-range interactions gives rise to intralayer as well interlayer long-range couplings. The generic antiferromagnetic Hamiltonian is then given by
where the first term couples the nearest-neighbour spins between the layers with coupling strength
, the second term describes the long-range intralayer coupling with
and the third term the long-range interlayer coupling of Heisenberg spins with
. The long-range interaction is given by
where
J is the coupling constant,
i and
j are the rung dimer indices, and
and
are the indices of the respective legs (layers) of the ladder (bilayer) model. In Eq. (
318) we use the notation
and
. The long-range interaction introduced is similar to the one introduced in the sections before, however with the difference that the alternating sign in the numerator gives rise to staggered, non-frustrating antiferromagnetic long-range interactions as we choose
. The exponent of the numerator is given by the one-norm
while the usual geometric distance in the denominator given by
can be identified as the two-norm
. When increasing the overall interaction strength
J of intralayer and interlayer coupling and likewise making the long-range decay exponent
smaller, the Heisenberg spins will antialign at odd distances and align at even distances inducing a Néel-ordered antiferromagnetic phase. For
the system becomes superextensive (analogous to the ferromagnetic LRTFIM in
Section 6.1) and for
we recover the antiferromagnetic short-range ladder or bilayer model.
In order to employ the pCUT+MC method we need an exactly soluble limit about which we can perform the series expansion. We introduce the perturbation parameter
rescaling the Hamiltonian by
and identify the first term of Eq. (
318) as the unperturbed part
which corresponds to a sum over Heisenberg dimers that can be easily diagonalised. For a single dimer, the lowest-lying state with total spin
is the antisymmetric combination of
spins
also called a rung singlet in our case with the associated energy
. The total spin
states
are three-fold degenerate with the energy
and are called triplets. Alternatively, we can use an
-symmetric basis, where the singlet and triplet states read
There is also a convenient mapping from the spin operators to bosonic,
-symmetric operators creating and annihilating these states introduced in Ref. [
312] that can be readily adapted to hard-core bosonic operators for triplet excitations [
313]. We use the mapping
where
is the Levi-Civita symbol, and insert these expressions into Eq. (
318). After some straightforward manipulations we obtain
with
is the number of rungs,
, and
. The second and third line constitute the perturbation
associated with the physical perturbation parameter
. The
and
operators are absent, as terms contributing to these operators cancel out due to the reflection symmetry about the center of the rung dimers.
For
, the ground state is given by a trivial product state of rung singlets and for small but finite
the ground state is still adiabatically connected to this product state. We refer to this adiabatically connected ground state as rung-singlet ground state. Elementary excitations in this phase are called triplons [
314] corresponding to dressed triplet excitations. For large
we expect the non-frustrating antiferromagnetic long-range interaction to induce an antiferromagnetic phase giving rise to a quantum phase transition (QPT) between these two phases. In contrast to the transverse-field Ising model and the anisotropic XY model where the ground state in the ordered phase spontaneously breaks the discrete
symmetry of the Hamiltonian, the Néel-ordered antiferromagnetic ground state for strong long-range couplings must break the continuous
symmetry of the Heisenberg Hamiltonian. Related to this observation, we present in the following two prominent theorems of great importance that apply to Hamiltonians with continuous symmetries:
First, there is the Hohenberg-Mermin-Wagner (HMW) theorem [
129,
130,
131,
294] that rules out the spontaneous breaking of continuous symmetries in one- and two dimensional systems for
. From the quantum-classical correspondence (see Ref. [
113]), we can infer that for one-dimensional systems a QPT at
breaking such symmetry should be ruled out as well. Indeed, it was shown rigorously by Pitaevskii and Stringari in 1991 [
295] that it is prohibited in one dimension for
as well. However, there is the restriction that the interaction must be sufficiently short-range, i.e., the condition
for
must hold for the long-range decay exponent. A stronger condition was given by Ref. [
315] with
where
d is the spatial dimension of the system. For the quantum case it was shown in Ref. [
316] for a staggered antiferromagnetic long-range Heisenberg chain (
) that long-range order is absent for all
.
Second, Goldstone’s theorem [
317,
318,
319] applies. The theorem states, when spontaneous breaking of a continuous symmetry occurs, it gives rise to massless excitations also known as Nambu-Goldstone modes [
317,
318,
319]. For instance, magnons that are quantised spin-wave excitations are a manifestation of such gapless Nambu-Goldstones modes inside the ordered phase of Heisenberg systems. The recent findings in Ref. [
320] actually go beyond the conventional scenario of Goldstone modes due to long-range interactions. The authors identified three regimes depending on the decay exponent
: For ferromagnetic (antiferromagnetic) interactions, they found standard Goldstone modes for
(
) and an anomalous Goldstone regime with
and
(
) for
. Interestingly, a third regime for strong long-range interactions
(
) was found where the Goldstone modes become gapped via a generalised Higgs mechanism [
320]. A recent large-scale QMC study [
38] investigating the dynamic properties of the non-frustrating staggered antiferromagnetic square lattice Heisenberg model confirmed these three scenarios and found evidence that the Higgs regime already occurs in the regime
when the Hamiltonian is still extensive [
38]. Let us mention at this point that previous results from linear spin-wave theory [
22,
321] already indicated the existence of a sublinear dispersion in the staggered antiferromagnetic long-range Heisenberg chain a decade earlier. Remarkably, the existence of gapped Goldstone modes in the Heisenberg model with long-range interactions was already pointed out in the 1960s by Refs. [
322,
323].
The HMW theorem as well as Nambu-Goldstone modes are a distinguishing feature of quantum systems with continuous symmetries like Heisenberg antiferromagnets. The quantum field theory describing dimerised Heisenberg antiferromagnets [
113] is given by the action
which is the same
n-component
theory as the one describing the transverse-field Ising model. Here, the order-parameter field
is now a 3-component instead of one-component field. We can readily include long-range interactions by adding the term
to the action (
325). For a more detailed discussion of the short-range
action we refer to the
Appendix A and to
Section 6.1.1 where the implications of long-range interactions were already discussed for
. The classical equivalent of the
action can describe the finite temperature phase transition of the ferromagnetic Heisenberg model [
1,
36,
113,
324]. Interestingly, the action of zero-temperature Heisenberg ferromagnets is not given by the above action [
113]. In fact, the quantum field theory describes a class of dimerised Heisenberg antiferromagnets that can be appropriately described by Eq. (
325) since their low-energy physics can be mapped onto the
quantum rotor model because pairwise anitferromagnetically coupled Heisenberg spins are an effective low-energy representation of quantum rotors [
113]. We can see this correspondence by looking at the Hamiltonian of the
quantum rotor model
where the first part
with
and
the angular momentum operator gives the kinetic energy of the rotors. The second part
with
is the interaction between the 3-component quantum rotors inducing parallel ordering of the rotors [
113]. The physical picture is that the kinetic energy is minimised when the orientation of the rotors is maximally uncertain and the energy of the interaction term is minimised when the rotors are aligned [
113]. The eigenvalues of the kinetic term of a single quantum rotor are given by
with
corresponding to an
-fold degenerate state. Now, given two spins
S interacting antiferromagnetically with coupling strength
forming a dimer, the total spin is
and the eigenenergies are given by
with an
-fold degeneracy. This gives a one-to-one correspondence to the kinetic energy of a rotor but with an upper bound for the energy. Thus, this mapping is only valid when considering the low-energy properties of the models [
113] and holds certainly for large
. We just introduced a powerful mapping from the low-energy properties of the antiferromagnetic Heisenberg systems like ladder and bilayer models forming Heisenberg dimers on the rungs (see Eq. (
318)) to an
quantum rotor model that is described by the action (
325). It turns out that this mapping is not only valid for large
K/
J but also in the ordered phase and at the quantum critical point [
113]. To include algebraically decaying long-range interactions, the term (
326) must be added to the action. The implications of long-range interactions for a
n-component order-parameter field in the
field theory was already extensively discussed in the 1970s [
1,
2,
3] for the classical action but it took until 2001 until the quantum analog was studied by Ref. [
20]. More studies followed by N. Defenu et al. employing functional RG approaches [
21,
243,
325].
The predictions for long-range Heisenberg models are in large parts the same as for the LRTFIM. We expect long-range mean-field behaviour with a Gaussian fixed point for
, a regime of nearest-neighbour criticality for
, and a non-trivial regime with continuously varying critical exponents in-between [
20,
21,
243,
325]. For
however, there is the major difference that a QPT associated with the spontaneous breaking of continuous symmetries is ruled out by the HMW theorem [
129,
130,
131,
294,
295] at least for
[
21,
243,
316,
325]. The boundary is then called lower critical exponent
. For the Heisenberg model, the previously discussed action (
325) with long-range interactions (
326) holds, which together with the HMW theorem rules out the breaking of the order parameter field for sufficiently short-ranged interactions and therefore a QPT in the regime
. For
a QPT breaking the continuous symmetry is predicted [
20,
21,
243,
325]. Indeed, it is long confirmed by several numerical studies [
22,
296,
297,
298,
299,
300,
301,
302,
303,
304,
305] that the HMW theorem can be circumvented if the long-range interaction is sufficiently strong.
So far, most of the studies considered only Heisenberg Hamiltonians with long-range interactions in one-dimensional systems [
22,
35,
296,
297,
298,
299,
300,
301,
302,
303,
304,
305], however often considering various modifications to the Hamiltonian. Some studies considered a one-dimensional Heisenberg chain with non-frustrating (staggered) antiferromagnetic long-range interactions [
22,
299,
304] while others included an anisotropy along the
z-components resulting in the
-model [
301,
302]. An interesting example with frustrating long-range interactions is the excatly solvable Haldane-Shastry model [
326,
327] with
that is known to show quasi-long-range order (QLRO) just like the conventional Heisenberg chain but with vanishing logarithmic corrections [
328]. For the dimerisation transition between QLRO and a valence bond solid (VBS), the frustrated long-range interaction seems to play only a minor role compared to the
-
model [
213]. Based on this, in Refs. [
213,
296,
297,
298,
329] a model combining the frustrated
-
model with additional non-frustrating staggered long-range interaction was investigated. The initial intention was to realise a one-dimensional analog of deconfined criticality [
330,
331,
332,
333] between a Néel-order phase and a VBS. As the transition was found to be of first order, the authors in Ref. [
334] later turned to the one-dimensional long-range
J-
Q model indeed finding evidence for a continuous deconfined QPT. There are also studies considering larger spin
. In Ref. [
303] a Heisenberg chain with single-ion anisotropy was studied while in Ref. [
300] the XXZ chain was under scrutiny. In both cases, a rich phase diagram was found with intriguing quantum critical properties. For instance, continuous symmmetry breaking due to long-range interactions gives rise to a new, possibly exotic, tricritical point in the XXZ chain with no analog in short-range 1d spin systems [
300].
In
Section 6 and
Section 7 we reviewed the results of various numerical studies investigating the quantum critical properties of the LRTFM [
23,
24,
25,
26,
28,
29,
30,
32,
33,
34] and the LRTFAXYM [
31] with focus on extracting the critical exponent as a function of the long-range decay
. In comparison to the LRTFIM, for long-range Heisenberg models just recently, efforts have been made to extract the critical exponents associated with the QPT as a function of the long-range decay exponent [
22,
35,
37,
206]. Ref. [
22] studying a one-dimensional long-range Heisenberg chain is one notable exception in the sense that it was ahead of its time performing large scale SSE QMC simulations for long-range Heisenberg systems already in 2005 way before the rapid progress in implementing quantum simulators in quantum optics reignited the interest in long-range interacting systems.
In the following, we will first review the results from Refs. [
37,
206] for the long-range square-lattice Heisenberg bilayer model in
Section 8.1. Then, we proceed with the discussion of the results from Ref. [
35] for long-range Heisenberg ladders in
Section 8.2. Finally, we conclude this section with the discussion of the long-range Heisenberg chain from Ref. [
22] in
Section 8.3.
8.1. Staggered Antiferromagnetic Long-Range Heisenberg Square Lattice Bilayer Model
In this section we consider two square lattices stacked directly on top of each other, where the Heisenberg spins on each lattice site interact with their nearest neighbours. When the two spins on top of each other form a rung dimer interacting with coupling strength
, the Hamiltonian describing the bilayer system is given by Eq. (
318). Here, we neglect the long-range interlayer interactions which results in the Hamiltonian
We consider staggered non-frustrating long-range interactions along the layers
of the form (
319). The short-range model (
) was subject of several studies from the 1990s onward [
162,
183,
184,
335,
336,
337,
338,
339,
340,
341] investigating its quantum-critical properties. For a critical coupling ratio
, the system undergoes a QPT breaking the continuous
symmetry of the Hamiltonian from a ground state adiabatically connected to the product singlet state towards an antiferromagnetic Néel ground state with gapless magnon excitations (Nambu-Goldstone modes). For long-range interactions we expect that these Goldstone modes are altered [
38,
320] and that the criticality of the system changes as a function of the decay exponent
. We expect short-range criticality until
, a non-trivial regime of continuously varying criticality in
with
, and a long-range mean-field regime for
[
20,
21,
243,
325].
Until recently, to the best of our knowledge, there was a complete absence of numerical confirmation of this scenario. The first data for the critical point and exponents as a function of
was published by Song et al. [
37] using SSE QMC and is now supplemented by pCUT+MC data from Adelhardt and Schmidt [
206].
In
Figure 32 the critical values from Refs. [
37,
206] are plotted together with results from functional renormalisation group (FRG) calculations from Ref. [
21] for the
quantum rotor model. The critical points
for both methods are in excellent agreement over the entire
-range. However, the critical exponents cannot be directly compared since the FSS of the magnetisation curves from SSE give
and
while pCUT+MC for the gap and spectral weight give
and
. While the QMC data shows relatively large error bars in the long-range mean-field regime, the pCUT approach overestimates the critical exponents in the short-range regime. For the two-dimensional long-range Heisenberg bilayer, the deviation is larger than for the LRTFIM in two dimensions. This is no surprise as a previous series expansion for the nearest-neighbour square lattice Heisenberg bilayer model showed similar deviations [
344]. Also, the critical exponents from QMC are better at capturing the boundaries of the long-range mean-field and the short-range regime. In general, both approaches show good agreement with FRG results from Ref. [
21] for the non-trivial intermediate regime (which apparently has its own shortcomings at the boundary to the short-range regime).
In conclusion, we summarised the results from SSE QMC and pCUT+MC for the square-lattice Heisenberg bilayer model with staggered antiferromagnetic interactions. Both approaches are in good agreement and confirm the three critical regimes [
1,
2,
3,
20,
21,
243,
325] from the two-dimensional
quantum rotor model within their limitations.
8.2. Staggered Antiferromagnetic Long-range Heisenberg Ladder Models
After the discussion of the square lattice bilayer model, we can imagine the Heisenberg ladder models as effectively reducing the dimension of the square-lattice bilayer model by one now considering two linear chains coupled by rung interactions. The long-range ladder Hamiltonian then reads
with long-range coupling
along the legs of the ladder and
between spins of different legs. Recalling the long-range interactions of Eq. (
319) and the definitions of
and
thereafter, we define two distinct Hamiltonians
and
, where the first one includes long-range interactions only along the legs while the second one is the original Hamiltonian including long-range interactions both along and in between the legs.
A sketch of these ladders is provided in
Figure 33. As for the bilayer model, for small
(c. f. Eq. (
319)) the ground state is adiabatically connected to the product state of rung singlets (rung-singlet ground state), while for strong coupling ratios the long-range interactions wants to induce an antiferromagnetic ground state. However, in contrast to the bilayer model in
Section 8.1, there is no QPT for the nearest-neighbour Heisenberg ladder [
345,
346,
347], due to the HMW theorem ruling out continuous symmetry breaking for one-dimensional quantum models [
295]. Note, the HMW theorem only rules out a QPT with continuous symmetry breaking and not a QPT in general. For instance, the nearest-neighbour isotropic XY model in a transverse field exhibits a QPT without breaking the
symmetry [
283,
284,
285,
286,
287,
288,
289,
290,
291,
292] while in the nearest-neighbour Heisenberg ladder there is no QPT at all [
345,
346,
347]. In fact, a QPT was ruled out until
in another one-dimensional model, namely the staggered antiferromagnetic long-range Heisenberg chain in Ref. [
316]. As there is a one-to-one correspondence between antiferromagnetic Heisenberg ladders and the low-energy properties of the one-dimensional
quantum rotor model [
113], we can expect a QPT predicted by the quantum field theory given by the action Eq. (
325) with Eq. (
326) for
with
[
20,
21,
243]. The existence of a lower decay critical exponent
has the consequence that there are only two quantum critical regimes. One is the long-range mean-field regime for
and the other one is a regime of continuously varying critical exponent with a non-trivial fixed point. The third regime in these models is non-critical.
The only study investigating the full parameter space of this model is Ref. [
35] using pCUT+MC and complementary linear spin-wave calculations. A previous study [
305] using QMC and DMRG investigated the
parameter line for
. There are also known limiting cases of decoupled staggered long-range Heisenberg chains at
where a QPT from a QLRO phase towards the same Néel-ordered antiferromagnetic phase occurs. We can compare the critical points determined by pCUT+MC for
and the linear spin-wave results to the SSE QMC data from Ref. [
22] and also to linear spin-wave calculations in Refs. [
22,
321] of the long-range Heisenberg chain. In fact, the linear spinwave calculations for
can be seen as a generalisation of the ones for the Heisenberg chain and therefore exactly includes them as a limiting case.
We can find a plot in
Figure 34 showing a ground-state phase diagram for both
and
from Ref. [
35]. The figure also includes other known values from Refs. [
22,
305], which fit into the overall picture of the pCUT+MC results. We also show the critical exponents
,
, and
determined by the pCUT+MC approach. It is easy to identify the two critical regimes of long-range mean-field behaviour and continuously varying critical exponents as predicted by quantum field theory [
20,
21,
243]. In the non-trivial regime, the exponents shown here diverge when approaching the lower critical dimension
which is in reasonable agreement with FRG results in Ref. [
21,
243]. The largest deviation can be seen for
which is the hardest exponent to extract from DlogPadé extrapolations. Using the (hyper-) scaling relations (
3)-(
9), the remaining critical exponents can be determined. See
Figure A2 in
Appendix D showing all critical exponents for the Heisenberg ladders. It should be noted that there are three main difficulties discussed in Ref. [
35]. First, as just stated the
exponent is difficult to determine. Second, it becomes increasingly hard to extrapolate the perturbative series in the regions
(
) for
(
) Third, the presence of logarithmic corrections to the dominant power-law behaviour about the critical point spoils the exponents around the upper critical dimension at
. In the end, all factors plays a role when determining all critical exponents due to error propagation. We can observe in
Figure A2 that several exponents from the pCUT+MC approach deviate from the FRG exponents significantly in the non-trivial regime. For instance the
exponent seems to approach a constant value
for
for pCUT+MC while the FRG predicts a diverging exponent. Another important finding is that the lower critical decay exponent
is apparently not universal in these two models and considerably smaller than the predicted value
from quantum field theory [
21,
243]. This claim is made in Ref. [
35] due to the known limiting case of the long-range Heisenberg chain [
22] from SSE QMC and due to linear-spin wave calculations [
35].
Beyond these interesting discrepancies, there was speculation about a possible deconfined quantum critical point along the
parameter line for
in Ref. [
305]. The reason for this was the fact that the staggered long-range Heisenberg ladder undergoes a QPT from a disordered phase with a non-local string order parameter towards a Néel-ordered phase with conventional order. Also, they found a sharp peak and a gap in the dynamic structure factor at the ordering momentum
in the ordered phase that could be indicative of deconfined excitations in terms of spinons [
305,
348]. Usually, when there is a QPT between two competing ordered phases the system undergoes a first-order phase transition or there must be a coexistence phase. There is also a much more exotic scenario beyond the Landau-Ginzburg-Wilson theory of phase transitions [
330]. A deconfined quantum critical point [
330,
331,
332,
333] is a second-order QPT that is not described by a "confining" order parameter but by an emergent
-symmetric gauge field with "deconfined" degrees of freedom accompanied by a fractionalisation of the order parameters [
330]. A paradigmatic example is the deconfined QPT between a Néel-ordered anitferromagnetic ground-state and a VBS state on a two-dimensional lattice [
330,
331,
333,
349,
350,
351,
352,
353,
354]. While deconfined criticality was originally proposed in two dimensions, similarities in terms of a conventional Luttinger Liquid theory description have been drawn in one dimension [
355,
356,
357,
358,
359,
360,
361,
362]. Interestingly, a very close analogy to a two-dimensional deconfined critical point was found in Ref. [
334] using a toy model with six-spin Heisenberg interactions and long-range two-spin interactions inducing a continuous phase transition between a VBS phase and an antiferromagnetic phase. A scenario between a conventional order and non-local string order as proposed by Ref. [
305] would go even beyond the one-dimensional scenarios found so far. However, it was argued in Ref. [
35] that there should be no such deconfined critical point in the above long-range Heisenberg ladders due to the critical exponents found and the fact that the rung-singlet phase is adiabatically connected to the trivial product state of rung singlets not falling into the category of symmetry-protected topological phases despite the presence of a non-local string order parameter [
35,
363]. Another possible interpretation of the finding in Ref. [
305] is probably along the lines of Refs. [
38,
320]. The observed gap in the dynamic structure factor is in agreement with strong finite-size artifacts arising from the altered dispersion
of sublinear behaviour
in the anomalous Goldstone regime
.
In this subsection we have seen that the antiferromagnetic Heisenberg ladders with staggered long-range interactions show two quantum critical regimes. One regime with long-range mean-field behaviour and a second non-trivial regime with continuously varying critical exponents. There is also a third regime which does not show any QPT because continuous symmetry breaking is ruled out by the HMW theorem. Despite the numerical confirmation of the two critical regimes predicted from quantum field theory [
20,
21,
243,
325], a discrepancy between the upper bounds of the Néel-ordered phase and the predicted lower critical exponent
was identified by Ref. [
35]. On the other hand the data in Ref. [
35] is in agreement with other literature for the limiting case of long-range Heisenberg spin chains [
22,
321]. We also discussed briefly the possibility of a deconfined QPT and the presence of anomalous Goldstone modes in one dimension for the above ladder models. To put it briefly, the Hamiltonian (
329) hosts some intriguing physics with several aspects that need further clarification.
8.3. Staggered Antiferromagnetic Long-Range Heisenberg Chain
Last, we consider another one-dimensional system, the Heisenberg chain with staggered non-frustrating antiferromagnetic long-range interactions
In the previous two models (bilayer and ladders) the unperturbed part at
consisted of uncoupled dimers with a trivial product singlet state as its ground state and local triplet excitations above. Here, the unperturbed part is the nearest-neighbour Heisenberg chain with a ground state exhibiting quasi long-range order (QLRO) and fractionalised elementary excitations. These excitations are referred to as spinons and can be seen as propagating domain walls carrying
degrees of freedom. The perturbation consists of long-range interactions that couple sites beyond its nearest-neighbours. This interaction is of the same algebraic form as Eq. (
319). Because of the non-frustrating nature of the antiferromagnetic long-range interactions, it induces a Néel-ordered antiferromagnetic phase upon increasing its coupling strength. Again, a QPT breaking the
-symmetry of the Hamiltonian is only allowed when the long-range decay exponent satisfies
due to the HMW theorem [
22,
316] and thus such a transition can be ruled out for larger decay exponents. In this model, a QPT from a QLRO towards an ordered phase is expected to occur and therefore the
theory of Eq. (
325) does not apply. The
Wess-Zumino-Witten non-linear
model [
364,
365] is known to describe the low-energy physics of Heisenberg chains and includes topological coupling to account for the presence of QLRO in the lattice model [
22,
213]. As for the
theory a long-range coupling analogous to Eq. (
326) can be added to describe the Hamiltonian (
330) [
22].
Ref. [
22] is a comprehensive study of the Heisenberg chain (
330) using large-scale SSE QMC simulations to extract the critical properties of the QPT.
The results for the critical point as well as the critical exponents
and
z can be found in
Figure 35. The overall behaviour of the critical point as a function of the decay exponent
is very similar to the one found in the previous subsection for the Heisenberg ladder. Here, the critical point diverges at about
when approaching the lower critical exponent
. The hard boundary for a QPT is again given by the HMW theorem. Yet, there is a significant difference. While for the Heisenberg ladders the disordered rung-singlet phase exists for any
, for the Heisenberg chain the QLRO phase only exists for
. Thus, for the Heisenberg chain the critical line terminates in a marginal point at
and
. For any
the perturbation parameter
becomes irrelevant and the system is always in the antiferromagnetic phase [
22]. The long-range Heisenberg chain (
330) was also studied in Ref. [
213,
296,
329] in the context of a Heisenberg chain with frustrated next-nearest neighbour and non-frustrating long-range interactions, i.e., the
-
chain with staggered long-range interactions. In both models the QLRO-Néel QPT can be identified by a level crossing between a triplet
and quintuplet
excitations.
10 The critical point from finite-size scaling of the level crossing from ED [
296] and DMRG [
329] is in very good agreement with the QMC results at
(see
Figure 35). Proceeding with the critical exponents, we can see that the critical exponent
matches well with the field-theoretical expectations
(linear in at least leading order but also a simple scaling argument predicts the linear behaviour [
22]) in the range
. For
the
from SSE QMC starts to deviate from the linear behaviour. One interesting observation was pointed out in Ref. [
35]. In both, the Heisenberg ladders and the Heisenberg chain, the linear behaviour
is expected from the underlying quantum field theory, yet the data from pCUT+MC and SSE QMC indicate a deviation from this with
for the Heisenberg ladders and
for the Heisenberg chain. It should be noted however, that the exponent from the pCUT+MC approach is determined using scaling relations and therefore can suffer from unfavourable error propagation, especially when the
exponent is involved. The dynamical exponent
z of the Heisenberg chain was extracted from SSE QMC as well. The exponent
z is one at the marginal point
and then quickly drops to
where it seems to be constant within error bars up to
. This finding is also in contrast to the RG prediction where
in leading order and the exponent is expected to be constant even in higher orders [
22]. In Ref. [
213] for
the dynamic exponent
z was determined in excellent agreement with the QMC results (see
Figure 35). Also, the results for the
model with non-frustrating long-range interactions, where the same QLRO-Néel transition is realised, the exponent is in agreement with
[
213,
296]. Further, the RG analysis in Ref. [
22] gave also a prediction for
. However, this could not be compared with QMC as it was not possible to obtain accurate estimates of
[
22]. Note also that neither the long-range transverse-field XY chain nor the Heisenberg chain show long-range mean-field behaviour.
The Heisenberg chain with staggered long-range interactions is another prime example of long-range models hosting intriguing critical behaviour. The remaining discrepancy between the numerical SSE QMC results and the underlying field-theoretical description shows that the critical properties are not yet fully settled and further exploration of the model is necessary.
9. Summary and Outlook
In this review we gave an overview of recent advancements in the investigation of quantum-critical properties of quantum magnets with long-range interactions focusing on two techniques, both based on Monte Carlo integration but complementary in spirit. On the one hand, we described pCUT+MC, where classical Monte Carlo integration is decisive in the embedding scheme of white graphs. This allows to extract series expansions of relevant physical quantities directly in the thermodynamic limit. On the other hand, SSE QMC enables calculations on large finite systems where finite-size scaling can be used to determine physical properties of the infinite system. Both quantitative and unbiased approaches take the full long-range interaction into account and can be used a priori in any spatial dimension for any geometry.
In recent years, both techniques, alongside other methods, have been applied successfully to one- and two-dimensional quantum magnets involving long-range Ising, XY, and Heisenberg interactions on various bipartite and non-bipartite lattices. In this work, we have summarised the obtained quantum-critical properties including quantum phase diagrams and the (full sets of) critical exponents for all these systems coherently. Further, we reviewed how long-range interactions are used to study quantum phase transitions above the upper critical dimension and how the scaling techniques are extended to extract these quantum critical properties from the numerical calculations. This is indeed generically the case for all unfrustrated systems in this review with the exception of the one-dimensional isotropic XY and Heisenberg chain. For frustrated systems, one can apply both MC techniques successfully for Ising interactions while in general only the pCUT+MC method is applicable (if an appropriate perturbative limit exists) due to the sign problem of SSE QMC. Nevertheless, in all frustrated cases the small- regime of long-range interactions is challenging for both approaches and further technical developments are desirable.
In the future, several extensions and research directions are interesting. As mentioned in
Section 6.3 of this review, the interplay between long-range interactions and geometric frustration is a vibrant research field at the moment. It has been demonstrated numerically and experimentally that this interplay provides a great resource to engineer exotic phases of matter [
54,
55,
109,
282,
366] with the most spectacular example being the
quantum spin liquid on the Ruby lattice [
55,
109,
366]. We expect further rapid development in the field since many promising theoretical proposals can be realised in analogue quantum simulation platforms (e. g. programmable Rydberg atom quantum simulators [
52,
55]).
In terms of methods, pCUT+MC is yet to be extended to arbitrary unit cells, larger spin values, and multi-spin interactions. The access to larger unit cells will enable the investigation of the interplay between long-range interactions and frustration on even more relevant lattice structures, e. g. the Kagome or Ruby lattice. There are systems with multi-spin interactions hosting deconfined quantum criticality [
334,
349,
350,
351,
352,
353] and an introduction of long-range interactions to this type of systems seems to be an interesting research topic [
334]. We hope to spark further interest in the development and application of pCUT+MC by other users.
The SSE QMC approach is a widely used numerical tool for the calculation of unbiased thermal averages of observables. A large variety of distinct QPT is potentially accessible by these QMC simulations. We envision that for all systems that do not suffer from a sign problem, SSE QMC, in combination with appropriate zero temperature protocols and finite-size scaling, can be used to study how long-range interactions affect QPTs beyond the standard
symmetry. The SSE QMC method has also been extended to tackle frustrated systems in a more efficient way [
259,
260,
367]. However, an efficient treatment of both, the long-range interaction and frustration, has not been introduced yet [
32,
40]. Ref. [
368] developed an SSE QMC approach to access the toric code quantum spin liquid regime [
55,
109,
366]. An application of the SSE QMC approach to extended long-range interacting Bose-Hubbard models (see
Section 5.5) along the lines of directed loop updates [
215,
216,
217] would also be a natural development. This would enable to numerically calculate observables with SSE QMC for ultracold gas experiments with optical lattices. A possible application could be the study of complex crystalline phases and their breakdown in frustrated Bose-Hubbard systems with long-range interactions [
64,
73,
111,
228].
Finally, in the context of the long-range mean-field regime above the upper critical dimension, a lot of research has been conducted regarding finite-size scaling in classical systems [
12,
13,
14,
15,
16,
17,
18,
19,
41,
124,
127,
142,
145], including the study of multiplicative logarithmic corrections for the characteristic length scale at the upper critical dimension [
17] and the investigation of the role of Fourier modes and boundary conditions [
19]. On the contrary, its quantum counterpart has only been treated successfully in recent years [
32,
34]. Besides the transfer of established concepts from classical to quantum Q-FSS, one interesting open question is a detailed understanding of the crossover regime between classical and quantum Q-FSS for small temperatures. Even though we focused on the quantum version of Q-FSS of Ref. [
34] due to the quantum nature of the models analysed in this review, the ground work has been conducted by the inventors of classical Q-FSS (Refs. [
15,
16,
17,
18,
19,
41,
124]) and, in general, many other researchers who provided valuable insight into the scaling above the upper critical dimension, e.g. Refs. [
10,
12,
13,
14,
126,
127,
142]. Overall, it is exciting that the abstract concept of dangerous irrelevant variables and the physics above the upper critical dimension is accessible in quantum-optical platforms realising long-range interactions.
Figure 1.
Susceptibility of the long-range transverse-field Ising chain for different linear system sizes from to . The smaller the system, the farther away from the critical point the susceptibility starts to deviate from the thermodynamic limit and the farther the peak position shifts away from the critical point.
Figure 1.
Susceptibility of the long-range transverse-field Ising chain for different linear system sizes from to . The smaller the system, the farther away from the critical point the susceptibility starts to deviate from the thermodynamic limit and the farther the peak position shifts away from the critical point.
Figure 2.
Rescaled correlation length according to Eq. (37) for a model above the upper critical dimension with (to be specific, the long-range transverse-field Ising chain with ). The control parameter is proportional to the transverse field. The collapse of the data around the critical point verifies the scaling Eq. (37) and therefore demonstrates that is indeed – in contrast to prior belief – not bound by the linear system size but .
Figure 2.
Rescaled correlation length according to Eq. (37) for a model above the upper critical dimension with (to be specific, the long-range transverse-field Ising chain with ). The control parameter is proportional to the transverse field. The collapse of the data around the critical point verifies the scaling Eq. (37) and therefore demonstrates that is indeed – in contrast to prior belief – not bound by the linear system size but .
Figure 3.
Schematic illustration of cluster additivity. The contribution of a disconnected cluster (clusters within dashed circle) made up of individual connected clusters A and B (yellow areas) is the sum of its individual parts.
Figure 3.
Schematic illustration of cluster additivity. The contribution of a disconnected cluster (clusters within dashed circle) made up of individual connected clusters A and B (yellow areas) is the sum of its individual parts.
Figure 4.
Illustration of the method of perturbative continuous unitary transformations (pCUT) transforming the original Hamiltonian on the left to a block-diagonal quasiparticle-conserving effective Hamiltonian on the right. The desired effective Hamiltonian is given in the limit of the flow parameter ℓ of the continuous unitary transformation . While the different quasiparticle sectors interact with each other by the offdiagonal blocks in the original Hamiltonian the offdiagonal blocks are zero in the effective Hamiltonian as they are "rotated away" during the flow.
Figure 4.
Illustration of the method of perturbative continuous unitary transformations (pCUT) transforming the original Hamiltonian on the left to a block-diagonal quasiparticle-conserving effective Hamiltonian on the right. The desired effective Hamiltonian is given in the limit of the flow parameter ℓ of the continuous unitary transformation . While the different quasiparticle sectors interact with each other by the offdiagonal blocks in the original Hamiltonian the offdiagonal blocks are zero in the effective Hamiltonian as they are "rotated away" during the flow.
Figure 5.
Schematic illustration of cluster additivity a and additivity in the one-particle basis b on a disconnected cluster (grey area) consisting of individual connected clusters A and B (yellow areas). a Cluster additivity in the one-particle basis translates to one particle being on cluster A and zero on B and vice versa. To calculate the contribution on the disconnected cluster both contributions need to be considered including the cases of zero occupancy. b Particle additivity is fulfilled when the one-particle contribution on a disconnected cluster is simply the sum of one-particle contributions on the connected clusters A and B.
Figure 5.
Schematic illustration of cluster additivity a and additivity in the one-particle basis b on a disconnected cluster (grey area) consisting of individual connected clusters A and B (yellow areas). a Cluster additivity in the one-particle basis translates to one particle being on cluster A and zero on B and vice versa. To calculate the contribution on the disconnected cluster both contributions need to be considered including the cases of zero occupancy. b Particle additivity is fulfilled when the one-particle contribution on a disconnected cluster is simply the sum of one-particle contributions on the connected clusters A and B.
Figure 6.
For operators that are not cluster additive, the contribution on the disconnected cluster originates not only from the sum of the contributions where a single particle is on A or B but also from contributions where the particle can hop between the two connected clusters.
Figure 6.
For operators that are not cluster additive, the contribution on the disconnected cluster originates not only from the sum of the contributions where a single particle is on A or B but also from contributions where the particle can hop between the two connected clusters.
Figure 7.
Illustration of a graph isomorphism and automorphism map for an example graph. a The mapping , and identity for the remaining vertices is a graph isomorphism preserving the adjacency of vertices. If such a isomorphism exists between two graphs, they are topologically equivalent. b Under a graph automorphism the edge set remains invariant, i.e., the graph is mapped onto itself. Here it is exemplified for the mapping , and identity for the remaining vertices. A graph automorphism is therefore special case of a graph isomorphism which lives the edge set invariant.
Figure 7.
Illustration of a graph isomorphism and automorphism map for an example graph. a The mapping , and identity for the remaining vertices is a graph isomorphism preserving the adjacency of vertices. If such a isomorphism exists between two graphs, they are topologically equivalent. b Under a graph automorphism the edge set remains invariant, i.e., the graph is mapped onto itself. Here it is exemplified for the mapping , and identity for the remaining vertices. A graph automorphism is therefore special case of a graph isomorphism which lives the edge set invariant.
Figure 8.
Depiction of a graph monomorphism
and a graph with colour attributes.
a An example showing of a graph monomorphism. The smaller graph on the left is mapped onto the bigger graph on the right, which is the same graph as in the
Figure 7. Explicitly, the mapping
is given by
,
,
, and
.
b A coloured graph with "green" colour attribute assigned to vertices 0 and 2 (
). If mappings are applied to colour graphs, the colour set must be left invariant, i.e., vertices with a colour must be mapped onto each other and vertices with no colour as well.
Figure 8.
Depiction of a graph monomorphism
and a graph with colour attributes.
a An example showing of a graph monomorphism. The smaller graph on the left is mapped onto the bigger graph on the right, which is the same graph as in the
Figure 7. Explicitly, the mapping
is given by
,
,
, and
.
b A coloured graph with "green" colour attribute assigned to vertices 0 and 2 (
). If mappings are applied to colour graphs, the colour set must be left invariant, i.e., vertices with a colour must be mapped onto each other and vertices with no colour as well.
Figure 9.
In contrast to the conventional approach using coloured graphs (left), where different expansion parameters or different interaction types are associated with an edge colour, for white graphs (center) the edge colour is ignored in the topological classification of graphs. Instead additional information is tracked, e.g., by associating each link with abstract expansion parameters and only substituting these abstract contribution during the embedding procedure, reintroducing the correct colour information (right), hence the name white graphs.
a For the problem of Eq. (
137) on a linear graph with two edges there are three distinct graphs as the expansion parameters
and
are associated with individual edge colours (left), but there is only one white-graph as we associated one abstract expansion parameter for each edge (center). When substituting the abstract expansion parameters with the physical one, reintroducing the correct colour, we can recover the polynomial contributions of the conventional approach (right) (c.f. Ref. [
152]).
b For the problem of Eq. (
138) on the same graph, there are also three topologically distinct graphs (left), but for the white graph contribution we have to introduce multiple abstract expansion parameters for each edge, due to the three flavours
(center). The substitution works analogously to recover the polynomial contribution form the conventional approach (right). Parameters that are not explicitly set are set to zero (c.f. Ref. [
204]).
Figure 9.
In contrast to the conventional approach using coloured graphs (left), where different expansion parameters or different interaction types are associated with an edge colour, for white graphs (center) the edge colour is ignored in the topological classification of graphs. Instead additional information is tracked, e.g., by associating each link with abstract expansion parameters and only substituting these abstract contribution during the embedding procedure, reintroducing the correct colour information (right), hence the name white graphs.
a For the problem of Eq. (
137) on a linear graph with two edges there are three distinct graphs as the expansion parameters
and
are associated with individual edge colours (left), but there is only one white-graph as we associated one abstract expansion parameter for each edge (center). When substituting the abstract expansion parameters with the physical one, reintroducing the correct colour, we can recover the polynomial contributions of the conventional approach (right) (c.f. Ref. [
152]).
b For the problem of Eq. (
138) on the same graph, there are also three topologically distinct graphs (left), but for the white graph contribution we have to introduce multiple abstract expansion parameters for each edge, due to the three flavours
(center). The substitution works analogously to recover the polynomial contribution form the conventional approach (right). Parameters that are not explicitly set are set to zero (c.f. Ref. [
204]).
Figure 10.
The graph with three edges given in
Figure 8 is embedded on the infinite lattice. We consider a coloured graph with additional colour attribute (yellow and green) for two specific vertices. The reason behind colouring the vertices might by to simply fix the graph due to translational and rotational symmetry of the lattice as we do when calculating the ground-state energy or due to the presence of a one-quasiparticle process from one coloured site to the other. For the ground-state energy contribution the embedding factor would be
. There are six possible embeddings (monomorphisms) on the infinite lattice, when the coloured vertices on the graph are correctly mapped to the coloured sites on the lattice. There are only three geometrically distinct embeddings but the number of monomorphisms is two times bigger due to an ambiguity of mapping the subgraph on the graph illustrated by the arrows.
Figure 10.
The graph with three edges given in
Figure 8 is embedded on the infinite lattice. We consider a coloured graph with additional colour attribute (yellow and green) for two specific vertices. The reason behind colouring the vertices might by to simply fix the graph due to translational and rotational symmetry of the lattice as we do when calculating the ground-state energy or due to the presence of a one-quasiparticle process from one coloured site to the other. For the ground-state energy contribution the embedding factor would be
. There are six possible embeddings (monomorphisms) on the infinite lattice, when the coloured vertices on the graph are correctly mapped to the coloured sites on the lattice. There are only three geometrically distinct embeddings but the number of monomorphisms is two times bigger due to an ambiguity of mapping the subgraph on the graph illustrated by the arrows.
Figure 12.
The workflow of the presented pCUT+MC method consists of several steps. There are three major steps. First, there is the graph generation that only has to be done once. Second, the calculation of the graph contributions with the pCUT method and third the Monte Carlo algorithm to embed the contributions on the lattice to determine the perturbative series of the quantity of interest in the thermodynamic limit.
Figure 12.
The workflow of the presented pCUT+MC method consists of several steps. There are three major steps. First, there is the graph generation that only has to be done once. Second, the calculation of the graph contributions with the pCUT method and third the Monte Carlo algorithm to embed the contributions on the lattice to determine the perturbative series of the quantity of interest in the thermodynamic limit.
Figure 13.
a) An SSE configuration of a transverse-field Ising chain with sites and a sequence length . The spatial spin direction goes from left to right and the imaginary-time dimension goes from bottom to top. The number of trivial operators in the operator sequence is 8. Filled (empty) circles represent spins aligned in direction. The propagated states correspond to the p-th row from below with the lowest row being state . b) A depiction of the possible vertices for field operators. c) A depiction of the possible vertices for constant operators. d) A depiction of the allowed vertices for ferromagnetic Ising operators. Note, a ferromagnetic Ising vertex can only connect sites which are connected by a ferromagnetic bond in the Hamiltonian. e) A depiction of the allowed vertices for antiferromagnetic Ising operators. Note, an antiferromagnetic Ising vertex can only connect sites which are connected by an antiferromagnetic bond in the Hamiltonian.
Figure 13.
a) An SSE configuration of a transverse-field Ising chain with sites and a sequence length . The spatial spin direction goes from left to right and the imaginary-time dimension goes from bottom to top. The number of trivial operators in the operator sequence is 8. Filled (empty) circles represent spins aligned in direction. The propagated states correspond to the p-th row from below with the lowest row being state . b) A depiction of the possible vertices for field operators. c) A depiction of the possible vertices for constant operators. d) A depiction of the allowed vertices for ferromagnetic Ising operators. Note, a ferromagnetic Ising vertex can only connect sites which are connected by a ferromagnetic bond in the Hamiltonian. e) A depiction of the allowed vertices for antiferromagnetic Ising operators. Note, an antiferromagnetic Ising vertex can only connect sites which are connected by an antiferromagnetic bond in the Hamiltonian.
Figure 14.
Assignment of leg numbers to operator legs for an operator at propagation step p. a) Illustration of a two-site operator with four real vertex legs. b) Illustration of a single-site operator with only two real vertex legs and two ghost legs.
Figure 14.
Assignment of leg numbers to operator legs for an operator at propagation step p. a) Illustration of a two-site operator with four real vertex legs. b) Illustration of a single-site operator with only two real vertex legs and two ghost legs.
Figure 15.
Illustration for the segmentation and construction of the doubly-linked list for an exemplary configuration of the ferromagnetic transverse-field Ising model. a) Segmentation of a configuration into disjoint clusters including the numbers of the operator legs. Ghost legs are not depicted. The coloured lines with ellipses at each end depict the operator legs that are linked. Each colour represents one cluster in the off-diagonal update. b) Depiction of a doubly-linked list for the configuration shown in a). The left column represents the entry numbers in the list and the right column the corresponding legs to which the entry is connected to. The colour represents the clusters in a) to which the connection belongs to. Ghost legs are linked to the value and are shaded in grey.
Figure 15.
Illustration for the segmentation and construction of the doubly-linked list for an exemplary configuration of the ferromagnetic transverse-field Ising model. a) Segmentation of a configuration into disjoint clusters including the numbers of the operator legs. Ghost legs are not depicted. The coloured lines with ellipses at each end depict the operator legs that are linked. Each colour represents one cluster in the off-diagonal update. b) Depiction of a doubly-linked list for the configuration shown in a). The left column represents the entry numbers in the list and the right column the corresponding legs to which the entry is connected to. The colour represents the clusters in a) to which the connection belongs to. Ghost legs are linked to the value and are shaded in grey.
Figure 16.
a) An SSE configuration of a Heisenberg chain with ten sites and a sequence length of 27. The number of trivial operators is 5. Filled (empty) circles represent spins in direction. The propagated states correspond to the p-th row in the configuration. b) A depiction of the allowed vertices for ferromagnetic diagonal operators. c) A depiction of the allowed vertices for ferromagnetic off-diagonal operators. d) A depiction of the allowed vertices for antiferromagnetic diagonal operators. e) A depiction of the allowed vertices for antiferromagnetic off-diagonal operators. Note that (anti)ferromagnetic vertices can only connect sites which are connected by (anti)ferromagnetic bonds in the Hamiltonian.
Figure 16.
a) An SSE configuration of a Heisenberg chain with ten sites and a sequence length of 27. The number of trivial operators is 5. Filled (empty) circles represent spins in direction. The propagated states correspond to the p-th row in the configuration. b) A depiction of the allowed vertices for ferromagnetic diagonal operators. c) A depiction of the allowed vertices for ferromagnetic off-diagonal operators. d) A depiction of the allowed vertices for antiferromagnetic diagonal operators. e) A depiction of the allowed vertices for antiferromagnetic off-diagonal operators. Note that (anti)ferromagnetic vertices can only connect sites which are connected by (anti)ferromagnetic bonds in the Hamiltonian.
Figure 17.
Illustration showing the construction of the doubly-linked vertex list and the off-diagonal deterministic loop update for unfrustrated Heisenberg models. As an example, a Heisenberg chain with periodic boundary conditions and nearest-neighbour antiferromagnetic and next-nearest-neighbour ferromagnetic interactions is considered. a) Illustration of a configuration including the numbers of the operator legs belonging to the respective operators. The coloured lines with ellipses at each end depict the operator legs which are linked. Each colour represents one loop in the off-diagonal update. b) Depiction of a doubly-linked list for the configuration shown in a). The left column represents the entry numbers in the list and the right column the corresponding legs to which the entry is connected to. The colour represents the loops in a) to which the connection is belonging to.
Figure 17.
Illustration showing the construction of the doubly-linked vertex list and the off-diagonal deterministic loop update for unfrustrated Heisenberg models. As an example, a Heisenberg chain with periodic boundary conditions and nearest-neighbour antiferromagnetic and next-nearest-neighbour ferromagnetic interactions is considered. a) Illustration of a configuration including the numbers of the operator legs belonging to the respective operators. The coloured lines with ellipses at each end depict the operator legs which are linked. Each colour represents one loop in the off-diagonal update. b) Depiction of a doubly-linked list for the configuration shown in a). The left column represents the entry numbers in the list and the right column the corresponding legs to which the entry is connected to. The colour represents the loops in a) to which the connection is belonging to.
Figure 18.
Illustration of the beta doubling method for the one-dimensional LRTFIM in the short-range regime with decay exponent at a transverse field of . The simulation starts at (leftmost points) for every system size. The inverse temperature is then doubled in every beta-doubling step until the maximum is reached. All of the shown magnetisation curves seem to be converged to zero temperature. Larger systems with were discarded as they do not appear to be fully converged yet.
Figure 18.
Illustration of the beta doubling method for the one-dimensional LRTFIM in the short-range regime with decay exponent at a transverse field of . The simulation starts at (leftmost points) for every system size. The inverse temperature is then doubled in every beta-doubling step until the maximum is reached. All of the shown magnetisation curves seem to be converged to zero temperature. Larger systems with were discarded as they do not appear to be fully converged yet.
Figure 19.
Sketch of the three distinct critical regimes of the QPT in the ferromagnetic LRTFIM. For one- and two-dimensional systems, all three regimes exist and the boundaries can be obtained using the expressions in the figure. For there is only the long-range mean-field and the nearest-neighbour mean-field regime with a boundary at .
Figure 19.
Sketch of the three distinct critical regimes of the QPT in the ferromagnetic LRTFIM. For one- and two-dimensional systems, all three regimes exist and the boundaries can be obtained using the expressions in the figure. For there is only the long-range mean-field and the nearest-neighbour mean-field regime with a boundary at .
Figure 20.
Critical field values and exponents from numerical studies of the ferromagnetic LRTFIM on the linear chain. The panels display
(top),
(second row left),
(second row right),
(third row left),
(third row right),
(fourth row left),
(fourth row right), and
z (bottom). The labels refer to references in the following way: ’SSE QMC (2021) [
32,
247]’, ’SSE QMC (2022) [
34,
246]’, ’pCUT+MC (2022) [
34,
246]’, ’pCUT+MC
from SSE (2022) [
34,
246]’, ’DMRG (2018) [
28]’, ’FRG (2017) [
21]’, ’PI QMC (2021) [
33]’ and ’QMC+SPO (2023)’ [
248]. The values for the critical exponents of the transition in the nearest-neighbour model
,
,
,
,
,
, and
[
245,
249,
250] and in the long-range mean-field regime
,
,
,
,
,
, and
[
20,
21] are given by the dashed lines.
Figure 20.
Critical field values and exponents from numerical studies of the ferromagnetic LRTFIM on the linear chain. The panels display
(top),
(second row left),
(second row right),
(third row left),
(third row right),
(fourth row left),
(fourth row right), and
z (bottom). The labels refer to references in the following way: ’SSE QMC (2021) [
32,
247]’, ’SSE QMC (2022) [
34,
246]’, ’pCUT+MC (2022) [
34,
246]’, ’pCUT+MC
from SSE (2022) [
34,
246]’, ’DMRG (2018) [
28]’, ’FRG (2017) [
21]’, ’PI QMC (2021) [
33]’ and ’QMC+SPO (2023)’ [
248]. The values for the critical exponents of the transition in the nearest-neighbour model
,
,
,
,
,
, and
[
245,
249,
250] and in the long-range mean-field regime
,
,
,
,
,
, and
[
20,
21] are given by the dashed lines.
Figure 21.
Critical field values and exponents from numerical studies of the ferromagnetic LRTFIM on the square lattice. The panels display
(top),
(middle left),
z (middle right),
(bottom left),
(bottom right). The data points ’SSE QMC (2021)’ for
,
and
are from Ref. [
32,
247]. The data points ’pCUT+MC (2019)’ for
and
are from Ref. [
29]. The data points ’FRG (2017)’ for
and
z originate from the functional RG study in Ref. [
21]. The values for the critical exponents of the transition in the nearest-neighbour model
,
, and
[
251] and in the long-range mean-field regime
,
, and
[
20,
21] are given by the dashed lines.
Figure 21.
Critical field values and exponents from numerical studies of the ferromagnetic LRTFIM on the square lattice. The panels display
(top),
(middle left),
z (middle right),
(bottom left),
(bottom right). The data points ’SSE QMC (2021)’ for
,
and
are from Ref. [
32,
247]. The data points ’pCUT+MC (2019)’ for
and
are from Ref. [
29]. The data points ’FRG (2017)’ for
and
z originate from the functional RG study in Ref. [
21]. The values for the critical exponents of the transition in the nearest-neighbour model
,
, and
[
251] and in the long-range mean-field regime
,
, and
[
20,
21] are given by the dashed lines.
Figure 22.
Exponent Ϙ extracted by the data collapse of
and
for different decay exponents of the LRTFIM on the linear chain [
34]. The black dashed line depicts the prediction by Q-FSS
, while the grey dashed line shows the prediction from standard FSS
. In the regime above the upper critical dimension for
, the predictions start to deviate and the extracted values for Ϙ are clearly in line with the Q-FSS scenario. Figure adapted from Ref. [
34].
Figure 22.
Exponent Ϙ extracted by the data collapse of
and
for different decay exponents of the LRTFIM on the linear chain [
34]. The black dashed line depicts the prediction by Q-FSS
, while the grey dashed line shows the prediction from standard FSS
. In the regime above the upper critical dimension for
, the predictions start to deviate and the extracted values for Ϙ are clearly in line with the Q-FSS scenario. Figure adapted from Ref. [
34].
Figure 23.
Critical field values and exponents from numerical studies of the antiferromagnetic LRTFIM on the chain. The upper panel displays critical field values
, the lower left panel displays values for the critical exponents
and
, and the lower right panel values for the critical exponents
. The data points ’SSE QMC (2021)’ and ’SSE QMC
(2021)’ for
,
and
are from Ref. [
32,
247]. The data points ’pCUT+MC (2019)’ for
and
are from Ref. [
29]. The data points ’TDVP (2012)’ for
,
and
originate from Ref. [
23]. The data points ’DMRG (2016)’ for
are from Ref. [
24]. The data points ’DMRG (2017)’ for
and
originate from Ref. [
26]. The values for the critical exponents of the transition in the nearest-neighbour model
,
and
[
245,
249,
250,
261] are given by the black dashed lines.
Figure 23.
Critical field values and exponents from numerical studies of the antiferromagnetic LRTFIM on the chain. The upper panel displays critical field values
, the lower left panel displays values for the critical exponents
and
, and the lower right panel values for the critical exponents
. The data points ’SSE QMC (2021)’ and ’SSE QMC
(2021)’ for
,
and
are from Ref. [
32,
247]. The data points ’pCUT+MC (2019)’ for
and
are from Ref. [
29]. The data points ’TDVP (2012)’ for
,
and
originate from Ref. [
23]. The data points ’DMRG (2016)’ for
are from Ref. [
24]. The data points ’DMRG (2017)’ for
and
originate from Ref. [
26]. The values for the critical exponents of the transition in the nearest-neighbour model
,
and
[
245,
249,
250,
261] are given by the black dashed lines.
Figure 24.
Critical field values and exponents from numerical studies of the antiferromagnetic LRTFIM on the square lattice. The upper panel displays critical field values
, the lower left panel displays values for the critical exponents
and
, and the lower right panel values for the critical exponents
. The data points ’SSE QMC (2021)’ for
,
and
are from Ref. [
32,
247]. The data points ’pCUT+MC (2019)’ for
and
originate from Ref. [
29]. The values for the critical exponents of the transition in the nearest-neighbour model
,
and
[
251] are given by the dashed lines.
Figure 24.
Critical field values and exponents from numerical studies of the antiferromagnetic LRTFIM on the square lattice. The upper panel displays critical field values
, the lower left panel displays values for the critical exponents
and
, and the lower right panel values for the critical exponents
. The data points ’SSE QMC (2021)’ for
,
and
are from Ref. [
32,
247]. The data points ’pCUT+MC (2019)’ for
and
originate from Ref. [
29]. The values for the critical exponents of the transition in the nearest-neighbour model
,
and
[
251] are given by the dashed lines.
Figure 25.
Illustration of the four local spin configurations called flippable plaquettes. The arrows denote the local spin orientation in z-direction. The grey lines in the back visualise the triangular lattice. Solid (dotted) black lines are depicted on (anti)ferromagnetically aligned bonds. Note, flipping the spin in the center of each configuration maps the left to the right configuration in each row and vice versa.
Figure 25.
Illustration of the four local spin configurations called flippable plaquettes. The arrows denote the local spin orientation in z-direction. The grey lines in the back visualise the triangular lattice. Solid (dotted) black lines are depicted on (anti)ferromagnetically aligned bonds. Note, flipping the spin in the center of each configuration maps the left to the right configuration in each row and vice versa.
Figure 26.
Phase diagrams for the long-range transverse-field Ising model on the triangular lattice (left panel) and triangular lattice cylinders with infinite extent in
x-direction and a circumference of 6 sites in
y-direction (YC6) (right panel). Right panel: The critical field values
for the triangular lattice for the QPT between the high-field polarised phase originate from pCUT+MC calculations [
29]. The inset of the right panel represents the critical exponents
determined from the series expansion [
29]. The black dashed line represents the critical value
of the 3D-XY universality class [
280,
281]. The critical exponent
confirms the
D-XY universality class within the limitations of the series expansion. Left panel: The transitions values between the high-field
x-polarised phase and the
-clock-ordered phase are from the gap closing of the high-field series obtained from pCUT+MC [
30]. The transition between the plain stripe low-field phase and the
-clock-orderd phase is termined via a level crossing of both ground-state energies which were calculated perturbatively [
30]. The phase diagrams for small
values are not yet conclusively determined [
27,
29,
30].
Figure 26.
Phase diagrams for the long-range transverse-field Ising model on the triangular lattice (left panel) and triangular lattice cylinders with infinite extent in
x-direction and a circumference of 6 sites in
y-direction (YC6) (right panel). Right panel: The critical field values
for the triangular lattice for the QPT between the high-field polarised phase originate from pCUT+MC calculations [
29]. The inset of the right panel represents the critical exponents
determined from the series expansion [
29]. The black dashed line represents the critical value
of the 3D-XY universality class [
280,
281]. The critical exponent
confirms the
D-XY universality class within the limitations of the series expansion. Left panel: The transitions values between the high-field
x-polarised phase and the
-clock-ordered phase are from the gap closing of the high-field series obtained from pCUT+MC [
30]. The transition between the plain stripe low-field phase and the
-clock-orderd phase is termined via a level crossing of both ground-state energies which were calculated perturbatively [
30]. The phase diagrams for small
values are not yet conclusively determined [
27,
29,
30].
Figure 27.
Sketch of a generic phase diagram of a LRFTIM in for which a degenerate subspace at
and
breaks down in an order-by-disorder scenario for
and into a different crystalline state due to the long-range interactions (LRIs)
. An example ist the LRTFIM on the triangular lattice where the order-by-disorder phase is the clock-ordered phase and the crystalline phase is the plain stripe phase [
27,
29,
30,
110,
276,
278,
279]. The transition between the crystalline phase and the order-by-disorder phase is believed to be a first-order lever-crossing transition [
29,
30,
110,
276].
Figure 27.
Sketch of a generic phase diagram of a LRFTIM in for which a degenerate subspace at
and
breaks down in an order-by-disorder scenario for
and into a different crystalline state due to the long-range interactions (LRIs)
. An example ist the LRTFIM on the triangular lattice where the order-by-disorder phase is the clock-ordered phase and the crystalline phase is the plain stripe phase [
27,
29,
30,
110,
276,
278,
279]. The transition between the crystalline phase and the order-by-disorder phase is believed to be a first-order lever-crossing transition [
29,
30,
110,
276].
Figure 28.
Quantum phase diagram of the ferromagnetic nearest-neighbour XY chain in a transverse-field (see Ref. [
283]).
and
denote multicritical points. The phase transition between the symmetric high-field polarised phases and magnetically ordered low-field phases are of (1+1)D Ising universality for
. For
the transition at the multicritical points
and
has critical exponents
and
. The transition between the
and the
is of (Ising)
2 type.
Figure 28.
Quantum phase diagram of the ferromagnetic nearest-neighbour XY chain in a transverse-field (see Ref. [
283]).
and
denote multicritical points. The phase transition between the symmetric high-field polarised phases and magnetically ordered low-field phases are of (1+1)D Ising universality for
. For
the transition at the multicritical points
and
has critical exponents
and
. The transition between the
and the
is of (Ising)
2 type.
Figure 29.
Critical exponents
z and
for the ferromagnetic long-range transverse-field XY model as a function of
. The data points
and
were determined in Ref. [
31] by studying the dispersion of the elementary excitations at the critical point that can be determined exactly in first-order perturbation theory from the high-field limit. The lines
and
are theoretical predictions from the QFT investigated in Ref. [
31]. The region with the blue-shaded background denotes the regime in which the nearest-neighbour XY universality occurs [
283], while for
the QPT is in a long-range regime with continuously varying exponents.
Figure 29.
Critical exponents
z and
for the ferromagnetic long-range transverse-field XY model as a function of
. The data points
and
were determined in Ref. [
31] by studying the dispersion of the elementary excitations at the critical point that can be determined exactly in first-order perturbation theory from the high-field limit. The lines
and
are theoretical predictions from the QFT investigated in Ref. [
31]. The region with the blue-shaded background denotes the regime in which the nearest-neighbour XY universality occurs [
283], while for
the QPT is in a long-range regime with continuously varying exponents.
Figure 30.
Critical gap exponent
and critical values of
(see inset) as a function of
for the anisotropic ferromagnetic transverse-field XY model. The data points ’pCUT+MC’ are improved results from Ref. [
31] and that data point ’ED’ is from Ref. [
306]. The anisotropy parameter
is tuned from
(Ising) to
(isotropic XY). As discussed in
Section 7.1 the isotropic case is analytically solvable with
and
. The black lines denote the values of
if the ferromagnetic LRTFIM in the nearest-neighbour and the long-range mean-field regime.
Figure 30.
Critical gap exponent
and critical values of
(see inset) as a function of
for the anisotropic ferromagnetic transverse-field XY model. The data points ’pCUT+MC’ are improved results from Ref. [
31] and that data point ’ED’ is from Ref. [
306]. The anisotropy parameter
is tuned from
(Ising) to
(isotropic XY). As discussed in
Section 7.1 the isotropic case is analytically solvable with
and
. The black lines denote the values of
if the ferromagnetic LRTFIM in the nearest-neighbour and the long-range mean-field regime.
Figure 31.
Critical gap exponent
and critical values of
(see inset) as a function of
for the anisotropic antiferromagnetic transverse-field XY model [
31]. The anisotropy parameter
is tuned from
(Ising) to
(isotropic XY). As discussed in
Section 7.1 the isotropic case is analytically solvable with
and
.
Figure 31.
Critical gap exponent
and critical values of
(see inset) as a function of
for the anisotropic antiferromagnetic transverse-field XY model [
31]. The anisotropy parameter
is tuned from
(Ising) to
(isotropic XY). As discussed in
Section 7.1 the isotropic case is analytically solvable with
and
.
Figure 32.
Critical values and exponents from numerical studies of the Néel- ordering transition in the unfrustrated antiferromagnetic long-range Heisenberg square lattice bilayer model. The upper panel shows the critical values. The middle left panel displays critical exponent values
, the middle right panel the exponent
, the lower right panel the exponent
, and the lower right panel the one-particle spectral weight exponent
. The data points ’QMC (2024)’ for
and
originate from Ref. [
37]. The data points ’pCUT+MC (2024)’ for
and
originate from Ref. [
206]. The data points ’FRG (2017)’ for
and
originate from Ref. [
21]. The black dashed lines denote the critical exponents in the regime of short-range
criticality (
,
,
, and
[
342,
343]) and long-range mean-field criticality (
,
,
, and
[
20,
21]).
Figure 32.
Critical values and exponents from numerical studies of the Néel- ordering transition in the unfrustrated antiferromagnetic long-range Heisenberg square lattice bilayer model. The upper panel shows the critical values. The middle left panel displays critical exponent values
, the middle right panel the exponent
, the lower right panel the exponent
, and the lower right panel the one-particle spectral weight exponent
. The data points ’QMC (2024)’ for
and
originate from Ref. [
37]. The data points ’pCUT+MC (2024)’ for
and
originate from Ref. [
206]. The data points ’FRG (2017)’ for
and
originate from Ref. [
21]. The black dashed lines denote the critical exponents in the regime of short-range
criticality (
,
,
, and
[
342,
343]) and long-range mean-field criticality (
,
,
, and
[
20,
21]).
Figure 33.
Illustration of the quantum spin ladders with long-range interactions. For nearest-neighbour interactions (
) both long-range ladders
(left) and
(right) reduce to the same Heisenberg ladder. The coupling on the rungs
is illustrated with black lines and the long-range coupling along the legs
and in between the legs
is depicted in blue for
and in purple for
. The figure is adapted from Ref. [
35].
Figure 33.
Illustration of the quantum spin ladders with long-range interactions. For nearest-neighbour interactions (
) both long-range ladders
(left) and
(right) reduce to the same Heisenberg ladder. The coupling on the rungs
is illustrated with black lines and the long-range coupling along the legs
and in between the legs
is depicted in blue for
and in purple for
. The figure is adapted from Ref. [
35].
Figure 34.
Critical values and exponents from numerical studies of the Néel-ordering transition in the unfrustrated antiferromagnetic long-range Heisenberg ladders. The panels show: critical values
(upper left),
(upper right),
(lower left), and
(lower right). The data points ’pCUT+MC
’, ’SW
’, ’pCUT+MC
’, ’SW
’ originate from Ref. [
35] and refer to parallel (||) and parallel + diagonal (⋈) interactions. The ’SSE QMC (2005)’ data point from Ref. [
22] shows a
value on the long-range Heisenberg chain which corresponds to the limiting case of decoupled legs. The ’QMC (2022)’ data show a
value for bowtie ladders [
305]. The data points ’FRG 2020’ are from Ref. [
243] and show the critical exponents for the one-dimensional
quantum rotor model. The blue shaded region denotes the
regime in which long-range mean-field criticality is expected. The black dashed lines denote long-range mean field critical exponents.
Figure 34.
Critical values and exponents from numerical studies of the Néel-ordering transition in the unfrustrated antiferromagnetic long-range Heisenberg ladders. The panels show: critical values
(upper left),
(upper right),
(lower left), and
(lower right). The data points ’pCUT+MC
’, ’SW
’, ’pCUT+MC
’, ’SW
’ originate from Ref. [
35] and refer to parallel (||) and parallel + diagonal (⋈) interactions. The ’SSE QMC (2005)’ data point from Ref. [
22] shows a
value on the long-range Heisenberg chain which corresponds to the limiting case of decoupled legs. The ’QMC (2022)’ data show a
value for bowtie ladders [
305]. The data points ’FRG 2020’ are from Ref. [
243] and show the critical exponents for the one-dimensional
quantum rotor model. The blue shaded region denotes the
regime in which long-range mean-field criticality is expected. The black dashed lines denote long-range mean field critical exponents.
Figure 35.
Critical values and exponents from numerical studies of the QLRO-Néel transition in the unfrustrated antiferromagnetic long-range Heisenberg chain. The upper panel shows critical values
, the lower left shows the critical exponent
z and the lower right shows
. The data points ’SSE QMC 2005’ are from Ref. [
22], ’ED 2010’ are from Refs. [
213,
296] and the single data point ’DMRG 2018’ is from Ref. [
329]. The dashed line is the for
is the prediction from first order RG and scaling arguments provided in Ref. [
22].
Figure 35.
Critical values and exponents from numerical studies of the QLRO-Néel transition in the unfrustrated antiferromagnetic long-range Heisenberg chain. The upper panel shows critical values
, the lower left shows the critical exponent
z and the lower right shows
. The data points ’SSE QMC 2005’ are from Ref. [
22], ’ED 2010’ are from Refs. [
213,
296] and the single data point ’DMRG 2018’ is from Ref. [
329]. The dashed line is the for
is the prediction from first order RG and scaling arguments provided in Ref. [
22].
Table 1.
Definitions of critical exponents by means of the singularities of thermodynamic quantities for a magnetic phase transition. The free energy density is denoted by
f. Note that the control parameter susceptibility associated with the critical exponent
coincides with the heat capacity only for thermal phase transitions, where
, while for QPTs the meaning depends on the control parameter triggering the phase transition [
128].
Table 1.
Definitions of critical exponents by means of the singularities of thermodynamic quantities for a magnetic phase transition. The free energy density is denoted by
f. Note that the control parameter susceptibility associated with the critical exponent
coincides with the heat capacity only for thermal phase transitions, where
, while for QPTs the meaning depends on the control parameter triggering the phase transition [
128].
Observable |
Definition |
Crit. Exp. |
Singularity |
Characteristic length
|
via
|
|
|
Energy gap
Charact. time scale
|
via
|
|
|
Order parameter m
|
|
|
|
|
|
Order-parameter
susceptibility
|
|
|
|
Control-parameter
susceptibility
|
|
|
|
Correlation
function
|
|
|
|