Chapter 0. Introduction

0. Learning objectives

Upon completion of this chapter, the student will understand

  • The motivation of enhanced sampling methods.
  • Key terminologies such as phase/configuration space, ensemble, Boltzmann distribution, collective variable, free energy profile, and potential of mean force.
  • The derivation of the Boltzmann distribution, potential of mean force, and free energy profile.

1. Motivation of enhanced sampling methods

As supercomputing resources have become increasingly available in recent years1, molecular dynamics (MD) has become a standard tool for studying atomistic-level behaviors of molecular systems in various scientific disciplines2. However, an MD simulation can accurately describe the thermodynamic and structural properties of a system only if it achieves ergodicity during its simulation timescale. This means that the simulation must explore the phase space comprehensively, or equivalently, ensure that all possible configurations are sampled. This is usually a challenging task in real-world systems, as the metastable states of interest are frequently separated by substantial free energy barriers that are difficult to overcome by random fluctuations within a reasonable amount of time (see Figure 1 below, or the leftmost panel of Figure 1 in Chapter 2).

Figure 1. A demonstration of sampling a 2D free energy surface f(x1,x2)=x146x12+x1+x22, with the dots representing the position of the system in the 2D space (x1,x2) at different simulation steps. The figure shows that the system was unable to cross the free energy barrier to reach the other metastable state if the standard MD approach is used.

To overcome this limitation, researchers have developed various enhanced sampling methods in recent decades. One subset of these methods involves sampling along a set of predefined collective variables (CVs). These CVs, as we denote as a d-dimensional vector ξ=(ξ1(x),ξ2(x),,ξd(x)), can be any functions of the system coordinates (x), such as torsions, coordination numbers, or radii of gyration. Ideally, CVs should represent a small number of degrees of freedom that capture the slowest degrees of freedom of the system and distinguish metastable states of interest. Some examples of CV-based methods include metadynamics, adaptive biasing force, umbrella sampling, and replica-exchange umbrella sampling, which will all be covered in this mini-course.

2. Terminologies

Before diving into the following chapters, we will review some key concepts in statistical mechanics by introducing a few terminologies one may come across in enhanced sampling methods. This section serves as a quick review of the essential concepts to better understand the tutorials in later chapters. Readers are encouraged to visit Section 2.2 (Glossary of essential notions) in the review by Hénin et al.3 for a more exhaustive list of terms.

Timescale

The real timescale of a molecular system can vary significantly depending on the size and complexity of the system, as well as the process of interest. As illustrated in Figure 2, the timescales of molecular motions can range from femtoseconds to seconds or even longer. For instance, bond vibrations, the fastest motion in a molecular system, typically occur on the order of femtoseconds. In contrast, biological processes such as allosteric transitions or protein-ligand binding have timescales considerably longer than bond vibrations, making the transitions between metastable states of interest rare events. This long timescale is due to the high free energy barriers separating metastable states, as discussed in the previous section.

Importantly, to ensure that the motion of atoms in a system is captured and no short-lived events are missed, such as bond vibrations and molecular rotations, the time step used in molecular dynamics simulations must be smaller than the shortest timescale in the system. This inevitably small time step makes simulations that study biological processes computationally expensive, easily requiring 108 steps or more. To reduce the computational cost of simulating long-timescale processes, researchers frequently use enhanced sampling methods to bias the system towards the metastable states of interest. These methods make rare events more probable in a shorter timescale, thus reducing the number of required simulation steps.

Figure 2. The timescales of different molecular motions.

Phase space and configuration space

In statistical mechanics, the phase space of a system is a space where all the possible states of the system can be represented. In three-dimensional Cartesian coordinate systems, a state is characterized by the coordinates xR3N and momenta pR3N of the system, where N is the number of particles of the system. Consequently, each state can be represented by a unique point in the phase space, which is the joint space (x,p).

In contrast, the configuration space is defined by all the possible configurations (defined solely by x) of the system, so its dimensionality (3N) is half of that of the phase space (6N). When simulating long-timescale events using molecular dynamics, it is crucial to have sufficient sampling in the configuration space to capture the distribution of all possible configurations of the system. This ensures that rare events and transitions between different configurations are adequately represented. It is this need for comprehensive sampling in the configuration space motivates the development of enhanced sampling methods, which are the focus of this mini-course.

Microstate and macrostate

In the case where a system is composed of a larger number of particles, such as atoms or molecules, a microstate is a particular spatial arrangement of the particles in the system, i.e., a configuration. On the other hand, a macrostate, which can be defined by macroscopic thermodynamics variables (e.g., temperature, pressure, or total energy), is a collection of many possible microstates. For example, in a system composed of particles A and B that can either have an energy of 0 or 1, there are only three possible macrostates, which correspond to total energies of 0, 1, and 2, respectively. In the macrostate having a total energy of 2, there are two possible microstates: particles A and B having energies of 0 and 1, and the other way around.

Boltzmann distribution

To accurately calculate thermodynamic properties from molecular dynamics, we need to sample the Boltzmann distribution so that the system is at equilibrium. The Boltzmann distribution, sometimes called a Gibbs distribution, is a probability distribution that characterizes the distribution of states of a system in thermal equilibrium at a given temperature. It describes the probability of finding a particle in a particular energy state given its energy and the temperature of the system. Specifically, in the case where the energy levels are discrete, the normalized probability pi of finding the system being in state i can be expressed as follows: (1)pi=eβEiieβEieβEi where Ei is the total energy (the sum of potential and kinetic energy) of state i, and β=1/kBT is the inverse temperature (where kB is the Boltzmann constant (1.38×1023 JK1), and T is the system temperature). Frequently, we denote the normalization constant as Q=ieβEi, which is the so-called partition function.

In a phase space, where a state is defined by its continuous variables, coordinates x and momenta p, we can rewrite the Boltzmann distribution as (2)μ(x,p)=eβ(U(x)+K(p))eβ(U(x)+K(p))dxdpeβ(U(x)+K(p)) where μ(x,p), analogous to pi in the discrete space, is the probability of finding the system to have coordinates x and momenta p. In this case, we have the partition function of the phase space Q=eβ(U(x)+K(p))dxdp=(eβU(x)dx)(eβK(p)dp). Note that in Equation 2, we express the total energy as the sum of the potential energy U(x) and the kinetic energy K(p).

From Equation 2, we can derive the configurational distribution (the Boltzmann distribution in the configuration space) by integrating μ(x,p) with respect to the momenta p: (3)ν(x)=μ(x,p)dp=1Qeβ(U(x)+K(p))dp=eβU(x)(eβK(p)dp)(eβU(x)dx)(eβK(p)dp) Finally, we have (4)ν(x)=eβU(x)eβU(x)dxeβU(x) Frequently, Equation 4 is written as ν(x)=1ZeβU(x), where Z=eβU(x)dx is the configurational partition function.

Supplementary note: Maxwell-Boltzmann distribution

Different from the Boltzmann distribution, the Maxwell-Boltzmann distribution is a special case of the Boltzmann distribution. It is a probability distribution that describes the speed (v) of non-interacting gas particles in thermal equilibrium at a given temperature. In a three-dimensional velocity space, the Maxwell-Boltzmann distribution is given by (S1)f(v)=(m2πkBT)3/2emv22kBTemv22kBT

Supplementary note: Derivation of the Boltzmann distribution

The derivation of the Boltzmann distribution (Equation 2) is based on the second law of thermodynamics, which states that any spontaneous process will tend to maximize the total entropy of the system. By definition, the entropy of a system can be expressed as (S2)S=kBlnW where W is the multiplicity, i.e., the number of microstates corresponding to a particular macrostate of a thermodynamic system. Now, we consider a system of N particles, where ni particles have an energy of Ei, i.e., ini=N. Then, the multiplicity W can be expressed as W=N!n1!n2!n3!. Hence, (S3)lnW=lnN!ilnni! By Stirling’s approximation (lnn!=nlnnn), we have (S4)lnW=(NlnNN)(inilnniini)=NlnNinilnni Let the probability of finding a particle having an energy Ei be pi=ni/N, we then have lnW=NlnNi(Npi)ln(Npi)=NlnNNipi(lnN+lnpi) (S5)=NlnNNlnNipiNipilnpi=Nipilnpi Given the entropy S=kBlnW, we have (S6)S/kBN=ipilnpi Now, the goal is to find pi that maximizes the entropy term S/kBN under the following two constraints:

  • The sum of the probabilities should be 1, i.e., ipi=1
  • The average energy of the particles should be E¯, i.e., ipiEi=E¯

To solve for pi that maximize f, we consider the following function: (S7)f(pi)=ipilnpiα(ipi)β(ipiEi) where α, β are Language multipliers. (Note that Ei’s are a bunch of constants, not variables as pi.) Then, we equate the partial derivatives of f with respect to pi to 0, namely, fpi=0ilnpiipi1piαβ(iEi)=0 (S8)i(lnpi1αβEi)=0pi=e(α+1)eβEieβEi Applying the normalization constraint ipi=1, we have (S9)e(α+1)ieβEi=1e(α+1)=1eβEi which means that e(α+1) is the normalization constant. To get the expression of β, we consider the Maxwell relation (SU)V,N=1T.

Specifically, we have (S10)dU=d(NE¯)=Nd(ipiEi)=NiEidpi dS=kBNd(ipilnpi)=kBNi(lnpi+1)dpi (S11)dS=kBNi(αβEi)dpi Given that ipi=1idpi=0, dS can be simplied as follows: (S12)dS=αkBNidpi0+βkB(NiEidpi)=βkBdU Hence, (S13)(SU)V,N=βkB=1Tβ=1kBT Altogether, Equations S8 and S13 derive the expression of the Boltzmann distribution, i.e., Equation 1.

Ensemble

In statistical mechanics, an ensemble is a collection of hypothetical copies of a system that all share the same macroscopic properties (such as temperature, pressure, volume, and chemical potential) but may differ in their microscopic details (such as the positions and velocities of individual particles). Some of the most common thermodynamic ensembles include

  • NVT ensemble: The NVT ensemble is also called the canonical ensemble, where the number of particles (N), volume (V), and temperature (T) are fixed. The probability distribution and the corresponding partition function for the NVT ensemble are given by the Boltzmann distribution, namely, (5)pi=eβEiieβEi, ZNVT=ieβEi
  • NPT ensemble: The NPT ensemble is also called the isothermal-isobaric ensemble, or the Gibbs ensemble, where the number of particles (N), pressure(P), and temperature (T) are fixed. The probability distribution and partition function of the NPT ensemble is given below: (6)pi=eβEieβPViieβEieβPVi, ZNPT=ieβEieβPVi
  • NVE ensemble: The NVE ensemble is also called the microcanonical ensemble. The NVE ensemble is defined by assigning an equal probability to every microstate whose energy falls within a range centered at E, and all the other microstates are given a probability of 0. That is, the probability in the NVE ensemble can be written as pi=1/W, where W is the number of microstates within the range of energy.
  • μVT ensemble: The μVT ensemble is also called the grand canonical ensemble, where the chemical potential (μ), volume (V) and temperature (T) are fixed. The probability distribution and partition function of the μVT ensemble is given below: (7)pi=eβEieβμNiieβEieβμNi, ZμVT=ieβEieβμNi

Notably, given the copies of the system in an ensemble, we can calculate the ensemble average of an observable O(x,p) of interest, which is simply taking the average of O(x,p) considering all points (x,p) in the phase space: (8)O=O(x,p)μ(x,p)dxdp

Ergodicity

The dynamics of a system is said to be ergodic if the ensemble average of the observable O(x,p) of interest is equal to the time average of the observable over an infinitely long trajectory, namely, (9)O=O(x,p)μ(x,p)dxdp=limMt=1MO(xt,pt) In simple terms, if one monitors an obervable of interest for a single particle over a sufficiently long period of time and take the average of the observable at all time frames, the outcome would be equivalent to the average of the same observable computed over all possible configurations of the system at a single instant. In practice, given that it is impractical to identify all possible configurations at any given time frame, reaching ergodicity in molecular dynamics is essential as it allows one to estimate the ensemble average by computing the time average over a sufficiently long trajectory.

Metastable state

Metastable states are the states having a higher potential energy than the ground state but are stable enough for the system to reside for an extended (but finite) amount of time. For example, there are two metastable states in Figure 1. One has values of x2 around -1.8, while the other has values of x2 around 1.8.

Collective variable (CV)

Here, we refer to collective variables (CVs) as a d-dimensional vector ξ=(ξ1(x),ξ2(x),,ξd(x)), which can be seen as a lower-dimensional representation reduced/projected from the full 3N-dimensional configurations x via the mapping function Φ(x)=ξ, with d3N. Traditionally, CVs can be defined as any functions of the coordinates of the system, but could additionally include the alchemical variable, as shown in a recent paper proposing alchemical metadynamics4, the method we will explore in Chapter 7.

Ideally, an optimal set of CVs should be able to capture the slowest dynamics of the system and distinguish the metastable states of interest. For example, in Figure 1, x1 is a reliable CV for distinguishing the two metastable states of interest. Conversely, x2 is orthogonal to the slowest degree of freedom of the system and, therefore, is not suitable for describing transitions between the two metastable states. Identifying the optimal set of CVs is often challenging in systems where the slowest-relaxing coordinates are not intuitive. In such cases, machine learning-based algorithms like RAVE5 or SGOOP6 can be helpful.

Free energy

In the canonical (NVT) ensemble, Helmholtz free energy is defined as (10)F=1βlnZNVT=1βeβU(x)dx while in the Gibbs (NPT) ensemble, the Gibbs free energy can be expessed as (11)G=F+PV=1β[lnZNVT(lnZNVTlnV)] Practically, in a soluted system where the solvent is nearly imcompressible, i.e., VP0, the difference in Gibbs free energy is almost equal to the difference in Helmholtz free energy.

Free energy profile

A free energy profile is meaningful only if there is a collective variable ξ of interest. A free energy profile, sometimes called a free energy landscape or a free energy surface, can be computed as follows: (12)F(ξ)=1βlnP(ξ)=1βlnδ(ξΦ(x))ν(x)dx=1βδ(ξξ(x)) where ν(x)=eβU(x)/eβU(x)dx is the Boltzmann distribution and Φ is a function mapping the 3N-dimensional vector x to the d-dimensional CV vector ξ. The function δ is the multivariate Dirac delta function, i.e., (13)δ(ξΦ(x))=idδ(ξiΦi(x)), δ(ξiΦi(x))={+, if ξi=Φi(x)0, otherwise where δ is a single-variable Dirac delta function and Φi is a function mapping x to the i-th CV, ξi. While Equations 12 and 13 may seem daunting, they essentially mean that the free energy profile along the CVs can be computed from the marginal probability density P(x) along the CVs, which is obtained by integrating the Boltzmann distribution ν(x) over all variables except for the CVs Φ(x). This interpretation may be understood more easily with the example in the Supplementary note below.

Supplementary note: An example of analytical free energy profile

In this example, given the potential energy U(x1,x2)=x146x12+x1+x22 (the same functional form as the one in the free energy surface in Figure 1), we want to determine its free energy profile along the collective variable x1. First, we can calculate the Boltzmann distribution (S14)ν(x)=1Zexp(β(x146x12+x1))exp(βx22) where Z=exp(β(x146x12+x1))exp(βx22)dx1dx2 is a spearable integral. To save space, we let f(x1)=x146x12+x1, so ν(x)=1Zeβf(x1)eβx22 and Z=(eβf(x1)dx1)(eβx22dx2). Then, given that x1 is the only CV, we can compute the free energy profile F(x1) as follows: F(x1)=1βln(δ(x1Φ1(x))ν(x1,x2)dx1dx2) (S15)=1βln[(+δ(x1Φ1(x))eβf(x1)dx1)(+eβx22dx2)] where Φ1(x) takes in x (and maps it to a specific value of x1), but is a function of x1 and therefore, can be grouped into the first integral in Eequation S15. Importantly, the value x1 runs from to + and whenever it is not equal to Φ1(x), the detal function δ(x1Φ1(x)) returns 0. In fact, given f(x)δ(xy)dx=f(y) (and (+eβx22dx2)=π/β), we have (S16)F(x1)=1βln(eβf(x1)πβ)F(x1)=x146x12+x11βln(πβ) The same result can be reached by integrating the Boltzmann distribution overall all variables but not x1, i.e., (S17)F(x1)=1βln(eβU(x1,x2)dx2)=x146x12+x11βln(πβ)

Free energy estimator

A free energy estimator is an expression or algorithm that is used to estimate free energy differences or free energy surfaces from simulation data, such as configurations, energies, and forces. Other than estimating free energy by directly measuring the probability ratios or from transition matrices, common free energy estimators include weighted histogram analysis method (WHAM)7, thermodynamic integration (TI)8, Bennett acceptance ratio (BAR)9, and multistate Bennett acceptance ratio (MBAR)10. These free energy estimators will be briefly introduced in the following chapters.

Potential of mean force (PMF)

While the term potential of mean force (PMF) is frequently used interchangeably with free energy surface, the historical definition of PMF is defined between two particles 1 and 2 and represents the effective potential energy that particle 1 experiences due to the presence of particle 2, when the positions of all other particles in the system are averaged over. It is defined based on the radial distribution function g(r): (14)W(r)=1βlng(r)=F(r)+2βlnr+C where F(r) is the free energy surface along the interparticle distance and C is an arbitrary constant. That is, the historic definition of PMF is related to the free energy surface but is still a distinct quantity. Check the Supplementary note below to see how PMF relates to the radial distribution function. In this mini-course, we treat the term “potential of mean force” equivalently as the term “free energy profile”.

Supplementary note: Relationship between the potential of mean force and distribution/correlation functions 1. General distribution functions
Here we consider a canonical (NVT) ensemble characterized by N particles in a volume V (hence the average number density ρ=N/V) at temperature T. (Note that the theory also works for other thermodynamic ensembles.) Given UN=UN(r1,...,rN), the potential energy due to the interaction between particles, the configurational partition function is given by (S18)ZN=...eβUN(r1,...,rN)dr1...drN where r1,...,rN are the position vectors of the particles. With this, the probability of finding particle 1 in the volume element dr1 at point r1, etc., can be expressed as (S19)P(N)(r1,...,rN)dr1...drN=eβUNZNdr1...drN To determine the probability for the case where the positions of n<N particles are fixed in the volume elements dr1,...,drn at the point r1,...,rn, with the remaining Nn particles unconstrained, we simply integrate Equation S19 over the particles with indices n+1,...,N, namely,

(S20)P(n)(r1,...,rn)=...eβUNZNdrn+1...drN

Assuming that the particles are indistinguishable, here we define the n-particle density function: ρ(n)(dr1...drN)=N!(Nn)!P(N)(r1,...,rn) (S21)ρ(n)(dr1...drN)=N!(Nn)!...eβUNZNdrn+1...drN which represents the probability density of finding any particle at r1, any particle at r2, ..., and any particle at rn. (Note that for n=1 (one-particle density), we have ρ(1)(r)=ρ, the bulk number density. See Wikipedia or LibreTexts for the proof.)

2. General correlation functions
A general correlation function can be defined in terms of the distribution function ρ(n)(r1,...,rn) as g(n)(r1,...,rn)=ρ(n)(r1,...,rn)/ρn (S22)g(n)(r1,...,rn)=(VN)nN!(Nn)!...eβUNZNdrn+1...drN Specifically, for n=2, we have the so-called pair correlation functions. Note that g(2)(r1,r2) can be rewritten as g(r)=g(|r1r2|), so that a pair of particles (with one of them being at the origin) separated by a distance r is considered. (S23)g(2)(r1,r2)=g(r)=1ρ2N(N1)...eβUNZNdr3...drN 3. Potential of mean force (PMF)
Finally, here we relate the pair correlation function/radial distribution function g(r) to the mean force F, which is the negative derivative of the potential of mean force (PMF) W. Here, we consider the mean force that particle 1 experiences due to the presence of particle 2: F1=ddr1UN(r1,...,rN)r1,r2=...(dUNdr1)eβUNdr3...drN...eβUdr3...drN (S24)=1βddr1(...eβUNdr3...drN)...eβUNdr3...drN=1βddr1ln(...eβUNdr3...drN) From Equation S23, we have (S25)...eβUNdr3...drN=ρ2ZNN(N1)g(2)(r1,r2) Hence (S26)F1=1βddr1[ln(ρ2ZNN(N1))+ln(g(2)(r1,r2))]=1βdr1lng(r) Finally, we have (S27)W(r)=F1dr1=(1βddr1lng(r))dr1W(r)=1βlng(r) (Check this handout to see how the radial distribution function g(r) can be used to calculate other thermodynamic quantities such as the energy or pressure.)

3. Exercises

👨‍💻 Exercise 1

Given the functional form of the free energy surface in Figure 1 (f(x1,x2)=x146x12+x1+x22), estimate the timescale of transitioning from the left to right metastable states at 300 K. (Hint: One can use the transition state theory (TST) to estimate the rate constant of the process.)

Click here for the solution.

First, we need to figure out the height of the free energy barrier going from left to right metastable states. This requires determining the local minima and maxima of the free energy surface by equating the partial derivative of f(x1,x2) with respect to x1: fx1=04x1312x1+1=0 As a result, at x11.7723, f(x1) has the global minimum of 10.7524 kBT, while a local maximum of 0.0417 kBT occurs at x10.0835, so the barrier height is 10.7941 kBT. Given the free energy barrier height (ΔG=10.7941 kBT), now we can use transition state theory (TST) to determine the rate constant as follows: (E1)k=kBThexp(ΔGkBT) Accordingly, we have (E2)k=1.38×1023 J/K×300 K6.626×1034 Jsexp(10.7524)1.34×108 s1 Therefore, the timescale of the process is t=1/k7.5×109 s=7.5 ns. That is, with a time step of 2 femtoseconds, it would require at least 3.7 million steps to simulate this process in a standard molecular dynamics simulation. Given currently available high-end computing resources, such a computational cost is not too demanding. However, for an even larger free energy barrier, say, 20 kBT, the timescale quickly goes up to 77.6 microseconds. For a barrier of 30 kBT, which is not uncommon,

👨‍💻 Exercise 2

A one-dimensional simple harmonic potential with a force constant K can be expressed as U(x)=12K(xx0)2, where x0 is the equilibrium position on the x-axis. Given this, derive the probability density distribution and Helmholtz free energy expressions for the system.

Click here for the solution.

The one-dimensional harmonic potential is given by (E3)U(x)=12K(xx0)2 Substituting Equation E3 into Equation 4, we have (E4)ν(x)=exp(12βK(xx0)2)exp(12βK(xx0)2)dx Let a=βK2(xx0)da=βK2dx, then the denominator Z of Equation E4 can be calculated as below: (E5)Z=exp(a2)dx=2βKexp(a2)da=2πβK Therefore, the probability distribution of a one-dimensional harmonic oscillator is (E6)ν(x)=βK2πexp(βK2(xx0)2) Importantly, this is a Gaussian distribution centered at x0, with the standard deviation being σ=1/βK. Then, using F=1βlnZ, we can compute the free energy as (E7)F=1βln(2πβK)=12βln(2πσ)


  1. Padua, D. Encyclopedia of Parallel Computing; Springer Science & Business Media, 2011. ↩︎

  2. Gelpi, J.; Hospital, A.; Goñi, R.; Orozco, M. Molecular Dynamics Simulations: Advances and Applications. Adv. Appl. Bioinform. Chem. 2015, 8, 37–47. ↩︎

  3. HéninJ.; LelièvreT.; Shirts, M. R.; Valsson, O.; Delemotte, L. Enhanced Sampling Methods for Molecular Dynamics Simulations. Living J. Comput. Mol. Sci. 2022, 4 (1). ↩︎

  4. Hsu, W.-T.; Piomponi, V.; Merz, P. T.; Bussi, G.; Shirts, M. R. Alchemical Metadynamics: Adding Alchemical Variables to Metadynamics to Enhance Sampling in Free Energy Calculations. Journal of Chemical Theory and Computation 2023. ↩︎

  5. Ribeiro, J. M. L.; Bravo, P.; Wang, Y.; Tiwary, P. Reweighted Autoencoded Variational Bayes for Enhanced Sampling (RAVE). The Journal of Chemical Physics 2018, 149 (7), 072301. ↩︎

  6. Tiwary, P.; Berne, B. J. Spectral Gap Optimization of Order Parameters for Sampling Complex Molecular Systems. Proceedings of the National Academy of Sciences 2016, 113 (11), 2839–2844. ↩︎

  7. Kumar, S.; Rosenberg, J. M.; Bouzida, D.; Swendsen, R. H.; Kollman, P. A. The Weighted Histogram Analysis Method for Free-Energy Calculations on Biomolecules. I. The Method. J. Comput. Chem. 1992, 13 (8), 1011–1021. ↩︎

  8. Kirkwood, J. G. Statistical Mechanics of Fluid Mixtures. J. Chem. Phys. 1935, 3 (5), 300–313. ↩︎

  9. Bennett, C. H. Efficient Estimation of Free Energy Differences from Monte Carlo Data. J. Comput. Phys. 1976, 22 (2), 245–268. ↩︎

  10. Shirts, M. R.; Chodera, J. D. Statistically Optimal Analysis of Samples from Multiple Equilibrium States. J. Chem. Phys. 2008, 129 (12), 124105. ↩︎

Next