1. Introduction
As out-of-equilibrium phases, glasses exhibit a complex disordered atomistic structure that can virtually accommodate the entire periodic table [Mauro 2018; Musgraves et al. 2019; Zanotto and Coutinho 2004]. This structural complexity has historically limited our ability to unveil structure–property relationships in glasses [Bapst et al. 2020; Tanaka et al. 2019, 2010], but, in turn, offers a vast design space to discover new glasses with unusual properties [Liu et al. 2019d; Onbaşlı et al. 2018; Ravinder et al. 2020]. Although different experimental protocols have been developed to investigate the nature of the linkages between glass composition, structure, and properties [Almeida and Santos 2015; Kamitsos 2015; Salmon and Zeidler 2015; Youngman 2018], they yield a series of structural fingerprints that are experimentally accessible—e.g., pair distribution function (PDF) computed by diffraction experiments or coordination numbers accessed by nuclear magnetic resonance (NMR) [Fischer et al. 2005; Kroeker 2015; Wright 1988; Youngman 2018]—rather than offering a direct access to the three-dimensional atomic structure itself [Affatigato 2015; Greaves and Sen 2007; Huang et al. 2013; Yang et al. 2021; Zhou et al. 2021]. In that regard, atomistic simulations have become a routine tool to easily access the atomistic structure of glasses [Bauchy 2019; Du 2019; Pedone 2009]—which is otherwise invisible from conventional experiments [Pandey et al. 2016b, a]—and to decipher the physical nature of the linkages between glass structure and properties [Binder and Kob 2011; Massobrio 2015; Onbaşlı and Mauro 2020]. Atomistic simulations of glasses can be broadly divided into two different families, namely, knowledge-based and data-based.
1.1. Physical-knowledge-based simulations
As a primary goal, atomistic simulations aim to reveal the atomistic structure of glasses [Mauro et al. 2016; Takada 2021]. This can be achieved by explicitly simulating the time-dependent formation process of glasses—e.g., melt-quenching [Debenedetti and Stillinger 2001], sol–gel transition [Du et al. 2018], vapor deposition [Wang et al. 2020], irradiation [Krishnan et al. 2017a, b], or annealing [Grigoriev et al. 2016]. Such simulations rely on our knowledge of the physics governing the interaction between atoms, i.e., the interatomic forcefields (see Figure 1) [Massobrio 2015; Mauro et al. 2016; Onbaşlı and Mauro 2020]. In detail, starting from the sole knowledge of first-principles electron interactions [Hohenberg and Kohn 1964; Kohn and Sham 1965], one can compute the interatomic interactions acting in a glass system [Boero et al. 2015; Hafner 2008]. In turn, this interatomic forcefield drives the motions of the atoms as per Newton’s law of motion—which is the basic principle behind molecular dynamics (MD) simulations [Alder and Wainwright 1959; Takada 2021]. In other words, MD simulations predict the spontaneous motions of the atoms in glasses (or liquids) [Durandurdu and Drabold 2002; Micoulaut et al. 2013]. However, since MD simulations are generally limited to short timescales (typically up to a few nanoseconds) [Lane 2015], it is intrinsically unable to match experimental timescale (up to days) [Li et al. 2017]. In particular, as the most common method to prepare a glass [Mauro and Zanotto 2014], the melt-quenching process consists in melting a liquid that is then cooling into a glass. However, the cooling rates that are accessible to MD simulations (1014–109 K/s) are orders of magnitude larger than those experienced in conventional experiments (102–100 K/s) [Li et al. 2017; Vollmayr et al. 1996a, b]. Note that, although the short timescale accessible to MD simulations challenges the comparison between glasses prepared by melt-quenching experiments and MD simulations, it should be pointed out that direct comparisons are possible when the simulated system is at equilibrium (liquid) or metastable equilibrium (supercooled liquid) since the structure and properties of such systems do not depend on their thermal history (see Figure 1) [Le Losq et al. 2017].
To overcome the short timescale accessible to MD simulations [Li et al. 2017], various simulation techniques have been proposed to accelerate the exploration of a glass’s potential energy landscape (PEL)—which represents the topography of the potential energy of a system as a function of its atom positions [Debenedetti and Stillinger 2001; Lacks 2001; Lacks and Osborne 2004]. Although MD simulates the real, spontaneous evolution of the system within its PEL, its limited timescale typically prevents large energy barriers to be overcome [Yu et al. 2017b, 2015]. In contrast, accelerated simulation techniques tend to push systems across energy barriers to accelerate their dynamics, but such accelerated dynamics may not always match the spontaneous dynamics of the atoms [Bauchy et al. 2017; Fullerton and Berthier 2020; Liu et al. 2019a]. As one of the most simple sampling technique, energy-based Monte Carlo (MC) simulations aim to explore the PEL of a glass by performing a series of random “moves” (e.g., by displacing a randomly selected atom) [Allen and Tildesley 2017; Utz et al. 2000] so as to discover lower-energy states in the PEL [Arceri et al. 2020; Vollmayr-Lee et al. 2013; Welch et al. 2013]. However, since MC moves do not necessarily reproduce the spontaneous dynamics of a glass as it relaxes toward lower-energy states, it is not guaranteed that the simulated glass matches that formed by experiments [Berthier and Ediger 2020].
1.2. Experimental-data-based simulations
To overcome the limitations of simulations based on physical knowledge [Berthier and Ediger 2020; Li et al. 2017; Wright 2020], simulations based on experimental data—e.g., reverse Monte Carlo (RMC) simulations [Biswas et al. 2004; Keen and McGreevy 1990] or empirical potential structure refinement (EPSR) simulations [Nienhuis et al. 2021; Soper 2005; Weigel et al. 2008]—have been proposed to invert experimental data (i.e., structural signatures of the real underlying atomic structure) into a three-dimensional atomic structure. In detail, RMC simulations adopt an MC search algorithm [Allen and Tildesley 2017] wherein a series of MC moves (e.g., random displacement of atoms) is performed so as to eventually obtain a glass structure that satisfies the structural constraints provided by experiments [McGreevy 2001; McGreevy and Pusztai 1988]. In the same spirit, EPSR simulations reduce the difference between simulation and experimental data by iteratively tuning the coefficients of a predefined empirical potential until the structure simulated by MD matches experiments. As compared to RMC, EPSR simulations tend to yield more realistic structures (which satisfy some basic stability constraints imposed by the empirical potential). However, the accuracy of EPSR simulations can be limited by the choice of the analytical form that is adopted for the empirical potential [Nienhuis et al. 2021; Soper 2005; Weigel et al. 2008].
Figure 1 shows a schematic illustrating the interactions between experiments and simulations. On the one hand, experiments can extract various signatures (e.g., PDF computed by diffraction experiments [Wright 1988]) of the atomic structure of glasses [Affatigato 2015], which can be used to validate knowledge-based simulations. On the other hand, RMC simulations can utilize these structural signatures as constraints to “inversely” construct an atomistic structure that satisfies available experimental data [McGreevy 2001]. It is notable that simulations based on experimental data effectively bypass any explicit glass formation process (e.g., melt-quenching) and, hence, are not affected by the limited timescale accessible to atomistic simulations. However, such simulations cannot offer any physical insights in the dynamics of the atoms in a glass [Bottaro and Lindorff-Larsen 2018; Pandey et al. 2016b; Yu et al. 2015].
In the following, we briefly review the state of the art in atomistic simulations of glasses in Section 2. The present challenges facing these simulations are discussed in Section 3. To address these challenges, we then highlight several new opportunities to advance the atomistic modeling of glasses in Section 4. Finally, we offer some concluding remarks in Section 5.
2. Overview of the state of the art in atomistic simulations of glasses
2.1. Basic principles of atomistic simulations of glasses
2.1.1. Newton’s law of motion
As an alternative classification, simulations can be broadly divided in terms of their description of the atomic motion. Namely, (i) MD simulations offer a direct description of the spontaneous dynamics of the atoms as per Newton’s law of motion, whereas (ii) other types of simulations (e.g., MC or RMC simulations) simply aim to construct an atomic structure based on a target objective (e.g., minimizing energy and maximizing agreement with experiments) without any explicit description of the dynamics of the atoms [Allen and Tildesley 2017; Takada 2021]. MD simulations predict the motion of the atoms that is predicted by Newton’s law of motion (see Figure 2a) [Alder and Wainwright 1959; Pedone 2009]. This requires the knowledge of the interatomic forcefield, that is, the force experienced by each atom. Such forces can comprise radial 2-body interactions, angular 3-body interactions, and/or many-body interactions [Daw and Baskes 1983; Stillinger and Weber 1985] and play a key role in predicting atom trajectories [Du 2015; Huang and Kieffer 2015]. In practice, the interatomic forcefield can be accurately computed using first-principles electron-level methods (e.g., ab initio MD simulation [Boero et al. 2015; Hafner 2008]) or can be approximately estimated by some empirical functions (i.e., classical MD simulation [Carré et al. 2016, 2008]). More technical details are provided in Sections 2.3 to 2.6.
2.1.2. Minimum search in a cost function landscape
Unlike MD simulations [Massobrio 2015], RMC [McGreevy 2001], energy-based MC [Allen and Tildesley 2017; Liu et al. 2019f; Utz et al. 2000], or energy minimization based on gradient descent [Bitzek et al. 2006; Shewchuk 1994] do not follow Newton’s law of motion. Rather, these approaches rely on exploring and finding the minimum position of a given “cost landscape” (e.g., potential energy for MC and energy minimization, or a loss difference between simulated and experimental fingerprints for RMC) via a series of structural modifications (e.g., displacing atoms) [McGreevy 2001; Takada 2021]. For instance, Figure 2b illustrates the principle behind energy-based MC simulations [Allen and Tildesley 2017], wherein the MC simulation searches the global minimum within the PEL by performing a series of tentative MC moves. Each move is either accepted or rejected according to a given acceptance probability defined in the MC algorithm (see Section 2.2) [Liu et al. 2019f; McGreevy and Pusztai 1988]. In contrast to MD simulations wherein large energy barriers are unlikely to be overcome [Debenedetti and Stillinger 2001; Liu et al. 2019a], simulations based on sampling a PEL can “accelerate” atomic motion to (i) jump over energy barriers and (ii) move toward the minimum energy state [Utz et al. 2000; Yu et al. 2015], which corresponds to the most stable energy state that the glass relaxes toward upon aging/relaxation [Welch et al. 2013]. Note that the landscape to explore (i.e., the function to minimize) is not always the potential energy [Debenedetti and Stillinger 2001; Tang et al. 2021], but, rather, can take the form of any function, e.g., a loss function capturing the structural difference between simulated and experimental glass in the case of RMC simulations [McGreevy 2001]. More technical details on RMC are provided in Section 2.2.
2.2. Reverse Monte Carlo (RMC) simulations
2.2.1. Cost function
RMC simulations rely on the MC search algorithm to generate an atomistic structure satisfying experimental structural constraints (i.e., measured structural fingerprints) [Keen and McGreevy 1990; McGreevy 2001; McGreevy and Pusztai 1988]. In detail, the simulation searches for the global minimum in a cost function landscape as a function of the system’s atom positions (see upper panel in Figure 3a) by performing a series of random atomic displacements (“MC move”, see lower panel in Figure 3a) [McGreevy 2001]. Here, the cost function R𝜒 is defined as the magnitude of difference between a simulation result and its experimental reference—often, the simulated and experimental PDFs g(r), formulated as follows [Wright 1993; Zhou et al. 2020]:
(1) |
2.2.2. Acceptance rate of RMC search
In analogy to the conventional energy-based Metropolis MC algorithm [Allen and Tildesley 2017], the acceptance probability P of each MC move can be expressed as [Zhou et al. 2020]:
(2) |
2.2.3. Structural match to experimental data
RMC simulations generally yield atomistic structures that exhibit an excellent agreement with the experimental data that are used to define the cost function (e.g., PDF) [Keen and McGreevy 1990], which is not surprising since RMC solely aims to minimize the difference between simulated and experimental structural data. Figure 3b shows the evolution of the cost function R𝜒 as a function of the number of MC search steps during the RMC simulation of a sodium silicate glass ((Na2O)30(SiO2)70) [Zhou et al. 2020], wherein the cost function is defined based on the neutron PDF. As expected, R𝜒 gradually decreases and, finally, the simulation generates an atomistic structure that exhibits an excellent match with experimental data. Further, Figure 3c shows the PDF g(r) computed both at the start and end of the RMC simulation. As expected, the final configuration exhibits an excellent agreement with the experimental reference, i.e., the measured g(r) [McGreevy 2001; Pandey et al. 2015; Zhou et al. 2020].
RMC simulations have successfully been used to reveal the structure of a variety of glasses or amorphous solids, especially in the case of systems that are too complex to be simulated by MD or for which no reliable empirical forcefield is available. Example of simulated systems include nuclear waste glasses [Bouty et al. 2014], metal-organic frameworks (MOFs) [Beake et al. 2013; Gaillac et al. 2017], calcium carbonate [Fernandez-Martinez et al. 2013; Goodwin et al. 2010], etc. Overall, RMC simulations offer a powerful tool to model glasses when accurate experimental data are available [Playford et al. 2014].
2.3. Molecular dynamics (MD) simulations
MD simulations rely on interatomic forcefields to predict the trajectory of the atoms according to Newton’s law of motion [Alder and Wainwright 1960, 1959; Rahman and Stillinger 1971; Stillinger and Rahman 1974]. Despite their short timescale [Li et al. 2017], MD simulations reproduce the formation process of glasses (e.g., melt quenching process under high cooling rate) to form the atomistic structure of a glass [Bauchy 2014; Micoulaut et al. 2013]. In this section, we provide some general guidelines to implement MD simulations of glasses.
2.3.1. Numerical algorithm
The algorithm of MD simulations consists of a loop of four successive steps (see Figure 4a) [Allen and Tildesley 2017; Takada 2021], namely, (i) computing the system’s potential energy U({ri}) by summing up all interatomic interactions for the current atom positions {ri}, (ii) calculating the resultant force {Fi} experienced by each atom i via energy differentiation (i.e., Fi = −∂U∕∂ri), (iii) obtaining each atom’s acceleration {ai} from {Fi} as per Newton’s law of motion, that is, ai = Fi∕mi, where mi is the mass of atom i, and finally, (iv) updating the atom positions and velocities after a small, fixed timestep via numerical integration (e.g., Verlet or leapfrog algorithm) [Allen and Tildesley 2017]. Eventually, this four-step loop yields the position of the atom as a function of time, that is, the trajectory of each atom.
It should be pointed out that the accuracy of an MD simulation depends on (i) the realistic nature of the initial configuration (i.e., initial positions and velocities), (ii) the value of the integration timestep, and, importantly, (iii) the accuracy of the interatomic potential energy. First, in the case of the glasses, the initial configuration is usually constructed by randomly placing some atoms in a cubic box while ensuring the absence of any unrealistic overlap (see below). The configuration is then relaxed at elevated temperature until the system loses the memory of its initial configuration. At this point, the outcome of the simulation does not depend any longer on how realistic the initial configuration was. Second, the integration timestep needs to be small enough (∼1 fs) to ensure the accuracy of the numerical integration. In practice, one can check that the timestep is small enough by ensuring the conservation of the system’s total energy and linear/angular momentum during a simulated dynamics in the microcanonical (NVE) ensemble [Grubmüller et al. 1991; Levesque and Verlet 1993; Omelyan et al. 2002]). Third, if the integration timestep is small enough, the accuracy of the MD simulation primarily depends on that of the underlying potential energy (see Figure 4a) [Du 2015; Huang and Kieffer 2015].
Relying on the most fundamental quantum mechanics describing electron interactions (i.e., first-principles calculation of electron interactions) [Hohenberg and Kohn 1964; Kohn and Sham 1965], the interatomic interactions can be accurately computed by conducting first-principles MD simulation (i.e., ab initio MD simulation [Boero et al. 2015], see Section 2.4). Such simulations are computationally expensive but can be used to validate or inform classical MD simulations relying on simplified, empirical potential energy functionals (see Section 2.5) [Carré et al. 2008].
2.3.2. Thermodynamical ensembles
A direct numerical solution of Newton’s law of motion yields the microcanonical NVE dynamics of the N atoms, that is, wherein the total energy and volume of the system are conserved. However, the NVE ensemble does not always mimic experimental conditions. In practice, a system can either be isolated from the rest of the universe (i.e., presenting a constant energy) or be in contact with a thermostat (wherein the thermostat can provide or receive energy so as to fix the temperature T of the system). The volume V of the system can also either be fixed or variable based on the pressure P imposed by a barostat. As such, the most common thermodynamic ensembles used in MD simulations are the microcanonical (NVE), canonical (NVT), and isothermal–isobaric (NPT) ensembles [Du 2019; Leach 2001]. Simulating the motion of the atoms in a non-NVE ensemble (e.g., NVT or NPT) requires slight modification to the constitutive equations used to calculate the acceleration of each atom [Allen and Tildesley 2017; Du 2019]. On the one hand, simulating a constant-temperature dynamics requires the use of a thermostat-based MD algorithm (e.g., Nosé–Hoover thermostat [Hoover 1985; Nosé 1984,?]) to adjust on-the-fly the atomic velocities so as to achieve a target temperature [Andersen 1980; Martyna et al. 1996]. On the other hand, simulating a constant-pressure dynamics involves the use of a barostat-based MD algorithm [Allen and Tildesley 2017; Leach 2001] (e.g., Nose–Hoover barostat [Martyna et al. 1992]) to adjust on-the-fly the system’s volume so as to achieve a target pressure (or state of stress) [Martyna et al. 1996; Tuckerman et al. 2006]. More details can be found in Allen and Tildesley [2017], Leach [2001].
2.3.3. Melt-quenching simulations
The most common method used in MD to prepare a glass consists in mimicking the experimental melt-quenching process [Soules 1990]. The melt-quenching process is detailed in the following and illustrated by Figure 4b for the case of a silica glass [Carré et al. 2008; Liu et al. 2019e].
First, an initial configuration must be created. To this end, one option is to start with a crystalline configuration (e.g., α-quartz in the case of a silica glass). However, this is often not the preferred option as (i) there might not exist a crystal with the same composition as that of the target glass, (ii) the crystal might have a very different density from that of the glass, (iii) the shape of the unit cell of the crystal (e.g., triclinic unit cell) might result in an unnecessarily complicated simulation box geometry, and (iv) melting the initial crystalline structure might require high temperature and/or long simulation time. As an alternative—often preferred—option, the initial configuration can be prepared by randomly placing atoms or molecules (e.g., SiO2 molecules in the case of a silica glass) into a cubic simulation box with periodic boundary conditions (PBC) [Leach 2001]. The box size can either be fixed so as to match the experimental density of the glass (if when the quenching is performed in the NVT ensemble) or to an arbitrary value (if the quenching is performed in the NPT ensemble). Note that, when creating the initial configuration, care must be taken to avoid any unrealistic atom overlap that might cause spurious fast atom motions at the beginning of the simulation [Martínez et al. 2009]. This can be achieved by imposing a cutoff threshold when randomly placing the atoms, e.g., by ensuring that the distance between a pair of atoms is never lower than the sum of their radii or that a pair of cations (or anions) are never close to each other.
Second, the initial configuration must be melted at high temperature so as to fully lose the memory of the initial configuration. The choice of the melting temperature depends on the type of glass that is considered. On the one hand, the temperature must be high enough to ensure that the atoms move fast enough to “reset” the initial configuration or fully melt the initial crystal. On the other hand, a temperature that is too high might induce some spurious instabilities (e.g., due to the “Buckingham catastrophe” [Carré et al. 2008] or if the system approaches the vaporization point) and increase the length of the simulation (since, at fixed cooling rate, a higher initial temperature increases the duration of cooling). The fact that the initial melting is long enough to ensure that the system has lost the memory of its initial configuration can be checked by computing the intermediate scattering function (ISF) or, as an approximate rule of thumb, by ensuring that the atoms have, on average, diffused by at least half of the simulation box.
Third, the melt is quenched to the glassy state by decreasing temperature. This is typically achieved by linearly decreasing temperature over time, while other more complicated thermal routes can be considered (e.g., nonconstant cooling rate or step-by-step cooling). The cooling process can either be performed in the NVT or NPT ensemble. The NPT ensemble is usually desirable as it mimics the experimental synthesis of glass (i.e., wherein the melt is quenched under atmospheric pressure). In turn, the NVT ensemble can yield some spurious effects as, since density is fixed, the pressure changes as a function of temperature. As a result, when using the NVT ensemble during cooling, the glass tends to “freeze” under positive (compressive) pressure at the glass transition temperature, which can impact the resulting structure. However, some interatomic forcefields are unable to yield the correct final glass density, so that using the NVT ensemble and fixing the volume to achieve the experimental glass density might, in some cases, be necessary. Due to the high computing cost of MD, the cooling rate (typically 1014–109 K/s) is much faster than those achieved in conventional experiments (102–100 K/s) [Li et al. 2017]. From the viewpoint of PEL [Goldstein 1969; Wilkinson and Mauro 2021], the system starts, upon cooling, from a high-energy, ergodic liquid state and gradually evolves to some nonergodic, lower-energy state, before eventually getting trapped in a local minimum corresponding to a glassy state (see Figure 4b) [Debenedetti and Stillinger 2001]. At the end of the cooling process, it is also common to perform an additional relaxation run to ensure that the system has reached a plateau in energy, and volume or pressure (in the NPT and NVT ensembles, respectively).
Fourth, once formed, the glass can be investigated or subjected to additional simulations [Du 2019; Pedone 2009]. Typical investigations include various structural analysis (e.g., coordination number) [Bauchy 2014, 2012], mechanical properties (e.g., strain–stress curve) [Liu et al. 2019b; Yang et al. 2019], thermodynamics (e.g., potential energy) [Bauchy and Micoulaut 2015], and vibrational/dynamical properties (e.g., atom diffusivity) [Bauchy et al. 2013; Bauchy and Micoulaut 2011]. More details can be found in Levchenko et al. [2020], Massobrio [2015]. Note that any statistical analysis must be averaged over a trajectory that is long enough to filter out the effect of statistical fluctuations (the magnitude thereof depending on the size of the simulated system). It is also worth noting that, upon quenching, the glass may end up in notably different regions of the energy landscape [Allen and Tildesley 2017]. As such, it is a good practice to simulate a series of independent melt-quenching simulations for each glass (typically around 3 to 6) and to average results over these simulations.
2.4. Ab initio molecular dynamics simulations
2.4.1. First-principles calculation of interatomic interactions
Ab initio MD simulations estimate the interatomic interactions based on first-principles calculation of electron interactions within the framework of density function theory (DFT) [Kohn and Sham 1965; Marx and Hutter 2009; Parr 1980]. Figure 5a illustrates the basic idea of ab initio MD, wherein a glass system’s potential energy U is computed as a function of both the electron wave functions {𝛹e} (i.e., the electron state) and the nuclei positions {rn} (i.e., the nuclei state) [Hohenberg and Kohn 1964; Kohn and Sham 1965]. The potential energy U is calculated by summing up three types of interactions in the system [Boero et al. 2015; Marx and Hutter 2009]:
(3) |
2.4.2. Choice of exchange-correlation (XC) potential
Ue−e is a key ingredient in first-principles simulation. It can be further described by three types of energy terms [Boero et al. 2015; Kohn and Sham 1965]:
(4) |
2.4.3. Numerical algorithms
Based on all these formulations of first-principles interatomic interactions [Boero et al. 2015; Marx and Hutter 2009], the system’s ground-state potential energy U({𝛹e},{rn}) can be calculated via optimizing {𝛹e} so as to minimize U for the current nuclei positions {rn} (see Figure 5a) [Hafner 2008; Pople et al. 1989]. After updating the new nuclei positions {rn} after a small timestep, the system’s potential energy U is then recomputed through optimizing {𝛹e} [Boero et al. 2015]. Two algorithms have been developed to optimize {𝛹e} and update {rn} [Niklasson et al. 2006], viz., (i) the Born–Oppenheimer (BO) approximation to disentangle the updates of {𝛹e} and {rn} [Boero et al. 2015; Niklasson et al. 2006], wherein {𝛹e} is recomputed at each timestep and offers an accurate estimation of the potential energy U to update the nuclei positions {rn}, and (ii) the Car–Parrinello (CP) method to construct a Lagrangian LCP({𝛹e},{rn}) that updates both quantities on-the-fly based on Lagrangian dynamics [Car and Parrinello 1985], where LCP includes a constructed kinetic energy term of {𝛹e} by assigning to {𝛹e} a fictitious electronic mass that dictates the update inertia of {𝛹e} [Boero et al. 2015]. In practice, systems with slow change of nuclei positions tend to require a large fictitious mass to delay the update of {𝛹e} as well as a large timestep to accelerate the update of {rn} [Car and Parrinello 1985]. In other words, when running a CP-MD simulation, one needs to select a proper set of timestep and fictitious mass large enough to (i) reduce the computation cost of the update of {𝛹e} and {rn} but also small enough to (ii) numerically conserve the system’s energy and linear/angular momentum [Boero et al. 2015]. More details can be found in Boero et al. [2015], Marx and Hutter [2009].
2.4.4. Applications and limitations
Since ab initio MD simulations are typically able to accurately describe interatomic interactions [Pople et al. 1989], they are the method of choice to simulate complex glasses for which no robust empirical forcefields are available (see below). This is typically the case for glasses exhibiting complex bonds (i.e., with a mixed ionic/covalent/metallic character or varying electronic delocalization) or flexible local structures (e.g., varying coordination numbers) [Massobrio 2015]. For example, chalcogenide glasses (e.g., Ge–Se glasses) typically exhibit complex structural features that cannot be accurately reproduced by classical MD [Mauro and Varshneya 2006], including under- and over-coordinated atoms, edge-sharing structures, and homopolar bonds (see insert of Figure 5b) [Petri et al. 2000]. Figure 5b shows a melt-quenched GeSe2 glass prepared by ab initio MD simulation, wherein its computed PDF g(r) is compared with neutron diffraction data [Micoulaut et al. 2013]. Relying on CP-MD method and GGA-XC potential [Car and Parrinello 1985; Perdew et al. 1996a], the simulation offers a glass structure that is in excellent agreement with experimental data [Micoulaut et al. 2009].
However, despite their unparalleled predictive power [Pople et al. 1989], ab initio simulations still come with a series of limitations. First, although they rely on a fundamental description of electronic interactions, they still rely on a number of assumptions (e.g., choice of the pseudopotential and exchange-correlation function). Second, ab initio MD simulations come with a computational cost that is orders of magnitude larger than that of classical MD [Marx and Hutter 2009]. This limits the timescale that is accessible to ab initio MD simulations (typically to 100s of ps) [Boero et al. 2015]. This restricts their use to high cooling rates, which, in turn, tends to yield glasses that are more disordered (higher fictive temperature) than their experimental counterparts [Li et al. 2017]. Finally, the computational cost of ab initio MD simulations typically scales with the cube of the number of electronic degrees of freedom [Hafner 2008]. As a result, ab initio MD simulations are usually limited to very small systems (up to a few hundred atoms) [Massobrio 2015]. This prevents their use in capturing small compositional effects (since replacing one atom by another largely affects the overall composition in a small system) or extended structural features (e.g., large rings or compositional heterogeneity) [Du and Corrales 2006; Nakano et al. 1994].
Thanks to their accuracy, ab initio MD simulations have been applied to predict the structure and dynamics of various types of glasses or amorphous solids, including amorphous silicon [Car and Parrinello 1988], chalcogenide [Micoulaut et al. 2013], Ge–Sb–Te phase-change materials [Caravati et al. 2009], MOFs [Darby et al. 2020; Sillar et al. 2009], silicates [Baral et al. 2017; Tilocca and de Leeuw 2006], borates [Scherer et al. 2019], etc. Notably, by accurately describing the interatomic interactions, ab initio MD offers a powerful tool to simulate complex glasses that cannot be properly described by empirical forcefields.
2.5. Classical molecular dynamics simulations
2.5.1. Empirical forcefields
In contrast to the explicit description of electronic effects offered by ab initio MD [Marx and Hutter 2009], classical MD relies on empirical forcefields to describe interatomic interactions—via physics/intuition-based, computationally efficient functionals [Massobrio 2015]. For example, Figure 6a shows the PDF g(r) of a melt-quenched (Na2O)30(SiO2)70 glass prepared by classical MD simulation using an empirical forcefield [Cormack et al. 2002; Du and Cormack 2004]. The neutron diffraction data is added as a reference for comparison [Wright et al. 1991]. Notably, this empirical forcefield offers an atomistic structure that is in good agreement with experimental data. Note that the simulated and experimental glasses are prepared with different cooling rate [Li et al. 2017]. The good agreement between simulated and experimental PDFs (see Figure 6a) suggests that, although the empirical forcefield is a simplification of the real interatomic interactions [Huang and Kieffer 2015; van Beest et al. 1990], it can accurately describe them while relying on a significantly reduced computational cost as compared to first-principles calculations [Bauchy et al. 2013; Carré et al. 2008]. Thanks to its much faster execution time as compared to ab initio MD simulation, classical MD simulation can extend to much longer timescales (up to a few nanoseconds) and larger length scales (up to millions of atoms) [Lane 2015; Plimpton 1995a].
2.5.2. Forcefield functionals
The optimal functional form of the empirical interatomic potential depends on the type of glass that is simulated (e.g., ionic, covalent, metallic glass, etc.) and no “universal” empirical potential is available to date [Du 2015; Huang and Kieffer 2015].
In practice, ionocovalent glasses (e.g., silicate glasses, which feature ionocovalent bonds [Huang and Kieffer 2015]) can be well described by a combination of: (i) long-range coulombic interactions and (ii) a Buckingham-format functional description of short-range interactions to compute the potential energy U(rij) between atom i and atom j at a distance rij (see inset of Figure 6a) [Carré et al. 2008; Du and Corrales 2006; van Beest et al. 1990]:
(5) |
Note that, in the case of interactions that rapidly converge to zero upon increasing distance (e.g., Van der Waals interactions, which is proportional to 1/r6), the energy error associated with the fact of using a finite cutoff converges to zero as the cutoff increases. However, this is not the case for Coulombic interactions, which exhibit a slow decrease upon increasing distance r (i.e., proportional to 1/r). In such a case, the energy contribution that is neglected when using a cutoff (arising from long-distance interactions between atoms) does not converge to zero upon increasing cutoff. As such, long-range Coulombic interactions (including across PBC) must be explicitly considered to ensure the accuracy of the simulation. In practice, computing Coulombic interactions is often the most computationally expensive step when simulating ionocovalent glasses. In that regard, various summation methods have been developed to speed up the computation of electrostatic interactions, such as the Ewald method [Ewald 1921], particle–particle–particle–mesh (PPPM) method [Hockney and Eastwood 1988], and damped-shifted-force (DSF) method [Fennell and Gezelter 2006]. In detail, both the Ewald and PPPM methods are based on the idea that the summation of long-range interactions can be efficiently calculated in the reciprocal space using a Fourier transform [Allen and Tildesley 2017]. Unlike the Ewald or PPPM methods, the DSF method is conducted in real space and is based on the rapidly converging summation of a short-range potential—whose cutoff is large enough—to approach the effective long-range Coulombic interactions [Fennell and Gezelter 2006]. In practice, both the Ewald and PPPM summation methods offer accurate calculations of the Coulombic interactions. In many cases, electrostatic interactions can also be reasonably well approximated by the DSF method [Carré et al. 2007; Fennell and Gezelter 2006], which, in turn, is significantly more computationally efficient than Ewald and PPPM since it does not involve any Fourier transform [Fennell and Gezelter 2006].
It is worth pointing out that, as the accuracy of classical MD simulations significantly relies on the analytical formulation of the empirical forcefield, it is necessary, when developing empirical forcefield functionals, to account for all essential components of the interaction (e.g., Coulombic component of ionocovalent interaction) that affect the targeted glass properties. Indeed, although select properties are not very sensitive to the details of a forcefield (e.g., density or PDF), other properties can be strongly affected by small variations in the parameters or analytical formulation of the empirical potential. In such cases, it may be necessary to develop advanced forcefields that are able to capture some complex chemical or physical behaviors, including bond formation/breaking, charge transfers, polarization effects, short-range repulsion, etc. [Jahn et al. 2006; Jahn and Madden 2007; Serva et al. 2020]. In that regard, some advanced forcefields such as the aspherical ion model (AIM)—which is more complex than the Buckingham potential [Liu et al. 2020a]—have been developed to account for these essential, complex interaction behaviors in studying the dynamics of complex oxides [Jahn et al. 2006; Jahn and Madden 2007; Serva et al. 2020]. Note that, although such potentials are more computationally expensive, they generally show an improved level of agreement with experiments and a high transferability upon composition, temperature, and pressure changes [Jahn et al. 2006; Jahn and Madden 2007; Serva et al. 2020].
In contrast to ionocovalent glasses, covalent glasses (e.g., amorphous silicon [Stillinger and Weber 1985] or chalcogenide glasses [Micoulaut et al. 2013]) cannot be described by 2-body interatomic potentials due to the directional nature of the covalent bonds they form [Phillips 1981, 1979]. As such, describing covalent glasses requires the use of 3-body potential interactions—e.g., Stillinger–Weber (SW) potential [Ding and Andersen 1986; Stillinger and Weber 1985]—to constraint the values of the interatomic angular interactions between a central atom i and its two neighbor atoms j and k [Du 2015; Mauro and Varshneya 2006]:
(6) |
Finally, the simulation of metallic glasses typically requires the description of many-body effects [Pelletier and Qiao 2019; Sheng et al. 2006], which can be well described by the Embedded Atom Method (EAM) potential [Daw et al. 1993]. In the EAM approach, the potential energy Ui of a central atom i is formulated as [Daw and Baskes 1984, 1983]:
(7) |
Note that, to reduce the computational cost associated with the calculation of the empirical forcefields, a cutoff rc is typically adopted—wherein the interaction energy between a pair of atoms is assumed to be zero if their distance exceeds the cutoff [Chialvo and Debenedetti 1990; Du 2015]. In addition, a neighbor list algorithm is typically adopted to reduce the number of times the distance between a pair of atoms is calculated [Allen and Tildesley 2017; Leach 2001].
2.5.3. Forcefield parameterization
In addition to the choice of its functional form [Du 2015], the accuracy of an empirical forcefield strongly depends on its parameterization [Sundararaman et al. 2020, 2018; Wang et al. 2018]. Figure 6b illustrates the general idea behind the parameterization of an empirical potential. Such parameterization can be formulated as an optimization problem, wherein the forcefield parameters are optimized so as to minimize a given cost function. This problem can be illustrated in terms of a cost function landscape, which represents the topographical evolution of the cost function as a function of the forcefield parameters. The cost function captures the level of mismatch between a simulated metric and a given reference value. The simulated metric can be the interatomic energy/force, some structural features (e.g., PDF), or some other macroscopic properties (e.g., density, stiffness, etc.).
For each type of metric, reference values can be offered by experimental data or ab initio simulations. On the one hand, defining the cost function in terms of the level of mismatch between computed and experimental data tends to yield a good agreement between simulated and experimental glass properties—since the parameterization scheme specifically aims to minimize this difference. However, on the other hand, it is not always meaningful to compare a simulated glass with its experimental counterpart. Indeed, glasses simulated by MD are prepared with a cooling rate that is significantly higher than those experienced in traditional experiments and, therefore, should be different (typically more disordered) than their experimental counterparts [Carré et al. 2016; Li et al. 2017; Vollmayr et al. 1996b]. Hence, parameterizing an empirical forcefield so as to “force” a simulated glass to exhibit properties that match experimental data may yield an unrealistic forcefield. That is, forcing simulated and experimental glasses to feature similar properties can typically only be achieved by “mutual compensations of errors”, that is, wherein the forcefield is deformed so as to compensate the fact that the simulated glass is prepared with an extremely high cooling rate. As such, although such parameterization may offer an apparent agreement between simulations and experiments for the properties that are included in the cost function, the resulting forcefield, due to its nonphysical nature, may dramatically fail at predicting properties that are not included in its cost function. In contrast, parameterizing a forcefield so as to match with ab initio data is expected to yield a more realistic description of the true interatomic potential. However, glasses prepared with such a realistic forcefield should be compared to hyperquenched glasses (i.e., prepared under high cooling rate) and, hence, may not exhibit a good agreement with experimental data prepared under slower cooling rates.
In terms of the metric to be considered in the cost function, it has recently been suggested that, in the case of glasses or liquids, using a structural property as reference (e.g., PDF) tends to offer better results than directly forcing the parameterized forcefield to match reference force values (as computed by ab initio simulations) [Carré et al. 2016, 2008]. In the example shown in Figure 6c, the cost function R𝜒 is defined as the magnitude of the difference between the PDF g(r) computed by classical MD and ab initio MD for a liquid silica system [Liu et al. 2019e]. Note that parameterizing this forcefield based on liquid (rather than glassy) configurations effectively removes any spurious effects arising from the thermal history of the system.
Various optimization methods are available to “navigate” the cost function landscape, that is, to identify the optimal forcefield parameters that minimize the cost function [Carré et al. 2016], including MC search [Iype et al. 2013], Bayesian optimization [Liu et al. 2019c, e], particle swarm optimization [Christensen et al. 2021], or gradient descent search [Shewchuk 1994]. Note that this optimization problem is typically ill-defined since several degenerate sets of forcefield parameters can often minimize the cost function, that is, several competitive minimum positions can be found in the cost function landscape [Liu et al. 2020a, 2019c]. Figure 6c illustrates two competitive forcefields based on the same Buckingham-format functional [van Beest et al. 1990] (see (5)). These forcefields can be described as “soft” (featuring weak coulombic interactions) and “hard” (featuring more intense coulombic interactions) and both offer a competitive minimum value of the cost function R𝜒 for silica liquids [Liu et al. 2020a]. Nevertheless, despite the apparent harmony, the atomistic structures offered by the soft and hard potential exhibit pronounced differences for structural features that are not included in the cost function (e.g., bond angular distribution or ring size distribution) [Liu et al. 2020a]. This exemplifies the need to a posteriori validate forcefields that validate the accuracy of forcefields after their parameterization [Liu et al. 2019c].
Finally, it is worth pointing out that both the choice of forcefield functionals and forcefield parameterization can affect the forcefield’s transferability as a function of, for instance, glass composition, pressure, or temperature [Hernandez et al. 2019; Jahn and Madden 2007]. Indeed, forcefields are often developed for a specific glass composition and range temperature/pressure (e.g., liquid silica). In general, there is no guarantee that such forcefields can offer realistic predictions far from the conditions used during their training [Sundararaman et al. 2020, 2018]. For instance, certain chemical behaviors—e.g., charge transfers—may not be similar under various conditions of temperature or pressure and, hence, may require some adjustments in the parameterization of empirical forces to be properly accounted for [Liu et al. 2020a]. The transferability of forcefields is also likely to be poor when the forcefield is used under conditions wherein the atoms exhibit some local environments (e.g., varying coordination numbers) that are different from those observed during the training of the forcefield. This limitation is especially pronounced for borate glasses (wherein the average coordination number of boron depends on temperature) and pressurized glasses (wherein coordination numbers tend to increase under pressure) [Salmon and Zeidler 2015]. In order to overcome these limitations and develop transferable forcefields, one first needs to propose a forcefield functional that is complex enough to capture the nature of interatomic interactions in all the conditions of interest (e.g., Buckingham-format potential for ionocovalent bond interactions in silicate glasses), but generic enough to avoid any unrealistic extrapolations due to “overfitting” [Du 2015; Huang and Kieffer 2015]. Then, the transferability of a forcefield greatly depends on its forcefield parameterization scheme, where the cost function should be designed so as to consider the structure and properties of glasses under different conditions [Liu et al. 2019c; Sundararaman et al. 2018]. Note that there always exists a balance between (i) the ability of a forcefield to accurately predict the unique, detailed properties of a specific system and (ii) the capability of a forcefield to offer a robust transferability for a wide range of compositions and conditions [Liu et al. 2019c]. To address the issue of transferability, several recent works have focussed on developing generic forcefields that can be applied to a large compositional envelope at the expense of potentially being unable to capture the fine structural details of a specific composition [Hernandez et al. 2019; Jahn and Madden 2007], including the well-established Teter potential for modified silicate glasses [Du and Corrales 2006], the Wang–Bauchy potential for borosilicate glasses [Wang et al. 2018], etc.
Thanks to their high computational efficiency as compared to ab initio simulations [Carré et al. 2008; Huang and Kieffer 2015], classical MD simulations relying on empirical forcefields have been extensively used to model glasses [Cormier et al. 2003; Mead and Mountjoy 2006; Tanguy et al. 1998]. Simulated systems include glassy silica [Carré et al. 2008; Liu et al. 2020a; van Beest et al. 1990], modified (e.g., alkali) silicate glasses [Cormack and Du 2001], aluminosilicate glasses [Bouhadja et al. 2013], borates [Kieu et al. 2011; Sundararaman et al. 2020; Wang et al. 2018], etc. It should be pointed out that, as the number of elements in the glass increases, it becomes extremely challenging to properly describe all the distinct interactions between pair of elements (since the number of parameters that need to be adjusted scales with the square of the number of elements) [Du 2015].
2.6. Reactive molecular dynamics simulations
2.6.1. Gap between ab initio and classical MD simulations
In contrast to classical MD, ab initio MD can accurately describe a chemical reaction process but is limited to small system size (up to hundreds of atoms) [Boero et al. 2015], while, in turn, classical MD can extend to large systems (up to millions of atoms) but lack an accurate description of chemical reaction processes [Plimpton 1995a; Senftle et al. 2016]. Although classical MD simulations exhibit high computation efficiency [Marx and Hutter 2009; Massobrio 2015], they rely on simplified empirical forcefields, which may accurately predict certain properties (e.g., PDF g(r)) of glasses but may not offer robust predictions for more complex properties that are more sensitive to the details of the forcefields [Ganster et al. 2007; Ispas et al. 2002]. In addition, classical MD simulations relying on static forcefields and fixed charges are often unable to robustly account for charge transfer mechanisms and defective coordinated states, which prevents such simulations from properly describing the breaking and formation of bonds during chemical reactions [Buehler et al. 2006; Deng et al. 2021; Du et al. 2019a; Yu et al. 2018]. To address this limitation, reactive MD simulations—i.e., MD relying on reactive forcefields (e.g., ReaxFF) [van Duin et al. 2003, 2001]—have been proposed as an intermediate option, in between classical and ab initio MD. The promise of reactive MD is to offer an accuracy that approaches that of ab initio MD while involving a computational burden that is more comparable to that of classical MD. This makes it possible for reactive MD to simulate fairly large systems (up to tens of thousands of atoms) over extended time scales (up to a few nanoseconds) with an enhanced accuracy as compared to classical MD (see Figure 7a) [Senftle et al. 2016]. However, reactive forcefields typically rely on hundreds-to-thousands of parameters and, hence, are extremely challenging to parameterize. This has thus far limited their applications to a small number of glass families [Senftle et al. 2016].
2.6.2. ReaxFF forcefield
The ReaxFF forcefield is one of the most popular implementations of reactive MD [Leven et al. 2021]. As detailed in the following, the key advantages of ReaxFF are that it (i) explicitly describes the dynamical formations and breaking of bonds, (ii) accounts for charge transfers between atoms, and (iii) dynamically adjusts the interatomic interactions as a function of the local environment of each atom [van Duin et al. 2003, 2001]. This makes it possible to describe phase transitions, defect formations, or chemical reactions between atoms (e.g., to simulate the reactivity of a glass surface with water) [Buehler et al. 2006; Du et al. 2018].
In detail, the ReaxFF forcefield calculates the system’s potential energy Usys by summing up the following energy terms [van Duin et al. 2003]:
(8) |
2.6.3. Glass reactivity
Reactive MD simulations offer an ideal tool to investigate the chemical reactivity of glasses [Senftle et al. 2016]. For instance, Figure 7b illustrates the sol–gel formation process of glassy silica (in terms of the condensation of Si(OH)4 precursors) from the viewpoint of an energy state transition in PEL [Du et al. 2018; Steinfeld et al. 1999]. Classical MD relying on empirical forcefield (e.g., Buckingham potential [van Beest et al. 1990]) generally offers an accurate description of near-equilibrium properties (i.e., around local minimum energy states) but typically fails at properly describing far-from-equilibrium behaviors like those experienced in transition states [Liu et al. 2019e]. In particular, classical MD typically does not offer realistic energy barrier predictions [Du et al. 2018]. In contrast, reactive MD (e.g., based on ReaxFF) can offer more accurate description of transition energy barriers, which makes it possible to simulate sol-to-gel transitions [Du et al. 2019b, 2018; Zhao et al. 2021, 2020]. Reactive potentials also make it possible to model the reactivity of glass with aqueous solutions [Deng et al. 2019; Du et al. 2019a; Fogarty et al. 2010; Mahadevan and Du 2020; Yu et al. 2018]. The description of bond formation/breaking also make reactive MD simulations an ideal tool to study fracture processes [Bauchy et al. 2016, 2015; To et al. 2020] or vapor deposition processes [Wang et al. 2020].
2.6.4. Glass structure
The fact that ReaxFF can dynamically adjust interatomic interactions as a function of the local environment of each atom is also a key advantage in glass simulations—since glasses can exhibit a large variety of local structures (e.g., varying coordination states). As such, when properly parameterized, reactive forcefields have the potential to offer an improved description of the atomic structure of glasses as compared to classical MD [Yu et al. 2017a, 2016]. Figure 7c shows a comparison between the neutron structure factor SN(Q) computed by classical and reactive MD simulation for a (Na2O)30(SiO2)70 glass [Yu et al. 2017a]. Taking the experimental SN(Q) as a reference [Grimley et al. 1990], reactive MD relying on the ReaxFF forcefield offers an improved description of the medium-range structure (e.g., improved left-skewed symmetry of SN(Q) at low-Q region [Yu et al. 2017a]) as compared to that offered by classical MD relying on a Buckingham forcefield [Du and Cormack 2004]. The ability of ReaxFF to properly handle coordination defects also makes it an ideal tool to study irradiation-induced vitrification [Krishnan et al. 2017a, c; Wang et al. 2017].
Although its application to glassy systems has thus far remained fairly limited, the ReaxFF forcefield has been used to model several noncrystalline systems, including amorphous silicon [Buehler et al. 2006], glassy silica [Yu et al. 2016], sodium silicate glasses [Deng et al. 2020], modified aluminosilicate glasses [Dongol et al. 2018; Liu et al. 2020c; Mahadevan and Du 2021], organosilicate glasses [Rimsza et al. 2016], and zeolitic imidazolate frameworks (ZIFs) [Yang et al. 2018], etc. Despite the unique advantages offered by ReaxFF, its applications are presently limited by the range of elements that have been parameterized [Leven et al. 2021; Senftle et al. 2016].
3. Grand challenges in atomistic simulations of glasses
3.1. Parameterization of empirical forcefields
3.1.1. High dimensionality
Empirical forcefields involve many parameters that need to be optimized [Liu et al. 2019c]. For example, ReaxFF forcefields generally comprise hundreds (or thousands) of parameters [van Duin et al. 2003]. Although classical forcefields are usually simpler, they still involve dozens of parameters—a number that increases upon increasing number of elements [Huang and Kieffer 2015]. Such high dimensionality makes it challenging to identify optimal values for the parameters, that is, to reliably find the minimum position in the cost function landscape [Carré et al. 2016; Shewchuk 1994].
3.1.2. Roughness of cost function landscape
Figure 8a illustrates a cost function landscape as a function of arbitrary forcefield parameters, where the cost function R𝜒 is defined herein as the magnitude of difference between the PDF g(r) computed by classical MD and its ab initio reference [Liu et al. 2019e]. Since the landscape is rough and features many local minima, traditional gradient-descent-based optimization algorithms are usually largely inefficient since they tend to get stuck in local minima—so that the outcome of the optimization strongly depends on the initial starting point [Shewchuk 1994].
As an illustrative example, Figure 8b shows the cost function landscape of a Buckingham forcefield for glassy silica as a function of two forcefield parameters (i.e., the silicon partial charge qsi and the short-range Si–O interaction intensity ASiO) [Liu et al. 2019e]. Even for this simple system, the cost function landscape is rough and exhibits several local minima. As a result, optimizations relying on conjugated gradient descent [Shewchuk 1994] tend to get easily stuck at the positions nearby the start point (see Figure 8b) [Liu et al. 2019e]. Indeed, although the cost function looks fairly smooth at a high level, a closer inspection reveals that the cost function locally features a very rough landscape showing a large number of local minima (see bottom panel in Figure 8b) [Liu et al. 2019e]. Overall, as a consequence of the rough nature of typical cost function landscapes, forcefield parameterization schemes are often strongly biased, heavily relying on intuition, or requiring a large number of independent optimization (with different starting points) [Liu et al. 2019e, c].
3.2. Effect of the cooling rate
3.2.1. Importance of thermal history
Due to the short timescale that is accessible to MD simulations (up to a few nanoseconds trajectories) [Plimpton 1995a], glass simulations relying on the melt-quenching approach are limited to ultrafast cooling rates (1014–109 K/s), which far exceed the cooling rates achieved in conventional experiments (102–100 K/s) [Li et al. 2017]. This is important since, glasses being out-of-equilibrium phases, their structure and properties depend on their thermal history. Specifically, glasses tend to reach deeper positions in their PEL upon decreasing cooling rates (i.e., they become more stable, see Figure 9a) [Debenedetti and Stillinger 2001].
Figure 9b illustrates the evolution of the potential energy U of a glass with respect to the decreasing temperature during melt-quenching with varying cooling rates: (i) MD simulation with a large cooling rate (typically 10 K/ps), (ii) MD simulation with a slow cooling rate (typically 10−3 K/ps), and (iii) experimental melt-quench (typically 10−12 K/ps) [Debenedetti and Stillinger 2001]. At high temperature, the three systems are at the equilibrium liquid state and, hence, exhibit the same potential energy U [Debenedetti and Stillinger 2001]. As temperature decreases, U decreases (the slope depending on the heat capacity of the liquid) and all three systems enter into the metastable supercooled liquid state following the same linear master curve [Debenedetti and Stillinger 2001]. From this point, the transition from the metastable supercooled liquid state to the out-of-equilibrium glassy state (i.e., at the glass fictive temperature Tf) occurs when the relaxation time of the system exceeds the observation time. Although, at fixed temperature, all the three systems exhibit the same relaxation time, the observation time is significantly lower in MD simulations, especially upon large cooling rate. As such, glasses simulated by MD using a large cooling rate enter the glassy state at higher temperature and remain stuck is high-energy states in the PEL (see Figures 9a and b) [Debenedetti and Stillinger 2001]. As a result, glasses simulated by MD strongly depend on the choice of the cooling rate and tend to be more stable (i.e., lower final energy U0, more ordered) upon decreasing cooling rates.
3.2.2. Gap between simulated versus experimental glasses
Figure 9c shows the evolution of the glass fictive temperature Tf and final potential energy U0 as a function of the cooling rate for both melt-quenching experiments and MD simulations—using the case of a (Na2O)30(SiO2)70 glass as an archetypical example [Li et al. 2017; Zhou et al. 2020]. As expected, slower cooling rates result in lower Tf and U0. However, due to the huge gap between the cooling rates that are accessible to experiments and simulations [Li et al. 2017], experimental values of Tf and U0 tend to be much lower than their simulation counterparts [Li et al. 2017; Zhou et al. 2020]. This difference in the order of the magnitude of the cooling rate is a critical limitation of MD simulations since it results in a systematic difference between experimental and simulated glasses [Li et al. 2017; Vollmayr et al. 1996b]. Note that different types of glasses may exhibit various dependence on the cooling rate, so that certain glasses are more sensitive to variations in the cooling rate than others [Liu et al. 2018]. It should also be noted that certain glass properties or structural features (e.g., medium-range order structure) are more sensitive to the cooling rate than others (e.g., short-range order structure) [Li et al. 2017].
3.3. Finite size effects
Due to their high computational cost, atomistic simulations are limited to fairly small systems (i.e., small number of atoms). Although PBC are typically used to mitigate any spurious effects arising from the presence of surfaces, the limited size of simulated glasses (from hundreds to millions of atoms) can affect the accuracy of the simulation [Du 2015]. Difficulties related to the system size include: (i) lack of statistical sampling if the number of atoms is too low [Berthier et al. 2012; Horbach et al. 1996], (ii) enhanced fluctuations in the thermodynamic properties of the simulated system (e.g., temperature or pressure) since the magnitude of fluctuations inversely scale with the square root of the number of atoms [Du 2019; Leach 2001], and (iii) systematic errors arising from the limited size of the simulation box (e.g., absence of large rings or extended collective vibrational modes) [Ganster et al. 2004; Nakano et al. 1994].
Figure 10 shows the computed neutron structure factor SN(Q) of a glassy silica system prepared by melt-quenching MD simulations while using three different system sizes N, namely, 648 (small system), 5184 (intermediate system), and 41,472 atoms (large system) [Nakano et al. 1994]. The computed structure factor SN(Q) is found to depend on the system size since, especially in the low-Q region (which captures the medium-range order structure of the glass)—before the structure factor eventually converges for large systems [Nakano et al. 1994]. This indicates that glass simulations relying on small simulation boxes (and, hence, small numbers of atoms) do not always properly capture the structure and properties of glasses [Du 2015; Ganster et al. 2004; Nakano et al. 1994]. In practice, one needs to select a system size that is large enough to mitigate finite size effects, but small enough to ensure a reasonable computational cost [Du 2015].
3.4. Limitations of reverse Monte Carlo (RMC) simulations
3.4.1. Need for structural data
Reverse Monte Carlo (RMC) simulations aim to construct atomistic structures that match one or several structural fingerprints provided by experiments (often, the PDF) [McGreevy 2001]. As such, RMC simulations offer an attractive alternative to MD simulations since such simulations can effectively bypass the melt-quench route to form a glass and, hence, are not affected by high cooling rate effects. However, RMC simulations largely rely on the availability (and accuracy) of experimental data. This limits the predictive power of RMC simulations since this approach does not make it possible to simulate in silico glasses that have not been experimentally synthesized and characterized yet. This also limits the range of conditions (temperature, pressure, etc.) to the ones that have already been experimentally explored.
3.4.2. Ill-defined nature of RMC simulations
RMC simulations adopt the MC algorithm to search for an atomic structure that exhibits a minimum in the cost function landscape (see upper panel of Figure 11a) [Allen and Tildesley 2017], wherein the cost function R𝜒 is defined as the magnitude of difference between the simulated structural fingerprints and their experimental references [Zhou et al. 2020]. This can be formulated as an optimization problem, wherein some degrees of freedom (i.e., the positions of the atoms) are adjusted so as to minimize a cost function. However, the optimization problem at the core of RMC simulations is intrinsically ill-defined. Indeed, although a given three-dimensional structure yields unique values for the fingerprints, a given fingerprint can be associated with a large number of different three-dimensional structures. That is, the structure–fingerprint relationship is not reversible. For instance, due to the complex, disordered nature of glasses [Bunde and Havlin 2012], very different structures can be associated with the same PDF (see Figure 11b) [Pandey et al. 2016b]. This is a consequence of the fact that the PDF is simply a one-dimensional signature of a glass structure (averaged over various types of elements) and, therefore, only offers a highly compressed representation of a three-dimensional structure—which does not contain enough information to robustly reconstruct the structure itself. In short, RMC simulations do not exhibit a single solution, which manifests itself by the fact that the cost function landscape exhibits various competitive minima [Pandey et al. 2015]. Since traditional RMC simulations do not embed any knowledge of the interatomic interactions, the structures generated by RMC may not be thermodynamically stable [Zhou et al. 2020] and the regions of the PEL that are explored in RMC simulations may be very different from those that are accessed during the melt-quenching of a glass (see bottom panel in Figure 11a) [Pandey et al. 2015].
Figure 11c shows the evolution of the optimization cost function R𝜒 and the potential U with respect to the number of MC search steps during an RMC simulation for a (Na2O)30(SiO2)70 glass [Zhou et al. 2020]. Note that the values of R𝜒 and U obtained for the same glass prepared by a melt-quenching MD simulation are also added as references [Zhou et al. 2020]. As expected, R𝜒 gradually decreases and eventually becomes lower than the MD reference, that is, the RMC simulation offers an optimal structure that indeed matches the experimental structural fingerprints (see upper panel of Figure 11c) [Zhou et al. 2020]. However, upon RMC refinement, the potential energy U is not monotonically decreasing but, rather, tends to increase in the late stages of the MC search. Eventually, the final glass structure exhibits a notably higher potential energy U than that of the MD-simulated glass reference (see bottom panel of Figure 11c) [Zhou et al. 2020]. These results exemplify that an apparent excellent agreement with experimental data (e.g., based on the PDF) does not always translate into a realistic glass structure [Pandey et al. 2015; Zhou et al. 2020]. The ill-defined nature of RMC can be partially mitigated by the introduction of an energy penalty term U in the MC cost function to favor structures featuring low potential energy—an approach that is typically referred to as hybrid RMC [Bousige et al. 2015; Jain et al. 2006; Opletal et al. 2002]. However, hybrid RMC requires trial-and-error tests or intuition to adjust the weights of R𝜒 and U in the cost function. In addition, hybrid RMC simulations are significantly more computationally expensive than traditional RMC simulations since the potential energy of the system U must be computed at each MC step [Zhou et al. 2020].
4. New developments in atomistic modeling of glasses
4.1. Machine-learned forcefields
4.1.1. Types of machine-learned forcefields
To address the limitations associated with empirical forcefields relying on a fixed analytical form, machine learning (ML) offers a promising pathway to develop new complex forcefields that rely on an ML model to map a given atomic configuration to its potential energy. The promise of machine-learned forcefields is to approach the accuracy of ab initio simulations with a computational burden that is more comparable to (although typically higher than) that of classical empirical forcefields. In detail, machine-learned forcefields adopt ML regression models to fit the system’s PEL as a function of the atom positions [Behler 2016; Mishin 2021; Mueller et al. 2020], without the need to rely on any physical or chemical intuition to define a functional format of the empirical forcefield [Du 2015; Huang and Kieffer 2015]. The ground-truth PEL datapoints are often provided by ab initio MD simulations [Boero et al. 2015]. Several types of regression models have been developed so far to interpolate the PEL information provided by ab initio simulations [Mishin 2021], i.e., to map the system’s atom positions {ri} to its potential energy U{ri} (see Figure 12a). This includes (i) physically informed neural networks (PINN) or PINN-type regression models, which use as inputs several structural features (e.g., translation- and rotation-invariant angular 3-body and radial 2-body order parameters [Bartók et al. 2013; Chmiela et al. 2018; Pun et al. 2019]) to predict the system’s potential energy U [Huan et al. 2017; Pun et al. 2019; Schütt et al. 2018], (ii) graph neural networks (GNN), which directly use as input the atom positions (modeled as a graph, wherein some nodes, the atoms, are connected by some edges, the chemical interactions) to predict U [Gilmer et al. 2017; Park et al. 2021], and (iii) Gaussian approximation potentials (GAP), which, in contrast to the neural network fitting used in PINN and GNN [Russell et al. 2010], adopt a nonparametric fitting technique (i.e., Gaussian process regression (GPR) [Rasmussen and Williams 2008]) to offer not only the PEL prediction but also the confidence interval of the prediction [Bartók et al. 2010; Bartók and Csányi 2015]. Importantly, this confidence interval can be used to train forcefields by active learning, that is, wherein new ground-truth ab initio simulations are dynamically conducted for structures associated with high confidence intervals—which are then added to the training set to retrain the GAP forcefield. Such active learning makes it possible to dynamically retrain the forcefield based on the very structures that are the most informative to improve its accuracy [Byggmästar et al. 2019; Caro 2019].
4.1.2. Accuracy of machine-learned forcefields
As compared to empirical forcefields based on fixed functional formats [Du 2019], machine-learned forcefields exhibit more flexibility to fit the PEL and, therefore, tend to offer more accurate predictions (when properly trained) [Mishin 2021]. Figure 12b illustrates, from the viewpoint of a PEL [Lacks 2001], the amorphous-to-crystalline transition of a silicon solid upon increasing pressure [McMillan et al. 2005; Pandey et al. 2011]. The reference PEL datapoints are provided by ab initio MD simulations and are fitted by both an analytical classical and a machine-learned forcefield, respectively [Deringer et al. 2021]. The evolution of the average atomic volume as a function of pressure predicted by both the classical and machine-learned forcefields are provided in Figure 12c [Deringer et al. 2021]. On the one hand, the classical forcefield based on a SW potential offers a good description of the close-to-equilibrium states (amorphous and crystalline states, corresponding to local minima in the PEL) [Stillinger and Weber 1985], but fails to properly predict the pressure-driven amorphous-to-crystalline transition [Deringer et al. 2021] as it overestimates the energy barrier between these two states [Debenedetti and Stillinger 2001]. On the other hand, the GAP machine-learned forcefield accurately fits the reference PEL datapoints and offers an accurate prediction of the energy of the transition state upon increasing pressure [Deringer et al. 2021]. This illustrates the ability of machine-learned forcefields to offer accurate predictions, both for equilibrium and out-of-equilibrium systems [Mishin 2021]. It should nevertheless be noted that, like any ML model, the accuracy of a machine-learned forcefield largely depends on the size, accuracy, and diversity of the training set—that is, the ensemble of configurations (and their ground-truth potential energy) that are used to train the forcefield.
As an emerging family of forcefields [Chmiela et al. 2018], machine-learned forcefields have been successfully applied to simulate select glasses or disordered solids [Friederich et al. 2021], including amorphous Si [Deringer et al. 2021], silica [Li and Ando 2018], Ge–Sb–Te (GST) phase-change materials [Mocanu et al. 2020, 2018], MOFs [Eckhoff and Behler 2019], etc. On the one hand, these forcefields yield very accurate predictions, whose accuracy approaches that of ab initio MD [Chmiela et al. 2018]. On the other hand, in contrast to the high computation cost of ab initio MD [Hafner 2008], the MD simulations relying on machine-learned forcefields are more efficient, which unlocks the prediction of more complex properties—e.g., properties involving long-term timescales or extended length scales [Friederich et al. 2021], such as thermal conductivity [Sosso et al. 2018], glass reactivity [Erlebach et al. 2021], and phase transformations [Deringer et al. 2021]. Clearly, machine-learned forcefields are an extremely promising pathway to address the challenges facing the simulation of glasses, but their challenging parameterization remains a key bottleneck.
4.2. Development of analytical empirical forcefields by machine learning
4.2.1. Advantages of analytical empirical forcefields
Despite the fact that, thanks to their flexibility, machine-learned forcefields have the potential to offer very high accuracy [Mishin 2021], analytical empirical forcefields exhibit several key advantages [Du 2015; Huang and Kieffer 2015]. As an important feature, the functional forms are selected based on physical and chemical knowledge/intuition regarding the nature of the interatomic interactions [van Beest et al. 1990]. For instance, ion–ion interactions are modeled based on the Coulomb’s law, forcefields can explicitly account for Van der Waals interactions (with a term proportional to 1∕r6), or the dipole of an atom can explicitly be modeled as a spring in polarizable forcefields [Allen and Tildesley 2017; Carré et al. 2016; Du 2015; Fennell and Gezelter 2006; Huang and Kieffer 2015; Musgraves et al. 2019]. This tends to render empirical forcefields more physics-based, but also more interpretable than their “black-box” machine-learned counterparts [Cranmer et al. 2020; Hernandez et al. 2019]. Such fixed analytical form also has several advantages. First, it imposes a strong constraint on the predicted potential energy—for instance, the selected analytical form can ensure that the energy diverges to infinity at close distance and converges to zero at long distance [Senftle et al. 2016]. This is important as, in contrast, due to their large flexibility, machine-learned forcefields can sometimes offer unrealistic, nonphysical predictions. This is a consequence of the fact, like any ML model, machine-learned forcefields can excel at interpolating the regions of the PEL that they have been trained on, but tend to offer very unrealistic predictions far from their training set. In turn, the fixed analytical form of analytical forcefields limit their ability to accurately interpolate complex PEL, but makes them more robust and more likely to offer reasonable transferability to systems that are different from those included in the training set [Liu et al. 2019c]. Second, machine-learned forcefields typically rely on complex models (e.g., artificial neural networks) that present a very large number of tunable parameters. Such a large dimensionality must be balanced by a large training set. As a result, machine-learned forcefields typically require a significantly larger training set (that is, a larger number of configurations with their ground-truth potential energy) than analytical empirical forcefields (which typically involve orders of magnitude fewer parameters) [Liu et al. 2020a]. In addition, although machine-learned forcefields have the flexibility to interpolate any PEL without requiring any intuition regarding the nature of the interactions, they still rely on a certain number of assumptions regarding the complexity of the PEL (e.g., to determine the optimal number of neurons or layers in an artificial neural network model). Finally, although machine-learned forcefields are orders of magnitude faster than ab initio simulations, they still typically involve an increased computational burden as compared to simple analytical forcefields [Cranmer et al. 2020; Yaseen et al. 2016].
4.2.2. Machine-learning-aided parameterization scheme
As an alternative to machine-learned forcefields, machine-learning-aided parameterization can combine the power of ML with the robustness of analytical forcefields. Indeed, as compared to traditional gradient-descent-based parameterization schemes (see Figure 8) [Carré et al. 2016], ML offers a new paradigm to parameterize analytical forcefields in a more efficient and unbiased fashion [Liu et al. 2019e]. For instance, Figure 13 shows an ML-aided parameterization scheme based on Gaussian process regression (GPR) and Bayesian optimization (BO) [Frazier and Wang 2016; Ueno et al. 2016], which is applied to parameterize a Buckingham forcefield (see (5)) for glassy silica [Liu et al. 2019e]. The ML approach aims to find the global minimum (i.e., the optimal set of parameters) in the cost function landscape as a function of the forcefield parameters [Liu et al. 2019c], where the cost function R𝜒 is here defined as the magnitude of difference of the PDF g(r) computed by classical MD simulation and its ab initio reference [Wright 1993].
Figure 13a illustrates the basic idea behind the use of GPR and BO to parameterize empirical forcefields [Liu et al. 2019c]. For illustration purposes, only one forcefield parameter (i.e., silicon charge qsi) is here selected to plot the cost function landscape [Liu et al. 2019c], while the other parameters are fixed to their values in the well-established BKS potential [van Beest et al. 1990]. First, the GPR model fits the known (i.e., ground-truth) points in the cost function landscape—so as to offer not only a smooth interpolation of the landscape profile but also its confidence interval (see upper panel of Figure 13a) [Liu et al. 2019c]. Then, BO prescribes the set of parameters (i.e., qsi herein) that exhibits the highest probability to yield the global minimum for R𝜒 (see bottom panel of Figure 13a) [Liu et al. 2019c]. This is achieved by computing the expected improvement (EI) metric that offers the best balance between (i) the “exploitation” of the forcefields parameters that are predicted by the GPR model to yield a minimum R𝜒 value and (ii) the “exploration” of the domains of forcefield parameter values that are associated with high confidence interval (i.e., high uncertainty) [Frazier and Wang 2016]. This parameterization scheme ensures that the optimal forcefield parameters (i.e., those that minimize the cost function) can quickly be identified, while mitigating the risk for the optimization scheme to remain stuck in a local minimum of the cost function landscape.
4.2.3. Efficient search of global minimum
Figure 13b shows the search path that is prescribed by ML to find the optimal forcefield parameters (i.e., those yielding a global minimum for the cost function R𝜒) [Liu et al. 2019e]. For illustration purposes, the search is conducted within a two-dimensional parameter space (i.e., comprising the silicon partial charge qSi and the short-range SiO interaction intensity ASiO) [Liu et al. 2019e], while the other forcefield parameters are fixed to their values in the BKS potential [van Beest et al. 1990]. Notably, this ML-aided parameterization approach can quickly identify the optimal set of forcefield parameters after only a few iterations [Liu et al. 2019e]. Importantly, this optimization scheme is unbiased to the start position in the search space, that is, the outcome of the optimization does not depend on the initial values of the forcefield parameters [Liu et al. 2019e]. This approach offers a promising route to efficiently refine existing empirical forcefields at a fraction of the computational cost that is needed to develop a machine-learned forcefield [Liu et al. 2020a, 2019c, e].
4.3. Constructing stable glass structures by force-enhanced atomic refinement
4.3.1. Bypassing the melt-quenching route
Due to the huge gap of cooling rate between melt-quenching experiments and MD simulations (see Figure 9) [Li et al. 2017], glasses prepared by MD simulations tend to exhibit higher fictive temperature (i.e., to be less stable) than experimentally synthesized glasses [Vollmayr et al. 1996b]. Alternatively, RMC simulations can bypass this melt-quenching route by directly constructing atomic structures that match experimental data (see Section 2.2) [McGreevy 2001]. However, due to the ill-defined nature of RMC (see Section 3.4), RMC-based glasses are often thermodynamically unstable (see Figure 11) [Pandey et al. 2015; Zhou et al. 2020]. To mitigate this issue, force-enhanced atomic refinement (FEAR) has been proposed as a powerful, information-based approach that simultaneously leverages the knowledge of experimental data (which are used to constraint the glass structure) and interatomic potential (which is used to ensure the thermodynamic stability of the simulated glass) [Pandey et al. 2016a, b, 2015].
4.3.2. FEAR algorithm
Figure 14a illustrates how the FEAR algorithm searches for a realistic atomistic structure that simultaneously matches experimental data and features minimum potential energy [Pandey et al. 2016a, 2015]. In detail, each FEAR iteration consists of a given number (e.g., 5000) of RMC search steps in the cost function landscape R𝜒 (i.e., difference between simulated and experimental data), followed by one energy minimization step in the PEL [Zhou et al. 2020]. The cost function R𝜒 is defined herein as the magnitude of the difference between the PDFs g(r) computed by FEAR simulation and neutron diffraction experiment [Zhou et al. 2020]. RMC and energy minimalization steps are then literately repeated until both the cost function R𝜒 and the potential energy converge. In this scheme, the RMC steps are used to slightly deform the atomic structure (so as to escape local minima in the PEL), while the energy minimization is used to ensure that the simulated structure never drifts away from the stable region of the PEL that is physically accessible [Pandey et al. 2015; Zhou et al. 2020]. Eventually, by simultaneously leveraging both experimental and interatomic energy information, FEAR simulations can yield glass structures that simultaneously exhibit enhanced agreement with available experimental data and a larger thermodynamic stability than glasses prepared by MD or RMC [Pandey et al. 2015]. It is worth pointing out that, in contrast to MD simulations that heavily rely on the accuracy of interatomic forcefields [Takada 2021], the accuracy of FEAR simulation is fairly insensitive to the quality of interatomic forcefield that is used during the energy minimization step [Zhou et al. 2021]. That is, both high- and low-quality forcefields can guide the system toward energy-stable regions in the PEL, while, eventually, the details of the simulated structure are mainly controlled by the RMC steps [Zhou et al. 2021]. Importantly, FEAR simulations are extremely computationally efficient since the potential energy of the system only needs to be computed during the energy minimization step—whereas it needs to be computed at any step during MD or hybrid RMC simulations.
4.3.3. Realistic prediction of atomistic structure
Figure 14b shows the evolution of the cost function R𝜒 as a function of the number of optimization steps for both RMC and FEAR simulations [Zhou et al. 2020]. The R𝜒 values of glasses prepared by melt-quenching MD simulations at different cooling rate are added as references [Zhou et al. 2020]. Both FEAR and RMC eventually offer atomistic structures exhibiting small R𝜒 values, which are notably lower than that of the MD references [Zhou et al. 2020]. Indeed, the PDFs g(r) provided by both FEAR and RMC simulations are in excellent agreement with the experimental g(r) (see Figure 14c) [Zhou et al. 2020], while MD simulation slightly deviates from the experimental g(r) due to the large cooling rate [Li et al. 2017]. Interestingly, the energy minimizations steps upon FEAR refinement yield a faster convergence (i.e., faster decrease in the cost function R𝜒) as compared to RMC. Further, Figure 14d shows the evolution of the potential energy U as a function of the number of optimization steps both upon FEAR and RMC simulations [Zhou et al. 2020]. MD simulation results are also added as references [Zhou et al. 2020]. It is notable that RMC simulation yields unrealistic atomic structures that are unstable as they present a potential energy that is significantly higher than those achieved by MD [Zhou et al. 2020]. In comparison, the FEAR simulation offers a notably more realistic, energy-stable atomistic structure that still satisfies the experimental constraints as well as the RMC-based structure [Zhou et al. 2020]. Importantly, the FEAR-generated structure features a potential energy that is notably lower than those achieved by melt-quenching MD, even in the case of the slowest cooling rate [Zhou et al. 2020]. This highlights the fact that the structure generated by FEAR exhibits a lower fictive temperature than the MD-based configurations—that is, FEAR can efficiently bypass limitations arising from the fast cooling rate used in MD simulations [Debenedetti and Stillinger 2001; Li et al. 2017]. This establishes FEAR as a promising technique to generate glass structures that are very stable and that can directly be quantitatively compared with experimental glasses.
5. Conclusions and future opportunities
Overall, atomistic simulations provide an efficient tool to access the atomic structure of glasses, which is otherwise challenging to directly visualize by conventional experiments. However, simulations remain plagued by a series of challenges, including (i) slow execution runtime, (ii) small accessible timescales and/or length scales, (iii) cooling rate difference between experiments and simulations, (iv) availability and accuracy of empirical forcefields, and (v) ill-defined nature of simulation protocols (e.g., RMC simulations). To overcome these limitations, one can resort to either (i) simulation protocols involving more physical information (e.g., FEAR simulations, or more accurate empirical forcefield functionals) or (ii) ML approaches (e.g., machine-learned forcefields, or ML-aided forcefield parameterization). Indeed, with the aid of ML, one can develop more accurate and computationally efficient forcefields to speed up MD simulations, thereby unlocking longer timescales or larger system sizes.
As a future opportunity, we envision that smart closed-loop integrations of ML modeling and MD simulations will leapfrog glass modeling (see Figure 15). Example of such opportunities are listed in the following:
(i) Deciphering complex simulation data by ML. Atomic trajectories generated by MD simulations contains all the structural information that govern glass properties. However, due to the complex, disordered nature of glass structures, it is challenging to “separate the wheat from the chaff”, that is, to pinpoint the key structural features that govern glass properties [Tanaka et al. 2019]. In that regard, owing to its ability to discover hidden pattern in complex, multi-dimensional data, ML offers a new opportunity to identify relevant structural patterns in simulated glassy structures [Biroli 2020; Tanaka et al. 2019]. Prominent examples include the recently developed softness approach based on classification-based ML [Cubuk et al. 2017; Liu et al. 2021c, b] and graph neural networks (GNN)—a new and powerful family of ML models—that represent atomistic structures as nodes-and-edges graphs [Bapst et al. 2020]. Inputting complex simulation data into such advanced ML models makes it possible to decipher previously hidden structure–property relationships and capture the key structural features that govern glass properties. The ability to learn new physics from ML models depends on their “interpretability”. For instance, one can analyze the weight associated to each input structural feature in the softness approach [Cubuk et al. 2017; Liu et al. 2021c, b] or monitor the model responses for various graph inputs in the GNN method [Bapst et al. 2020]. These approaches make it possible to unravel the key structural patterns, that is, the ones that are the most influential in governing the properties of glasses. Finally, it is worth mentioning that, in addition to ML, topological analyzes represent an alternative, traditionally challenging but direct approach to investigate simulated glassy structures [Bauchy 2019]. For instance, some recently developed topological approaches (e.g., persistent homology [Nakamura et al. 2015]) have been demonstrated to be effective tools in identifying influential medium-range order structural features in disordered phases [Sørensen et al. 2020].
(ii) Combining simulations and ML for glasses’ inverse design. Owing to its economical nature as compared to systematic experiments, high-throughput virtual screening (HTVS, that is, the systematic simulation of a large number of glasses) offers an efficient route to identify in silico an optimal glass composition featuring an optimal characteristics within a given compositional domain [Noh et al. 2020; Sanchez-Lengeling and Aspuru-Guzik 2018]. However, although simulations excel at predicting the properties of a given glass as a function of its composition (i.e., “forward prediction”), their application to “inverse design” problems (that is, given an optimal property target, find the best glass composition) remains limited by their high computational cost—which prevents the systematic exploration of large compositional spaces [Sanchez-Lengeling and Aspuru-Guzik 2018]. To address this issue, ML offers an ideal companion to MD simulations—since an ML model can learn from a series of MD simulations and, based on this, recommend what should be the next glass composition to simulate. Such closed-loop integrations of MD and ML could greatly accelerate the discovery of novel glasses featuring desirable properties or functionalities [Liu et al. 2020b, 2019g; Noh et al. 2020]. Note that, when applicable, it is desirable to integrate experimental data into the closed loop of MD and ML, wherein such “ground-truth” experimental data can either inform or validate the models (see Figure 15). This is essential to ensure the reliability of glass modeling approaches so as to avoid any “bubble effect” arising from the error accumulation within the closed-loop integration of MD and ML (wherein MD and ML would mutually comfort and amplify each other in their improper predictions) [Bagnoli et al. 2022; Yang et al. 2019].
(iii) Leveraging differentiable programing platforms. When integrating simulations and ML models within unified pipelines, different programing languages can present a communication barrier between ML and simulation packages, which often rely on Python and C++/Fortran, respectively. In addition, most MD packages are still rooted in fairly ancient computing paradigms (e.g., with no automated differentiation), which is reminiscent of the state of ML before automatic hardware acceleration and differentiation became popular. To overcome these frictions, automatic differentiable (auto-diff) programing platforms (e.g., Python JAX [Bradbury et al. 2018]) have been recently developed to seamlessly integrate ML and simulations within unified pipelines. In contrast to traditional programing platforms that rely on handwritten derivatives (e.g., C++ [Plimpton 1995b]), auto-diff platforms excel at computing on-the-fly the backward gradient of any quantities with no additional computation burden associated with differentiation [Bradbury et al. 2018]—an operation that comes with a notable computing time in traditional simulators (e.g., force calculation in MD simulations [Allen and Tildesley 2017], see Figure 4a). Moreover, simulations built on auto-diff platforms gain backward differentiability, which makes it possible to use their outcomes to directly train ML models using gradient back propagation [Schoenholz and Cubuk 2020]. This create new opportunities to train a ML model directly based on differentiable physical knowledge rather than on data [Liu et al. 2020b]. Finally, the auto-diff platforms generally enable native “just-in-time” compilation on high-performance dedicated hardware accelerators [Bradbury et al. 2018; Schoenholz and Cubuk 2020], such as graphics processing units (GPU) and tensor processing units (TPU) [Wang et al. 2019]. Specifically, TPUs are specifically designed as matrix processors and, thanks to their tailored architecture, offer unparalleled performances in deep learning problems (up to 200X faster than GPUs). Note that MD simulations (e.g., JAX-MD [Schoenholz and Cubuk 2020]) can natively run on GPU- or TPU-based auto-diff platforms, which is key to avoid any speed bottleneck arising due to moving from one hardware to another within simulation pipelines [Plimpton 1995b; Schoenholz and Cubuk 2020]. This could greatly accelerate MD simulations relying on artificial neural networks potentials [Liu et al. 2020b; Yang et al. 2019].
(iv) Replacing slow simulations by faster surrogate ML simulation engines. Although the development of auto-diff platforms enables differentiable simulations and native hardware acceleration, the computational efficiency of numerical simulations is still limited by the intrinsic computation cost associated with the underlying numerical algorithms (e.g., Newton’s law of motion in MD simulations [Allen and Tildesley 2017]). The numerical algorithms behind scientific numerical simulations are likely to remain their bottleneck [Lane 2015; Yang et al. 2019]. To mitigate this issue, surrogate ML simulation engines offer a unique, largely untapped opportunity to replace slow simulations so as to accelerate their execution without compromising accuracy [Kochkov et al. 2021; Liu et al. 2021a]. Surrogate ML engines can be implemented as artificial neural network (ANN) models, such as convolutional neural network (CNN) [Kochkov et al. 2021] or GNN [Bapst et al. 2020].
Overall, we envision that the “fusion” of experiments, simulations, and ML models (see Figure 15) will unlock a new era in glass modeling (and materials modeling in general)—wherein traditional boundaries between physics and empirical models, knowledge and data, forward and inverse predictions, or experimental and simulation data would eventually fade. We hope that the present perspective will modestly contribute to stimulating new developments in that direction.
Conflicts of interest
Authors have no conflict of interest to declare.
Acknowledgments
This work was supported by the National Science Foundation under Grants No. CMMI-1762292, CMMI-1826420, DMR-1928538, DMREF-1922167, and CAREER-1944510. Part of this work was also supported by the Department of Energy’s Nuclear Energy University Program (DOE-NEUP: DE-NE18-15020).